Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
13,979
10,569,163,098
IssuesEvent
2019-10-06 17:46:44
visualworks/portaljal.com.br
https://api.github.com/repos/visualworks/portaljal.com.br
opened
Retry failed network request
bug infrastructure thirdparty ux
When a network request fails either to GlobalBus or to the user's base, there's no option to retrigger unless doing a browser refresh, therefore retriggering all the request. We want to retrigger only the failed refresh.
1.0
Retry failed network request - When a network request fails either to GlobalBus or to the user's base, there's no option to retrigger unless doing a browser refresh, therefore retriggering all the request. We want to retrigger only the failed refresh.
infrastructure
retry failed network request when a network request fails either to globalbus or to the user s base there s no option to retrigger unless doing a browser refresh therefore retriggering all the request we want to retrigger only the failed refresh
1
11,269
9,040,170,751
IssuesEvent
2019-02-10 14:09:29
coq/coq
https://api.github.com/repos/coq/coq
opened
dune and coqtop.byte
kind: bug kind: infrastructure
<!-- Thank you for reporting a bug to Coq! --> #### Description of the problem <!-- If you can, it's helpful to provide self-contained example of some code that reproduces the bug. If not, a link to a larger example is also helpful. --> I was trying to use coqtop.byte built with dune to debug something in equations and failed. First issue is that just running coqtop.byte failed, one has to ``` export LD_LIBRARY_PATH=PATH_TO_COQ/_build/install/default//lib/stublibs:$LD_LIBRARY_PATH ``` For this to run, otherwise the dynamic library linker fails to find a DLL. After this, processing works, and I can drop to the toplevel, however `#use "include";;` fails, and after adding the PATH_TO_COQ/dev directory to the source files it succeeds but cannot find the object files it needs. Adding #directory "PATH_TO_COQ/_build/default/dev" it still fails with: ``` File "/Users/sozeau/dev/coq/coq/_build/default/dev/top_printers.ml", line 13, characters 5-10: Error: Unbound module Sorts Hint: Did you mean Sort? File "/Users/sozeau/dev/coq/coq/_build/default/dev/vm_printers.ml", line 2, characters 5-9: Error: Unbound module Term Unbound value ppid. ... ```
1.0
dune and coqtop.byte - <!-- Thank you for reporting a bug to Coq! --> #### Description of the problem <!-- If you can, it's helpful to provide self-contained example of some code that reproduces the bug. If not, a link to a larger example is also helpful. --> I was trying to use coqtop.byte built with dune to debug something in equations and failed. First issue is that just running coqtop.byte failed, one has to ``` export LD_LIBRARY_PATH=PATH_TO_COQ/_build/install/default//lib/stublibs:$LD_LIBRARY_PATH ``` For this to run, otherwise the dynamic library linker fails to find a DLL. After this, processing works, and I can drop to the toplevel, however `#use "include";;` fails, and after adding the PATH_TO_COQ/dev directory to the source files it succeeds but cannot find the object files it needs. Adding #directory "PATH_TO_COQ/_build/default/dev" it still fails with: ``` File "/Users/sozeau/dev/coq/coq/_build/default/dev/top_printers.ml", line 13, characters 5-10: Error: Unbound module Sorts Hint: Did you mean Sort? File "/Users/sozeau/dev/coq/coq/_build/default/dev/vm_printers.ml", line 2, characters 5-9: Error: Unbound module Term Unbound value ppid. ... ```
infrastructure
dune and coqtop byte description of the problem if you can it s helpful to provide self contained example of some code that reproduces the bug if not a link to a larger example is also helpful i was trying to use coqtop byte built with dune to debug something in equations and failed first issue is that just running coqtop byte failed one has to export ld library path path to coq build install default lib stublibs ld library path for this to run otherwise the dynamic library linker fails to find a dll after this processing works and i can drop to the toplevel however use include fails and after adding the path to coq dev directory to the source files it succeeds but cannot find the object files it needs adding directory path to coq build default dev it still fails with file users sozeau dev coq coq build default dev top printers ml line characters error unbound module sorts hint did you mean sort file users sozeau dev coq coq build default dev vm printers ml line characters error unbound module term unbound value ppid
1
92,574
26,719,440,960
IssuesEvent
2023-01-29 00:03:11
envoyproxy/envoy
https://api.github.com/repos/envoyproxy/envoy
closed
build/mobile: RBE fails with arm64 versions of Bazel on macOS
area/build stale mobile
*Title*: build/mobile: RBE fails with arm64 versions of Bazel on macOS *Description*: When building something in the `mobile/` workspace with an arm64 bazel binary on an M1 Mac with RBE enabled, the `protoc` with the wrong architecture is attempted to be run, leading to [this error](https://envoy.cluster.engflow.com/invocation/30cc6fcc-8c30-46b0-b616-48af66cec8a9#console): > sandbox-exec: execvp() of 'bazel-out/.../com_google_protobuf/protoc' failed: Bad CPU type in executable *Repro steps*: On an M1 Mac with EngFlow certs installed at `/path/to/certs/engflow.crt` and `/path/to/certs/engflow.key`: ``` git clone https://github.com/envoyproxy/envoy.git cd envoy/mobile ./bazelw build \ --tls_client_certificate=/path/to/certs/engflow.crt \ --tls_client_key=/path/to/certs/engflow.key \ --config remote-ci-macos \ //library/common:envoy_main_interface_lib ``` The build succeeds if this line gets hardcoded to `amd64`: https://github.com/envoyproxy/envoy/blob/367763e5c63c003e495981e98fd9967bbcb86d20/mobile/bazelw#L16 *Admin and Stats Output*: N/A *Config*: N/A *Logs*: N/A *Call Stack*: N/A
1.0
build/mobile: RBE fails with arm64 versions of Bazel on macOS - *Title*: build/mobile: RBE fails with arm64 versions of Bazel on macOS *Description*: When building something in the `mobile/` workspace with an arm64 bazel binary on an M1 Mac with RBE enabled, the `protoc` with the wrong architecture is attempted to be run, leading to [this error](https://envoy.cluster.engflow.com/invocation/30cc6fcc-8c30-46b0-b616-48af66cec8a9#console): > sandbox-exec: execvp() of 'bazel-out/.../com_google_protobuf/protoc' failed: Bad CPU type in executable *Repro steps*: On an M1 Mac with EngFlow certs installed at `/path/to/certs/engflow.crt` and `/path/to/certs/engflow.key`: ``` git clone https://github.com/envoyproxy/envoy.git cd envoy/mobile ./bazelw build \ --tls_client_certificate=/path/to/certs/engflow.crt \ --tls_client_key=/path/to/certs/engflow.key \ --config remote-ci-macos \ //library/common:envoy_main_interface_lib ``` The build succeeds if this line gets hardcoded to `amd64`: https://github.com/envoyproxy/envoy/blob/367763e5c63c003e495981e98fd9967bbcb86d20/mobile/bazelw#L16 *Admin and Stats Output*: N/A *Config*: N/A *Logs*: N/A *Call Stack*: N/A
non_infrastructure
build mobile rbe fails with versions of bazel on macos title build mobile rbe fails with versions of bazel on macos description when building something in the mobile workspace with an bazel binary on an mac with rbe enabled the protoc with the wrong architecture is attempted to be run leading to sandbox exec execvp of bazel out com google protobuf protoc failed bad cpu type in executable repro steps on an mac with engflow certs installed at path to certs engflow crt and path to certs engflow key git clone cd envoy mobile bazelw build tls client certificate path to certs engflow crt tls client key path to certs engflow key config remote ci macos library common envoy main interface lib the build succeeds if this line gets hardcoded to admin and stats output n a config n a logs n a call stack n a
0
20,924
14,239,812,627
IssuesEvent
2020-11-18 20:44:07
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
Cannot browse source code in Visual Studio
area-infrastructure
### Describe the bug When opening the source code from this repository in Visual Studio, there are two issues. First, the packages cannot be restored fully which is probably the reason why all code files are mostly underlined in red. Second, I repeatedly get modal Microsoft login dialogs that want me to login somewhere. When I do log in, it has no effect at all. Visual Studio is completely unusable with these dialogs popping up all the time! ### To Reproduce 1. Check out the source code from this repository. 2. Switch to the tag v3.1.6 3. Change global.json to match the installed .NET SDK version 4. Open src\Mvc\Mvc.sln in Visual Studio 2019 Community Edition 16.6.4 ### Exceptions (if any) This is the package manager output: ``` Alle Pakete sind bereits installiert. Nichts muss wiederhergestellt werden. Verstrichene Zeit: 00:01:46.2440102 ========== Abgeschlossen ========== Fehler beim Wiederherstellen von NuGet-Paketen: Der Dienstindex für die Quelle "https://pkgs.dev.azure.com/dnceng/_packaging/darc-int-dotnet-extensions-be18161f/nuget/v3/index.json" konnte nicht geladen werden. ``` ### Further technical details dotnet --info: ``` .NET Core SDK (gemäß "global.json"): Version: 3.1.302 Commit: 41faccf259 Laufzeitumgebung: OS Name: Windows OS Version: 10.0.19041 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\3.1.302\ Host (useful for support): Version: 3.1.6 Commit: 3acd9b0cd1 .NET Core SDKs installed: 3.1.302 [C:\Program Files\dotnet\sdk] .NET Core runtimes installed: Microsoft.AspNetCore.All 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] To install additional .NET Core runtimes or SDKs: https://aka.ms/dotnet-download ```
1.0
Cannot browse source code in Visual Studio - ### Describe the bug When opening the source code from this repository in Visual Studio, there are two issues. First, the packages cannot be restored fully which is probably the reason why all code files are mostly underlined in red. Second, I repeatedly get modal Microsoft login dialogs that want me to login somewhere. When I do log in, it has no effect at all. Visual Studio is completely unusable with these dialogs popping up all the time! ### To Reproduce 1. Check out the source code from this repository. 2. Switch to the tag v3.1.6 3. Change global.json to match the installed .NET SDK version 4. Open src\Mvc\Mvc.sln in Visual Studio 2019 Community Edition 16.6.4 ### Exceptions (if any) This is the package manager output: ``` Alle Pakete sind bereits installiert. Nichts muss wiederhergestellt werden. Verstrichene Zeit: 00:01:46.2440102 ========== Abgeschlossen ========== Fehler beim Wiederherstellen von NuGet-Paketen: Der Dienstindex für die Quelle "https://pkgs.dev.azure.com/dnceng/_packaging/darc-int-dotnet-extensions-be18161f/nuget/v3/index.json" konnte nicht geladen werden. ``` ### Further technical details dotnet --info: ``` .NET Core SDK (gemäß "global.json"): Version: 3.1.302 Commit: 41faccf259 Laufzeitumgebung: OS Name: Windows OS Version: 10.0.19041 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\3.1.302\ Host (useful for support): Version: 3.1.6 Commit: 3acd9b0cd1 .NET Core SDKs installed: 3.1.302 [C:\Program Files\dotnet\sdk] .NET Core runtimes installed: Microsoft.AspNetCore.All 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 2.1.20 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.2.8 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.1.6 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] To install additional .NET Core runtimes or SDKs: https://aka.ms/dotnet-download ```
infrastructure
cannot browse source code in visual studio describe the bug when opening the source code from this repository in visual studio there are two issues first the packages cannot be restored fully which is probably the reason why all code files are mostly underlined in red second i repeatedly get modal microsoft login dialogs that want me to login somewhere when i do log in it has no effect at all visual studio is completely unusable with these dialogs popping up all the time to reproduce check out the source code from this repository switch to the tag change global json to match the installed net sdk version open src mvc mvc sln in visual studio community edition exceptions if any this is the package manager output alle pakete sind bereits installiert nichts muss wiederhergestellt werden verstrichene zeit abgeschlossen fehler beim wiederherstellen von nuget paketen der dienstindex für die quelle konnte nicht geladen werden further technical details dotnet info net core sdk gemäß global json version commit laufzeitumgebung os name windows os version os platform windows rid base path c program files dotnet sdk host useful for support version commit net core sdks installed net core runtimes installed microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft windowsdesktop app to install additional net core runtimes or sdks
1
786,055
27,632,852,562
IssuesEvent
2023-03-10 12:16:26
TESTARtool/TESTAR_dev
https://api.github.com/repos/TESTARtool/TESTAR_dev
closed
Update OrientDB CI Actions URL
type: technical feature request priority: high status: not started
Seems that OrientDB migrated the hosting of its databases from: https://s3.us-east-2.amazonaws.com/orientdb3/releases/3.0.34 to: https://repo1.maven.org/maven2/com/orientechnologies/orientdb-community/3.0.34/ Change the related Gradle Tasks to fix the CI Actions
1.0
Update OrientDB CI Actions URL - Seems that OrientDB migrated the hosting of its databases from: https://s3.us-east-2.amazonaws.com/orientdb3/releases/3.0.34 to: https://repo1.maven.org/maven2/com/orientechnologies/orientdb-community/3.0.34/ Change the related Gradle Tasks to fix the CI Actions
non_infrastructure
update orientdb ci actions url seems that orientdb migrated the hosting of its databases from to change the related gradle tasks to fix the ci actions
0
297,071
25,597,536,646
IssuesEvent
2022-12-01 17:17:43
elastic/kibana
https://api.github.com/repos/elastic/kibana
reopened
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js - logstash pipeline create new cancel button discards the pipeline and redirects to the list
blocker Feature:Logstash Pipelines Team:Logstash failed-test skipped-test v7.11.0
A test failed on a tracked branch ``` Error: retry.try timeout: Error: expected [ { selected: false, id: 'cuhvoemuemat', description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.', lastModified: 'A few seconds ago', username: 'test_user' }, { selected: false, id: 'empty_pipeline_1', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_10', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_11', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_12', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_13', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_14', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_15', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_16', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_17', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_18', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_19', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_2', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_20', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_21', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_3', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_4', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_5', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_6', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_7', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' } ] to sort of equal [ { selected: false, id: 'cuhvoemuemat', description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.', lastModified: 'A minute ago', username: 'test_user' }, { selected: false, id: 'empty_pipeline_1', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_10', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_11', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_12', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_13', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_14', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_15', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_16', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_17', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_18', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_19', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_2', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_20', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_21', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_3', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_4', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_5', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_6', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_7', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' } ] at Assertion.assert (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:100:11) at Assertion.eql (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:244:8) at retry.try (test/functional/apps/logstash/pipeline_create.js:98:35) at process._tickCallback (internal/process/next_tick.js:68:7) at onFailure (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:28:9) at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:68:13) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/9586/) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js","test.name":"logstash pipeline create new cancel button discards the pipeline and redirects to the list","test.failCount":3}} -->
2.0
Failing test: Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js - logstash pipeline create new cancel button discards the pipeline and redirects to the list - A test failed on a tracked branch ``` Error: retry.try timeout: Error: expected [ { selected: false, id: 'cuhvoemuemat', description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.', lastModified: 'A few seconds ago', username: 'test_user' }, { selected: false, id: 'empty_pipeline_1', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_10', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_11', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_12', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_13', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_14', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_15', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_16', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_17', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_18', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_19', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_2', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_20', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_21', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_3', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_4', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_5', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_6', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_7', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' } ] to sort of equal [ { selected: false, id: 'cuhvoemuemat', description: 'Ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse.', lastModified: 'A minute ago', username: 'test_user' }, { selected: false, id: 'empty_pipeline_1', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_10', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_11', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_12', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_13', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_14', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_15', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_16', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_17', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_18', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_19', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_2', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_20', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_21', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_3', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_4', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_5', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_6', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' }, { selected: false, id: 'empty_pipeline_7', description: 'an empty pipeline', lastModified: '3 years ago', username: 'elastic' } ] at Assertion.assert (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:100:11) at Assertion.eql (/dev/shm/workspace/parallel/16/kibana/packages/kbn-expect/expect.js:244:8) at retry.try (test/functional/apps/logstash/pipeline_create.js:98:35) at process._tickCallback (internal/process/next_tick.js:68:7) at onFailure (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:28:9) at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:68:13) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/9586/) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests.x-pack/test/functional/apps/logstash/pipeline_create·js","test.name":"logstash pipeline create new cancel button discards the pipeline and redirects to the list","test.failCount":3}} -->
non_infrastructure
failing test chrome x pack ui functional tests x pack test functional apps logstash pipeline create·js logstash pipeline create new cancel button discards the pipeline and redirects to the list a test failed on a tracked branch error retry try timeout error expected selected false id cuhvoemuemat description ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse lastmodified a few seconds ago username test user selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic to sort of equal selected false id cuhvoemuemat description ohuji wirisaw gowporef ga ug hu ga fi parvucofu evu vocup vafse lastmodified a minute ago username test user selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic selected false id empty pipeline description an empty pipeline lastmodified years ago username elastic at assertion assert dev shm workspace parallel kibana packages kbn expect expect js at assertion eql dev shm workspace parallel kibana packages kbn expect expect js at retry try test functional apps logstash pipeline create js at process tickcallback internal process next tick js at onfailure dev shm workspace parallel kibana test common services retry retry for success ts at retryforsuccess dev shm workspace parallel kibana test common services retry retry for success ts first failure
0
9,087
7,807,788,640
IssuesEvent
2018-06-11 18:01:50
ampproject/docs
https://api.github.com/repos/ampproject/docs
closed
Build fails on "Error 502"
P0: Critical Type: Bug Type: Site Infrastructure
This error started today and I'm not sure what the problem is. The 3 other parallel jobs worked. ``` Warning: Don't know how to categorize amp-subscriptions-google.md, skipping.. /home/travis/build/ampproject/docs/scripts/import_docs.js:32 throw err; ^ Error: Error 502 at ChildProcess.exithandler (child_process.js:275:12) at emitTwo (events.js:126:13) at ChildProcess.emit (events.js:214:7) at maybeClose (internal/child_process.js:925:16) at Socket.stream.socket.on (internal/child_process.js:346:11) at emitOne (events.js:116:13) at Socket.emit (events.js:211:7) at Pipe._handle.close [as _onclose] (net.js:557:12) [17:14:48] 'build' errored after 9.51 s [17:14:48] The following tasks did not complete: optimize-images, sass [17:14:48] Did you forget to signal async completion? ``` Log: https://travis-ci.org/ampproject/docs/jobs/390843966 I've restarted the job several times and it either just fails on the script or produces a 502 error with warnings about categorization --- which shouldn't occur. I'm thinking it's flakiness but not sure. Can't reproduce locally.
1.0
Build fails on "Error 502" - This error started today and I'm not sure what the problem is. The 3 other parallel jobs worked. ``` Warning: Don't know how to categorize amp-subscriptions-google.md, skipping.. /home/travis/build/ampproject/docs/scripts/import_docs.js:32 throw err; ^ Error: Error 502 at ChildProcess.exithandler (child_process.js:275:12) at emitTwo (events.js:126:13) at ChildProcess.emit (events.js:214:7) at maybeClose (internal/child_process.js:925:16) at Socket.stream.socket.on (internal/child_process.js:346:11) at emitOne (events.js:116:13) at Socket.emit (events.js:211:7) at Pipe._handle.close [as _onclose] (net.js:557:12) [17:14:48] 'build' errored after 9.51 s [17:14:48] The following tasks did not complete: optimize-images, sass [17:14:48] Did you forget to signal async completion? ``` Log: https://travis-ci.org/ampproject/docs/jobs/390843966 I've restarted the job several times and it either just fails on the script or produces a 502 error with warnings about categorization --- which shouldn't occur. I'm thinking it's flakiness but not sure. Can't reproduce locally.
infrastructure
build fails on error this error started today and i m not sure what the problem is the other parallel jobs worked warning don t know how to categorize amp subscriptions google md skipping home travis build ampproject docs scripts import docs js throw err error error at childprocess exithandler child process js at emittwo events js at childprocess emit events js at maybeclose internal child process js at socket stream socket on internal child process js at emitone events js at socket emit events js at pipe handle close net js build errored after s the following tasks did not complete optimize images sass did you forget to signal async completion log i ve restarted the job several times and it either just fails on the script or produces a error with warnings about categorization which shouldn t occur i m thinking it s flakiness but not sure can t reproduce locally
1
26,967
20,969,088,494
IssuesEvent
2022-03-28 09:38:53
tempus-finance/tempus-app
https://api.github.com/repos/tempus-finance/tempus-app
opened
Setup project base for v3
enhancement infrastructure
that may include but not limited to: - start with CRA - pick tech stack, e.g. `@hookstate`, i18n libraries, wallet dependencies, etc - set up `react-router` - folder strcuture - anything you can think of
1.0
Setup project base for v3 - that may include but not limited to: - start with CRA - pick tech stack, e.g. `@hookstate`, i18n libraries, wallet dependencies, etc - set up `react-router` - folder strcuture - anything you can think of
infrastructure
setup project base for that may include but not limited to start with cra pick tech stack e g hookstate libraries wallet dependencies etc set up react router folder strcuture anything you can think of
1
21,473
14,589,107,822
IssuesEvent
2020-12-19 00:32:15
KaTeX/KaTeX
https://api.github.com/repos/KaTeX/KaTeX
closed
Safari screenshotter
bug infrastructure
The Safari screenshotter is broken. I have three PRs that currently are shown as having errors purely due to Safari screenshotter disfunction. Even when the Safari screenshotter works, the process is clumsy and tedious. I think we should get rid of it.
1.0
Safari screenshotter - The Safari screenshotter is broken. I have three PRs that currently are shown as having errors purely due to Safari screenshotter disfunction. Even when the Safari screenshotter works, the process is clumsy and tedious. I think we should get rid of it.
infrastructure
safari screenshotter the safari screenshotter is broken i have three prs that currently are shown as having errors purely due to safari screenshotter disfunction even when the safari screenshotter works the process is clumsy and tedious i think we should get rid of it
1
3,425
4,311,296,346
IssuesEvent
2016-07-21 22:09:41
dart-lang/www.dartlang.org
https://api.github.com/repos/dart-lang/www.dartlang.org
closed
Highlighted TOC section doesn't quite match the currently visible section
bug Infrastructure P2 PatchesWelcome
From https://code.google.com/p/dart/issues/detail?id=21272: > Hello team :) > > https://www.dartlang.org/tools/editor/ > > Your topic selection list on the left does not work correctly. > > All but the first and last selections do not change their own font to bold when selected. > > I believe this happens when at certain scroll points. > > I am using Google Chrome. > > Kindest Regards, > > Daniel Couillard
1.0
Highlighted TOC section doesn't quite match the currently visible section - From https://code.google.com/p/dart/issues/detail?id=21272: > Hello team :) > > https://www.dartlang.org/tools/editor/ > > Your topic selection list on the left does not work correctly. > > All but the first and last selections do not change their own font to bold when selected. > > I believe this happens when at certain scroll points. > > I am using Google Chrome. > > Kindest Regards, > > Daniel Couillard
infrastructure
highlighted toc section doesn t quite match the currently visible section from hello team your topic selection list on the left does not work correctly all but the first and last selections do not change their own font to bold when selected i believe this happens when at certain scroll points i am using google chrome kindest regards daniel couillard
1
35,124
30,774,216,654
IssuesEvent
2023-07-31 04:26:37
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
closed
Code Analyzer: Use of wp-env is problematic
status: prioritization tool: monorepo infrastructure
<!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. --> **Prerequisites (mark completed items with an [x]):** - [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose - [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is). **Issue Description:** When running Code Analyzer we run into a lot of issues. ``` Command failed: wp-env start ``` wp-env is responsible for calling two functions vi WP CLI https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/src/Internal/DataStores/Orders/OrdersTableDataStore.php#L2321 https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/bin/wc-get-schema.php#L12-L20 Calling `wp-env start` programatically appears to cause problems, especially with Docker. We need to either: 1. Find a safer way of handling wp-env 2. Find a replacement for wp-env 3. Determine a different way of detecting schema changes, perhaps through static analysis
1.0
Code Analyzer: Use of wp-env is problematic - <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. --> **Prerequisites (mark completed items with an [x]):** - [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose - [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is). **Issue Description:** When running Code Analyzer we run into a lot of issues. ``` Command failed: wp-env start ``` wp-env is responsible for calling two functions vi WP CLI https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/src/Internal/DataStores/Orders/OrdersTableDataStore.php#L2321 https://github.com/woocommerce/woocommerce/blob/3828aabbef631b8e6ce1778ef8a1cee80e60b83f/plugins/woocommerce/bin/wc-get-schema.php#L12-L20 Calling `wp-env start` programatically appears to cause problems, especially with Docker. We need to either: 1. Find a safer way of handling wp-env 2. Find a replacement for wp-env 3. Determine a different way of detecting schema changes, perhaps through static analysis
infrastructure
code analyzer use of wp env is problematic prerequisites mark completed items with an i have checked that my issue type is not listed here my issue is not a security issue support request bug report enhancement or feature request please use the link above if it is issue description when running code analyzer we run into a lot of issues command failed wp env start wp env is responsible for calling two functions vi wp cli calling wp env start programatically appears to cause problems especially with docker we need to either find a safer way of handling wp env find a replacement for wp env determine a different way of detecting schema changes perhaps through static analysis
1
24,799
17,786,452,842
IssuesEvent
2021-08-31 11:40:57
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Serve built static files in production
type: bug work: infrastructure status: ready
## Problem <!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.--> Currently, the frontend is only set up for local development and depends on the Vite server running at port 3000 to serve the frontend. This will not work for production, since we'll have pre-built static files. ## Proposed solution <!-- A clear and concise description of your proposed solution or feature. --> We need to update the Django templates to use the existing setup for development environments and serve pre-built static files for production environments (or if the vite server is not running). ## Additional context <!-- Add any other context or screenshots about the feature request here.--> This is required for https://github.com/centerofci/mathesar/issues/225
1.0
Serve built static files in production - ## Problem <!-- Please provide a clear and concise description of the problem that this feature request is designed to solve.--> Currently, the frontend is only set up for local development and depends on the Vite server running at port 3000 to serve the frontend. This will not work for production, since we'll have pre-built static files. ## Proposed solution <!-- A clear and concise description of your proposed solution or feature. --> We need to update the Django templates to use the existing setup for development environments and serve pre-built static files for production environments (or if the vite server is not running). ## Additional context <!-- Add any other context or screenshots about the feature request here.--> This is required for https://github.com/centerofci/mathesar/issues/225
infrastructure
serve built static files in production problem currently the frontend is only set up for local development and depends on the vite server running at port to serve the frontend this will not work for production since we ll have pre built static files proposed solution we need to update the django templates to use the existing setup for development environments and serve pre built static files for production environments or if the vite server is not running additional context this is required for
1
29,733
24,222,510,132
IssuesEvent
2022-09-26 12:04:45
arduino/arduino-ide
https://api.github.com/repos/arduino/arduino-ide
closed
Spurious IDE update notifications when using snapshot or nightly builds
topic: infrastructure type: imperfection
### Describe the problem On every startup, the Arduino IDE checks for new versions of the IDE (https://github.com/arduino/arduino-ide/pull/797). If a newer version is available, a notification/dialog is shown offering an update. "Newer" is determined by comparing the version of the user's IDE to the latest available version on the update channel. This comparison is done according to [semver](https://semver.org/). In order to facilitate beta testing, builds are generated of the Arduino IDE at the current stage in development. These builds are given an identifying version of the following form: - `<version>-snapshot-<short hash>` - builds generated for every push and pull request that modifies relevant files - `<version>-nightly-<YYYYMMDD>` - daily builds of the tip of the default branch The established project management practices cause the `<version>` component of these to be the version of the most recent release. During the pre-release phase of the project development, all releases had [a pre-release suffix](https://semver.org/#spec-item-9) (e.g., `2.0.0-rc9.4`). Appending the "snapshot" or "nightly" suffix to that pre-release version caused these builds to have the correct precedence (e.g., `2.0.0-rc9.2.snapshot-20cc34c` > `2.0.0-rc9.2`). This situation has changed now that the project is using production release versions (e.g., `2.0.0-nightly-20220915` < `2.0.0`) 🐛 Users of "snapshot" or "nightly" builds are presented with a spurious update notification on startup. ### To reproduce #### Setup Make sure you have the default values of the relevant settings: ```json { "arduino.checkForUpdates": true, "arduino.ide.updateChannel": "stable" } ``` #### Steps 1. Download the latest nightly build of Arduino IDE 2.x from the links on the "Software" page: https://www.arduino.cc/en/software#nightly-builds 1. Start the nightly build. 🐛 A spurious IDE update notification dialog appears: ![image](https://user-images.githubusercontent.com/8572152/190379778-28f8d1d1-4f92-4121-9f50-dbc057735cba.png) ### Expected behavior IDE update notification is only shown when a newer version of the Arduino IDE is available. ### Arduino IDE version 2.0.0-snapshot-5f82577 ### Operating system Windows ### Operating system version 10 ### Additional context I think the solution will be to add a step to the end of the [release procedure](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md) to bump the version metadata after completing a release, following the same procedure as documented under [the "Create the release on GitHub" section of that document](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md#-create-the-release-on-github). I see two possible approaches: - Use a pre-release of a patch bump (e.g., the metadata should currently be set to `2.0.1-rc.1`) - Use a patch bump (e.g., the metadata should currently be set to `2.0.1`) With the latter approach, the release manager could skip the metadata update step of the release unless the release was determined to be a minor or major version bump (according to the nature of the changes being shipped). ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds) - [X] My report contains all necessary details
1.0
Spurious IDE update notifications when using snapshot or nightly builds - ### Describe the problem On every startup, the Arduino IDE checks for new versions of the IDE (https://github.com/arduino/arduino-ide/pull/797). If a newer version is available, a notification/dialog is shown offering an update. "Newer" is determined by comparing the version of the user's IDE to the latest available version on the update channel. This comparison is done according to [semver](https://semver.org/). In order to facilitate beta testing, builds are generated of the Arduino IDE at the current stage in development. These builds are given an identifying version of the following form: - `<version>-snapshot-<short hash>` - builds generated for every push and pull request that modifies relevant files - `<version>-nightly-<YYYYMMDD>` - daily builds of the tip of the default branch The established project management practices cause the `<version>` component of these to be the version of the most recent release. During the pre-release phase of the project development, all releases had [a pre-release suffix](https://semver.org/#spec-item-9) (e.g., `2.0.0-rc9.4`). Appending the "snapshot" or "nightly" suffix to that pre-release version caused these builds to have the correct precedence (e.g., `2.0.0-rc9.2.snapshot-20cc34c` > `2.0.0-rc9.2`). This situation has changed now that the project is using production release versions (e.g., `2.0.0-nightly-20220915` < `2.0.0`) 🐛 Users of "snapshot" or "nightly" builds are presented with a spurious update notification on startup. ### To reproduce #### Setup Make sure you have the default values of the relevant settings: ```json { "arduino.checkForUpdates": true, "arduino.ide.updateChannel": "stable" } ``` #### Steps 1. Download the latest nightly build of Arduino IDE 2.x from the links on the "Software" page: https://www.arduino.cc/en/software#nightly-builds 1. Start the nightly build. 🐛 A spurious IDE update notification dialog appears: ![image](https://user-images.githubusercontent.com/8572152/190379778-28f8d1d1-4f92-4121-9f50-dbc057735cba.png) ### Expected behavior IDE update notification is only shown when a newer version of the Arduino IDE is available. ### Arduino IDE version 2.0.0-snapshot-5f82577 ### Operating system Windows ### Operating system version 10 ### Additional context I think the solution will be to add a step to the end of the [release procedure](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md) to bump the version metadata after completing a release, following the same procedure as documented under [the "Create the release on GitHub" section of that document](https://github.com/arduino/arduino-ide/blob/main/docs/internal/release-procedure.md#-create-the-release-on-github). I see two possible approaches: - Use a pre-release of a patch bump (e.g., the metadata should currently be set to `2.0.1-rc.1`) - Use a patch bump (e.g., the metadata should currently be set to `2.0.1`) With the latter approach, the release manager could skip the metadata update step of the release unless the release was determined to be a minor or major version bump (according to the nature of the changes being shipped). ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds) - [X] My report contains all necessary details
infrastructure
spurious ide update notifications when using snapshot or nightly builds describe the problem on every startup the arduino ide checks for new versions of the ide if a newer version is available a notification dialog is shown offering an update newer is determined by comparing the version of the user s ide to the latest available version on the update channel this comparison is done according to in order to facilitate beta testing builds are generated of the arduino ide at the current stage in development these builds are given an identifying version of the following form snapshot builds generated for every push and pull request that modifies relevant files nightly daily builds of the tip of the default branch the established project management practices cause the component of these to be the version of the most recent release during the pre release phase of the project development all releases had e g appending the snapshot or nightly suffix to that pre release version caused these builds to have the correct precedence e g snapshot this situation has changed now that the project is using production release versions e g nightly 🐛 users of snapshot or nightly builds are presented with a spurious update notification on startup to reproduce setup make sure you have the default values of the relevant settings json arduino checkforupdates true arduino ide updatechannel stable steps download the latest nightly build of arduino ide x from the links on the software page start the nightly build 🐛 a spurious ide update notification dialog appears expected behavior ide update notification is only shown when a newer version of the arduino ide is available arduino ide version snapshot operating system windows operating system version additional context i think the solution will be to add a step to the end of the to bump the version metadata after completing a release following the same procedure as documented under i see two possible approaches use a pre release of a patch bump e g the metadata should currently be set to rc use a patch bump e g the metadata should currently be set to with the latter approach the release manager could skip the metadata update step of the release unless the release was determined to be a minor or major version bump according to the nature of the changes being shipped issue checklist i searched for previous reports in i verified the problem still occurs when using the latest my report contains all necessary details
1
114,610
9,744,504,814
IssuesEvent
2019-06-03 07:10:40
ITISFoundation/osparc-simcore
https://api.github.com/repos/ITISFoundation/osparc-simcore
closed
ISAN Use Case UCD 1D, 2D: Movie Not Shown
ISAN Prio 1 Bug UX_improvements tester_review
## Long story short Enter CC 2D and open the cc-2d-viewer. Then retrieve and refresh -> the movie should properly show within the Jupyter notebook. -> The movie is not shown, does not work. ## Expected behaviour The movie in the Jupyter notebook should properly show. ## Actual behaviour Please see recording as follows: ![2019-05-10_20-49-34](https://user-images.githubusercontent.com/32800795/57550253-1eaca480-7366-11e9-8da5-f48feba4191e.gif) ## Steps to reproduce Please see above. ## Your environment ```json [ { "name": "osparc-simcore", "version": "cfc73c1", "url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/" }, { "name": "osparc-simcore UI ['clean']", "version": "cfc73c1", "url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/services/web/client/" }, { "name": "qooxdoo-compiler", "version": "0.3.0-alpha.20190423-1234", "url": "https://github.com/qooxdoo/qooxdoo-compiler" }, { "name": "qxapp", "version": "1.0.0", "url": "" }, { "name": "qooxdoo framework", "version": "6.0.0-alpha-20190409-f79cc78", "url": "http://qooxdoo.org" }, { "name": "qx-iconfont-fontawesome5", "version": "0.0.6", "url": "https://github.com/ITISFoundation/qx-iconfont-fontawesome5" }, { "name": "qx-iconfont-material", "version": "0.1.4", "url": "https://github.com/ITISFoundation/qx-osparc-theme" }, { "name": "qx-osparc-theme", "version": "0.4.6", "url": "https://github.com/ITISFoundation/qx-osparc-theme" }, { "name": "socket.io", "version": "2.1.2", "url": "https://github.com/socketio/socket.io" }, { "name": "jsondiffpatch", "version": "0.3.11", "url": "https://github.com/benjamine/jsondiffpatch" }, { "name": "jsonTreeViewer", "version": "0.6.0", "url": "https://github.com/summerstyle/jsonTreeViewer" }, { "name": "Ajv", "version": "6.5.0", "url": "https://github.com/epoberezkin/ajv" }, { "name": "plotly", "version": "1.43.2", "url": "https://github.com/plotly/plotly.js" }, { "name": "gridster", "version": "0.7.0", "url": "https://github.com/dsmorse/gridster.js" }, { "name": "svg.js", "version": "2.6.4", "url": "https://github.com/svgdotjs/svg.js" } ] ```
1.0
ISAN Use Case UCD 1D, 2D: Movie Not Shown - ## Long story short Enter CC 2D and open the cc-2d-viewer. Then retrieve and refresh -> the movie should properly show within the Jupyter notebook. -> The movie is not shown, does not work. ## Expected behaviour The movie in the Jupyter notebook should properly show. ## Actual behaviour Please see recording as follows: ![2019-05-10_20-49-34](https://user-images.githubusercontent.com/32800795/57550253-1eaca480-7366-11e9-8da5-f48feba4191e.gif) ## Steps to reproduce Please see above. ## Your environment ```json [ { "name": "osparc-simcore", "version": "cfc73c1", "url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/" }, { "name": "osparc-simcore UI ['clean']", "version": "cfc73c1", "url": "https://github.com/ITISFoundation/osparc-simcore/tree/cfc73c1/services/web/client/" }, { "name": "qooxdoo-compiler", "version": "0.3.0-alpha.20190423-1234", "url": "https://github.com/qooxdoo/qooxdoo-compiler" }, { "name": "qxapp", "version": "1.0.0", "url": "" }, { "name": "qooxdoo framework", "version": "6.0.0-alpha-20190409-f79cc78", "url": "http://qooxdoo.org" }, { "name": "qx-iconfont-fontawesome5", "version": "0.0.6", "url": "https://github.com/ITISFoundation/qx-iconfont-fontawesome5" }, { "name": "qx-iconfont-material", "version": "0.1.4", "url": "https://github.com/ITISFoundation/qx-osparc-theme" }, { "name": "qx-osparc-theme", "version": "0.4.6", "url": "https://github.com/ITISFoundation/qx-osparc-theme" }, { "name": "socket.io", "version": "2.1.2", "url": "https://github.com/socketio/socket.io" }, { "name": "jsondiffpatch", "version": "0.3.11", "url": "https://github.com/benjamine/jsondiffpatch" }, { "name": "jsonTreeViewer", "version": "0.6.0", "url": "https://github.com/summerstyle/jsonTreeViewer" }, { "name": "Ajv", "version": "6.5.0", "url": "https://github.com/epoberezkin/ajv" }, { "name": "plotly", "version": "1.43.2", "url": "https://github.com/plotly/plotly.js" }, { "name": "gridster", "version": "0.7.0", "url": "https://github.com/dsmorse/gridster.js" }, { "name": "svg.js", "version": "2.6.4", "url": "https://github.com/svgdotjs/svg.js" } ] ```
non_infrastructure
isan use case ucd movie not shown long story short enter cc and open the cc viewer then retrieve and refresh the movie should properly show within the jupyter notebook the movie is not shown does not work expected behaviour the movie in the jupyter notebook should properly show actual behaviour please see recording as follows steps to reproduce please see above your environment json name osparc simcore version url name osparc simcore ui version url name qooxdoo compiler version alpha url name qxapp version url name qooxdoo framework version alpha url name qx iconfont version url name qx iconfont material version url name qx osparc theme version url name socket io version url name jsondiffpatch version url name jsontreeviewer version url name ajv version url name plotly version url name gridster version url name svg js version url
0
124,861
4,934,574,169
IssuesEvent
2016-11-28 19:28:29
devinit/D-Portal
https://api.github.com/repos/devinit/D-Portal
closed
Update homepage text - Search, and notes
priority
1 - Add text at top introducing d-portal and new Search function 2 - Update text at the bottom of the page
1.0
Update homepage text - Search, and notes - 1 - Add text at top introducing d-portal and new Search function 2 - Update text at the bottom of the page
non_infrastructure
update homepage text search and notes add text at top introducing d portal and new search function update text at the bottom of the page
0
15,309
11,455,432,015
IssuesEvent
2020-02-06 19:05:20
enarx/enarx
https://api.github.com/repos/enarx/enarx
closed
Ensure all files have up-to-date Apache license headers
infrastructure
With the new year and all the new project files that come with it, we should ensure everything contains proper license headers.
1.0
Ensure all files have up-to-date Apache license headers - With the new year and all the new project files that come with it, we should ensure everything contains proper license headers.
infrastructure
ensure all files have up to date apache license headers with the new year and all the new project files that come with it we should ensure everything contains proper license headers
1
468,467
13,483,450,860
IssuesEvent
2020-09-11 03:57:02
Apexal/late
https://api.github.com/repos/Apexal/late
closed
Calendar sometimes duplicates work blocks
Priority: High Type: Bug
When manually adding a work block on the calendar, it sometimes displays two matching blocks after requesting one to be added (ex. two "Homework 3" blocks for the same class within the same time period). It doesn't go away- it has to be manually deleted. To duplicate: 1. Add work block to calendar. 2. See if duplicate assignment appears. Device: Windows PC using Google Chrome
1.0
Calendar sometimes duplicates work blocks - When manually adding a work block on the calendar, it sometimes displays two matching blocks after requesting one to be added (ex. two "Homework 3" blocks for the same class within the same time period). It doesn't go away- it has to be manually deleted. To duplicate: 1. Add work block to calendar. 2. See if duplicate assignment appears. Device: Windows PC using Google Chrome
non_infrastructure
calendar sometimes duplicates work blocks when manually adding a work block on the calendar it sometimes displays two matching blocks after requesting one to be added ex two homework blocks for the same class within the same time period it doesn t go away it has to be manually deleted to duplicate add work block to calendar see if duplicate assignment appears device windows pc using google chrome
0
71,248
15,190,323,461
IssuesEvent
2021-02-15 17:44:59
gigantum/gigantum-client
https://api.github.com/repos/gigantum/gigantum-client
opened
Add support for user defined HTTPS certs
enhancement feedback:request security team:backend
## Description With the improved authentication workflow and future Team Server improvements, it should be easy to run a multi-tenant Client as an externally hosted service with proper TLS certificates. An initial implementation could be to simply provide a cert and key via the config file that is then installed into the webserver automatically for the user. Future improvements could automatically run Let's Encrypt certbot.
True
Add support for user defined HTTPS certs - ## Description With the improved authentication workflow and future Team Server improvements, it should be easy to run a multi-tenant Client as an externally hosted service with proper TLS certificates. An initial implementation could be to simply provide a cert and key via the config file that is then installed into the webserver automatically for the user. Future improvements could automatically run Let's Encrypt certbot.
non_infrastructure
add support for user defined https certs description with the improved authentication workflow and future team server improvements it should be easy to run a multi tenant client as an externally hosted service with proper tls certificates an initial implementation could be to simply provide a cert and key via the config file that is then installed into the webserver automatically for the user future improvements could automatically run let s encrypt certbot
0
252,748
21,628,987,244
IssuesEvent
2022-05-05 07:40:56
storj/gateway-mt
https://api.github.com/repos/storj/gateway-mt
closed
Custom metadata is not saved when S3 MPU is used
bug s3-tests blocked
Custom metadata is not saved when S3 MPU is used (in this example the threshold is 8MB) As you can see in the second case the metadata is empty, but with real AWS it's foo: bar in both cases. ``` > aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 1mb s3://bucket1/1mb upload: ./1mb to s3://bucket1/1mb > aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 20mb s3://bucket1/20mb upload: ./20mb to s3://bucket1/20mb > aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 1mb x { "AcceptRanges": "bytes", "LastModified": "Thu, 30 Sep 2021 11:36:10 GMT", "ContentLength": 1048576, "ETag": "\"f5f16eee8ad5a86a3439dc75429831ee\"", "ContentType": "application/octet-stream", "Metadata": { "foo": "bar" } } > aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 20mb x { "AcceptRanges": "bytes", "LastModified": "Thu, 30 Sep 2021 11:36:21 GMT", "ContentLength": 20971520, "ETag": "\"80f096bdc4c7fd9639307cd1174b398a-3\"", "ContentType": "application/octet-stream", "Metadata": {} } ``` Credit: The problem is identified by the rclone s3 unit tests during the conversation in rclone/rclone#5616 where @olefrost suggested to re-run s3 unit tests with storj gateway-mt and @ncw drew attention to this problem from many failures.
1.0
Custom metadata is not saved when S3 MPU is used - Custom metadata is not saved when S3 MPU is used (in this example the threshold is 8MB) As you can see in the second case the metadata is empty, but with real AWS it's foo: bar in both cases. ``` > aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 1mb s3://bucket1/1mb upload: ./1mb to s3://bucket1/1mb > aws s3 cp --endpoint http://192.168.80.4:7777 --metadata foo=bar 20mb s3://bucket1/20mb upload: ./20mb to s3://bucket1/20mb > aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 1mb x { "AcceptRanges": "bytes", "LastModified": "Thu, 30 Sep 2021 11:36:10 GMT", "ContentLength": 1048576, "ETag": "\"f5f16eee8ad5a86a3439dc75429831ee\"", "ContentType": "application/octet-stream", "Metadata": { "foo": "bar" } } > aws s3api get-object --endpoint http://192.168.80.4:7777 --bucket bucket1 --key 20mb x { "AcceptRanges": "bytes", "LastModified": "Thu, 30 Sep 2021 11:36:21 GMT", "ContentLength": 20971520, "ETag": "\"80f096bdc4c7fd9639307cd1174b398a-3\"", "ContentType": "application/octet-stream", "Metadata": {} } ``` Credit: The problem is identified by the rclone s3 unit tests during the conversation in rclone/rclone#5616 where @olefrost suggested to re-run s3 unit tests with storj gateway-mt and @ncw drew attention to this problem from many failures.
non_infrastructure
custom metadata is not saved when mpu is used custom metadata is not saved when mpu is used in this example the threshold is as you can see in the second case the metadata is empty but with real aws it s foo bar in both cases aws cp endpoint metadata foo bar upload to aws cp endpoint metadata foo bar upload to aws get object endpoint bucket key x acceptranges bytes lastmodified thu sep gmt contentlength etag contenttype application octet stream metadata foo bar aws get object endpoint bucket key x acceptranges bytes lastmodified thu sep gmt contentlength etag contenttype application octet stream metadata credit the problem is identified by the rclone unit tests during the conversation in rclone rclone where olefrost suggested to re run unit tests with storj gateway mt and ncw drew attention to this problem from many failures
0
12,313
9,692,739,967
IssuesEvent
2019-05-24 14:28:21
aspnet/AspNetCore
https://api.github.com/repos/aspnet/AspNetCore
closed
Symbols from daily builds of master brancn are not published to symbol server
2 - Working area-infrastructure question
## Summary The C# extension test team recently did a test of symbol server support using a preview6 build of the framework and found that one assembly, Microsoft.AspNetCore.Routing.Abstractions, had a PDB file, but the PDB wasn't published to symbol server. ## Environment data `dotnet --info` output: .NET Core SDK (reflecting any global.json): Version: 3.0.100-preview6-011799 Commit: 3816213dde Runtime Environment: OS Name: Mac OS X OS Version: 10.14 OS Platform: Darwin RID: osx.10.14-x64 Base Path: /usr/local/share/dotnet/sdk/3.0.100-preview6-011799/ Host (useful for support): Version: 3.0.0-preview6-27707-07 Commit: c9e7f134dc .NET Core SDKs installed: 3.0.100-preview6-011799 [/usr/local/share/dotnet/sdk] .NET Core runtimes installed: Microsoft.AspNetCore.App 3.0.0-preview6-19257-04 [/usr/local/share/dotnet/shared/Microsoft.AspNetCore.App] Microsoft.NETCore.App 3.0.0-preview6-27707-07 [/usr/local/share/dotnet/shared/Microsoft.NETCore.App] To install additional .NET Core runtimes or SDKs: https://aka.ms/dotnet-download VS Code version: 1.33.1 C# Extension version: 1.19 (v1.19.1) ## Steps to reproduce 1. Create a new ASP.NET Razor project by doing mkdir razor cd razor dotnet new razor 2. Open razor with VScode 3. It may ask at the bottom right to install assets to build and debug, say yes. 4. After it has done this it will have created a .vscode directory in the ‘razor’ directory and a ‘launch.json’ file in that directory. Open the ‘.vscode’ directory, and double click on the ‘launch.json to edit it. 5. Add the following lines after the “name” line of the “Configurations” object. (This allows debugging through the framework. "justMyCode": false, "symbolOptions": { "searchMicrosoftSymbolServer": true }, "suppressJITOptimizations": true, 6. Double click on the ‘Startup.cs’ file in the ‘Explorer’ pane to bring up that source code. 7. Add the following line to the Startup.cs file at the beginning of the ‘Configure’ method (at the bottom of the file). Console.WriteLine("x " + 3); 8. Place a breakpoint on that line (F9). 9. Run the application until it hit that breakpoint (F5). ## Expected behavior It will load symbols from the Microsoft symbol server successfully. ## Actual behavior One assembly, Microsoft.AspNetCore.Routing.Abstractions, didn't have symbols.
1.0
Symbols from daily builds of master brancn are not published to symbol server - ## Summary The C# extension test team recently did a test of symbol server support using a preview6 build of the framework and found that one assembly, Microsoft.AspNetCore.Routing.Abstractions, had a PDB file, but the PDB wasn't published to symbol server. ## Environment data `dotnet --info` output: .NET Core SDK (reflecting any global.json): Version: 3.0.100-preview6-011799 Commit: 3816213dde Runtime Environment: OS Name: Mac OS X OS Version: 10.14 OS Platform: Darwin RID: osx.10.14-x64 Base Path: /usr/local/share/dotnet/sdk/3.0.100-preview6-011799/ Host (useful for support): Version: 3.0.0-preview6-27707-07 Commit: c9e7f134dc .NET Core SDKs installed: 3.0.100-preview6-011799 [/usr/local/share/dotnet/sdk] .NET Core runtimes installed: Microsoft.AspNetCore.App 3.0.0-preview6-19257-04 [/usr/local/share/dotnet/shared/Microsoft.AspNetCore.App] Microsoft.NETCore.App 3.0.0-preview6-27707-07 [/usr/local/share/dotnet/shared/Microsoft.NETCore.App] To install additional .NET Core runtimes or SDKs: https://aka.ms/dotnet-download VS Code version: 1.33.1 C# Extension version: 1.19 (v1.19.1) ## Steps to reproduce 1. Create a new ASP.NET Razor project by doing mkdir razor cd razor dotnet new razor 2. Open razor with VScode 3. It may ask at the bottom right to install assets to build and debug, say yes. 4. After it has done this it will have created a .vscode directory in the ‘razor’ directory and a ‘launch.json’ file in that directory. Open the ‘.vscode’ directory, and double click on the ‘launch.json to edit it. 5. Add the following lines after the “name” line of the “Configurations” object. (This allows debugging through the framework. "justMyCode": false, "symbolOptions": { "searchMicrosoftSymbolServer": true }, "suppressJITOptimizations": true, 6. Double click on the ‘Startup.cs’ file in the ‘Explorer’ pane to bring up that source code. 7. Add the following line to the Startup.cs file at the beginning of the ‘Configure’ method (at the bottom of the file). Console.WriteLine("x " + 3); 8. Place a breakpoint on that line (F9). 9. Run the application until it hit that breakpoint (F5). ## Expected behavior It will load symbols from the Microsoft symbol server successfully. ## Actual behavior One assembly, Microsoft.AspNetCore.Routing.Abstractions, didn't have symbols.
infrastructure
symbols from daily builds of master brancn are not published to symbol server summary the c extension test team recently did a test of symbol server support using a build of the framework and found that one assembly microsoft aspnetcore routing abstractions had a pdb file but the pdb wasn t published to symbol server environment data dotnet info output net core sdk reflecting any global json version commit runtime environment os name mac os x os version os platform darwin rid osx base path usr local share dotnet sdk host useful for support version commit net core sdks installed net core runtimes installed microsoft aspnetcore app microsoft netcore app to install additional net core runtimes or sdks vs code version c extension version steps to reproduce create a new asp net razor project by doing mkdir razor cd razor dotnet new razor open razor with vscode it may ask at the bottom right to install assets to build and debug say yes after it has done this it will have created a vscode directory in the ‘razor’ directory and a ‘launch json’ file in that directory open the ‘ vscode’ directory and double click on the ‘launch json to edit it add the following lines after the “name” line of the “configurations” object this allows debugging through the framework justmycode false symboloptions searchmicrosoftsymbolserver true suppressjitoptimizations true double click on the ‘startup cs’ file in the ‘explorer’ pane to bring up that source code add the following line to the startup cs file at the beginning of the ‘configure’ method at the bottom of the file console writeline x place a breakpoint on that line run the application until it hit that breakpoint expected behavior it will load symbols from the microsoft symbol server successfully actual behavior one assembly microsoft aspnetcore routing abstractions didn t have symbols
1
107,014
16,743,776,674
IssuesEvent
2021-06-11 13:12:33
Polkadex-Substrate/polkadexTEE-worker
https://api.github.com/repos/Polkadex-Substrate/polkadexTEE-worker
opened
Exchange websocket client with rustls client
security
Exchange websocket client with rustls client as soon as openfinex supports tls
True
Exchange websocket client with rustls client - Exchange websocket client with rustls client as soon as openfinex supports tls
non_infrastructure
exchange websocket client with rustls client exchange websocket client with rustls client as soon as openfinex supports tls
0
26,489
20,156,851,176
IssuesEvent
2022-02-09 17:13:49
SasView/sasview
https://api.github.com/repos/SasView/sasview
opened
Decide what to do about the SourceForge and danse source code archives and the archive repo
Admin and support infrastructure task
_This issue has been written as an aide memoir for @butlerpd and @smk78!_ A long-standing issue has been to sunset the old SANSView/SasView SourceForge site. A pre-requisite for this is ensuring important content is securely archived elsewhere, and GitHub would be the logical place. A similar, but arguably less urgent, argument applies to legacy content on the danse server. The SasView\archive repo was created for this purpose, and is functional. However, a problem has arisen. It transpires that GitHub will not track files >~50Mb, or >~25Mb if using the browser interface. For such files, GitHub recommends doing one of two things: - either, adding these files using [Git Large File Support (Git-LFS)](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-git-large-file-storage), - or, by adding these files as releases. For simplicity, @smk78 migrated the legacy .gz, .tgz, and .zip distribution archives to \archive using Git-LFS. This has worked well, except the volume of data has slightly exceeded (1.06Gb) the free 1Gb LFS limit that GitHub makes available per organization. There is no immediate threat to the content in \archive (see [here](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-storage-and-bandwidth-usage)) and the repo can still be cloned, but no additional content _anywhere_ in the SasView organization can now be added by LFS **(adding by normal commits is unaffected!)**. To add further content by LFS would require removing existing content, or purchasing a monthly Git-LFS 'data pack' (at the time of writing $5 gets 50Gb). Alternatively, we need to figure out another way to archive our legacy content.
1.0
Decide what to do about the SourceForge and danse source code archives and the archive repo - _This issue has been written as an aide memoir for @butlerpd and @smk78!_ A long-standing issue has been to sunset the old SANSView/SasView SourceForge site. A pre-requisite for this is ensuring important content is securely archived elsewhere, and GitHub would be the logical place. A similar, but arguably less urgent, argument applies to legacy content on the danse server. The SasView\archive repo was created for this purpose, and is functional. However, a problem has arisen. It transpires that GitHub will not track files >~50Mb, or >~25Mb if using the browser interface. For such files, GitHub recommends doing one of two things: - either, adding these files using [Git Large File Support (Git-LFS)](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-git-large-file-storage), - or, by adding these files as releases. For simplicity, @smk78 migrated the legacy .gz, .tgz, and .zip distribution archives to \archive using Git-LFS. This has worked well, except the volume of data has slightly exceeded (1.06Gb) the free 1Gb LFS limit that GitHub makes available per organization. There is no immediate threat to the content in \archive (see [here](https://docs.github.com/en/repositories/working-with-files/managing-large-files/about-storage-and-bandwidth-usage)) and the repo can still be cloned, but no additional content _anywhere_ in the SasView organization can now be added by LFS **(adding by normal commits is unaffected!)**. To add further content by LFS would require removing existing content, or purchasing a monthly Git-LFS 'data pack' (at the time of writing $5 gets 50Gb). Alternatively, we need to figure out another way to archive our legacy content.
infrastructure
decide what to do about the sourceforge and danse source code archives and the archive repo this issue has been written as an aide memoir for butlerpd and a long standing issue has been to sunset the old sansview sasview sourceforge site a pre requisite for this is ensuring important content is securely archived elsewhere and github would be the logical place a similar but arguably less urgent argument applies to legacy content on the danse server the sasview archive repo was created for this purpose and is functional however a problem has arisen it transpires that github will not track files or if using the browser interface for such files github recommends doing one of two things either adding these files using or by adding these files as releases for simplicity migrated the legacy gz tgz and zip distribution archives to archive using git lfs this has worked well except the volume of data has slightly exceeded the free lfs limit that github makes available per organization there is no immediate threat to the content in archive see and the repo can still be cloned but no additional content anywhere in the sasview organization can now be added by lfs adding by normal commits is unaffected to add further content by lfs would require removing existing content or purchasing a monthly git lfs data pack at the time of writing gets alternatively we need to figure out another way to archive our legacy content
1
24,398
17,196,122,245
IssuesEvent
2021-07-16 17:39:26
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
opened
Network layer changes for hooking up the backend for PlatformParameters
Priority: Essential Type: Task Where: Infrastructure
PlatformParameter API & Gae Models, also a MockPlatformParameterSerivice for fake response. PlatformParameterService GaePlatformParameters GaePlatformParameter MockPlatformParameterService Target PR date: 17 July 2021 Target completion date: 24 July 2021
1.0
Network layer changes for hooking up the backend for PlatformParameters - PlatformParameter API & Gae Models, also a MockPlatformParameterSerivice for fake response. PlatformParameterService GaePlatformParameters GaePlatformParameter MockPlatformParameterService Target PR date: 17 July 2021 Target completion date: 24 July 2021
infrastructure
network layer changes for hooking up the backend for platformparameters platformparameter api gae models also a mockplatformparameterserivice for fake response platformparameterservice gaeplatformparameters gaeplatformparameter mockplatformparameterservice target pr date july target completion date july
1
466,893
13,436,452,446
IssuesEvent
2020-09-07 14:24:13
scikit-image/scikit-image
https://api.github.com/repos/scikit-image/scikit-image
closed
Memory corruption in flood_fill
action: solution identified good first issue priority: critical type: bug
## Description When using `flood_fill`, there is a consistent double free error and a dying kernel (as shown in the output from Jupyter) when using a negative row coordinate for the point where the fill starts. ## Way to reproduce ```python import numpy as np from skimage import segmentation a = np.random.randint(0, 2, (100,100)) #for r in range(a.shape[0]): # segmentation.flood_fill(a, (r, 0), 1, in_place=True) # segmentation.flood_fill(a, (r, a.shape[1]-1), 1, in_place=True) # segmentation.flood_fill(a, (r, -1), 1, in_place=True) for c in range(a.shape[1]): # segmentation.flood_fill(a, ( 0, c), 1, in_place=True) # segmentation.flood_fill(a, (a.shape[0]-1, c), 1, in_place=True) segmentation.flood_fill(a, (-1, c), 1, in_place=True) ``` All commented lines work, except the last one, which consistently triggers the error. The `flood` function also fails likewise with `segmentation.flood(a, (-1, c))`. ## Version information ```python # Paste the output of the following python commands from __future__ import print_function import sys; print(sys.version) import platform; print(platform.platform()) import skimage; print("scikit-image version: {}".format(skimage.__version__)) import numpy; print("numpy version: {}".format(numpy.__version__)) ``` ```python 3.7.7 (default, May 7 2020, 21:25:33) [GCC 7.3.0] Linux-5.4.0-42-generic-x86_64-with-debian-bullseye-sid scikit-image version: 0.17.2 numpy version: 1.19.1 ``` This error was present in version 0.16.2 too.
1.0
Memory corruption in flood_fill - ## Description When using `flood_fill`, there is a consistent double free error and a dying kernel (as shown in the output from Jupyter) when using a negative row coordinate for the point where the fill starts. ## Way to reproduce ```python import numpy as np from skimage import segmentation a = np.random.randint(0, 2, (100,100)) #for r in range(a.shape[0]): # segmentation.flood_fill(a, (r, 0), 1, in_place=True) # segmentation.flood_fill(a, (r, a.shape[1]-1), 1, in_place=True) # segmentation.flood_fill(a, (r, -1), 1, in_place=True) for c in range(a.shape[1]): # segmentation.flood_fill(a, ( 0, c), 1, in_place=True) # segmentation.flood_fill(a, (a.shape[0]-1, c), 1, in_place=True) segmentation.flood_fill(a, (-1, c), 1, in_place=True) ``` All commented lines work, except the last one, which consistently triggers the error. The `flood` function also fails likewise with `segmentation.flood(a, (-1, c))`. ## Version information ```python # Paste the output of the following python commands from __future__ import print_function import sys; print(sys.version) import platform; print(platform.platform()) import skimage; print("scikit-image version: {}".format(skimage.__version__)) import numpy; print("numpy version: {}".format(numpy.__version__)) ``` ```python 3.7.7 (default, May 7 2020, 21:25:33) [GCC 7.3.0] Linux-5.4.0-42-generic-x86_64-with-debian-bullseye-sid scikit-image version: 0.17.2 numpy version: 1.19.1 ``` This error was present in version 0.16.2 too.
non_infrastructure
memory corruption in flood fill description when using flood fill there is a consistent double free error and a dying kernel as shown in the output from jupyter when using a negative row coordinate for the point where the fill starts way to reproduce python import numpy as np from skimage import segmentation a np random randint for r in range a shape segmentation flood fill a r in place true segmentation flood fill a r a shape in place true segmentation flood fill a r in place true for c in range a shape segmentation flood fill a c in place true segmentation flood fill a a shape c in place true segmentation flood fill a c in place true all commented lines work except the last one which consistently triggers the error the flood function also fails likewise with segmentation flood a c version information python paste the output of the following python commands from future import print function import sys print sys version import platform print platform platform import skimage print scikit image version format skimage version import numpy print numpy version format numpy version python default may linux generic with debian bullseye sid scikit image version numpy version this error was present in version too
0
61,620
17,023,741,078
IssuesEvent
2021-07-03 03:35:35
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Wrong parent region in Austria
Component: nominatim Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 11.36am, Friday, 19th August 2011]** When searching adresses in two administrative boundaries (Schrding and Ried) the parent region indicated is "Vcklabruck" which has type: place: region; This is wrong; all three are administrative entities (counties) on the same level with parent "Obersterreich: boundaries: administrative
1.0
Wrong parent region in Austria - **[Submitted to the original trac issue database at 11.36am, Friday, 19th August 2011]** When searching adresses in two administrative boundaries (Schrding and Ried) the parent region indicated is "Vcklabruck" which has type: place: region; This is wrong; all three are administrative entities (counties) on the same level with parent "Obersterreich: boundaries: administrative
non_infrastructure
wrong parent region in austria when searching adresses in two administrative boundaries schrding and ried the parent region indicated is vcklabruck which has type place region this is wrong all three are administrative entities counties on the same level with parent obersterreich boundaries administrative
0
309,210
26,656,594,690
IssuesEvent
2023-01-25 17:18:58
getsentry/sentry-javascript
https://api.github.com/repos/getsentry/sentry-javascript
closed
Add integration tests for NextJS dev server
Package: Nextjs Status: Backlog Type: Tests
We currently only run integration test for our nextjs sdk against production builds, but we should we should also run integration tests against the dev server for nextjs. We want to make sure our future changes don't break dev workflows. https://github.com/getsentry/sentry-javascript/issues/4115 was an issue that was not caught on release because it affected only dev server env.
1.0
Add integration tests for NextJS dev server - We currently only run integration test for our nextjs sdk against production builds, but we should we should also run integration tests against the dev server for nextjs. We want to make sure our future changes don't break dev workflows. https://github.com/getsentry/sentry-javascript/issues/4115 was an issue that was not caught on release because it affected only dev server env.
non_infrastructure
add integration tests for nextjs dev server we currently only run integration test for our nextjs sdk against production builds but we should we should also run integration tests against the dev server for nextjs we want to make sure our future changes don t break dev workflows was an issue that was not caught on release because it affected only dev server env
0
12,169
9,608,700,856
IssuesEvent
2019-05-12 09:00:33
eclipse/vorto
https://api.github.com/repos/eclipse/vorto
closed
Upgrade db schema on vorto-dev
Infrastructure Repository
Confirmations: - AWS RDS DB Schema contains changes for multi - tenancy
1.0
Upgrade db schema on vorto-dev - Confirmations: - AWS RDS DB Schema contains changes for multi - tenancy
infrastructure
upgrade db schema on vorto dev confirmations aws rds db schema contains changes for multi tenancy
1
248,599
7,934,299,283
IssuesEvent
2018-07-08 17:32:28
Kheiden/RPi-tankbot
https://api.github.com/repos/Kheiden/RPi-tankbot
closed
Single photo disparity map
Priority 2 core
Depends on #43 I want to create a single photo disparity map image.
1.0
Single photo disparity map - Depends on #43 I want to create a single photo disparity map image.
non_infrastructure
single photo disparity map depends on i want to create a single photo disparity map image
0
26,396
20,071,797,457
IssuesEvent
2022-02-04 08:03:10
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Build failure: "...nativeaot\Runtime\Full\AsmOffsets.cs' could not be found"
area-Infrastructure-coreclr untriaged in-pr
Building with: ``` build.cmd clr.runtime+Clr.NativeAotLibs+clr.corelib+clr.nativecorelib+clr.tools+clr.iltools+clr.alljits+libs -arch x64 -c Debug -librariesConfiguration Release ``` fails with: ``` CSC : error CS2001: Source file 'C:\gh\runtime3\artifacts\obj\\coreclr\windows.x64.Debug\nativeaot\Runtime\Full\AsmOffsets.cs' could not be found. [C:\gh\runtime3\src\coreclr\nativeaot\Test.CoreLib\src\Test.CoreLib.csproj] ``` However, building with: ``` build.cmd clr+libs -arch x64 -c Debug -librariesConfiguration Release ``` succeeds. This seems to indicate that there's some issue with the Clr.NativeAotLibs subset not building enough. @MichalStrehovsky ?
1.0
Build failure: "...nativeaot\Runtime\Full\AsmOffsets.cs' could not be found" - Building with: ``` build.cmd clr.runtime+Clr.NativeAotLibs+clr.corelib+clr.nativecorelib+clr.tools+clr.iltools+clr.alljits+libs -arch x64 -c Debug -librariesConfiguration Release ``` fails with: ``` CSC : error CS2001: Source file 'C:\gh\runtime3\artifacts\obj\\coreclr\windows.x64.Debug\nativeaot\Runtime\Full\AsmOffsets.cs' could not be found. [C:\gh\runtime3\src\coreclr\nativeaot\Test.CoreLib\src\Test.CoreLib.csproj] ``` However, building with: ``` build.cmd clr+libs -arch x64 -c Debug -librariesConfiguration Release ``` succeeds. This seems to indicate that there's some issue with the Clr.NativeAotLibs subset not building enough. @MichalStrehovsky ?
infrastructure
build failure nativeaot runtime full asmoffsets cs could not be found building with build cmd clr runtime clr nativeaotlibs clr corelib clr nativecorelib clr tools clr iltools clr alljits libs arch c debug librariesconfiguration release fails with csc error source file c gh artifacts obj coreclr windows debug nativeaot runtime full asmoffsets cs could not be found however building with build cmd clr libs arch c debug librariesconfiguration release succeeds this seems to indicate that there s some issue with the clr nativeaotlibs subset not building enough michalstrehovsky
1
13,750
16,502,892,037
IssuesEvent
2021-05-25 15:57:23
googleapis/google-auth-library-python
https://api.github.com/repos/googleapis/google-auth-library-python
closed
'testing/constraints-*.txt' not used
priority: p2 type: process
From #757. In [`noxfile.py`](https://github.com/googleapis/google-auth-library-python/blob/a9234423cb2b69068fc0d30a5a0ee86a599ab8b7/noxfile.py#L87-L89), the constraints are not being used, which is why the one populated file, [`testing/constraints-3.6.txt](https://github.com/googleapis/google-auth-library-python/blob/master/testing/constraints-3.6.txt) is in an invalid state (multiple entries for `rsa`, maybe more issues).
1.0
'testing/constraints-*.txt' not used - From #757. In [`noxfile.py`](https://github.com/googleapis/google-auth-library-python/blob/a9234423cb2b69068fc0d30a5a0ee86a599ab8b7/noxfile.py#L87-L89), the constraints are not being used, which is why the one populated file, [`testing/constraints-3.6.txt](https://github.com/googleapis/google-auth-library-python/blob/master/testing/constraints-3.6.txt) is in an invalid state (multiple entries for `rsa`, maybe more issues).
non_infrastructure
testing constraints txt not used from in the constraints are not being used which is why the one populated file is in an invalid state multiple entries for rsa maybe more issues
0
6,327
6,334,663,947
IssuesEvent
2017-07-26 17:08:57
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
closed
Update the publish definition to match what is in release/2.0.0 branch
area-Infrastructure
Need to port https://github.com/dotnet/coreclr/commit/ba86da4a6c821b0f042e105a72fd901c7f7b94c3 and https://github.com/dotnet/coreclr/commit/8ab4fb4260bdf1dafdf4b050bcc01102437f875d to master once the builds become stable.
1.0
Update the publish definition to match what is in release/2.0.0 branch - Need to port https://github.com/dotnet/coreclr/commit/ba86da4a6c821b0f042e105a72fd901c7f7b94c3 and https://github.com/dotnet/coreclr/commit/8ab4fb4260bdf1dafdf4b050bcc01102437f875d to master once the builds become stable.
infrastructure
update the publish definition to match what is in release branch need to port and to master once the builds become stable
1
85,825
15,755,286,339
IssuesEvent
2021-03-31 01:30:36
ChenLuigi/GitHubScannerBower4
https://api.github.com/repos/ChenLuigi/GitHubScannerBower4
opened
WS-2017-0247 (Low) detected in ms-0.7.1.tgz
security vulnerability
## WS-2017-0247 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.7.1.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.7.1.tgz">https://registry.npmjs.org/ms/-/ms-0.7.1.tgz</a></p> <p>Path to dependency file: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/package.json</p> <p>Path to vulnerable library: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/node_modules/ms/package.json</p> <p> Dependency Hierarchy: - grunt-sass-0.17.0.tgz (Root Library) - node-sass-1.2.3.tgz - mocha-2.5.3.tgz - debug-2.2.0.tgz - :x: **ms-0.7.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS). <p>Publish Date: 2017-04-12 <p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p> <p>Release Date: 2017-04-12</p> <p>Fix Resolution: 2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2017-0247 (Low) detected in ms-0.7.1.tgz - ## WS-2017-0247 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ms-0.7.1.tgz</b></p></summary> <p>Tiny ms conversion utility</p> <p>Library home page: <a href="https://registry.npmjs.org/ms/-/ms-0.7.1.tgz">https://registry.npmjs.org/ms/-/ms-0.7.1.tgz</a></p> <p>Path to dependency file: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/package.json</p> <p>Path to vulnerable library: GitHubScannerBower4/GoldenPanel_Lighter/GoldenPanel/c3-0.4.10/package/node_modules/ms/package.json</p> <p> Dependency Hierarchy: - grunt-sass-0.17.0.tgz (Root Library) - node-sass-1.2.3.tgz - mocha-2.5.3.tgz - debug-2.2.0.tgz - :x: **ms-0.7.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of this package are vulnerable to Regular Expression Denial of Service (ReDoS). <p>Publish Date: 2017-04-12 <p>URL: <a href=https://github.com/zeit/ms/commit/305f2ddcd4eff7cc7c518aca6bb2b2d2daad8fef>WS-2017-0247</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>3.4</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/vercel/ms/pull/89">https://github.com/vercel/ms/pull/89</a></p> <p>Release Date: 2017-04-12</p> <p>Fix Resolution: 2.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
ws low detected in ms tgz ws low severity vulnerability vulnerable library ms tgz tiny ms conversion utility library home page a href path to dependency file goldenpanel lighter goldenpanel package package json path to vulnerable library goldenpanel lighter goldenpanel package node modules ms package json dependency hierarchy grunt sass tgz root library node sass tgz mocha tgz debug tgz x ms tgz vulnerable library vulnerability details affected versions of this package are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
583
2,776,011,819
IssuesEvent
2015-05-04 19:19:52
WP-API/WP-API
https://api.github.com/repos/WP-API/WP-API
closed
Write permission discoverability for endpoints
#Blessed Infrastructure
As a client developer writing an application, I want to have my application discover which endpoints are readable / writable through response values, and without having to know the intricacies of WordPress' capabilities system. Currently, we can't easily support this use case because capability is determined at run time. We'll need some sort of framework for a `WP_JSON_Controller` to expose permission scope for its endpoints.
1.0
Write permission discoverability for endpoints - As a client developer writing an application, I want to have my application discover which endpoints are readable / writable through response values, and without having to know the intricacies of WordPress' capabilities system. Currently, we can't easily support this use case because capability is determined at run time. We'll need some sort of framework for a `WP_JSON_Controller` to expose permission scope for its endpoints.
infrastructure
write permission discoverability for endpoints as a client developer writing an application i want to have my application discover which endpoints are readable writable through response values and without having to know the intricacies of wordpress capabilities system currently we can t easily support this use case because capability is determined at run time we ll need some sort of framework for a wp json controller to expose permission scope for its endpoints
1
938
2,521,170,027
IssuesEvent
2015-01-19 12:19:05
ChrisCummins/msc-thesis
https://api.github.com/repos/ChrisCummins/msc-thesis
closed
Build error on Ubuntu using distro automake install.
Bug Testing & Tooling
Ubuntu distro version of automake is 1.11.3. The `AM_EXTRA_RECURSIVE_TARGETS` macro used in `configure.ac` is undefined in automake < 1.13. See GNU [Automake 1.13 release notes](http://lwn.net/Articles/531373/).
1.0
Build error on Ubuntu using distro automake install. - Ubuntu distro version of automake is 1.11.3. The `AM_EXTRA_RECURSIVE_TARGETS` macro used in `configure.ac` is undefined in automake < 1.13. See GNU [Automake 1.13 release notes](http://lwn.net/Articles/531373/).
non_infrastructure
build error on ubuntu using distro automake install ubuntu distro version of automake is the am extra recursive targets macro used in configure ac is undefined in automake see gnu
0
8,597
7,519,415,515
IssuesEvent
2018-04-12 11:32:25
djavorszky/ddn
https://api.github.com/repos/djavorszky/ddn
closed
Containeraize agents
enhancement infrastructure new feature question
Currently, most connectors are running inside containers, but in a manual fashion, that is, the container is started and then the connector is copied inside and then started manually. I'd like to have containers that already have these (mostly) set up from the moment someone pulls them and starts them. This would allow us to also do sub-second updates on them. Will require us writing a Dockerfile for each. - [x] mysql-55 - [x] mysql-56 - [x] mysql-57 - [x] postgres-94
1.0
Containeraize agents - Currently, most connectors are running inside containers, but in a manual fashion, that is, the container is started and then the connector is copied inside and then started manually. I'd like to have containers that already have these (mostly) set up from the moment someone pulls them and starts them. This would allow us to also do sub-second updates on them. Will require us writing a Dockerfile for each. - [x] mysql-55 - [x] mysql-56 - [x] mysql-57 - [x] postgres-94
infrastructure
containeraize agents currently most connectors are running inside containers but in a manual fashion that is the container is started and then the connector is copied inside and then started manually i d like to have containers that already have these mostly set up from the moment someone pulls them and starts them this would allow us to also do sub second updates on them will require us writing a dockerfile for each mysql mysql mysql postgres
1
22,948
15,691,005,849
IssuesEvent
2021-03-25 17:21:07
Alv-no/alv-website
https://api.github.com/repos/Alv-no/alv-website
closed
Setup deployment pipeline on Azure
infrastructure
- Setup Azure DevOps pipeline inspired [by this](https://github.com/Alv-no/alvtime-monorepo/tree/master/.AzureDevOps/pipelines) - Use Terraform and Docker
1.0
Setup deployment pipeline on Azure - - Setup Azure DevOps pipeline inspired [by this](https://github.com/Alv-no/alvtime-monorepo/tree/master/.AzureDevOps/pipelines) - Use Terraform and Docker
infrastructure
setup deployment pipeline on azure setup azure devops pipeline inspired use terraform and docker
1
77,423
9,565,076,341
IssuesEvent
2019-05-05 09:32:34
gitthermal/thermal
https://api.github.com/repos/gitthermal/thermal
opened
Commit related information to repository settings
Design Front end Git discussion 💡 Feature
## Description Add commit related information to the repository settings. #### For example - General - Last open - Last commit - User description - Working copy - Current Branch - Status
1.0
Commit related information to repository settings - ## Description Add commit related information to the repository settings. #### For example - General - Last open - Last commit - User description - Working copy - Current Branch - Status
non_infrastructure
commit related information to repository settings description add commit related information to the repository settings for example general last open last commit user description working copy current branch status
0
71,221
23,492,766,477
IssuesEvent
2022-08-17 20:33:14
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[Bug :bug:]: ModuleNotFoundError: No module named 'certifi'
C-py I-defect needs-triaging
### What happened? Error in [this build](https://github.com/alexmojaki/pyodide-worker-runner/runs/7885917987?check_suite_focus=true). ### How can we reproduce the issue? ```shell from selenium import webdriver ``` ### Relevant log output ```shell test.py:8: in <module> from selenium import webdriver /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/__init__.py:18: in <module> from .firefox.webdriver import WebDriver as Firefox # noqa /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/firefox/webdriver.py:26: in <module> from selenium.webdriver.remote.webdriver import WebDriver as RemoteWebDriver /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py:52: in <module> from .remote_connection import RemoteConnection /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/remote_connection.py:25: in <module> import certifi E ModuleNotFoundError: No module named 'certifi' ``` ### Operating System Ubuntu ### Selenium version Python 4.4.1 ### What are the browser(s) and version(s) where you see this issue? None ### What are the browser driver(s) and version(s) where you see this issue? None ### Are you using Selenium Grid? _No response_
1.0
[Bug :bug:]: ModuleNotFoundError: No module named 'certifi' - ### What happened? Error in [this build](https://github.com/alexmojaki/pyodide-worker-runner/runs/7885917987?check_suite_focus=true). ### How can we reproduce the issue? ```shell from selenium import webdriver ``` ### Relevant log output ```shell test.py:8: in <module> from selenium import webdriver /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/__init__.py:18: in <module> from .firefox.webdriver import WebDriver as Firefox # noqa /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/firefox/webdriver.py:26: in <module> from selenium.webdriver.remote.webdriver import WebDriver as RemoteWebDriver /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/webdriver.py:52: in <module> from .remote_connection import RemoteConnection /opt/hostedtoolcache/Python/3.9.13/x64/lib/python3.9/site-packages/selenium/webdriver/remote/remote_connection.py:25: in <module> import certifi E ModuleNotFoundError: No module named 'certifi' ``` ### Operating System Ubuntu ### Selenium version Python 4.4.1 ### What are the browser(s) and version(s) where you see this issue? None ### What are the browser driver(s) and version(s) where you see this issue? None ### Are you using Selenium Grid? _No response_
non_infrastructure
modulenotfounderror no module named certifi what happened error in how can we reproduce the issue shell from selenium import webdriver relevant log output shell test py in from selenium import webdriver opt hostedtoolcache python lib site packages selenium webdriver init py in from firefox webdriver import webdriver as firefox noqa opt hostedtoolcache python lib site packages selenium webdriver firefox webdriver py in from selenium webdriver remote webdriver import webdriver as remotewebdriver opt hostedtoolcache python lib site packages selenium webdriver remote webdriver py in from remote connection import remoteconnection opt hostedtoolcache python lib site packages selenium webdriver remote remote connection py in import certifi e modulenotfounderror no module named certifi operating system ubuntu selenium version python what are the browser s and version s where you see this issue none what are the browser driver s and version s where you see this issue none are you using selenium grid no response
0
3,988
4,754,951,420
IssuesEvent
2016-10-24 09:11:42
robotology/QA
https://api.github.com/repos/robotology/QA
closed
New GUI ethLoader
firmware icub-ethernet infrastructure
Hi, I wanted to update the robot, and as I've understood from issues #134 and #135, there is a new gui for the `ethLoader`, and all the updated instructions are referred to the new gui. However, even if I updated everything, I still see the old gui; therefore, is there any option or so that I had to activate to see the new gui? (I updated the firmware anyway with the old one, but for future and also students it would be better to have the one that matches with the instructions). Thanks!
1.0
New GUI ethLoader - Hi, I wanted to update the robot, and as I've understood from issues #134 and #135, there is a new gui for the `ethLoader`, and all the updated instructions are referred to the new gui. However, even if I updated everything, I still see the old gui; therefore, is there any option or so that I had to activate to see the new gui? (I updated the firmware anyway with the old one, but for future and also students it would be better to have the one that matches with the instructions). Thanks!
infrastructure
new gui ethloader hi i wanted to update the robot and as i ve understood from issues and there is a new gui for the ethloader and all the updated instructions are referred to the new gui however even if i updated everything i still see the old gui therefore is there any option or so that i had to activate to see the new gui i updated the firmware anyway with the old one but for future and also students it would be better to have the one that matches with the instructions thanks
1
9,060
3,834,080,298
IssuesEvent
2016-04-01 08:11:28
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Joomla 3.5 Admin Problem
No Code Attached Yet
Just upgraded to 3.5, now I can only see active position items under the "Select Position" drop down menu, all other positions are gone. See attached screen shot. HELP!![screen shot 2016-03-31 at 20 50 28](https://issues.joomla.org/uploads/1/bf8d1ef0b652f083f03c74060dcb2d9c.png)
1.0
Joomla 3.5 Admin Problem - Just upgraded to 3.5, now I can only see active position items under the "Select Position" drop down menu, all other positions are gone. See attached screen shot. HELP!![screen shot 2016-03-31 at 20 50 28](https://issues.joomla.org/uploads/1/bf8d1ef0b652f083f03c74060dcb2d9c.png)
non_infrastructure
joomla admin problem just upgraded to now i can only see active position items under the select position drop down menu all other positions are gone see attached screen shot help
0
11,336
9,311,762,219
IssuesEvent
2019-03-25 22:21:56
aspnet/AspNetCore
https://api.github.com/repos/aspnet/AspNetCore
closed
Logged-in state is not consistent between Fetch data page and account management pages in React app with authentication in ASP.NET Core 3 Preview 3
area-mvc identity-service investigate
### Describe the bug In a React app with authentication created using ASP.NET Core 3 Preview 3 when I click "Logout" on my account profile page I seem to be logged out there, but still logged in on the home page and "Fetch data" page. ### To Reproduce 1. Using ASP.NET Core 3.0 Preview 3 create a new project with `dotnet new react --auth Individual` 2. Apply the various workarounds at https://github.com/dotnet/core/blob/master/release-notes/3.0/preview/3.0.0-preview-known-issues.md 3. Run the app. 4. Click "Register" and complete the registration form (or log in with an existing account). 5. Click on "Hello youremail@..." in the header to go to /Identity/Account/Manage 6. Click "Logout" in the header on that page. You should be redirected back the homepage. 7. Observe that the header still says "Hello youremail@..." and shows a "Logout" link. 8. Click "Fetch data" in the header - the Weather forecast page is displayed, so you are indeed still logged in. 9. Click on "Hello youremail@..." in the header again. You are now prompted to log in, so you are not logged in on /Identity/Account/Manage. ### Expected behavior When I click "Logout" on any page I am logged out for the entire application. Note that the reverse is not true - when I log out from the home page and browse to /Identity/Account/Manage I am prompted to log in, as expected.
1.0
Logged-in state is not consistent between Fetch data page and account management pages in React app with authentication in ASP.NET Core 3 Preview 3 - ### Describe the bug In a React app with authentication created using ASP.NET Core 3 Preview 3 when I click "Logout" on my account profile page I seem to be logged out there, but still logged in on the home page and "Fetch data" page. ### To Reproduce 1. Using ASP.NET Core 3.0 Preview 3 create a new project with `dotnet new react --auth Individual` 2. Apply the various workarounds at https://github.com/dotnet/core/blob/master/release-notes/3.0/preview/3.0.0-preview-known-issues.md 3. Run the app. 4. Click "Register" and complete the registration form (or log in with an existing account). 5. Click on "Hello youremail@..." in the header to go to /Identity/Account/Manage 6. Click "Logout" in the header on that page. You should be redirected back the homepage. 7. Observe that the header still says "Hello youremail@..." and shows a "Logout" link. 8. Click "Fetch data" in the header - the Weather forecast page is displayed, so you are indeed still logged in. 9. Click on "Hello youremail@..." in the header again. You are now prompted to log in, so you are not logged in on /Identity/Account/Manage. ### Expected behavior When I click "Logout" on any page I am logged out for the entire application. Note that the reverse is not true - when I log out from the home page and browse to /Identity/Account/Manage I am prompted to log in, as expected.
non_infrastructure
logged in state is not consistent between fetch data page and account management pages in react app with authentication in asp net core preview describe the bug in a react app with authentication created using asp net core preview when i click logout on my account profile page i seem to be logged out there but still logged in on the home page and fetch data page to reproduce using asp net core preview create a new project with dotnet new react auth individual apply the various workarounds at run the app click register and complete the registration form or log in with an existing account click on hello youremail in the header to go to identity account manage click logout in the header on that page you should be redirected back the homepage observe that the header still says hello youremail and shows a logout link click fetch data in the header the weather forecast page is displayed so you are indeed still logged in click on hello youremail in the header again you are now prompted to log in so you are not logged in on identity account manage expected behavior when i click logout on any page i am logged out for the entire application note that the reverse is not true when i log out from the home page and browse to identity account manage i am prompted to log in as expected
0
22,954
15,696,571,661
IssuesEvent
2021-03-26 02:22:16
seattle-uat/civiform
https://api.github.com/repos/seattle-uat/civiform
closed
enable civiform.com for our AWS instance
dev infrastructure
Can we point civiform.com to our AWS instance. I'm holding the domain under my personal GoDaddy account. Please provide me the configuration that I need to set. Gerardo
1.0
enable civiform.com for our AWS instance - Can we point civiform.com to our AWS instance. I'm holding the domain under my personal GoDaddy account. Please provide me the configuration that I need to set. Gerardo
infrastructure
enable civiform com for our aws instance can we point civiform com to our aws instance i m holding the domain under my personal godaddy account please provide me the configuration that i need to set gerardo
1
28,925
23,613,870,699
IssuesEvent
2022-08-24 14:25:37
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Include upper level classes plus bridging axioms in go-plus import chain
Infrastructure
Note: this is prompted by the desire to retire go-upper.obo https://github.com/geneontology/go-site/issues/129 Including upper level classes (primarily BFO but also CARO) in the import chain is itself straightforward. Note that we may want to limit confusing annotation assertions. However, this is not much use unless bindings are present. An early attempt for GO: http://purl.obolibrary.org/obo/go/extensions/go-bfo-bridge.owl this is probably overkill and can be replaced by 3 binding/bridge axioms (MF,BP->P, CC->CARO:?) However, we will also need to ensure that bindings are present for all imported ontologies, to prevent them hanging out at root. There are a number of issues here: - we might need to make ontological commitments on others behalf, e.g. SO - import dependencies are hard to manage, see issue 38 in the [robot tracker](https://github.com/ontodev/robot/issues). but this is something we need to tackle anyway This may wait til after: #12211
1.0
Include upper level classes plus bridging axioms in go-plus import chain - Note: this is prompted by the desire to retire go-upper.obo https://github.com/geneontology/go-site/issues/129 Including upper level classes (primarily BFO but also CARO) in the import chain is itself straightforward. Note that we may want to limit confusing annotation assertions. However, this is not much use unless bindings are present. An early attempt for GO: http://purl.obolibrary.org/obo/go/extensions/go-bfo-bridge.owl this is probably overkill and can be replaced by 3 binding/bridge axioms (MF,BP->P, CC->CARO:?) However, we will also need to ensure that bindings are present for all imported ontologies, to prevent them hanging out at root. There are a number of issues here: - we might need to make ontological commitments on others behalf, e.g. SO - import dependencies are hard to manage, see issue 38 in the [robot tracker](https://github.com/ontodev/robot/issues). but this is something we need to tackle anyway This may wait til after: #12211
infrastructure
include upper level classes plus bridging axioms in go plus import chain note this is prompted by the desire to retire go upper obo including upper level classes primarily bfo but also caro in the import chain is itself straightforward note that we may want to limit confusing annotation assertions however this is not much use unless bindings are present an early attempt for go this is probably overkill and can be replaced by binding bridge axioms mf bp p cc caro however we will also need to ensure that bindings are present for all imported ontologies to prevent them hanging out at root there are a number of issues here we might need to make ontological commitments on others behalf e g so import dependencies are hard to manage see issue in the but this is something we need to tackle anyway this may wait til after
1
566,583
16,824,772,206
IssuesEvent
2021-06-17 17:01:25
heyset/deskinar
https://api.github.com/repos/heyset/deskinar
opened
User should be able to answer questions
3 points high priority new feature
Any user (for now) should be able to answer questions, and add them to the thread.
1.0
User should be able to answer questions - Any user (for now) should be able to answer questions, and add them to the thread.
non_infrastructure
user should be able to answer questions any user for now should be able to answer questions and add them to the thread
0
13,911
10,543,972,778
IssuesEvent
2019-10-02 15:57:27
mp-access/Backend
https://api.github.com/repos/mp-access/Backend
opened
User Acceptance Environment
Infrastructure
For our new release procedure we want a new user acceptance environment on digital ocean. - We have a release candidate branch that is deployed on digital ocean. - RC feature freeze is on Wednesday an is testes one week until the release one week later.
1.0
User Acceptance Environment - For our new release procedure we want a new user acceptance environment on digital ocean. - We have a release candidate branch that is deployed on digital ocean. - RC feature freeze is on Wednesday an is testes one week until the release one week later.
infrastructure
user acceptance environment for our new release procedure we want a new user acceptance environment on digital ocean we have a release candidate branch that is deployed on digital ocean rc feature freeze is on wednesday an is testes one week until the release one week later
1
450,648
31,934,154,239
IssuesEvent
2023-09-19 09:22:07
FuelLabs/sway-libs
https://api.github.com/repos/FuelLabs/sway-libs
closed
Specification missing function parameters
Documentation Good First Issue Improvement
### Motivation In all of the libraries i.e. [NFT](https://github.com/FuelLabs/sway-libs/blob/master/libs/nft/SPECIFICATION.md) are missing function parameters that tell the user how to use the specific function. It has been brought up in the forum [here](https://forum.fuel.network/t/sway-libs-nft-library/2893/5)
1.0
Specification missing function parameters - ### Motivation In all of the libraries i.e. [NFT](https://github.com/FuelLabs/sway-libs/blob/master/libs/nft/SPECIFICATION.md) are missing function parameters that tell the user how to use the specific function. It has been brought up in the forum [here](https://forum.fuel.network/t/sway-libs-nft-library/2893/5)
non_infrastructure
specification missing function parameters motivation in all of the libraries i e are missing function parameters that tell the user how to use the specific function it has been brought up in the forum
0
52,663
13,224,888,279
IssuesEvent
2020-08-17 20:03:04
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
i3_add_pybindings in top level CMakeLists.txt -> wrong lib location (Trac #120)
Migrated from Trac cmake defect
It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file, the build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib A work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory() But Troy claims they should happily coexist at top level CMakeLists.txt <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/120">https://code.icecube.wisc.edu/projects/icecube/ticket/120</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "_ts": "1416713876900096", "description": "It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file,\nthe build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib\n\nA work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory()\n\nBut Troy claims they should happily coexist at top level CMakeLists.txt", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2008-09-03T15:55:07", "component": "cmake", "summary": "i3_add_pybindings in top level CMakeLists.txt -> wrong lib location", "priority": "major", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
i3_add_pybindings in top level CMakeLists.txt -> wrong lib location (Trac #120) - It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file, the build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib A work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory() But Troy claims they should happily coexist at top level CMakeLists.txt <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/120">https://code.icecube.wisc.edu/projects/icecube/ticket/120</a>, reported by blaufussand owned by troy</em></summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:56", "_ts": "1416713876900096", "description": "It seems when the i3_add_pybindings voodoo is in the toplevel CMakeLists.txt file,\nthe build paths of the standard libs is changed to $I3_BUILD/lib/icecube instead of the standard $I3_BUILD/lib\n\nA work around is to use a separate CMakeLists.txt file in private/pybindings and add the add_subdirectory()\n\nBut Troy claims they should happily coexist at top level CMakeLists.txt", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "time": "2008-09-03T15:55:07", "component": "cmake", "summary": "i3_add_pybindings in top level CMakeLists.txt -> wrong lib location", "priority": "major", "keywords": "", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
non_infrastructure
add pybindings in top level cmakelists txt wrong lib location trac it seems when the add pybindings voodoo is in the toplevel cmakelists txt file the build paths of the standard libs is changed to build lib icecube instead of the standard build lib a work around is to use a separate cmakelists txt file in private pybindings and add the add subdirectory but troy claims they should happily coexist at top level cmakelists txt migrated from json status closed changetime ts description it seems when the add pybindings voodoo is in the toplevel cmakelists txt file nthe build paths of the standard libs is changed to build lib icecube instead of the standard build lib n na work around is to use a separate cmakelists txt file in private pybindings and add the add subdirectory n nbut troy claims they should happily coexist at top level cmakelists txt reporter blaufuss cc resolution fixed time component cmake summary add pybindings in top level cmakelists txt wrong lib location priority major keywords milestone owner troy type defect
0
329,877
10,025,935,443
IssuesEvent
2019-07-17 04:33:34
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
none: Some fatal errors occurred: [ERROR Port-8443]: Port 8443 is in use
cause/port-conflict co/none-driver kind/bug priority/backlog r/2019q2 triage/needs-information
<!-- Please use this template while reporting an issue, providing as much information as possible. Failure to do so may result in a delayed response. Thank you! --> **The exact command to reproduce the issue**: ``` export MINIKUBE_WANTUPDATENOTIFICATION=false export MINIKUBE_WANTREPORTERRORPROMPT=false export CHANGE_MINIKUBE_NONE_USER=true sudo -E minikube start --vm-driver=none ``` **The full output of the command that failed**: ``` 😄 minikube v1.0.1 on linux (amd64) 🔥 Creating none VM (CPUs=2, Memory=2048MB, Disk=20000MB) ... 📶 "minikube" IP address is 192.168.8.159 🐳 Configuring Docker as the container runtime ... 🐳 Version of container runtime is 18.09.6 ✨ Preparing Kubernetes environment ... ❌ Unable to load cached images: loading cached images: loading image /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: stat /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: no such file or directory 💾 Downloading kubeadm v1.14.1 💾 Downloading kubelet v1.14.1 🚜 Pulling images required by Kubernetes v1.14.1 ... 🚀 Launching Kubernetes v1.14.1 using kubeadm ... 💣 Error starting cluster: kubeadm init: sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI : running command: sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI output: [init] Using Kubernetes version: v1.14.1 [preflight] Running pre-flight checks [WARNING FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml]: /etc/kubernetes/manifests/kube-apiserver.yaml already exists [WARNING FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml]: /etc/kubernetes/manifests/kube-controller-manager.yaml already exists [WARNING FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml]: /etc/kubernetes/manifests/kube-scheduler.yaml already exists [WARNING FileAvailable--etc-kubernetes-manifests-etcd.yaml]: /etc/kubernetes/manifests/etcd.yaml already exists [WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/ [WARNING Swap]: running with swap on is not supported. Please disable swap [WARNING FileExisting-ebtables]: ebtables not found in system path [WARNING FileExisting-ethtool]: ethtool not found in system path [WARNING FileExisting-socat]: socat not found in system path [WARNING Hostname]: hostname "minikube" could not be reached [WARNING Hostname]: hostname "minikube": lookup minikube on 127.0.0.53:53: server misbehaving [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service' [WARNING Port-10250]: Port 10250 is in use [WARNING DirAvailable--data-minikube]: /data/minikube is not empty error execution phase preflight: [preflight] Some fatal errors occurred: [ERROR Port-8443]: Port 8443 is in use [ERROR Port-10251]: Port 10251 is in use [ERROR Port-10252]: Port 10252 is in use [ERROR Port-2379]: Port 2379 is in use [ERROR Port-2380]: Port 2380 is in use [preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...` : running command: sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI .: exit status 1 😿 Sorry that minikube crashed. If this was unexpected, we would love to hear from you: 👉 https://github.com/kubernetes/minikube/issues/new ❌ Problems detected in "kube-addon-manager": error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused ``` **The output of the `minikube logs` command**: ``` ==> dmesg <== [May11 16:28] secureboot: Secure boot could not be determined (mode 0) [ +0.000000] pmd_set_huge: Cannot satisfy [mem 0xf8000000-0xf8200000] with a huge-page mapping due to MTRR override. [ +1.100563] r8169 0000:02:00.0: can't disable ASPM; OS doesn't have ASPM control [ +0.333285] ata5.00: supports DRM functions and may not be fully accessible [ +0.000140] ata5.00: READ LOG DMA EXT failed, trying PIO [ +0.018473] ata5.00: supports DRM functions and may not be fully accessible [ +0.650433] usb 3-5.2: device descriptor read/64, error -32 [ +9.859326] kauditd_printk_skb: 58 callbacks suppressed [ +0.489483] aufs au_opts_verify:1609:dockerd[1937]: dirperm1 breaks the protection by the permission bits on the lower branch [May12 03:01] IRQ 16: no longer affine to CPU1 [ +0.000007] IRQ 29: no longer affine to CPU1 [ +0.024017] IRQ 23: no longer affine to CPU2 [ +0.000007] IRQ 27: no longer affine to CPU2 [ +0.032058] IRQ 26: no longer affine to CPU3 [ +0.000010] IRQ 28: no longer affine to CPU3 [ +0.012294] cache: parent cpu1 should not be sleeping [ +0.002224] cache: parent cpu2 should not be sleeping [ +0.002101] cache: parent cpu3 should not be sleeping [ +0.375470] ata5.00: supports DRM functions and may not be fully accessible [ +0.018803] ata5.00: supports DRM functions and may not be fully accessible [ +5.010804] ata1: link is slow to respond, please be patient (ready=0) [ +4.651973] ata1: COMRESET failed (errno=-16) [May12 06:52] sd 6:0:0:0: [sdc] No Caching mode page found [ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through [ +0.012436] sd 6:0:0:1: [sdd] No Caching mode page found [ +0.000009] sd 6:0:0:1: [sdd] Assuming drive cache: write through [May12 11:07] sd 6:0:0:0: [sdc] No Caching mode page found [ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through [ +0.001174] sd 6:0:0:1: [sdd] No Caching mode page found [ +0.000003] sd 6:0:0:1: [sdd] Assuming drive cache: write through [May12 12:03] kauditd_printk_skb: 32 callbacks suppressed [May12 18:13] IRQ 16: no longer affine to CPU1 [ +0.000006] IRQ 29: no longer affine to CPU1 [ +0.024149] IRQ 23: no longer affine to CPU2 [ +0.000006] IRQ 27: no longer affine to CPU2 [ +0.032001] IRQ 26: no longer affine to CPU3 [ +0.000007] IRQ 28: no longer affine to CPU3 [ +0.011035] cache: parent cpu1 should not be sleeping [ +0.002241] cache: parent cpu2 should not be sleeping [ +0.002119] cache: parent cpu3 should not be sleeping [ +0.369614] ata5.00: supports DRM functions and may not be fully accessible [ +0.019113] ata5.00: supports DRM functions and may not be fully accessible [ +5.037481] ata1: link is slow to respond, please be patient (ready=0) [ +4.676005] ata1: COMRESET failed (errno=-16) [May13 02:54] kauditd_printk_skb: 37 callbacks suppressed ==> kernel <== 20:38:38 up 2 days, 4:10, 1 user, load average: 0.96, 0.86, 0.82 Linux desktop 4.18.0-18-generic #19~18.04.1-Ubuntu SMP Fri Apr 5 10:22:13 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux ==> kube-addon-manager <== INFO: == Kubernetes addon manager started at 2019-05-14T00:33:24+00:00 with ADDON_CHECK_INTERVAL_SEC=60 == error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused WRN: == Failed to start /opt/namespace.yaml in namespace at 2019-05-14T00:33:24+00:00. 99 tries remaining. == INFO: == Default service account in the kube-system namespace has token default-token-hr5j9 == find: '/etc/kubernetes/admission-controls': No such file or directory INFO: == Entering periodical apply loop at 2019-05-14T00:33:29+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:33:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:33:31+00:00 == namespace/kube-system unchanged INFO: == Successfully started /opt/namespace.yaml in namespace at 2019-05-14T00:33:34+00:00 INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:34:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:34:30+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:35:30+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:35:31+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:36:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:36:31+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:37:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:37:31+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:38:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:38:30+00:00 == ==> kube-apiserver <== I0514 00:38:34.831507 1 log.go:172] http: TLS handshake error from 127.0.0.1:49094: remote error: tls: bad certificate I0514 00:38:35.021488 1 log.go:172] http: TLS handshake error from 127.0.0.1:49096: remote error: tls: bad certificate I0514 00:38:35.235536 1 log.go:172] http: TLS handshake error from 127.0.0.1:49100: remote error: tls: bad certificate I0514 00:38:35.436229 1 log.go:172] http: TLS handshake error from 127.0.0.1:49102: remote error: tls: bad certificate I0514 00:38:35.632126 1 log.go:172] http: TLS handshake error from 127.0.0.1:49104: remote error: tls: bad certificate I0514 00:38:35.708998 1 log.go:172] http: TLS handshake error from 127.0.0.1:49106: remote error: tls: bad certificate I0514 00:38:35.711809 1 log.go:172] http: TLS handshake error from 127.0.0.1:49108: remote error: tls: bad certificate I0514 00:38:35.711850 1 log.go:172] http: TLS handshake error from 127.0.0.1:49110: remote error: tls: bad certificate I0514 00:38:35.711876 1 log.go:172] http: TLS handshake error from 127.0.0.1:49112: remote error: tls: bad certificate I0514 00:38:35.713858 1 log.go:172] http: TLS handshake error from 127.0.0.1:49114: remote error: tls: bad certificate I0514 00:38:35.714111 1 log.go:172] http: TLS handshake error from 127.0.0.1:49116: remote error: tls: bad certificate I0514 00:38:35.715785 1 log.go:172] http: TLS handshake error from 127.0.0.1:49118: remote error: tls: bad certificate I0514 00:38:35.716377 1 log.go:172] http: TLS handshake error from 127.0.0.1:49124: remote error: tls: bad certificate I0514 00:38:35.716620 1 log.go:172] http: TLS handshake error from 127.0.0.1:49120: remote error: tls: bad certificate I0514 00:38:35.716870 1 log.go:172] http: TLS handshake error from 127.0.0.1:49122: remote error: tls: bad certificate I0514 00:38:35.722733 1 log.go:172] http: TLS handshake error from 127.0.0.1:49126: remote error: tls: bad certificate I0514 00:38:35.840634 1 log.go:172] http: TLS handshake error from 127.0.0.1:49128: remote error: tls: bad certificate I0514 00:38:36.027863 1 log.go:172] http: TLS handshake error from 127.0.0.1:49130: remote error: tls: bad certificate I0514 00:38:36.243197 1 log.go:172] http: TLS handshake error from 127.0.0.1:49132: remote error: tls: bad certificate I0514 00:38:36.299517 1 log.go:172] http: TLS handshake error from 127.0.0.1:49134: remote error: tls: bad certificate I0514 00:38:36.446919 1 log.go:172] http: TLS handshake error from 127.0.0.1:49136: remote error: tls: bad certificate I0514 00:38:36.640301 1 log.go:172] http: TLS handshake error from 127.0.0.1:49138: remote error: tls: bad certificate I0514 00:38:36.717558 1 log.go:172] http: TLS handshake error from 127.0.0.1:49140: remote error: tls: bad certificate I0514 00:38:36.720112 1 log.go:172] http: TLS handshake error from 127.0.0.1:49150: remote error: tls: bad certificate I0514 00:38:36.720144 1 log.go:172] http: TLS handshake error from 127.0.0.1:49144: remote error: tls: bad certificate I0514 00:38:36.720168 1 log.go:172] http: TLS handshake error from 127.0.0.1:49142: remote error: tls: bad certificate I0514 00:38:36.721405 1 log.go:172] http: TLS handshake error from 127.0.0.1:49146: remote error: tls: bad certificate I0514 00:38:36.721835 1 log.go:172] http: TLS handshake error from 127.0.0.1:49152: remote error: tls: bad certificate I0514 00:38:36.721908 1 log.go:172] http: TLS handshake error from 127.0.0.1:49148: remote error: tls: bad certificate I0514 00:38:36.722280 1 log.go:172] http: TLS handshake error from 127.0.0.1:49154: remote error: tls: bad certificate I0514 00:38:36.723077 1 log.go:172] http: TLS handshake error from 127.0.0.1:49156: remote error: tls: bad certificate I0514 00:38:36.723116 1 log.go:172] http: TLS handshake error from 127.0.0.1:49158: remote error: tls: bad certificate I0514 00:38:36.843667 1 log.go:172] http: TLS handshake error from 127.0.0.1:49160: remote error: tls: bad certificate I0514 00:38:37.032639 1 log.go:172] http: TLS handshake error from 127.0.0.1:49164: remote error: tls: bad certificate I0514 00:38:37.253082 1 log.go:172] http: TLS handshake error from 127.0.0.1:49166: remote error: tls: bad certificate I0514 00:38:37.461185 1 log.go:172] http: TLS handshake error from 127.0.0.1:49168: remote error: tls: bad certificate I0514 00:38:37.647529 1 log.go:172] http: TLS handshake error from 127.0.0.1:49170: remote error: tls: bad certificate I0514 00:38:37.726883 1 log.go:172] http: TLS handshake error from 127.0.0.1:49172: remote error: tls: bad certificate I0514 00:38:37.731172 1 log.go:172] http: TLS handshake error from 127.0.0.1:49178: remote error: tls: bad certificate I0514 00:38:37.731797 1 log.go:172] http: TLS handshake error from 127.0.0.1:49174: remote error: tls: bad certificate I0514 00:38:37.731833 1 log.go:172] http: TLS handshake error from 127.0.0.1:49184: remote error: tls: bad certificate I0514 00:38:37.731857 1 log.go:172] http: TLS handshake error from 127.0.0.1:49182: remote error: tls: bad certificate I0514 00:38:37.731881 1 log.go:172] http: TLS handshake error from 127.0.0.1:49176: remote error: tls: bad certificate I0514 00:38:37.733898 1 log.go:172] http: TLS handshake error from 127.0.0.1:49186: remote error: tls: bad certificate I0514 00:38:37.733938 1 log.go:172] http: TLS handshake error from 127.0.0.1:49180: remote error: tls: bad certificate I0514 00:38:37.734391 1 log.go:172] http: TLS handshake error from 127.0.0.1:49190: remote error: tls: bad certificate I0514 00:38:37.734678 1 log.go:172] http: TLS handshake error from 127.0.0.1:49188: remote error: tls: bad certificate I0514 00:38:37.847115 1 log.go:172] http: TLS handshake error from 127.0.0.1:49194: remote error: tls: bad certificate I0514 00:38:38.035683 1 log.go:172] http: TLS handshake error from 127.0.0.1:49198: remote error: tls: bad certificate I0514 00:38:38.202140 1 log.go:172] http: TLS handshake error from 127.0.0.1:49200: remote error: tls: bad certificate ==> kube-scheduler <== E0514 00:38:33.696199 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolumeClaim: Get https://localhost:8443/api/v1/persistentvolumeclaims?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.696962 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolume: Get https://localhost:8443/api/v1/persistentvolumes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.696988 1 reflector.go:126] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:223: Failed to list *v1.Pod: Get https://localhost:8443/api/v1/pods?fieldSelector=status.phase%3DFailed%!C(MISSING)status.phase%3DSucceeded&limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.699704 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Service: Get https://localhost:8443/api/v1/services?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700091 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1beta1.PodDisruptionBudget: Get https://localhost:8443/apis/policy/v1beta1/poddisruptionbudgets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700353 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StatefulSet: Get https://localhost:8443/apis/apps/v1/statefulsets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700372 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicaSet: Get https://localhost:8443/apis/apps/v1/replicasets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700394 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StorageClass: Get https://localhost:8443/apis/storage.k8s.io/v1/storageclasses?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.701981 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Node: Get https://localhost:8443/api/v1/nodes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") ``` **The operating system version**: `Linux Mint 19.1 Tessa`
1.0
none: Some fatal errors occurred: [ERROR Port-8443]: Port 8443 is in use - <!-- Please use this template while reporting an issue, providing as much information as possible. Failure to do so may result in a delayed response. Thank you! --> **The exact command to reproduce the issue**: ``` export MINIKUBE_WANTUPDATENOTIFICATION=false export MINIKUBE_WANTREPORTERRORPROMPT=false export CHANGE_MINIKUBE_NONE_USER=true sudo -E minikube start --vm-driver=none ``` **The full output of the command that failed**: ``` 😄 minikube v1.0.1 on linux (amd64) 🔥 Creating none VM (CPUs=2, Memory=2048MB, Disk=20000MB) ... 📶 "minikube" IP address is 192.168.8.159 🐳 Configuring Docker as the container runtime ... 🐳 Version of container runtime is 18.09.6 ✨ Preparing Kubernetes environment ... ❌ Unable to load cached images: loading cached images: loading image /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: stat /home/vadim/.minikube/cache/images/k8s.gcr.io/kube-proxy_v1.14.1: no such file or directory 💾 Downloading kubeadm v1.14.1 💾 Downloading kubelet v1.14.1 🚜 Pulling images required by Kubernetes v1.14.1 ... 🚀 Launching Kubernetes v1.14.1 using kubeadm ... 💣 Error starting cluster: kubeadm init: sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI : running command: sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI output: [init] Using Kubernetes version: v1.14.1 [preflight] Running pre-flight checks [WARNING FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml]: /etc/kubernetes/manifests/kube-apiserver.yaml already exists [WARNING FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml]: /etc/kubernetes/manifests/kube-controller-manager.yaml already exists [WARNING FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml]: /etc/kubernetes/manifests/kube-scheduler.yaml already exists [WARNING FileAvailable--etc-kubernetes-manifests-etcd.yaml]: /etc/kubernetes/manifests/etcd.yaml already exists [WARNING IsDockerSystemdCheck]: detected "cgroupfs" as the Docker cgroup driver. The recommended driver is "systemd". Please follow the guide at https://kubernetes.io/docs/setup/cri/ [WARNING Swap]: running with swap on is not supported. Please disable swap [WARNING FileExisting-ebtables]: ebtables not found in system path [WARNING FileExisting-ethtool]: ethtool not found in system path [WARNING FileExisting-socat]: socat not found in system path [WARNING Hostname]: hostname "minikube" could not be reached [WARNING Hostname]: hostname "minikube": lookup minikube on 127.0.0.53:53: server misbehaving [WARNING Service-Kubelet]: kubelet service is not enabled, please run 'systemctl enable kubelet.service' [WARNING Port-10250]: Port 10250 is in use [WARNING DirAvailable--data-minikube]: /data/minikube is not empty error execution phase preflight: [preflight] Some fatal errors occurred: [ERROR Port-8443]: Port 8443 is in use [ERROR Port-10251]: Port 10251 is in use [ERROR Port-10252]: Port 10252 is in use [ERROR Port-2379]: Port 2379 is in use [ERROR Port-2380]: Port 2380 is in use [preflight] If you know what you are doing, you can make a check non-fatal with `--ignore-preflight-errors=...` : running command: sudo /usr/bin/kubeadm init --config /var/lib/kubeadm.yaml --ignore-preflight-errors=DirAvailable--etc-kubernetes-manifests --ignore-preflight-errors=DirAvailable--data-minikube --ignore-preflight-errors=Port-10250 --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-scheduler.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-apiserver.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-kube-controller-manager.yaml --ignore-preflight-errors=FileAvailable--etc-kubernetes-manifests-etcd.yaml --ignore-preflight-errors=Swap --ignore-preflight-errors=CRI .: exit status 1 😿 Sorry that minikube crashed. If this was unexpected, we would love to hear from you: 👉 https://github.com/kubernetes/minikube/issues/new ❌ Problems detected in "kube-addon-manager": error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused ``` **The output of the `minikube logs` command**: ``` ==> dmesg <== [May11 16:28] secureboot: Secure boot could not be determined (mode 0) [ +0.000000] pmd_set_huge: Cannot satisfy [mem 0xf8000000-0xf8200000] with a huge-page mapping due to MTRR override. [ +1.100563] r8169 0000:02:00.0: can't disable ASPM; OS doesn't have ASPM control [ +0.333285] ata5.00: supports DRM functions and may not be fully accessible [ +0.000140] ata5.00: READ LOG DMA EXT failed, trying PIO [ +0.018473] ata5.00: supports DRM functions and may not be fully accessible [ +0.650433] usb 3-5.2: device descriptor read/64, error -32 [ +9.859326] kauditd_printk_skb: 58 callbacks suppressed [ +0.489483] aufs au_opts_verify:1609:dockerd[1937]: dirperm1 breaks the protection by the permission bits on the lower branch [May12 03:01] IRQ 16: no longer affine to CPU1 [ +0.000007] IRQ 29: no longer affine to CPU1 [ +0.024017] IRQ 23: no longer affine to CPU2 [ +0.000007] IRQ 27: no longer affine to CPU2 [ +0.032058] IRQ 26: no longer affine to CPU3 [ +0.000010] IRQ 28: no longer affine to CPU3 [ +0.012294] cache: parent cpu1 should not be sleeping [ +0.002224] cache: parent cpu2 should not be sleeping [ +0.002101] cache: parent cpu3 should not be sleeping [ +0.375470] ata5.00: supports DRM functions and may not be fully accessible [ +0.018803] ata5.00: supports DRM functions and may not be fully accessible [ +5.010804] ata1: link is slow to respond, please be patient (ready=0) [ +4.651973] ata1: COMRESET failed (errno=-16) [May12 06:52] sd 6:0:0:0: [sdc] No Caching mode page found [ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through [ +0.012436] sd 6:0:0:1: [sdd] No Caching mode page found [ +0.000009] sd 6:0:0:1: [sdd] Assuming drive cache: write through [May12 11:07] sd 6:0:0:0: [sdc] No Caching mode page found [ +0.000003] sd 6:0:0:0: [sdc] Assuming drive cache: write through [ +0.001174] sd 6:0:0:1: [sdd] No Caching mode page found [ +0.000003] sd 6:0:0:1: [sdd] Assuming drive cache: write through [May12 12:03] kauditd_printk_skb: 32 callbacks suppressed [May12 18:13] IRQ 16: no longer affine to CPU1 [ +0.000006] IRQ 29: no longer affine to CPU1 [ +0.024149] IRQ 23: no longer affine to CPU2 [ +0.000006] IRQ 27: no longer affine to CPU2 [ +0.032001] IRQ 26: no longer affine to CPU3 [ +0.000007] IRQ 28: no longer affine to CPU3 [ +0.011035] cache: parent cpu1 should not be sleeping [ +0.002241] cache: parent cpu2 should not be sleeping [ +0.002119] cache: parent cpu3 should not be sleeping [ +0.369614] ata5.00: supports DRM functions and may not be fully accessible [ +0.019113] ata5.00: supports DRM functions and may not be fully accessible [ +5.037481] ata1: link is slow to respond, please be patient (ready=0) [ +4.676005] ata1: COMRESET failed (errno=-16) [May13 02:54] kauditd_printk_skb: 37 callbacks suppressed ==> kernel <== 20:38:38 up 2 days, 4:10, 1 user, load average: 0.96, 0.86, 0.82 Linux desktop 4.18.0-18-generic #19~18.04.1-Ubuntu SMP Fri Apr 5 10:22:13 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux ==> kube-addon-manager <== INFO: == Kubernetes addon manager started at 2019-05-14T00:33:24+00:00 with ADDON_CHECK_INTERVAL_SEC=60 == error: unable to recognize "STDIN": Get https://localhost:8443/api?timeout=32s: dial tcp 127.0.0.1:8443: connect: connection refused WRN: == Failed to start /opt/namespace.yaml in namespace at 2019-05-14T00:33:24+00:00. 99 tries remaining. == INFO: == Default service account in the kube-system namespace has token default-token-hr5j9 == find: '/etc/kubernetes/admission-controls': No such file or directory INFO: == Entering periodical apply loop at 2019-05-14T00:33:29+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:33:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:33:31+00:00 == namespace/kube-system unchanged INFO: == Successfully started /opt/namespace.yaml in namespace at 2019-05-14T00:33:34+00:00 INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:34:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:34:30+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:35:30+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:35:31+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:36:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:36:31+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:37:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:37:31+00:00 == INFO: Leader is desktop INFO: == Kubernetes addon ensure completed at 2019-05-14T00:38:29+00:00 == INFO: == Reconciling with deprecated label == error: no objects passed to apply INFO: == Reconciling with addon-manager label == serviceaccount/storage-provisioner unchanged INFO: == Kubernetes addon reconcile completed at 2019-05-14T00:38:30+00:00 == ==> kube-apiserver <== I0514 00:38:34.831507 1 log.go:172] http: TLS handshake error from 127.0.0.1:49094: remote error: tls: bad certificate I0514 00:38:35.021488 1 log.go:172] http: TLS handshake error from 127.0.0.1:49096: remote error: tls: bad certificate I0514 00:38:35.235536 1 log.go:172] http: TLS handshake error from 127.0.0.1:49100: remote error: tls: bad certificate I0514 00:38:35.436229 1 log.go:172] http: TLS handshake error from 127.0.0.1:49102: remote error: tls: bad certificate I0514 00:38:35.632126 1 log.go:172] http: TLS handshake error from 127.0.0.1:49104: remote error: tls: bad certificate I0514 00:38:35.708998 1 log.go:172] http: TLS handshake error from 127.0.0.1:49106: remote error: tls: bad certificate I0514 00:38:35.711809 1 log.go:172] http: TLS handshake error from 127.0.0.1:49108: remote error: tls: bad certificate I0514 00:38:35.711850 1 log.go:172] http: TLS handshake error from 127.0.0.1:49110: remote error: tls: bad certificate I0514 00:38:35.711876 1 log.go:172] http: TLS handshake error from 127.0.0.1:49112: remote error: tls: bad certificate I0514 00:38:35.713858 1 log.go:172] http: TLS handshake error from 127.0.0.1:49114: remote error: tls: bad certificate I0514 00:38:35.714111 1 log.go:172] http: TLS handshake error from 127.0.0.1:49116: remote error: tls: bad certificate I0514 00:38:35.715785 1 log.go:172] http: TLS handshake error from 127.0.0.1:49118: remote error: tls: bad certificate I0514 00:38:35.716377 1 log.go:172] http: TLS handshake error from 127.0.0.1:49124: remote error: tls: bad certificate I0514 00:38:35.716620 1 log.go:172] http: TLS handshake error from 127.0.0.1:49120: remote error: tls: bad certificate I0514 00:38:35.716870 1 log.go:172] http: TLS handshake error from 127.0.0.1:49122: remote error: tls: bad certificate I0514 00:38:35.722733 1 log.go:172] http: TLS handshake error from 127.0.0.1:49126: remote error: tls: bad certificate I0514 00:38:35.840634 1 log.go:172] http: TLS handshake error from 127.0.0.1:49128: remote error: tls: bad certificate I0514 00:38:36.027863 1 log.go:172] http: TLS handshake error from 127.0.0.1:49130: remote error: tls: bad certificate I0514 00:38:36.243197 1 log.go:172] http: TLS handshake error from 127.0.0.1:49132: remote error: tls: bad certificate I0514 00:38:36.299517 1 log.go:172] http: TLS handshake error from 127.0.0.1:49134: remote error: tls: bad certificate I0514 00:38:36.446919 1 log.go:172] http: TLS handshake error from 127.0.0.1:49136: remote error: tls: bad certificate I0514 00:38:36.640301 1 log.go:172] http: TLS handshake error from 127.0.0.1:49138: remote error: tls: bad certificate I0514 00:38:36.717558 1 log.go:172] http: TLS handshake error from 127.0.0.1:49140: remote error: tls: bad certificate I0514 00:38:36.720112 1 log.go:172] http: TLS handshake error from 127.0.0.1:49150: remote error: tls: bad certificate I0514 00:38:36.720144 1 log.go:172] http: TLS handshake error from 127.0.0.1:49144: remote error: tls: bad certificate I0514 00:38:36.720168 1 log.go:172] http: TLS handshake error from 127.0.0.1:49142: remote error: tls: bad certificate I0514 00:38:36.721405 1 log.go:172] http: TLS handshake error from 127.0.0.1:49146: remote error: tls: bad certificate I0514 00:38:36.721835 1 log.go:172] http: TLS handshake error from 127.0.0.1:49152: remote error: tls: bad certificate I0514 00:38:36.721908 1 log.go:172] http: TLS handshake error from 127.0.0.1:49148: remote error: tls: bad certificate I0514 00:38:36.722280 1 log.go:172] http: TLS handshake error from 127.0.0.1:49154: remote error: tls: bad certificate I0514 00:38:36.723077 1 log.go:172] http: TLS handshake error from 127.0.0.1:49156: remote error: tls: bad certificate I0514 00:38:36.723116 1 log.go:172] http: TLS handshake error from 127.0.0.1:49158: remote error: tls: bad certificate I0514 00:38:36.843667 1 log.go:172] http: TLS handshake error from 127.0.0.1:49160: remote error: tls: bad certificate I0514 00:38:37.032639 1 log.go:172] http: TLS handshake error from 127.0.0.1:49164: remote error: tls: bad certificate I0514 00:38:37.253082 1 log.go:172] http: TLS handshake error from 127.0.0.1:49166: remote error: tls: bad certificate I0514 00:38:37.461185 1 log.go:172] http: TLS handshake error from 127.0.0.1:49168: remote error: tls: bad certificate I0514 00:38:37.647529 1 log.go:172] http: TLS handshake error from 127.0.0.1:49170: remote error: tls: bad certificate I0514 00:38:37.726883 1 log.go:172] http: TLS handshake error from 127.0.0.1:49172: remote error: tls: bad certificate I0514 00:38:37.731172 1 log.go:172] http: TLS handshake error from 127.0.0.1:49178: remote error: tls: bad certificate I0514 00:38:37.731797 1 log.go:172] http: TLS handshake error from 127.0.0.1:49174: remote error: tls: bad certificate I0514 00:38:37.731833 1 log.go:172] http: TLS handshake error from 127.0.0.1:49184: remote error: tls: bad certificate I0514 00:38:37.731857 1 log.go:172] http: TLS handshake error from 127.0.0.1:49182: remote error: tls: bad certificate I0514 00:38:37.731881 1 log.go:172] http: TLS handshake error from 127.0.0.1:49176: remote error: tls: bad certificate I0514 00:38:37.733898 1 log.go:172] http: TLS handshake error from 127.0.0.1:49186: remote error: tls: bad certificate I0514 00:38:37.733938 1 log.go:172] http: TLS handshake error from 127.0.0.1:49180: remote error: tls: bad certificate I0514 00:38:37.734391 1 log.go:172] http: TLS handshake error from 127.0.0.1:49190: remote error: tls: bad certificate I0514 00:38:37.734678 1 log.go:172] http: TLS handshake error from 127.0.0.1:49188: remote error: tls: bad certificate I0514 00:38:37.847115 1 log.go:172] http: TLS handshake error from 127.0.0.1:49194: remote error: tls: bad certificate I0514 00:38:38.035683 1 log.go:172] http: TLS handshake error from 127.0.0.1:49198: remote error: tls: bad certificate I0514 00:38:38.202140 1 log.go:172] http: TLS handshake error from 127.0.0.1:49200: remote error: tls: bad certificate ==> kube-scheduler <== E0514 00:38:33.696199 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolumeClaim: Get https://localhost:8443/api/v1/persistentvolumeclaims?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.696962 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.PersistentVolume: Get https://localhost:8443/api/v1/persistentvolumes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.696988 1 reflector.go:126] k8s.io/kubernetes/cmd/kube-scheduler/app/server.go:223: Failed to list *v1.Pod: Get https://localhost:8443/api/v1/pods?fieldSelector=status.phase%3DFailed%!C(MISSING)status.phase%3DSucceeded&limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.699704 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Service: Get https://localhost:8443/api/v1/services?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700091 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1beta1.PodDisruptionBudget: Get https://localhost:8443/apis/policy/v1beta1/poddisruptionbudgets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700353 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StatefulSet: Get https://localhost:8443/apis/apps/v1/statefulsets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700372 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.ReplicaSet: Get https://localhost:8443/apis/apps/v1/replicasets?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.700394 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.StorageClass: Get https://localhost:8443/apis/storage.k8s.io/v1/storageclasses?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") E0514 00:38:33.701981 1 reflector.go:126] k8s.io/client-go/informers/factory.go:133: Failed to list *v1.Node: Get https://localhost:8443/api/v1/nodes?limit=500&resourceVersion=0: x509: certificate signed by unknown authority (possibly because of "crypto/rsa: verification error" while trying to verify candidate authority certificate "minikubeCA") ``` **The operating system version**: `Linux Mint 19.1 Tessa`
non_infrastructure
none some fatal errors occurred port is in use the exact command to reproduce the issue export minikube wantupdatenotification false export minikube wantreporterrorprompt false export change minikube none user true sudo e minikube start vm driver none the full output of the command that failed 😄 minikube on linux 🔥 creating none vm cpus memory disk 📶 minikube ip address is 🐳 configuring docker as the container runtime 🐳 version of container runtime is ✨ preparing kubernetes environment ❌ unable to load cached images loading cached images loading image home vadim minikube cache images gcr io kube proxy stat home vadim minikube cache images gcr io kube proxy no such file or directory 💾 downloading kubeadm 💾 downloading kubelet 🚜 pulling images required by kubernetes 🚀 launching kubernetes using kubeadm 💣 error starting cluster kubeadm init sudo usr bin kubeadm init config var lib kubeadm yaml ignore preflight errors diravailable etc kubernetes manifests ignore preflight errors diravailable data minikube ignore preflight errors port ignore preflight errors fileavailable etc kubernetes manifests kube scheduler yaml ignore preflight errors fileavailable etc kubernetes manifests kube apiserver yaml ignore preflight errors fileavailable etc kubernetes manifests kube controller manager yaml ignore preflight errors fileavailable etc kubernetes manifests etcd yaml ignore preflight errors swap ignore preflight errors cri running command sudo usr bin kubeadm init config var lib kubeadm yaml ignore preflight errors diravailable etc kubernetes manifests ignore preflight errors diravailable data minikube ignore preflight errors port ignore preflight errors fileavailable etc kubernetes manifests kube scheduler yaml ignore preflight errors fileavailable etc kubernetes manifests kube apiserver yaml ignore preflight errors fileavailable etc kubernetes manifests kube controller manager yaml ignore preflight errors fileavailable etc kubernetes manifests etcd yaml ignore preflight errors swap ignore preflight errors cri output using kubernetes version running pre flight checks etc kubernetes manifests kube apiserver yaml already exists etc kubernetes manifests kube controller manager yaml already exists etc kubernetes manifests kube scheduler yaml already exists etc kubernetes manifests etcd yaml already exists detected cgroupfs as the docker cgroup driver the recommended driver is systemd please follow the guide at running with swap on is not supported please disable swap ebtables not found in system path ethtool not found in system path socat not found in system path hostname minikube could not be reached hostname minikube lookup minikube on server misbehaving kubelet service is not enabled please run systemctl enable kubelet service port is in use data minikube is not empty error execution phase preflight some fatal errors occurred port is in use port is in use port is in use port is in use port is in use if you know what you are doing you can make a check non fatal with ignore preflight errors running command sudo usr bin kubeadm init config var lib kubeadm yaml ignore preflight errors diravailable etc kubernetes manifests ignore preflight errors diravailable data minikube ignore preflight errors port ignore preflight errors fileavailable etc kubernetes manifests kube scheduler yaml ignore preflight errors fileavailable etc kubernetes manifests kube apiserver yaml ignore preflight errors fileavailable etc kubernetes manifests kube controller manager yaml ignore preflight errors fileavailable etc kubernetes manifests etcd yaml ignore preflight errors swap ignore preflight errors cri exit status 😿 sorry that minikube crashed if this was unexpected we would love to hear from you 👉 ❌ problems detected in kube addon manager error unable to recognize stdin get dial tcp connect connection refused the output of the minikube logs command dmesg secureboot secure boot could not be determined mode pmd set huge cannot satisfy with a huge page mapping due to mtrr override can t disable aspm os doesn t have aspm control supports drm functions and may not be fully accessible read log dma ext failed trying pio supports drm functions and may not be fully accessible usb device descriptor read error kauditd printk skb callbacks suppressed aufs au opts verify dockerd breaks the protection by the permission bits on the lower branch irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to cache parent should not be sleeping cache parent should not be sleeping cache parent should not be sleeping supports drm functions and may not be fully accessible supports drm functions and may not be fully accessible link is slow to respond please be patient ready comreset failed errno sd no caching mode page found sd assuming drive cache write through sd no caching mode page found sd assuming drive cache write through sd no caching mode page found sd assuming drive cache write through sd no caching mode page found sd assuming drive cache write through kauditd printk skb callbacks suppressed irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to irq no longer affine to cache parent should not be sleeping cache parent should not be sleeping cache parent should not be sleeping supports drm functions and may not be fully accessible supports drm functions and may not be fully accessible link is slow to respond please be patient ready comreset failed errno kauditd printk skb callbacks suppressed kernel up days user load average linux desktop generic ubuntu smp fri apr utc gnu linux kube addon manager info kubernetes addon manager started at with addon check interval sec error unable to recognize stdin get dial tcp connect connection refused wrn failed to start opt namespace yaml in namespace at tries remaining info default service account in the kube system namespace has token default token find etc kubernetes admission controls no such file or directory info entering periodical apply loop at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at namespace kube system unchanged info successfully started opt namespace yaml in namespace at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at info leader is desktop info kubernetes addon ensure completed at info reconciling with deprecated label error no objects passed to apply info reconciling with addon manager label serviceaccount storage provisioner unchanged info kubernetes addon reconcile completed at kube apiserver log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate log go http tls handshake error from remote error tls bad certificate kube scheduler reflector go io client go informers factory go failed to list persistentvolumeclaim get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list persistentvolume get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io kubernetes cmd kube scheduler app server go failed to list pod get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list service get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list poddisruptionbudget get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list statefulset get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list replicaset get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list storageclass get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca reflector go io client go informers factory go failed to list node get certificate signed by unknown authority possibly because of crypto rsa verification error while trying to verify candidate authority certificate minikubeca the operating system version linux mint tessa
0
24,046
7,451,634,520
IssuesEvent
2018-03-29 04:12:35
cerner/terra-core
https://api.github.com/repos/cerner/terra-core
closed
Update Express Dependency to ^4.16.0
build-tools
# Issue Description When visiting the `express` doc site, there is a notice stating: > Express 4.16.0 contains important security updates. For more information on what was added in this release, see the [4.16.0 changelog](http://expressjs.com/en/changelog/4x.html#4.16.0). ## Issue Type <!-- Is this a new feature request, enhancement, bug report, other? --> - [ ] New Feature - [ ] Enhancement - [ ] Bug - [x] Other
1.0
Update Express Dependency to ^4.16.0 - # Issue Description When visiting the `express` doc site, there is a notice stating: > Express 4.16.0 contains important security updates. For more information on what was added in this release, see the [4.16.0 changelog](http://expressjs.com/en/changelog/4x.html#4.16.0). ## Issue Type <!-- Is this a new feature request, enhancement, bug report, other? --> - [ ] New Feature - [ ] Enhancement - [ ] Bug - [x] Other
non_infrastructure
update express dependency to issue description when visiting the express doc site there is a notice stating express contains important security updates for more information on what was added in this release see the issue type new feature enhancement bug other
0
9,615
8,053,612,178
IssuesEvent
2018-08-02 00:01:47
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
opened
Change the different types of links to be a single [Link] attribute
interface/infrastructure refactor
Currently we have [Link], [LinkByPath], [ScopedLinkByName], [ScopedLink], [ChildLink], [ChildLinkByName], [ParentLink]. I think we need a single [Link] class with perhaps an enum argument e.g. [Link(type=LinkType.Path, path="xyz")] [Link(type=LinkType.Scoped, name="xyz")] [Link(type=LinkType.Child, name="xyz")] [Link(type=LinkType.Parent, name="xyz")]
1.0
Change the different types of links to be a single [Link] attribute - Currently we have [Link], [LinkByPath], [ScopedLinkByName], [ScopedLink], [ChildLink], [ChildLinkByName], [ParentLink]. I think we need a single [Link] class with perhaps an enum argument e.g. [Link(type=LinkType.Path, path="xyz")] [Link(type=LinkType.Scoped, name="xyz")] [Link(type=LinkType.Child, name="xyz")] [Link(type=LinkType.Parent, name="xyz")]
infrastructure
change the different types of links to be a single attribute currently we have i think we need a single class with perhaps an enum argument e g
1
1,690
3,343,384,822
IssuesEvent
2015-11-15 13:11:44
rackt/redux
https://api.github.com/repos/rackt/redux
closed
[React Native] Upgrading to Babel 6 and/or adding .babelrc to .npmignore
help wanted infrastructure
React Native will be shipping with Babel 6 in about three weeks. One consequence of this is that Babel finds .babelrc in `node_modules/redux` and tries to consume it -- since Redux is using Babel 5's configuration, this causes Babel 6 to crash. The current verdict is that packages will have to exclude .babelrc when distributed through npm: https://github.com/facebook/react-native/issues/4062.
1.0
[React Native] Upgrading to Babel 6 and/or adding .babelrc to .npmignore - React Native will be shipping with Babel 6 in about three weeks. One consequence of this is that Babel finds .babelrc in `node_modules/redux` and tries to consume it -- since Redux is using Babel 5's configuration, this causes Babel 6 to crash. The current verdict is that packages will have to exclude .babelrc when distributed through npm: https://github.com/facebook/react-native/issues/4062.
infrastructure
upgrading to babel and or adding babelrc to npmignore react native will be shipping with babel in about three weeks one consequence of this is that babel finds babelrc in node modules redux and tries to consume it since redux is using babel s configuration this causes babel to crash the current verdict is that packages will have to exclude babelrc when distributed through npm
1
32,661
26,877,842,432
IssuesEvent
2023-02-05 08:56:18
aimclub/FEDOT
https://api.github.com/repos/aimclub/FEDOT
closed
Make experiment for calculating overhead in using multiprocessing
research infrastructure
Perform experiment: 1) Run fitting n pipelines in multiprocessing mode 2) Run fitting n pipelines in single process mode 3) Calculate overhead for creating a process
1.0
Make experiment for calculating overhead in using multiprocessing - Perform experiment: 1) Run fitting n pipelines in multiprocessing mode 2) Run fitting n pipelines in single process mode 3) Calculate overhead for creating a process
infrastructure
make experiment for calculating overhead in using multiprocessing perform experiment run fitting n pipelines in multiprocessing mode run fitting n pipelines in single process mode calculate overhead for creating a process
1
30,901
25,153,802,446
IssuesEvent
2022-11-10 12:03:30
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Improve concurrency of `runtest.sh`
os-linux test-enhancement area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate
Our unix test script `runtest.sh` runs TCs in parallel. However it looks a little inefficient for some scenarios. For example, the script decides to run 4 processes concurrently. Then it starts first 4 tests. Let's number them 0 to 3. It waits for process 0. So if process 0 is long running but 1, 2 and 3 are not, process 0 could be a bottleneck. Even if 1, 2 and 3 are finished the script cannot run more processes before 0 is done. Because it waits for process sequentially. I think it would improve concurrency of unit tests if it can wait for any of child processes, not a in sequential order.
1.0
Improve concurrency of `runtest.sh` - Our unix test script `runtest.sh` runs TCs in parallel. However it looks a little inefficient for some scenarios. For example, the script decides to run 4 processes concurrently. Then it starts first 4 tests. Let's number them 0 to 3. It waits for process 0. So if process 0 is long running but 1, 2 and 3 are not, process 0 could be a bottleneck. Even if 1, 2 and 3 are finished the script cannot run more processes before 0 is done. Because it waits for process sequentially. I think it would improve concurrency of unit tests if it can wait for any of child processes, not a in sequential order.
infrastructure
improve concurrency of runtest sh our unix test script runtest sh runs tcs in parallel however it looks a little inefficient for some scenarios for example the script decides to run processes concurrently then it starts first tests let s number them to it waits for process so if process is long running but and are not process could be a bottleneck even if and are finished the script cannot run more processes before is done because it waits for process sequentially i think it would improve concurrency of unit tests if it can wait for any of child processes not a in sequential order
1
147,124
19,500,433,672
IssuesEvent
2021-12-28 01:31:33
talevy013/TestTal
https://api.github.com/repos/talevy013/TestTal
opened
CVE-2020-10968 (High) detected in jackson-databind-2.6.7.1.jar
security vulnerability
## CVE-2020-10968 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.7.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.7.1/jackson-databind-2.6.7.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.7.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/talevy013/TestTal/commit/f001f7f069d9289dded859e738eb111a8fd2e984">f001f7f069d9289dded859e738eb111a8fd2e984</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.aoju.bus.proxy.provider.remoting.RmiProvider (aka bus-proxy). <p>Publish Date: 2020-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10968>CVE-2020-10968</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10968">https://nvd.nist.gov/vuln/detail/CVE-2020-10968</a></p> <p>Release Date: 2020-03-26</p> <p>Fix Resolution: jackson-databind-2.9.10.4</p> </p> </details> <p></p>
True
CVE-2020-10968 (High) detected in jackson-databind-2.6.7.1.jar - ## CVE-2020-10968 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.6.7.1.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /m2/repository/com/fasterxml/jackson/core/jackson-databind/2.6.7.1/jackson-databind-2.6.7.1.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.6.7.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/talevy013/TestTal/commit/f001f7f069d9289dded859e738eb111a8fd2e984">f001f7f069d9289dded859e738eb111a8fd2e984</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.aoju.bus.proxy.provider.remoting.RmiProvider (aka bus-proxy). <p>Publish Date: 2020-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10968>CVE-2020-10968</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10968">https://nvd.nist.gov/vuln/detail/CVE-2020-10968</a></p> <p>Release Date: 2020-03-26</p> <p>Fix Resolution: jackson-databind-2.9.10.4</p> </p> </details> <p></p>
non_infrastructure
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org aoju bus proxy provider remoting rmiprovider aka bus proxy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind
0
1,382
3,179,709,491
IssuesEvent
2015-09-25 04:03:39
t3kt/vjzual2
https://api.github.com/repos/t3kt/vjzual2
opened
support modules disabling header buttons
enhancement infrastructure ui
like not showing the toggle for advanced parameters (or viewers) when the module doesn't have any
1.0
support modules disabling header buttons - like not showing the toggle for advanced parameters (or viewers) when the module doesn't have any
infrastructure
support modules disabling header buttons like not showing the toggle for advanced parameters or viewers when the module doesn t have any
1
27,765
30,338,864,610
IssuesEvent
2023-07-11 11:24:41
ICTU/quality-time
https://api.github.com/repos/ICTU/quality-time
closed
Notify the user why editing is not possible
Usability
It would be useful to notify the user when the user tries to edit a field when it's not editable. This can occur in the following situations: - the user is not logged in - the user is using the 'time machine' functionality where he/she looks into a view of the past, fields are not editable on purpose
True
Notify the user why editing is not possible - It would be useful to notify the user when the user tries to edit a field when it's not editable. This can occur in the following situations: - the user is not logged in - the user is using the 'time machine' functionality where he/she looks into a view of the past, fields are not editable on purpose
non_infrastructure
notify the user why editing is not possible it would be useful to notify the user when the user tries to edit a field when it s not editable this can occur in the following situations the user is not logged in the user is using the time machine functionality where he she looks into a view of the past fields are not editable on purpose
0
244,001
20,601,998,836
IssuesEvent
2022-03-06 12:10:36
einTosti/BuildSystem
https://api.github.com/repos/einTosti/BuildSystem
closed
`NullPointerException` while opening GameRule inventory
type: bug status: needs testing version: 1.18 priority: high
``` [12:50:35 ERROR]: Could not pass event InventoryClickEvent to BuildSystem v2.18 java.lang.NullPointerException: Cannot invoke "com.eintosti.buildsystem.version.GameRules.getInventory(org.bukkit.entity.Player, org.bukkit.World)" because "this.gameRules" is null at com.eintosti.buildsystem.inventory.GameRuleInventory.openInventory(GameRuleInventory.java:45) ~[BuildSystem-2.18.jar:?] at com.eintosti.buildsystem.inventory.EditInventory.onInventoryClick(EditInventory.java:285) ~[BuildSystem-2.18.jar:?] at com.destroystokyo.paper.event.executor.asm.generated.GeneratedEventExecutor72.execute(Unknown Source) ~[?:?] at org.bukkit.plugin.EventExecutor.lambda$create$1(EventExecutor.java:75) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?] at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:80) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:git-Paper-215] at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?] at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:629) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?] at net.minecraft.server.network.ServerGamePacketListenerImpl.handleContainerClick(ServerGamePacketListenerImpl.java:2838) ~[?:?] at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:58) ~[?:?] at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:13) ~[?:?] at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:56) ~[?:?] at net.minecraft.server.TickTask.run(TickTask.java:18) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?] at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?] at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1423) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.c(MinecraftServer.java:189) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?] at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1400) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1393) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:132) ~[?:?] at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1371) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1277) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:317) ~[paper-1.18.1.jar:git-Paper-215] at java.lang.Thread.run(Thread.java:833) ~[?:?] ```
1.0
`NullPointerException` while opening GameRule inventory - ``` [12:50:35 ERROR]: Could not pass event InventoryClickEvent to BuildSystem v2.18 java.lang.NullPointerException: Cannot invoke "com.eintosti.buildsystem.version.GameRules.getInventory(org.bukkit.entity.Player, org.bukkit.World)" because "this.gameRules" is null at com.eintosti.buildsystem.inventory.GameRuleInventory.openInventory(GameRuleInventory.java:45) ~[BuildSystem-2.18.jar:?] at com.eintosti.buildsystem.inventory.EditInventory.onInventoryClick(EditInventory.java:285) ~[BuildSystem-2.18.jar:?] at com.destroystokyo.paper.event.executor.asm.generated.GeneratedEventExecutor72.execute(Unknown Source) ~[?:?] at org.bukkit.plugin.EventExecutor.lambda$create$1(EventExecutor.java:75) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?] at co.aikar.timings.TimedEventExecutor.execute(TimedEventExecutor.java:80) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:git-Paper-215] at org.bukkit.plugin.RegisteredListener.callEvent(RegisteredListener.java:70) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?] at org.bukkit.plugin.SimplePluginManager.callEvent(SimplePluginManager.java:629) ~[paper-api-1.18.1-R0.1-SNAPSHOT.jar:?] at net.minecraft.server.network.ServerGamePacketListenerImpl.handleContainerClick(ServerGamePacketListenerImpl.java:2838) ~[?:?] at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:58) ~[?:?] at net.minecraft.network.protocol.game.ServerboundContainerClickPacket.handle(ServerboundContainerClickPacket.java:13) ~[?:?] at net.minecraft.network.protocol.PacketUtils.lambda$ensureRunningOnSameThread$1(PacketUtils.java:56) ~[?:?] at net.minecraft.server.TickTask.run(TickTask.java:18) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.util.thread.BlockableEventLoop.doRunTask(BlockableEventLoop.java:149) ~[?:?] at net.minecraft.util.thread.ReentrantBlockableEventLoop.doRunTask(ReentrantBlockableEventLoop.java:23) ~[?:?] at net.minecraft.server.MinecraftServer.doRunTask(MinecraftServer.java:1423) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.c(MinecraftServer.java:189) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.util.thread.BlockableEventLoop.pollTask(BlockableEventLoop.java:122) ~[?:?] at net.minecraft.server.MinecraftServer.pollTaskInternal(MinecraftServer.java:1400) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.pollTask(MinecraftServer.java:1393) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.util.thread.BlockableEventLoop.managedBlock(BlockableEventLoop.java:132) ~[?:?] at net.minecraft.server.MinecraftServer.waitUntilNextTick(MinecraftServer.java:1371) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.runServer(MinecraftServer.java:1277) ~[paper-1.18.1.jar:git-Paper-215] at net.minecraft.server.MinecraftServer.lambda$spin$0(MinecraftServer.java:317) ~[paper-1.18.1.jar:git-Paper-215] at java.lang.Thread.run(Thread.java:833) ~[?:?] ```
non_infrastructure
nullpointerexception while opening gamerule inventory could not pass event inventoryclickevent to buildsystem java lang nullpointerexception cannot invoke com eintosti buildsystem version gamerules getinventory org bukkit entity player org bukkit world because this gamerules is null at com eintosti buildsystem inventory gameruleinventory openinventory gameruleinventory java at com eintosti buildsystem inventory editinventory oninventoryclick editinventory java at com destroystokyo paper event executor asm generated execute unknown source at org bukkit plugin eventexecutor lambda create eventexecutor java at co aikar timings timedeventexecutor execute timedeventexecutor java at org bukkit plugin registeredlistener callevent registeredlistener java at org bukkit plugin simplepluginmanager callevent simplepluginmanager java at net minecraft server network servergamepacketlistenerimpl handlecontainerclick servergamepacketlistenerimpl java at net minecraft network protocol game serverboundcontainerclickpacket handle serverboundcontainerclickpacket java at net minecraft network protocol game serverboundcontainerclickpacket handle serverboundcontainerclickpacket java at net minecraft network protocol packetutils lambda ensurerunningonsamethread packetutils java at net minecraft server ticktask run ticktask java at net minecraft util thread blockableeventloop doruntask blockableeventloop java at net minecraft util thread reentrantblockableeventloop doruntask reentrantblockableeventloop java at net minecraft server minecraftserver doruntask minecraftserver java at net minecraft server minecraftserver c minecraftserver java at net minecraft util thread blockableeventloop polltask blockableeventloop java at net minecraft server minecraftserver polltaskinternal minecraftserver java at net minecraft server minecraftserver polltask minecraftserver java at net minecraft util thread blockableeventloop managedblock blockableeventloop java at net minecraft server minecraftserver waituntilnexttick minecraftserver java at net minecraft server minecraftserver runserver minecraftserver java at net minecraft server minecraftserver lambda spin minecraftserver java at java lang thread run thread java
0
130,151
18,042,014,278
IssuesEvent
2021-09-18 07:38:39
m93a/filtrex
https://api.github.com/repos/m93a/filtrex
closed
Make errors i18n-friendly
enhancement design implemented
All errors, both thrown and compile time and returned at runtime, should have an identifier which would make translation possible. It is not acceptable to show English errors in a non-English application.
1.0
Make errors i18n-friendly - All errors, both thrown and compile time and returned at runtime, should have an identifier which would make translation possible. It is not acceptable to show English errors in a non-English application.
non_infrastructure
make errors friendly all errors both thrown and compile time and returned at runtime should have an identifier which would make translation possible it is not acceptable to show english errors in a non english application
0
4,198
4,890,105,849
IssuesEvent
2016-11-18 12:41:39
SatelliteQE/robottelo
https://api.github.com/repos/SatelliteQE/robottelo
closed
Workaround upstream bug closure in downstream jobs
Infrastructure
Skip downstream tests if `bug status = 'verified' && QA whiteboard contains 'Verified in upstream'`.
1.0
Workaround upstream bug closure in downstream jobs - Skip downstream tests if `bug status = 'verified' && QA whiteboard contains 'Verified in upstream'`.
infrastructure
workaround upstream bug closure in downstream jobs skip downstream tests if bug status verified qa whiteboard contains verified in upstream
1
70,744
7,199,507,163
IssuesEvent
2018-02-05 16:08:50
ubtue/tuefind
https://api.github.com/repos/ubtue/tuefind
closed
Anzeige der "Weiteren Schlagwörter"
ready for testing
Bitte einleitende "gnd" usw. aus der Anzeige entfernen ![weitere_sww](https://user-images.githubusercontent.com/17004223/35636130-7e18724c-06b0-11e8-8c96-9b3af560eff6.png)
1.0
Anzeige der "Weiteren Schlagwörter" - Bitte einleitende "gnd" usw. aus der Anzeige entfernen ![weitere_sww](https://user-images.githubusercontent.com/17004223/35636130-7e18724c-06b0-11e8-8c96-9b3af560eff6.png)
non_infrastructure
anzeige der weiteren schlagwörter bitte einleitende gnd usw aus der anzeige entfernen
0
20,353
13,877,662,283
IssuesEvent
2020-10-17 05:22:53
esrlabs/chipmunk
https://api.github.com/repos/esrlabs/chipmunk
opened
Only produce artifacts we need in gh actions
infrastructure
Some generated artifacts are not needed and should not end up as artifacts at all ![image](https://user-images.githubusercontent.com/99541/96329037-65b5c380-1049-11eb-9fa8-b0f0b7e7a598.png)
1.0
Only produce artifacts we need in gh actions - Some generated artifacts are not needed and should not end up as artifacts at all ![image](https://user-images.githubusercontent.com/99541/96329037-65b5c380-1049-11eb-9fa8-b0f0b7e7a598.png)
infrastructure
only produce artifacts we need in gh actions some generated artifacts are not needed and should not end up as artifacts at all
1
83,092
3,626,320,370
IssuesEvent
2016-02-10 00:10:57
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
Make kubectl describers and resource printers more consistent
component/kubectl help-wanted kind/cleanup priority/P2 team/ux
Noticed when reviewing #20886 https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/describe.go Describers should all print object metadata consistently (name, namespace if relevant, labels). All controllers should print at least pod templates the same way -- using DescribePodTemplate. I'm sure there are other opportunities to improve consistency. I just noticed these differences at a glance. The resource printers (for get) are also inconsistent. @AdoHe has been improving on them (e.g., #20557, #20409). https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/resource_printer.go cc @kubernetes/kubectl @mqliang @erictune @mikedanese @janetkuo @kargakis @madhusudancs @AdoHe
1.0
Make kubectl describers and resource printers more consistent - Noticed when reviewing #20886 https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/describe.go Describers should all print object metadata consistently (name, namespace if relevant, labels). All controllers should print at least pod templates the same way -- using DescribePodTemplate. I'm sure there are other opportunities to improve consistency. I just noticed these differences at a glance. The resource printers (for get) are also inconsistent. @AdoHe has been improving on them (e.g., #20557, #20409). https://github.com/kubernetes/kubernetes/blob/master/pkg/kubectl/resource_printer.go cc @kubernetes/kubectl @mqliang @erictune @mikedanese @janetkuo @kargakis @madhusudancs @AdoHe
non_infrastructure
make kubectl describers and resource printers more consistent noticed when reviewing describers should all print object metadata consistently name namespace if relevant labels all controllers should print at least pod templates the same way using describepodtemplate i m sure there are other opportunities to improve consistency i just noticed these differences at a glance the resource printers for get are also inconsistent adohe has been improving on them e g cc kubernetes kubectl mqliang erictune mikedanese janetkuo kargakis madhusudancs adohe
0
13,279
10,180,241,540
IssuesEvent
2019-08-09 09:47:55
raiden-network/raiden
https://api.github.com/repos/raiden-network/raiden
opened
Check if "Decreasing linting errors" script works
infrastructure
I've never seen the "Decreasing linting errors" step in CI fail. Either we're really good at increasing mypy errors, or the script doesn't work as expected.
1.0
Check if "Decreasing linting errors" script works - I've never seen the "Decreasing linting errors" step in CI fail. Either we're really good at increasing mypy errors, or the script doesn't work as expected.
infrastructure
check if decreasing linting errors script works i ve never seen the decreasing linting errors step in ci fail either we re really good at increasing mypy errors or the script doesn t work as expected
1
3,884
4,691,431,734
IssuesEvent
2016-10-11 10:29:29
elmsln/elmsln
https://api.github.com/repos/elmsln/elmsln
closed
During Amazon spin up. PHP Warning: Error while sending QUERY packet
bug Developer Experience infrastructure medium minor testing
PHP Warning: Error while sending QUERY packet. PID=18550 in /var/www/elmsln/core/dslmcode/shared/drupal-7.x/modules/ulmus/apdqc/apdqc.mysql.inc on line 860 Not sure what this is will investigate at some point...
1.0
During Amazon spin up. PHP Warning: Error while sending QUERY packet - PHP Warning: Error while sending QUERY packet. PID=18550 in /var/www/elmsln/core/dslmcode/shared/drupal-7.x/modules/ulmus/apdqc/apdqc.mysql.inc on line 860 Not sure what this is will investigate at some point...
infrastructure
during amazon spin up php warning error while sending query packet php warning error while sending query packet pid in var www elmsln core dslmcode shared drupal x modules ulmus apdqc apdqc mysql inc on line not sure what this is will investigate at some point
1
26,354
20,022,491,403
IssuesEvent
2022-02-01 17:40:06
cloud-native-toolkit/automation-modules
https://api.github.com/repos/cloud-native-toolkit/automation-modules
closed
Request new module: cp-truststore
new_module type:gitops category:infrastructure approved
**Related user story** Provide the issue number for the user story related to this request - https://github.ibm.com/gsi-labs/industry-cloud-planning/issues/550 **Type of module** gitops **If requesting a terraform module, what is the target cloud platform?** N/A **What is the purpose of the module?** Please describe what this automation module will do **What is the category of the module?** infrastructure
1.0
Request new module: cp-truststore - **Related user story** Provide the issue number for the user story related to this request - https://github.ibm.com/gsi-labs/industry-cloud-planning/issues/550 **Type of module** gitops **If requesting a terraform module, what is the target cloud platform?** N/A **What is the purpose of the module?** Please describe what this automation module will do **What is the category of the module?** infrastructure
infrastructure
request new module cp truststore related user story provide the issue number for the user story related to this request type of module gitops if requesting a terraform module what is the target cloud platform n a what is the purpose of the module please describe what this automation module will do what is the category of the module infrastructure
1
27,627
22,045,308,504
IssuesEvent
2022-05-30 00:20:25
dart-lang/site-www
https://api.github.com/repos/dart-lang/site-www
closed
Consider extracting link check from stage/deploy to catch link issues in PRs
enhancement infrastructure p2-medium e1-hours links
### Describe the problem The link check is tied to the build/stage jobs, which don't run in every user's PRs. We should consider extracting it from those jobs so it will catch link issues for all pull requests. ### Expected fix Perhaps create a separate job to run the link checker. If possible avoid rebuilding the website again just for the link check. ### Additional context _No response_
1.0
Consider extracting link check from stage/deploy to catch link issues in PRs - ### Describe the problem The link check is tied to the build/stage jobs, which don't run in every user's PRs. We should consider extracting it from those jobs so it will catch link issues for all pull requests. ### Expected fix Perhaps create a separate job to run the link checker. If possible avoid rebuilding the website again just for the link check. ### Additional context _No response_
infrastructure
consider extracting link check from stage deploy to catch link issues in prs describe the problem the link check is tied to the build stage jobs which don t run in every user s prs we should consider extracting it from those jobs so it will catch link issues for all pull requests expected fix perhaps create a separate job to run the link checker if possible avoid rebuilding the website again just for the link check additional context no response
1
25,830
19,273,655,903
IssuesEvent
2021-12-10 09:18:40
chatwoot/chatwoot
https://api.github.com/repos/chatwoot/chatwoot
closed
dockerhub builds failing
infrastructure devops
**Describe the bug** Dockerhub builds are failing since today morning. **To Reproduce** Steps to reproduce the behavior: 1. Go to https://hub.docker.com/repository/docker/chatwoot/chatwoot/builds **Expected behavior** docker builds should pass **Screenshots** n/a **Browser logs** n/a **Server logs** n/a **Environment** n/a **Desktop (please complete the following information):** n/a **Smartphone (please complete the following information):** n/a **Additional context** n/a
1.0
dockerhub builds failing - **Describe the bug** Dockerhub builds are failing since today morning. **To Reproduce** Steps to reproduce the behavior: 1. Go to https://hub.docker.com/repository/docker/chatwoot/chatwoot/builds **Expected behavior** docker builds should pass **Screenshots** n/a **Browser logs** n/a **Server logs** n/a **Environment** n/a **Desktop (please complete the following information):** n/a **Smartphone (please complete the following information):** n/a **Additional context** n/a
infrastructure
dockerhub builds failing describe the bug dockerhub builds are failing since today morning to reproduce steps to reproduce the behavior go to expected behavior docker builds should pass screenshots n a browser logs n a server logs n a environment n a desktop please complete the following information n a smartphone please complete the following information n a additional context n a
1
25,583
18,877,176,314
IssuesEvent
2021-11-14 07:20:40
w3c/aria-practices
https://api.github.com/repos/w3c/aria-practices
closed
Remove dependency "npm-merge-driver"
Infrastructure
See #1766 When running `npm audit fix --force` it changes the version for `npm-merge-driver` to `^0.0.0`, which seems weird at first. Changing it back to `^2.3.6` and running `npm audit` again it finds 7 low severity vulnerabilities. I'd like to get `npm audit` to report 0 vulnerabilities and keep it there. Until `npm-merge-driver` gets a new version, I suggest we drop this dependency.
1.0
Remove dependency "npm-merge-driver" - See #1766 When running `npm audit fix --force` it changes the version for `npm-merge-driver` to `^0.0.0`, which seems weird at first. Changing it back to `^2.3.6` and running `npm audit` again it finds 7 low severity vulnerabilities. I'd like to get `npm audit` to report 0 vulnerabilities and keep it there. Until `npm-merge-driver` gets a new version, I suggest we drop this dependency.
infrastructure
remove dependency npm merge driver see when running npm audit fix force it changes the version for npm merge driver to which seems weird at first changing it back to and running npm audit again it finds low severity vulnerabilities i d like to get npm audit to report vulnerabilities and keep it there until npm merge driver gets a new version i suggest we drop this dependency
1
21,037
14,286,729,056
IssuesEvent
2020-11-23 15:30:02
sunpy/sunpy
https://api.github.com/repos/sunpy/sunpy
closed
Intermittent image-rotation test failures on OS X when using conda
Effort High Infrastructure Package Expert Priority High Tests image
We have intermittent failures of our image rotation tests and these failures appear to be isolated to OS X when using conda(-forge). #4235 added raw output for these failures occurred, and some of the output is truly bizarre. I will add investigative stuff in separate posts. My current conjecture is there is nothing wrong with SunPy code, but rather a C extension in the `numpy`/`scipy`/`scikit-image` ecosystem is not being compiled for conda(-forge) with the correct compile options for OS X such that there's the intermittent potential for bad memory access of arrays. --- Edit: go down to https://github.com/sunpy/sunpy/issues/4290#issuecomment-676573472 for a summary of the current understanding
1.0
Intermittent image-rotation test failures on OS X when using conda - We have intermittent failures of our image rotation tests and these failures appear to be isolated to OS X when using conda(-forge). #4235 added raw output for these failures occurred, and some of the output is truly bizarre. I will add investigative stuff in separate posts. My current conjecture is there is nothing wrong with SunPy code, but rather a C extension in the `numpy`/`scipy`/`scikit-image` ecosystem is not being compiled for conda(-forge) with the correct compile options for OS X such that there's the intermittent potential for bad memory access of arrays. --- Edit: go down to https://github.com/sunpy/sunpy/issues/4290#issuecomment-676573472 for a summary of the current understanding
infrastructure
intermittent image rotation test failures on os x when using conda we have intermittent failures of our image rotation tests and these failures appear to be isolated to os x when using conda forge added raw output for these failures occurred and some of the output is truly bizarre i will add investigative stuff in separate posts my current conjecture is there is nothing wrong with sunpy code but rather a c extension in the numpy scipy scikit image ecosystem is not being compiled for conda forge with the correct compile options for os x such that there s the intermittent potential for bad memory access of arrays edit go down to for a summary of the current understanding
1
17,319
12,300,903,737
IssuesEvent
2020-05-11 14:39:31
hotosm/tasking-manager
https://api.github.com/repos/hotosm/tasking-manager
closed
Maintenance site
Component: Frontend Component: Infrastructure Difficulty: Medium Priority: High Status: In Progress Type: Enhancement
When deploying changes for the TM now with the frontend being separated from the backend. The frontend usually persists, while the backend is not reachable. For the user it looks like the tool is working well, but it doesn't behave well. Lets find a solution to put the frontend in maintenance mode during deployments and also if there are any unexpected errors.
1.0
Maintenance site - When deploying changes for the TM now with the frontend being separated from the backend. The frontend usually persists, while the backend is not reachable. For the user it looks like the tool is working well, but it doesn't behave well. Lets find a solution to put the frontend in maintenance mode during deployments and also if there are any unexpected errors.
infrastructure
maintenance site when deploying changes for the tm now with the frontend being separated from the backend the frontend usually persists while the backend is not reachable for the user it looks like the tool is working well but it doesn t behave well lets find a solution to put the frontend in maintenance mode during deployments and also if there are any unexpected errors
1
34,964
30,635,149,049
IssuesEvent
2023-07-24 17:10:55
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Wasm.Build.Tests.Blazor.MiscTests.NativeBuild_WithDeployOnBuild_UsedByVS failing in CI due to SIGKILL
arch-wasm area-Infrastructure-mono
e.g. ``` [] Linking with emcc with -O0. This may take a while ... [] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-ld" @/datadisks/disk1/work/B6EC09B5/t/emscripten_3kui486x.rsp.utf-8 [] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize" -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features [] emcc : error : '/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features' failed (received SIGKILL (-9)) [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj] [] /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.WebAssembly.Sdk/8.0.0-ci/Sdk/WasmApp.Native.targets(431,5): error MSB3073: The command "emcc "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-default.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-link.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/emcc-link.rsp"" exited with code 1. [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj] ``` https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-78580-merge-c362e9e2c2db48c69b/Workloads-Wasm.Build.Tests.Blazor.MiscTests/1/console.d18bdf9a.log?helixlogtype=result
1.0
Wasm.Build.Tests.Blazor.MiscTests.NativeBuild_WithDeployOnBuild_UsedByVS failing in CI due to SIGKILL - e.g. ``` [] Linking with emcc with -O0. This may take a while ... [] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-ld" @/datadisks/disk1/work/B6EC09B5/t/emscripten_3kui486x.rsp.utf-8 [] "/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize" -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features [] emcc : error : '/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.Emscripten.3.1.12.Sdk.linux-x64/8.0.0-alpha.1.22558.2/tools/bin/wasm-emscripten-finalize -g --dyncalls-i64 --dwarf /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm -o /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/dotnet.wasm --detect-features' failed (received SIGKILL (-9)) [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj] [] /datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NET.Runtime.WebAssembly.Sdk/8.0.0-ci/Sdk/WasmApp.Native.targets(431,5): error MSB3073: The command "emcc "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-default.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/dotnet-net7+latest/packs/Microsoft.NETCore.App.Runtime.Mono.browser-wasm/8.0.0-ci/runtimes/browser-wasm/native/src/emcc-link.rsp" "@/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/obj/Debug/net7.0/wasm/for-publish/emcc-link.rsp"" exited with code 1. [/datadisks/disk1/work/B6EC09B5/w/9FD80866/e/wbt/blz_deploy_on_build_Debug_False_cv4azeoz.a0d/blz_deploy_on_build_Debug_False_cv4azeoz.a0d.csproj] ``` https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-78580-merge-c362e9e2c2db48c69b/Workloads-Wasm.Build.Tests.Blazor.MiscTests/1/console.d18bdf9a.log?helixlogtype=result
infrastructure
wasm build tests blazor misctests nativebuild withdeployonbuild usedbyvs failing in ci due to sigkill e g linking with emcc with this may take a while datadisks work w e dotnet latest packs microsoft net runtime emscripten sdk linux alpha tools bin wasm ld datadisks work t emscripten rsp utf datadisks work w e dotnet latest packs microsoft net runtime emscripten sdk linux alpha tools bin wasm emscripten finalize g dyncalls dwarf datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm o datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm detect features emcc error datadisks work w e dotnet latest packs microsoft net runtime emscripten sdk linux alpha tools bin wasm emscripten finalize g dyncalls dwarf datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm o datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish dotnet wasm detect features failed received sigkill datadisks work w e dotnet latest packs microsoft net runtime webassembly sdk ci sdk wasmapp native targets error the command emcc datadisks work w e dotnet latest packs microsoft netcore app runtime mono browser wasm ci runtimes browser wasm native src emcc default rsp datadisks work w e dotnet latest packs microsoft netcore app runtime mono browser wasm ci runtimes browser wasm native src emcc link rsp datadisks work w e wbt blz deploy on build debug false obj debug wasm for publish emcc link rsp exited with code
1
241,737
18,473,005,416
IssuesEvent
2021-10-18 01:28:23
edgexfoundry/edgex-docs
https://api.github.com/repos/edgexfoundry/edgex-docs
closed
Listing of supported device protocols
documentation enhancement help wanted good first issue 1-low jakarta
We need some documentation on the supported device services and protocols. In order not to have to keep updating the page with each new DS added, it would be ok to show how to find which device services we support (for example through GitHub). This question comes up a lot in Slack and via email forums. Need a good place to send people to see the list.
1.0
Listing of supported device protocols - We need some documentation on the supported device services and protocols. In order not to have to keep updating the page with each new DS added, it would be ok to show how to find which device services we support (for example through GitHub). This question comes up a lot in Slack and via email forums. Need a good place to send people to see the list.
non_infrastructure
listing of supported device protocols we need some documentation on the supported device services and protocols in order not to have to keep updating the page with each new ds added it would be ok to show how to find which device services we support for example through github this question comes up a lot in slack and via email forums need a good place to send people to see the list
0
10,856
8,759,715,714
IssuesEvent
2018-12-15 19:14:00
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Optimize performance of push notifications code in test suite
area: notifications (messages) area: testing-infrastructure good first issue help wanted in progress
If you add a `return` statement at the top of `handle_push_notifications`, running `test-backend test_notifications` goes from taking 4.3s to 3.6s on my laptop. I suspect there are other tests where we could save significant time as well, but it's hard to run the full test suite that way, because `handle_push_notifications` does useful stuff. What I'm thinking is that we can/should add an early-return in that function is `push_notification_enabled()` is False, and see if that helps improve the test suite performance. This issue doesn't require a ton of Zulip background, but it may require some good Python test system debugging skills (the tests that are likely to fail as one works on this involve a decent amount of mocks). @HarshitOnGitHub this may be an issue of interest for you when you have time again if nobody else does it first.
1.0
Optimize performance of push notifications code in test suite - If you add a `return` statement at the top of `handle_push_notifications`, running `test-backend test_notifications` goes from taking 4.3s to 3.6s on my laptop. I suspect there are other tests where we could save significant time as well, but it's hard to run the full test suite that way, because `handle_push_notifications` does useful stuff. What I'm thinking is that we can/should add an early-return in that function is `push_notification_enabled()` is False, and see if that helps improve the test suite performance. This issue doesn't require a ton of Zulip background, but it may require some good Python test system debugging skills (the tests that are likely to fail as one works on this involve a decent amount of mocks). @HarshitOnGitHub this may be an issue of interest for you when you have time again if nobody else does it first.
infrastructure
optimize performance of push notifications code in test suite if you add a return statement at the top of handle push notifications running test backend test notifications goes from taking to on my laptop i suspect there are other tests where we could save significant time as well but it s hard to run the full test suite that way because handle push notifications does useful stuff what i m thinking is that we can should add an early return in that function is push notification enabled is false and see if that helps improve the test suite performance this issue doesn t require a ton of zulip background but it may require some good python test system debugging skills the tests that are likely to fail as one works on this involve a decent amount of mocks harshitongithub this may be an issue of interest for you when you have time again if nobody else does it first
1
297,831
22,405,829,672
IssuesEvent
2022-06-18 00:56:59
Equipment-and-Tool-Institute/j1939-84
https://api.github.com/repos/Equipment-and-Tool-Institute/j1939-84
closed
Create and Affirm A-7 Tables for Minimum Test Results Support
documentation
Update A-7 Tables For 2024 MY Effects Create with user community and review with CARB. Table A-7 provides minimum test results content for DM4 and DM7. Continuous tests for Misfire and Fuel Quantity obviate discrete test results for DM7. 2019MY+ with continuous monitors are failing. See #1125 for posted background on misfire.
1.0
Create and Affirm A-7 Tables for Minimum Test Results Support - Update A-7 Tables For 2024 MY Effects Create with user community and review with CARB. Table A-7 provides minimum test results content for DM4 and DM7. Continuous tests for Misfire and Fuel Quantity obviate discrete test results for DM7. 2019MY+ with continuous monitors are failing. See #1125 for posted background on misfire.
non_infrastructure
create and affirm a tables for minimum test results support update a tables for my effects create with user community and review with carb table a provides minimum test results content for and continuous tests for misfire and fuel quantity obviate discrete test results for with continuous monitors are failing see for posted background on misfire
0
183,715
21,781,981,599
IssuesEvent
2022-05-13 20:04:30
kira0x1/mikaela
https://api.github.com/repos/kira0x1/mikaela
closed
discord.js-12.3.1.tgz: 1 vulnerabilities (highest severity is: 6.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>discord.js-12.3.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-0235](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | node-fetch-2.6.1.tgz | Transitive | 12.5.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0235</summary> ### Vulnerable Library - <b>node-fetch-2.6.1.tgz</b></p> <p>A light-weight module that brings window.fetch to node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p> <p> Dependency Hierarchy: - discord.js-12.3.1.tgz (Root Library) - :x: **node-fetch-2.6.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor <p>Publish Date: 2022-01-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235>CVE-2022-0235</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-r683-j2x4-v87g">https://github.com/advisories/GHSA-r683-j2x4-v87g</a></p> <p>Release Date: 2022-01-16</p> <p>Fix Resolution (node-fetch): 2.6.7</p> <p>Direct dependency fix Resolution (discord.js): 12.5.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"discord.js","packageVersion":"12.3.1","packageFilePaths":["/tmp/ws-scm/MikaelaBot/package.json"],"isTransitiveDependency":false,"dependencyTree":"discord.js:12.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"12.5.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-0235","vulnerabilityDetails":"node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> -->
True
discord.js-12.3.1.tgz: 1 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>discord.js-12.3.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2022-0235](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | node-fetch-2.6.1.tgz | Transitive | 12.5.3 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0235</summary> ### Vulnerable Library - <b>node-fetch-2.6.1.tgz</b></p> <p>A light-weight module that brings window.fetch to node.js</p> <p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-2.6.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/MikaelaBot/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/MikaelaBot/node_modules/node-fetch/package.json</p> <p> Dependency Hierarchy: - discord.js-12.3.1.tgz (Root Library) - :x: **node-fetch-2.6.1.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor <p>Publish Date: 2022-01-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235>CVE-2022-0235</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-r683-j2x4-v87g">https://github.com/advisories/GHSA-r683-j2x4-v87g</a></p> <p>Release Date: 2022-01-16</p> <p>Fix Resolution (node-fetch): 2.6.7</p> <p>Direct dependency fix Resolution (discord.js): 12.5.3</p> </p> <p></p> Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details> <!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"discord.js","packageVersion":"12.3.1","packageFilePaths":["/tmp/ws-scm/MikaelaBot/package.json"],"isTransitiveDependency":false,"dependencyTree":"discord.js:12.3.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"12.5.3","isBinary":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2022-0235","vulnerabilityDetails":"node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0235","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> -->
non_infrastructure
discord js tgz vulnerabilities highest severity is vulnerable library discord js tgz path to dependency file tmp ws scm mikaelabot package json path to vulnerable library tmp ws scm mikaelabot node modules node fetch package json vulnerabilities cve severity cvss dependency type fixed in remediation available medium node fetch tgz transitive details cve vulnerable library node fetch tgz a light weight module that brings window fetch to node js library home page a href path to dependency file tmp ws scm mikaelabot package json path to vulnerable library tmp ws scm mikaelabot node modules node fetch package json dependency hierarchy discord js tgz root library x node fetch tgz vulnerable library found in base branch main vulnerability details node fetch is vulnerable to exposure of sensitive information to an unauthorized actor publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node fetch direct dependency fix resolution discord js step up your open source security game with whitesource istransitivedependency false dependencytree discord js isminimumfixversionavailable true minimumfixversion isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails node fetch is vulnerable to exposure of sensitive information to an unauthorized actor vulnerabilityurl
0
255,019
8,102,473,709
IssuesEvent
2018-08-13 01:58:25
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
closed
Deleting error-report not replicating properly
Priority: 2 - Medium Status: 5 - Ready Type: Bug UI/UX
I deleted a report in couchdb while still open in the UI. Back to UI, the LHS still shows the report… **Steps to reproduce**: -_ Login `alpha.dev` as admin (or other instance)_ - _Create a report via sms with an invalid data: eg send sms (using admin console or otherwise): `1!C!125#Romeo` _ -_Notice that the report appears on the LHS in the UI ![image](https://user-images.githubusercontent.com/6979995/43229257-50e3c8fc-90b8-11e8-84a2-2b33612cb249.png) _ -_Click on the report and notice is opens in the RHS with correct error message_ ![image](https://user-images.githubusercontent.com/6979995/43229432-d4da66d4-90b8-11e8-80cf-ca4e43bd5ac0.png) _copy the report uuid, delete the doc in Futon (or maybe other non-UI ways to delete the report)_ _Look back at the UI_ **What should happen**: - _the report should be completely deleted from the UI_ -_The browser should redirect to `/_design/medic/_rewrite/#/reports/`_ **What actually happens**: - _The report still shows in the LHS (until a manual refresh)_ -_The url still has the report id_ ![image](https://user-images.githubusercontent.com/6979995/43229723-e423a21c-90b9-11e8-9bc9-d1d98b7627fa.png) - _Error in console_ ![image](https://user-images.githubusercontent.com/6979995/43229893-6bd3ef14-90ba-11e8-98ef-7b7b21176732.png) **Environment**: _Fill in as many of these as apply. It's especially useful if some environments work and some don't (ie works in Chrome but not in FF)._ - Instance: _(tested alpha.dev.medicmobile.org, and beta-old)_ - Browser: _(Firefox, Chrome)_ - Client platform: _(MacOS)_ - App: _(webapp)_ - Version: _(2.16.1-beta.1, 3.1.0)_ **Other**: _The UI `delete` works as expected in this case_
1.0
Deleting error-report not replicating properly - I deleted a report in couchdb while still open in the UI. Back to UI, the LHS still shows the report… **Steps to reproduce**: -_ Login `alpha.dev` as admin (or other instance)_ - _Create a report via sms with an invalid data: eg send sms (using admin console or otherwise): `1!C!125#Romeo` _ -_Notice that the report appears on the LHS in the UI ![image](https://user-images.githubusercontent.com/6979995/43229257-50e3c8fc-90b8-11e8-84a2-2b33612cb249.png) _ -_Click on the report and notice is opens in the RHS with correct error message_ ![image](https://user-images.githubusercontent.com/6979995/43229432-d4da66d4-90b8-11e8-80cf-ca4e43bd5ac0.png) _copy the report uuid, delete the doc in Futon (or maybe other non-UI ways to delete the report)_ _Look back at the UI_ **What should happen**: - _the report should be completely deleted from the UI_ -_The browser should redirect to `/_design/medic/_rewrite/#/reports/`_ **What actually happens**: - _The report still shows in the LHS (until a manual refresh)_ -_The url still has the report id_ ![image](https://user-images.githubusercontent.com/6979995/43229723-e423a21c-90b9-11e8-9bc9-d1d98b7627fa.png) - _Error in console_ ![image](https://user-images.githubusercontent.com/6979995/43229893-6bd3ef14-90ba-11e8-98ef-7b7b21176732.png) **Environment**: _Fill in as many of these as apply. It's especially useful if some environments work and some don't (ie works in Chrome but not in FF)._ - Instance: _(tested alpha.dev.medicmobile.org, and beta-old)_ - Browser: _(Firefox, Chrome)_ - Client platform: _(MacOS)_ - App: _(webapp)_ - Version: _(2.16.1-beta.1, 3.1.0)_ **Other**: _The UI `delete` works as expected in this case_
non_infrastructure
deleting error report not replicating properly i deleted a report in couchdb while still open in the ui back to ui the lhs still shows the report… steps to reproduce login alpha dev as admin or other instance create a report via sms with an invalid data eg send sms using admin console or otherwise c romeo notice that the report appears on the lhs in the ui click on the report and notice is opens in the rhs with correct error message copy the report uuid delete the doc in futon or maybe other non ui ways to delete the report look back at the ui what should happen the report should be completely deleted from the ui the browser should redirect to design medic rewrite reports what actually happens the report still shows in the lhs until a manual refresh the url still has the report id error in console environment fill in as many of these as apply it s especially useful if some environments work and some don t ie works in chrome but not in ff instance tested alpha dev medicmobile org and beta old browser firefox chrome client platform macos app webapp version beta other the ui delete works as expected in this case
0
87,341
3,750,058,792
IssuesEvent
2016-03-11 03:46:09
GLolol/PyLink
https://api.github.com/repos/GLolol/PyLink
opened
opercmds: 'trace'/'massban' command
feature priority:wishlist
Similar to Atheme contrib's `os_trace` module, this would be something that can match users by various criteria (`nick!user@host` globs or targets mentioned in #170) and then perform actions on them: `KILL`, `KLINE/ZLINE` (#139), `LIST`, etc.
1.0
opercmds: 'trace'/'massban' command - Similar to Atheme contrib's `os_trace` module, this would be something that can match users by various criteria (`nick!user@host` globs or targets mentioned in #170) and then perform actions on them: `KILL`, `KLINE/ZLINE` (#139), `LIST`, etc.
non_infrastructure
opercmds trace massban command similar to atheme contrib s os trace module this would be something that can match users by various criteria nick user host globs or targets mentioned in and then perform actions on them kill kline zline list etc
0
19,472
5,889,452,962
IssuesEvent
2017-05-17 12:56:51
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Menu reset in filter after creating new menu item
No Code Attached Yet
### Steps to reproduce the issue I have a menu with about 40 items. I go to Menus > My menu and see list of items: ![screen shot 2017-05-17 at 15 48 29](https://cloud.githubusercontent.com/assets/3432048/26154505/607cfd7a-3b18-11e7-83d3-1304be8041de.png) After creating new menu item I returning to the list of all items and no menu is selected: ![screen shot 2017-05-17 at 15 49 57](https://cloud.githubusercontent.com/assets/3432048/26154550/9bce9956-3b18-11e7-9e56-5cf70ee5ada7.png) That's very comfortless when you create mass menu items. ### Expected result After creating new menu item we must be returned to the same menu. ### Actual result I returning to All menu items. ### System information (as much as possible) Joomla 3.7.0 PHP Version 7.0.17
1.0
Menu reset in filter after creating new menu item - ### Steps to reproduce the issue I have a menu with about 40 items. I go to Menus > My menu and see list of items: ![screen shot 2017-05-17 at 15 48 29](https://cloud.githubusercontent.com/assets/3432048/26154505/607cfd7a-3b18-11e7-83d3-1304be8041de.png) After creating new menu item I returning to the list of all items and no menu is selected: ![screen shot 2017-05-17 at 15 49 57](https://cloud.githubusercontent.com/assets/3432048/26154550/9bce9956-3b18-11e7-9e56-5cf70ee5ada7.png) That's very comfortless when you create mass menu items. ### Expected result After creating new menu item we must be returned to the same menu. ### Actual result I returning to All menu items. ### System information (as much as possible) Joomla 3.7.0 PHP Version 7.0.17
non_infrastructure
menu reset in filter after creating new menu item steps to reproduce the issue i have a menu with about items i go to menus my menu and see list of items after creating new menu item i returning to the list of all items and no menu is selected that s very comfortless when you create mass menu items expected result after creating new menu item we must be returned to the same menu actual result i returning to all menu items system information as much as possible joomla php version
0
160,051
12,501,029,276
IssuesEvent
2020-06-02 00:00:09
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
opened
Flaky test PublicExposureOfPubternalTypeProducesPUB0001
test-investigation
https://dev.azure.com/dnceng/public/_build/results?buildId=663272&view=results PublicExposureOfPubternalTypeProducesPUB0001 ``` [xUnit.net 00:00:08.07] Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(member: "public /*MM*/CD c { get; }") [FAIL] [xUnit.net 00:00:08.08] The collection was expected to contain a single element, but it was empty. [xUnit.net 00:00:08.08] Stack Trace: [xUnit.net 00:00:08.08] /_/src/Analyzers/Internal.AspNetCore.Analyzers/test/PubternabilityAnalyzerTests.cs(47,0): at Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(String member) [xUnit.net 00:00:08.08] --- End of stack trace from previous location --- [xUnit.net 00:00:08.08] Output: [xUnit.net 00:00:08.08] Adding file: Test0 [xUnit.net 00:00:08.08] using A.Internal.Namespace; [xUnit.net 00:00:08.08] namespace A.Internal.Namespace [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public class C {} [xUnit.net 00:00:08.08] public delegate C CD (); [xUnit.net 00:00:08.08] public class CAAttribute: System.Attribute {} [xUnit.net 00:00:08.08] [xUnit.net 00:00:08.08] public class Program [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public static void Main() {} [xUnit.net 00:00:08.08] } [xUnit.net 00:00:08.08] } [xUnit.net 00:00:08.08] namespace A [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public class T [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public CD c { get; } [xUnit.net 00:00:08.08] } [xUnit.net 00:00:08.08] } ```
1.0
Flaky test PublicExposureOfPubternalTypeProducesPUB0001 - https://dev.azure.com/dnceng/public/_build/results?buildId=663272&view=results PublicExposureOfPubternalTypeProducesPUB0001 ``` [xUnit.net 00:00:08.07] Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(member: "public /*MM*/CD c { get; }") [FAIL] [xUnit.net 00:00:08.08] The collection was expected to contain a single element, but it was empty. [xUnit.net 00:00:08.08] Stack Trace: [xUnit.net 00:00:08.08] /_/src/Analyzers/Internal.AspNetCore.Analyzers/test/PubternabilityAnalyzerTests.cs(47,0): at Internal.AspNetCore.Analyzers.Tests.PubternabilityAnalyzerTests.PublicExposureOfPubternalTypeProducesPUB0001(String member) [xUnit.net 00:00:08.08] --- End of stack trace from previous location --- [xUnit.net 00:00:08.08] Output: [xUnit.net 00:00:08.08] Adding file: Test0 [xUnit.net 00:00:08.08] using A.Internal.Namespace; [xUnit.net 00:00:08.08] namespace A.Internal.Namespace [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public class C {} [xUnit.net 00:00:08.08] public delegate C CD (); [xUnit.net 00:00:08.08] public class CAAttribute: System.Attribute {} [xUnit.net 00:00:08.08] [xUnit.net 00:00:08.08] public class Program [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public static void Main() {} [xUnit.net 00:00:08.08] } [xUnit.net 00:00:08.08] } [xUnit.net 00:00:08.08] namespace A [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public class T [xUnit.net 00:00:08.08] { [xUnit.net 00:00:08.08] public CD c { get; } [xUnit.net 00:00:08.08] } [xUnit.net 00:00:08.08] } ```
non_infrastructure
flaky test internal aspnetcore analyzers tests pubternabilityanalyzertests member public mm cd c get the collection was expected to contain a single element but it was empty stack trace src analyzers internal aspnetcore analyzers test pubternabilityanalyzertests cs at internal aspnetcore analyzers tests pubternabilityanalyzertests string member end of stack trace from previous location output adding file using a internal namespace namespace a internal namespace public class c public delegate c cd public class caattribute system attribute public class program public static void main namespace a public class t public cd c get
0
2,228
3,576,276,087
IssuesEvent
2016-01-27 18:55:59
dotnet/roslyn-analyzers
https://api.github.com/repos/dotnet/roslyn-analyzers
closed
"Project Site" link on NuGet.org points to wrong site.
Area-Infrastructure Bug
For example, the "Project Site" link on https://www.nuget.org/packages/System.Runtime.InteropServices.Analyzers/ points to http://msdn.com/roslyn, which redirects to http://github.com/dotnet/roslyn. We should be directing users back to http://github.com/dotnet/roslyn-analyzers instead.
1.0
"Project Site" link on NuGet.org points to wrong site. - For example, the "Project Site" link on https://www.nuget.org/packages/System.Runtime.InteropServices.Analyzers/ points to http://msdn.com/roslyn, which redirects to http://github.com/dotnet/roslyn. We should be directing users back to http://github.com/dotnet/roslyn-analyzers instead.
infrastructure
project site link on nuget org points to wrong site for example the project site link on points to which redirects to we should be directing users back to instead
1
28,032
22,776,041,889
IssuesEvent
2022-07-08 14:33:17
Zilliqa/scilla
https://api.github.com/repos/Zilliqa/scilla
opened
Move to Ubuntu 20.04
infrastructure
Things that we need to change for this: - [ ] Docker and Docker.slim files - [ ] Make sure Travis CI uses Ubuntu 20.04 - [ ] Change `INSTALL.md` file accordingly (including the WSL installation)
1.0
Move to Ubuntu 20.04 - Things that we need to change for this: - [ ] Docker and Docker.slim files - [ ] Make sure Travis CI uses Ubuntu 20.04 - [ ] Change `INSTALL.md` file accordingly (including the WSL installation)
infrastructure
move to ubuntu things that we need to change for this docker and docker slim files make sure travis ci uses ubuntu change install md file accordingly including the wsl installation
1
30,986
25,225,542,303
IssuesEvent
2022-11-14 15:43:35
SonarSource/sonar-scanner-msbuild
https://api.github.com/repos/SonarSource/sonar-scanner-msbuild
closed
Fix broken ITs: Change tested plugin versions
Infrastructure
Integration tests uses LATEST_RELEASE of plugins with old SQ instances. The assumption that it should work is not valid anymore since the release strategy changed few years ago. We need to update Orchestration to: * use plugins that were shipped with the given version of SQ * add IT with LATEST_RELEASE of SQ and use analyzer LATEST_RELEASE only there
1.0
Fix broken ITs: Change tested plugin versions - Integration tests uses LATEST_RELEASE of plugins with old SQ instances. The assumption that it should work is not valid anymore since the release strategy changed few years ago. We need to update Orchestration to: * use plugins that were shipped with the given version of SQ * add IT with LATEST_RELEASE of SQ and use analyzer LATEST_RELEASE only there
infrastructure
fix broken its change tested plugin versions integration tests uses latest release of plugins with old sq instances the assumption that it should work is not valid anymore since the release strategy changed few years ago we need to update orchestration to use plugins that were shipped with the given version of sq add it with latest release of sq and use analyzer latest release only there
1
56,186
13,770,942,545
IssuesEvent
2020-10-07 21:04:58
GoogleCloudPlatform/ruby-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/ruby-docs-samples
closed
Minitest::Result: test_0009_in_query_without_array failed
:rotating_light: api: firestore buildcop: flaky buildcop: issue priority: p1 samples type: bug
This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: d52320406beaa73ab9b4e668d742acac43f33461 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/23be4cb1-8456-4efd-a366-3b5d38760de8), [Sponge](http://sponge2/23be4cb1-8456-4efd-a366-3b5d38760de8) status: failed <details><summary>Test output</summary><br><pre>/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:183:in `assert' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:264:in `assert_includes' /tmpfs/src/github/ruby-docs-samples/firestore/acceptance/query_data_test.rb:121:in `block (2 levels) in <top (required)>' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:98:in `block (3 levels) in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:195:in `capture_exceptions' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:95:in `block (2 levels) in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:272:in `time_it' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:94:in `block in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:211:in `with_info_handler' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:93:in `run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:1029:in `run_one_method' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:341:in `run_one_method' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:328:in `block (2 levels) in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `each' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `block in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:354:in `with_info_handler' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:326:in `run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `block in __run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `map' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `__run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:141:in `run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:68:in `block in autorun'</pre></details>
2.0
Minitest::Result: test_0009_in_query_without_array failed - This test failed! To configure my behavior, see [the Build Cop Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/buildcop). If I'm commenting on this issue too often, add the `buildcop: quiet` label and I will stop commenting. --- commit: d52320406beaa73ab9b4e668d742acac43f33461 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/23be4cb1-8456-4efd-a366-3b5d38760de8), [Sponge](http://sponge2/23be4cb1-8456-4efd-a366-3b5d38760de8) status: failed <details><summary>Test output</summary><br><pre>/usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:183:in `assert' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/assertions.rb:264:in `assert_includes' /tmpfs/src/github/ruby-docs-samples/firestore/acceptance/query_data_test.rb:121:in `block (2 levels) in <top (required)>' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:98:in `block (3 levels) in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:195:in `capture_exceptions' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:95:in `block (2 levels) in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:272:in `time_it' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:94:in `block in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:211:in `with_info_handler' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest/test.rb:93:in `run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:1029:in `run_one_method' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:341:in `run_one_method' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:328:in `block (2 levels) in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `each' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:327:in `block in run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:367:in `on_signal' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:354:in `with_info_handler' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:326:in `run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `block in __run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `map' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:164:in `__run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:141:in `run' /usr/local/bundle/gems/minitest-5.14.2/lib/minitest.rb:68:in `block in autorun'</pre></details>
non_infrastructure
minitest result test in query without array failed this test failed to configure my behavior see if i m commenting on this issue too often add the buildcop quiet label and i will stop commenting commit buildurl status failed test output usr local bundle gems minitest lib minitest assertions rb in assert usr local bundle gems minitest lib minitest assertions rb in assert includes tmpfs src github ruby docs samples firestore acceptance query data test rb in block levels in usr local bundle gems minitest lib minitest test rb in block levels in run usr local bundle gems minitest lib minitest test rb in capture exceptions usr local bundle gems minitest lib minitest test rb in block levels in run usr local bundle gems minitest lib minitest rb in time it usr local bundle gems minitest lib minitest test rb in block in run usr local bundle gems minitest lib minitest rb in on signal usr local bundle gems minitest lib minitest test rb in with info handler usr local bundle gems minitest lib minitest test rb in run usr local bundle gems minitest lib minitest rb in run one method usr local bundle gems minitest lib minitest rb in run one method usr local bundle gems minitest lib minitest rb in block levels in run usr local bundle gems minitest lib minitest rb in each usr local bundle gems minitest lib minitest rb in block in run usr local bundle gems minitest lib minitest rb in on signal usr local bundle gems minitest lib minitest rb in with info handler usr local bundle gems minitest lib minitest rb in run usr local bundle gems minitest lib minitest rb in block in run usr local bundle gems minitest lib minitest rb in map usr local bundle gems minitest lib minitest rb in run usr local bundle gems minitest lib minitest rb in run usr local bundle gems minitest lib minitest rb in block in autorun
0
538,502
15,770,299,488
IssuesEvent
2021-03-31 19:17:03
ArchitectSMP/k8s-gitops
https://api.github.com/repos/ArchitectSMP/k8s-gitops
closed
Proxy Service for improving American connection to server
enhancement feature/cluster priority/medium
Look into Proxy Service for improving American connection to server
1.0
Proxy Service for improving American connection to server - Look into Proxy Service for improving American connection to server
non_infrastructure
proxy service for improving american connection to server look into proxy service for improving american connection to server
0
48,013
5,931,076,637
IssuesEvent
2017-05-24 04:35:58
pump-io/pump.io
https://api.github.com/repos/pump-io/pump.io
opened
Make sure multiple tests don't have the same title
test suite
I just went to debug "User stream API test" and discovered that there are _three tests_ with this suite description. This seems really unfortunate.
1.0
Make sure multiple tests don't have the same title - I just went to debug "User stream API test" and discovered that there are _three tests_ with this suite description. This seems really unfortunate.
non_infrastructure
make sure multiple tests don t have the same title i just went to debug user stream api test and discovered that there are three tests with this suite description this seems really unfortunate
0
110,851
13,943,790,536
IssuesEvent
2020-10-23 00:06:04
1aurabrown/o2-monde-shopify
https://api.github.com/repos/1aurabrown/o2-monde-shopify
closed
Collection Page Re-Design
design
- [x] small text - [x] text blocks - similar to homepage quotes - [x] pagination - [x] product card: - [x] style on top center - [x] color just below - [x] price and sizes bottom - [x] option with "size" label - [x] text appears on hover - [x] preorder tag Mock Up States: - [x] sale - [x] some sizes sold out - [x] hover - [x] all sizes sold out - [x] sizes label - [x] 4px grid gutter option
1.0
Collection Page Re-Design - - [x] small text - [x] text blocks - similar to homepage quotes - [x] pagination - [x] product card: - [x] style on top center - [x] color just below - [x] price and sizes bottom - [x] option with "size" label - [x] text appears on hover - [x] preorder tag Mock Up States: - [x] sale - [x] some sizes sold out - [x] hover - [x] all sizes sold out - [x] sizes label - [x] 4px grid gutter option
non_infrastructure
collection page re design small text text blocks similar to homepage quotes pagination product card style on top center color just below price and sizes bottom option with size label text appears on hover preorder tag mock up states sale some sizes sold out hover all sizes sold out sizes label grid gutter option
0
29,116
23,729,578,497
IssuesEvent
2022-08-30 23:40:39
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
WorkspacesNetCore.MetadataReference should not need to wait for all asynchronous operations
Bug Area-Infrastructure Integration-Test
**Version Used**: #34265 The following wait operation should not be required: https://github.com/dotnet/roslyn/blob/8c5af921afd4fc1a4e949052571e6503b0ca8288/src/VisualStudio/IntegrationTest/IntegrationTests/Workspace/WorkspacesNetCore.cs#L42-L44 It should only need to wait for `FeatureAttribute.Workspace`.
1.0
WorkspacesNetCore.MetadataReference should not need to wait for all asynchronous operations - **Version Used**: #34265 The following wait operation should not be required: https://github.com/dotnet/roslyn/blob/8c5af921afd4fc1a4e949052571e6503b0ca8288/src/VisualStudio/IntegrationTest/IntegrationTests/Workspace/WorkspacesNetCore.cs#L42-L44 It should only need to wait for `FeatureAttribute.Workspace`.
infrastructure
workspacesnetcore metadatareference should not need to wait for all asynchronous operations version used the following wait operation should not be required it should only need to wait for featureattribute workspace
1
24,732
17,662,597,104
IssuesEvent
2021-08-21 20:32:25
pathfinder-for-autonomous-navigation/FlightSoftware
https://api.github.com/repos/pathfinder-for-autonomous-navigation/FlightSoftware
opened
CI Timing Out Due to Poor PTest Performance
bug ptest infrastructure
TL;DR; PTest sucks and isn't shutting down the flight software binary processes properly. Running all the mission checkouts locally on my desktop eventually pinned all 24 cores and grabbed 32 gigs of RAM plus 38 of swap crashing other processes running in the background -- like my chrome tabs. Upgrade PTest to properly shut things down.
1.0
CI Timing Out Due to Poor PTest Performance - TL;DR; PTest sucks and isn't shutting down the flight software binary processes properly. Running all the mission checkouts locally on my desktop eventually pinned all 24 cores and grabbed 32 gigs of RAM plus 38 of swap crashing other processes running in the background -- like my chrome tabs. Upgrade PTest to properly shut things down.
infrastructure
ci timing out due to poor ptest performance tl dr ptest sucks and isn t shutting down the flight software binary processes properly running all the mission checkouts locally on my desktop eventually pinned all cores and grabbed gigs of ram plus of swap crashing other processes running in the background like my chrome tabs upgrade ptest to properly shut things down
1
293,926
9,011,172,812
IssuesEvent
2019-02-05 14:04:19
raster-foundry/raster-foundry-python-client
https://api.github.com/repos/raster-foundry/raster-foundry-python-client
opened
Add example of creating uploads from existing S3 bucket
client requested priority
A lot of times we have a set of files in S3 (hundreds or thousands) that we want to import, but there isn't a great way or an example for how to do this.
1.0
Add example of creating uploads from existing S3 bucket - A lot of times we have a set of files in S3 (hundreds or thousands) that we want to import, but there isn't a great way or an example for how to do this.
non_infrastructure
add example of creating uploads from existing bucket a lot of times we have a set of files in hundreds or thousands that we want to import but there isn t a great way or an example for how to do this
0
20,515
13,961,869,926
IssuesEvent
2020-10-25 06:17:53
SirJosh3917/smiley-face-game
https://api.github.com/repos/SirJosh3917/smiley-face-game
opened
Allow users to change their password
client enhancement infrastructure server ui
This would require - Way to send emails to the afflicted account email - Generate short-living JWT specifically for resetting a password - UI for clicking "reset my password", and UI for resetting the password
1.0
Allow users to change their password - This would require - Way to send emails to the afflicted account email - Generate short-living JWT specifically for resetting a password - UI for clicking "reset my password", and UI for resetting the password
infrastructure
allow users to change their password this would require way to send emails to the afflicted account email generate short living jwt specifically for resetting a password ui for clicking reset my password and ui for resetting the password
1
9,186
7,858,093,440
IssuesEvent
2018-06-21 12:59:31
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
serve/warns_on_assets_paths_test is failing on the Windows bot, but not locally
Type: bug area-infrastructure
I just checked in a new test for pub. It passes everywhere, including on my local Windows machine. But it's failing on the bot with: &nbsp;Results of running 'pub serve': &nbsp;&nbsp;Process exited with exit code -1. &nbsp;&nbsp;Standard output: &nbsp;&nbsp;| MSG : Serving myapp on http://127.0.0.1:59461 &nbsp;&nbsp;| MSG : Build completed successfully &nbsp;&nbsp;Standard error: &nbsp;&nbsp;| FINE: Pub 0.1.2+3 &nbsp;&nbsp;| IO : Spawning cmd /c ver &nbsp;&nbsp;| IO : Finished ver. Exit code 0.stdout:Microsoft Windows [Version 6.1.7601]Nothing output on stderr. &nbsp;&nbsp;| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web. &nbsp;&nbsp;| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo. &nbsp;&nbsp;| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo\assets. I didn't think a negative exit code was even possible. As you can see, no exception is thrown. It looks like the process just dies. http://chromegw.corp.google.com/i/client.dart/builders/pub-win-be/builds/2695/steps/pub%20tests/logs/stdio Any idea what's going on here?
1.0
serve/warns_on_assets_paths_test is failing on the Windows bot, but not locally - I just checked in a new test for pub. It passes everywhere, including on my local Windows machine. But it's failing on the bot with: &nbsp;Results of running 'pub serve': &nbsp;&nbsp;Process exited with exit code -1. &nbsp;&nbsp;Standard output: &nbsp;&nbsp;| MSG : Serving myapp on http://127.0.0.1:59461 &nbsp;&nbsp;| MSG : Build completed successfully &nbsp;&nbsp;Standard error: &nbsp;&nbsp;| FINE: Pub 0.1.2+3 &nbsp;&nbsp;| IO : Spawning cmd /c ver &nbsp;&nbsp;| IO : Finished ver. Exit code 0.stdout:Microsoft Windows [Version 6.1.7601]Nothing output on stderr. &nbsp;&nbsp;| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web. &nbsp;&nbsp;| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo. &nbsp;&nbsp;| IO : Listing directory C:\Users\CHROME~2\AppData\Local\Temp\pub_3ddf4566-68f5-11e3-adfc-005056030208\myapp\web\foo\assets. I didn't think a negative exit code was even possible. As you can see, no exception is thrown. It looks like the process just dies. http://chromegw.corp.google.com/i/client.dart/builders/pub-win-be/builds/2695/steps/pub%20tests/logs/stdio Any idea what's going on here?
infrastructure
serve warns on assets paths test is failing on the windows bot but not locally i just checked in a new test for pub it passes everywhere including on my local windows machine but it s failing on the bot with nbsp results of running pub serve nbsp nbsp process exited with exit code nbsp nbsp standard output nbsp nbsp msg serving myapp on nbsp nbsp msg build completed successfully nbsp nbsp standard error nbsp nbsp fine pub nbsp nbsp io spawning cmd c ver nbsp nbsp io finished ver exit code stdout microsoft windows nothing output on stderr nbsp nbsp io listing directory c users chrome appdata local temp pub adfc myapp web nbsp nbsp io listing directory c users chrome appdata local temp pub adfc myapp web foo nbsp nbsp io listing directory c users chrome appdata local temp pub adfc myapp web foo assets i didn t think a negative exit code was even possible as you can see no exception is thrown it looks like the process just dies any idea what s going on here
1
225,889
17,928,905,050
IssuesEvent
2021-09-10 06:16:38
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
opened
Fix intermittent debugger test failure in beta3 staging branch
Type/Task Area/Debugger Area/IntegrationTest Points/0.25 SwanLakeDump
**Description:** The following debugger integration test case seems to be failing intermittently due to an timeout issue. ![Screenshot 2021-09-10 at 11 43 30](https://user-images.githubusercontent.com/29032600/132808470-54c81ab9-a9ae-4bca-9786-a3dbfff935db.png) **Describe your task(s)** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
Fix intermittent debugger test failure in beta3 staging branch - **Description:** The following debugger integration test case seems to be failing intermittently due to an timeout issue. ![Screenshot 2021-09-10 at 11 43 30](https://user-images.githubusercontent.com/29032600/132808470-54c81ab9-a9ae-4bca-9786-a3dbfff935db.png) **Describe your task(s)** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
non_infrastructure
fix intermittent debugger test failure in staging branch description the following debugger integration test case seems to be failing intermittently due to an timeout issue describe your task s related issues optional suggested labels optional suggested assignees optional
0
781,644
27,445,126,303
IssuesEvent
2023-03-02 13:48:21
Testaustime/testaustime-frontend
https://api.github.com/repos/Testaustime/testaustime-frontend
closed
If there is no data, the day chart is broken
bug priority-medium
The chart should show 0s for all days (a line at the bottom instead of middle) ![image](https://user-images.githubusercontent.com/38920928/210190551-945bbb93-371f-4d05-992a-1a759b4a38f5.png)
1.0
If there is no data, the day chart is broken - The chart should show 0s for all days (a line at the bottom instead of middle) ![image](https://user-images.githubusercontent.com/38920928/210190551-945bbb93-371f-4d05-992a-1a759b4a38f5.png)
non_infrastructure
if there is no data the day chart is broken the chart should show for all days a line at the bottom instead of middle
0
48,731
13,184,725,713
IssuesEvent
2020-08-12 19:58:54
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py (Trac #110)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/110 , reported by icecube and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2007-09-06T17:23:26", "description": "I see a different output running in icerec (trunk, RC 1-6-0)\n'make test -C TWRDistillery' --> Failed\n(Troy recommended 'ctest -VV -R TWRDistillery::TWRrecoTest' --> Failed with \n/Users/schlenst/prog/ice/icerec/trunk/src/TWRDistillery/private/twr-distillery/I3TWRReTrigger.cxx:127: FATAL: No DMADD file) \nand \n'python TWRDistillery/resources/scripts/TWRrecoTest.py' --> OK\n\nWhats wrong? Thanks, Stefan", "reporter": "icecube", "cc": "", "resolution": "invalid", "_ts": "1189099406000000", "component": "cmake", "summary": "make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py", "priority": "normal", "keywords": "running tests", "time": "2007-09-06T15:05:10", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py (Trac #110) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/110 , reported by icecube and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2007-09-06T17:23:26", "description": "I see a different output running in icerec (trunk, RC 1-6-0)\n'make test -C TWRDistillery' --> Failed\n(Troy recommended 'ctest -VV -R TWRDistillery::TWRrecoTest' --> Failed with \n/Users/schlenst/prog/ice/icerec/trunk/src/TWRDistillery/private/twr-distillery/I3TWRReTrigger.cxx:127: FATAL: No DMADD file) \nand \n'python TWRDistillery/resources/scripts/TWRrecoTest.py' --> OK\n\nWhats wrong? Thanks, Stefan", "reporter": "icecube", "cc": "", "resolution": "invalid", "_ts": "1189099406000000", "component": "cmake", "summary": "make test -C TWRDistillery and python TWRDistillery/resources/scripts/TWRrecoTest.py", "priority": "normal", "keywords": "running tests", "time": "2007-09-06T15:05:10", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
non_infrastructure
make test c twrdistillery and python twrdistillery resources scripts twrrecotest py trac migrated from reported by icecube and owned by troy json status closed changetime description i see a different output running in icerec trunk rc n make test c twrdistillery failed n troy recommended ctest vv r twrdistillery twrrecotest failed with n users schlenst prog ice icerec trunk src twrdistillery private twr distillery cxx fatal no dmadd file nand n python twrdistillery resources scripts twrrecotest py ok n nwhats wrong thanks stefan reporter icecube cc resolution invalid ts component cmake summary make test c twrdistillery and python twrdistillery resources scripts twrrecotest py priority normal keywords running tests time milestone owner troy type defect
0
69,324
8,393,438,250
IssuesEvent
2018-10-09 20:33:36
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
icon on Brave Rewards Panel needs link or purpose
QA/Test-Plan-Specified QA/Yes design feature/rewards
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. NOTE THAT THIS IS THE REPOSITORY FOR THE UPCOMING VERSION OF BRAVE. SEE [browser-laptop](https://github.com/brave/browser-laptop)FOR THE CURRENT PRODUCTION VERSION OF BRAVE ON MACOS, WINDOWS AND LINUX. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description There is an icon of an exclamation point in a circle on the Brave Rewards Panel that doesn't seem to have a purpose or link (it's not mentioned in the spec but it is implemented). Need to determine what this icon should link to or what it's purpose should be (if it's not always visible, when does it display?). ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. If not enabled, enable Rewards. 2. Navigate to a site. 3. Click on the BAT logo. ## Actual result: Next to 'Include in Brave Auto-Contribute' text note the icon. Clicking on it doesn't do anything. ![screen shot 2018-10-09 at 4 07 17 pm](https://user-images.githubusercontent.com/28145373/46696715-d5824f80-cbe0-11e8-8b8b-bfe0c7e68346.png) ## Expected result: Clicking on the icon or hovering over it should provide some information. Unless it's not always supposed to display in which case this should be documented. ## Reproduces how often: easy ## Brave version (chrome://version info) Brave | 0.55.11 Chromium: 70.0.3538.35 (Official Build) beta(64-bit) -- | -- Revision | 28dcb499844fa40c28d5f62e337876cb936f79f5-refs/branch-heads/3538@{#678} OS | Mac OS X ### Reproducible on current release: - Does it reproduce on brave-browser dev/beta builds? yes - Does it reproduce on browser-laptop? n/a ### Website problems only: - Does the issue resolve itself when disabling Brave Shields? n/a - Is the issue reproducible on the latest version of Chrome? n/a ### Additional Information
1.0
icon on Brave Rewards Panel needs link or purpose - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. NOTE THAT THIS IS THE REPOSITORY FOR THE UPCOMING VERSION OF BRAVE. SEE [browser-laptop](https://github.com/brave/browser-laptop)FOR THE CURRENT PRODUCTION VERSION OF BRAVE ON MACOS, WINDOWS AND LINUX. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description There is an icon of an exclamation point in a circle on the Brave Rewards Panel that doesn't seem to have a purpose or link (it's not mentioned in the spec but it is implemented). Need to determine what this icon should link to or what it's purpose should be (if it's not always visible, when does it display?). ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. If not enabled, enable Rewards. 2. Navigate to a site. 3. Click on the BAT logo. ## Actual result: Next to 'Include in Brave Auto-Contribute' text note the icon. Clicking on it doesn't do anything. ![screen shot 2018-10-09 at 4 07 17 pm](https://user-images.githubusercontent.com/28145373/46696715-d5824f80-cbe0-11e8-8b8b-bfe0c7e68346.png) ## Expected result: Clicking on the icon or hovering over it should provide some information. Unless it's not always supposed to display in which case this should be documented. ## Reproduces how often: easy ## Brave version (chrome://version info) Brave | 0.55.11 Chromium: 70.0.3538.35 (Official Build) beta(64-bit) -- | -- Revision | 28dcb499844fa40c28d5f62e337876cb936f79f5-refs/branch-heads/3538@{#678} OS | Mac OS X ### Reproducible on current release: - Does it reproduce on brave-browser dev/beta builds? yes - Does it reproduce on browser-laptop? n/a ### Website problems only: - Does the issue resolve itself when disabling Brave Shields? n/a - Is the issue reproducible on the latest version of Chrome? n/a ### Additional Information
non_infrastructure
icon on brave rewards panel needs link or purpose have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue note that this is the repository for the upcoming version of brave see the current production version of brave on macos windows and linux please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description there is an icon of an exclamation point in a circle on the brave rewards panel that doesn t seem to have a purpose or link it s not mentioned in the spec but it is implemented need to determine what this icon should link to or what it s purpose should be if it s not always visible when does it display steps to reproduce if not enabled enable rewards navigate to a site click on the bat logo actual result next to include in brave auto contribute text note the icon clicking on it doesn t do anything expected result clicking on the icon or hovering over it should provide some information unless it s not always supposed to display in which case this should be documented reproduces how often easy brave version chrome version info brave chromium   official build  beta bit revision refs branch heads os mac os x reproducible on current release does it reproduce on brave browser dev beta builds yes does it reproduce on browser laptop n a website problems only does the issue resolve itself when disabling brave shields n a is the issue reproducible on the latest version of chrome n a additional information
0
675,371
23,092,430,126
IssuesEvent
2022-07-26 16:15:13
Kong/kubernetes-ingress-controller
https://api.github.com/repos/Kong/kubernetes-ingress-controller
closed
Regression Testing with Kuma
priority/medium area/ci
### Is there an existing issue for this? - [X] I have searched the existing issues ### Problem Statement Changes to helm chart will occasionally break deployment of KIC when a service mesh is present. Having a regression tests greatly reduces risk of breaking changes. ### Proposed Solution Create an automated deployment of GW, KIC and Kuma when a PR is created. If the environment stands up, proceed to next phase of the release cycle. ### Additional information - [ ] needs https://github.com/Kong/kubernetes-testing-framework/issues/303 ### Acceptance Criteria As an end user, I consider it acceptable when you can deploy GW, KIC and Kuma without having to make unreasonable changes to the default values.yaml.
1.0
Regression Testing with Kuma - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Problem Statement Changes to helm chart will occasionally break deployment of KIC when a service mesh is present. Having a regression tests greatly reduces risk of breaking changes. ### Proposed Solution Create an automated deployment of GW, KIC and Kuma when a PR is created. If the environment stands up, proceed to next phase of the release cycle. ### Additional information - [ ] needs https://github.com/Kong/kubernetes-testing-framework/issues/303 ### Acceptance Criteria As an end user, I consider it acceptable when you can deploy GW, KIC and Kuma without having to make unreasonable changes to the default values.yaml.
non_infrastructure
regression testing with kuma is there an existing issue for this i have searched the existing issues problem statement changes to helm chart will occasionally break deployment of kic when a service mesh is present having a regression tests greatly reduces risk of breaking changes proposed solution create an automated deployment of gw kic and kuma when a pr is created if the environment stands up proceed to next phase of the release cycle additional information needs acceptance criteria as an end user i consider it acceptable when you can deploy gw kic and kuma without having to make unreasonable changes to the default values yaml
0
8,282
7,324,877,288
IssuesEvent
2018-03-03 01:41:44
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
[uap] CoreFx build is failing since 'RemotelyInvokable' does not contain a definition for 'LongWait'
area-Infrastructure
https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1436514 ```text 2018-03-02T21:25:47.5255290Z Build FAILED. 2018-03-02T21:25:47.5268140Z 2018-03-02T21:25:47.5269301Z ProcessTestBase.Uap.cs(68,57): error CS0117: 'RemotelyInvokable' does not contain a definition for 'LongWait' [E:\A\_work\36\s\corefx\src\System.Diagnostics.Process\tests\System.Diagnostics.Process.Tests.csproj] 2018-03-02T21:25:47.5272048Z 0 Warning(s) 2018-03-02T21:25:47.5273015Z 1 Error(s) ``` cc @jkotas, @stephentoub
1.0
[uap] CoreFx build is failing since 'RemotelyInvokable' does not contain a definition for 'LongWait' - https://devdiv.visualstudio.com/DefaultCollection/DevDiv/_build?_a=summary&buildId=1436514 ```text 2018-03-02T21:25:47.5255290Z Build FAILED. 2018-03-02T21:25:47.5268140Z 2018-03-02T21:25:47.5269301Z ProcessTestBase.Uap.cs(68,57): error CS0117: 'RemotelyInvokable' does not contain a definition for 'LongWait' [E:\A\_work\36\s\corefx\src\System.Diagnostics.Process\tests\System.Diagnostics.Process.Tests.csproj] 2018-03-02T21:25:47.5272048Z 0 Warning(s) 2018-03-02T21:25:47.5273015Z 1 Error(s) ``` cc @jkotas, @stephentoub
infrastructure
corefx build is failing since remotelyinvokable does not contain a definition for longwait text build failed processtestbase uap cs error remotelyinvokable does not contain a definition for longwait warning s error s cc jkotas stephentoub
1
35,710
14,852,460,350
IssuesEvent
2021-01-18 08:37:31
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Need clarification whether the feature can be disabled/removed from a cluster
Pri2 container-service/svc cxp doc-enhancement triaged
There is no mention of whether the feature can be removed from a cluster and I saw that there's no ` --disable-azure-rbac` command in the CLI. Can this be clarified in the docs please? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 42afd8de-b697-100c-98c7-ae8e6a6f1b92 * Version Independent ID: 81674054-0ea2-210b-2e65-18a0f15617c1 * Content: [Manage Azure RBAC in Kubernetes From Azure - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/manage-azure-rbac) * Content Source: [articles/aks/manage-azure-rbac.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/manage-azure-rbac.md) * Service: **container-service** * GitHub Login: @palma21 * Microsoft Alias: **jpalma**
1.0
Need clarification whether the feature can be disabled/removed from a cluster - There is no mention of whether the feature can be removed from a cluster and I saw that there's no ` --disable-azure-rbac` command in the CLI. Can this be clarified in the docs please? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 42afd8de-b697-100c-98c7-ae8e6a6f1b92 * Version Independent ID: 81674054-0ea2-210b-2e65-18a0f15617c1 * Content: [Manage Azure RBAC in Kubernetes From Azure - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/manage-azure-rbac) * Content Source: [articles/aks/manage-azure-rbac.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/manage-azure-rbac.md) * Service: **container-service** * GitHub Login: @palma21 * Microsoft Alias: **jpalma**
non_infrastructure
need clarification whether the feature can be disabled removed from a cluster there is no mention of whether the feature can be removed from a cluster and i saw that there s no disable azure rbac command in the cli can this be clarified in the docs please document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login microsoft alias jpalma
0
619,618
19,531,195,562
IssuesEvent
2021-12-30 17:09:04
jshmrtn/hygeia
https://api.github.com/repos/jshmrtn/hygeia
closed
Befreiung von der Quarantäne Text anpassen
enhancement high-priority
## Is your feature request related to a problem? Please describe. In der Übersicht von einer KP steht dass man sich aus der Quarantäne entlassen kann, wenn man innerhalb von 6 Monaten bereits erkrankt ist. (siehe Screenshot) ## Describe the solution you'd like Bitte den ersten Satz anpassen auf: "Sie können sich von der Quarantäne befreien lassen, wenn Sie im Besitz eines gültigen Covid-Zertifikates sind." ![2021-12-16 08_18_12-Corinne Fräfel - Person - Hygeia](https://user-images.githubusercontent.com/82145566/146636183-dd8c16ef-aedb-4368-b9e5-3edf2e1cd1b3.png) ## Describe alternatives you've considered <!-- A clear and concise description of any alternative solutions or features you've considered. --> ## Additional context <!-- Add any other context or screenshots about the feature request here. --> ## References * https://app.forecast.it/project/P-205/scoping/T3283
1.0
Befreiung von der Quarantäne Text anpassen - ## Is your feature request related to a problem? Please describe. In der Übersicht von einer KP steht dass man sich aus der Quarantäne entlassen kann, wenn man innerhalb von 6 Monaten bereits erkrankt ist. (siehe Screenshot) ## Describe the solution you'd like Bitte den ersten Satz anpassen auf: "Sie können sich von der Quarantäne befreien lassen, wenn Sie im Besitz eines gültigen Covid-Zertifikates sind." ![2021-12-16 08_18_12-Corinne Fräfel - Person - Hygeia](https://user-images.githubusercontent.com/82145566/146636183-dd8c16ef-aedb-4368-b9e5-3edf2e1cd1b3.png) ## Describe alternatives you've considered <!-- A clear and concise description of any alternative solutions or features you've considered. --> ## Additional context <!-- Add any other context or screenshots about the feature request here. --> ## References * https://app.forecast.it/project/P-205/scoping/T3283
non_infrastructure
befreiung von der quarantäne text anpassen is your feature request related to a problem please describe in der übersicht von einer kp steht dass man sich aus der quarantäne entlassen kann wenn man innerhalb von monaten bereits erkrankt ist siehe screenshot describe the solution you d like bitte den ersten satz anpassen auf sie können sich von der quarantäne befreien lassen wenn sie im besitz eines gültigen covid zertifikates sind describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here references
0
37,619
10,041,088,038
IssuesEvent
2019-07-18 21:41:06
gunrock/gunrock
https://api.github.com/repos/gunrock/gunrock
opened
Too many warnings during build
bug build issue enhancement
**Summary:** We have way too many warnings building Gunrock. They've lost their utility. **Reproducibility:** Always **To Reproduce:** Build Gunrock as usual. **Expected Behavior:** There should be absolutely no warnings.
1.0
Too many warnings during build - **Summary:** We have way too many warnings building Gunrock. They've lost their utility. **Reproducibility:** Always **To Reproduce:** Build Gunrock as usual. **Expected Behavior:** There should be absolutely no warnings.
non_infrastructure
too many warnings during build summary we have way too many warnings building gunrock they ve lost their utility reproducibility always to reproduce build gunrock as usual expected behavior there should be absolutely no warnings
0