Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 853 | labels stringlengths 4 898 | body stringlengths 2 262k | index stringclasses 13 values | text_combine stringlengths 96 262k | label stringclasses 2 values | text stringlengths 96 250k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
89,465 | 25,805,746,456 | IssuesEvent | 2022-12-11 11:49:21 | NFFT/nfft | https://api.github.com/repos/NFFT/nfft | closed | [Build routine] Progress bar for visualisation? | enhancement question build routine | How about adding a progress bar such as the one `apt` has to the build routine? This would be an enhancement for the users who build `NFFT` for the first time since they would know what is already done in comparison to the whole build step.
As a measure of progress, I would like to suggest the count of already executed instructions. | 1.0 | [Build routine] Progress bar for visualisation? - How about adding a progress bar such as the one `apt` has to the build routine? This would be an enhancement for the users who build `NFFT` for the first time since they would know what is already done in comparison to the whole build step.
As a measure of progress, I would like to suggest the count of already executed instructions. | build | progress bar for visualisation how about adding a progress bar such as the one apt has to the build routine this would be an enhancement for the users who build nfft for the first time since they would know what is already done in comparison to the whole build step as a measure of progress i would like to suggest the count of already executed instructions | 1 |
87,913 | 25,248,929,981 | IssuesEvent | 2022-11-15 13:19:05 | bcpierce00/unison | https://api.github.com/repos/bcpierce00/unison | closed | support dune properly, remove usage of c_names | effort-low impact-low build | ````
[ 70s] + /usr/bin/mkdir /home/abuild/rpmbuild/BUILDROOT/unison-2.51.2.26a29f7-62.opt_409.1.x86_64
[ 70s] + cd unison-2.51.2.26a29f7
[ 70s] + dune_release_pkgs=unison
[ 70s] + echo 2.51.2.26a29f7
[ 70s] + tee VERSION
[ 70s] 2.51.2.26a29f7
[ 70s] + dune_for_release=
[ 70s] + : dune_release_pkgs
[ 70s] + test -n unison
[ 70s] + echo unison
[ 70s] + dune_for_release=--for-release-of-packages=unison
[ 70s] + dune installed-libraries
[ 70s] bigarray (version: OCaml 4.09.0)
[ 70s] bytes (version: OCaml 4.09.0)
[ 70s] compiler-libs (version: OCaml 4.09.0)
[ 70s] compiler-libs.bytecomp (version: OCaml 4.09.0)
[ 70s] compiler-libs.common (version: OCaml 4.09.0)
[ 70s] compiler-libs.optcomp (version: OCaml 4.09.0)
[ 70s] compiler-libs.toplevel (version: OCaml 4.09.0)
[ 70s] dynlink (version: OCaml 4.09.0)
[ 70s] lablgtk2 (version: 2.18.8)
[ 70s] lablgtk2.auto-init (version: n/a)
[ 70s] lablgtk2.glade (version: n/a)
[ 70s] lablgtk2.gnomecanvas (version: n/a)
[ 70s] lablgtk2.gtkspell (version: n/a)
[ 70s] lablgtk2.rsvg (version: n/a)
[ 70s] lablgtk2.sourceview2 (version: n/a)
[ 70s] raw_spacetime (version: OCaml 4.09.0)
[ 70s] result (version: [distributed with Ocaml])
[ 70s] seq (version: OCaml 4.09.0)
[ 70s] stdlib (version: OCaml 4.09.0)
[ 70s] str (version: OCaml 4.09.0)
[ 70s] threads (version: OCaml 4.09.0)
[ 70s] threads.none (version: OCaml 4.09.0)
[ 70s] threads.posix (version: OCaml 4.09.0)
[ 70s] uchar (version: OCaml 4.09.0)
[ 70s] unix (version: OCaml 4.09.0)
[ 70s] + dune external-lib-deps --for-release-of-packages=unison @install
[ 70s] Info: Creating file dune-project with this contents:
[ 70s] | (lang dune 2.2)
[ 70s] File "dune", line 10, characters 1-41:
[ 70s] 10 | (c_names bytearray_stubs osxsupport pty)
[ 70s] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ 70s] Error: 'c_names' was deleted in version 2.0 of the dune language. Use the
[ 70s] (foreign_stubs ...) field instead.
[ 70s] error: Bad exit status from /var/tmp/rpm-tmp.cw95L5 (%build)
```` | 1.0 | support dune properly, remove usage of c_names - ````
[ 70s] + /usr/bin/mkdir /home/abuild/rpmbuild/BUILDROOT/unison-2.51.2.26a29f7-62.opt_409.1.x86_64
[ 70s] + cd unison-2.51.2.26a29f7
[ 70s] + dune_release_pkgs=unison
[ 70s] + echo 2.51.2.26a29f7
[ 70s] + tee VERSION
[ 70s] 2.51.2.26a29f7
[ 70s] + dune_for_release=
[ 70s] + : dune_release_pkgs
[ 70s] + test -n unison
[ 70s] + echo unison
[ 70s] + dune_for_release=--for-release-of-packages=unison
[ 70s] + dune installed-libraries
[ 70s] bigarray (version: OCaml 4.09.0)
[ 70s] bytes (version: OCaml 4.09.0)
[ 70s] compiler-libs (version: OCaml 4.09.0)
[ 70s] compiler-libs.bytecomp (version: OCaml 4.09.0)
[ 70s] compiler-libs.common (version: OCaml 4.09.0)
[ 70s] compiler-libs.optcomp (version: OCaml 4.09.0)
[ 70s] compiler-libs.toplevel (version: OCaml 4.09.0)
[ 70s] dynlink (version: OCaml 4.09.0)
[ 70s] lablgtk2 (version: 2.18.8)
[ 70s] lablgtk2.auto-init (version: n/a)
[ 70s] lablgtk2.glade (version: n/a)
[ 70s] lablgtk2.gnomecanvas (version: n/a)
[ 70s] lablgtk2.gtkspell (version: n/a)
[ 70s] lablgtk2.rsvg (version: n/a)
[ 70s] lablgtk2.sourceview2 (version: n/a)
[ 70s] raw_spacetime (version: OCaml 4.09.0)
[ 70s] result (version: [distributed with Ocaml])
[ 70s] seq (version: OCaml 4.09.0)
[ 70s] stdlib (version: OCaml 4.09.0)
[ 70s] str (version: OCaml 4.09.0)
[ 70s] threads (version: OCaml 4.09.0)
[ 70s] threads.none (version: OCaml 4.09.0)
[ 70s] threads.posix (version: OCaml 4.09.0)
[ 70s] uchar (version: OCaml 4.09.0)
[ 70s] unix (version: OCaml 4.09.0)
[ 70s] + dune external-lib-deps --for-release-of-packages=unison @install
[ 70s] Info: Creating file dune-project with this contents:
[ 70s] | (lang dune 2.2)
[ 70s] File "dune", line 10, characters 1-41:
[ 70s] 10 | (c_names bytearray_stubs osxsupport pty)
[ 70s] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
[ 70s] Error: 'c_names' was deleted in version 2.0 of the dune language. Use the
[ 70s] (foreign_stubs ...) field instead.
[ 70s] error: Bad exit status from /var/tmp/rpm-tmp.cw95L5 (%build)
```` | build | support dune properly remove usage of c names usr bin mkdir home abuild rpmbuild buildroot unison opt cd unison dune release pkgs unison echo tee version dune for release dune release pkgs test n unison echo unison dune for release for release of packages unison dune installed libraries bigarray version ocaml bytes version ocaml compiler libs version ocaml compiler libs bytecomp version ocaml compiler libs common version ocaml compiler libs optcomp version ocaml compiler libs toplevel version ocaml dynlink version ocaml version auto init version n a glade version n a gnomecanvas version n a gtkspell version n a rsvg version n a version n a raw spacetime version ocaml result version seq version ocaml stdlib version ocaml str version ocaml threads version ocaml threads none version ocaml threads posix version ocaml uchar version ocaml unix version ocaml dune external lib deps for release of packages unison install info creating file dune project with this contents lang dune file dune line characters c names bytearray stubs osxsupport pty error c names was deleted in version of the dune language use the foreign stubs field instead error bad exit status from var tmp rpm tmp build | 1 |
1,360 | 2,733,794,352 | IssuesEvent | 2015-04-17 15:54:35 | gregorio-project/gregorio | https://api.github.com/repos/gregorio-project/gregorio | opened | Build error | bug build/install | In trying to build `develop` and `release-3.0` from a freshly cloned repository on a Mac I get the following error:
`bison: gabc-score-determination-y.c: cannot open: Permission denied`
Checking the file, it's permissions don't seem to be any different from the other files. Does anyone have any idea as to why this error might be cropping up? I can obviously get around it by running the build with root permissions, but I thought we had fixed things so that wasn't necessary. | 1.0 | Build error - In trying to build `develop` and `release-3.0` from a freshly cloned repository on a Mac I get the following error:
`bison: gabc-score-determination-y.c: cannot open: Permission denied`
Checking the file, it's permissions don't seem to be any different from the other files. Does anyone have any idea as to why this error might be cropping up? I can obviously get around it by running the build with root permissions, but I thought we had fixed things so that wasn't necessary. | build | build error in trying to build develop and release from a freshly cloned repository on a mac i get the following error bison gabc score determination y c cannot open permission denied checking the file it s permissions don t seem to be any different from the other files does anyone have any idea as to why this error might be cropping up i can obviously get around it by running the build with root permissions but i thought we had fixed things so that wasn t necessary | 1 |
151,773 | 23,871,224,128 | IssuesEvent | 2022-09-07 14:58:59 | nevermined-io/defi-marketplace | https://api.github.com/repos/nevermined-io/defi-marketplace | closed | Subscription purchase flow | design | - [x] Create subscription page desing
- [x] Defin susbscription purchase flow | 1.0 | Subscription purchase flow - - [x] Create subscription page desing
- [x] Defin susbscription purchase flow | non_build | subscription purchase flow create subscription page desing defin susbscription purchase flow | 0 |
12,563 | 5,215,374,400 | IssuesEvent | 2017-01-26 04:32:17 | grpc/grpc | https://api.github.com/repos/grpc/grpc | closed | linux/dbg: json_run_localhost timeout | BUILDPONY c++ flaky test P0 | `c++_linux_dbg.bins/dbg/json_run_localhost --scenarios_json '{"scenarios": [{"name": "cpp_protobuf_async_client_sync_server_streaming_qps_unconstrained_secure", "warmup_seconds": 0, "benchmark_seconds": 1, "num_servers": 1, "server_config": {"async_server_threads": 0, "core_limit": 0, "security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "server_type": "SYNC_SERVER"}, "num_clients": 0, "client_config": {"client_type": "ASYNC_CLIENT", "security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "payload_config": {"simple_params": {"resp_size": 0, "req_size": 0}}, "client_channels": 64, "async_client_threads": 0, "outstanding_rpcs_per_channel": 100, "rpc_type": "STREAMING", "load_params": {"closed_loop": {}}, "histogram_params": {"max_possible": 60000000000.0, "resolution": 0.01}}}]}' GRPC_POLL_STRATEGY=poll-cv`
https://gist.githubusercontent.com/dgquintas/81d81ad215048fe43a473ed7434b9c76/raw/7fa9401c238c1d1f9ba2b5f913e4784b3fbf0199/gistfile1.txt
| 1.0 | linux/dbg: json_run_localhost timeout - `c++_linux_dbg.bins/dbg/json_run_localhost --scenarios_json '{"scenarios": [{"name": "cpp_protobuf_async_client_sync_server_streaming_qps_unconstrained_secure", "warmup_seconds": 0, "benchmark_seconds": 1, "num_servers": 1, "server_config": {"async_server_threads": 0, "core_limit": 0, "security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "server_type": "SYNC_SERVER"}, "num_clients": 0, "client_config": {"client_type": "ASYNC_CLIENT", "security_params": {"use_test_ca": true, "server_host_override": "foo.test.google.fr"}, "payload_config": {"simple_params": {"resp_size": 0, "req_size": 0}}, "client_channels": 64, "async_client_threads": 0, "outstanding_rpcs_per_channel": 100, "rpc_type": "STREAMING", "load_params": {"closed_loop": {}}, "histogram_params": {"max_possible": 60000000000.0, "resolution": 0.01}}}]}' GRPC_POLL_STRATEGY=poll-cv`
https://gist.githubusercontent.com/dgquintas/81d81ad215048fe43a473ed7434b9c76/raw/7fa9401c238c1d1f9ba2b5f913e4784b3fbf0199/gistfile1.txt
| build | linux dbg json run localhost timeout c linux dbg bins dbg json run localhost scenarios json scenarios grpc poll strategy poll cv | 1 |
94,568 | 27,237,433,393 | IssuesEvent | 2023-02-21 17:21:15 | dotnet/arcade | https://api.github.com/repos/dotnet/arcade | closed | Build failed: Maestro Build Promotion/main #Promoting dotnet-windowsdesktop build 20230215.2 (167131) to channel(s) '.NET 7' # | First Responder Build Failed | Build [#Promoting dotnet-windowsdesktop build 20230215.2 (167131) to channel(s) '.NET 7' #](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2114840) failed
## :x: : internal / Maestro Build Promotion failed
### Summary
**Finished** - Wed, 15 Feb 2023 16:43:30 GMT
**Duration** - 5 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Publishing
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\75cfd413-cea0-4696-9250-4c7dea5793cd\Microsoft.WindowsDesktop.App.Ref.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Ref.7.0.4.symbols.nupkg.sha512'
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\873a646a-2cae-4bf6-9a6d-82ceca6a79a4\Microsoft.WindowsDesktop.App.Runtime.win-arm64.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Runtime.win-arm64.7.0.4.symbols.nupkg.sha512'
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\1e1ffe07-22e1-48b5-93d2-634792494483\Microsoft.WindowsDesktop.App.Runtime.win-x64.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Runtime.win-x64.7.0.4.symbols.nupkg.sha512'
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\6aac4d7b-d943-4682-9da4-3a78c92c20a7\Microsoft.WindowsDesktop.App.Runtime.win-x86.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Runtime.win-x86.7.0.4.symbols.nupkg.sha512'
### Changes
| 1.0 | Build failed: Maestro Build Promotion/main #Promoting dotnet-windowsdesktop build 20230215.2 (167131) to channel(s) '.NET 7' # - Build [#Promoting dotnet-windowsdesktop build 20230215.2 (167131) to channel(s) '.NET 7' #](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_build/results?buildId=2114840) failed
## :x: : internal / Maestro Build Promotion failed
### Summary
**Finished** - Wed, 15 Feb 2023 16:43:30 GMT
**Duration** - 5 minutes
**Requested for** - DotNet Bot
**Reason** - manual
### Details
#### Publishing
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\75cfd413-cea0-4696-9250-4c7dea5793cd\Microsoft.WindowsDesktop.App.Ref.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Ref.7.0.4.symbols.nupkg.sha512'
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\873a646a-2cae-4bf6-9a6d-82ceca6a79a4\Microsoft.WindowsDesktop.App.Runtime.win-arm64.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Runtime.win-arm64.7.0.4.symbols.nupkg.sha512'
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\1e1ffe07-22e1-48b5-93d2-634792494483\Microsoft.WindowsDesktop.App.Runtime.win-x64.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Runtime.win-x64.7.0.4.symbols.nupkg.sha512'
- :x: - [[Log]](https://dev.azure.com/dnceng/7ea9116e-9fac-403d-b258-b31fcf1bb293/_apis/build/builds/2114840/logs/28) - .packages\microsoft.dotnet.arcade.sdk\8.0.0-beta.23113.1\tools\SdkTasks\PublishArtifactsInManifest.proj(144,5): error : Asset 'D:\a\_work\1\a\6aac4d7b-d943-4682-9da4-3a78c92c20a7\Microsoft.WindowsDesktop.App.Runtime.win-x86.7.0.4.symbols.nupkg.sha512' already exists with different contents at 'assets/symbols/Microsoft.WindowsDesktop.App.Runtime.win-x86.7.0.4.symbols.nupkg.sha512'
### Changes
| build | build failed maestro build promotion main promoting dotnet windowsdesktop build to channel s net build failed x internal maestro build promotion failed summary finished wed feb gmt duration minutes requested for dotnet bot reason manual details publishing x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error asset d a work a microsoft windowsdesktop app ref symbols nupkg already exists with different contents at assets symbols microsoft windowsdesktop app ref symbols nupkg x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error asset d a work a microsoft windowsdesktop app runtime win symbols nupkg already exists with different contents at assets symbols microsoft windowsdesktop app runtime win symbols nupkg x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error asset d a work a microsoft windowsdesktop app runtime win symbols nupkg already exists with different contents at assets symbols microsoft windowsdesktop app runtime win symbols nupkg x packages microsoft dotnet arcade sdk beta tools sdktasks publishartifactsinmanifest proj error asset d a work a microsoft windowsdesktop app runtime win symbols nupkg already exists with different contents at assets symbols microsoft windowsdesktop app runtime win symbols nupkg changes | 1 |
90,215 | 26,010,133,112 | IssuesEvent | 2022-12-21 00:20:28 | benthevining/Limes | https://api.github.com/repos/benthevining/Limes | closed | [BUG] Files undefined symbols with gcc | bug os/macOS lib/limes_files Build | MacOS, GCC 12.
The linker cannot see `files::module_path::get_impl()`. See [this build log](https://my.cdash.org/viewBuildError.php?buildid=2234491). | 1.0 | [BUG] Files undefined symbols with gcc - MacOS, GCC 12.
The linker cannot see `files::module_path::get_impl()`. See [this build log](https://my.cdash.org/viewBuildError.php?buildid=2234491). | build | files undefined symbols with gcc macos gcc the linker cannot see files module path get impl see | 1 |
28,919 | 12,990,787,223 | IssuesEvent | 2020-07-23 01:14:28 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | this page doesn't seems to work | Pri2 container-service/svc cxp needs-more-info triaged |
[Enter feedback here]
I followed all the steps but the demo doesn't seem to work.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: fd0e7d5a-37e9-07dc-f395-9cb5bb580e45
* Version Independent ID: 7e4faf70-9724-7e7b-832b-1cd99a974920
* Content: [Create an ingress controller - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-basic)
* Content Source: [articles/aks/ingress-basic.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-basic.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | 1.0 | this page doesn't seems to work -
[Enter feedback here]
I followed all the steps but the demo doesn't seem to work.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: fd0e7d5a-37e9-07dc-f395-9cb5bb580e45
* Version Independent ID: 7e4faf70-9724-7e7b-832b-1cd99a974920
* Content: [Create an ingress controller - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/ingress-basic)
* Content Source: [articles/aks/ingress-basic.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/ingress-basic.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | non_build | this page doesn t seems to work i followed all the steps but the demo doesn t seem to work document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned | 0 |
80,973 | 23,348,983,250 | IssuesEvent | 2022-08-09 21:01:11 | PowerShell/PowerShell | https://api.github.com/repos/PowerShell/PowerShell | closed | MSI installer doesn't enable remoting even after selecting it | Area-Maintainers-Build Resolution-By Design | Steps to reproduce
------------------
1. Check the current available endpoints using **Windows PowerShell**
```powershell
$PSVersionTable
Name Value
---- -----
PSVersion 5.1.17134.112
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.17134.112
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
# Lists the default endpoints
Get-PSSessionConfiguration | ft Name
Name
----
Microsoft.PowerShell
Microsoft.Powershell.Workflow
Microsoft.PowerShell32
```
2. Install the PS Core 6.1.0-preview 3 using the MSI and select _Enable PowerShell remoting_

Expected behavior
-----------------
New remoting endpoint(s) are **PRSENT** on the system as done by _Install-PowerShellRemoting.ps1_ script in $PSHome
```powershell
# From Windows PowerShell
Get-PSSessionConfiguration | ft name
Name
----
Microsoft.PowerShell
Microsoft.Powershell.Workflow
Microsoft.PowerShell32
PowerShell.6
PowerShell.6.1.0-preview.3
# From PowerShell Core
Get-PSSessionConfiguration |ft name
Name
----
PowerShell.6
PowerShell.6.1.0-preview.3
```
Actual behavior
---------------
New remoting endpoint(s) are **ABSENT**
```powershell
# From Windows PowerShell
Get-PSSessionConfiguration | ft name
Name
----
Microsoft.PowerShell
Microsoft.Powershell.Workflow
Microsoft.PowerShell32
# From PowerShell Core
Get-PSSessionConfiguration
```
| 1.0 | MSI installer doesn't enable remoting even after selecting it - Steps to reproduce
------------------
1. Check the current available endpoints using **Windows PowerShell**
```powershell
$PSVersionTable
Name Value
---- -----
PSVersion 5.1.17134.112
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.17134.112
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
# Lists the default endpoints
Get-PSSessionConfiguration | ft Name
Name
----
Microsoft.PowerShell
Microsoft.Powershell.Workflow
Microsoft.PowerShell32
```
2. Install the PS Core 6.1.0-preview 3 using the MSI and select _Enable PowerShell remoting_

Expected behavior
-----------------
New remoting endpoint(s) are **PRSENT** on the system as done by _Install-PowerShellRemoting.ps1_ script in $PSHome
```powershell
# From Windows PowerShell
Get-PSSessionConfiguration | ft name
Name
----
Microsoft.PowerShell
Microsoft.Powershell.Workflow
Microsoft.PowerShell32
PowerShell.6
PowerShell.6.1.0-preview.3
# From PowerShell Core
Get-PSSessionConfiguration |ft name
Name
----
PowerShell.6
PowerShell.6.1.0-preview.3
```
Actual behavior
---------------
New remoting endpoint(s) are **ABSENT**
```powershell
# From Windows PowerShell
Get-PSSessionConfiguration | ft name
Name
----
Microsoft.PowerShell
Microsoft.Powershell.Workflow
Microsoft.PowerShell32
# From PowerShell Core
Get-PSSessionConfiguration
```
| build | msi installer doesn t enable remoting even after selecting it steps to reproduce check the current available endpoints using windows powershell powershell psversiontable name value psversion psedition desktop pscompatibleversions buildversion clrversion wsmanstackversion psremotingprotocolversion serializationversion lists the default endpoints get pssessionconfiguration ft name name microsoft powershell microsoft powershell workflow microsoft install the ps core preview using the msi and select enable powershell remoting expected behavior new remoting endpoint s are prsent on the system as done by install powershellremoting script in pshome powershell from windows powershell get pssessionconfiguration ft name name microsoft powershell microsoft powershell workflow microsoft powershell powershell preview from powershell core get pssessionconfiguration ft name name powershell powershell preview actual behavior new remoting endpoint s are absent powershell from windows powershell get pssessionconfiguration ft name name microsoft powershell microsoft powershell workflow microsoft from powershell core get pssessionconfiguration | 1 |
231,720 | 7,642,437,029 | IssuesEvent | 2018-05-08 09:15:21 | bitshares/bitshares-ui | https://api.github.com/repos/bitshares/bitshares-ui | closed | [3] Exchange not loading for new, zero-balance accounts | bug high priority | When a new account is created via cloud login, the exchange tab doesnt load anymore for some markets.
prominent example:
does not load
https://wallet.bitshares.org/#/market/USD_BTS
does load
https://wallet.bitshares.org/#/market/BTS_USD
The issue is in https://github.com/bitshares/bitshares-ui/blob/c69f297ee1cea480b531f8f38542baf1b17a5736/app/components/Exchange/Exchange.jsx#L288 that the feeStatus does not get loaded, which causes a return null in the render method. | 1.0 | [3] Exchange not loading for new, zero-balance accounts - When a new account is created via cloud login, the exchange tab doesnt load anymore for some markets.
prominent example:
does not load
https://wallet.bitshares.org/#/market/USD_BTS
does load
https://wallet.bitshares.org/#/market/BTS_USD
The issue is in https://github.com/bitshares/bitshares-ui/blob/c69f297ee1cea480b531f8f38542baf1b17a5736/app/components/Exchange/Exchange.jsx#L288 that the feeStatus does not get loaded, which causes a return null in the render method. | non_build | exchange not loading for new zero balance accounts when a new account is created via cloud login the exchange tab doesnt load anymore for some markets prominent example does not load does load the issue is in that the feestatus does not get loaded which causes a return null in the render method | 0 |
248,835 | 7,936,722,239 | IssuesEvent | 2018-07-09 10:20:39 | telstra/open-kilda | https://api.github.com/repos/telstra/open-kilda | closed | HTTP Status 500 from NB on null latency in Neo4j | bug priority/2-high | ```
<!doctype html><html lang="en"><head><title>HTTP Status 500 – Internal Server Error</title><style type="text/css">h1 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:22px;} h2 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:16px;} h3 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:14px;} body {font-family:Tahoma,Arial,sans-serif;color:black;background-color:white;} b {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;} p {font-family:Tahoma,Arial,sans-serif;background:white;color:black;font-size:12px;} a {color:black;} a.name {color:black;} .line {height:1px;background-color:#525D76;border:none;}</style></head><body><h1>HTTP Status 500 – Internal Server Error</h1><hr class="line" /><p><b>Type</b> Exception Report</p><p><b>Message</b> Request processing failed; nested exception is org.springframework.web.client.RestClientException: Could not extract response: no suitable HttpMessageConverter found for response type [class [Lorg.openkilda.messaging.info.event.IslInfoData;] and content type [text/html;charset=utf-8]</p><p><b>Description</b> The server encountered an unexpected condition that prevented it from fulfilling the request.</p><p><b>Exception</b></p><pre>org.springframework.web.util.NestedServletException: Request processing failed; nested exception is org.springframework.web.client.RestClientException: Could not extract response: no suitable HttpMessageConverter found for response type [class [Lorg.openkilda.messaging.info.event.IslInfoData;] and content type [text/html;charset=utf-8]
org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:982)
org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:861)
javax.servlet.http.HttpServlet.service(HttpServlet.java:635)
org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846)
javax.servlet.http.HttpServlet.service(HttpServlet.java:742)
org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
org.openkilda.northbound.utils.RequestCorrelationFilter.doFilterInternal(RequestCorrelationFilter.java:57)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:170)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:63)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.www.BasicAuthenticationFilter.doFilterInternal(BasicAuthenticationFilter.java:215)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.logout.LogoutFilter.doFilter(LogoutFilter.java:116)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.header.HeaderWriterFilter.doFilterInternal(HeaderWriterFilter.java:64)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.SecurityContextPersistenceFilter.doFilter(SecurityContextPersistenceFilter.java:105)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter.doFilterInternal(WebAsyncManagerIntegrationFilter.java:56)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.FilterChainProxy.doFilterInternal(FilterChainProxy.java:214)
org.springframework.security.web.FilterChainProxy.doFilter(FilterChainProxy.java:177)
org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:346)
org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:262)
org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
</pre><p><b>Root Cause</b></p><pre>org.springframework.web.client.RestClientException: Could not extract response: no suitable HttpMessageConverter found for response type [class [Lorg.openkilda.messaging.info.event.IslInfoData;] and content type [text/html;charset=utf-8]
org.springframework.web.client.HttpMessageConverterExtractor.extractData(HttpMessageConverterExtractor.java:110)
org.springframework.web.client.RestTemplate$ResponseEntityResponseExtractor.extractData(RestTemplate.java:917)
org.springframework.web.client.RestTemplate$ResponseEntityResponseExtractor.extractData(RestTemplate.java:901)
org.springframework.web.client.RestTemplate.doExecute(RestTemplate.java:655)
org.springframework.web.client.RestTemplate.execute(RestTemplate.java:613)
org.springframework.web.client.RestTemplate.exchange(RestTemplate.java:531)
org.openkilda.northbound.service.impl.LinkServiceImpl.getLinks(LinkServiceImpl.java:83)
org.openkilda.northbound.controller.LinkController.getLinks(LinkController.java:55)
sun.reflect.GeneratedMethodAccessor148.invoke(Unknown Source)
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
java.lang.reflect.Method.invoke(Method.java:498)
org.springframework.web.method.support.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:205)
org.springframework.web.method.support.InvocableHandlerMethod.invokeForRequest(InvocableHandlerMethod.java:133)
org.springframework.web.servlet.mvc.method.annotation.ServletInvocableHandlerMethod.invokeAndHandle(ServletInvocableHandlerMethod.java:97)
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.invokeHandlerMethod(RequestMappingHandlerAdapter.java:827)
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.handleInternal(RequestMappingHandlerAdapter.java:738)
org.springframework.web.servlet.mvc.method.AbstractHandlerMethodAdapter.handle(AbstractHandlerMethodAdapter.java:85)
org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:967)
org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:901)
org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:970)
org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:861)
javax.servlet.http.HttpServlet.service(HttpServlet.java:635)
org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846)
javax.servlet.http.HttpServlet.service(HttpServlet.java:742)
org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
org.openkilda.northbound.utils.RequestCorrelationFilter.doFilterInternal(RequestCorrelationFilter.java:57)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:170)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:63)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.www.BasicAuthenticationFilter.doFilterInternal(BasicAuthenticationFilter.java:215)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.logout.LogoutFilter.doFilter(LogoutFilter.java:116)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.header.HeaderWriterFilter.doFilterInternal(HeaderWriterFilter.java:64)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.SecurityContextPersistenceFilter.doFilter(SecurityContextPersistenceFilter.java:105)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter.doFilterInternal(WebAsyncManagerIntegrationFilter.java:56)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.FilterChainProxy.doFilterInternal(FilterChainProxy.java:214)
org.springframework.security.web.FilterChainProxy.doFilter(FilterChainProxy.java:177)
org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:346)
org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:262)
org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
</pre><p><b>Note</b> The full stack trace of the root cause is available in the server logs.</p><hr class="line" /><h3>Apache Tomcat/8.5.16</h3></body></html>
```
| 1.0 | HTTP Status 500 from NB on null latency in Neo4j - ```
<!doctype html><html lang="en"><head><title>HTTP Status 500 – Internal Server Error</title><style type="text/css">h1 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:22px;} h2 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:16px;} h3 {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;font-size:14px;} body {font-family:Tahoma,Arial,sans-serif;color:black;background-color:white;} b {font-family:Tahoma,Arial,sans-serif;color:white;background-color:#525D76;} p {font-family:Tahoma,Arial,sans-serif;background:white;color:black;font-size:12px;} a {color:black;} a.name {color:black;} .line {height:1px;background-color:#525D76;border:none;}</style></head><body><h1>HTTP Status 500 – Internal Server Error</h1><hr class="line" /><p><b>Type</b> Exception Report</p><p><b>Message</b> Request processing failed; nested exception is org.springframework.web.client.RestClientException: Could not extract response: no suitable HttpMessageConverter found for response type [class [Lorg.openkilda.messaging.info.event.IslInfoData;] and content type [text/html;charset=utf-8]</p><p><b>Description</b> The server encountered an unexpected condition that prevented it from fulfilling the request.</p><p><b>Exception</b></p><pre>org.springframework.web.util.NestedServletException: Request processing failed; nested exception is org.springframework.web.client.RestClientException: Could not extract response: no suitable HttpMessageConverter found for response type [class [Lorg.openkilda.messaging.info.event.IslInfoData;] and content type [text/html;charset=utf-8]
org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:982)
org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:861)
javax.servlet.http.HttpServlet.service(HttpServlet.java:635)
org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846)
javax.servlet.http.HttpServlet.service(HttpServlet.java:742)
org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
org.openkilda.northbound.utils.RequestCorrelationFilter.doFilterInternal(RequestCorrelationFilter.java:57)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:170)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:63)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.www.BasicAuthenticationFilter.doFilterInternal(BasicAuthenticationFilter.java:215)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.logout.LogoutFilter.doFilter(LogoutFilter.java:116)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.header.HeaderWriterFilter.doFilterInternal(HeaderWriterFilter.java:64)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.SecurityContextPersistenceFilter.doFilter(SecurityContextPersistenceFilter.java:105)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter.doFilterInternal(WebAsyncManagerIntegrationFilter.java:56)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.FilterChainProxy.doFilterInternal(FilterChainProxy.java:214)
org.springframework.security.web.FilterChainProxy.doFilter(FilterChainProxy.java:177)
org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:346)
org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:262)
org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
</pre><p><b>Root Cause</b></p><pre>org.springframework.web.client.RestClientException: Could not extract response: no suitable HttpMessageConverter found for response type [class [Lorg.openkilda.messaging.info.event.IslInfoData;] and content type [text/html;charset=utf-8]
org.springframework.web.client.HttpMessageConverterExtractor.extractData(HttpMessageConverterExtractor.java:110)
org.springframework.web.client.RestTemplate$ResponseEntityResponseExtractor.extractData(RestTemplate.java:917)
org.springframework.web.client.RestTemplate$ResponseEntityResponseExtractor.extractData(RestTemplate.java:901)
org.springframework.web.client.RestTemplate.doExecute(RestTemplate.java:655)
org.springframework.web.client.RestTemplate.execute(RestTemplate.java:613)
org.springframework.web.client.RestTemplate.exchange(RestTemplate.java:531)
org.openkilda.northbound.service.impl.LinkServiceImpl.getLinks(LinkServiceImpl.java:83)
org.openkilda.northbound.controller.LinkController.getLinks(LinkController.java:55)
sun.reflect.GeneratedMethodAccessor148.invoke(Unknown Source)
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
java.lang.reflect.Method.invoke(Method.java:498)
org.springframework.web.method.support.InvocableHandlerMethod.doInvoke(InvocableHandlerMethod.java:205)
org.springframework.web.method.support.InvocableHandlerMethod.invokeForRequest(InvocableHandlerMethod.java:133)
org.springframework.web.servlet.mvc.method.annotation.ServletInvocableHandlerMethod.invokeAndHandle(ServletInvocableHandlerMethod.java:97)
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.invokeHandlerMethod(RequestMappingHandlerAdapter.java:827)
org.springframework.web.servlet.mvc.method.annotation.RequestMappingHandlerAdapter.handleInternal(RequestMappingHandlerAdapter.java:738)
org.springframework.web.servlet.mvc.method.AbstractHandlerMethodAdapter.handle(AbstractHandlerMethodAdapter.java:85)
org.springframework.web.servlet.DispatcherServlet.doDispatch(DispatcherServlet.java:967)
org.springframework.web.servlet.DispatcherServlet.doService(DispatcherServlet.java:901)
org.springframework.web.servlet.FrameworkServlet.processRequest(FrameworkServlet.java:970)
org.springframework.web.servlet.FrameworkServlet.doGet(FrameworkServlet.java:861)
javax.servlet.http.HttpServlet.service(HttpServlet.java:635)
org.springframework.web.servlet.FrameworkServlet.service(FrameworkServlet.java:846)
javax.servlet.http.HttpServlet.service(HttpServlet.java:742)
org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52)
org.openkilda.northbound.utils.RequestCorrelationFilter.doFilterInternal(RequestCorrelationFilter.java:57)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:317)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.invoke(FilterSecurityInterceptor.java:127)
org.springframework.security.web.access.intercept.FilterSecurityInterceptor.doFilter(FilterSecurityInterceptor.java:91)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.access.ExceptionTranslationFilter.doFilter(ExceptionTranslationFilter.java:114)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.session.SessionManagementFilter.doFilter(SessionManagementFilter.java:137)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.AnonymousAuthenticationFilter.doFilter(AnonymousAuthenticationFilter.java:111)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.servletapi.SecurityContextHolderAwareRequestFilter.doFilter(SecurityContextHolderAwareRequestFilter.java:170)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.savedrequest.RequestCacheAwareFilter.doFilter(RequestCacheAwareFilter.java:63)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.www.BasicAuthenticationFilter.doFilterInternal(BasicAuthenticationFilter.java:215)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.authentication.logout.LogoutFilter.doFilter(LogoutFilter.java:116)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.header.HeaderWriterFilter.doFilterInternal(HeaderWriterFilter.java:64)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.SecurityContextPersistenceFilter.doFilter(SecurityContextPersistenceFilter.java:105)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.context.request.async.WebAsyncManagerIntegrationFilter.doFilterInternal(WebAsyncManagerIntegrationFilter.java:56)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
org.springframework.security.web.FilterChainProxy$VirtualFilterChain.doFilter(FilterChainProxy.java:331)
org.springframework.security.web.FilterChainProxy.doFilterInternal(FilterChainProxy.java:214)
org.springframework.security.web.FilterChainProxy.doFilter(FilterChainProxy.java:177)
org.springframework.web.filter.DelegatingFilterProxy.invokeDelegate(DelegatingFilterProxy.java:346)
org.springframework.web.filter.DelegatingFilterProxy.doFilter(DelegatingFilterProxy.java:262)
org.springframework.web.filter.CharacterEncodingFilter.doFilterInternal(CharacterEncodingFilter.java:197)
org.springframework.web.filter.OncePerRequestFilter.doFilter(OncePerRequestFilter.java:107)
</pre><p><b>Note</b> The full stack trace of the root cause is available in the server logs.</p><hr class="line" /><h3>Apache Tomcat/8.5.16</h3></body></html>
```
| non_build | http status from nb on null latency in http status – internal server error font family tahoma arial sans serif color white background color font size font family tahoma arial sans serif color white background color font size font family tahoma arial sans serif color white background color font size body font family tahoma arial sans serif color black background color white b font family tahoma arial sans serif color white background color p font family tahoma arial sans serif background white color black font size a color black a name color black line height background color border none http status – internal server error type exception report message request processing failed nested exception is org springframework web client restclientexception could not extract response no suitable httpmessageconverter found for response type and content type description the server encountered an unexpected condition that prevented it from fulfilling the request exception org springframework web util nestedservletexception request processing failed nested exception is org springframework web client restclientexception could not extract response no suitable httpmessageconverter found for response type and content type org springframework web servlet frameworkservlet processrequest frameworkservlet java org springframework web servlet frameworkservlet doget frameworkservlet java javax servlet http httpservlet service httpservlet java org springframework web servlet frameworkservlet service frameworkservlet java javax servlet http httpservlet service httpservlet java org apache tomcat websocket server wsfilter dofilter wsfilter java org openkilda northbound utils requestcorrelationfilter dofilterinternal requestcorrelationfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web access intercept filtersecurityinterceptor invoke filtersecurityinterceptor java org springframework security web access intercept filtersecurityinterceptor dofilter filtersecurityinterceptor java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web access exceptiontranslationfilter dofilter exceptiontranslationfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web session sessionmanagementfilter dofilter sessionmanagementfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web authentication anonymousauthenticationfilter dofilter anonymousauthenticationfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web servletapi securitycontextholderawarerequestfilter dofilter securitycontextholderawarerequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web savedrequest requestcacheawarefilter dofilter requestcacheawarefilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web authentication org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web authentication logout logoutfilter dofilter logoutfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web header headerwriterfilter dofilterinternal headerwriterfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web context securitycontextpersistencefilter dofilter securitycontextpersistencefilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web context request async webasyncmanagerintegrationfilter dofilterinternal webasyncmanagerintegrationfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web filterchainproxy dofilterinternal filterchainproxy java org springframework security web filterchainproxy dofilter filterchainproxy java org springframework web filter delegatingfilterproxy invokedelegate delegatingfilterproxy java org springframework web filter delegatingfilterproxy dofilter delegatingfilterproxy java org springframework web filter characterencodingfilter dofilterinternal characterencodingfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java root cause org springframework web client restclientexception could not extract response no suitable httpmessageconverter found for response type and content type org springframework web client httpmessageconverterextractor extractdata httpmessageconverterextractor java org springframework web client resttemplate responseentityresponseextractor extractdata resttemplate java org springframework web client resttemplate responseentityresponseextractor extractdata resttemplate java org springframework web client resttemplate doexecute resttemplate java org springframework web client resttemplate execute resttemplate java org springframework web client resttemplate exchange resttemplate java org openkilda northbound service impl linkserviceimpl getlinks linkserviceimpl java org openkilda northbound controller linkcontroller getlinks linkcontroller java sun reflect invoke unknown source sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java java lang reflect method invoke method java org springframework web method support invocablehandlermethod doinvoke invocablehandlermethod java org springframework web method support invocablehandlermethod invokeforrequest invocablehandlermethod java org springframework web servlet mvc method annotation servletinvocablehandlermethod invokeandhandle servletinvocablehandlermethod java org springframework web servlet mvc method annotation requestmappinghandleradapter invokehandlermethod requestmappinghandleradapter java org springframework web servlet mvc method annotation requestmappinghandleradapter handleinternal requestmappinghandleradapter java org springframework web servlet mvc method abstracthandlermethodadapter handle abstracthandlermethodadapter java org springframework web servlet dispatcherservlet dodispatch dispatcherservlet java org springframework web servlet dispatcherservlet doservice dispatcherservlet java org springframework web servlet frameworkservlet processrequest frameworkservlet java org springframework web servlet frameworkservlet doget frameworkservlet java javax servlet http httpservlet service httpservlet java org springframework web servlet frameworkservlet service frameworkservlet java javax servlet http httpservlet service httpservlet java org apache tomcat websocket server wsfilter dofilter wsfilter java org openkilda northbound utils requestcorrelationfilter dofilterinternal requestcorrelationfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web access intercept filtersecurityinterceptor invoke filtersecurityinterceptor java org springframework security web access intercept filtersecurityinterceptor dofilter filtersecurityinterceptor java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web access exceptiontranslationfilter dofilter exceptiontranslationfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web session sessionmanagementfilter dofilter sessionmanagementfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web authentication anonymousauthenticationfilter dofilter anonymousauthenticationfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web servletapi securitycontextholderawarerequestfilter dofilter securitycontextholderawarerequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web savedrequest requestcacheawarefilter dofilter requestcacheawarefilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web authentication org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web authentication logout logoutfilter dofilter logoutfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web header headerwriterfilter dofilterinternal headerwriterfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web context securitycontextpersistencefilter dofilter securitycontextpersistencefilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web context request async webasyncmanagerintegrationfilter dofilterinternal webasyncmanagerintegrationfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java org springframework security web filterchainproxy virtualfilterchain dofilter filterchainproxy java org springframework security web filterchainproxy dofilterinternal filterchainproxy java org springframework security web filterchainproxy dofilter filterchainproxy java org springframework web filter delegatingfilterproxy invokedelegate delegatingfilterproxy java org springframework web filter delegatingfilterproxy dofilter delegatingfilterproxy java org springframework web filter characterencodingfilter dofilterinternal characterencodingfilter java org springframework web filter onceperrequestfilter dofilter onceperrequestfilter java note the full stack trace of the root cause is available in the server logs apache tomcat | 0 |
88,262 | 25,356,786,449 | IssuesEvent | 2022-11-20 12:41:28 | nodejs/node | https://api.github.com/repos/nodejs/node | opened | Unqualified call to `std::move` in node_http2.cc | build | Warning (from clang-cl on Windows) that I don't know how to fix:
```
src\node_http2.cc(647,32): warning : unqualified call to 'std::move' [-Wunqualified-std-cast-call] [D:\Git\nodejs\node\libnode.vcxproj]
```
https://github.com/nodejs/node/blob/4bee69a8c4a699c23342977b670ae90b870ac00d/src/node_http2.cc#L647-L650 | 1.0 | Unqualified call to `std::move` in node_http2.cc - Warning (from clang-cl on Windows) that I don't know how to fix:
```
src\node_http2.cc(647,32): warning : unqualified call to 'std::move' [-Wunqualified-std-cast-call] [D:\Git\nodejs\node\libnode.vcxproj]
```
https://github.com/nodejs/node/blob/4bee69a8c4a699c23342977b670ae90b870ac00d/src/node_http2.cc#L647-L650 | build | unqualified call to std move in node cc warning from clang cl on windows that i don t know how to fix src node cc warning unqualified call to std move | 1 |
51,721 | 21,784,301,200 | IssuesEvent | 2022-05-13 23:49:33 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | opened | Automate cleanup of Argo Workflows | ops and shared services | **Describe the issue**
Argo Workflows do not go away after they are run. The system generates warning messages when there are over 100 of them.
Implement an automated system to remove old workflows.
**Additional context**
In the past, these were periodically removed manually.
**Definition of done**
- [ ] Create a build with the `argo` CLI
- [ ] Create a config map with the necessary environment variables
- [ ] Create a cronjob to run daily and remove workflows that are more than X days old
| 1.0 | Automate cleanup of Argo Workflows - **Describe the issue**
Argo Workflows do not go away after they are run. The system generates warning messages when there are over 100 of them.
Implement an automated system to remove old workflows.
**Additional context**
In the past, these were periodically removed manually.
**Definition of done**
- [ ] Create a build with the `argo` CLI
- [ ] Create a config map with the necessary environment variables
- [ ] Create a cronjob to run daily and remove workflows that are more than X days old
| non_build | automate cleanup of argo workflows describe the issue argo workflows do not go away after they are run the system generates warning messages when there are over of them implement an automated system to remove old workflows additional context in the past these were periodically removed manually definition of done create a build with the argo cli create a config map with the necessary environment variables create a cronjob to run daily and remove workflows that are more than x days old | 0 |
29,303 | 8,318,631,262 | IssuesEvent | 2018-09-25 15:06:46 | elastic/runbld | https://api.github.com/repos/elastic/runbld | closed | optionally add timestamps to output | enhancement logging ~builds-team | Since runbld wraps other process(es), some builds do a good job and have timestamps in their output (others do not).
It would be nice to be able to inject timestamps at the beginning of the log messages optionally. Or another stance is to push this dependency down to the actual code/projects/builds that are being run by runbld.
Regardless maybe add timestamps to the runbld specific messages.
| 1.0 | optionally add timestamps to output - Since runbld wraps other process(es), some builds do a good job and have timestamps in their output (others do not).
It would be nice to be able to inject timestamps at the beginning of the log messages optionally. Or another stance is to push this dependency down to the actual code/projects/builds that are being run by runbld.
Regardless maybe add timestamps to the runbld specific messages.
| build | optionally add timestamps to output since runbld wraps other process es some builds do a good job and have timestamps in their output others do not it would be nice to be able to inject timestamps at the beginning of the log messages optionally or another stance is to push this dependency down to the actual code projects builds that are being run by runbld regardless maybe add timestamps to the runbld specific messages | 1 |
195,180 | 14,706,285,107 | IssuesEvent | 2021-01-04 19:35:18 | PRI-Illinois/WARM-PDD | https://api.github.com/repos/PRI-Illinois/WARM-PDD | closed | [BUG] "Link to Dashboard" doesn't appear on either calculator | Browser: Chrome READY FOR RETEST Severity 4: Minor Type: Bug | **Actual Behavior**
There was no link to "choose different calculator" found on the Destination Page for theSpecialty and Commodity Crop Calculators.
**Expected behavior**
When opening the Destination Page for Specialty and Commodity Crop Calculators, I expected to see a link to return to the dashboard labelled "choose different calculator".
**To Reproduce**
On the Dashboard under "Choose Calculator" - select "Commodity Crop Pests" - select "Go to Calculator" - the link labelled "choose different calculator" should be absent. Repeat for "Specialty Crop Pests".
**Browser and Version**
Google Chrome, Version86.0.4240.183
**Platform**
Intel Desktop, Windows 10
**Screenshots**
If applicable, add screenshots to help explain your problem.

| 1.0 | [BUG] "Link to Dashboard" doesn't appear on either calculator - **Actual Behavior**
There was no link to "choose different calculator" found on the Destination Page for theSpecialty and Commodity Crop Calculators.
**Expected behavior**
When opening the Destination Page for Specialty and Commodity Crop Calculators, I expected to see a link to return to the dashboard labelled "choose different calculator".
**To Reproduce**
On the Dashboard under "Choose Calculator" - select "Commodity Crop Pests" - select "Go to Calculator" - the link labelled "choose different calculator" should be absent. Repeat for "Specialty Crop Pests".
**Browser and Version**
Google Chrome, Version86.0.4240.183
**Platform**
Intel Desktop, Windows 10
**Screenshots**
If applicable, add screenshots to help explain your problem.

| non_build | link to dashboard doesn t appear on either calculator actual behavior there was no link to choose different calculator found on the destination page for thespecialty and commodity crop calculators expected behavior when opening the destination page for specialty and commodity crop calculators i expected to see a link to return to the dashboard labelled choose different calculator to reproduce on the dashboard under choose calculator select commodity crop pests select go to calculator the link labelled choose different calculator should be absent repeat for specialty crop pests browser and version google chrome platform intel desktop windows screenshots if applicable add screenshots to help explain your problem | 0 |
82,623 | 23,834,768,283 | IssuesEvent | 2022-09-06 04:01:29 | google/mediapipe | https://api.github.com/repos/google/mediapipe | opened | 【MAC】An error occurred building 'hello world' example | type:build/install | I'm trying to build 'hello_world' on MAC, and following error occurred.
**System information** (Please provide as much relevant information as possible)
Error in download_and_extract: java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.32.0/rules_apple.0.32.0.tar.gz] to /private/var/tmp/_bazel_wangjunjie/95eda4f7c99f7ddce46cd8d36dd7bed7/external/build_bazel_rules_apple/temp11825797914784900860/rules_apple.0.32.0.tar.gz: Connection refused (Connection refused)
ERROR: no such package '@build_bazel_rules_apple//apple': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.32.0/rules_apple.0.32.0.tar.gz] to /private/var/tmp/_bazel_wangjunjie/95eda4f7c99f7ddce46cd8d36dd7bed7/external/build_bazel_rules_apple/temp11825797914784900860/rules_apple.0.32.0.tar.gz: Connection refused (Connection refused)
INFO: Elapsed time: 12.881s
INFO: 0 processes.
**Describe the problem**:
seems like the package `build_bazel_rules_apple//apple` i can't download. But when i copy the link to the web browser, i can visit. Need help~
| 1.0 | 【MAC】An error occurred building 'hello world' example - I'm trying to build 'hello_world' on MAC, and following error occurred.
**System information** (Please provide as much relevant information as possible)
Error in download_and_extract: java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.32.0/rules_apple.0.32.0.tar.gz] to /private/var/tmp/_bazel_wangjunjie/95eda4f7c99f7ddce46cd8d36dd7bed7/external/build_bazel_rules_apple/temp11825797914784900860/rules_apple.0.32.0.tar.gz: Connection refused (Connection refused)
ERROR: no such package '@build_bazel_rules_apple//apple': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.32.0/rules_apple.0.32.0.tar.gz] to /private/var/tmp/_bazel_wangjunjie/95eda4f7c99f7ddce46cd8d36dd7bed7/external/build_bazel_rules_apple/temp11825797914784900860/rules_apple.0.32.0.tar.gz: Connection refused (Connection refused)
INFO: Elapsed time: 12.881s
INFO: 0 processes.
**Describe the problem**:
seems like the package `build_bazel_rules_apple//apple` i can't download. But when i copy the link to the web browser, i can visit. Need help~
| build | 【mac】an error occurred building hello world example i m trying to build hello world on mac and following error occurred system information please provide as much relevant information as possible error in download and extract java io ioexception error downloading to private var tmp bazel wangjunjie external build bazel rules apple rules apple tar gz connection refused connection refused error no such package build bazel rules apple apple java io ioexception error downloading to private var tmp bazel wangjunjie external build bazel rules apple rules apple tar gz connection refused connection refused info elapsed time info processes describe the problem seems like the package build bazel rules apple apple i can t download but when i copy the link to the web browser i can visit need help | 1 |
191,217 | 6,826,984,178 | IssuesEvent | 2017-11-08 15:44:20 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | Lost/Stolen: MAC address in user email is not populated | Priority: Medium Type: Bug | The placeholder for the MAC address doesn't seem to be valid as seen in the screenshot below

| 1.0 | Lost/Stolen: MAC address in user email is not populated - The placeholder for the MAC address doesn't seem to be valid as seen in the screenshot below

| non_build | lost stolen mac address in user email is not populated the placeholder for the mac address doesn t seem to be valid as seen in the screenshot below | 0 |
59,399 | 14,580,932,818 | IssuesEvent | 2020-12-18 09:56:43 | intellij-rust/intellij-rust | https://api.github.com/repos/intellij-rust/intellij-rust | closed | Failed to render build output error | bug subsystem::build & run subsystem::tools | <!--
Hello and thank you for the issue!
If you would like to report a bug, we have added some points below that you can fill out.
Feel free to remove all the irrelevant text to request a new feature.
-->
## Environment
* **IntelliJ Rust plugin version: `0.2.116.2844-193-nightly`**
* **Rust toolchain version: `rustc 1.43.0-nightly (58b834344 2020-02-05)`**
* **IDE name and version: CLion 19.3.3**
* **Operating system: Win 10**
## Problem description
Failed to render error message
## Steps to reproduce
This problem pops up only under some specific circumstances that I cannot reproduce, so I cannot provide something like a minimal erroneous sample code easily. However, I attached the error message below that might be helpful.
Sample Build Output (irrelevant details are hidden):
```
C:/Users/?/.cargo/bin/cargo.exe build --color=always --package ? --example ? --message-format=json
Compiling ? v0.1.0 (?)
{"reason":"compiler-message","package_id":"? 0.1.0 (path+file:///?)","target":{"kind":["lib"],"crate_types":["lib"],"name":"?","src_path":"?\\src\\lib.rs","edition":"2018","doctest":true},"message":{"rendered":"error[E0505]: cannot move out of `stream` because it is borrowed\n --> src\\lib.rs:74:50\n |\n73 | poll.registry().register(\n | -------- borrow later used by call\n74 | &mut stream, conn_mgr.generate_token(stream),\n | ----------- ^^^^^^ move out of `stream` occurs here\n | |\n | borrow of `stream` occurs here\n\n","children":[],"code":{"code":"E0505","explanation":"A value was moved out while it was still borrowed.\n\nErroneous code example:\n\n```compile_fail,E0505\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: Value) {}\n\nfn main() {\n let x = Value{};\n let _ref_to_val: &Value = &x;\n eat(x);\n borrow(_ref_to_val);\n}\n```\n\nHere, the function `eat` takes ownership of `x`. However,\n`x` cannot be moved because the borrow to `_ref_to_val`\nneeds to last till the function `borrow`.\nTo fix that you can do a few different things:\n\n* Try to avoid moving the variable.\n* Release borrow before move.\n* Implement the `Copy` trait on the type.\n\nExamples:\n\n```\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: &Value) {}\n\nfn main() {\n let x = Value{};\n\n let ref_to_val: &Value = &x;\n eat(&x); // pass by reference, if it's possible\n borrow(ref_to_val);\n}\n```\n\nOr:\n\n```\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: Value) {}\n\nfn main() {\n let x = Value{};\n\n let ref_to_val: &Value = &x;\n borrow(ref_to_val);\n // ref_to_val is no longer used.\n eat(x);\n}\n```\n\nOr:\n\n```\n#[derive(Clone, Copy)] // implement Copy trait\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: Value) {}\n\nfn main() {\n let x = Value{};\n let ref_to_val: &Value = &x;\n eat(x); // it will be copied here.\n borrow(ref_to_val);\n}\n```\n\nYou can find more information about borrowing in the rust-book:\nhttp://doc.rust-lang.org/book/ch04-02-references-and-borrowing.html\n"},"level":"error","message":"cannot move out of `stream` because it is borrowed","spans":[{"byte_end":2018,"byte_start":2007,"column_end":24,"column_start":13,"expansion":nu
ll,"file_name":"src\\lib.rs","is_primary":false,"label":"borrow of `stream` occurs here","line_end":74,"line_start":74,"suggested_replacement":null,"suggestion_applicability":null,"text":[{"highlight_end":24,"highlight_start":13,"text":" &mut stream, conn_mgr.generate_token(stream),"}]},{"byte_end":2050,"byte_start":2044,"column_end":56,"column_start":50,"expansion":null,"file_name":"src\\lib.rs","is_primary":true,"label":"move out of `stream` occurs here","line_end":74,"line_start":74,"suggested_replacement":null,"suggestion_applicability":null,"text":[{"highlight_end":56,"highlight_start":50,"text":" &mut stream, conn_mgr.generate_token(stream),"}]},{"byte_end":1992,"byte_start":1984,"column_end":35,"column_start":27,"expansion":null,"file_name":"src\\lib.rs","is_primary":false,"label":"borrow later used by call","line_end":73,"line_start":73,"suggested_replacement":null,"suggestion_applicability":null,"text":[{"highlight_end":35,"highlight_start":27,"text":" poll.registry().register("}]}]}}
error: aborting due to previous error
For more information about this error, try `rustc --explain E0505`.
error: could not compile `hello_server`.
To learn more, run the command again with --verbose.
Process finished with exit code 101
``` | 1.0 | Failed to render build output error - <!--
Hello and thank you for the issue!
If you would like to report a bug, we have added some points below that you can fill out.
Feel free to remove all the irrelevant text to request a new feature.
-->
## Environment
* **IntelliJ Rust plugin version: `0.2.116.2844-193-nightly`**
* **Rust toolchain version: `rustc 1.43.0-nightly (58b834344 2020-02-05)`**
* **IDE name and version: CLion 19.3.3**
* **Operating system: Win 10**
## Problem description
Failed to render error message
## Steps to reproduce
This problem pops up only under some specific circumstances that I cannot reproduce, so I cannot provide something like a minimal erroneous sample code easily. However, I attached the error message below that might be helpful.
Sample Build Output (irrelevant details are hidden):
```
C:/Users/?/.cargo/bin/cargo.exe build --color=always --package ? --example ? --message-format=json
Compiling ? v0.1.0 (?)
{"reason":"compiler-message","package_id":"? 0.1.0 (path+file:///?)","target":{"kind":["lib"],"crate_types":["lib"],"name":"?","src_path":"?\\src\\lib.rs","edition":"2018","doctest":true},"message":{"rendered":"error[E0505]: cannot move out of `stream` because it is borrowed\n --> src\\lib.rs:74:50\n |\n73 | poll.registry().register(\n | -------- borrow later used by call\n74 | &mut stream, conn_mgr.generate_token(stream),\n | ----------- ^^^^^^ move out of `stream` occurs here\n | |\n | borrow of `stream` occurs here\n\n","children":[],"code":{"code":"E0505","explanation":"A value was moved out while it was still borrowed.\n\nErroneous code example:\n\n```compile_fail,E0505\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: Value) {}\n\nfn main() {\n let x = Value{};\n let _ref_to_val: &Value = &x;\n eat(x);\n borrow(_ref_to_val);\n}\n```\n\nHere, the function `eat` takes ownership of `x`. However,\n`x` cannot be moved because the borrow to `_ref_to_val`\nneeds to last till the function `borrow`.\nTo fix that you can do a few different things:\n\n* Try to avoid moving the variable.\n* Release borrow before move.\n* Implement the `Copy` trait on the type.\n\nExamples:\n\n```\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: &Value) {}\n\nfn main() {\n let x = Value{};\n\n let ref_to_val: &Value = &x;\n eat(&x); // pass by reference, if it's possible\n borrow(ref_to_val);\n}\n```\n\nOr:\n\n```\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: Value) {}\n\nfn main() {\n let x = Value{};\n\n let ref_to_val: &Value = &x;\n borrow(ref_to_val);\n // ref_to_val is no longer used.\n eat(x);\n}\n```\n\nOr:\n\n```\n#[derive(Clone, Copy)] // implement Copy trait\nstruct Value {}\n\nfn borrow(val: &Value) {}\n\nfn eat(val: Value) {}\n\nfn main() {\n let x = Value{};\n let ref_to_val: &Value = &x;\n eat(x); // it will be copied here.\n borrow(ref_to_val);\n}\n```\n\nYou can find more information about borrowing in the rust-book:\nhttp://doc.rust-lang.org/book/ch04-02-references-and-borrowing.html\n"},"level":"error","message":"cannot move out of `stream` because it is borrowed","spans":[{"byte_end":2018,"byte_start":2007,"column_end":24,"column_start":13,"expansion":nu
ll,"file_name":"src\\lib.rs","is_primary":false,"label":"borrow of `stream` occurs here","line_end":74,"line_start":74,"suggested_replacement":null,"suggestion_applicability":null,"text":[{"highlight_end":24,"highlight_start":13,"text":" &mut stream, conn_mgr.generate_token(stream),"}]},{"byte_end":2050,"byte_start":2044,"column_end":56,"column_start":50,"expansion":null,"file_name":"src\\lib.rs","is_primary":true,"label":"move out of `stream` occurs here","line_end":74,"line_start":74,"suggested_replacement":null,"suggestion_applicability":null,"text":[{"highlight_end":56,"highlight_start":50,"text":" &mut stream, conn_mgr.generate_token(stream),"}]},{"byte_end":1992,"byte_start":1984,"column_end":35,"column_start":27,"expansion":null,"file_name":"src\\lib.rs","is_primary":false,"label":"borrow later used by call","line_end":73,"line_start":73,"suggested_replacement":null,"suggestion_applicability":null,"text":[{"highlight_end":35,"highlight_start":27,"text":" poll.registry().register("}]}]}}
error: aborting due to previous error
For more information about this error, try `rustc --explain E0505`.
error: could not compile `hello_server`.
To learn more, run the command again with --verbose.
Process finished with exit code 101
``` | build | failed to render build output error hello and thank you for the issue if you would like to report a bug we have added some points below that you can fill out feel free to remove all the irrelevant text to request a new feature environment intellij rust plugin version nightly rust toolchain version rustc nightly ide name and version clion operating system win problem description failed to render error message steps to reproduce this problem pops up only under some specific circumstances that i cannot reproduce so i cannot provide something like a minimal erroneous sample code easily however i attached the error message below that might be helpful sample build output irrelevant details are hidden c users cargo bin cargo exe build color always package example message format json compiling reason compiler message package id path file target kind crate types name src path src lib rs edition doctest true message rendered error cannot move out of stream because it is borrowed n src lib rs n poll registry register n borrow later used by call mut stream conn mgr generate token stream n move out of stream occurs here n n borrow of stream occurs here n n children code code explanation a value was moved out while it was still borrowed n nerroneous code example n n compile fail nstruct value n nfn borrow val value n nfn eat val value n nfn main n let x value n let ref to val value x n eat x n borrow ref to val n n n nhere the function eat takes ownership of x however n x cannot be moved because the borrow to ref to val nneeds to last till the function borrow nto fix that you can do a few different things n n try to avoid moving the variable n release borrow before move n implement the copy trait on the type n nexamples n n nstruct value n nfn borrow val value n nfn eat val value n nfn main n let x value n n let ref to val value x n eat x pass by reference if it s possible n borrow ref to val n n n nor n n nstruct value n nfn borrow val value n nfn eat val value n nfn main n let x value n n let ref to val value x n borrow ref to val n ref to val is no longer used n eat x n n n nor n n n implement copy trait nstruct value n nfn borrow val value n nfn eat val value n nfn main n let x value n let ref to val value x n eat x it will be copied here n borrow ref to val n n n nyou can find more information about borrowing in the rust book n move out of stream because it is borrowed spans byte end byte start column end column start expansion nu ll file name src lib rs is primary false label borrow of stream occurs here line end line start suggested replacement null suggestion applicability null text byte end byte start column end column start expansion null file name src lib rs is primary true label move out of stream occurs here line end line start suggested replacement null suggestion applicability null text byte end byte start column end column start expansion null file name src lib rs is primary false label borrow later used by call line end line start suggested replacement null suggestion applicability null text error aborting due to previous error for more information about this error try rustc explain error could not compile hello server to learn more run the command again with verbose process finished with exit code | 1 |
13,032 | 10,082,496,267 | IssuesEvent | 2019-07-25 11:24:17 | moosetechnology/Moose | https://api.github.com/repos/moosetechnology/Moose | closed | The #allWithType: method is not scalable | FAME FAMIX Infrastructure | This method consumes more than 99% of my importer time.
The problem is the #updateCacheOnAddingOf: method that:
- call the #mooseTypes method to obtain the class + all famix traits of the entity class in a really unefficient way
- create one collection for each of those types and add the entity to those collections
A better strategy should be applied, like having only collection for real classes | 1.0 | The #allWithType: method is not scalable - This method consumes more than 99% of my importer time.
The problem is the #updateCacheOnAddingOf: method that:
- call the #mooseTypes method to obtain the class + all famix traits of the entity class in a really unefficient way
- create one collection for each of those types and add the entity to those collections
A better strategy should be applied, like having only collection for real classes | non_build | the allwithtype method is not scalable this method consumes more than of my importer time the problem is the updatecacheonaddingof method that call the moosetypes method to obtain the class all famix traits of the entity class in a really unefficient way create one collection for each of those types and add the entity to those collections a better strategy should be applied like having only collection for real classes | 0 |
5,704 | 30,081,740,273 | IssuesEvent | 2023-06-29 04:30:16 | Apple-Dev-Projects/RoamX | https://api.github.com/repos/Apple-Dev-Projects/RoamX | closed | Configure CI/CD build action | maintain | Setup gated check-ins and create hosted build action for repository | True | Configure CI/CD build action - Setup gated check-ins and create hosted build action for repository | non_build | configure ci cd build action setup gated check ins and create hosted build action for repository | 0 |
341,960 | 10,310,137,311 | IssuesEvent | 2019-08-29 14:36:49 | yalla-coop/earwig | https://api.github.com/repos/yalla-coop/earwig | opened | Change these four button labels to 'Skip' on the status check pages | priority-2 | Screenshots of how it should look here: https://drive.google.com/open?id=1GMooSdX9GpDdkkZcOq5VuCi0dKAKHJW_
| 1.0 | Change these four button labels to 'Skip' on the status check pages - Screenshots of how it should look here: https://drive.google.com/open?id=1GMooSdX9GpDdkkZcOq5VuCi0dKAKHJW_
| non_build | change these four button labels to skip on the status check pages screenshots of how it should look here | 0 |
79,988 | 7,735,663,638 | IssuesEvent | 2018-05-27 17:33:40 | futtta/ao_critcss_aas | https://api.github.com/repos/futtta/ao_critcss_aas | closed | advanced option; force noptimize | to be tested | in some (very) rare cases ccss.com has issues parsing AO's CSS, in which case having an (advanced) option to force the plugin to feed ccss.com a URL with `?ao_noptimize=1` appended to the to be CCSS'ed URL would make sense. | 1.0 | advanced option; force noptimize - in some (very) rare cases ccss.com has issues parsing AO's CSS, in which case having an (advanced) option to force the plugin to feed ccss.com a URL with `?ao_noptimize=1` appended to the to be CCSS'ed URL would make sense. | non_build | advanced option force noptimize in some very rare cases ccss com has issues parsing ao s css in which case having an advanced option to force the plugin to feed ccss com a url with ao noptimize appended to the to be ccss ed url would make sense | 0 |
237,224 | 7,757,607,305 | IssuesEvent | 2018-05-31 16:50:40 | JiscRDSS/rdss-canonical-data-model | https://api.github.com/repos/JiscRDSS/rdss-canonical-data-model | closed | data admin workflow | alpha pilot feedback 1 priority:High use case | There doesn't seem to be any support for data admin workflow. This was one of the areas where we identified a weakness in Pure and built something around it into our 'RDYork' middleware/dashboard. Pure has a workflow status (in progress/for validation/validated/for revalidation) but we are finding it helpful to be able to record specific actions that have happened to a dataset, or need to happen, (e.g. waiting for info from researcher; dataset will be deposited but hasn't yet). Since everyone will have their own workflow, I can see this being a tricky area, though!
Examples
• Researcher contacted - awaiting info
• Researcher contacted - awaiting data
• DOI created
• DOI not yet created
• PhD student deposit
• Data externally held
• Data held by Dept
• Upload in progress [system generated - i.e. upload to Archivematica. Ditto the following three bullet points]
• Data uploaded
• Upload problem
• Data not uploaded
• Metadata held for publication [i.e. not to be released until associated paper published - this is common]
• Data held for publication
• Data available [for public download]
• Waiting for approval [i.e. someone has requested access and a manual step is required to approve it]
• Request approved, data in progress
• Data not available [I think this means we've refused an access request]
[Yor] | 1.0 | data admin workflow - There doesn't seem to be any support for data admin workflow. This was one of the areas where we identified a weakness in Pure and built something around it into our 'RDYork' middleware/dashboard. Pure has a workflow status (in progress/for validation/validated/for revalidation) but we are finding it helpful to be able to record specific actions that have happened to a dataset, or need to happen, (e.g. waiting for info from researcher; dataset will be deposited but hasn't yet). Since everyone will have their own workflow, I can see this being a tricky area, though!
Examples
• Researcher contacted - awaiting info
• Researcher contacted - awaiting data
• DOI created
• DOI not yet created
• PhD student deposit
• Data externally held
• Data held by Dept
• Upload in progress [system generated - i.e. upload to Archivematica. Ditto the following three bullet points]
• Data uploaded
• Upload problem
• Data not uploaded
• Metadata held for publication [i.e. not to be released until associated paper published - this is common]
• Data held for publication
• Data available [for public download]
• Waiting for approval [i.e. someone has requested access and a manual step is required to approve it]
• Request approved, data in progress
• Data not available [I think this means we've refused an access request]
[Yor] | non_build | data admin workflow there doesn t seem to be any support for data admin workflow this was one of the areas where we identified a weakness in pure and built something around it into our rdyork middleware dashboard pure has a workflow status in progress for validation validated for revalidation but we are finding it helpful to be able to record specific actions that have happened to a dataset or need to happen e g waiting for info from researcher dataset will be deposited but hasn t yet since everyone will have their own workflow i can see this being a tricky area though examples • researcher contacted awaiting info • researcher contacted awaiting data • doi created • doi not yet created • phd student deposit • data externally held • data held by dept • upload in progress • data uploaded • upload problem • data not uploaded • metadata held for publication • data held for publication • data available • waiting for approval • request approved data in progress • data not available | 0 |
77,309 | 26,910,166,030 | IssuesEvent | 2023-02-06 22:46:14 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | [Content organization] Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page. (02.04.3) | content design 508/Accessibility ia 508-defect-2 collab-cycle-feedback Staging CCIssue02.04 CC-Dashboard my-education-benefits education-letters | ### General Information
#### VFS team name
DGIB - My Education Benefits
#### VFS product name
VA Education Letters
#### VFS feature name
#### Point of Contact/Reviewers
Brian DeConinck - @briandeconinck - Accessibility
*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
---
### Platform Issue
Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page.
### Issue Details
In the "How do I download..." section, three of the four steps given are just a heading with no additional content following. Headings are typically used to indicate sections of content, and screen reader users often navigate page content by jumping to the heading of the section they want to read. Headings with no content could be confusing.
### Link, screenshot or steps to recreate
### VA.gov Experience Standard
[Category Number 02, Issue Number 04](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
### Other References
WCAG SC 1.3.1_A
WCAG SC 2.4.6_AA
### Platform Recommendation
I think the process list component might not be a good fit for the instructions given here. The component works really well when you have high-level steps that can be described briefly with a heading, but require a paragraph or two of explanation to actually follow the steps.
Consider reworking the steps to be a little broader (eg. combine the last three steps into one step with a brief heading and details), or else just use a standard `ol` ordered list rather than the process list component.
If it's important to use a process list component with these steps specifically, you may be able to use `p` elements in place of the headings for each step. I'm not aware of any other places on VA.gov that do this, but it would avoid the heading navigation issue while preserving the visual style.
---
### VFS Guidance
- Close the ticket when the issue has been resolved or validated by your Product Owner
- If your team has additional questions or needs Platform help validating the issue, please comment on the ticket
- Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue.
- If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up. | 1.0 | [Content organization] Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page. (02.04.3) - ### General Information
#### VFS team name
DGIB - My Education Benefits
#### VFS product name
VA Education Letters
#### VFS feature name
#### Point of Contact/Reviewers
Brian DeConinck - @briandeconinck - Accessibility
*For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website.
---
### Platform Issue
Your page doesn't follow heading hierarchy. Headings aren't properly nested, and headings and/or labels aren't relevant to the page.
### Issue Details
In the "How do I download..." section, three of the four steps given are just a heading with no additional content following. Headings are typically used to indicate sections of content, and screen reader users often navigate page content by jumping to the heading of the section they want to read. Headings with no content could be confusing.
### Link, screenshot or steps to recreate
### VA.gov Experience Standard
[Category Number 02, Issue Number 04](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html)
### Other References
WCAG SC 1.3.1_A
WCAG SC 2.4.6_AA
### Platform Recommendation
I think the process list component might not be a good fit for the instructions given here. The component works really well when you have high-level steps that can be described briefly with a heading, but require a paragraph or two of explanation to actually follow the steps.
Consider reworking the steps to be a little broader (eg. combine the last three steps into one step with a brief heading and details), or else just use a standard `ol` ordered list rather than the process list component.
If it's important to use a process list component with these steps specifically, you may be able to use `p` elements in place of the headings for each step. I'm not aware of any other places on VA.gov that do this, but it would avoid the heading navigation issue while preserving the visual style.
---
### VFS Guidance
- Close the ticket when the issue has been resolved or validated by your Product Owner
- If your team has additional questions or needs Platform help validating the issue, please comment on the ticket
- Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue.
- If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up. | non_build | your page doesn t follow heading hierarchy headings aren t properly nested and headings and or labels aren t relevant to the page general information vfs team name dgib my education benefits vfs product name va education letters vfs feature name point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue your page doesn t follow heading hierarchy headings aren t properly nested and headings and or labels aren t relevant to the page issue details in the how do i download section three of the four steps given are just a heading with no additional content following headings are typically used to indicate sections of content and screen reader users often navigate page content by jumping to the heading of the section they want to read headings with no content could be confusing link screenshot or steps to recreate va gov experience standard other references wcag sc a wcag sc aa platform recommendation i think the process list component might not be a good fit for the instructions given here the component works really well when you have high level steps that can be described briefly with a heading but require a paragraph or two of explanation to actually follow the steps consider reworking the steps to be a little broader eg combine the last three steps into one step with a brief heading and details or else just use a standard ol ordered list rather than the process list component if it s important to use a process list component with these steps specifically you may be able to use p elements in place of the headings for each step i m not aware of any other places on va gov that do this but it would avoid the heading navigation issue while preserving the visual style vfs guidance close the ticket when the issue has been resolved or validated by your product owner if your team has additional questions or needs platform help validating the issue please comment on the ticket some feedback provided may be out of scope for your iteration of the product however platform s octo leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue if you do not believe that this staging review issue ticket is the responsibility of your team comment below providing an explanation and who you believe is responsible please tag the point of contact reviewers governance team will research and will follow up | 0 |
12,959 | 5,279,279,581 | IssuesEvent | 2017-02-07 10:48:03 | sandaemc/house | https://api.github.com/repos/sandaemc/house | opened | Buy initial materials | building | - [ ] 4 cu. sand (Php3,000)
- [ ]3 cu. Gravel 3/4 (Php2,500)
- 500 pcs CHB (Php6,000)
- 16 pcs 16mm RSB (Php4,480)
- 8 pcs 12mm RSB (Php1,220)
- 60 pcs 10mm RSB (Php6,240)
- 30 pcs Cement (Php6,840)
- 8 pcs plywood 1/4 ordinary (Php1,150)
- 5 kilos c.w nails assorted
- 10 kilos tie wire
- 40 pcs 2x2x8 coco lumber | 1.0 | Buy initial materials - - [ ] 4 cu. sand (Php3,000)
- [ ]3 cu. Gravel 3/4 (Php2,500)
- 500 pcs CHB (Php6,000)
- 16 pcs 16mm RSB (Php4,480)
- 8 pcs 12mm RSB (Php1,220)
- 60 pcs 10mm RSB (Php6,240)
- 30 pcs Cement (Php6,840)
- 8 pcs plywood 1/4 ordinary (Php1,150)
- 5 kilos c.w nails assorted
- 10 kilos tie wire
- 40 pcs 2x2x8 coco lumber | build | buy initial materials cu sand cu gravel pcs chb pcs rsb pcs rsb pcs rsb pcs cement pcs plywood ordinary kilos c w nails assorted kilos tie wire pcs coco lumber | 1 |
1,134 | 2,532,289,990 | IssuesEvent | 2015-01-23 15:09:37 | AAndharia/ZIMS-School-Mgmt | https://api.github.com/repos/AAndharia/ZIMS-School-Mgmt | closed | student fee receipt | bug Tested & Verified | when I select user Enrollment Date , Total fee and Outstanding should be filled but it is not see screenshot:-

after it when I trying to submit I got following error see screen shot below

| 1.0 | student fee receipt - when I select user Enrollment Date , Total fee and Outstanding should be filled but it is not see screenshot:-

after it when I trying to submit I got following error see screen shot below

| non_build | student fee receipt when i select user enrollment date total fee and outstanding should be filled but it is not see screenshot after it when i trying to submit i got following error see screen shot below | 0 |
161,543 | 20,154,098,556 | IssuesEvent | 2022-02-09 15:01:05 | kapseliboi/crowdfunding-frontend | https://api.github.com/repos/kapseliboi/crowdfunding-frontend | opened | CVE-2018-3728 (High) detected in hoek-2.16.3.tgz | security vulnerability | ## CVE-2018-3728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hoek/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.11.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/crowdfunding-frontend/commit/3c5fb7d3f2932ae2023353faafd51bf99b953a5f">3c5fb7d3f2932ae2023353faafd51bf99b953a5f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution (hoek): 4.2.0</p>
<p>Direct dependency fix Resolution (nodemon): 1.11.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-3728 (High) detected in hoek-2.16.3.tgz - ## CVE-2018-3728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hoek/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-1.11.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/kapseliboi/crowdfunding-frontend/commit/3c5fb7d3f2932ae2023353faafd51bf99b953a5f">3c5fb7d3f2932ae2023353faafd51bf99b953a5f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16082</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution (hoek): 4.2.0</p>
<p>Direct dependency fix Resolution (nodemon): 1.11.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_build | cve high detected in hoek tgz cve high severity vulnerability vulnerable library hoek tgz general purpose node utilities library home page a href path to dependency file package json path to vulnerable library node modules hoek package json dependency hierarchy nodemon tgz root library chokidar tgz fsevents tgz node pre gyp tgz request tgz hawk tgz x hoek tgz vulnerable library found in head commit a href found in base branch master vulnerability details hoek node module before and x before suffers from a modification of assumed immutable data maid vulnerability via merge and applytodefaults functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hoek direct dependency fix resolution nodemon step up your open source security game with whitesource | 0 |
88,999 | 25,558,951,543 | IssuesEvent | 2022-11-30 09:18:42 | firecracker-microvm/firecracker | https://api.github.com/repos/firecracker-microvm/firecracker | closed | [Bug] `cargo update` results in un-buildable workspace | Codebase: Build System | # Describe the bug
Running `cargo update` or deleting `Cargo.lock` results in an un-buildable workspace in the sense that `cargo build` fails with compilation errors due to incompatible dependencies.
## To Reproduce
1. Run `cargo update` or delete `Cargo.lock`
2. Run `cargo build` (or `./tools/devtools build`)
## Expected behaviour
The dependencies specified in the various `Cargo.toml` files are compatible and the build succeeds
## Additional context
This is due to us using `>=` instead of `=` to specify dependency versions for crates that do not follow semver (e.g. crates prior to version one, especially `rust-vmm`). For example, `vmm-sys-utils=0.10.0` does not export `ioctl_expr` anymore, but since `utils/Cargo.toml` specifies `vmm-sys-utils>=0.8.0`, this version is downloaded by `cargo update`.
To fix this, we need to go through `Cargo.lock` and lock our dependencies to versions that are proven to compile.
## Checks
- [x] Have you searched the Firecracker Issues database for similar problems?
- [x] Have you read the existing relevant Firecracker documentation?
- [x] Are you certain the bug being reported is a Firecracker issue?
| 1.0 | [Bug] `cargo update` results in un-buildable workspace - # Describe the bug
Running `cargo update` or deleting `Cargo.lock` results in an un-buildable workspace in the sense that `cargo build` fails with compilation errors due to incompatible dependencies.
## To Reproduce
1. Run `cargo update` or delete `Cargo.lock`
2. Run `cargo build` (or `./tools/devtools build`)
## Expected behaviour
The dependencies specified in the various `Cargo.toml` files are compatible and the build succeeds
## Additional context
This is due to us using `>=` instead of `=` to specify dependency versions for crates that do not follow semver (e.g. crates prior to version one, especially `rust-vmm`). For example, `vmm-sys-utils=0.10.0` does not export `ioctl_expr` anymore, but since `utils/Cargo.toml` specifies `vmm-sys-utils>=0.8.0`, this version is downloaded by `cargo update`.
To fix this, we need to go through `Cargo.lock` and lock our dependencies to versions that are proven to compile.
## Checks
- [x] Have you searched the Firecracker Issues database for similar problems?
- [x] Have you read the existing relevant Firecracker documentation?
- [x] Are you certain the bug being reported is a Firecracker issue?
| build | cargo update results in un buildable workspace describe the bug running cargo update or deleting cargo lock results in an un buildable workspace in the sense that cargo build fails with compilation errors due to incompatible dependencies to reproduce run cargo update or delete cargo lock run cargo build or tools devtools build expected behaviour the dependencies specified in the various cargo toml files are compatible and the build succeeds additional context this is due to us using instead of to specify dependency versions for crates that do not follow semver e g crates prior to version one especially rust vmm for example vmm sys utils does not export ioctl expr anymore but since utils cargo toml specifies vmm sys utils this version is downloaded by cargo update to fix this we need to go through cargo lock and lock our dependencies to versions that are proven to compile checks have you searched the firecracker issues database for similar problems have you read the existing relevant firecracker documentation are you certain the bug being reported is a firecracker issue | 1 |
101,276 | 30,988,976,752 | IssuesEvent | 2023-08-09 01:51:48 | tensorflow/tfjs | https://api.github.com/repos/tensorflow/tfjs | closed | @tensorflow/tfjs-tflite@0.0.1-alpha.9" has incorrect peer dependency "@tensorflow/tfjs-core@3.20.0 | type:build/install stat:awaiting response comp:tfjs-tflite stale | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04):
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow.js installed from (npm or script link):
- TensorFlow.js version:
- CUDA/cuDNN version:
**Describe the problem**
**Provide the exact sequence of commands / steps that you executed before running into the problem**
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| 1.0 | @tensorflow/tfjs-tflite@0.0.1-alpha.9" has incorrect peer dependency "@tensorflow/tfjs-core@3.20.0 - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04):
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow.js installed from (npm or script link):
- TensorFlow.js version:
- CUDA/cuDNN version:
**Describe the problem**
**Provide the exact sequence of commands / steps that you executed before running into the problem**
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| build | tensorflow tfjs tflite alpha has incorrect peer dependency tensorflow tfjs core please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow js installed from npm or script link tensorflow js version cuda cudnn version describe the problem provide the exact sequence of commands steps that you executed before running into the problem any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached | 1 |
15,792 | 6,039,459,688 | IssuesEvent | 2017-06-10 03:20:11 | nodejs/node | https://api.github.com/repos/nodejs/node | closed | Empty list of builds in /download/v8-canary/index.json | build | List of builds in `index.js` file [here](https://nodejs.org/download/v8-canary/) is empty so it's currently not possible to install v8-canary builds via nvm like so:
```bash
NVM_NODEJS_ORG_MIRROR=https://nodejs.org/download/v8-canary/ nvm i node
``` | 1.0 | Empty list of builds in /download/v8-canary/index.json - List of builds in `index.js` file [here](https://nodejs.org/download/v8-canary/) is empty so it's currently not possible to install v8-canary builds via nvm like so:
```bash
NVM_NODEJS_ORG_MIRROR=https://nodejs.org/download/v8-canary/ nvm i node
``` | build | empty list of builds in download canary index json list of builds in index js file is empty so it s currently not possible to install canary builds via nvm like so bash nvm nodejs org mirror nvm i node | 1 |
86,318 | 24,817,720,763 | IssuesEvent | 2022-10-25 14:16:51 | netlify/build | https://api.github.com/repos/netlify/build | closed | Inputs API draft | Discussion type: feature project/build-plugins-v1 theme/plugin-author-experience theme/plugin-user-experience stale | ## Overview
`inputs` is a property that can be defined in a plugin's `manifest.yml` file. An "input" is an expected value that can be provided to the plugin at runtime. Here's an example input definition:
```yml
inputs:
- name: phone
label: Twilio Phone Number
required: true
pattern: /^([0-9\(\)\/\+ \-]*)$/
sensitive: true
```
Not all inputs are required. An input definition can be used to validate user input, as well as configure the UI that might be presented for the user to provide their input. For example, a "select" type input might result in an HTML `<select>` element for the user to select a value. Note that inputs might come from anywhere (Eg., another plugin, CLI, etc.), so a UI may not always be involved.
## Why this is important
This API is meant to drive a dynamic UI in the Netlify dashboard, completely configured through a plugin's `manifest.yml`. It is similar in concept to the editor pane in Netlify CMS, except:
- we're only configuring a form
- nothing is previewed
- only simple values are handled (no crazy markdown widgets)
## Default input properties
These properties are available to all inputs, regardless of type.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------|
| `name` |**yes** | |name for internal reference purposes |
| `label` |**yes** | |proper display name for UI |
| `type` | |`"string"` |UI input type (see Available types below) |
| `description` | | |Longer description |
| `required` | |`false` |Set field as required |
| `default` | | |Default value used if none supplied |
| `repeat` | |`false` |Whether the field can be repeated - produces an array of values if `true`|
| `repeat_min` | | |Minimum repeat count (if `repeat` is `true`) |
| `repeat_max` | | |Maximum repeat count (if `repeat` is `true`) |
Here's an example use case for the `repeat` property, where a plugin accepts a list of email addresses. The user could add as many email addresses as necessary since `repeat` is set to `true`.
```yml
inputs:
- name: emails
label: Email Addresses
required: true
repeat: true
repeat_min: 1
```
## Available types
A list of types that we should allow, and type-specific properties that should be accepted for each.
### `string`
Accepts a string value.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `pattern` | | |Regular expression that this string should match. |
| `sensitive` | |`false` |Whether to mask the value to prevent it from being displayed in logs, CLI, or API. |
### `number`
Accepts a numeric value.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `min` | | |Minimum acceptable value. |
| `max` | | |Maximum acceptable value. |
### `boolean`
Accepts a boolean value.
***No type-specific properties.***
### `text`
Longer string value.
***No type-specific properties.***
### `select`
Accepts one or more values from a predefined list.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `options` |**yes** | |List of values that can be selected. |
| `multiple` | |`false` |Allow multiple values to be selected |
### `object`
Accepts a map of inputs.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `inputs` |**yes** | |Map of input definitions |
Here's an example input configuration that accepts an array of address objects.
```yml
inputs:
- name: addresses
label: Addresses
type: object
required: true
repeat: true
repeat_min: 1
inputs:
- name: street_address
label: Street Address
required: true
- name: city
label: City
required: true
- name: state
label: State
required: true
``` | 1.0 | Inputs API draft - ## Overview
`inputs` is a property that can be defined in a plugin's `manifest.yml` file. An "input" is an expected value that can be provided to the plugin at runtime. Here's an example input definition:
```yml
inputs:
- name: phone
label: Twilio Phone Number
required: true
pattern: /^([0-9\(\)\/\+ \-]*)$/
sensitive: true
```
Not all inputs are required. An input definition can be used to validate user input, as well as configure the UI that might be presented for the user to provide their input. For example, a "select" type input might result in an HTML `<select>` element for the user to select a value. Note that inputs might come from anywhere (Eg., another plugin, CLI, etc.), so a UI may not always be involved.
## Why this is important
This API is meant to drive a dynamic UI in the Netlify dashboard, completely configured through a plugin's `manifest.yml`. It is similar in concept to the editor pane in Netlify CMS, except:
- we're only configuring a form
- nothing is previewed
- only simple values are handled (no crazy markdown widgets)
## Default input properties
These properties are available to all inputs, regardless of type.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------|
| `name` |**yes** | |name for internal reference purposes |
| `label` |**yes** | |proper display name for UI |
| `type` | |`"string"` |UI input type (see Available types below) |
| `description` | | |Longer description |
| `required` | |`false` |Set field as required |
| `default` | | |Default value used if none supplied |
| `repeat` | |`false` |Whether the field can be repeated - produces an array of values if `true`|
| `repeat_min` | | |Minimum repeat count (if `repeat` is `true`) |
| `repeat_max` | | |Maximum repeat count (if `repeat` is `true`) |
Here's an example use case for the `repeat` property, where a plugin accepts a list of email addresses. The user could add as many email addresses as necessary since `repeat` is set to `true`.
```yml
inputs:
- name: emails
label: Email Addresses
required: true
repeat: true
repeat_min: 1
```
## Available types
A list of types that we should allow, and type-specific properties that should be accepted for each.
### `string`
Accepts a string value.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `pattern` | | |Regular expression that this string should match. |
| `sensitive` | |`false` |Whether to mask the value to prevent it from being displayed in logs, CLI, or API. |
### `number`
Accepts a numeric value.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `min` | | |Minimum acceptable value. |
| `max` | | |Maximum acceptable value. |
### `boolean`
Accepts a boolean value.
***No type-specific properties.***
### `text`
Longer string value.
***No type-specific properties.***
### `select`
Accepts one or more values from a predefined list.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `options` |**yes** | |List of values that can be selected. |
| `multiple` | |`false` |Allow multiple values to be selected |
### `object`
Accepts a map of inputs.
|key |required |default |description |
|:---------------|:---------|:----------|:--------------------------------------------------|
| `inputs` |**yes** | |Map of input definitions |
Here's an example input configuration that accepts an array of address objects.
```yml
inputs:
- name: addresses
label: Addresses
type: object
required: true
repeat: true
repeat_min: 1
inputs:
- name: street_address
label: Street Address
required: true
- name: city
label: City
required: true
- name: state
label: State
required: true
``` | build | inputs api draft overview inputs is a property that can be defined in a plugin s manifest yml file an input is an expected value that can be provided to the plugin at runtime here s an example input definition yml inputs name phone label twilio phone number required true pattern sensitive true not all inputs are required an input definition can be used to validate user input as well as configure the ui that might be presented for the user to provide their input for example a select type input might result in an html element for the user to select a value note that inputs might come from anywhere eg another plugin cli etc so a ui may not always be involved why this is important this api is meant to drive a dynamic ui in the netlify dashboard completely configured through a plugin s manifest yml it is similar in concept to the editor pane in netlify cms except we re only configuring a form nothing is previewed only simple values are handled no crazy markdown widgets default input properties these properties are available to all inputs regardless of type key required default description name yes name for internal reference purposes label yes proper display name for ui type string ui input type see available types below description longer description required false set field as required default default value used if none supplied repeat false whether the field can be repeated produces an array of values if true repeat min minimum repeat count if repeat is true repeat max maximum repeat count if repeat is true here s an example use case for the repeat property where a plugin accepts a list of email addresses the user could add as many email addresses as necessary since repeat is set to true yml inputs name emails label email addresses required true repeat true repeat min available types a list of types that we should allow and type specific properties that should be accepted for each string accepts a string value key required default description pattern regular expression that this string should match sensitive false whether to mask the value to prevent it from being displayed in logs cli or api number accepts a numeric value key required default description min minimum acceptable value max maximum acceptable value boolean accepts a boolean value no type specific properties text longer string value no type specific properties select accepts one or more values from a predefined list key required default description options yes list of values that can be selected multiple false allow multiple values to be selected object accepts a map of inputs key required default description inputs yes map of input definitions here s an example input configuration that accepts an array of address objects yml inputs name addresses label addresses type object required true repeat true repeat min inputs name street address label street address required true name city label city required true name state label state required true | 1 |
16,649 | 6,259,771,850 | IssuesEvent | 2017-07-14 18:52:40 | grpc/grpc | https://api.github.com/repos/grpc/grpc | closed | c++_linux_opt.bins/opt/hybrid_end2end_test --gtest_filter=HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService GRPC_POLL_STRATEGY=poll (from grpc.c++_linux_opt) | infra/BUILDPONY lang/c++ priority/P0 | Standard Output
D1215 16:29:11.170558720 28875 test_config.c:290] test slowdown: machine=1.000000 build=1.000000 poll=1.000000 total=1.000000
Note: Google Test filter = HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService
[==========] Running 1 test from 1 test case.
[----------] Global test environment set-up.
[----------] 1 test from HybridEnd2endTest
[ RUN ] HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService
D1215 16:29:11.170822380 28875 ev_posix.c:105] Using polling engine: poll
D1215 16:29:11.171998810 28875 ev_posix.c:105] Using polling engine: poll
I1215 16:29:11.172032957 28875 server_builder.cc:247] Synchronous server. Num CQs: 16, Min pollers: 1, Max Pollers: 2147483647, CQ timeout (msec): 1000
I1215 16:29:11.234500675 30624 test_service_impl.cc:305] recv msg hello0
I1215 16:29:11.234682417 30624 test_service_impl.cc:305] recv msg hello1
I1215 16:29:11.234818595 30624 test_service_impl.cc:305] recv msg hello2
test/cpp/end2end/hybrid_end2end_test.cc:70: Failure
Value of: VerifyReturnSuccess(cq, i)
Actual: false
Expected: expect_ok
Which is: true
[ FAILED ] HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService (1118 ms)
[----------] 1 test from HybridEnd2endTest (1118 ms total)
[----------] Global test environment tear-down
[==========] 1 test from 1 test case ran. (1118 ms total)
[ PASSED ] 0 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService
1 FAILED TEST
D1215 16:29:12.288335972 28875 ev_posix.c:105] Using polling engine: poll | 1.0 | c++_linux_opt.bins/opt/hybrid_end2end_test --gtest_filter=HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService GRPC_POLL_STRATEGY=poll (from grpc.c++_linux_opt) - Standard Output
D1215 16:29:11.170558720 28875 test_config.c:290] test slowdown: machine=1.000000 build=1.000000 poll=1.000000 total=1.000000
Note: Google Test filter = HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService
[==========] Running 1 test from 1 test case.
[----------] Global test environment set-up.
[----------] 1 test from HybridEnd2endTest
[ RUN ] HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService
D1215 16:29:11.170822380 28875 ev_posix.c:105] Using polling engine: poll
D1215 16:29:11.171998810 28875 ev_posix.c:105] Using polling engine: poll
I1215 16:29:11.172032957 28875 server_builder.cc:247] Synchronous server. Num CQs: 16, Min pollers: 1, Max Pollers: 2147483647, CQ timeout (msec): 1000
I1215 16:29:11.234500675 30624 test_service_impl.cc:305] recv msg hello0
I1215 16:29:11.234682417 30624 test_service_impl.cc:305] recv msg hello1
I1215 16:29:11.234818595 30624 test_service_impl.cc:305] recv msg hello2
test/cpp/end2end/hybrid_end2end_test.cc:70: Failure
Value of: VerifyReturnSuccess(cq, i)
Actual: false
Expected: expect_ok
Which is: true
[ FAILED ] HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService (1118 ms)
[----------] 1 test from HybridEnd2endTest (1118 ms total)
[----------] Global test environment tear-down
[==========] 1 test from 1 test case ran. (1118 ms total)
[ PASSED ] 0 tests.
[ FAILED ] 1 test, listed below:
[ FAILED ] HybridEnd2endTest.AsyncRequestStreamResponseStream_AsyncDupService
1 FAILED TEST
D1215 16:29:12.288335972 28875 ev_posix.c:105] Using polling engine: poll | build | c linux opt bins opt hybrid test gtest filter asyncrequeststreamresponsestream asyncdupservice grpc poll strategy poll from grpc c linux opt standard output test config c test slowdown machine build poll total note google test filter asyncrequeststreamresponsestream asyncdupservice running test from test case global test environment set up test from asyncrequeststreamresponsestream asyncdupservice ev posix c using polling engine poll ev posix c using polling engine poll server builder cc synchronous server num cqs min pollers max pollers cq timeout msec test service impl cc recv msg test service impl cc recv msg test service impl cc recv msg test cpp hybrid test cc failure value of verifyreturnsuccess cq i actual false expected expect ok which is true asyncrequeststreamresponsestream asyncdupservice ms test from ms total global test environment tear down test from test case ran ms total tests test listed below asyncrequeststreamresponsestream asyncdupservice failed test ev posix c using polling engine poll | 1 |
258,602 | 22,331,578,605 | IssuesEvent | 2022-06-14 14:54:36 | filecoin-project/builtin-actors | https://api.github.com/repos/filecoin-project/builtin-actors | closed | market / miner withdrawal integration test | P1 area/test | refer to https://github.com/filecoin-project/specs-actors/blob/master/actors/test/market_miner_withdrawal_test.go
- [x] TestMarketWithdraw
- [x] TestMinerWithdraw | 1.0 | market / miner withdrawal integration test - refer to https://github.com/filecoin-project/specs-actors/blob/master/actors/test/market_miner_withdrawal_test.go
- [x] TestMarketWithdraw
- [x] TestMinerWithdraw | non_build | market miner withdrawal integration test refer to testmarketwithdraw testminerwithdraw | 0 |
10,158 | 2,618,939,610 | IssuesEvent | 2015-03-03 00:03:35 | marmarek/test | https://api.github.com/repos/marmarek/test | closed | Dom0 kernel: better nouveau suport | C: xen P: major R: fixed T: defect | **Reported by joanna on 27 Apr 40264910 04:00 UTC**
The currently used pvops0 kernel (2.6.32.9-7.qubes) is based on xen/stable-2.6.32.9 branch and Fedora backported nouveau's patches. This kernel, when run as Dom0 under Xen, doesn't support HW accel for nouveau driver, the default driver for NVidia graphics cards.
The users can still get around the problem, but this requires complicated setup, and also doesn't let the user to use various "desktop effects" in Dom0. | 1.0 | Dom0 kernel: better nouveau suport - **Reported by joanna on 27 Apr 40264910 04:00 UTC**
The currently used pvops0 kernel (2.6.32.9-7.qubes) is based on xen/stable-2.6.32.9 branch and Fedora backported nouveau's patches. This kernel, when run as Dom0 under Xen, doesn't support HW accel for nouveau driver, the default driver for NVidia graphics cards.
The users can still get around the problem, but this requires complicated setup, and also doesn't let the user to use various "desktop effects" in Dom0. | non_build | kernel better nouveau suport reported by joanna on apr utc the currently used kernel qubes is based on xen stable branch and fedora backported nouveau s patches this kernel when run as under xen doesn t support hw accel for nouveau driver the default driver for nvidia graphics cards the users can still get around the problem but this requires complicated setup and also doesn t let the user to use various desktop effects in | 0 |
239,302 | 7,788,511,190 | IssuesEvent | 2018-06-07 05:11:30 | robinagist/ezo | https://api.github.com/repos/robinagist/ezo | closed | view contracts and deployments | enhancement high priority | + view all contracts
+ view all compilations of a contract
+ view deployments of a contract | 1.0 | view contracts and deployments - + view all contracts
+ view all compilations of a contract
+ view deployments of a contract | non_build | view contracts and deployments view all contracts view all compilations of a contract view deployments of a contract | 0 |
27,110 | 4,281,088,848 | IssuesEvent | 2016-07-15 00:28:24 | bitwalker/bundler | https://api.github.com/repos/bitwalker/bundler | opened | Test suite | help wanted testing | Need a solid test suite which can be run in CI, with coverage reporting.
Tests should run/work on Erlang 18.3 and 19.x, using Elixir 1.3+ | 1.0 | Test suite - Need a solid test suite which can be run in CI, with coverage reporting.
Tests should run/work on Erlang 18.3 and 19.x, using Elixir 1.3+ | non_build | test suite need a solid test suite which can be run in ci with coverage reporting tests should run work on erlang and x using elixir | 0 |
75,173 | 20,710,565,865 | IssuesEvent | 2022-03-12 00:10:19 | gakonst/foundry | https://api.github.com/repos/gakonst/foundry | closed | forge installs prerelease instead of release versions of required solidity compiler | T-bug C-forge Cmd-forge-build | ### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
forge 0.1.0 (7ed07e7 2022-03-10T00:04:06.029112+00:00)
### What command(s) is the bug in?
forge build
### Operating System
macOS (M1)
### Describe the bug
Even after an empty `.svm` directory forge always installs a prerelease version of the requested compiler `0.8.11-develop` instead of `0.8.11`.
`Ayush` on telegram suggested the error might be related to the fact, that ethers-rs installs the compiler via the following repo:
[`https://github.com/roynalnaruto/solc-builds/tree/master/macosx/aarch6](https://github.com/roynalnaruto/solc-builds/tree/master/macosx/aarch64)
Could be only arm mac related, afaik.
```
~/git/contracts-fresh main*
❯ forge build
Compiling...
installing solc version "0.8.11"
Successfully installed solc 0.8.11
Compiling 82 files with 0.8.11
Error:
0: Compiler run failed
Warning: This is a pre-release compiler version, please do not use it in production.
ParserError: Source file requires different compiler version (current compiler is 0.8.11-develop.2022.2.16+commit.d7f03943.Darwin.appleclang) - note that nightly builds are considered to be strictly less than the released version
--> /Users/peet/git/contracts-fresh/contracts/compound/strategies/AlpacaERC4626.sol:2:1:
|
2 | pragma solidity 0.8.11;
| ^^^^^^^^^^^^^^^^^^^^^^^
``` | 1.0 | forge installs prerelease instead of release versions of required solidity compiler - ### Component
Forge
### Have you ensured that all of these are up to date?
- [X] Foundry
- [X] Foundryup
### What version of Foundry are you on?
forge 0.1.0 (7ed07e7 2022-03-10T00:04:06.029112+00:00)
### What command(s) is the bug in?
forge build
### Operating System
macOS (M1)
### Describe the bug
Even after an empty `.svm` directory forge always installs a prerelease version of the requested compiler `0.8.11-develop` instead of `0.8.11`.
`Ayush` on telegram suggested the error might be related to the fact, that ethers-rs installs the compiler via the following repo:
[`https://github.com/roynalnaruto/solc-builds/tree/master/macosx/aarch6](https://github.com/roynalnaruto/solc-builds/tree/master/macosx/aarch64)
Could be only arm mac related, afaik.
```
~/git/contracts-fresh main*
❯ forge build
Compiling...
installing solc version "0.8.11"
Successfully installed solc 0.8.11
Compiling 82 files with 0.8.11
Error:
0: Compiler run failed
Warning: This is a pre-release compiler version, please do not use it in production.
ParserError: Source file requires different compiler version (current compiler is 0.8.11-develop.2022.2.16+commit.d7f03943.Darwin.appleclang) - note that nightly builds are considered to be strictly less than the released version
--> /Users/peet/git/contracts-fresh/contracts/compound/strategies/AlpacaERC4626.sol:2:1:
|
2 | pragma solidity 0.8.11;
| ^^^^^^^^^^^^^^^^^^^^^^^
``` | build | forge installs prerelease instead of release versions of required solidity compiler component forge have you ensured that all of these are up to date foundry foundryup what version of foundry are you on forge what command s is the bug in forge build operating system macos describe the bug even after an empty svm directory forge always installs a prerelease version of the requested compiler develop instead of ayush on telegram suggested the error might be related to the fact that ethers rs installs the compiler via the following repo could be only arm mac related afaik git contracts fresh main ❯ forge build compiling installing solc version successfully installed solc compiling files with error compiler run failed warning this is a pre release compiler version please do not use it in production parsererror source file requires different compiler version current compiler is develop commit darwin appleclang note that nightly builds are considered to be strictly less than the released version users peet git contracts fresh contracts compound strategies sol pragma solidity | 1 |
43,669 | 11,276,989,503 | IssuesEvent | 2020-01-15 01:08:13 | streamnative/bookkeeper | https://api.github.com/repos/streamnative/bookkeeper | opened | ISSUE-1923: src package is missing stream/clients/python/scripts/docker_build.sh | area/build area/tests triage/week-7 type/bug | Original Issue: apache/bookkeeper#1923
---
This prevents integration tests to be run against the src package. | 1.0 | ISSUE-1923: src package is missing stream/clients/python/scripts/docker_build.sh - Original Issue: apache/bookkeeper#1923
---
This prevents integration tests to be run against the src package. | build | issue src package is missing stream clients python scripts docker build sh original issue apache bookkeeper this prevents integration tests to be run against the src package | 1 |
496,296 | 14,344,617,718 | IssuesEvent | 2020-11-28 15:18:00 | super-cooper/memebot | https://api.github.com/repos/super-cooper/memebot | opened | Command callbacks are called regardless of whether or not command fails | bug good first issue low-priority | **Description**
If a Command fails for whatever reason, there is nothing to communicate the outcome of the command and the callback is called regardless.
**To Reproduce**
1. Run `!poll`
2. Observe that the callback runs even though the command fails gracefully
**Expected behavior**
On command failure, the callback should be skipped.
**Additional context**
We could maybe optionally allow a callback to run if commands fail under certain circumstances, but I can't imagine a scenario where regular cleanup in `Command.exec` would not suffice. | 1.0 | Command callbacks are called regardless of whether or not command fails - **Description**
If a Command fails for whatever reason, there is nothing to communicate the outcome of the command and the callback is called regardless.
**To Reproduce**
1. Run `!poll`
2. Observe that the callback runs even though the command fails gracefully
**Expected behavior**
On command failure, the callback should be skipped.
**Additional context**
We could maybe optionally allow a callback to run if commands fail under certain circumstances, but I can't imagine a scenario where regular cleanup in `Command.exec` would not suffice. | non_build | command callbacks are called regardless of whether or not command fails description if a command fails for whatever reason there is nothing to communicate the outcome of the command and the callback is called regardless to reproduce run poll observe that the callback runs even though the command fails gracefully expected behavior on command failure the callback should be skipped additional context we could maybe optionally allow a callback to run if commands fail under certain circumstances but i can t imagine a scenario where regular cleanup in command exec would not suffice | 0 |
659,235 | 21,919,884,206 | IssuesEvent | 2022-05-22 12:09:23 | alan-turing-institute/environmental-ds-book | https://api.github.com/repos/alan-turing-institute/environmental-ds-book | closed | [DOC] It's quite hard to actually find content in the book | enhancement medium priority | <!--- Thanks for taking the time to request a feature! -->
<!--- Please go through the sections below -->
## Feature Request
<!--- Is your feature request related to a problem? Please describe -->
<!--- A clear and concise description of what the problem is. Ex. I have an issue when [...] -->
I was recommended the book during a presentation about it and was told this was the link to browse.
https://the-environmental-ds-book.netlify.app/welcome.html
When I browsed the book I first clicked on Oceans but it says "under construction" I then clicked "Forests" it said the same. I think clicked on Forest: Sensors. It also said under construction.
As a casual reader it was difficult to find the non-blank parts of the book. To be honest, if I had had less time I would have simply assumed there was an error and the book was completely empty.
## Describe the solution you'd like
<!--- A clear and concise description of what you want to happen (where possible provide a link to a working version of the feature). Add any considered drawbacks -->
Is there some way the filled in sections could be highlighted more prominently or the generic sections (E.g. Forest, Ocean) could contains pointers to actual content.
## Describe alternatives you've considered
<!--- A clear and concise description of any alternative solutions or features you've considered -->
Any mechanism that made the filled in sections more prominent than the non filled in ones would work. But I'm not familiar enough with the technology that turns text into the book to make a coherent suggestion.
## Additional information
<!--- Add any other context or screenshots about the feature request here --> | 1.0 | [DOC] It's quite hard to actually find content in the book - <!--- Thanks for taking the time to request a feature! -->
<!--- Please go through the sections below -->
## Feature Request
<!--- Is your feature request related to a problem? Please describe -->
<!--- A clear and concise description of what the problem is. Ex. I have an issue when [...] -->
I was recommended the book during a presentation about it and was told this was the link to browse.
https://the-environmental-ds-book.netlify.app/welcome.html
When I browsed the book I first clicked on Oceans but it says "under construction" I then clicked "Forests" it said the same. I think clicked on Forest: Sensors. It also said under construction.
As a casual reader it was difficult to find the non-blank parts of the book. To be honest, if I had had less time I would have simply assumed there was an error and the book was completely empty.
## Describe the solution you'd like
<!--- A clear and concise description of what you want to happen (where possible provide a link to a working version of the feature). Add any considered drawbacks -->
Is there some way the filled in sections could be highlighted more prominently or the generic sections (E.g. Forest, Ocean) could contains pointers to actual content.
## Describe alternatives you've considered
<!--- A clear and concise description of any alternative solutions or features you've considered -->
Any mechanism that made the filled in sections more prominent than the non filled in ones would work. But I'm not familiar enough with the technology that turns text into the book to make a coherent suggestion.
## Additional information
<!--- Add any other context or screenshots about the feature request here --> | non_build | it s quite hard to actually find content in the book feature request i was recommended the book during a presentation about it and was told this was the link to browse when i browsed the book i first clicked on oceans but it says under construction i then clicked forests it said the same i think clicked on forest sensors it also said under construction as a casual reader it was difficult to find the non blank parts of the book to be honest if i had had less time i would have simply assumed there was an error and the book was completely empty describe the solution you d like is there some way the filled in sections could be highlighted more prominently or the generic sections e g forest ocean could contains pointers to actual content describe alternatives you ve considered any mechanism that made the filled in sections more prominent than the non filled in ones would work but i m not familiar enough with the technology that turns text into the book to make a coherent suggestion additional information | 0 |
52,314 | 12,949,234,295 | IssuesEvent | 2020-07-19 08:15:28 | BEEmod/BEE2.4 | https://api.github.com/repos/BEEmod/BEE2.4 | closed | Compiler crashes on Python 3.8: required field "type_ignores" missing from Module | Compile Error Development Build Fixed | Building and running the compiler on Python 3.8 causes it to crash with the following error:
```
Traceback (most recent call last):
File "conditions\__init__.py", line 279, in setup_result
File "<string>", line 1, in <lambda>
File "conditions\python.py", line 207, in res_python_setup
TypeError: required field "type_ignores" missing from Module
```
I reported this to TeamSpen a while back but it hasn't been fixed, so I figured I'd open this as a reminder. | 1.0 | Compiler crashes on Python 3.8: required field "type_ignores" missing from Module - Building and running the compiler on Python 3.8 causes it to crash with the following error:
```
Traceback (most recent call last):
File "conditions\__init__.py", line 279, in setup_result
File "<string>", line 1, in <lambda>
File "conditions\python.py", line 207, in res_python_setup
TypeError: required field "type_ignores" missing from Module
```
I reported this to TeamSpen a while back but it hasn't been fixed, so I figured I'd open this as a reminder. | build | compiler crashes on python required field type ignores missing from module building and running the compiler on python causes it to crash with the following error traceback most recent call last file conditions init py line in setup result file line in file conditions python py line in res python setup typeerror required field type ignores missing from module i reported this to teamspen a while back but it hasn t been fixed so i figured i d open this as a reminder | 1 |
334,008 | 10,134,672,250 | IssuesEvent | 2019-08-02 08:12:31 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | mail.rajasthan.gov.in - site is not usable | browser-firefox engine-gecko priority-normal | <!-- @browser: Firefox 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://mail.rajasthan.gov.in/owa/#path=/mail
**Browser / Version**: Firefox 69.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Site is not usable
**Description**: SCREEN IS BUZZING
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/8/1a5c0146-f25a-40f3-bf1d-af355f83740f.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190712011116</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | mail.rajasthan.gov.in - site is not usable - <!-- @browser: Firefox 69.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; rv:69.0) Gecko/20100101 Firefox/69.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://mail.rajasthan.gov.in/owa/#path=/mail
**Browser / Version**: Firefox 69.0
**Operating System**: Windows 7
**Tested Another Browser**: Unknown
**Problem type**: Site is not usable
**Description**: SCREEN IS BUZZING
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/8/1a5c0146-f25a-40f3-bf1d-af355f83740f.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190712011116</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_build | mail rajasthan gov in site is not usable url browser version firefox operating system windows tested another browser unknown problem type site is not usable description screen is buzzing steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta from with ❤️ | 0 |
91,603 | 26,433,727,619 | IssuesEvent | 2023-01-15 05:13:57 | logseq/logseq | https://api.github.com/repos/logseq/logseq | closed | Security issue | build :status/automatic-stale | ### What happened?
How can I report a vulnerability without disclosing it publicly in a GH issue?
Best
### Reproduce the Bug
.
### Expected Behavior
_No response_
### Screenshots
_No response_
### Desktop Platform Information
_No response_
### Mobile Platform Information
_No response_
### Additional Context
_No response_ | 1.0 | Security issue - ### What happened?
How can I report a vulnerability without disclosing it publicly in a GH issue?
Best
### Reproduce the Bug
.
### Expected Behavior
_No response_
### Screenshots
_No response_
### Desktop Platform Information
_No response_
### Mobile Platform Information
_No response_
### Additional Context
_No response_ | build | security issue what happened how can i report a vulnerability without disclosing it publicly in a gh issue best reproduce the bug expected behavior no response screenshots no response desktop platform information no response mobile platform information no response additional context no response | 1 |
24,802 | 17,787,869,864 | IssuesEvent | 2021-08-31 13:13:21 | Opentrons/opentrons | https://api.github.com/repos/Opentrons/opentrons | closed | chore(python): add Black as Python formatter | chore infrastructure cpx | ## Overview
We've been testing out [Black](https://github.com/psf/black) in other Opentrons projects as a Python code formatter. While the project remains in a beta prerelease state (and has been for some time), we're comfortable enough with it that we'd like to bring it in as a Python formatter to ensure consistent style and reduce cognitive load
For Opentrons folks, see [discussion in Slack](https://opentrons.slack.com/archives/CTLS4MBB4/p1618317957006500)
## Acceptance criteria
- [ ] Black added as dev dependency to Python projects
- [x] robot-server
- [ ] api
- [ ] Project Makefiles have `format` targets added
- [x] robot-server
- [ ] api
- [x] Top-level Makefile has `format-py` target added as a dependency to a `format` target
- [x] CI lint tasks checks formatting via `black --check` | 1.0 | chore(python): add Black as Python formatter - ## Overview
We've been testing out [Black](https://github.com/psf/black) in other Opentrons projects as a Python code formatter. While the project remains in a beta prerelease state (and has been for some time), we're comfortable enough with it that we'd like to bring it in as a Python formatter to ensure consistent style and reduce cognitive load
For Opentrons folks, see [discussion in Slack](https://opentrons.slack.com/archives/CTLS4MBB4/p1618317957006500)
## Acceptance criteria
- [ ] Black added as dev dependency to Python projects
- [x] robot-server
- [ ] api
- [ ] Project Makefiles have `format` targets added
- [x] robot-server
- [ ] api
- [x] Top-level Makefile has `format-py` target added as a dependency to a `format` target
- [x] CI lint tasks checks formatting via `black --check` | non_build | chore python add black as python formatter overview we ve been testing out in other opentrons projects as a python code formatter while the project remains in a beta prerelease state and has been for some time we re comfortable enough with it that we d like to bring it in as a python formatter to ensure consistent style and reduce cognitive load for opentrons folks see acceptance criteria black added as dev dependency to python projects robot server api project makefiles have format targets added robot server api top level makefile has format py target added as a dependency to a format target ci lint tasks checks formatting via black check | 0 |
94,915 | 27,326,961,209 | IssuesEvent | 2023-02-25 05:43:34 | arrayfire/arrayfire-rust | https://api.github.com/repos/arrayfire/arrayfire-rust | closed | [Build] Half 2.2.0 is not compatible | Bug Build | Description
===========
The latest version of half 2.2.0 is not compatible with arrayfire-rust 3.8.0
Build Environment
-----------------
Compiler version: <!-- MSVC v140 or gcc 9.3.2 -->
Operating system: <!-- Windows 10; Ubuntu 18.04 -->
Build environment: <!-- Environment variables; Installed software -->
Error Log
---------
```
use half;
use half::f16;
use half::*;
fn main() {
let arr = vec![half::f16::from_f32(5.4), half::f16::from_f32(-2.1),half::f16::from_f32(-2.1) ];
let mut test = arrayfire::Array::new(&arr, arrayfire::Dim4::new(&[3, 1, 1, 1]));
}
```
error[E0277]: the trait bound `f16: HasAfEnum` is not satisfied
--> examples/test.rs:27:17
|
27 | let mut test = arrayfire::Array::new(&arr, arrayfire::Dim4::new(&[3, 1, 1, 1]));
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the trait `HasAfEnum` is not implemented for `f16`
|
= help: the following other types implement trait `HasAfEnum`:
bool
f32
f64
half::binary16::f16
i16
i32
i64
num_complex::Complex<f32>
and 5 others
note: required by a bound in `Array`
--> /home/test/.cargo/registry/src/github.com-1ecc6299db9ec823/arrayfire-3.8.0/src/core/array.rs:163:21
|
163 | pub struct Array<T: HasAfEnum> {
| ^^^^^^^^^ required by this bound in `Array`
| 1.0 | [Build] Half 2.2.0 is not compatible - Description
===========
The latest version of half 2.2.0 is not compatible with arrayfire-rust 3.8.0
Build Environment
-----------------
Compiler version: <!-- MSVC v140 or gcc 9.3.2 -->
Operating system: <!-- Windows 10; Ubuntu 18.04 -->
Build environment: <!-- Environment variables; Installed software -->
Error Log
---------
```
use half;
use half::f16;
use half::*;
fn main() {
let arr = vec![half::f16::from_f32(5.4), half::f16::from_f32(-2.1),half::f16::from_f32(-2.1) ];
let mut test = arrayfire::Array::new(&arr, arrayfire::Dim4::new(&[3, 1, 1, 1]));
}
```
error[E0277]: the trait bound `f16: HasAfEnum` is not satisfied
--> examples/test.rs:27:17
|
27 | let mut test = arrayfire::Array::new(&arr, arrayfire::Dim4::new(&[3, 1, 1, 1]));
| ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ the trait `HasAfEnum` is not implemented for `f16`
|
= help: the following other types implement trait `HasAfEnum`:
bool
f32
f64
half::binary16::f16
i16
i32
i64
num_complex::Complex<f32>
and 5 others
note: required by a bound in `Array`
--> /home/test/.cargo/registry/src/github.com-1ecc6299db9ec823/arrayfire-3.8.0/src/core/array.rs:163:21
|
163 | pub struct Array<T: HasAfEnum> {
| ^^^^^^^^^ required by this bound in `Array`
| build | half is not compatible description the latest version of half is not compatible with arrayfire rust build environment compiler version operating system build environment error log use half use half use half fn main let arr vec let mut test arrayfire array new arr arrayfire new error the trait bound hasafenum is not satisfied examples test rs let mut test arrayfire array new arr arrayfire new the trait hasafenum is not implemented for help the following other types implement trait hasafenum bool half num complex complex and others note required by a bound in array home test cargo registry src github com arrayfire src core array rs pub struct array required by this bound in array | 1 |
90,272 | 11,362,149,080 | IssuesEvent | 2020-01-26 19:25:54 | Techie09/readr | https://api.github.com/repos/Techie09/readr | opened | On Login, Begin a Session | 5pt - Design & Work Core functionality | After the end-user logs in, the start of a new session should prompt the end-user for an ISBN. The End-user may enter in the ISBN and can optionally enter a description of the session as well, or if the End-user cannot provide an ISBN, then the end-user can skip the ISBN requirement, but the Description field will be required.
*[ ] Prompt User for an ISBN
*[ ] If an ISBN is provided, optionally allow the user to set a description
*[ ] if the ISBN is not provided, require the user to set a description. | 1.0 | On Login, Begin a Session - After the end-user logs in, the start of a new session should prompt the end-user for an ISBN. The End-user may enter in the ISBN and can optionally enter a description of the session as well, or if the End-user cannot provide an ISBN, then the end-user can skip the ISBN requirement, but the Description field will be required.
*[ ] Prompt User for an ISBN
*[ ] If an ISBN is provided, optionally allow the user to set a description
*[ ] if the ISBN is not provided, require the user to set a description. | non_build | on login begin a session after the end user logs in the start of a new session should prompt the end user for an isbn the end user may enter in the isbn and can optionally enter a description of the session as well or if the end user cannot provide an isbn then the end user can skip the isbn requirement but the description field will be required prompt user for an isbn if an isbn is provided optionally allow the user to set a description if the isbn is not provided require the user to set a description | 0 |
221,181 | 16,996,414,264 | IssuesEvent | 2021-07-01 07:06:53 | Orbital-Mulgo/Project | https://api.github.com/repos/Orbital-Mulgo/Project | opened | Task: Requirement Analysis | documentation extension feature testing | <!-- Issue title should mirror the Task Title. -->
# Task Title
Task: Requirement Analysis
## Task Description
This task will look to state the requirements of what is to be tested.
## Main Issue
<!-- The link below should link to its Main Issue. -->
This task is part of [Feature: Automated Testing](https://github.com/Orbital-Mulgo/Project/issues/34).
| 1.0 | Task: Requirement Analysis - <!-- Issue title should mirror the Task Title. -->
# Task Title
Task: Requirement Analysis
## Task Description
This task will look to state the requirements of what is to be tested.
## Main Issue
<!-- The link below should link to its Main Issue. -->
This task is part of [Feature: Automated Testing](https://github.com/Orbital-Mulgo/Project/issues/34).
| non_build | task requirement analysis task title task requirement analysis task description this task will look to state the requirements of what is to be tested main issue this task is part of | 0 |
44,089 | 11,364,858,009 | IssuesEvent | 2020-01-27 09:21:28 | bbtsoftware/BBT.Maybe | https://api.github.com/repos/bbtsoftware/BBT.Maybe | opened | Local build fails | Build | At the moment, locally running `./build.ps1` fails for me:
> PS D:\_repo\BBT.Maybe> ./build.ps1
> Preparing to run build script...
> Unable to find version '0.32.1' of package 'Cake'.
> C:\Users\xx\.nuget\packages\: Package 'Cake.0.32.1' is not found on source 'C:\Users\xx\.nuget\packages\'.
https://www.myget.org/F/cake/api/v3/index.json: Package 'Cake.0.32.1' is not found on source 'https://www.myget.org/F/cake/api/v3/index.json'.
> An error occured while restoring NuGet tools.
> In D:\_repo\BBT.Maybe\build.ps1:166 Zeichen:9
> Throw "An error occured while restoring NuGet tools."
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : OperationStopped: (An error occure...ng NuGet tools.:String) [], RuntimeException
+ FullyQualifiedErrorId : An error occured while restoring NuGet tools.
| 1.0 | Local build fails - At the moment, locally running `./build.ps1` fails for me:
> PS D:\_repo\BBT.Maybe> ./build.ps1
> Preparing to run build script...
> Unable to find version '0.32.1' of package 'Cake'.
> C:\Users\xx\.nuget\packages\: Package 'Cake.0.32.1' is not found on source 'C:\Users\xx\.nuget\packages\'.
https://www.myget.org/F/cake/api/v3/index.json: Package 'Cake.0.32.1' is not found on source 'https://www.myget.org/F/cake/api/v3/index.json'.
> An error occured while restoring NuGet tools.
> In D:\_repo\BBT.Maybe\build.ps1:166 Zeichen:9
> Throw "An error occured while restoring NuGet tools."
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : OperationStopped: (An error occure...ng NuGet tools.:String) [], RuntimeException
+ FullyQualifiedErrorId : An error occured while restoring NuGet tools.
| build | local build fails at the moment locally running build fails for me ps d repo bbt maybe build preparing to run build script unable to find version of package cake c users xx nuget packages package cake is not found on source c users xx nuget packages package cake is not found on source an error occured while restoring nuget tools in d repo bbt maybe build zeichen throw an error occured while restoring nuget tools categoryinfo operationstopped an error occure ng nuget tools string runtimeexception fullyqualifiederrorid an error occured while restoring nuget tools | 1 |
53,725 | 13,198,209,515 | IssuesEvent | 2020-08-14 01:42:18 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | GCC truncation warnings with --enable-debug on Fedora 32 | Type: Building good first issue |
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Fedora
Distribution Version | 32
Linux Kernel | 5.7.14-200.fc32.x86_64
Architecture | x86-64
ZFS Version | master (faa296c73c7ccd535c0874d4f7e8f7c4ea43eea6)
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
I'm seeing truncation warnings when building ZFS using `./configure --enable-debug` on Fedora 32. If I build without `--enable-debug`, I don't see the warnings.
### Describe how to reproduce the problem
```
./autogen.sh && ./configure --enable-debug && make
```
### Include any warning/errors/backtraces from the system logs
```
CC os/linux/libzfs_mount_os.lo
CC os/linux/libzfs_pool_os.lo
CC os/linux/libzfs_sendrecv_os.lo
os/linux/libzfs_mount_os.c: In function ‘zfs_selinux_setcontext’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4093 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
268 | append_mntopt(name, context, mntopts, mtabopt, B_TRUE);
| ~~~~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output 5 or more bytes (assuming 4100) into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
os/linux/libzfs_mount_os.c: In function ‘zfs_adjust_mount_options’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4086 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
297 | append_mntopt(MNTOPT_CONTEXT, prop,
| ~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output between 12 and 4107 bytes into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
make[3]: *** [Makefile:966: os/linux/libzfs_mount_os.lo] Error 1
make[3]: *** Waiting for unfinished jobs....
make[3]: Leaving directory '/home/fedora/zfs/lib/libzfs'
make[2]: *** [Makefile:660: all-recursive] Error 1
make[2]: Leaving directory '/home/fedora/zfs/lib'
make[1]: *** [Makefile:870: all-recursive] Error 1
make[1]: Leaving directory '/home/fedora/zfs'
make: *** [Makefile:733: all] Error 2
``` | 1.0 | GCC truncation warnings with --enable-debug on Fedora 32 -
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Fedora
Distribution Version | 32
Linux Kernel | 5.7.14-200.fc32.x86_64
Architecture | x86-64
ZFS Version | master (faa296c73c7ccd535c0874d4f7e8f7c4ea43eea6)
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
I'm seeing truncation warnings when building ZFS using `./configure --enable-debug` on Fedora 32. If I build without `--enable-debug`, I don't see the warnings.
### Describe how to reproduce the problem
```
./autogen.sh && ./configure --enable-debug && make
```
### Include any warning/errors/backtraces from the system logs
```
CC os/linux/libzfs_mount_os.lo
CC os/linux/libzfs_pool_os.lo
CC os/linux/libzfs_sendrecv_os.lo
os/linux/libzfs_mount_os.c: In function ‘zfs_selinux_setcontext’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4093 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
268 | append_mntopt(name, context, mntopts, mtabopt, B_TRUE);
| ~~~~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output 5 or more bytes (assuming 4100) into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
os/linux/libzfs_mount_os.c: In function ‘zfs_adjust_mount_options’:
os/linux/libzfs_mount_os.c:250:45: error: ‘%s’ directive output may be truncated writing up to 4095 bytes into a region of size 4086 [-Werror=format-truncation=]
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~
......
297 | append_mntopt(MNTOPT_CONTEXT, prop,
| ~~~~
os/linux/libzfs_mount_os.c:250:2: note: ‘snprintf’ output between 12 and 4107 bytes into a destination of size 4096
250 | snprintf(tmp, MNT_LINE_MAX, quote ? ",%s=\"%s\"" : ",%s=%s", name, val);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
cc1: all warnings being treated as errors
make[3]: *** [Makefile:966: os/linux/libzfs_mount_os.lo] Error 1
make[3]: *** Waiting for unfinished jobs....
make[3]: Leaving directory '/home/fedora/zfs/lib/libzfs'
make[2]: *** [Makefile:660: all-recursive] Error 1
make[2]: Leaving directory '/home/fedora/zfs/lib'
make[1]: *** [Makefile:870: all-recursive] Error 1
make[1]: Leaving directory '/home/fedora/zfs'
make: *** [Makefile:733: all] Error 2
``` | build | gcc truncation warnings with enable debug on fedora system information type version name distribution name fedora distribution version linux kernel architecture zfs version master commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing i m seeing truncation warnings when building zfs using configure enable debug on fedora if i build without enable debug i don t see the warnings describe how to reproduce the problem autogen sh configure enable debug make include any warning errors backtraces from the system logs cc os linux libzfs mount os lo cc os linux libzfs pool os lo cc os linux libzfs sendrecv os lo os linux libzfs mount os c in function ‘zfs selinux setcontext’ os linux libzfs mount os c error ‘ s’ directive output may be truncated writing up to bytes into a region of size snprintf tmp mnt line max quote s s s s name val append mntopt name context mntopts mtabopt b true os linux libzfs mount os c note ‘snprintf’ output or more bytes assuming into a destination of size snprintf tmp mnt line max quote s s s s name val os linux libzfs mount os c in function ‘zfs adjust mount options’ os linux libzfs mount os c error ‘ s’ directive output may be truncated writing up to bytes into a region of size snprintf tmp mnt line max quote s s s s name val append mntopt mntopt context prop os linux libzfs mount os c note ‘snprintf’ output between and bytes into a destination of size snprintf tmp mnt line max quote s s s s name val all warnings being treated as errors make error make waiting for unfinished jobs make leaving directory home fedora zfs lib libzfs make error make leaving directory home fedora zfs lib make error make leaving directory home fedora zfs make error | 1 |
19,482 | 3,212,015,883 | IssuesEvent | 2015-10-06 13:51:43 | ariya/phantomjs | https://api.github.com/repos/ariya/phantomjs | closed | Phantom crashing on CentOS even when linked against local QT | old.Priority-Medium old.Status-New old.Type-Defect | _**[sa...@pivotallabs.com](http://code.google.com/u/105663441621740193392/) commented:**_
> <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b>
1.6.1
>
> <b>What steps will reproduce the problem?</b>
N/A - occurs when running the test suite for our closed source application.
>
> <b>What is the expected output? What do you see instead?</b>
N/A
>
> <b>Which operating system are you using?</b>
CentOS release 5.6 (Final)
>
> <b>Did you use binary PhantomJS or did you compile it from source?</b>
Compiled from source - downloaded zip file, unzip, cd in, run build.sh. Compile completed with no errors. ldd shows that the binary is linking against the right version of qt (the one within the phantomjs build directory).
>
> <b>Please provide any additional information below.</b>
> I am running poltergeist which uses phantom.js as a capybara driver. When I run the specs, random specs will fail with the "PhantomJS has crashed. Please file a bug report at https://code.google.com/p/phantomjs/issues/entry and attach the crash dump file" error. When I run it multiple times, different specs fail each time, but always the same way.
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #699](http://code.google.com/p/phantomjs/issues/detail?id=699).
:star2: **3** people had starred this issue at the time of migration. | 1.0 | Phantom crashing on CentOS even when linked against local QT - _**[sa...@pivotallabs.com](http://code.google.com/u/105663441621740193392/) commented:**_
> <b>Which version of PhantomJS are you using? Tip: run 'phantomjs --version'.</b>
1.6.1
>
> <b>What steps will reproduce the problem?</b>
N/A - occurs when running the test suite for our closed source application.
>
> <b>What is the expected output? What do you see instead?</b>
N/A
>
> <b>Which operating system are you using?</b>
CentOS release 5.6 (Final)
>
> <b>Did you use binary PhantomJS or did you compile it from source?</b>
Compiled from source - downloaded zip file, unzip, cd in, run build.sh. Compile completed with no errors. ldd shows that the binary is linking against the right version of qt (the one within the phantomjs build directory).
>
> <b>Please provide any additional information below.</b>
> I am running poltergeist which uses phantom.js as a capybara driver. When I run the specs, random specs will fail with the "PhantomJS has crashed. Please file a bug report at https://code.google.com/p/phantomjs/issues/entry and attach the crash dump file" error. When I run it multiple times, different specs fail each time, but always the same way.
**Disclaimer:**
This issue was migrated on 2013-03-15 from the project's former issue tracker on Google Code, [Issue #699](http://code.google.com/p/phantomjs/issues/detail?id=699).
:star2: **3** people had starred this issue at the time of migration. | non_build | phantom crashing on centos even when linked against local qt commented which version of phantomjs are you using tip run phantomjs version what steps will reproduce the problem n a occurs when running the test suite for our closed source application what is the expected output what do you see instead n a which operating system are you using centos release final did you use binary phantomjs or did you compile it from source compiled from source downloaded zip file unzip cd in run build sh compile completed with no errors ldd shows that the binary is linking against the right version of qt the one within the phantomjs build directory please provide any additional information below i am running poltergeist which uses phantom js as a capybara driver when i run the specs random specs will fail with the quot phantomjs has crashed please file a bug report at and attach the crash dump file quot error when i run it multiple times different specs fail each time but always the same way disclaimer this issue was migrated on from the project s former issue tracker on google code nbsp people had starred this issue at the time of migration | 0 |
68,724 | 17,388,697,060 | IssuesEvent | 2021-08-02 02:29:03 | openmsupply/application-manager-web-app | https://api.github.com/repos/openmsupply/application-manager-web-app | closed | Epic #29 - Template builder | EPIC Epic #29: Template builder | ## Description
The Admin user should be able to see all existing Template applications that Applicants can apply to, and be able to change existing ones - there is a versioning of Applications so all Applications finalised before the change would still be pointing to the previous versions. To be decided - What to do with existing applications?
The Admin user can also create a new Template, there are many steps in the process of creating an application (better defined when the Template is in place). The Questions plugins would also require to have a way to display in the Template builder mode. So all Question plugins might need updates at this point.
There are other things included in the Template builder: selecting the category for this application, who can apply to and review which sections and in what stage. The stages workflow that the application goes through and the actions that happen after each trigger are also configured in here.
## Epics (break-down)
- [Epic #67](523) Import and Export Template Info
- [Epic #68](525) Template expiry configuration
- [Epic #69](527) Versioning templates
- [Epic #70](528) Localisation
- [Epic #71](529) Template structure
- [Epic #72](530) Template configuration of Actions
- [Epic #73](531) Template configurations of Form Elements
- [Epic #74](532) Configuring permission policies and names
- [Epic #75](533) Configurations of template permissions
| 1.0 | Epic #29 - Template builder - ## Description
The Admin user should be able to see all existing Template applications that Applicants can apply to, and be able to change existing ones - there is a versioning of Applications so all Applications finalised before the change would still be pointing to the previous versions. To be decided - What to do with existing applications?
The Admin user can also create a new Template, there are many steps in the process of creating an application (better defined when the Template is in place). The Questions plugins would also require to have a way to display in the Template builder mode. So all Question plugins might need updates at this point.
There are other things included in the Template builder: selecting the category for this application, who can apply to and review which sections and in what stage. The stages workflow that the application goes through and the actions that happen after each trigger are also configured in here.
## Epics (break-down)
- [Epic #67](523) Import and Export Template Info
- [Epic #68](525) Template expiry configuration
- [Epic #69](527) Versioning templates
- [Epic #70](528) Localisation
- [Epic #71](529) Template structure
- [Epic #72](530) Template configuration of Actions
- [Epic #73](531) Template configurations of Form Elements
- [Epic #74](532) Configuring permission policies and names
- [Epic #75](533) Configurations of template permissions
| build | epic template builder description the admin user should be able to see all existing template applications that applicants can apply to and be able to change existing ones there is a versioning of applications so all applications finalised before the change would still be pointing to the previous versions to be decided what to do with existing applications the admin user can also create a new template there are many steps in the process of creating an application better defined when the template is in place the questions plugins would also require to have a way to display in the template builder mode so all question plugins might need updates at this point there are other things included in the template builder selecting the category for this application who can apply to and review which sections and in what stage the stages workflow that the application goes through and the actions that happen after each trigger are also configured in here epics break down import and export template info template expiry configuration versioning templates localisation template structure template configuration of actions template configurations of form elements configuring permission policies and names configurations of template permissions | 1 |
134,600 | 18,475,012,300 | IssuesEvent | 2021-10-18 05:54:41 | samq-ghdemo/JS-DEMO | https://api.github.com/repos/samq-ghdemo/JS-DEMO | closed | CVE-2017-16137 (Medium) detected in debug-2.2.0.tgz - autoclosed | security vulnerability | ## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: JS-DEMO/package.json</p>
<p>Path to vulnerable library: JS-DEMO/node_modules/nyc/node_modules/debug/package.json,JS-DEMO/node_modules/npm/node_modules/node-gyp/node_modules/path-array/node_modules/array-index/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- node-gyp-3.4.0.tgz
- path-array-1.0.1.tgz
- array-index-1.0.0.tgz
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-DEMO/commit/65c8bc8e7543b27d85a7808be0b83c7d68219bb7">65c8bc8e7543b27d85a7808be0b83c7d68219bb7</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;node-gyp:3.4.0;path-array:1.0.1;array-index:1.0.0;debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-16137 (Medium) detected in debug-2.2.0.tgz - autoclosed - ## CVE-2017-16137 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>debug-2.2.0.tgz</b></p></summary>
<p>small debugging utility</p>
<p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p>
<p>Path to dependency file: JS-DEMO/package.json</p>
<p>Path to vulnerable library: JS-DEMO/node_modules/nyc/node_modules/debug/package.json,JS-DEMO/node_modules/npm/node_modules/node-gyp/node_modules/path-array/node_modules/array-index/node_modules/debug/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- node-gyp-3.4.0.tgz
- path-array-1.0.1.tgz
- array-index-1.0.0.tgz
- :x: **debug-2.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-ghdemo/JS-DEMO/commit/65c8bc8e7543b27d85a7808be0b83c7d68219bb7">65c8bc8e7543b27d85a7808be0b83c7d68219bb7</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137>CVE-2017-16137</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137">https://nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-16137</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 2.6.9</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"debug","packageVersion":"2.2.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;node-gyp:3.4.0;path-array:1.0.1;array-index:1.0.0;debug:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.6.9"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2017-16137","vulnerabilityDetails":"The debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter. It takes around 50k characters to block for 2 seconds making this a low severity issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16137","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_build | cve medium detected in debug tgz autoclosed cve medium severity vulnerability vulnerable library debug tgz small debugging utility library home page a href path to dependency file js demo package json path to vulnerable library js demo node modules nyc node modules debug package json js demo node modules npm node modules node gyp node modules path array node modules array index node modules debug package json dependency hierarchy grunt npm install tgz root library npm tgz node gyp tgz path array tgz array index tgz x debug tgz vulnerable library found in head commit a href found in base branch main vulnerability details the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt npm install npm node gyp path array array index debug isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the debug module is vulnerable to regular expression denial of service when untrusted user input is passed into the o formatter it takes around characters to block for seconds making this a low severity issue vulnerabilityurl | 0 |
503,817 | 14,598,380,661 | IssuesEvent | 2020-12-21 00:34:19 | rubyforgood/casa | https://api.github.com/repos/rubyforgood/casa | closed | Move all date formats to il8n | :sparkles: :computer: Contributor Friendly / Devel Priority: Medium | Do this https://github.com/rubyforgood/casa/pull/1402#discussion_r531743219
Syntax we want to use: `format: :day_and_date`
**Why?**
- You can name the keys whatever you like using i18n
- One of the benefits of using i18n is that if you internationalize the application, then it's easy to have dates display in formats used by other countries/languages <- even if CASA doesn't need this, it's good for CASA contributors to get used to good ways of doing things. :)
| 1.0 | Move all date formats to il8n - Do this https://github.com/rubyforgood/casa/pull/1402#discussion_r531743219
Syntax we want to use: `format: :day_and_date`
**Why?**
- You can name the keys whatever you like using i18n
- One of the benefits of using i18n is that if you internationalize the application, then it's easy to have dates display in formats used by other countries/languages <- even if CASA doesn't need this, it's good for CASA contributors to get used to good ways of doing things. :)
| non_build | move all date formats to do this syntax we want to use format day and date why you can name the keys whatever you like using one of the benefits of using is that if you internationalize the application then it s easy to have dates display in formats used by other countries languages even if casa doesn t need this it s good for casa contributors to get used to good ways of doing things | 0 |
71,579 | 18,786,573,452 | IssuesEvent | 2021-11-08 12:48:23 | Scirra/Construct-3-bugs | https://api.github.com/repos/Scirra/Construct-3-bugs | closed | android export does not work | Build Service | ## Problem description
from last 24 hours I have been trying to export android apk files but it is literally hit & trial. Sometimes it works,sometimes not. Most of the time it displays "Build:Preparing Project" & then "connection lost" & then this cycle keeps on repeating. Sometimes it ends with a error message "Build timeout"
Latest stable version. Chrome browser
| 1.0 | android export does not work - ## Problem description
from last 24 hours I have been trying to export android apk files but it is literally hit & trial. Sometimes it works,sometimes not. Most of the time it displays "Build:Preparing Project" & then "connection lost" & then this cycle keeps on repeating. Sometimes it ends with a error message "Build timeout"
Latest stable version. Chrome browser
| build | android export does not work problem description from last hours i have been trying to export android apk files but it is literally hit trial sometimes it works sometimes not most of the time it displays build preparing project then connection lost then this cycle keeps on repeating sometimes it ends with a error message build timeout latest stable version chrome browser | 1 |
439,628 | 12,684,693,570 | IssuesEvent | 2020-06-19 23:44:08 | MyMICDS/MyMICDS-v2 | https://api.github.com/repos/MyMICDS/MyMICDS-v2 | opened | Switch from Travis to GitHub Actions | effort: medium priority: medium work length: short | If we get around to finishing https://github.com/MyMICDS/MyMICDS-v2-Angular/issues/140, we should probably do it here too. | 1.0 | Switch from Travis to GitHub Actions - If we get around to finishing https://github.com/MyMICDS/MyMICDS-v2-Angular/issues/140, we should probably do it here too. | non_build | switch from travis to github actions if we get around to finishing we should probably do it here too | 0 |
46,120 | 11,786,898,118 | IssuesEvent | 2020-03-17 13:09:01 | SpecFlowOSS/SpecFlow | https://api.github.com/repos/SpecFlowOSS/SpecFlow | opened | SpecFlow doesn't work with .NET Core 3.1.200 SDK | .NET Core Support bug generator msbuild | <!-- PLEASE GIVE YOUR ISSUE A SENSIBLE NAME. This makes it easier to identify issues at a glance -->
<!-- PLEASE CHECK THE OPTIONS THAT APPLY TO YOU BY ADDING AN 'x' TO THE CORRESPONDING CHECKBOX ('[ ]') -->
### SpecFlow Version:
- [x] 3.1
- [ ] 3.0
- [ ] 2.4
- [ ] 2.3
- [ ] 2.2
- [ ] 2.1
- [ ] 2.0
- [ ] 1.9
### Used Test Runner
- [x] SpecFlow+Runner
- [x] MSTest
- [x] NUnit
- [x] Xunit
<!-- PLEASE INCLUDE THE VERSION NUMBER OF YOUR TEST RUNNER -->
Version number:
### Project Format of the SpecFlow project
- [ ] Classic project format using `packages.config`
- [ ] Classic project format using `<PackageReference>` tags
- [x] Sdk-style project format
### .feature.cs files are generated using
- [x] `SpecFlow.Tools.MsBuild.Generation` NuGet package
- [ ] `SpecFlowSingleFileGenerator` custom tool
### Visual Studio Version
- [x] VS 2019
- [ ] VS 2017
- [ ] VS 2015
### `Enable SpecFlowSingleFileGenerator Custom Tool` option in Visual Studio extension settings
- [ ] Enabled
- [x] Disabled
### Are the latest Visual Studio updates installed?
- [x] Yes
- [ ] No, I use Visual Studio version `<Major>.<Minor>.<Patch>` <!-- e.g. 16.1.0 -->
### .NET Framework:
- [ ] >= .NET 4.5
- [ ] before .NET 4.5
- [ ] .NET Core 2.0
- [ ] .NET Core 2.1
- [ ] .NET Core 2.2
- [x] .NET Core 3.0
### Test Execution Method:
- [ ] Visual Studio Test Explorer
- [ ] TFS/VSTS/Azure DevOps – Task – PLEASE SPECIFY THE NAME OF THE TASK
- [x] Command line – `dotnet test`
### Issue Description
<!-- PLEASE PROVIDE AS MUCH INFORMATION AS POSSIBLE ON THE ISSUE -->
When using the .NET Core 3.1.200 SDK, the MSBuild Generation fails with following error:
```
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : [SpecFlow] System.IO.FileNotFoundException: Could not load file or assembly 'TechTalk.SpecFlow, Version=3.1.0.0, Culture=neutral, PublicKeyToken=0778194805d6db41'. The system cannot find the file specified. [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : File name: 'TechTalk.SpecFlow, Version=3.1.0.0, Culture=neutral, PublicKeyToken=0778194805d6db41' [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
t32 typeToken, RuntimeTypeHandle[] typeInstantiationContext, RuntimeTypeHandle[] methodInstantiationContext) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.FilterCustomAttributeRecord(MetadataToken caCtorToken, MetadataImport& scope, RuntimeModule decoratedModule, MetadataToken decoratedToken, RuntimeType attributeFilterType, Boolean mustBeInheritable, ListBuilder`1& derivedAttributes, RuntimeType& attributeType, IRuntimeMethodInfo& ctor, Boolean& ctorHasParameters, Boolean& isVarArg) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.AddCustomAttributes(ListBuilder`1& attributes, RuntimeModule decoratedModule, Int32 decoratedMetadataToken, RuntimeType attributeFilterType, Boolean mustBeInheritable, ListBuilder`1 derivedAttributes) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.GetCustomAttributes(RuntimeModule decoratedModule, Int32 decoratedMetadataToken, Int32 pcaCount, RuntimeType attributeFilterType) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.GetCustomAttributes(RuntimeAssembly assembly, RuntimeType caType) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.RuntimeAssembly.GetCustomAttributes(Type attributeType, Boolean inherit) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Attribute.GetCustomAttributes(Assembly element, Type attributeTC:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Attribute.GetCustomAttribute(Assembly element, Type attributeType, Boolean inherit) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Attribute.GetCustomAttribute(Assembly element, Type attributeType) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.Plugins.GeneratorPluginLoader.LoadPlugin(PluginDescriptor pluginDescriptor) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.GeneratorContainerBuilder.LoadPlugin(PluginDescriptor pluginDescriptor, IGeneratorPluginLoader pluginLoader, GeneratorPluginEvents generatorPluginEvents, UnitTestProviderConfiguration unitTestProviderConfigration) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.GeneratorContainerBuilder.LoadPlugins(ObjectContainer container, GeneratorPluginEvents generatorPluginEvents, UnitTestProviderConfiguration unitTestProviderConfigration, IEnumerable`1 generatorPlugins) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.GeneratorContainerBuilder.CreateContainer(SpecFlowConfigurationHolder configurationHolder, ProjectSettings projectSettings, IEnumerable`1 generatorPlugins) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at SpecFlow.Tools.MsBuild.Generation.GenerateFeatureFileCodeBehindTask.Execute() [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
```
### Steps to Reproduce
<!-- PLEASE DESCRIBE THE STEPS REQUIRED TO REPRODUCE THIS ISSUE, IF POSSIBLE -->
### Workaround
Pin the sdk- version to 3.1.102 via global.json
``` json
{
"sdk": {
"version": "3.1.102"
}
}
```
| 1.0 | SpecFlow doesn't work with .NET Core 3.1.200 SDK - <!-- PLEASE GIVE YOUR ISSUE A SENSIBLE NAME. This makes it easier to identify issues at a glance -->
<!-- PLEASE CHECK THE OPTIONS THAT APPLY TO YOU BY ADDING AN 'x' TO THE CORRESPONDING CHECKBOX ('[ ]') -->
### SpecFlow Version:
- [x] 3.1
- [ ] 3.0
- [ ] 2.4
- [ ] 2.3
- [ ] 2.2
- [ ] 2.1
- [ ] 2.0
- [ ] 1.9
### Used Test Runner
- [x] SpecFlow+Runner
- [x] MSTest
- [x] NUnit
- [x] Xunit
<!-- PLEASE INCLUDE THE VERSION NUMBER OF YOUR TEST RUNNER -->
Version number:
### Project Format of the SpecFlow project
- [ ] Classic project format using `packages.config`
- [ ] Classic project format using `<PackageReference>` tags
- [x] Sdk-style project format
### .feature.cs files are generated using
- [x] `SpecFlow.Tools.MsBuild.Generation` NuGet package
- [ ] `SpecFlowSingleFileGenerator` custom tool
### Visual Studio Version
- [x] VS 2019
- [ ] VS 2017
- [ ] VS 2015
### `Enable SpecFlowSingleFileGenerator Custom Tool` option in Visual Studio extension settings
- [ ] Enabled
- [x] Disabled
### Are the latest Visual Studio updates installed?
- [x] Yes
- [ ] No, I use Visual Studio version `<Major>.<Minor>.<Patch>` <!-- e.g. 16.1.0 -->
### .NET Framework:
- [ ] >= .NET 4.5
- [ ] before .NET 4.5
- [ ] .NET Core 2.0
- [ ] .NET Core 2.1
- [ ] .NET Core 2.2
- [x] .NET Core 3.0
### Test Execution Method:
- [ ] Visual Studio Test Explorer
- [ ] TFS/VSTS/Azure DevOps – Task – PLEASE SPECIFY THE NAME OF THE TASK
- [x] Command line – `dotnet test`
### Issue Description
<!-- PLEASE PROVIDE AS MUCH INFORMATION AS POSSIBLE ON THE ISSUE -->
When using the .NET Core 3.1.200 SDK, the MSBuild Generation fails with following error:
```
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : [SpecFlow] System.IO.FileNotFoundException: Could not load file or assembly 'TechTalk.SpecFlow, Version=3.1.0.0, Culture=neutral, PublicKeyToken=0778194805d6db41'. The system cannot find the file specified. [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : File name: 'TechTalk.SpecFlow, Version=3.1.0.0, Culture=neutral, PublicKeyToken=0778194805d6db41' [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
t32 typeToken, RuntimeTypeHandle[] typeInstantiationContext, RuntimeTypeHandle[] methodInstantiationContext) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.FilterCustomAttributeRecord(MetadataToken caCtorToken, MetadataImport& scope, RuntimeModule decoratedModule, MetadataToken decoratedToken, RuntimeType attributeFilterType, Boolean mustBeInheritable, ListBuilder`1& derivedAttributes, RuntimeType& attributeType, IRuntimeMethodInfo& ctor, Boolean& ctorHasParameters, Boolean& isVarArg) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.AddCustomAttributes(ListBuilder`1& attributes, RuntimeModule decoratedModule, Int32 decoratedMetadataToken, RuntimeType attributeFilterType, Boolean mustBeInheritable, ListBuilder`1 derivedAttributes) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.GetCustomAttributes(RuntimeModule decoratedModule, Int32 decoratedMetadataToken, Int32 pcaCount, RuntimeType attributeFilterType) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.CustomAttribute.GetCustomAttributes(RuntimeAssembly assembly, RuntimeType caType) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Reflection.RuntimeAssembly.GetCustomAttributes(Type attributeType, Boolean inherit) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Attribute.GetCustomAttributes(Assembly element, Type attributeTC:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Attribute.GetCustomAttribute(Assembly element, Type attributeType, Boolean inherit) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at System.Attribute.GetCustomAttribute(Assembly element, Type attributeType) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.Plugins.GeneratorPluginLoader.LoadPlugin(PluginDescriptor pluginDescriptor) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.GeneratorContainerBuilder.LoadPlugin(PluginDescriptor pluginDescriptor, IGeneratorPluginLoader pluginLoader, GeneratorPluginEvents generatorPluginEvents, UnitTestProviderConfiguration unitTestProviderConfigration) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.GeneratorContainerBuilder.LoadPlugins(ObjectContainer container, GeneratorPluginEvents generatorPluginEvents, UnitTestProviderConfiguration unitTestProviderConfigration, IEnumerable`1 generatorPlugins) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at TechTalk.SpecFlow.Generator.GeneratorContainerBuilder.CreateContainer(SpecFlowConfigurationHolder configurationHolder, ProjectSettings projectSettings, IEnumerable`1 generatorPlugins) [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : at SpecFlow.Tools.MsBuild.Generation.GenerateFeatureFileCodeBehindTask.Execute() [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
C:\Users\awi\.nuget\packages\specflow.tools.msbuild.generation\3.1.62\build\SpecFlow.Tools.MsBuild.Generation.targets(93,5): error : [C:\work\SpecFlow-Examples\NETCore Examples\BowlingKata-SpecFlowJson-xUnit\Bowling.SpecFlowXUnit\Bowling.SpecFlowXUnit.csproj]
```
### Steps to Reproduce
<!-- PLEASE DESCRIBE THE STEPS REQUIRED TO REPRODUCE THIS ISSUE, IF POSSIBLE -->
### Workaround
Pin the sdk- version to 3.1.102 via global.json
``` json
{
"sdk": {
"version": "3.1.102"
}
}
```
| build | specflow doesn t work with net core sdk specflow version used test runner specflow runner mstest nunit xunit version number project format of the specflow project classic project format using packages config classic project format using tags sdk style project format feature cs files are generated using specflow tools msbuild generation nuget package specflowsinglefilegenerator custom tool visual studio version vs vs vs enable specflowsinglefilegenerator custom tool option in visual studio extension settings enabled disabled are the latest visual studio updates installed yes no i use visual studio version net framework net before net net core net core net core net core test execution method visual studio test explorer tfs vsts azure devops – task – please specify the name of the task command line – dotnet test issue description when using the net core sdk the msbuild generation fails with following error c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error system io filenotfoundexception could not load file or assembly techtalk specflow version culture neutral publickeytoken the system cannot find the file specified c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error file name techtalk specflow version culture neutral publickeytoken typetoken runtimetypehandle typeinstantiationcontext runtimetypehandle methodinstantiationcontext c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system reflection customattribute filtercustomattributerecord metadatatoken cactortoken metadataimport scope runtimemodule decoratedmodule metadatatoken decoratedtoken runtimetype attributefiltertype boolean mustbeinheritable listbuilder derivedattributes runtimetype attributetype iruntimemethodinfo ctor boolean ctorhasparameters boolean isvararg c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system reflection customattribute addcustomattributes listbuilder attributes runtimemodule decoratedmodule decoratedmetadatatoken runtimetype attributefiltertype boolean mustbeinheritable listbuilder derivedattributes c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system reflection customattribute getcustomattributes runtimemodule decoratedmodule decoratedmetadatatoken pcacount runtimetype attributefiltertype c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system reflection customattribute getcustomattributes runtimeassembly assembly runtimetype catype c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system reflection runtimeassembly getcustomattributes type attributetype boolean inherit c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system attribute getcustomattributes assembly element type attributetc users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system attribute getcustomattribute assembly element type attributetype boolean inherit c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at system attribute getcustomattribute assembly element type attributetype c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at techtalk specflow generator plugins generatorpluginloader loadplugin plugindescriptor plugindescriptor c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at techtalk specflow generator generatorcontainerbuilder loadplugin plugindescriptor plugindescriptor igeneratorpluginloader pluginloader generatorpluginevents generatorpluginevents unittestproviderconfiguration unittestproviderconfigration c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at techtalk specflow generator generatorcontainerbuilder loadplugins objectcontainer container generatorpluginevents generatorpluginevents unittestproviderconfiguration unittestproviderconfigration ienumerable generatorplugins c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at techtalk specflow generator generatorcontainerbuilder createcontainer specflowconfigurationholder configurationholder projectsettings projectsettings ienumerable generatorplugins c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error at specflow tools msbuild generation generatefeaturefilecodebehindtask execute c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error c users awi nuget packages specflow tools msbuild generation build specflow tools msbuild generation targets error steps to reproduce workaround pin the sdk version to via global json json sdk version | 1 |
99,670 | 16,449,165,068 | IssuesEvent | 2021-05-21 01:14:41 | Killy85/LepsisBot | https://api.github.com/repos/Killy85/LepsisBot | opened | CVE-2019-10086 (High) detected in commons-beanutils-1.8.3.jar | security vulnerability | ## CVE-2019-10086 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.3.jar</b></p></summary>
<p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Path to dependency file: /LepsisBot/UT2004AFIA/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar</p>
<p>
Dependency Hierarchy:
- pogamut-ut2004-3.7.0.jar (Root Library)
- pogamut-base-3.7.0.jar
- amis-utils-3.7.0.jar
- :x: **commons-beanutils-1.8.3.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean.
<p>Publish Date: 2019-08-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086>CVE-2019-10086</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f">https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f</a></p>
<p>Release Date: 2019-08-20</p>
<p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-10086 (High) detected in commons-beanutils-1.8.3.jar - ## CVE-2019-10086 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-beanutils-1.8.3.jar</b></p></summary>
<p>BeanUtils provides an easy-to-use but flexible wrapper around reflection and introspection.</p>
<p>Path to dependency file: /LepsisBot/UT2004AFIA/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/commons-beanutils/commons-beanutils/1.8.3/commons-beanutils-1.8.3.jar</p>
<p>
Dependency Hierarchy:
- pogamut-ut2004-3.7.0.jar (Root Library)
- pogamut-base-3.7.0.jar
- amis-utils-3.7.0.jar
- :x: **commons-beanutils-1.8.3.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Apache Commons Beanutils 1.9.2, a special BeanIntrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all Java objects. We, however were not using this by default characteristic of the PropertyUtilsBean.
<p>Publish Date: 2019-08-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10086>CVE-2019-10086</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f">https://github.com/victims/victims-cve-db/commit/16a669c84d95bbbd4294f30e609049a36700847f</a></p>
<p>Release Date: 2019-08-20</p>
<p>Fix Resolution: commons-beanutils:commons-beanutils:1.9.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_build | cve high detected in commons beanutils jar cve high severity vulnerability vulnerable library commons beanutils jar beanutils provides an easy to use but flexible wrapper around reflection and introspection path to dependency file lepsisbot pom xml path to vulnerable library root repository commons beanutils commons beanutils commons beanutils jar dependency hierarchy pogamut jar root library pogamut base jar amis utils jar x commons beanutils jar vulnerable library vulnerability details in apache commons beanutils a special beanintrospector class was added which allows suppressing the ability for an attacker to access the classloader via the class property available on all java objects we however were not using this by default characteristic of the propertyutilsbean publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons beanutils commons beanutils step up your open source security game with whitesource | 0 |
85,551 | 24,621,718,650 | IssuesEvent | 2022-10-16 01:47:37 | spring-projects-experimental/spring-pulsar | https://api.github.com/repos/spring-projects-experimental/spring-pulsar | opened | Fix CI deprecation warnings | area/build | Each of our CI runs has notifications/warnings about several deprecated features we are using in our workflows.
List of deprecations: https://github.com/spring-projects-experimental/spring-pulsar/actions/runs/3238804365
| 1.0 | Fix CI deprecation warnings - Each of our CI runs has notifications/warnings about several deprecated features we are using in our workflows.
List of deprecations: https://github.com/spring-projects-experimental/spring-pulsar/actions/runs/3238804365
| build | fix ci deprecation warnings each of our ci runs has notifications warnings about several deprecated features we are using in our workflows list of deprecations | 1 |
100,412 | 30,696,061,251 | IssuesEvent | 2023-07-26 18:41:46 | discoverfinancial/a11y-theme-builder | https://api.github.com/repos/discoverfinancial/a11y-theme-builder | closed | [TB] UI issues with small window sizes | enhancement theme builder app | <!-- Issue Content -->
## Problem/Concern
If I size the Theme Builder application to fit in a smaller space (like 50% the size of the screen) I run into issues with some of the pages. For example on the Font Settings page.
<img width="986" alt="Screenshot 2023-05-11 at 1 58 40 PM" src="https://github.com/discoverfinancial/a11y-theme-builder/assets/32913837/4ddfcf2c-55e8-47d6-abe5-7d1e4b311c2e">
Notice that you can't see what the primary font is used for.
This will become much more important if we want to support Theme Builder on mobile devices, like tablets.
## Proposed Solution
| 1.0 | [TB] UI issues with small window sizes - <!-- Issue Content -->
## Problem/Concern
If I size the Theme Builder application to fit in a smaller space (like 50% the size of the screen) I run into issues with some of the pages. For example on the Font Settings page.
<img width="986" alt="Screenshot 2023-05-11 at 1 58 40 PM" src="https://github.com/discoverfinancial/a11y-theme-builder/assets/32913837/4ddfcf2c-55e8-47d6-abe5-7d1e4b311c2e">
Notice that you can't see what the primary font is used for.
This will become much more important if we want to support Theme Builder on mobile devices, like tablets.
## Proposed Solution
| build | ui issues with small window sizes problem concern if i size the theme builder application to fit in a smaller space like the size of the screen i run into issues with some of the pages for example on the font settings page img width alt screenshot at pm src notice that you can t see what the primary font is used for this will become much more important if we want to support theme builder on mobile devices like tablets proposed solution | 1 |
76,848 | 21,579,839,887 | IssuesEvent | 2022-05-02 17:28:49 | Polymer/tools | https://api.github.com/repos/Polymer/tools | closed | Option to extend parser |OR| Ignore fragments / custom indentation ? | Package: build Status: Available Priority: Medium Type: Question wontfix | Is there an option to extend the parser or set regex for ignoring special characters/indentations from being parsed and manipulated ?
### e.g. Currently:
Example HTML with indentation for server-side rendering:
```
{%= `<${arg['layoutElement']} name="layout">` %}
{%= `</${arg['layoutElement']}>` %}
```
In a simple task:
```
gulp.src(source)
.pipe(sourcesHtmlSplitter.split())
.pipe(sourcesHtmlSplitter.rejoin())
.pipe(destination)
```
becomes:
```
{%= `<${arg['layoutElement']} name="layout">` %}
{%= `<!--${arg['layoutElement']}-->` %}
```
I would like an ability to set regex for ignoring fargments of the code:
`ignoreCustomFragments: [ /{%[\s\S]*?%}/, /<%[\s\S]*?%>/, /<\?[\s\S]*?\?>/ ]`
Thanks | 1.0 | Option to extend parser |OR| Ignore fragments / custom indentation ? - Is there an option to extend the parser or set regex for ignoring special characters/indentations from being parsed and manipulated ?
### e.g. Currently:
Example HTML with indentation for server-side rendering:
```
{%= `<${arg['layoutElement']} name="layout">` %}
{%= `</${arg['layoutElement']}>` %}
```
In a simple task:
```
gulp.src(source)
.pipe(sourcesHtmlSplitter.split())
.pipe(sourcesHtmlSplitter.rejoin())
.pipe(destination)
```
becomes:
```
{%= `<${arg['layoutElement']} name="layout">` %}
{%= `<!--${arg['layoutElement']}-->` %}
```
I would like an ability to set regex for ignoring fargments of the code:
`ignoreCustomFragments: [ /{%[\s\S]*?%}/, /<%[\s\S]*?%>/, /<\?[\s\S]*?\?>/ ]`
Thanks | build | option to extend parser or ignore fragments custom indentation is there an option to extend the parser or set regex for ignoring special characters indentations from being parsed and manipulated e g currently example html with indentation for server side rendering in a simple task gulp src source pipe sourceshtmlsplitter split pipe sourceshtmlsplitter rejoin pipe destination becomes lt arg name layout gt i would like an ability to set regex for ignoring fargments of the code ignorecustomfragments thanks | 1 |
90,676 | 26,166,772,348 | IssuesEvent | 2023-01-01 11:40:27 | Sendouc/sendou.ink | https://api.github.com/repos/Sendouc/sendou.ink | closed | Show % of ink consumed for sub weapons | build analyzer | It's simple to get the number for it but needs consideration where in the UI to put it.

| 1.0 | Show % of ink consumed for sub weapons - It's simple to get the number for it but needs consideration where in the UI to put it.

| build | show of ink consumed for sub weapons it s simple to get the number for it but needs consideration where in the ui to put it | 1 |
568,292 | 16,963,776,590 | IssuesEvent | 2021-06-29 08:25:14 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | unifiedportal-mem.epfindia.gov.in - see bug description | browser-firefox-ios os-ios priority-important | <!-- @browser: Firefox iOS 34.2 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU iPhone OS 14_6 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/34.2 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/78661 -->
<!-- @extra_labels: browser-firefox-ios -->
**URL**: https://unifiedportal-mem.epfindia.gov.in/memberinterface/eNomination/geteNominationPage?_HDIV_STATE_=16-27-9429DBD1B15568D07F24C6A1EAAA35AC
**Browser / Version**: Firefox iOS 34.2
**Operating System**: iOS 14.6
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Desktop site not visible
**Steps to Reproduce**:
Menu option isn’t opening and when I request desktop site, unexpected error occurs
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | unifiedportal-mem.epfindia.gov.in - see bug description - <!-- @browser: Firefox iOS 34.2 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU iPhone OS 14_6 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/34.2 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/78661 -->
<!-- @extra_labels: browser-firefox-ios -->
**URL**: https://unifiedportal-mem.epfindia.gov.in/memberinterface/eNomination/geteNominationPage?_HDIV_STATE_=16-27-9429DBD1B15568D07F24C6A1EAAA35AC
**Browser / Version**: Firefox iOS 34.2
**Operating System**: iOS 14.6
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Desktop site not visible
**Steps to Reproduce**:
Menu option isn’t opening and when I request desktop site, unexpected error occurs
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_build | unifiedportal mem epfindia gov in see bug description url browser version firefox ios operating system ios tested another browser yes safari problem type something else description desktop site not visible steps to reproduce menu option isn’t opening and when i request desktop site unexpected error occurs browser configuration none from with ❤️ | 0 |
292,547 | 21,996,244,282 | IssuesEvent | 2022-05-26 06:43:23 | REPfit-Retail-Price-Fit/backend-main | https://api.github.com/repos/REPfit-Retail-Price-Fit/backend-main | opened | pip install mysqlclient error mysql_config not found linux | documentation | I've found an error while installing the mysqlclient pip's package on my Linux machine. Here is what I've found to fix this: https://stackoverflow.com/questions/5178292/pip-install-mysql-python-fails-with-environmenterror-mysql-config-not-found
So in Ubuntu, we need to install a package called `libmysqlclient-dev`. After I installed it, the error was gone. Let's add it to the initial steps guide in the main `README.md`.
Also, there is some code snippet related to this problem:
https://www.codegrepper.com/code-examples/shell/install+mysqlclient+ubuntu+for+python+3.8
| 1.0 | pip install mysqlclient error mysql_config not found linux - I've found an error while installing the mysqlclient pip's package on my Linux machine. Here is what I've found to fix this: https://stackoverflow.com/questions/5178292/pip-install-mysql-python-fails-with-environmenterror-mysql-config-not-found
So in Ubuntu, we need to install a package called `libmysqlclient-dev`. After I installed it, the error was gone. Let's add it to the initial steps guide in the main `README.md`.
Also, there is some code snippet related to this problem:
https://www.codegrepper.com/code-examples/shell/install+mysqlclient+ubuntu+for+python+3.8
| non_build | pip install mysqlclient error mysql config not found linux i ve found an error while installing the mysqlclient pip s package on my linux machine here is what i ve found to fix this so in ubuntu we need to install a package called libmysqlclient dev after i installed it the error was gone let s add it to the initial steps guide in the main readme md also there is some code snippet related to this problem | 0 |
78,189 | 22,154,802,824 | IssuesEvent | 2022-06-03 21:06:46 | apache/beam | https://api.github.com/repos/apache/beam | opened | Get Jenkins being stable and supportive | P3 improvement build-system |
Imported from Jira [BEAM-3953](https://issues.apache.org/jira/browse/BEAM-3953). Original Jira may contain additional context.
Reported by: yifanzou. | 1.0 | Get Jenkins being stable and supportive -
Imported from Jira [BEAM-3953](https://issues.apache.org/jira/browse/BEAM-3953). Original Jira may contain additional context.
Reported by: yifanzou. | build | get jenkins being stable and supportive imported from jira original jira may contain additional context reported by yifanzou | 1 |
23,712 | 7,370,294,127 | IssuesEvent | 2018-03-13 07:51:12 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Confusing traceback on attempt to build without the platform option | enhancement topic:buildsystem | There have been multiple bug tickets added due to confusion caused by the following build time traceback (Win64 native, but will happen on all platforms):
```
C:\godot>scons
scons: Reading SConscript files ...
No valid target platform selected.
The following were detected:
windows
Please run scons again with argument: platform=<string>
NameError: name 'env' is not defined:
File "C:\godot_j1\SConstruct", line 507:
show_progress = env['progress']
```
This happens if the compulsory platform option is not specified, which is normal.
**What's wrong here is printing a traceback, which just confuses developers new to scons and/or Godot.**
I suggest avoiding the traceback by explicitly exiting from the build right after printing the error message. | 1.0 | Confusing traceback on attempt to build without the platform option - There have been multiple bug tickets added due to confusion caused by the following build time traceback (Win64 native, but will happen on all platforms):
```
C:\godot>scons
scons: Reading SConscript files ...
No valid target platform selected.
The following were detected:
windows
Please run scons again with argument: platform=<string>
NameError: name 'env' is not defined:
File "C:\godot_j1\SConstruct", line 507:
show_progress = env['progress']
```
This happens if the compulsory platform option is not specified, which is normal.
**What's wrong here is printing a traceback, which just confuses developers new to scons and/or Godot.**
I suggest avoiding the traceback by explicitly exiting from the build right after printing the error message. | build | confusing traceback on attempt to build without the platform option there have been multiple bug tickets added due to confusion caused by the following build time traceback native but will happen on all platforms c godot scons scons reading sconscript files no valid target platform selected the following were detected windows please run scons again with argument platform nameerror name env is not defined file c godot sconstruct line show progress env this happens if the compulsory platform option is not specified which is normal what s wrong here is printing a traceback which just confuses developers new to scons and or godot i suggest avoiding the traceback by explicitly exiting from the build right after printing the error message | 1 |
4,705 | 3,880,814,031 | IssuesEvent | 2016-04-13 00:02:58 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 21357458: Setting System Font Weight via IB doesn't work | classification:ui/usability reproducible:always status:open | #### Description
Summary:
Setting the font weight of a label (in a table view cell in my case) with system font does nothing.
Steps to Reproduce:
In a storyboard, set the font of a UILabel to "System Light 17.0".
Expected Results:
The font should be San Francisco UI Light.
Actual Results:
the font is:
<UICTFont: 0x7ffa39c647e0> font-family: ".SFUIText-Regular"; font-weight: normal; font-style: normal; font-size: 17.00pt
Regression:
Bug introduced in 9.0 beta 1.
Notes:
I tried various font weights including Light, Thin, Ultra Light, Medium. It does not matter what font weight is set in the storyboard. The label always uses .SFUIText-Regular.
-
Product Version: 9.0 beta 1
Created: 2015-06-12T14:10:29.154020
Originated: 2015-06-12T16:10:00
Open Radar Link: http://www.openradar.me/21357458 | True | 21357458: Setting System Font Weight via IB doesn't work - #### Description
Summary:
Setting the font weight of a label (in a table view cell in my case) with system font does nothing.
Steps to Reproduce:
In a storyboard, set the font of a UILabel to "System Light 17.0".
Expected Results:
The font should be San Francisco UI Light.
Actual Results:
the font is:
<UICTFont: 0x7ffa39c647e0> font-family: ".SFUIText-Regular"; font-weight: normal; font-style: normal; font-size: 17.00pt
Regression:
Bug introduced in 9.0 beta 1.
Notes:
I tried various font weights including Light, Thin, Ultra Light, Medium. It does not matter what font weight is set in the storyboard. The label always uses .SFUIText-Regular.
-
Product Version: 9.0 beta 1
Created: 2015-06-12T14:10:29.154020
Originated: 2015-06-12T16:10:00
Open Radar Link: http://www.openradar.me/21357458 | non_build | setting system font weight via ib doesn t work description summary setting the font weight of a label in a table view cell in my case with system font does nothing steps to reproduce in a storyboard set the font of a uilabel to system light expected results the font should be san francisco ui light actual results the font is font family sfuitext regular font weight normal font style normal font size regression bug introduced in beta notes i tried various font weights including light thin ultra light medium it does not matter what font weight is set in the storyboard the label always uses sfuitext regular product version beta created originated open radar link | 0 |
291,824 | 21,940,759,857 | IssuesEvent | 2022-05-23 17:50:34 | typescript-eslint/typescript-eslint | https://api.github.com/repos/typescript-eslint/typescript-eslint | closed | Website: add a "report an issue" button on the playground | documentation accepting prs package: website | ### Suggested Changes
From @bradzacher in https://github.com/typescript-eslint/typescript-eslint/pull/4897/#discussion_r866262233:
> In the future we will be able to add a "report an issue" button on the playground which takes all the state and pre-fills the issue template for the user.
We should merge this after #4897 goes in, so it can use the issue form.
### Affected URL(s)
https://typescript-eslint.io/play | 1.0 | Website: add a "report an issue" button on the playground - ### Suggested Changes
From @bradzacher in https://github.com/typescript-eslint/typescript-eslint/pull/4897/#discussion_r866262233:
> In the future we will be able to add a "report an issue" button on the playground which takes all the state and pre-fills the issue template for the user.
We should merge this after #4897 goes in, so it can use the issue form.
### Affected URL(s)
https://typescript-eslint.io/play | non_build | website add a report an issue button on the playground suggested changes from bradzacher in in the future we will be able to add a report an issue button on the playground which takes all the state and pre fills the issue template for the user we should merge this after goes in so it can use the issue form affected url s | 0 |
46,982 | 11,942,152,460 | IssuesEvent | 2020-04-02 19:45:21 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | Segfault in Google Colab | Build issues | I'm getting Segfault in Google Colab when I use `python runtests.py` when installed from source (vs `pip install`)
I've seen a lot of red herrings but it appears to be related to complex matrixes and it doesn't appear to trigger at the point the failure happens (e.g. if the problem is in testX it will often pass testX and die in testX+2).
[test_decomp.py::TestEigh](https://github.com/scipy/scipy/blob/master/scipy/linalg/tests/test_decomp.py#L823) seems to have several tests that will trigger it. `np.complex128` in `test_value_subsets` is sure to break the code.
```
$ python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -v; echo "inline $?"
Building, see build.log...
Build OK (0:00:03.508031 elapsed)
============================= test session starts ==============================
platform linux -- Python 3.6.9, pytest-3.6.4, py-1.8.1, pluggy-0.7.1 -- /usr/bin/python3
cachedir: ../../../../../.pytest_cache
rootdir: /content/scipy, inifile: pytest.ini
plugins: typeguard-2.7.1
collected 76 items
scipy/linalg/tests/test_decomp.py::TestEigh::test_wrong_inputs PASSED [ 1%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_nonpositive_b PASSED [ 2%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_value_subsets /bin/bash: line 1: 3598 Segmentation fault (core dumped) python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -v
inline 139
```
colab sets `LD_PRELOAD` to libtcmalloc.
```
$ echo $LD_PRELOAD
$ unset LD_PRELOAD && python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -vv; echo "inline $?"
/usr/lib/x86_64-linux-gnu/libtcmalloc.so.4
Building, see build.log...
Build OK (0:00:03.505405 elapsed)
============================= test session starts ==============================
platform linux -- Python 3.6.9, pytest-3.6.4, py-1.8.1, pluggy-0.7.1 -- /usr/bin/python3
cachedir: ../../../../../.pytest_cache
rootdir: /content/scipy, inifile: pytest.ini
plugins: typeguard-2.7.1
collected 76 items
scipy/linalg/tests/test_decomp.py::TestEigh::test_wrong_inputs PASSED [ 1%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_nonpositive_b PASSED [ 2%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_value_subsets /bin/bash: line 1: 3806 Aborted (core dumped) python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -vv
inline 134
```
I suspect something is broken with a double free because of the different when running with tcmalloc and with the error not happening immediately.
I finally managed to get gdb working (colab's `python` is a bash script so you need to use python3) but it didn't shed any light for me
```
Thread 1 "python3" received signal SIGSEGV, Segmentation fault.
0x00007f77f9e6e0ad in tc_malloc ()
from /usr/lib/x86_64-linux-gnu/libtcmalloc.so.4
(gdb) bt
#0 0x00007f77f9e6e0ad in tc_malloc ()
from /usr/lib/x86_64-linux-gnu/libtcmalloc.so.4
#1 0x00007f77ed479b75 in string_from_pyobj (str=str@entry=0x7ffcb32d00b0,
len=len@entry=0x7ffcb32d005c, inistr=inistr@entry=0x7f77ed4b860e "A",
obj=<optimized out>,
errmess=errmess@entry=0x7f77ed4f1e38 "string_from_pyobj failed in converting 3rd keyword `range' of _flapack.zhegvx to C string")
at build/src.linux-x86_64-3.6/build/src.linux-x86_64-3.6/scipy/linalg/_flapackmodule.c:442
#2 0x00007f77ed4947e8 in f2py_rout__flapack_zhegvx (
capi_self=<optimized out>, capi_args=<optimized out>,
capi_keywds=<optimized out>, f2py_func=0x7f77eda67bb0 <zhegvx_>)
at build/src.linux-x86_64-3.6/build/src.linux-x86_64-3.6/scipy/linalg/_flapackmodule.c:53165
#3 0x00000000005a067e in PyObject_Call ()
#4 0x000000000050d966 in _PyEval_EvalFrameDefault ()
#5 0x0000000000508245 in ?? ()
#6 0x000000000050a080 in ?? ()
#7 0x000000000050aa7d in ?? ()
#8 0x000000000050d390 in _PyEval_EvalFrameDefault ()
#9 0x0000000000509455 in _PyFunction_FastCallDict ()
#10 0x0000000000595311 in ?? ()
#11 0x00000000005a067e in PyObject_Call ()
---Type <return> to continue, or q <return> to quit---
#12 0x000000000050d966 in _PyEval_EvalFrameDefault ()
#13 0x0000000000508245 in ?? ()
#14 0x00000000005893bb in ?? ()
#15 0x00000000005a067e in PyObject_Call ()
#16 0x000000000050d966 in _PyEval_EvalFrameDefault ()
#17 0x0000000000508245 in ?? ()
#18 0x000000000050a080 in ?? ()
#19 0x000000000050aa7d in ?? ()
#20 0x000000000050d390 in _PyEval_EvalFrameDefault ()
#21 0x0000000000508245 in ?? ()
#22 0x000000000050a080 in ?? ()
#23 0x000000000050aa7d in ?? ()
#24 0x000000000050c5b9 in _PyEval_EvalFrameDefault ()
#25 0x0000000000509d48 in ?? ()
#26 0x000000000050aa7d in ?? ()
#27 0x000000000050c5b9 in _PyEval_EvalFrameDefault ()
#28 0x0000000000508245 in ?? ()
#29 0x0000000000509642 in _PyFunction_FastCallDict ()
#30 0x0000000000595311 in ?? ()
#31 0x000000000054b1e1 in ?? ()
#32 0x00000000005aa6ec in _PyObject_FastCallKeywords ()
#33 0x000000000050abb3 in ?? ()
...
#166 0x0000000000508245 in ?? ()
#167 0x000000000050a080 in ?? ()
#168 0x000000000050aa7d in ?? ()
#169 0x000000000050d390 in _PyEval_EvalFrameDefault ()
#170 0x0000000000508245 in ?? ()
#171 0x000000000050b403 in PyEval_EvalCode ()
#172 0x0000000000635222 in ?? ()
---Type <return> to continue, or q <return> to quit---
#173 0x00000000006352d7 in PyRun_FileExFlags ()
#174 0x0000000000638a8f in PyRun_SimpleFileExFlags ()
#175 0x0000000000639631 in Py_Main ()
#176 0x00000000004b0f40 in main ()
``` | 1.0 | Segfault in Google Colab - I'm getting Segfault in Google Colab when I use `python runtests.py` when installed from source (vs `pip install`)
I've seen a lot of red herrings but it appears to be related to complex matrixes and it doesn't appear to trigger at the point the failure happens (e.g. if the problem is in testX it will often pass testX and die in testX+2).
[test_decomp.py::TestEigh](https://github.com/scipy/scipy/blob/master/scipy/linalg/tests/test_decomp.py#L823) seems to have several tests that will trigger it. `np.complex128` in `test_value_subsets` is sure to break the code.
```
$ python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -v; echo "inline $?"
Building, see build.log...
Build OK (0:00:03.508031 elapsed)
============================= test session starts ==============================
platform linux -- Python 3.6.9, pytest-3.6.4, py-1.8.1, pluggy-0.7.1 -- /usr/bin/python3
cachedir: ../../../../../.pytest_cache
rootdir: /content/scipy, inifile: pytest.ini
plugins: typeguard-2.7.1
collected 76 items
scipy/linalg/tests/test_decomp.py::TestEigh::test_wrong_inputs PASSED [ 1%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_nonpositive_b PASSED [ 2%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_value_subsets /bin/bash: line 1: 3598 Segmentation fault (core dumped) python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -v
inline 139
```
colab sets `LD_PRELOAD` to libtcmalloc.
```
$ echo $LD_PRELOAD
$ unset LD_PRELOAD && python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -vv; echo "inline $?"
/usr/lib/x86_64-linux-gnu/libtcmalloc.so.4
Building, see build.log...
Build OK (0:00:03.505405 elapsed)
============================= test session starts ==============================
platform linux -- Python 3.6.9, pytest-3.6.4, py-1.8.1, pluggy-0.7.1 -- /usr/bin/python3
cachedir: ../../../../../.pytest_cache
rootdir: /content/scipy, inifile: pytest.ini
plugins: typeguard-2.7.1
collected 76 items
scipy/linalg/tests/test_decomp.py::TestEigh::test_wrong_inputs PASSED [ 1%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_nonpositive_b PASSED [ 2%]
scipy/linalg/tests/test_decomp.py::TestEigh::test_value_subsets /bin/bash: line 1: 3806 Aborted (core dumped) python runtests.py -t scipy/linalg/tests/test_decomp.py::TestEigh -- -vv
inline 134
```
I suspect something is broken with a double free because of the different when running with tcmalloc and with the error not happening immediately.
I finally managed to get gdb working (colab's `python` is a bash script so you need to use python3) but it didn't shed any light for me
```
Thread 1 "python3" received signal SIGSEGV, Segmentation fault.
0x00007f77f9e6e0ad in tc_malloc ()
from /usr/lib/x86_64-linux-gnu/libtcmalloc.so.4
(gdb) bt
#0 0x00007f77f9e6e0ad in tc_malloc ()
from /usr/lib/x86_64-linux-gnu/libtcmalloc.so.4
#1 0x00007f77ed479b75 in string_from_pyobj (str=str@entry=0x7ffcb32d00b0,
len=len@entry=0x7ffcb32d005c, inistr=inistr@entry=0x7f77ed4b860e "A",
obj=<optimized out>,
errmess=errmess@entry=0x7f77ed4f1e38 "string_from_pyobj failed in converting 3rd keyword `range' of _flapack.zhegvx to C string")
at build/src.linux-x86_64-3.6/build/src.linux-x86_64-3.6/scipy/linalg/_flapackmodule.c:442
#2 0x00007f77ed4947e8 in f2py_rout__flapack_zhegvx (
capi_self=<optimized out>, capi_args=<optimized out>,
capi_keywds=<optimized out>, f2py_func=0x7f77eda67bb0 <zhegvx_>)
at build/src.linux-x86_64-3.6/build/src.linux-x86_64-3.6/scipy/linalg/_flapackmodule.c:53165
#3 0x00000000005a067e in PyObject_Call ()
#4 0x000000000050d966 in _PyEval_EvalFrameDefault ()
#5 0x0000000000508245 in ?? ()
#6 0x000000000050a080 in ?? ()
#7 0x000000000050aa7d in ?? ()
#8 0x000000000050d390 in _PyEval_EvalFrameDefault ()
#9 0x0000000000509455 in _PyFunction_FastCallDict ()
#10 0x0000000000595311 in ?? ()
#11 0x00000000005a067e in PyObject_Call ()
---Type <return> to continue, or q <return> to quit---
#12 0x000000000050d966 in _PyEval_EvalFrameDefault ()
#13 0x0000000000508245 in ?? ()
#14 0x00000000005893bb in ?? ()
#15 0x00000000005a067e in PyObject_Call ()
#16 0x000000000050d966 in _PyEval_EvalFrameDefault ()
#17 0x0000000000508245 in ?? ()
#18 0x000000000050a080 in ?? ()
#19 0x000000000050aa7d in ?? ()
#20 0x000000000050d390 in _PyEval_EvalFrameDefault ()
#21 0x0000000000508245 in ?? ()
#22 0x000000000050a080 in ?? ()
#23 0x000000000050aa7d in ?? ()
#24 0x000000000050c5b9 in _PyEval_EvalFrameDefault ()
#25 0x0000000000509d48 in ?? ()
#26 0x000000000050aa7d in ?? ()
#27 0x000000000050c5b9 in _PyEval_EvalFrameDefault ()
#28 0x0000000000508245 in ?? ()
#29 0x0000000000509642 in _PyFunction_FastCallDict ()
#30 0x0000000000595311 in ?? ()
#31 0x000000000054b1e1 in ?? ()
#32 0x00000000005aa6ec in _PyObject_FastCallKeywords ()
#33 0x000000000050abb3 in ?? ()
...
#166 0x0000000000508245 in ?? ()
#167 0x000000000050a080 in ?? ()
#168 0x000000000050aa7d in ?? ()
#169 0x000000000050d390 in _PyEval_EvalFrameDefault ()
#170 0x0000000000508245 in ?? ()
#171 0x000000000050b403 in PyEval_EvalCode ()
#172 0x0000000000635222 in ?? ()
---Type <return> to continue, or q <return> to quit---
#173 0x00000000006352d7 in PyRun_FileExFlags ()
#174 0x0000000000638a8f in PyRun_SimpleFileExFlags ()
#175 0x0000000000639631 in Py_Main ()
#176 0x00000000004b0f40 in main ()
``` | build | segfault in google colab i m getting segfault in google colab when i use python runtests py when installed from source vs pip install i ve seen a lot of red herrings but it appears to be related to complex matrixes and it doesn t appear to trigger at the point the failure happens e g if the problem is in testx it will often pass testx and die in testx seems to have several tests that will trigger it np in test value subsets is sure to break the code python runtests py t scipy linalg tests test decomp py testeigh v echo inline building see build log build ok elapsed test session starts platform linux python pytest py pluggy usr bin cachedir pytest cache rootdir content scipy inifile pytest ini plugins typeguard collected items scipy linalg tests test decomp py testeigh test wrong inputs passed scipy linalg tests test decomp py testeigh test nonpositive b passed scipy linalg tests test decomp py testeigh test value subsets bin bash line segmentation fault core dumped python runtests py t scipy linalg tests test decomp py testeigh v inline colab sets ld preload to libtcmalloc echo ld preload unset ld preload python runtests py t scipy linalg tests test decomp py testeigh vv echo inline usr lib linux gnu libtcmalloc so building see build log build ok elapsed test session starts platform linux python pytest py pluggy usr bin cachedir pytest cache rootdir content scipy inifile pytest ini plugins typeguard collected items scipy linalg tests test decomp py testeigh test wrong inputs passed scipy linalg tests test decomp py testeigh test nonpositive b passed scipy linalg tests test decomp py testeigh test value subsets bin bash line aborted core dumped python runtests py t scipy linalg tests test decomp py testeigh vv inline i suspect something is broken with a double free because of the different when running with tcmalloc and with the error not happening immediately i finally managed to get gdb working colab s python is a bash script so you need to use but it didn t shed any light for me thread received signal sigsegv segmentation fault in tc malloc from usr lib linux gnu libtcmalloc so gdb bt in tc malloc from usr lib linux gnu libtcmalloc so in string from pyobj str str entry len len entry inistr inistr entry a obj errmess errmess entry string from pyobj failed in converting keyword range of flapack zhegvx to c string at build src linux build src linux scipy linalg flapackmodule c in rout flapack zhegvx capi self capi args capi keywds func at build src linux build src linux scipy linalg flapackmodule c in pyobject call in pyeval evalframedefault in in in in pyeval evalframedefault in pyfunction fastcalldict in in pyobject call type to continue or q to quit in pyeval evalframedefault in in in pyobject call in pyeval evalframedefault in in in in pyeval evalframedefault in in in in pyeval evalframedefault in in in pyeval evalframedefault in in pyfunction fastcalldict in in in pyobject fastcallkeywords in in in in in pyeval evalframedefault in in pyeval evalcode in type to continue or q to quit in pyrun fileexflags in pyrun simplefileexflags in py main in main | 1 |
189,573 | 15,191,297,646 | IssuesEvent | 2021-02-15 19:37:09 | numpy/numpy | https://api.github.com/repos/numpy/numpy | opened | Hard to navigate docs front page, refguide link is missing | 04 - Documentation | Looking at https://numpy.org/devdocs/, it took me quite a while to find the link to the single most important page: the API reference guide front page. There's a couple of issues:
1. Too many links:
<img width="401" alt="image" src="https://user-images.githubusercontent.com/98330/107986735-f9ce7200-6fcc-11eb-8f34-c8aa6ade3139.png">
2. Missing link to https://numpy.org/devdocs/reference/index.html (only can be reached from the header navbar, not obvious).
We're now using the same theme as Pandas, so why not borrow this beautiful and clear front page (https://pandas.pydata.org/docs/):
<img width="670" alt="image" src="https://user-images.githubusercontent.com/98330/107987006-86793000-6fcd-11eb-984e-f52f3a51d2ff.png">
| 1.0 | Hard to navigate docs front page, refguide link is missing - Looking at https://numpy.org/devdocs/, it took me quite a while to find the link to the single most important page: the API reference guide front page. There's a couple of issues:
1. Too many links:
<img width="401" alt="image" src="https://user-images.githubusercontent.com/98330/107986735-f9ce7200-6fcc-11eb-8f34-c8aa6ade3139.png">
2. Missing link to https://numpy.org/devdocs/reference/index.html (only can be reached from the header navbar, not obvious).
We're now using the same theme as Pandas, so why not borrow this beautiful and clear front page (https://pandas.pydata.org/docs/):
<img width="670" alt="image" src="https://user-images.githubusercontent.com/98330/107987006-86793000-6fcd-11eb-984e-f52f3a51d2ff.png">
| non_build | hard to navigate docs front page refguide link is missing looking at it took me quite a while to find the link to the single most important page the api reference guide front page there s a couple of issues too many links img width alt image src missing link to only can be reached from the header navbar not obvious we re now using the same theme as pandas so why not borrow this beautiful and clear front page img width alt image src | 0 |
54,743 | 13,441,479,802 | IssuesEvent | 2020-09-08 04:16:26 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Minimum requirement for Tensorflow Lite | comp:lite stalled stat:awaiting response type:build/install | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): source
- TensorFlow version: N/A
- Python version: N/A
- Installed using virtualenv? pip? conda?:
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version:
- GPU model and memory:
**Describe the problem**
I tried to make an inference program with C/C++ (TF Lite Microcontroller) in ARMv7(32-bit) environment at the very beginning, but the model that I converted includes complex bidirectional LSTM and DNN, and it was hard to use with the C/C++ environment.
I got errors when my codes load the converted model in C/C++ even though there was no error while conversion.
Before converting the model, I checked that simple exmaple binary such as hello_world (TF Lite example) is running well on c/c++ with armv7 environment, and the converted model is also running well on python with x86 environment.
I changed my plan into using Python, but my embedded board has no tools like pip, wget, python, and other basic binaries to use. (Thus, now I'm trying to build python for my board.) Additionally, there is only 300 MB that I can use in the board.
I want to know the minimum requirement of Tensorflow Lite for Python and Microcontroller (C/C++).
The board has nxp4330(ARM32) chipset and it has only minimum binaries to run, so I have no idea what I have to do first for using TF Lite with this board.
Thank you in advance.
**Provide the exact sequence of commands / steps that you executed before running into the problem**
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| 1.0 | Minimum requirement for Tensorflow Lite - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Ubuntu 18.04
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): source
- TensorFlow version: N/A
- Python version: N/A
- Installed using virtualenv? pip? conda?:
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version:
- GPU model and memory:
**Describe the problem**
I tried to make an inference program with C/C++ (TF Lite Microcontroller) in ARMv7(32-bit) environment at the very beginning, but the model that I converted includes complex bidirectional LSTM and DNN, and it was hard to use with the C/C++ environment.
I got errors when my codes load the converted model in C/C++ even though there was no error while conversion.
Before converting the model, I checked that simple exmaple binary such as hello_world (TF Lite example) is running well on c/c++ with armv7 environment, and the converted model is also running well on python with x86 environment.
I changed my plan into using Python, but my embedded board has no tools like pip, wget, python, and other basic binaries to use. (Thus, now I'm trying to build python for my board.) Additionally, there is only 300 MB that I can use in the board.
I want to know the minimum requirement of Tensorflow Lite for Python and Microcontroller (C/C++).
The board has nxp4330(ARM32) chipset and it has only minimum binaries to run, so I have no idea what I have to do first for using TF Lite with this board.
Thank you in advance.
**Provide the exact sequence of commands / steps that you executed before running into the problem**
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| build | minimum requirement for tensorflow lite please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu ubuntu mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary source tensorflow version n a python version n a installed using virtualenv pip conda bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory describe the problem i tried to make an inference program with c c tf lite microcontroller in bit environment at the very beginning but the model that i converted includes complex bidirectional lstm and dnn and it was hard to use with the c c environment i got errors when my codes load the converted model in c c even though there was no error while conversion before converting the model i checked that simple exmaple binary such as hello world tf lite example is running well on c c with environment and the converted model is also running well on python with environment i changed my plan into using python but my embedded board has no tools like pip wget python and other basic binaries to use thus now i m trying to build python for my board additionally there is only mb that i can use in the board i want to know the minimum requirement of tensorflow lite for python and microcontroller c c the board has chipset and it has only minimum binaries to run so i have no idea what i have to do first for using tf lite with this board thank you in advance provide the exact sequence of commands steps that you executed before running into the problem any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached | 1 |
4,935 | 11,858,386,665 | IssuesEvent | 2020-03-25 11:20:40 | ImisDevelopers/1_011_a_infektionsfall_uebermittellung | https://api.github.com/repos/ImisDevelopers/1_011_a_infektionsfall_uebermittellung | closed | Cloud SQL database not initialized | !!PRIORITY!! architecture backend bug deployment | Add init script to deployment or allow backend to create tables if not existing. | 1.0 | Cloud SQL database not initialized - Add init script to deployment or allow backend to create tables if not existing. | non_build | cloud sql database not initialized add init script to deployment or allow backend to create tables if not existing | 0 |
292,131 | 21,952,754,068 | IssuesEvent | 2022-05-24 09:21:09 | QCDIS/NaaVRE | https://api.github.com/repos/QCDIS/NaaVRE | closed | Add documntation for to get SDIA credetials | documentation enhancement | Users have no idea what SDIA is when asked to fill in the credentials.
Add button/doc on how to get access. | 1.0 | Add documntation for to get SDIA credetials - Users have no idea what SDIA is when asked to fill in the credentials.
Add button/doc on how to get access. | non_build | add documntation for to get sdia credetials users have no idea what sdia is when asked to fill in the credentials add button doc on how to get access | 0 |
122,996 | 4,848,239,171 | IssuesEvent | 2016-11-10 16:58:27 | Signbank/NGT-signbank | https://api.github.com/repos/Signbank/NGT-signbank | closed | Translations of field choices don't show | bug internationalisation top priority | If I switch interface language to Dutch, I keep seeing the English names for the field choices. | 1.0 | Translations of field choices don't show - If I switch interface language to Dutch, I keep seeing the English names for the field choices. | non_build | translations of field choices don t show if i switch interface language to dutch i keep seeing the english names for the field choices | 0 |
161,055 | 20,120,389,827 | IssuesEvent | 2022-02-08 01:14:17 | arohablue/BlockDockServer | https://api.github.com/repos/arohablue/BlockDockServer | closed | CVE-2016-10735 (Medium) detected in bootstrap-3.3.6.js - autoclosed | security vulnerability | ## CVE-2016-10735 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /BlockDockServer/grails-app/assets/javascripts/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.js** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0, 4.0.0-beta.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-10735 (Medium) detected in bootstrap-3.3.6.js - autoclosed - ## CVE-2016-10735 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.6.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.6/js/bootstrap.js</a></p>
<p>Path to vulnerable library: /BlockDockServer/grails-app/assets/javascripts/bootstrap.js</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-3.3.6.js** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap 3.x before 3.4.0 and 4.x-beta before 4.0.0-beta.2, XSS is possible in the data-target attribute, a different vulnerability than CVE-2018-14041.
<p>Publish Date: 2019-01-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10735>CVE-2016-10735</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-10735</a></p>
<p>Release Date: 2019-01-09</p>
<p>Fix Resolution: bootstrap - 3.4.0, 4.0.0-beta.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_build | cve medium detected in bootstrap js autoclosed cve medium severity vulnerability vulnerable library bootstrap js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to vulnerable library blockdockserver grails app assets javascripts bootstrap js dependency hierarchy x bootstrap js vulnerable library vulnerability details in bootstrap x before and x beta before beta xss is possible in the data target attribute a different vulnerability than cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap beta step up your open source security game with whitesource | 0 |
224,808 | 7,473,168,299 | IssuesEvent | 2018-04-03 14:40:39 | rathena/rathena | https://api.github.com/repos/rathena/rathena | closed | [ Suggestion ] getinventorylist should return random option | component:core mode:renewal priority:low type:enhancement | <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**: -
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**: -
<!-- Please specify the client date you used. -->
* **Server Mode**: -
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**:
* Result: <!-- Describe the issue that you experienced in detail. -->This command lack of the ability to return random option which is introduced for quite a long time.
* Expected Result: <!-- Describe what you would expect to happen in detail. -->Expect it return the random option value.
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->-
* Official Information:<!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->-
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
| 1.0 | [ Suggestion ] getinventorylist should return random option - <!-- NOTE: Anything within these brackets will be hidden on the preview of the Issue. -->
* **rAthena Hash**: -
<!-- Please specify the rAthena [GitHub hash](https://help.github.com/articles/autolinked-references-and-urls/#commit-shas) on which you encountered this issue.
How to get your GitHub Hash:
1. cd your/rAthena/directory/
2. git rev-parse --short HEAD
3. Copy the resulting hash.
-->
* **Client Date**: -
<!-- Please specify the client date you used. -->
* **Server Mode**: -
<!-- Which mode does your server use: Pre-Renewal or Renewal? -->
* **Description of Issue**:
* Result: <!-- Describe the issue that you experienced in detail. -->This command lack of the ability to return random option which is introduced for quite a long time.
* Expected Result: <!-- Describe what you would expect to happen in detail. -->Expect it return the random option value.
* How to Reproduce: <!-- If you have not stated in the description of the result already, please give us a short guide how we can reproduce your issue. -->-
* Official Information:<!-- If possible, provide information from official servers (kRO or other sources) which prove that the result is wrong. Please take into account that iRO (especially iRO Wiki) is not always the same as kRO. -->-
<!-- * _NOTE: Make sure you quote ``` `@atcommands` ``` just like this so that you do not tag uninvolved GitHub users!_ -->
* **Modifications that may affect results**:
<!-- * Please provide any information that could influence the expected result. -->
<!-- * This can be either configurations you changed, database values you changed, or even external source modifications. -->
| non_build | getinventorylist should return random option rathena hash please specify the rathena on which you encountered this issue how to get your github hash cd your rathena directory git rev parse short head copy the resulting hash client date server mode description of issue result this command lack of the ability to return random option which is introduced for quite a long time expected result expect it return the random option value how to reproduce official information modifications that may affect results | 0 |
201,983 | 15,818,194,166 | IssuesEvent | 2021-04-05 15:40:51 | jeffalo/my-ocular | https://api.github.com/repos/jeffalo/my-ocular | closed | Link in README doesnt redirect to myocular | documentation | Try putting // before link
sorry if I'm annoying you | 1.0 | Link in README doesnt redirect to myocular - Try putting // before link
sorry if I'm annoying you | non_build | link in readme doesnt redirect to myocular try putting before link sorry if i m annoying you | 0 |
42,125 | 10,886,386,754 | IssuesEvent | 2019-11-18 12:27:42 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Mono: Data template directory not found | bug topic:buildsystem topic:mono | **Godot version:**
98caeb635
**OS/device including version:**
Windows 10 1903
**Issue description:**
Gives me this error when trying to export for windows:

All files are properly placed and the data folders are properly named (to my understanding):
 | 1.0 | Mono: Data template directory not found - **Godot version:**
98caeb635
**OS/device including version:**
Windows 10 1903
**Issue description:**
Gives me this error when trying to export for windows:

All files are properly placed and the data folders are properly named (to my understanding):
 | build | mono data template directory not found godot version os device including version windows issue description gives me this error when trying to export for windows all files are properly placed and the data folders are properly named to my understanding | 1 |
73,141 | 19,580,591,674 | IssuesEvent | 2022-01-04 20:44:26 | nunit/nunit-console | https://api.github.com/repos/nunit/nunit-console | closed | Do we need a new label for "dev" builds? | Build Needs Discussion | In our myget feed or in any combined set of feeds which includes it, "dev" builds sometimes appear alongside "alpha" or "beta" builds. Because "dev" > either of the other two, it appears to be a better selection to nuget and possibly to some users. In fact, depending on when the dev build was created, it may actually be either > or < than the others.
Ways we can fix this...
1. Delete dev builds from myget whenever an alpha or beta is produced. This is what I have been doing. If this is done consistently, then any dev builds present are guaranteed to be later builds than an alpha or beta.
2. Find some other label alphabetically less than "alpha", like "adv" or "adev" or "aaa". My preference would be for something meaningful. Any ideas?
3. Find some other label fors alpha and beta builds alphabetically greater than "dev" and less than "rc".
Note that even if we do 2 or 3, we still need to do 1.
Of course, another option is to ignore the problem, since we have lived with it already for a long time.
Ideas? | 1.0 | Do we need a new label for "dev" builds? - In our myget feed or in any combined set of feeds which includes it, "dev" builds sometimes appear alongside "alpha" or "beta" builds. Because "dev" > either of the other two, it appears to be a better selection to nuget and possibly to some users. In fact, depending on when the dev build was created, it may actually be either > or < than the others.
Ways we can fix this...
1. Delete dev builds from myget whenever an alpha or beta is produced. This is what I have been doing. If this is done consistently, then any dev builds present are guaranteed to be later builds than an alpha or beta.
2. Find some other label alphabetically less than "alpha", like "adv" or "adev" or "aaa". My preference would be for something meaningful. Any ideas?
3. Find some other label fors alpha and beta builds alphabetically greater than "dev" and less than "rc".
Note that even if we do 2 or 3, we still need to do 1.
Of course, another option is to ignore the problem, since we have lived with it already for a long time.
Ideas? | build | do we need a new label for dev builds in our myget feed or in any combined set of feeds which includes it dev builds sometimes appear alongside alpha or beta builds because dev either of the other two it appears to be a better selection to nuget and possibly to some users in fact depending on when the dev build was created it may actually be either or than the others ways we can fix this delete dev builds from myget whenever an alpha or beta is produced this is what i have been doing if this is done consistently then any dev builds present are guaranteed to be later builds than an alpha or beta find some other label alphabetically less than alpha like adv or adev or aaa my preference would be for something meaningful any ideas find some other label fors alpha and beta builds alphabetically greater than dev and less than rc note that even if we do or we still need to do of course another option is to ignore the problem since we have lived with it already for a long time ideas | 1 |
68,274 | 17,203,025,510 | IssuesEvent | 2021-07-17 16:55:44 | Ansible2/Arma-3-Survival | https://api.github.com/repos/Ansible2/Arma-3-Survival | closed | CAS rocket strikes land very far from target | MASTER Build bug enhancement | Happened with CUP vehicles on United Sahrani. Gun Runs seemed unaffected. | 1.0 | CAS rocket strikes land very far from target - Happened with CUP vehicles on United Sahrani. Gun Runs seemed unaffected. | build | cas rocket strikes land very far from target happened with cup vehicles on united sahrani gun runs seemed unaffected | 1 |
68,368 | 17,261,839,315 | IssuesEvent | 2021-07-22 08:43:59 | appsmithorg/appsmith | https://api.github.com/repos/appsmithorg/appsmith | closed | [Bug] Table data is not populating when extra space inside the mustache binding | Bug Critical Table Widget UI Building Pod Widgets | Issues occurs when extra space is provided in mustache binding:
```
{{
currentRow.dob == null ? "no birth" : moment(currentRow.dob).add(24, "years") < moment() ? "small" : moment(currentRow.dob).add(18, "years") < moment() ? "very small" : "very very small"
}}
```
Worked fine for:
`{{currentRow.dob == null ? "no birth" : moment(currentRow.dob).add(24, "years") < moment() ? "small" : moment(currentRow.dob).add(18, "years") < moment() ? "very small" : "very very small"}}
`
- Version: Cloud
- OS: MacOSX
- Browser: chrome
- Environment: production
| 1.0 | [Bug] Table data is not populating when extra space inside the mustache binding - Issues occurs when extra space is provided in mustache binding:
```
{{
currentRow.dob == null ? "no birth" : moment(currentRow.dob).add(24, "years") < moment() ? "small" : moment(currentRow.dob).add(18, "years") < moment() ? "very small" : "very very small"
}}
```
Worked fine for:
`{{currentRow.dob == null ? "no birth" : moment(currentRow.dob).add(24, "years") < moment() ? "small" : moment(currentRow.dob).add(18, "years") < moment() ? "very small" : "very very small"}}
`
- Version: Cloud
- OS: MacOSX
- Browser: chrome
- Environment: production
| build | table data is not populating when extra space inside the mustache binding issues occurs when extra space is provided in mustache binding currentrow dob null no birth moment currentrow dob add years moment small moment currentrow dob add years moment very small very very small worked fine for currentrow dob null no birth moment currentrow dob add years moment small moment currentrow dob add years moment very small very very small version cloud os macosx browser chrome environment production | 1 |
5,413 | 3,588,098,888 | IssuesEvent | 2016-01-30 19:53:23 | Gheoan/gheoan | https://api.github.com/repos/Gheoan/gheoan | opened | consider dropping Typescript | build enhancement not sure | Use Babel + Flowtype
Why?
- Tied to Microsoft
- Babel has more plugins and better support for new features
Why not?
- VSCode support is lost
| 1.0 | consider dropping Typescript - Use Babel + Flowtype
Why?
- Tied to Microsoft
- Babel has more plugins and better support for new features
Why not?
- VSCode support is lost
| build | consider dropping typescript use babel flowtype why tied to microsoft babel has more plugins and better support for new features why not vscode support is lost | 1 |
77,918 | 22,041,085,976 | IssuesEvent | 2022-05-29 11:15:07 | atk4/data | https://api.github.com/repos/atk4/data | closed | Implement field grouping for models | enhancement Saasty.io :building_construction: | See also: https://github.com/atk4/ui/pull/600
Currently, you can add as many fields as you want to the model:
``` php
$model->addField('name');
$model->addField('surname');
```
However, it would be nice if we could also group fields. For instance:
``` php
$model->addFieldGroup('full_name', ['name','surname']);
```
Here we define fields first and then group name. Here are just some things you can do with groups:
``` php
$model->addFieldGroup('full_name', ['name', 'surname'], ['Your Full Name', 'ui'=>['class'=>['red']]);
```
It should be even possible to use sub-groups:
``` php
$model->addFieldGroup('personal_information', ['full_name', 'address']);
```
In theory nesting could be as deep as you wish and field can also be part of multiple groups. Group should be unique and it's name should not be same as the field. Once group is defined, it can be used in setModel:
```
$form->setModel($user, ['personal_information']);
```
Instead of listing fields, specifying group will display all the fields in the group. The following rules will be respected:
- if possible, fields will be laid out to respect the grouping. For instance first and last name will be on same line. full name and address will be grouped into sections.
- group 'ui' property can suggest which layout to use for the form, pretty much same way how fields can use ui/form property.
- it is a job of the UI element to understand and treat groups properly. If that's not being done and group name is passed to the Model (e.g. onlyfields etc) then fields will be replaced with all the fields it contains. That's to insure compatibility and simplify implementation for UI elements that are not recursive.
- atk API, and export can also make use of grouping to group data, for instance get(['id', 'personal_information']) would return array with two elements, but second element would be array etc.
| 1.0 | Implement field grouping for models - See also: https://github.com/atk4/ui/pull/600
Currently, you can add as many fields as you want to the model:
``` php
$model->addField('name');
$model->addField('surname');
```
However, it would be nice if we could also group fields. For instance:
``` php
$model->addFieldGroup('full_name', ['name','surname']);
```
Here we define fields first and then group name. Here are just some things you can do with groups:
``` php
$model->addFieldGroup('full_name', ['name', 'surname'], ['Your Full Name', 'ui'=>['class'=>['red']]);
```
It should be even possible to use sub-groups:
``` php
$model->addFieldGroup('personal_information', ['full_name', 'address']);
```
In theory nesting could be as deep as you wish and field can also be part of multiple groups. Group should be unique and it's name should not be same as the field. Once group is defined, it can be used in setModel:
```
$form->setModel($user, ['personal_information']);
```
Instead of listing fields, specifying group will display all the fields in the group. The following rules will be respected:
- if possible, fields will be laid out to respect the grouping. For instance first and last name will be on same line. full name and address will be grouped into sections.
- group 'ui' property can suggest which layout to use for the form, pretty much same way how fields can use ui/form property.
- it is a job of the UI element to understand and treat groups properly. If that's not being done and group name is passed to the Model (e.g. onlyfields etc) then fields will be replaced with all the fields it contains. That's to insure compatibility and simplify implementation for UI elements that are not recursive.
- atk API, and export can also make use of grouping to group data, for instance get(['id', 'personal_information']) would return array with two elements, but second element would be array etc.
| build | implement field grouping for models see also currently you can add as many fields as you want to the model php model addfield name model addfield surname however it would be nice if we could also group fields for instance php model addfieldgroup full name here we define fields first and then group name here are just some things you can do with groups php model addfieldgroup full name it should be even possible to use sub groups php model addfieldgroup personal information in theory nesting could be as deep as you wish and field can also be part of multiple groups group should be unique and it s name should not be same as the field once group is defined it can be used in setmodel form setmodel user instead of listing fields specifying group will display all the fields in the group the following rules will be respected if possible fields will be laid out to respect the grouping for instance first and last name will be on same line full name and address will be grouped into sections group ui property can suggest which layout to use for the form pretty much same way how fields can use ui form property it is a job of the ui element to understand and treat groups properly if that s not being done and group name is passed to the model e g onlyfields etc then fields will be replaced with all the fields it contains that s to insure compatibility and simplify implementation for ui elements that are not recursive atk api and export can also make use of grouping to group data for instance get would return array with two elements but second element would be array etc | 1 |
53,748 | 23,051,476,688 | IssuesEvent | 2022-07-24 17:44:03 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | request: aws_iam_policy_document pre-flight checks | enhancement service/iam stale | _This issue was originally opened by @therealbill as hashicorp/terraform#11444. It was migrated here as part of the [provider split](https://www.hashicorp.com/blog/upcoming-provider-changes-in-terraform-0-10/). The original body of the issue is below._
<hr>
# Affected Resource(s)
- aws_iam_policy_document
AWS has a hard limit on the size in characters of a policy document (the JSON it uses) of 2048 bytes. Trying to add a policy larger than that will result in an error on apply.
# Current Behavior
The plan "works", the apply will fail.
# Desired Behavior
I would rather see the plan indicate the policy it is going to write is too large so it can be caught in that stage
In this case, terraform knows the size of the JSON document it just created and will attempt to upload, so I think it reasonable for it to check since it does not need info from the provider (this is one of the hard limits, you can't get them to raise it for you).
Also, the value of "sid" can only be alphanumeric, so I'd like to see the plan report an error if non-alphanumeric characters are in it. Less important then the policy size, of course. | 1.0 | request: aws_iam_policy_document pre-flight checks - _This issue was originally opened by @therealbill as hashicorp/terraform#11444. It was migrated here as part of the [provider split](https://www.hashicorp.com/blog/upcoming-provider-changes-in-terraform-0-10/). The original body of the issue is below._
<hr>
# Affected Resource(s)
- aws_iam_policy_document
AWS has a hard limit on the size in characters of a policy document (the JSON it uses) of 2048 bytes. Trying to add a policy larger than that will result in an error on apply.
# Current Behavior
The plan "works", the apply will fail.
# Desired Behavior
I would rather see the plan indicate the policy it is going to write is too large so it can be caught in that stage
In this case, terraform knows the size of the JSON document it just created and will attempt to upload, so I think it reasonable for it to check since it does not need info from the provider (this is one of the hard limits, you can't get them to raise it for you).
Also, the value of "sid" can only be alphanumeric, so I'd like to see the plan report an error if non-alphanumeric characters are in it. Less important then the policy size, of course. | non_build | request aws iam policy document pre flight checks this issue was originally opened by therealbill as hashicorp terraform it was migrated here as part of the the original body of the issue is below affected resource s aws iam policy document aws has a hard limit on the size in characters of a policy document the json it uses of bytes trying to add a policy larger than that will result in an error on apply current behavior the plan works the apply will fail desired behavior i would rather see the plan indicate the policy it is going to write is too large so it can be caught in that stage in this case terraform knows the size of the json document it just created and will attempt to upload so i think it reasonable for it to check since it does not need info from the provider this is one of the hard limits you can t get them to raise it for you also the value of sid can only be alphanumeric so i d like to see the plan report an error if non alphanumeric characters are in it less important then the policy size of course | 0 |
85,104 | 15,731,207,263 | IssuesEvent | 2021-03-29 16:47:46 | wrbejar/bag-of-holding | https://api.github.com/repos/wrbejar/bag-of-holding | opened | CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz | security vulnerability | ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: bag-of-holding/package.json</p>
<p>Path to vulnerable library: bag-of-holding/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-1.3.3.tgz (Root Library)
- node-sass-2.1.1.tgz
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wrbejar/bag-of-holding/commit/6087cf643d57f8f112ae650913c59bfc0a1033d6">6087cf643d57f8f112ae650913c59bfc0a1033d6</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via shortcutMatch in fromUrl().
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 3.0.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp-sass:1.3.3;node-sass:2.1.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via shortcutMatch in fromUrl().","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: bag-of-holding/package.json</p>
<p>Path to vulnerable library: bag-of-holding/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- gulp-sass-1.3.3.tgz (Root Library)
- node-sass-2.1.1.tgz
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wrbejar/bag-of-holding/commit/6087cf643d57f8f112ae650913c59bfc0a1033d6">6087cf643d57f8f112ae650913c59bfc0a1033d6</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via shortcutMatch in fromUrl().
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/hosted-git-info/releases/tag/v3.0.8">https://github.com/npm/hosted-git-info/releases/tag/v3.0.8</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 3.0.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"gulp-sass:1.3.3;node-sass:2.1.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via shortcutMatch in fromUrl().","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_build | cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file bag of holding package json path to vulnerable library bag of holding node modules hosted git info package json dependency hierarchy gulp sass tgz root library node sass tgz meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via shortcutmatch in fromurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree gulp sass node sass meow normalize package data hosted git info isminimumfixversionavailable true minimumfixversion hosted git info basebranches vulnerabilityidentifier cve vulnerabilitydetails the package hosted git info before are vulnerable to regular expression denial of service redos via shortcutmatch in fromurl vulnerabilityurl | 0 |
214,130 | 16,564,430,762 | IssuesEvent | 2021-05-29 05:32:14 | DCS-LCSR/SignStream3 | https://api.github.com/repos/DCS-LCSR/SignStream3 | closed | Testing 3.3.6 - in Big Sur | TESTING_3.3.6 | I opened a file, and immediately got this error message. I was, however, able to open the video file by choosing Edit Macro Unit from the menu. I might also mention that "Resouce" is misspelled.
<img width="643" alt="Screen Shot 2021-02-26 at 2 38 17 PM" src="https://user-images.githubusercontent.com/13629122/109347023-80dae000-7840-11eb-879d-b4720e8deb33.png">
| 1.0 | Testing 3.3.6 - in Big Sur - I opened a file, and immediately got this error message. I was, however, able to open the video file by choosing Edit Macro Unit from the menu. I might also mention that "Resouce" is misspelled.
<img width="643" alt="Screen Shot 2021-02-26 at 2 38 17 PM" src="https://user-images.githubusercontent.com/13629122/109347023-80dae000-7840-11eb-879d-b4720e8deb33.png">
| non_build | testing in big sur i opened a file and immediately got this error message i was however able to open the video file by choosing edit macro unit from the menu i might also mention that resouce is misspelled img width alt screen shot at pm src | 0 |
19,067 | 11,138,290,081 | IssuesEvent | 2019-12-20 21:53:50 | Azure/azure-powershell | https://api.github.com/repos/Azure/azure-powershell | closed | Unable to save CDN Rule Conditions | CDN Service Attention customer-reported | ## Description
I'm attempting to migrate from AzureRm to Az. In AzureRm, I was using the following (likely overly verbose) code to create a CDN Rule Condition, specifically for adding a custom caching rule to bypass the default cache policy for the hypothetical file `/foo.bar`:
```powershell
$rule = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRule
$ruleConditions = New-Object System.Collections.Generic.List[Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRuleCondition]
$ruleCondition = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRuleUrlPathCondition
$ruleConditionParameters = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSUrlPathConditionParameters
$ruleConditionParameters.Path = '/foo.bar'
$ruleConditionParameters.MatchType = 'Wildcard'
$ruleCondition.Parameters = $ruleConditionParameters
$ruleConditions.Add($ruleCondition)
$rule.Conditions = $ruleConditions
```
However, in powershell Az, even with aliases, neither `PSDeliveryRuleUrlPathCondition` or `PSUrlPathConditionParameters` objects seem to exist.
I've tried refactoring this code a couple of ways including:
```powershell
#...
$ruleCondition = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRuleCondition
$ruleCondition.MatchVariable = 'UrlPath'
$matchValues = New-Object System.Collections.Generic.List[string]
$matchValues.Add('/foo.bar')
$ruleCondition.MatchValue = $matchValues
#...
```
Running in debug, this sends the json payload (snippet):
```json
"conditions": [
{
"name": "UrlPath",
"parameters": {
"matchValues": [
"/foo.bar"
],
"@odata.type": "#Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters"
}
}
]
```
which returns
```json
"error": {
"code": "InvalidResource",
"message": "The property 'matchValues' does not exist on type 'Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters'. Make sure to only use property names that are defined by the type."
}
```
I've also tried:
```powershell
#...
$ruleCondition = New-AzCdnDeliveryRuleCondition -MatchVariable UrlPath -Operator Any -MatchValue '/foo.bar'
#...
```
which sends the json payload (snippet):
```json
"conditions": [
{
"name": "UrlPath",
"parameters": {
"operator": "Any",
"negateCondition": false,
"matchValues": [
"/foo.bar"
],
"@odata.type": "#Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters"
}
}
]
```
which returns
```json
"error": {
"code": "InvalidResource",
"message": "The property 'operator' does not exist on type 'Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters'. Make sure to only use property names that are defined by the type."
}
```
I'm able to create these caching rules just fine in the Azure Portal following the instructions (https://docs.microsoft.com/en-us/azure/cdn/cdn-caching-rules), and inspecting the json object snippet looks like:
```json
"conditions": [{
"name": "UrlPath",
"parameters": {
"@odata.type": "#Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters",
"matchType": "Wildcard",
"path": "/foo.bar"
}
}
]
```
Other notes: I tried installing previous versions of the `Az.Cdn` module, and with `Enable-AzureRmAlias`, my original old code had been working up to and including version `1.2.0`. My original code broke with the release of version `1.3.0`, where the types I was using appear to have been removed.
Looking at the release notes (https://docs.microsoft.com/en-us/powershell/azure/release-notes-azureps?view=azps-3.0.0#azcdn-2) this appears to be with the introduction of rulesEngine.
The documentation (https://docs.microsoft.com/en-us/azure/cdn/cdn-verizon-premium-rules-engine) states that the rules engine is for Premium Verizon only. However, I'm only attempting to use standard caching rules which should be available in my tier Standard Akamai (https://docs.microsoft.com/en-us/azure/cdn/cdn-features)
## Steps to reproduce
Described above
## Environment data
<!-- Please run $PSVersionTable and paste the output in the below code block -->
```
PS > $PSVersionTable
Name Value
---- -----
PSVersion 5.1.18362.145
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.18362.145
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
```
## Module versions
<!-- Please run (Get-Module -ListAvailable) and paste the output in the below code block -->
```powershell
ModuleType Version Name ExportedCommands
---------- ------- ---- ----------------
Script 1.6.4 Az.Accounts {Disable-AzDataCollection, Disable-AzContextAutosave, Enable-AzDataCollection, Enable-AzContextAutosave...}
Script 1.1.0 Az.Advisor {Get-AzAdvisorRecommendation, Enable-AzAdvisorRecommendation, Disable-AzAdvisorRecommendation, Get-AzAdvisorConfiguration...}
Script 1.0.2 Az.Aks {Get-AzAks, New-AzAks, Remove-AzAks, Import-AzAksCredential...}
Script 1.1.1 Az.AnalysisServices {Resume-AzAnalysisServicesServer, Suspend-AzAnalysisServicesServer, Get-AzAnalysisServicesServer, Remove-AzAnalysisServicesServer...}
Script 1.3.2 Az.ApiManagement {Add-AzApiManagementApiToProduct, Add-AzApiManagementProductToGroup, Add-AzApiManagementRegion, Add-AzApiManagementUserToGroup...}
Script 1.0.2 Az.ApplicationInsights {Get-AzApplicationInsights, New-AzApplicationInsights, Remove-AzApplicationInsights, Set-AzApplicationInsightsPricingPlan...}
Script 1.3.4 Az.Automation {Get-AzAutomationHybridWorkerGroup, Remove-AzAutomationHybridWorkerGroup, Get-AzAutomationJobOutputRecord, Import-AzAutomationDscNodeConfiguration...}
Script 2.0.1 Az.Batch {Remove-AzBatchAccount, Get-AzBatchAccount, Get-AzBatchAccountKey, New-AzBatchAccount...}
Script 1.0.1 Az.Billing {Get-AzBillingInvoice, Get-AzBillingPeriod, Get-AzEnrollmentAccount, Get-AzConsumptionBudget...}
Script 1.4.0 Az.Cdn {Get-AzCdnProfile, Get-AzCdnProfileSsoUrl, New-AzCdnProfile, Remove-AzCdnProfile...}
Script 1.2.1 Az.CognitiveServices {Get-AzCognitiveServicesAccount, Get-AzCognitiveServicesAccountKey, Get-AzCognitiveServicesAccountSku, Get-AzCognitiveServicesAccountType...}
Script 3.0.0 Az.Compute {Remove-AzAvailabilitySet, Get-AzAvailabilitySet, New-AzAvailabilitySet, Update-AzAvailabilitySet...}
Script 1.0.1 Az.ContainerInstance {New-AzContainerGroup, Get-AzContainerGroup, Remove-AzContainerGroup, Get-AzContainerInstanceLog}
Script 1.1.0 Az.ContainerRegistry {New-AzContainerRegistry, Get-AzContainerRegistry, Update-AzContainerRegistry, Remove-AzContainerRegistry...}
Script 1.4.1 Az.DataFactory {Set-AzDataFactoryV2, Update-AzDataFactoryV2, Get-AzDataFactoryV2, Remove-AzDataFactoryV2...}
Script 1.0.1 Az.DataLakeAnalytics {Get-AzDataLakeAnalyticsDataSource, New-AzDataLakeAnalyticsCatalogCredential, Remove-AzDataLakeAnalyticsCatalogCredential, Set-AzDataLakeAnalyticsCatalogCredential...}
Script 1.2.4 Az.DataLakeStore {Get-AzDataLakeStoreTrustedIdProvider, Remove-AzDataLakeStoreTrustedIdProvider, Remove-AzDataLakeStoreFirewallRule, Set-AzDataLakeStoreTrustedIdProvider...}
Script 1.0.1 Az.DeploymentManager {Get-AzDeploymentManagerArtifactSource, New-AzDeploymentManagerArtifactSource, Set-AzDeploymentManagerArtifactSource, Remove-AzDeploymentManagerArtifactSource...}
Script 1.0.0 Az.DevTestLabs {Get-AzDtlAllowedVMSizesPolicy, Get-AzDtlAutoShutdownPolicy, Get-AzDtlAutoStartPolicy, Get-AzDtlVMsPerLabPolicy...}
Script 1.1.1 Az.Dns {Get-AzDnsRecordSet, New-AzDnsRecordConfig, Remove-AzDnsRecordSet, Set-AzDnsRecordSet...}
Script 1.2.2 Az.EventGrid {New-AzEventGridTopic, Get-AzEventGridTopic, Set-AzEventGridTopic, New-AzEventGridTopicKey...}
Script 1.4.0 Az.EventHub {New-AzEventHubNamespace, Get-AzEventHubNamespace, Set-AzEventHubNamespace, Remove-AzEventHubNamespace...}
Script 1.1.2 Az.FrontDoor {New-AzFrontDoor, Get-AzFrontDoor, Set-AzFrontDoor, Remove-AzFrontDoor...}
Script 3.0.0 Az.HDInsight {Get-AzHDInsightJob, New-AzHDInsightSqoopJobDefinition, Wait-AzHDInsightJob, New-AzHDInsightStreamingMapReduceJobDefinition...}
Script 1.0.0 Az.HealthcareApis {New-AzHealthcareApisService, Remove-AzHealthcareApisService, Set-AzHealthcareApisService, Get-AzHealthcareApisService}
Script 2.0.0 Az.IotHub {Add-AzIotHubKey, Get-AzIotHubEventHubConsumerGroup, Get-AzIotHubConnectionString, Get-AzIotHubJob...}
Script 1.3.1 Az.KeyVault {Add-AzKeyVaultCertificate, Update-AzKeyVaultCertificate, Stop-AzKeyVaultCertificateOperation, Get-AzKeyVaultCertificateOperation...}
Script 1.3.1 Az.LogicApp {Get-AzIntegrationAccountAgreement, Get-AzIntegrationAccountAssembly, Get-AzIntegrationAccountBatchConfiguration, Get-AzIntegrationAccountCallbackUrl...}
Script 1.1.1 Az.MachineLearning {Move-AzMlCommitmentAssociation, Get-AzMlCommitmentAssociation, Get-AzMlCommitmentPlanUsageHistory, Remove-AzMlCommitmentPlan...}
Script 1.0.1 Az.ManagedServices {Get-AzManagedServicesAssignment, New-AzManagedServicesAssignment, Remove-AzManagedServicesAssignment, Get-AzManagedServicesDefinition...}
Script 1.0.1 Az.MarketplaceOrdering {Get-AzMarketplaceTerms, Set-AzMarketplaceTerms}
Script 1.1.0 Az.Media {Sync-AzMediaServiceStorageKey, Set-AzMediaServiceKey, Get-AzMediaServiceKey, Get-AzMediaServiceNameAvailability...}
Script 1.4.0 Az.Monitor {Get-AzMetricDefinition, Get-AzMetric, Remove-AzLogProfile, Get-AzLogProfile...}
Script 2.0.0 Az.Network {Add-AzApplicationGatewayAuthenticationCertificate, Get-AzApplicationGatewayAuthenticationCertificate, New-AzApplicationGatewayAuthenticationCertificate, Remove-AzApplicationGatewayAuthenticationC...
Script 1.1.0 Az.NotificationHubs {Get-AzNotificationHub, Get-AzNotificationHubAuthorizationRule, Get-AzNotificationHubListKey, Get-AzNotificationHubPNSCredential...}
Script 1.3.3 Az.OperationalInsights {New-AzOperationalInsightsAzureActivityLogDataSource, New-AzOperationalInsightsCustomLogDataSource, Disable-AzOperationalInsightsLinuxCustomLogCollection, Disable-AzOperationalInsightsIISLogCollec...
Script 1.1.3 Az.PolicyInsights {Get-AzPolicyEvent, Get-AzPolicyState, Get-AzPolicyStateSummary, Get-AzPolicyRemediation...}
Script 1.1.0 Az.PowerBIEmbedded {Remove-AzPowerBIWorkspaceCollection, Get-AzPowerBIWorkspaceCollection, Get-AzPowerBIWorkspaceCollectionAccessKey, Get-AzPowerBIWorkspace...}
Script 1.0.0 Az.PrivateDns {Get-AzPrivateDnsZone, Remove-AzPrivateDnsZone, Set-AzPrivateDnsZone, New-AzPrivateDnsZone...}
Script 2.0.1 Az.RecoveryServices {Get-AzRecoveryServicesBackupProperty, Get-AzRecoveryServicesVault, Get-AzRecoveryServicesVaultSettingsFile, New-AzRecoveryServicesVault...}
Script 1.1.1 Az.RedisCache {Remove-AzRedisCachePatchSchedule, New-AzRedisCacheScheduleEntry, Get-AzRedisCachePatchSchedule, New-AzRedisCachePatchSchedule...}
Script 1.0.2 Az.Relay {New-AzRelayNamespace, Get-AzRelayNamespace, Set-AzRelayNamespace, Remove-AzRelayNamespace...}
Script 1.7.1 Az.Resources {Get-AzProviderOperation, Remove-AzRoleAssignment, Get-AzRoleAssignment, New-AzRoleAssignment...}
Script 1.4.0 Az.ServiceBus {New-AzServiceBusNamespace, Get-AzServiceBusNamespace, Set-AzServiceBusNamespace, Remove-AzServiceBusNamespace...}
Script 2.0.0 Az.ServiceFabric {Add-AzServiceFabricClientCertificate, Add-AzServiceFabricClusterCertificate, Add-AzServiceFabricNode, Add-AzServiceFabricNodeType...}
Script 1.1.0 Az.SignalR {New-AzSignalR, Get-AzSignalR, Get-AzSignalRKey, New-AzSignalRKey...}
Script 2.0.0 Az.Sql {Get-AzSqlDatabaseTransparentDataEncryption, Get-AzSqlDatabaseTransparentDataEncryptionActivity, Set-AzSqlDatabaseTransparentDataEncryption, Get-AzSqlDatabaseUpgradeHint...}
Script 1.9.0 Az.Storage {Get-AzStorageAccount, Get-AzStorageAccountKey, New-AzStorageAccount, New-AzStorageAccountKey...}
Script 1.2.1 Az.StorageSync {Invoke-AzStorageSyncCompatibilityCheck, New-AzStorageSyncService, Get-AzStorageSyncService, Remove-AzStorageSyncService...}
Script 1.0.0 Az.StreamAnalytics {Get-AzStreamAnalyticsFunction, Get-AzStreamAnalyticsDefaultFunctionDefinition, New-AzStreamAnalyticsFunction, Remove-AzStreamAnalyticsFunction...}
Script 1.0.2 Az.TrafficManager {Add-AzTrafficManagerCustomHeaderToEndpoint, Remove-AzTrafficManagerCustomHeaderFromEndpoint, Add-AzTrafficManagerCustomHeaderToProfile, Remove-AzTrafficManagerCustomHeaderFromProfile...}
Script 1.5.0 Az.Websites {Get-AzAppServicePlan, Set-AzAppServicePlan, New-AzAppServicePlan, Remove-AzAppServicePlan...}
```
## Debug output
<!-- Set $DebugPreference='Continue' before running the repro and paste the resulting debug stream in the below code block -->
I've posted the relevant snippets above.
## Error output
<!-- Please run Resolve-AzError and paste the output in the below code block -->
```
Message : Operation returned an invalid status code 'BadRequest'
StackTrace : at Microsoft.Azure.Management.Cdn.EndpointsOperations.<BeginCreateWithHttpMessagesAsync>d__16.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.ConfiguredTaskAwaitable`1.ConfiguredTaskAwaiter.GetResult()
at Microsoft.Azure.Management.Cdn.EndpointsOperations.<CreateWithHttpMessagesAsync>d__7.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.ConfiguredTaskAwaitable`1.ConfiguredTaskAwaiter.GetResult()
at Microsoft.Azure.Management.Cdn.EndpointsOperationsExtensions.<CreateAsync>d__5.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult()
at Microsoft.Azure.Management.Cdn.EndpointsOperationsExtensions.Create(IEndpointsOperations operations, String resourceGroupName, String profileName, String endpointName, Endpoint endpoint)
at Microsoft.Azure.Commands.Cdn.Endpoint.NewAzureRmCdnEndpoint.NewEndpoint()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action)
at Microsoft.Azure.Commands.Cdn.Endpoint.NewAzureRmCdnEndpoint.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : Microsoft.Azure.Management.Cdn.Models.ErrorResponseException
InvocationInfo : {New-AzCdnEndpoint}
Line : New-AzCdnEndpoint `
```
| 1.0 | Unable to save CDN Rule Conditions - ## Description
I'm attempting to migrate from AzureRm to Az. In AzureRm, I was using the following (likely overly verbose) code to create a CDN Rule Condition, specifically for adding a custom caching rule to bypass the default cache policy for the hypothetical file `/foo.bar`:
```powershell
$rule = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRule
$ruleConditions = New-Object System.Collections.Generic.List[Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRuleCondition]
$ruleCondition = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRuleUrlPathCondition
$ruleConditionParameters = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSUrlPathConditionParameters
$ruleConditionParameters.Path = '/foo.bar'
$ruleConditionParameters.MatchType = 'Wildcard'
$ruleCondition.Parameters = $ruleConditionParameters
$ruleConditions.Add($ruleCondition)
$rule.Conditions = $ruleConditions
```
However, in powershell Az, even with aliases, neither `PSDeliveryRuleUrlPathCondition` or `PSUrlPathConditionParameters` objects seem to exist.
I've tried refactoring this code a couple of ways including:
```powershell
#...
$ruleCondition = New-Object Microsoft.Azure.Commands.Cdn.Models.Endpoint.PSDeliveryRuleCondition
$ruleCondition.MatchVariable = 'UrlPath'
$matchValues = New-Object System.Collections.Generic.List[string]
$matchValues.Add('/foo.bar')
$ruleCondition.MatchValue = $matchValues
#...
```
Running in debug, this sends the json payload (snippet):
```json
"conditions": [
{
"name": "UrlPath",
"parameters": {
"matchValues": [
"/foo.bar"
],
"@odata.type": "#Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters"
}
}
]
```
which returns
```json
"error": {
"code": "InvalidResource",
"message": "The property 'matchValues' does not exist on type 'Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters'. Make sure to only use property names that are defined by the type."
}
```
I've also tried:
```powershell
#...
$ruleCondition = New-AzCdnDeliveryRuleCondition -MatchVariable UrlPath -Operator Any -MatchValue '/foo.bar'
#...
```
which sends the json payload (snippet):
```json
"conditions": [
{
"name": "UrlPath",
"parameters": {
"operator": "Any",
"negateCondition": false,
"matchValues": [
"/foo.bar"
],
"@odata.type": "#Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters"
}
}
]
```
which returns
```json
"error": {
"code": "InvalidResource",
"message": "The property 'operator' does not exist on type 'Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters'. Make sure to only use property names that are defined by the type."
}
```
I'm able to create these caching rules just fine in the Azure Portal following the instructions (https://docs.microsoft.com/en-us/azure/cdn/cdn-caching-rules), and inspecting the json object snippet looks like:
```json
"conditions": [{
"name": "UrlPath",
"parameters": {
"@odata.type": "#Microsoft.Azure.Cdn.Models.DeliveryRuleUrlPathConditionParameters",
"matchType": "Wildcard",
"path": "/foo.bar"
}
}
]
```
Other notes: I tried installing previous versions of the `Az.Cdn` module, and with `Enable-AzureRmAlias`, my original old code had been working up to and including version `1.2.0`. My original code broke with the release of version `1.3.0`, where the types I was using appear to have been removed.
Looking at the release notes (https://docs.microsoft.com/en-us/powershell/azure/release-notes-azureps?view=azps-3.0.0#azcdn-2) this appears to be with the introduction of rulesEngine.
The documentation (https://docs.microsoft.com/en-us/azure/cdn/cdn-verizon-premium-rules-engine) states that the rules engine is for Premium Verizon only. However, I'm only attempting to use standard caching rules which should be available in my tier Standard Akamai (https://docs.microsoft.com/en-us/azure/cdn/cdn-features)
## Steps to reproduce
Described above
## Environment data
<!-- Please run $PSVersionTable and paste the output in the below code block -->
```
PS > $PSVersionTable
Name Value
---- -----
PSVersion 5.1.18362.145
PSEdition Desktop
PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...}
BuildVersion 10.0.18362.145
CLRVersion 4.0.30319.42000
WSManStackVersion 3.0
PSRemotingProtocolVersion 2.3
SerializationVersion 1.1.0.1
```
## Module versions
<!-- Please run (Get-Module -ListAvailable) and paste the output in the below code block -->
```powershell
ModuleType Version Name ExportedCommands
---------- ------- ---- ----------------
Script 1.6.4 Az.Accounts {Disable-AzDataCollection, Disable-AzContextAutosave, Enable-AzDataCollection, Enable-AzContextAutosave...}
Script 1.1.0 Az.Advisor {Get-AzAdvisorRecommendation, Enable-AzAdvisorRecommendation, Disable-AzAdvisorRecommendation, Get-AzAdvisorConfiguration...}
Script 1.0.2 Az.Aks {Get-AzAks, New-AzAks, Remove-AzAks, Import-AzAksCredential...}
Script 1.1.1 Az.AnalysisServices {Resume-AzAnalysisServicesServer, Suspend-AzAnalysisServicesServer, Get-AzAnalysisServicesServer, Remove-AzAnalysisServicesServer...}
Script 1.3.2 Az.ApiManagement {Add-AzApiManagementApiToProduct, Add-AzApiManagementProductToGroup, Add-AzApiManagementRegion, Add-AzApiManagementUserToGroup...}
Script 1.0.2 Az.ApplicationInsights {Get-AzApplicationInsights, New-AzApplicationInsights, Remove-AzApplicationInsights, Set-AzApplicationInsightsPricingPlan...}
Script 1.3.4 Az.Automation {Get-AzAutomationHybridWorkerGroup, Remove-AzAutomationHybridWorkerGroup, Get-AzAutomationJobOutputRecord, Import-AzAutomationDscNodeConfiguration...}
Script 2.0.1 Az.Batch {Remove-AzBatchAccount, Get-AzBatchAccount, Get-AzBatchAccountKey, New-AzBatchAccount...}
Script 1.0.1 Az.Billing {Get-AzBillingInvoice, Get-AzBillingPeriod, Get-AzEnrollmentAccount, Get-AzConsumptionBudget...}
Script 1.4.0 Az.Cdn {Get-AzCdnProfile, Get-AzCdnProfileSsoUrl, New-AzCdnProfile, Remove-AzCdnProfile...}
Script 1.2.1 Az.CognitiveServices {Get-AzCognitiveServicesAccount, Get-AzCognitiveServicesAccountKey, Get-AzCognitiveServicesAccountSku, Get-AzCognitiveServicesAccountType...}
Script 3.0.0 Az.Compute {Remove-AzAvailabilitySet, Get-AzAvailabilitySet, New-AzAvailabilitySet, Update-AzAvailabilitySet...}
Script 1.0.1 Az.ContainerInstance {New-AzContainerGroup, Get-AzContainerGroup, Remove-AzContainerGroup, Get-AzContainerInstanceLog}
Script 1.1.0 Az.ContainerRegistry {New-AzContainerRegistry, Get-AzContainerRegistry, Update-AzContainerRegistry, Remove-AzContainerRegistry...}
Script 1.4.1 Az.DataFactory {Set-AzDataFactoryV2, Update-AzDataFactoryV2, Get-AzDataFactoryV2, Remove-AzDataFactoryV2...}
Script 1.0.1 Az.DataLakeAnalytics {Get-AzDataLakeAnalyticsDataSource, New-AzDataLakeAnalyticsCatalogCredential, Remove-AzDataLakeAnalyticsCatalogCredential, Set-AzDataLakeAnalyticsCatalogCredential...}
Script 1.2.4 Az.DataLakeStore {Get-AzDataLakeStoreTrustedIdProvider, Remove-AzDataLakeStoreTrustedIdProvider, Remove-AzDataLakeStoreFirewallRule, Set-AzDataLakeStoreTrustedIdProvider...}
Script 1.0.1 Az.DeploymentManager {Get-AzDeploymentManagerArtifactSource, New-AzDeploymentManagerArtifactSource, Set-AzDeploymentManagerArtifactSource, Remove-AzDeploymentManagerArtifactSource...}
Script 1.0.0 Az.DevTestLabs {Get-AzDtlAllowedVMSizesPolicy, Get-AzDtlAutoShutdownPolicy, Get-AzDtlAutoStartPolicy, Get-AzDtlVMsPerLabPolicy...}
Script 1.1.1 Az.Dns {Get-AzDnsRecordSet, New-AzDnsRecordConfig, Remove-AzDnsRecordSet, Set-AzDnsRecordSet...}
Script 1.2.2 Az.EventGrid {New-AzEventGridTopic, Get-AzEventGridTopic, Set-AzEventGridTopic, New-AzEventGridTopicKey...}
Script 1.4.0 Az.EventHub {New-AzEventHubNamespace, Get-AzEventHubNamespace, Set-AzEventHubNamespace, Remove-AzEventHubNamespace...}
Script 1.1.2 Az.FrontDoor {New-AzFrontDoor, Get-AzFrontDoor, Set-AzFrontDoor, Remove-AzFrontDoor...}
Script 3.0.0 Az.HDInsight {Get-AzHDInsightJob, New-AzHDInsightSqoopJobDefinition, Wait-AzHDInsightJob, New-AzHDInsightStreamingMapReduceJobDefinition...}
Script 1.0.0 Az.HealthcareApis {New-AzHealthcareApisService, Remove-AzHealthcareApisService, Set-AzHealthcareApisService, Get-AzHealthcareApisService}
Script 2.0.0 Az.IotHub {Add-AzIotHubKey, Get-AzIotHubEventHubConsumerGroup, Get-AzIotHubConnectionString, Get-AzIotHubJob...}
Script 1.3.1 Az.KeyVault {Add-AzKeyVaultCertificate, Update-AzKeyVaultCertificate, Stop-AzKeyVaultCertificateOperation, Get-AzKeyVaultCertificateOperation...}
Script 1.3.1 Az.LogicApp {Get-AzIntegrationAccountAgreement, Get-AzIntegrationAccountAssembly, Get-AzIntegrationAccountBatchConfiguration, Get-AzIntegrationAccountCallbackUrl...}
Script 1.1.1 Az.MachineLearning {Move-AzMlCommitmentAssociation, Get-AzMlCommitmentAssociation, Get-AzMlCommitmentPlanUsageHistory, Remove-AzMlCommitmentPlan...}
Script 1.0.1 Az.ManagedServices {Get-AzManagedServicesAssignment, New-AzManagedServicesAssignment, Remove-AzManagedServicesAssignment, Get-AzManagedServicesDefinition...}
Script 1.0.1 Az.MarketplaceOrdering {Get-AzMarketplaceTerms, Set-AzMarketplaceTerms}
Script 1.1.0 Az.Media {Sync-AzMediaServiceStorageKey, Set-AzMediaServiceKey, Get-AzMediaServiceKey, Get-AzMediaServiceNameAvailability...}
Script 1.4.0 Az.Monitor {Get-AzMetricDefinition, Get-AzMetric, Remove-AzLogProfile, Get-AzLogProfile...}
Script 2.0.0 Az.Network {Add-AzApplicationGatewayAuthenticationCertificate, Get-AzApplicationGatewayAuthenticationCertificate, New-AzApplicationGatewayAuthenticationCertificate, Remove-AzApplicationGatewayAuthenticationC...
Script 1.1.0 Az.NotificationHubs {Get-AzNotificationHub, Get-AzNotificationHubAuthorizationRule, Get-AzNotificationHubListKey, Get-AzNotificationHubPNSCredential...}
Script 1.3.3 Az.OperationalInsights {New-AzOperationalInsightsAzureActivityLogDataSource, New-AzOperationalInsightsCustomLogDataSource, Disable-AzOperationalInsightsLinuxCustomLogCollection, Disable-AzOperationalInsightsIISLogCollec...
Script 1.1.3 Az.PolicyInsights {Get-AzPolicyEvent, Get-AzPolicyState, Get-AzPolicyStateSummary, Get-AzPolicyRemediation...}
Script 1.1.0 Az.PowerBIEmbedded {Remove-AzPowerBIWorkspaceCollection, Get-AzPowerBIWorkspaceCollection, Get-AzPowerBIWorkspaceCollectionAccessKey, Get-AzPowerBIWorkspace...}
Script 1.0.0 Az.PrivateDns {Get-AzPrivateDnsZone, Remove-AzPrivateDnsZone, Set-AzPrivateDnsZone, New-AzPrivateDnsZone...}
Script 2.0.1 Az.RecoveryServices {Get-AzRecoveryServicesBackupProperty, Get-AzRecoveryServicesVault, Get-AzRecoveryServicesVaultSettingsFile, New-AzRecoveryServicesVault...}
Script 1.1.1 Az.RedisCache {Remove-AzRedisCachePatchSchedule, New-AzRedisCacheScheduleEntry, Get-AzRedisCachePatchSchedule, New-AzRedisCachePatchSchedule...}
Script 1.0.2 Az.Relay {New-AzRelayNamespace, Get-AzRelayNamespace, Set-AzRelayNamespace, Remove-AzRelayNamespace...}
Script 1.7.1 Az.Resources {Get-AzProviderOperation, Remove-AzRoleAssignment, Get-AzRoleAssignment, New-AzRoleAssignment...}
Script 1.4.0 Az.ServiceBus {New-AzServiceBusNamespace, Get-AzServiceBusNamespace, Set-AzServiceBusNamespace, Remove-AzServiceBusNamespace...}
Script 2.0.0 Az.ServiceFabric {Add-AzServiceFabricClientCertificate, Add-AzServiceFabricClusterCertificate, Add-AzServiceFabricNode, Add-AzServiceFabricNodeType...}
Script 1.1.0 Az.SignalR {New-AzSignalR, Get-AzSignalR, Get-AzSignalRKey, New-AzSignalRKey...}
Script 2.0.0 Az.Sql {Get-AzSqlDatabaseTransparentDataEncryption, Get-AzSqlDatabaseTransparentDataEncryptionActivity, Set-AzSqlDatabaseTransparentDataEncryption, Get-AzSqlDatabaseUpgradeHint...}
Script 1.9.0 Az.Storage {Get-AzStorageAccount, Get-AzStorageAccountKey, New-AzStorageAccount, New-AzStorageAccountKey...}
Script 1.2.1 Az.StorageSync {Invoke-AzStorageSyncCompatibilityCheck, New-AzStorageSyncService, Get-AzStorageSyncService, Remove-AzStorageSyncService...}
Script 1.0.0 Az.StreamAnalytics {Get-AzStreamAnalyticsFunction, Get-AzStreamAnalyticsDefaultFunctionDefinition, New-AzStreamAnalyticsFunction, Remove-AzStreamAnalyticsFunction...}
Script 1.0.2 Az.TrafficManager {Add-AzTrafficManagerCustomHeaderToEndpoint, Remove-AzTrafficManagerCustomHeaderFromEndpoint, Add-AzTrafficManagerCustomHeaderToProfile, Remove-AzTrafficManagerCustomHeaderFromProfile...}
Script 1.5.0 Az.Websites {Get-AzAppServicePlan, Set-AzAppServicePlan, New-AzAppServicePlan, Remove-AzAppServicePlan...}
```
## Debug output
<!-- Set $DebugPreference='Continue' before running the repro and paste the resulting debug stream in the below code block -->
I've posted the relevant snippets above.
## Error output
<!-- Please run Resolve-AzError and paste the output in the below code block -->
```
Message : Operation returned an invalid status code 'BadRequest'
StackTrace : at Microsoft.Azure.Management.Cdn.EndpointsOperations.<BeginCreateWithHttpMessagesAsync>d__16.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.ConfiguredTaskAwaitable`1.ConfiguredTaskAwaiter.GetResult()
at Microsoft.Azure.Management.Cdn.EndpointsOperations.<CreateWithHttpMessagesAsync>d__7.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.ConfiguredTaskAwaitable`1.ConfiguredTaskAwaiter.GetResult()
at Microsoft.Azure.Management.Cdn.EndpointsOperationsExtensions.<CreateAsync>d__5.MoveNext()
--- End of stack trace from previous location where exception was thrown ---
at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw()
at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task)
at System.Runtime.CompilerServices.TaskAwaiter`1.GetResult()
at Microsoft.Azure.Management.Cdn.EndpointsOperationsExtensions.Create(IEndpointsOperations operations, String resourceGroupName, String profileName, String endpointName, Endpoint endpoint)
at Microsoft.Azure.Commands.Cdn.Endpoint.NewAzureRmCdnEndpoint.NewEndpoint()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ConfirmAction(String processMessage, String target, Action action)
at Microsoft.Azure.Commands.Cdn.Endpoint.NewAzureRmCdnEndpoint.ExecuteCmdlet()
at Microsoft.WindowsAzure.Commands.Utilities.Common.AzurePSCmdlet.ProcessRecord()
Exception : Microsoft.Azure.Management.Cdn.Models.ErrorResponseException
InvocationInfo : {New-AzCdnEndpoint}
Line : New-AzCdnEndpoint `
```
| non_build | unable to save cdn rule conditions description i m attempting to migrate from azurerm to az in azurerm i was using the following likely overly verbose code to create a cdn rule condition specifically for adding a custom caching rule to bypass the default cache policy for the hypothetical file foo bar powershell rule new object microsoft azure commands cdn models endpoint psdeliveryrule ruleconditions new object system collections generic list rulecondition new object microsoft azure commands cdn models endpoint psdeliveryruleurlpathcondition ruleconditionparameters new object microsoft azure commands cdn models endpoint psurlpathconditionparameters ruleconditionparameters path foo bar ruleconditionparameters matchtype wildcard rulecondition parameters ruleconditionparameters ruleconditions add rulecondition rule conditions ruleconditions however in powershell az even with aliases neither psdeliveryruleurlpathcondition or psurlpathconditionparameters objects seem to exist i ve tried refactoring this code a couple of ways including powershell rulecondition new object microsoft azure commands cdn models endpoint psdeliveryrulecondition rulecondition matchvariable urlpath matchvalues new object system collections generic list matchvalues add foo bar rulecondition matchvalue matchvalues running in debug this sends the json payload snippet json conditions name urlpath parameters matchvalues foo bar odata type microsoft azure cdn models deliveryruleurlpathconditionparameters which returns json error code invalidresource message the property matchvalues does not exist on type microsoft azure cdn models deliveryruleurlpathconditionparameters make sure to only use property names that are defined by the type i ve also tried powershell rulecondition new azcdndeliveryrulecondition matchvariable urlpath operator any matchvalue foo bar which sends the json payload snippet json conditions name urlpath parameters operator any negatecondition false matchvalues foo bar odata type microsoft azure cdn models deliveryruleurlpathconditionparameters which returns json error code invalidresource message the property operator does not exist on type microsoft azure cdn models deliveryruleurlpathconditionparameters make sure to only use property names that are defined by the type i m able to create these caching rules just fine in the azure portal following the instructions and inspecting the json object snippet looks like json conditions name urlpath parameters odata type microsoft azure cdn models deliveryruleurlpathconditionparameters matchtype wildcard path foo bar other notes i tried installing previous versions of the az cdn module and with enable azurermalias my original old code had been working up to and including version my original code broke with the release of version where the types i was using appear to have been removed looking at the release notes this appears to be with the introduction of rulesengine the documentation states that the rules engine is for premium verizon only however i m only attempting to use standard caching rules which should be available in my tier standard akamai steps to reproduce described above environment data ps psversiontable name value psversion psedition desktop pscompatibleversions buildversion clrversion wsmanstackversion psremotingprotocolversion serializationversion module versions powershell moduletype version name exportedcommands script az accounts disable azdatacollection disable azcontextautosave enable azdatacollection enable azcontextautosave script az advisor get azadvisorrecommendation enable azadvisorrecommendation disable azadvisorrecommendation get azadvisorconfiguration script az aks get azaks new azaks remove azaks import azakscredential script az analysisservices resume azanalysisservicesserver suspend azanalysisservicesserver get azanalysisservicesserver remove azanalysisservicesserver script az apimanagement add azapimanagementapitoproduct add azapimanagementproducttogroup add azapimanagementregion add azapimanagementusertogroup script az applicationinsights get azapplicationinsights new azapplicationinsights remove azapplicationinsights set azapplicationinsightspricingplan script az automation get azautomationhybridworkergroup remove azautomationhybridworkergroup get azautomationjoboutputrecord import azautomationdscnodeconfiguration script az batch remove azbatchaccount get azbatchaccount get azbatchaccountkey new azbatchaccount script az billing get azbillinginvoice get azbillingperiod get azenrollmentaccount get azconsumptionbudget script az cdn get azcdnprofile get azcdnprofilessourl new azcdnprofile remove azcdnprofile script az cognitiveservices get azcognitiveservicesaccount get azcognitiveservicesaccountkey get azcognitiveservicesaccountsku get azcognitiveservicesaccounttype script az compute remove azavailabilityset get azavailabilityset new azavailabilityset update azavailabilityset script az containerinstance new azcontainergroup get azcontainergroup remove azcontainergroup get azcontainerinstancelog script az containerregistry new azcontainerregistry get azcontainerregistry update azcontainerregistry remove azcontainerregistry script az datafactory set update get remove script az datalakeanalytics get azdatalakeanalyticsdatasource new azdatalakeanalyticscatalogcredential remove azdatalakeanalyticscatalogcredential set azdatalakeanalyticscatalogcredential script az datalakestore get azdatalakestoretrustedidprovider remove azdatalakestoretrustedidprovider remove azdatalakestorefirewallrule set azdatalakestoretrustedidprovider script az deploymentmanager get azdeploymentmanagerartifactsource new azdeploymentmanagerartifactsource set azdeploymentmanagerartifactsource remove azdeploymentmanagerartifactsource script az devtestlabs get azdtlallowedvmsizespolicy get azdtlautoshutdownpolicy get azdtlautostartpolicy get azdtlvmsperlabpolicy script az dns get azdnsrecordset new azdnsrecordconfig remove azdnsrecordset set azdnsrecordset script az eventgrid new azeventgridtopic get azeventgridtopic set azeventgridtopic new azeventgridtopickey script az eventhub new azeventhubnamespace get azeventhubnamespace set azeventhubnamespace remove azeventhubnamespace script az frontdoor new azfrontdoor get azfrontdoor set azfrontdoor remove azfrontdoor script az hdinsight get azhdinsightjob new azhdinsightsqoopjobdefinition wait azhdinsightjob new azhdinsightstreamingmapreducejobdefinition script az healthcareapis new azhealthcareapisservice remove azhealthcareapisservice set azhealthcareapisservice get azhealthcareapisservice script az iothub add aziothubkey get aziothubeventhubconsumergroup get aziothubconnectionstring get aziothubjob script az keyvault add azkeyvaultcertificate update azkeyvaultcertificate stop azkeyvaultcertificateoperation get azkeyvaultcertificateoperation script az logicapp get azintegrationaccountagreement get azintegrationaccountassembly get azintegrationaccountbatchconfiguration get azintegrationaccountcallbackurl script az machinelearning move azmlcommitmentassociation get azmlcommitmentassociation get azmlcommitmentplanusagehistory remove azmlcommitmentplan script az managedservices get azmanagedservicesassignment new azmanagedservicesassignment remove azmanagedservicesassignment get azmanagedservicesdefinition script az marketplaceordering get azmarketplaceterms set azmarketplaceterms script az media sync azmediaservicestoragekey set azmediaservicekey get azmediaservicekey get azmediaservicenameavailability script az monitor get azmetricdefinition get azmetric remove azlogprofile get azlogprofile script az network add azapplicationgatewayauthenticationcertificate get azapplicationgatewayauthenticationcertificate new azapplicationgatewayauthenticationcertificate remove azapplicationgatewayauthenticationc script az notificationhubs get aznotificationhub get aznotificationhubauthorizationrule get aznotificationhublistkey get aznotificationhubpnscredential script az operationalinsights new azoperationalinsightsazureactivitylogdatasource new azoperationalinsightscustomlogdatasource disable azoperationalinsightslinuxcustomlogcollection disable azoperationalinsightsiislogcollec script az policyinsights get azpolicyevent get azpolicystate get azpolicystatesummary get azpolicyremediation script az powerbiembedded remove azpowerbiworkspacecollection get azpowerbiworkspacecollection get azpowerbiworkspacecollectionaccesskey get azpowerbiworkspace script az privatedns get azprivatednszone remove azprivatednszone set azprivatednszone new azprivatednszone script az recoveryservices get azrecoveryservicesbackupproperty get azrecoveryservicesvault get azrecoveryservicesvaultsettingsfile new azrecoveryservicesvault script az rediscache remove azrediscachepatchschedule new azrediscachescheduleentry get azrediscachepatchschedule new azrediscachepatchschedule script az relay new azrelaynamespace get azrelaynamespace set azrelaynamespace remove azrelaynamespace script az resources get azprovideroperation remove azroleassignment get azroleassignment new azroleassignment script az servicebus new azservicebusnamespace get azservicebusnamespace set azservicebusnamespace remove azservicebusnamespace script az servicefabric add azservicefabricclientcertificate add azservicefabricclustercertificate add azservicefabricnode add azservicefabricnodetype script az signalr new azsignalr get azsignalr get azsignalrkey new azsignalrkey script az sql get azsqldatabasetransparentdataencryption get azsqldatabasetransparentdataencryptionactivity set azsqldatabasetransparentdataencryption get azsqldatabaseupgradehint script az storage get azstorageaccount get azstorageaccountkey new azstorageaccount new azstorageaccountkey script az storagesync invoke azstoragesynccompatibilitycheck new azstoragesyncservice get azstoragesyncservice remove azstoragesyncservice script az streamanalytics get azstreamanalyticsfunction get azstreamanalyticsdefaultfunctiondefinition new azstreamanalyticsfunction remove azstreamanalyticsfunction script az trafficmanager add aztrafficmanagercustomheadertoendpoint remove aztrafficmanagercustomheaderfromendpoint add aztrafficmanagercustomheadertoprofile remove aztrafficmanagercustomheaderfromprofile script az websites get azappserviceplan set azappserviceplan new azappserviceplan remove azappserviceplan debug output i ve posted the relevant snippets above error output message operation returned an invalid status code badrequest stacktrace at microsoft azure management cdn endpointsoperations d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices configuredtaskawaitable configuredtaskawaiter getresult at microsoft azure management cdn endpointsoperations d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices configuredtaskawaitable configuredtaskawaiter getresult at microsoft azure management cdn endpointsoperationsextensions d movenext end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices taskawaiter getresult at microsoft azure management cdn endpointsoperationsextensions create iendpointsoperations operations string resourcegroupname string profilename string endpointname endpoint endpoint at microsoft azure commands cdn endpoint newazurermcdnendpoint newendpoint at microsoft windowsazure commands utilities common azurepscmdlet confirmaction string processmessage string target action action at microsoft azure commands cdn endpoint newazurermcdnendpoint executecmdlet at microsoft windowsazure commands utilities common azurepscmdlet processrecord exception microsoft azure management cdn models errorresponseexception invocationinfo new azcdnendpoint line new azcdnendpoint | 0 |
810,587 | 30,249,261,319 | IssuesEvent | 2023-07-06 19:04:45 | LOG680-gr02-eq11/oxygen-cs-gr02-eq11 | https://api.github.com/repos/LOG680-gr02-eq11/oxygen-cs-gr02-eq11 | closed | Pipeline sur les repos | Priority: High feature | Pour vous assurer que chaque nouvelle version des applications sont fonctionnelles, vous devez mettre
en place un pipeline d’intégration continue à l’aide de la technologie de votre choix | 1.0 | Pipeline sur les repos - Pour vous assurer que chaque nouvelle version des applications sont fonctionnelles, vous devez mettre
en place un pipeline d’intégration continue à l’aide de la technologie de votre choix | non_build | pipeline sur les repos pour vous assurer que chaque nouvelle version des applications sont fonctionnelles vous devez mettre en place un pipeline d’intégration continue à l’aide de la technologie de votre choix | 0 |
76,792 | 21,570,441,851 | IssuesEvent | 2022-05-02 07:29:09 | Polymer/tools | https://api.github.com/repos/Polymer/tools | closed | Possible to create multiple HTMLSplitter/HTMLRejoiner instances | Package: build wontfix | See discussion here: https://github.com/Polymer/polymer-build/pull/118#discussion_r99653781
We want to support options for splitting and rejoining scripts/styles in the build process, but we may want to specifically disallow multiple splitters/rejoiners from being created if this becomes a 👟 🔫 for users. | 1.0 | Possible to create multiple HTMLSplitter/HTMLRejoiner instances - See discussion here: https://github.com/Polymer/polymer-build/pull/118#discussion_r99653781
We want to support options for splitting and rejoining scripts/styles in the build process, but we may want to specifically disallow multiple splitters/rejoiners from being created if this becomes a 👟 🔫 for users. | build | possible to create multiple htmlsplitter htmlrejoiner instances see discussion here we want to support options for splitting and rejoining scripts styles in the build process but we may want to specifically disallow multiple splitters rejoiners from being created if this becomes a 👟 🔫 for users | 1 |
17,108 | 6,359,767,271 | IssuesEvent | 2017-07-31 08:18:40 | spring-projects/spring-boot | https://api.github.com/repos/spring-projects/spring-boot | closed | Ensure that Boot applications built with Maven use the `-parameters` compiler flag by default | priority: normal theme: build type: chore type: enhancement | In #9297 we're considering adding jackson modules to our web starter. One of those modules, the [parameter-names](https://github.com/FasterXML/jackson-modules-java8/tree/master/parameter-names) one, is only useful if the JDK compiler has the `-parameters` option set.
This issue is about checking the various places where this option should be set. | 1.0 | Ensure that Boot applications built with Maven use the `-parameters` compiler flag by default - In #9297 we're considering adding jackson modules to our web starter. One of those modules, the [parameter-names](https://github.com/FasterXML/jackson-modules-java8/tree/master/parameter-names) one, is only useful if the JDK compiler has the `-parameters` option set.
This issue is about checking the various places where this option should be set. | build | ensure that boot applications built with maven use the parameters compiler flag by default in we re considering adding jackson modules to our web starter one of those modules the one is only useful if the jdk compiler has the parameters option set this issue is about checking the various places where this option should be set | 1 |
193,895 | 15,391,564,725 | IssuesEvent | 2021-03-03 14:42:20 | streetmix/streetmix | https://api.github.com/repos/streetmix/streetmix | closed | Rename "master" branch to "main" | documentation in progress | As of this issue's opening, we've already renamed the repository's "master" branch to "main". While the process itself is painless in GitHub, we now have to make sure that third-party services and documentation are pointing to the correct branch location. This issue is tracking any remaining todos.
### Updating local clones
If you have a local clone, you can update it by running:
```
git branch -m master main
git fetch origin
git branch -u origin/main main
```
### Next steps
- [x] update all third-party services, such as Heroku, which expect to pull code from the "master" branch - these services will need to be updated
- [x] Heroku
- [x] ReadTheDocs
- [x] update any code which specifically reference the "master" branch
- [x] update documentation (#2242)
| 1.0 | Rename "master" branch to "main" - As of this issue's opening, we've already renamed the repository's "master" branch to "main". While the process itself is painless in GitHub, we now have to make sure that third-party services and documentation are pointing to the correct branch location. This issue is tracking any remaining todos.
### Updating local clones
If you have a local clone, you can update it by running:
```
git branch -m master main
git fetch origin
git branch -u origin/main main
```
### Next steps
- [x] update all third-party services, such as Heroku, which expect to pull code from the "master" branch - these services will need to be updated
- [x] Heroku
- [x] ReadTheDocs
- [x] update any code which specifically reference the "master" branch
- [x] update documentation (#2242)
| non_build | rename master branch to main as of this issue s opening we ve already renamed the repository s master branch to main while the process itself is painless in github we now have to make sure that third party services and documentation are pointing to the correct branch location this issue is tracking any remaining todos updating local clones if you have a local clone you can update it by running git branch m master main git fetch origin git branch u origin main main next steps update all third party services such as heroku which expect to pull code from the master branch these services will need to be updated heroku readthedocs update any code which specifically reference the master branch update documentation | 0 |
47,257 | 11,992,050,230 | IssuesEvent | 2020-04-08 09:26:06 | syncthing/syncthing | https://api.github.com/repos/syncthing/syncthing | closed | Remove macOS 32 bit builds | build | As of currently our minimum macOS requirement is 10.10, which does not support any 32 bit hardware. The last macOS to support 32 bit hardware was 10.6, and our binaries haven't worked there for years. Hence there is no point in the 32 bit build.
(Issue for visibility and to document the reasoning & decision, not that it's specifically actionable.)
(ref #6516) | 1.0 | Remove macOS 32 bit builds - As of currently our minimum macOS requirement is 10.10, which does not support any 32 bit hardware. The last macOS to support 32 bit hardware was 10.6, and our binaries haven't worked there for years. Hence there is no point in the 32 bit build.
(Issue for visibility and to document the reasoning & decision, not that it's specifically actionable.)
(ref #6516) | build | remove macos bit builds as of currently our minimum macos requirement is which does not support any bit hardware the last macos to support bit hardware was and our binaries haven t worked there for years hence there is no point in the bit build issue for visibility and to document the reasoning decision not that it s specifically actionable ref | 1 |
8,720 | 3,783,478,151 | IssuesEvent | 2016-03-19 05:19:02 | ttimsmith/theboldreport.net | https://api.github.com/repos/ttimsmith/theboldreport.net | opened | Fix "LeadingZero" issue in _assets/stylesheets/_layout/_maincontent.scss | Code Climate | `0.9` should be written without a leading zero as `.9`
https://codeclimate.com/github/ttimsmith/theboldreport.net/_assets/stylesheets/_layout/_maincontent.scss#issue_56ecd62de693f0000102cc74 | 1.0 | Fix "LeadingZero" issue in _assets/stylesheets/_layout/_maincontent.scss - `0.9` should be written without a leading zero as `.9`
https://codeclimate.com/github/ttimsmith/theboldreport.net/_assets/stylesheets/_layout/_maincontent.scss#issue_56ecd62de693f0000102cc74 | non_build | fix leadingzero issue in assets stylesheets layout maincontent scss should be written without a leading zero as | 0 |
8,694 | 4,309,813,687 | IssuesEvent | 2016-07-21 17:11:18 | golang/go | https://api.github.com/repos/golang/go | closed | x/build/cmd/coordinator: watcher's git mirroring is slow/flaky | Builders | The "net" subrepo has stopped mirroring to github.
Note: https://github.com/golang/net/commits/master (stuck at Jul 7, 2016, f841c3)
Versus: https://go.googlesource.com/net (3 new commits since then, with e90d6d0 currently at top)
The watcher says:
http://farmer.golang.org/debug/watcher
```
! [remote rejected] ---40hexomitted--- -> ---40hexomitted--- (pre-receive hook declined)
error: failed to push some refs to 'https://gopherbot:---40hexomitted---@github.com/golang/net'
remote: error: GH002: Sorry, branch or tag names consisting of 40 hex characters are not allowed.
remote: error: Invalid branch or tag name "---40hexomitted---"
```
So Github doesn't accept some ref that Gerrit has?
/cc @spearce @adg @broady @quentinmit
Related: #11811 | 1.0 | x/build/cmd/coordinator: watcher's git mirroring is slow/flaky - The "net" subrepo has stopped mirroring to github.
Note: https://github.com/golang/net/commits/master (stuck at Jul 7, 2016, f841c3)
Versus: https://go.googlesource.com/net (3 new commits since then, with e90d6d0 currently at top)
The watcher says:
http://farmer.golang.org/debug/watcher
```
! [remote rejected] ---40hexomitted--- -> ---40hexomitted--- (pre-receive hook declined)
error: failed to push some refs to 'https://gopherbot:---40hexomitted---@github.com/golang/net'
remote: error: GH002: Sorry, branch or tag names consisting of 40 hex characters are not allowed.
remote: error: Invalid branch or tag name "---40hexomitted---"
```
So Github doesn't accept some ref that Gerrit has?
/cc @spearce @adg @broady @quentinmit
Related: #11811 | build | x build cmd coordinator watcher s git mirroring is slow flaky the net subrepo has stopped mirroring to github note stuck at jul versus new commits since then with currently at top the watcher says pre receive hook declined error failed to push some refs to remote error sorry branch or tag names consisting of hex characters are not allowed remote error invalid branch or tag name so github doesn t accept some ref that gerrit has cc spearce adg broady quentinmit related | 1 |
170,888 | 13,207,738,804 | IssuesEvent | 2020-08-15 00:22:52 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | opened | [Desktop] `Reset Brave Rewards data` is not translated in brave://settings/clearBrowserData | OS/Desktop QA/Test-Plan-Specified QA/Yes l10n | Follow up to https://github.com/brave/brave-browser/issues/926
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Open `brave://settings/clearBrowserData` in Polish locale
## Actual result:
<!--Please add screenshots if needed-->
`Reset Brave Rewards data` is not translated

## Expected result:
`Reset Brave Rewards data` is translated
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily reproduced
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.13.71 Chromium: 84.0.4147.125 (Official Build) dev (64-bit)
-- | --
Revision | d0784639447f2e10d32ebaf9861092b20cfde286-refs/branch-heads/4147@{#1059}
OS | Ubuntu 18.04 LTS
Brave | 1.14.42 Chromium: 85.0.4183.59 (Official Build) nightly (64-bit)
-- | --
Revision | 7cc9a5ee672625c38427a839cb1b15efd3128084-refs/branch-heads/4183@{#1216}
OS | Ubuntu 18.04 LTS
cc @brave/legacy_qa @mkarolin @NejcZdovc @rebron | 1.0 | [Desktop] `Reset Brave Rewards data` is not translated in brave://settings/clearBrowserData - Follow up to https://github.com/brave/brave-browser/issues/926
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Open `brave://settings/clearBrowserData` in Polish locale
## Actual result:
<!--Please add screenshots if needed-->
`Reset Brave Rewards data` is not translated

## Expected result:
`Reset Brave Rewards data` is translated
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily reproduced
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.13.71 Chromium: 84.0.4147.125 (Official Build) dev (64-bit)
-- | --
Revision | d0784639447f2e10d32ebaf9861092b20cfde286-refs/branch-heads/4147@{#1059}
OS | Ubuntu 18.04 LTS
Brave | 1.14.42 Chromium: 85.0.4183.59 (Official Build) nightly (64-bit)
-- | --
Revision | 7cc9a5ee672625c38427a839cb1b15efd3128084-refs/branch-heads/4183@{#1216}
OS | Ubuntu 18.04 LTS
cc @brave/legacy_qa @mkarolin @NejcZdovc @rebron | non_build | reset brave rewards data is not translated in brave settings clearbrowserdata follow up to steps to reproduce open brave settings clearbrowserdata in polish locale actual result reset brave rewards data is not translated expected result reset brave rewards data is translated reproduces how often easily reproduced brave version brave version info brave chromium official build dev bit revision refs branch heads os ubuntu lts brave chromium official build nightly bit revision refs branch heads os ubuntu lts cc brave legacy qa mkarolin nejczdovc rebron | 0 |
659,550 | 21,932,739,892 | IssuesEvent | 2022-05-23 11:13:23 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Saas application with TOTP as second authenticator fail | Priority/Highest Severity/Blocker bug Affected-6.0.0 QA-Reported | **Describe the issue:**
Saas application with TOTP as second authenticator fail.
console error:
`ERROR {org.wso2.carbon.identity.application.authentication.framework.handler.step.impl.DefaultStepHandler} - Authentication failed exception! Error when checking TOTP enabled for the user : achinij@carbon.super`
https://user-images.githubusercontent.com/39077751/168703327-e05c23a1-5d7a-4a92-b723-622e73346af8.mov
**How to reproduce:**
1. Login to console
2. Create tenant (admin : achinij, domain: wso2.com)
3. Create application and make it saas enable
<img width="987" alt="image" src="https://user-images.githubusercontent.com/39077751/168703668-a5ec03fe-c716-45bf-b422-95bddc2c41c6.png">
4. Add 1st factor as basic and second factor as TOTP.
5. Authenticate to the application with tenant user. achinij@wso2.com
**Expected behavior:**
Successfully authenticate with basic as first factor and TOTP as second factor
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: 5.12.0 Beta2
- OS: MAC
- Database: MSSQL
- Userstore: JDBC
---
### Optional Fields
**Related issues:**
<!-- Any related issues from this/other repositories-->
**Suggested labels:**
<!-- Only to be used by non-members -->
| 1.0 | Saas application with TOTP as second authenticator fail - **Describe the issue:**
Saas application with TOTP as second authenticator fail.
console error:
`ERROR {org.wso2.carbon.identity.application.authentication.framework.handler.step.impl.DefaultStepHandler} - Authentication failed exception! Error when checking TOTP enabled for the user : achinij@carbon.super`
https://user-images.githubusercontent.com/39077751/168703327-e05c23a1-5d7a-4a92-b723-622e73346af8.mov
**How to reproduce:**
1. Login to console
2. Create tenant (admin : achinij, domain: wso2.com)
3. Create application and make it saas enable
<img width="987" alt="image" src="https://user-images.githubusercontent.com/39077751/168703668-a5ec03fe-c716-45bf-b422-95bddc2c41c6.png">
4. Add 1st factor as basic and second factor as TOTP.
5. Authenticate to the application with tenant user. achinij@wso2.com
**Expected behavior:**
Successfully authenticate with basic as first factor and TOTP as second factor
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: 5.12.0 Beta2
- OS: MAC
- Database: MSSQL
- Userstore: JDBC
---
### Optional Fields
**Related issues:**
<!-- Any related issues from this/other repositories-->
**Suggested labels:**
<!-- Only to be used by non-members -->
| non_build | saas application with totp as second authenticator fail describe the issue saas application with totp as second authenticator fail console error error org carbon identity application authentication framework handler step impl defaultstephandler authentication failed exception error when checking totp enabled for the user achinij carbon super how to reproduce login to console create tenant admin achinij domain com create application and make it saas enable img width alt image src add factor as basic and second factor as totp authenticate to the application with tenant user achinij com expected behavior successfully authenticate with basic as first factor and totp as second factor environment information please complete the following information remove any unnecessary fields product version os mac database mssql userstore jdbc optional fields related issues suggested labels | 0 |
86,881 | 24,980,346,384 | IssuesEvent | 2022-11-02 11:10:30 | JuliaLang/julia | https://api.github.com/repos/JuliaLang/julia | closed | Full source build failing | build | Recent master is failing the full source build: https://buildkite.com/julialang/julia-master-scheduled/builds/141#01842c63-d05a-41c9-aea4-237e2d402f83
History of full source builds: https://buildkite.com/julialang/julia-master-scheduled
cc @DilumAluthge | 1.0 | Full source build failing - Recent master is failing the full source build: https://buildkite.com/julialang/julia-master-scheduled/builds/141#01842c63-d05a-41c9-aea4-237e2d402f83
History of full source builds: https://buildkite.com/julialang/julia-master-scheduled
cc @DilumAluthge | build | full source build failing recent master is failing the full source build history of full source builds cc dilumaluthge | 1 |
91,509 | 26,407,253,332 | IssuesEvent | 2023-01-13 09:09:57 | dotnet/command-line-api | https://api.github.com/repos/dotnet/command-line-api | closed | SourceLink needs updates for API changes | bug Area-Build & Infrastructure | The [Arcade update flow into installer](https://github.com/dotnet/installer/pull/15130) came up with some errors I think we've seen before:
```
/vmr/src/sourcelink/artifacts/source-build/self/src/src/dotnet-sourcelink/Program.cs(91,17): error CS0144: (NETCORE_ENGINEERING_TELEMETRY=Build) Cannot create an instance of the abstract type or interface 'Option'
/vmr/src/sourcelink/artifacts/source-build/self/src/src/dotnet-sourcelink/Program.cs(93,21): error CS0122: (NETCORE_ENGINEERING_TELEMETRY=Build) 'Option.Argument' is inaccessible due to its protection level
/vmr/src/sourcelink/artifacts/source-build/self/src/src/dotnet-sourcelink/Program.cs(95,17): error CS0144: (NETCORE_ENGINEERING_TELEMETRY=Build) Cannot create an instance of the abstract type or interface 'Option'
```
I think you've helped us fix these downstream errors before - could you help out again? Thanks! | 1.0 | SourceLink needs updates for API changes - The [Arcade update flow into installer](https://github.com/dotnet/installer/pull/15130) came up with some errors I think we've seen before:
```
/vmr/src/sourcelink/artifacts/source-build/self/src/src/dotnet-sourcelink/Program.cs(91,17): error CS0144: (NETCORE_ENGINEERING_TELEMETRY=Build) Cannot create an instance of the abstract type or interface 'Option'
/vmr/src/sourcelink/artifacts/source-build/self/src/src/dotnet-sourcelink/Program.cs(93,21): error CS0122: (NETCORE_ENGINEERING_TELEMETRY=Build) 'Option.Argument' is inaccessible due to its protection level
/vmr/src/sourcelink/artifacts/source-build/self/src/src/dotnet-sourcelink/Program.cs(95,17): error CS0144: (NETCORE_ENGINEERING_TELEMETRY=Build) Cannot create an instance of the abstract type or interface 'Option'
```
I think you've helped us fix these downstream errors before - could you help out again? Thanks! | build | sourcelink needs updates for api changes the came up with some errors i think we ve seen before vmr src sourcelink artifacts source build self src src dotnet sourcelink program cs error netcore engineering telemetry build cannot create an instance of the abstract type or interface option vmr src sourcelink artifacts source build self src src dotnet sourcelink program cs error netcore engineering telemetry build option argument is inaccessible due to its protection level vmr src sourcelink artifacts source build self src src dotnet sourcelink program cs error netcore engineering telemetry build cannot create an instance of the abstract type or interface option i think you ve helped us fix these downstream errors before could you help out again thanks | 1 |
71,674 | 18,842,919,739 | IssuesEvent | 2021-11-11 11:44:21 | o3de/o3de | https://api.github.com/repos/o3de/o3de | opened | 3rdParty dlls are not copied to output bin folder when they are deleted from output folder | kind/bug needs-triage sig/build | **Describe the bug**
If a user deletes 3rdParty dlls from output bin folder and builds again, it finishes to build successfully without realizing it has to copy the dlls again to the output folder. This leads to crashes when opening the editor because dlls are missing, even though the project was built successfully.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Build AutomatedTesting project on profile
2. Go to the `bin\profile` folder of your build directory
3. Delete a third party dll, for example PhysX_64.dll
4. Build AutomatedTesting project again
**Expected behavior**
The third party dll is copied to the output bin folder.
**Actual behavior**
Build finishes successfully but it didn't copy the third party dll to the output folder.
**Device**
Tested on windows, but it might occur in other platforms too.
| 1.0 | 3rdParty dlls are not copied to output bin folder when they are deleted from output folder - **Describe the bug**
If a user deletes 3rdParty dlls from output bin folder and builds again, it finishes to build successfully without realizing it has to copy the dlls again to the output folder. This leads to crashes when opening the editor because dlls are missing, even though the project was built successfully.
**Steps to reproduce**
Steps to reproduce the behavior:
1. Build AutomatedTesting project on profile
2. Go to the `bin\profile` folder of your build directory
3. Delete a third party dll, for example PhysX_64.dll
4. Build AutomatedTesting project again
**Expected behavior**
The third party dll is copied to the output bin folder.
**Actual behavior**
Build finishes successfully but it didn't copy the third party dll to the output folder.
**Device**
Tested on windows, but it might occur in other platforms too.
| build | dlls are not copied to output bin folder when they are deleted from output folder describe the bug if a user deletes dlls from output bin folder and builds again it finishes to build successfully without realizing it has to copy the dlls again to the output folder this leads to crashes when opening the editor because dlls are missing even though the project was built successfully steps to reproduce steps to reproduce the behavior build automatedtesting project on profile go to the bin profile folder of your build directory delete a third party dll for example physx dll build automatedtesting project again expected behavior the third party dll is copied to the output bin folder actual behavior build finishes successfully but it didn t copy the third party dll to the output folder device tested on windows but it might occur in other platforms too | 1 |
38,394 | 10,192,322,494 | IssuesEvent | 2019-08-12 10:47:33 | Haivision/srt | https://api.github.com/repos/Haivision/srt | closed | build failure stransmit with MinGW | Status: On Hold [build] | This might not be a problem of SRT, but MinGW seems not to provide `std::this_thread`.
However, it should be fixed or changed to use alternative to support MinGW (W32 cross compilation on Linux)
```
[ 85%] Building CXX object CMakeFiles/stransmit.dir/apps/stransmit.cpp.obj
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp: In member function ‘void BandwidthGuard::Checkpoint(size_t, size_t)’:
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp:366:14: error: ‘std::this_thread’ has not been declared
std::this_thread::sleep_for(expdur-dur);
^
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp: In member function ‘virtual bytevector SrtSource::Read(size_t)’:
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp:1094:17: error: ‘this_thread’ has not been declared
this_thread::sleep_for(chrono::milliseconds(10));
^
CMakeFiles/stransmit.dir/build.make:63: recipe for target 'CMakeFiles/stransmit.dir/apps/stransmit.cpp.obj' failed
make[2]: *** [CMakeFiles/stransmit.dir/apps/stransmit.cpp.obj] Error 1
CMakeFiles/Makefile2:104: recipe for target 'CMakeFiles/stransmit.dir/all' failed
make[1]: *** [CMakeFiles/stransmit.dir/all] Error 2
Makefile:127: recipe for target 'all' failed
make: *** [all] Error 2
``` | 1.0 | build failure stransmit with MinGW - This might not be a problem of SRT, but MinGW seems not to provide `std::this_thread`.
However, it should be fixed or changed to use alternative to support MinGW (W32 cross compilation on Linux)
```
[ 85%] Building CXX object CMakeFiles/stransmit.dir/apps/stransmit.cpp.obj
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp: In member function ‘void BandwidthGuard::Checkpoint(size_t, size_t)’:
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp:366:14: error: ‘std::this_thread’ has not been declared
std::this_thread::sleep_for(expdur-dur);
^
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp: In member function ‘virtual bytevector SrtSource::Read(size_t)’:
/home/jaykim/ssd/git/vlc/contrib/win32/srt/apps/stransmit.cpp:1094:17: error: ‘this_thread’ has not been declared
this_thread::sleep_for(chrono::milliseconds(10));
^
CMakeFiles/stransmit.dir/build.make:63: recipe for target 'CMakeFiles/stransmit.dir/apps/stransmit.cpp.obj' failed
make[2]: *** [CMakeFiles/stransmit.dir/apps/stransmit.cpp.obj] Error 1
CMakeFiles/Makefile2:104: recipe for target 'CMakeFiles/stransmit.dir/all' failed
make[1]: *** [CMakeFiles/stransmit.dir/all] Error 2
Makefile:127: recipe for target 'all' failed
make: *** [all] Error 2
``` | build | build failure stransmit with mingw this might not be a problem of srt but mingw seems not to provide std this thread however it should be fixed or changed to use alternative to support mingw cross compilation on linux building cxx object cmakefiles stransmit dir apps stransmit cpp obj home jaykim ssd git vlc contrib srt apps stransmit cpp in member function ‘void bandwidthguard checkpoint size t size t ’ home jaykim ssd git vlc contrib srt apps stransmit cpp error ‘std this thread’ has not been declared std this thread sleep for expdur dur home jaykim ssd git vlc contrib srt apps stransmit cpp in member function ‘virtual bytevector srtsource read size t ’ home jaykim ssd git vlc contrib srt apps stransmit cpp error ‘this thread’ has not been declared this thread sleep for chrono milliseconds cmakefiles stransmit dir build make recipe for target cmakefiles stransmit dir apps stransmit cpp obj failed make error cmakefiles recipe for target cmakefiles stransmit dir all failed make error makefile recipe for target all failed make error | 1 |
57,830 | 14,228,436,681 | IssuesEvent | 2020-11-18 03:48:03 | dxx-rebirth/dxx-rebirth | https://api.github.com/repos/dxx-rebirth/dxx-rebirth | closed | Alternative to GLU Dependency | build-failure | ### Environment
https://github.com/dxx-rebirth/dxx-rebirth/commit/93d3793b306cebb50be1dc86f974204d5d877ac0
#### Operating System Environment
* [ ] Arch
#### CPU environment
* [ ] x86\_64 (64-bit Intel/AMD)
### Description
The only program that needs "glu" installed I use personally, is dxx-rebirth, even when building the sdl2 port. Why is "glu" a hard depencency? Can an alternative to "glu" be used?
### Steps to Reproduce
Try to compile dxx-rebirth with "glu" uninstalled, and the compile error will occur.
```
similar/arch/ogl/gr.cpp:70:10: fatal error: GL/glu.h: No such file or directory
70 | #include <GL/glu.h>
| ^~~~~~~~~~
compilation terminated.
scons: *** [build/similar/arch/ogl/.d1x-rebirth.gr.o] Error 1
scons: building terminated because of errors.
Failed target count: total=1; targets with enable_build_failure_summary=1: 1
Failed node list:
build/similar/arch/ogl/.d1x-rebirth.gr.o
Failed command list:
x86_64-pc-linux-gnu-g++ -c -g -O2 -ftabstop=4 -Wall -Werror=extra -Werror=format=2 -Werror=missing-braces -Werror=missing-include-dirs -Werror=uninitialized -Werror=undef -Werror=pointer-arith -Werror=cast-qual -Werror=missing-declarations -Werror=redundant-decls -Werror=vla -pthread -funsigned-char -march=x86-64 -mtune=generic -O2 -pipe -fno-plt -flto -std=gnu++14 -Werror=unused -Werror=useless-cast -Wno-implicit-fallthrough -fvisibility=hidden -Wduplicated-branches -Wduplicated-cond -Wsuggest-attribute=noreturn -Wlogical-op -Wold-style-cast -Wno-sign-compare -D_FORTIFY_SOURCE=2 -DPHYSFS_DEPRECATED= -DNDEBUG -DRELEASE -D_REENTRANT -DDXX_BUILD_DESCENT_I -D__STDC_FORMAT_MACROS -DDXX_USE_SHAREPATH=1 -Icommon/include -Icommon/main -I. -Ibuild -I/usr/include/libpng16 -I/usr/include/SDL2 -Id1x-rebirth/main "similar/arch/ogl/gr.cpp" -o "build/similar/arch/ogl/.d1x-rebirth.gr.o"
==> ERROR: A failure occurred in build().
Aborting...
``` | 1.0 | Alternative to GLU Dependency - ### Environment
https://github.com/dxx-rebirth/dxx-rebirth/commit/93d3793b306cebb50be1dc86f974204d5d877ac0
#### Operating System Environment
* [ ] Arch
#### CPU environment
* [ ] x86\_64 (64-bit Intel/AMD)
### Description
The only program that needs "glu" installed I use personally, is dxx-rebirth, even when building the sdl2 port. Why is "glu" a hard depencency? Can an alternative to "glu" be used?
### Steps to Reproduce
Try to compile dxx-rebirth with "glu" uninstalled, and the compile error will occur.
```
similar/arch/ogl/gr.cpp:70:10: fatal error: GL/glu.h: No such file or directory
70 | #include <GL/glu.h>
| ^~~~~~~~~~
compilation terminated.
scons: *** [build/similar/arch/ogl/.d1x-rebirth.gr.o] Error 1
scons: building terminated because of errors.
Failed target count: total=1; targets with enable_build_failure_summary=1: 1
Failed node list:
build/similar/arch/ogl/.d1x-rebirth.gr.o
Failed command list:
x86_64-pc-linux-gnu-g++ -c -g -O2 -ftabstop=4 -Wall -Werror=extra -Werror=format=2 -Werror=missing-braces -Werror=missing-include-dirs -Werror=uninitialized -Werror=undef -Werror=pointer-arith -Werror=cast-qual -Werror=missing-declarations -Werror=redundant-decls -Werror=vla -pthread -funsigned-char -march=x86-64 -mtune=generic -O2 -pipe -fno-plt -flto -std=gnu++14 -Werror=unused -Werror=useless-cast -Wno-implicit-fallthrough -fvisibility=hidden -Wduplicated-branches -Wduplicated-cond -Wsuggest-attribute=noreturn -Wlogical-op -Wold-style-cast -Wno-sign-compare -D_FORTIFY_SOURCE=2 -DPHYSFS_DEPRECATED= -DNDEBUG -DRELEASE -D_REENTRANT -DDXX_BUILD_DESCENT_I -D__STDC_FORMAT_MACROS -DDXX_USE_SHAREPATH=1 -Icommon/include -Icommon/main -I. -Ibuild -I/usr/include/libpng16 -I/usr/include/SDL2 -Id1x-rebirth/main "similar/arch/ogl/gr.cpp" -o "build/similar/arch/ogl/.d1x-rebirth.gr.o"
==> ERROR: A failure occurred in build().
Aborting...
``` | build | alternative to glu dependency environment operating system environment arch cpu environment bit intel amd description the only program that needs glu installed i use personally is dxx rebirth even when building the port why is glu a hard depencency can an alternative to glu be used steps to reproduce try to compile dxx rebirth with glu uninstalled and the compile error will occur similar arch ogl gr cpp fatal error gl glu h no such file or directory include compilation terminated scons error scons building terminated because of errors failed target count total targets with enable build failure summary failed node list build similar arch ogl rebirth gr o failed command list pc linux gnu g c g ftabstop wall werror extra werror format werror missing braces werror missing include dirs werror uninitialized werror undef werror pointer arith werror cast qual werror missing declarations werror redundant decls werror vla pthread funsigned char march mtune generic pipe fno plt flto std gnu werror unused werror useless cast wno implicit fallthrough fvisibility hidden wduplicated branches wduplicated cond wsuggest attribute noreturn wlogical op wold style cast wno sign compare d fortify source dphysfs deprecated dndebug drelease d reentrant ddxx build descent i d stdc format macros ddxx use sharepath icommon include icommon main i ibuild i usr include i usr include rebirth main similar arch ogl gr cpp o build similar arch ogl rebirth gr o error a failure occurred in build aborting | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.