Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
4,485 | 5,110,358,838 | IssuesEvent | 2017-01-05 23:56:37 | sketch-city/project-ideas | https://api.github.com/repos/sketch-city/project-ideas | closed | Regional Server Network for Regional Apps | infrastructure | Please describe your project, the problem you're solving, and why it's important. Keep it brief! Link to further reading if necessary.
Local code irrespective of device downloadable from local servers / as close to backbone as possible.
| 1.0 | Regional Server Network for Regional Apps - Please describe your project, the problem you're solving, and why it's important. Keep it brief! Link to further reading if necessary.
Local code irrespective of device downloadable from local servers / as close to backbone as possible.
| infrastructure | regional server network for regional apps please describe your project the problem you re solving and why it s important keep it brief link to further reading if necessary local code irrespective of device downloadable from local servers as close to backbone as possible | 1 |
21,754 | 14,786,431,571 | IssuesEvent | 2021-01-12 05:29:43 | pol-is/polis | https://api.github.com/repos/pol-is/polis | opened | Use GitHub Container Registry to store pre-built containers for test workflows | ⚒️ infrastructure | Right now, we're pushing nightly builds to docker hub. In theory, this makes deploying quicker for new people, since can pull instead of building. In practice, I don't think it's used much.
Having said that, we're also building containers in order to run cypress tests. If we start building them for cross-browser testing of old browsers on BrowserStack, then we'll be building the containers at least twice per commit. This expends twice as many build minutes as we need, and each docker build takes about 8 minutes.
We could instead build the containers in one workflow, and push them to GitHub Container Registry. These would be set to private, so just for internal tests. We could then pull them in the workflows that needs to spin up an instance, without rebuilding them each time (e.g. cypress tests, browserstack tests, etc).
GitHub Container Registry vs GitHub Docker Registry: https://docs.github.com/en/free-pro-team@latest/packages/guides/migrating-to-github-container-registry-for-docker-images (main thing is that GHCR has more fine-grained permissions)
Enabling: https://docs.github.com/en/free-pro-team@latest/packages/guides/enabling-improved-container-support | 1.0 | Use GitHub Container Registry to store pre-built containers for test workflows - Right now, we're pushing nightly builds to docker hub. In theory, this makes deploying quicker for new people, since can pull instead of building. In practice, I don't think it's used much.
Having said that, we're also building containers in order to run cypress tests. If we start building them for cross-browser testing of old browsers on BrowserStack, then we'll be building the containers at least twice per commit. This expends twice as many build minutes as we need, and each docker build takes about 8 minutes.
We could instead build the containers in one workflow, and push them to GitHub Container Registry. These would be set to private, so just for internal tests. We could then pull them in the workflows that needs to spin up an instance, without rebuilding them each time (e.g. cypress tests, browserstack tests, etc).
GitHub Container Registry vs GitHub Docker Registry: https://docs.github.com/en/free-pro-team@latest/packages/guides/migrating-to-github-container-registry-for-docker-images (main thing is that GHCR has more fine-grained permissions)
Enabling: https://docs.github.com/en/free-pro-team@latest/packages/guides/enabling-improved-container-support | infrastructure | use github container registry to store pre built containers for test workflows right now we re pushing nightly builds to docker hub in theory this makes deploying quicker for new people since can pull instead of building in practice i don t think it s used much having said that we re also building containers in order to run cypress tests if we start building them for cross browser testing of old browsers on browserstack then we ll be building the containers at least twice per commit this expends twice as many build minutes as we need and each docker build takes about minutes we could instead build the containers in one workflow and push them to github container registry these would be set to private so just for internal tests we could then pull them in the workflows that needs to spin up an instance without rebuilding them each time e g cypress tests browserstack tests etc github container registry vs github docker registry main thing is that ghcr has more fine grained permissions enabling | 1 |
14,160 | 10,678,268,586 | IssuesEvent | 2019-10-21 16:56:24 | aspnet/AspNetCore | https://api.github.com/repos/aspnet/AspNetCore | closed | Unable to .\restore.cmd - TagBuilderWebSite.csproj | area-infrastructure | ### Describe the bug
Cannot restore from from clean master
### To Reproduce
Steps to reproduce the behavior:
1. Using this version of ASP.NET Core '418e35c4396682aea6ee6e9833ff3bd3bfe624eb'
2. Run this code '\.restore.cmd'
3. See error
### Expected behavior
Successful restore
### Screenshots
If applicable, add screenshots to help explain your problem.
### Additional context
dotnet --info:
```
C:\dev\AspNetCore>dotnet --info
A compatible installed .NET Core SDK for global.json version [5.0.100-alpha1-014696] from [C:\dev\AspNetCore\global.json] was not found
Install the [5.0.100-alpha1-014696] .NET Core SDK or update [C:\dev\AspNetCore\global.json] with an installed .NET Core SDK:
1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk]
1.0.0 [C:\Program Files\dotnet\sdk]
1.0.4 [C:\Program Files\dotnet\sdk]
2.1.4 [C:\Program Files\dotnet\sdk]
2.1.103 [C:\Program Files\dotnet\sdk]
2.1.104 [C:\Program Files\dotnet\sdk]
2.1.201 [C:\Program Files\dotnet\sdk]
2.1.202 [C:\Program Files\dotnet\sdk]
2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk]
2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk]
2.1.300 [C:\Program Files\dotnet\sdk]
2.1.401 [C:\Program Files\dotnet\sdk]
2.1.402 [C:\Program Files\dotnet\sdk]
2.1.403 [C:\Program Files\dotnet\sdk]
2.1.502 [C:\Program Files\dotnet\sdk]
2.1.503 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009426 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009472 [C:\Program Files\dotnet\sdk]
2.1.602 [C:\Program Files\dotnet\sdk]
2.1.800 [C:\Program Files\dotnet\sdk]
2.2.101 [C:\Program Files\dotnet\sdk]
3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk]
3.0.100 [C:\Program Files\dotnet\sdk]
3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk]
Host (useful for support):
Version: 5.0.0-alpha1.19514.1
Commit: 4ace84dbf9
.NET Core SDKs installed:
1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk]
1.0.0 [C:\Program Files\dotnet\sdk]
1.0.4 [C:\Program Files\dotnet\sdk]
2.1.4 [C:\Program Files\dotnet\sdk]
2.1.103 [C:\Program Files\dotnet\sdk]
2.1.104 [C:\Program Files\dotnet\sdk]
2.1.201 [C:\Program Files\dotnet\sdk]
2.1.202 [C:\Program Files\dotnet\sdk]
2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk]
2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk]
2.1.300 [C:\Program Files\dotnet\sdk]
2.1.401 [C:\Program Files\dotnet\sdk]
2.1.402 [C:\Program Files\dotnet\sdk]
2.1.403 [C:\Program Files\dotnet\sdk]
2.1.502 [C:\Program Files\dotnet\sdk]
2.1.503 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009426 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009472 [C:\Program Files\dotnet\sdk]
2.1.602 [C:\Program Files\dotnet\sdk]
2.1.800 [C:\Program Files\dotnet\sdk]
2.2.101 [C:\Program Files\dotnet\sdk]
3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk]
3.0.100 [C:\Program Files\dotnet\sdk]
3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.0.0-preview6.19307.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.1.0-preview1.19508.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 5.0.0-dev [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 1.0.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.0.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.1.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.1.2 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.0-preview1-26216-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.0-rc1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.3-servicing-26724-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 5.0.0-alpha1.19514.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
Console output:
```
C:\dev\AspNetCore>git status
On branch master
Your branch is up to date with 'origin/master'.
nothing to commit, working tree clean
C:\dev\AspNetCore>.\restore.cmd
Building of C# project is enabled and has dependencies on NodeJS projects. Building of NodeJS projects is enabled since node is detected in C:\Program Files.
Wiederherstellung in "41,94 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade
.sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process.
Wiederherstellung in "14,54 ms" für "C:\dev\AspNetCore\eng\tools\RepoTasks\RepoTasks.csproj
" abgeschlossen.
RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\netcoreapp5.0\RepoTasks.dll
RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\net472\RepoTasks.dll
Wiederherstellung in "41,11 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade
.sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process.
C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft.
Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN
etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge
laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T
agHelpersWebSite\TagHelpersWebSite.csproj]
Fehler beim Buildvorgang.
C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft.
Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN
etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge
laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T
agHelpersWebSite\TagHelpersWebSite.csproj]
0 Warnung(en)
1 Fehler
Verstrichene Zeit 00:00:15.58
Build failed.
```
I am new here, and have tried A LOT of things. I've got the restore and build running once, but installed VS 2019 Preview in an attempt to get targeting the locally built artifacts. I've uninstalled already, but am still getting these errors now.
How does the build pipeline decide which version of VS to use, and why is it that as soon as I have any other version of VS, it will always break? | 1.0 | Unable to .\restore.cmd - TagBuilderWebSite.csproj - ### Describe the bug
Cannot restore from from clean master
### To Reproduce
Steps to reproduce the behavior:
1. Using this version of ASP.NET Core '418e35c4396682aea6ee6e9833ff3bd3bfe624eb'
2. Run this code '\.restore.cmd'
3. See error
### Expected behavior
Successful restore
### Screenshots
If applicable, add screenshots to help explain your problem.
### Additional context
dotnet --info:
```
C:\dev\AspNetCore>dotnet --info
A compatible installed .NET Core SDK for global.json version [5.0.100-alpha1-014696] from [C:\dev\AspNetCore\global.json] was not found
Install the [5.0.100-alpha1-014696] .NET Core SDK or update [C:\dev\AspNetCore\global.json] with an installed .NET Core SDK:
1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk]
1.0.0 [C:\Program Files\dotnet\sdk]
1.0.4 [C:\Program Files\dotnet\sdk]
2.1.4 [C:\Program Files\dotnet\sdk]
2.1.103 [C:\Program Files\dotnet\sdk]
2.1.104 [C:\Program Files\dotnet\sdk]
2.1.201 [C:\Program Files\dotnet\sdk]
2.1.202 [C:\Program Files\dotnet\sdk]
2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk]
2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk]
2.1.300 [C:\Program Files\dotnet\sdk]
2.1.401 [C:\Program Files\dotnet\sdk]
2.1.402 [C:\Program Files\dotnet\sdk]
2.1.403 [C:\Program Files\dotnet\sdk]
2.1.502 [C:\Program Files\dotnet\sdk]
2.1.503 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009426 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009472 [C:\Program Files\dotnet\sdk]
2.1.602 [C:\Program Files\dotnet\sdk]
2.1.800 [C:\Program Files\dotnet\sdk]
2.2.101 [C:\Program Files\dotnet\sdk]
3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk]
3.0.100 [C:\Program Files\dotnet\sdk]
3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk]
Host (useful for support):
Version: 5.0.0-alpha1.19514.1
Commit: 4ace84dbf9
.NET Core SDKs installed:
1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk]
1.0.0 [C:\Program Files\dotnet\sdk]
1.0.4 [C:\Program Files\dotnet\sdk]
2.1.4 [C:\Program Files\dotnet\sdk]
2.1.103 [C:\Program Files\dotnet\sdk]
2.1.104 [C:\Program Files\dotnet\sdk]
2.1.201 [C:\Program Files\dotnet\sdk]
2.1.202 [C:\Program Files\dotnet\sdk]
2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk]
2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk]
2.1.300 [C:\Program Files\dotnet\sdk]
2.1.401 [C:\Program Files\dotnet\sdk]
2.1.402 [C:\Program Files\dotnet\sdk]
2.1.403 [C:\Program Files\dotnet\sdk]
2.1.502 [C:\Program Files\dotnet\sdk]
2.1.503 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009426 [C:\Program Files\dotnet\sdk]
2.1.600-preview-009472 [C:\Program Files\dotnet\sdk]
2.1.602 [C:\Program Files\dotnet\sdk]
2.1.800 [C:\Program Files\dotnet\sdk]
2.2.101 [C:\Program Files\dotnet\sdk]
3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk]
3.0.100 [C:\Program Files\dotnet\sdk]
3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk]
.NET Core runtimes installed:
Microsoft.AspNetCore.All 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.All 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All]
Microsoft.AspNetCore.App 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.0.0-preview6.19307.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 3.1.0-preview1.19508.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 5.0.0-dev [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 1.0.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.0.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.1.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 1.1.2 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.0.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.0-preview1-26216-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.0-rc1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.3-servicing-26724-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 5.0.0-alpha1.19514.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
To install additional .NET Core runtimes or SDKs:
https://aka.ms/dotnet-download
```
Console output:
```
C:\dev\AspNetCore>git status
On branch master
Your branch is up to date with 'origin/master'.
nothing to commit, working tree clean
C:\dev\AspNetCore>.\restore.cmd
Building of C# project is enabled and has dependencies on NodeJS projects. Building of NodeJS projects is enabled since node is detected in C:\Program Files.
Wiederherstellung in "41,94 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade
.sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process.
Wiederherstellung in "14,54 ms" für "C:\dev\AspNetCore\eng\tools\RepoTasks\RepoTasks.csproj
" abgeschlossen.
RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\netcoreapp5.0\RepoTasks.dll
RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\net472\RepoTasks.dll
Wiederherstellung in "41,11 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade
.sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process.
dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed.
dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process.
C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft.
Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN
etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge
laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T
agHelpersWebSite\TagHelpersWebSite.csproj]
Fehler beim Buildvorgang.
C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft.
Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN
etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge
laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T
agHelpersWebSite\TagHelpersWebSite.csproj]
0 Warnung(en)
1 Fehler
Verstrichene Zeit 00:00:15.58
Build failed.
```
I am new here, and have tried A LOT of things. I've got the restore and build running once, but installed VS 2019 Preview in an attempt to get targeting the locally built artifacts. I've uninstalled already, but am still getting these errors now.
How does the build pipeline decide which version of VS to use, and why is it that as soon as I have any other version of VS, it will always break? | infrastructure | unable to restore cmd tagbuilderwebsite csproj describe the bug cannot restore from from clean master to reproduce steps to reproduce the behavior using this version of asp net core run this code restore cmd see error expected behavior successful restore screenshots if applicable add screenshots to help explain your problem additional context dotnet info c dev aspnetcore dotnet info a compatible installed net core sdk for global json version from was not found install the net core sdk or update with an installed net core sdk preview preview host useful for support version commit net core sdks installed preview preview net core runtimes installed microsoft aspnetcore all final microsoft aspnetcore all final microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore app final microsoft aspnetcore app final microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app dev microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app servicing microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app to install additional net core runtimes or sdks console output c dev aspnetcore git status on branch master your branch is up to date with origin master nothing to commit working tree clean c dev aspnetcore restore cmd building of c project is enabled and has dependencies on nodejs projects building of nodejs projects is enabled since node is detected in c program files wiederherstellung in ms für c users admin nuget packages microsoft dotnet arcade sdk beta tools tools proj abgeschlossen dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process wiederherstellung in ms für c dev aspnetcore eng tools repotasks repotasks csproj abgeschlossen repotasks c dev aspnetcore artifacts bin repotasks release repotasks dll repotasks c dev aspnetcore artifacts bin repotasks release repotasks dll wiederherstellung in ms für c users admin nuget packages microsoft dotnet arcade sdk beta tools tools proj abgeschlossen dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process c program files microsoft visual studio enterprise msbuild current bin microsoft common currentversion targets error die importierte projektdatei c dev aspn etcore src mvc test websites taghelperswebsite taghelperswebsite csproj user konnte nicht ge laden werden das stammelement ist nicht vorhanden c dev aspnetcore src mvc test websites t aghelperswebsite taghelperswebsite csproj fehler beim buildvorgang c program files microsoft visual studio enterprise msbuild current bin microsoft common currentversion targets error die importierte projektdatei c dev aspn etcore src mvc test websites taghelperswebsite taghelperswebsite csproj user konnte nicht ge laden werden das stammelement ist nicht vorhanden c dev aspnetcore src mvc test websites t aghelperswebsite taghelperswebsite csproj warnung en fehler verstrichene zeit build failed i am new here and have tried a lot of things i ve got the restore and build running once but installed vs preview in an attempt to get targeting the locally built artifacts i ve uninstalled already but am still getting these errors now how does the build pipeline decide which version of vs to use and why is it that as soon as i have any other version of vs it will always break | 1 |
10,253 | 8,453,125,979 | IssuesEvent | 2018-10-20 12:33:09 | TeamBravo2018/cloned-rfid-card-detection | https://api.github.com/repos/TeamBravo2018/cloned-rfid-card-detection | opened | Setup MQTT Simulator | backlog item infrastructure messaging test production | ### Description ###
Simulate messaging between the applications by using MQTT simulator.
[https://dzone.com/articles/top-3-online-tools-to-simulate-an-mqtt-client]()
| 1.0 | Setup MQTT Simulator - ### Description ###
Simulate messaging between the applications by using MQTT simulator.
[https://dzone.com/articles/top-3-online-tools-to-simulate-an-mqtt-client]()
| infrastructure | setup mqtt simulator description simulate messaging between the applications by using mqtt simulator | 1 |
813,125 | 30,446,199,738 | IssuesEvent | 2023-07-15 17:48:03 | ncssar/radiolog | https://api.github.com/repos/ncssar/radiolog | opened | generated log PDFs should be searchable | enhancement Priority:Medium | not so important for generated clue report PDFs, but all others - radio log, team logs, clue log - should be searchable (using e.g. acrobat reader); surprised to see that they are currently not searchable | 1.0 | generated log PDFs should be searchable - not so important for generated clue report PDFs, but all others - radio log, team logs, clue log - should be searchable (using e.g. acrobat reader); surprised to see that they are currently not searchable | non_infrastructure | generated log pdfs should be searchable not so important for generated clue report pdfs but all others radio log team logs clue log should be searchable using e g acrobat reader surprised to see that they are currently not searchable | 0 |
716,244 | 24,626,321,471 | IssuesEvent | 2022-10-16 15:11:39 | AY2223S1-CS2103T-T10-1/tp | https://api.github.com/repos/AY2223S1-CS2103T-T10-1/tp | opened | As a forgetful student, I want to know which of my friends take a common module | type.Story priority.High | so that I know who to approach when I need help with that module's work. | 1.0 | As a forgetful student, I want to know which of my friends take a common module - so that I know who to approach when I need help with that module's work. | non_infrastructure | as a forgetful student i want to know which of my friends take a common module so that i know who to approach when i need help with that module s work | 0 |
293,730 | 25,318,877,925 | IssuesEvent | 2022-11-18 00:54:22 | dotnet/sdk | https://api.github.com/repos/dotnet/sdk | closed | dotnet test fail to pass MSBuild properties | Area-DotNet Test untriaged | ### Describe the bug
Until .Net 6.x, we were able to run
`
dotnet test App.sln -p Property=Value
`
and the property was adequately passed down to MSBuild, interpreted by project files and the like.
On upgrading to .Net 7.x, we found the property was no longer passed down. However, we have a workaround
`
dotnet test -p Property=Value App.sln
`
### To Reproduce
<!--
We ❤ code! Point us to a minimalistic repro project hosted in a GitHub repo, Gist snippet, or other means to see the isolated behavior.
We may close this issue if:
- the repro project you share with us is complex. We can't investigate custom projects, so don't point us to such, please.
- if we will not be able to repro the behavior you're reporting
-->
Fail - runs only one test
`
dotnet test TestProject1.csproj -p:CIBuild=Integration
`
Correct - runs only one test
`
dotnet test -p:CIBuild=Unit TestProject1.csproj
`
Correct - runs the two tests
`
dotnet test -p:CIBuild=Integration TestProject1.csproj
`
**TestProject1.csproj**
```xml
<Project Sdk="Microsoft.NET.Sdk">
<PropertyGroup>
<TargetFramework>net7.0</TargetFramework>
<ImplicitUsings>enable</ImplicitUsings>
<Nullable>enable</Nullable>
<IsPackable>false</IsPackable>
<ApplicationManifest>app.manifest</ApplicationManifest>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="Microsoft.NET.Test.Sdk" Version="17.3.2" />
<PackageReference Include="MSTest.TestAdapter" Version="2.2.10" />
<PackageReference Include="MSTest.TestFramework" Version="2.2.10" />
<PackageReference Include="coverlet.collector" Version="3.1.2" />
</ItemGroup>
</Project>
```
**Directory.Build.props**
```xml
<Project>
<PropertyGroup>
<LangVersion>11.0</LangVersion>
<TargetFramework>net7.0</TargetFramework>
<RunSettingsFilePath>$(MSBuildThisFileDirectory)Unit.runsettings</RunSettingsFilePath>
<RunSettingsFilePath Condition="$(CIBuild) != ''">$(MSBuildThisFileDirectory)$(CIBuild).runsettings</RunSettingsFilePath>
</PropertyGroup>
</Project>
```
**Unit.runsettings**
```xml
<RunSettings>
<RunConfiguration>
<TestCaseFilter>TestCategory=Unit</TestCaseFilter>
</RunConfiguration>
</RunSettings>
```
**Integration.runsettings**
```xml
<RunSettings>
<RunConfiguration>
<TestCaseFilter>TestCategory=Unit|TestCategory=Integration</TestCaseFilter>
</RunConfiguration>
</RunSettings>
```
**UnitTest1.cs**
```xml
namespace TestProject1
{
[TestClass]
public class UnitTest1
{
[TestMethod, TestCategory("Unit")]
public void TestMethod1()
{
}
[TestMethod, TestCategory("Integration")]
public void TestMethod2()
{
}
}
}
```
### Exceptions (if any)
No exceptions
### Further technical details
- Include the output of `dotnet --info`
`
.NET SDK:
Version: 7.0.100
Commit: e12b7af219
Runtime Environment:
OS Name: Windows
OS Version: 10.0.22621
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\7.0.100\
Host:
Version: 7.0.0
Architecture: x64
Commit: d099f075e4
.NET SDKs installed:
3.1.425 [C:\Program Files\dotnet\sdk]
6.0.202 [C:\Program Files\dotnet\sdk]
6.0.306 [C:\Program Files\dotnet\sdk]
7.0.100 [C:\Program Files\dotnet\sdk]
.NET runtimes installed:
Microsoft.AspNetCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Other architectures found:
x86 [C:\Program Files (x86)\dotnet]
registered at [HKLM\SOFTWARE\dotnet\Setup\InstalledVersions\x86\InstallLocation]
Environment variables:
Not set
global.json file:
Not found
Learn more:
https://aka.ms/dotnet/info
Download .NET:
https://aka.ms/dotnet/download
`
- VS 17.4
### Related
Passing properties seem to have worked since earlier versions, but it is not well documented.
#23198
| 1.0 | dotnet test fail to pass MSBuild properties - ### Describe the bug
Until .Net 6.x, we were able to run
`
dotnet test App.sln -p Property=Value
`
and the property was adequately passed down to MSBuild, interpreted by project files and the like.
On upgrading to .Net 7.x, we found the property was no longer passed down. However, we have a workaround
`
dotnet test -p Property=Value App.sln
`
### To Reproduce
<!--
We ❤ code! Point us to a minimalistic repro project hosted in a GitHub repo, Gist snippet, or other means to see the isolated behavior.
We may close this issue if:
- the repro project you share with us is complex. We can't investigate custom projects, so don't point us to such, please.
- if we will not be able to repro the behavior you're reporting
-->
Fail - runs only one test
`
dotnet test TestProject1.csproj -p:CIBuild=Integration
`
Correct - runs only one test
`
dotnet test -p:CIBuild=Unit TestProject1.csproj
`
Correct - runs the two tests
`
dotnet test -p:CIBuild=Integration TestProject1.csproj
`
**TestProject1.csproj**
```xml
<Project Sdk="Microsoft.NET.Sdk">
<PropertyGroup>
<TargetFramework>net7.0</TargetFramework>
<ImplicitUsings>enable</ImplicitUsings>
<Nullable>enable</Nullable>
<IsPackable>false</IsPackable>
<ApplicationManifest>app.manifest</ApplicationManifest>
</PropertyGroup>
<ItemGroup>
<PackageReference Include="Microsoft.NET.Test.Sdk" Version="17.3.2" />
<PackageReference Include="MSTest.TestAdapter" Version="2.2.10" />
<PackageReference Include="MSTest.TestFramework" Version="2.2.10" />
<PackageReference Include="coverlet.collector" Version="3.1.2" />
</ItemGroup>
</Project>
```
**Directory.Build.props**
```xml
<Project>
<PropertyGroup>
<LangVersion>11.0</LangVersion>
<TargetFramework>net7.0</TargetFramework>
<RunSettingsFilePath>$(MSBuildThisFileDirectory)Unit.runsettings</RunSettingsFilePath>
<RunSettingsFilePath Condition="$(CIBuild) != ''">$(MSBuildThisFileDirectory)$(CIBuild).runsettings</RunSettingsFilePath>
</PropertyGroup>
</Project>
```
**Unit.runsettings**
```xml
<RunSettings>
<RunConfiguration>
<TestCaseFilter>TestCategory=Unit</TestCaseFilter>
</RunConfiguration>
</RunSettings>
```
**Integration.runsettings**
```xml
<RunSettings>
<RunConfiguration>
<TestCaseFilter>TestCategory=Unit|TestCategory=Integration</TestCaseFilter>
</RunConfiguration>
</RunSettings>
```
**UnitTest1.cs**
```xml
namespace TestProject1
{
[TestClass]
public class UnitTest1
{
[TestMethod, TestCategory("Unit")]
public void TestMethod1()
{
}
[TestMethod, TestCategory("Integration")]
public void TestMethod2()
{
}
}
}
```
### Exceptions (if any)
No exceptions
### Further technical details
- Include the output of `dotnet --info`
`
.NET SDK:
Version: 7.0.100
Commit: e12b7af219
Runtime Environment:
OS Name: Windows
OS Version: 10.0.22621
OS Platform: Windows
RID: win10-x64
Base Path: C:\Program Files\dotnet\sdk\7.0.100\
Host:
Version: 7.0.0
Architecture: x64
Commit: d099f075e4
.NET SDKs installed:
3.1.425 [C:\Program Files\dotnet\sdk]
6.0.202 [C:\Program Files\dotnet\sdk]
6.0.306 [C:\Program Files\dotnet\sdk]
7.0.100 [C:\Program Files\dotnet\sdk]
.NET runtimes installed:
Microsoft.AspNetCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.AspNetCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App]
Microsoft.NETCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.NETCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App]
Microsoft.WindowsDesktop.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Microsoft.WindowsDesktop.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App]
Other architectures found:
x86 [C:\Program Files (x86)\dotnet]
registered at [HKLM\SOFTWARE\dotnet\Setup\InstalledVersions\x86\InstallLocation]
Environment variables:
Not set
global.json file:
Not found
Learn more:
https://aka.ms/dotnet/info
Download .NET:
https://aka.ms/dotnet/download
`
- VS 17.4
### Related
Passing properties seem to have worked since earlier versions, but it is not well documented.
#23198
| non_infrastructure | dotnet test fail to pass msbuild properties describe the bug until net x we were able to run dotnet test app sln p property value and the property was adequately passed down to msbuild interpreted by project files and the like on upgrading to net x we found the property was no longer passed down however we have a workaround dotnet test p property value app sln to reproduce we ❤ code point us to a minimalistic repro project hosted in a github repo gist snippet or other means to see the isolated behavior we may close this issue if the repro project you share with us is complex we can t investigate custom projects so don t point us to such please if we will not be able to repro the behavior you re reporting fail runs only one test dotnet test csproj p cibuild integration correct runs only one test dotnet test p cibuild unit csproj correct runs the two tests dotnet test p cibuild integration csproj csproj xml enable enable false app manifest directory build props xml msbuildthisfiledirectory unit runsettings msbuildthisfiledirectory cibuild runsettings unit runsettings xml testcategory unit integration runsettings xml testcategory unit testcategory integration cs xml namespace public class public void public void exceptions if any no exceptions further technical details include the output of dotnet info net sdk version commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk host version architecture commit net sdks installed net runtimes installed microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app other architectures found registered at environment variables not set global json file not found learn more download net vs related passing properties seem to have worked since earlier versions but it is not well documented | 0 |
6,370 | 6,361,319,796 | IssuesEvent | 2017-07-31 12:37:23 | warg-lang/warg | https://api.github.com/repos/warg-lang/warg | opened | Generate accessible static analysis diagnostics | ci enhancement infrastructure | [neovim](https://github.com/neovim/neovim) provides a nice [diagnostics overview](https://neovim.io/doc/reports/clang/) using Clang Static Analysis. While it is not completely transparent how to do that, making similar page would be great. | 1.0 | Generate accessible static analysis diagnostics - [neovim](https://github.com/neovim/neovim) provides a nice [diagnostics overview](https://neovim.io/doc/reports/clang/) using Clang Static Analysis. While it is not completely transparent how to do that, making similar page would be great. | infrastructure | generate accessible static analysis diagnostics provides a nice using clang static analysis while it is not completely transparent how to do that making similar page would be great | 1 |
6,826 | 6,657,418,989 | IssuesEvent | 2017-09-30 05:13:33 | nathantspencer/AtLAS-mobile | https://api.github.com/repos/nathantspencer/AtLAS-mobile | closed | UI + Request for Sign In/Out | Client-Server Security UI | Sign in request will require username and password. Sign out request will require username and authentication key. | True | UI + Request for Sign In/Out - Sign in request will require username and password. Sign out request will require username and authentication key. | non_infrastructure | ui request for sign in out sign in request will require username and password sign out request will require username and authentication key | 0 |
57,982 | 7,110,431,567 | IssuesEvent | 2018-01-17 10:36:10 | Userfeeds/Apps | https://api.github.com/repos/Userfeeds/Apps | closed | Inform that Metamask is disabled | design | Currently whenever my Metamask is not logged I can't see any functional buttons. I'd rather have them disabled and inform the user about this issue.

| 1.0 | Inform that Metamask is disabled - Currently whenever my Metamask is not logged I can't see any functional buttons. I'd rather have them disabled and inform the user about this issue.

| non_infrastructure | inform that metamask is disabled currently whenever my metamask is not logged i can t see any functional buttons i d rather have them disabled and inform the user about this issue | 0 |
2,631 | 2,699,148,226 | IssuesEvent | 2015-04-03 14:47:24 | itgsod-lukas-michanek/Neocache | https://api.github.com/repos/itgsod-lukas-michanek/Neocache | opened | Project cleanup | documentation | We are soon about to leave the stage of database modelling, but before we do, make sure that everything is properly commented, formatted and everything else. | 1.0 | Project cleanup - We are soon about to leave the stage of database modelling, but before we do, make sure that everything is properly commented, formatted and everything else. | non_infrastructure | project cleanup we are soon about to leave the stage of database modelling but before we do make sure that everything is properly commented formatted and everything else | 0 |
179,745 | 21,580,319,605 | IssuesEvent | 2022-05-02 17:59:45 | vincenzodistasio97/excel-to-json | https://api.github.com/repos/vincenzodistasio97/excel-to-json | opened | CVE-2020-28498 (Medium) detected in elliptic-6.4.0.tgz | security vulnerability | ## CVE-2020-28498 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.1.1.tgz (Root Library)
- webpack-3.8.1.tgz
- node-libs-browser-2.1.0.tgz
- crypto-browserify-3.12.0.tgz
- create-ecdh-4.0.0.tgz
- :x: **elliptic-6.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/excel-to-json/commit/e367d4db4134dc676344b2b9fb2443300bd3c9c7">e367d4db4134dc676344b2b9fb2443300bd3c9c7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution (elliptic): 6.5.4</p>
<p>Direct dependency fix Resolution (react-scripts): 1.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-28498 (Medium) detected in elliptic-6.4.0.tgz - ## CVE-2020-28498 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary>
<p>EC cryptography</p>
<p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/elliptic/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.1.1.tgz (Root Library)
- webpack-3.8.1.tgz
- node-libs-browser-2.1.0.tgz
- crypto-browserify-3.12.0.tgz
- create-ecdh-4.0.0.tgz
- :x: **elliptic-6.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/excel-to-json/commit/e367d4db4134dc676344b2b9fb2443300bd3c9c7">e367d4db4134dc676344b2b9fb2443300bd3c9c7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.
<p>Publish Date: 2021-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p>
<p>Release Date: 2021-02-02</p>
<p>Fix Resolution (elliptic): 6.5.4</p>
<p>Direct dependency fix Resolution (react-scripts): 1.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in elliptic tgz cve medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file client package json path to vulnerable library client node modules elliptic package json dependency hierarchy react scripts tgz root library webpack tgz node libs browser tgz crypto browserify tgz create ecdh tgz x elliptic tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution elliptic direct dependency fix resolution react scripts step up your open source security game with whitesource | 0 |
24,255 | 17,046,813,829 | IssuesEvent | 2021-07-06 00:59:33 | google/iree | https://api.github.com/repos/google/iree | closed | [build] cant build iree | bug 🐞 help wanted infrastructure 🛠️ support 🤗 | Hello,
I have some problems and errors while building IREE
I just followed ```getting started``` to build
https://google.github.io/iree/building-from-source/getting-started/
Could you give some hints or instructions for this problems?
Thanks!
## Trials
```
$ sudo apt-get install clang lld
$ sudo apt-get install clang++
$ sudo apt install python-clang-4.0
$ sudo apt install python-clang-5.0
$ sudo apt install python-clang-6.0
$ sudo apt install python-clang-7
$ sudo apt install python-clang-8
$ sudo apt install python-clang-9
$ sudo apt install python3-clang-10
$ sudo apt update
$ sudo apt upgrade
$ sudo apt autoremove
# g++ to clang++
$ sudo update-alternatives --config c++
# gcc to clang
$ sudo update-alternatives --config cc
```
```
$ export CC=clang
$ export CXX=clang++
$ sudo rm -r ../iree-build
$ cmake -B ../iree-build/ -DCMAKE_BUILD_TYPE=RelWithDebInfo . -GNinja
$ cmake --build ../iree-build/ -j6
```
## Machine Spec
- OS : Ubuntu 18.04.5 LTS
- CPU : Intel(R) Core(TM) i7-10700K CPU @ 3.80GHz
- RAM : 16GB
- SSD : nvme, samsung
## Verbose log
### Results of checking ```$CC```, ```$CXX``` after failed with errors

```
$ cmake --build ../iree-build/ -j6
[0/2] Re-checking globbed directories...
[2905/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o
FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o
/usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:47:24: error: only virtual member functions can be marked 'override'
void runOnFunction() override {
^~~~~~~~~
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:48:5: error: use of undeclared identifier 'getFunction'
getFunction().walk([&](WhileOp whileOp) { MatchAndRewrite(whileOp); });
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>' requested here
return std::make_unique<ControlFlowToScfPass>();
^
/home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'ControlFlowToScfPass'
virtual void runOnOperation() = 0;
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:125:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, const mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass &>' requested here
return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this));
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowToScfPassBase<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::clonePass' requested here
delete __ptr;
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::operator()' requested here
get_deleter()(__ptr);
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass> >::~unique_ptr' requested here
return std::make_unique<ControlFlowToScfPass>();
^
4 errors generated.
[2906/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o
FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o
/usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:187:24: error: only virtual member functions can be marked 'override'
void runOnFunction() override;
^~~~~~~~
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:205:38: error: use of undeclared identifier 'getFunction'
(void)applyPatternsAndFoldGreedily(getFunction(), std::move(patterns));
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>' requested here
return std::make_unique<LegalizeToStandardPass>();
^
/home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeToStandardPass'
virtual void runOnOperation() = 0;
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:229:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, const mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass &>' requested here
return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this));
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeToStandardPassBase<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::clonePass' requested here
delete __ptr;
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::operator()' requested here
get_deleter()(__ptr);
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass> >::~unique_ptr' requested here
return std::make_unique<LegalizeToStandardPass>();
^
4 errors generated.
[2907/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o
FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o
/usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:43:24: error: only virtual member functions can be marked 'override'
void runOnFunction() override;
^~~~~~~~
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:216:15: error: use of undeclared identifier 'getFunction'
auto func = getFunction();
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>' requested here
return std::make_unique<LegalizeControlFlowPass>();
^
/home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeControlFlowPass'
virtual void runOnOperation() = 0;
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:99:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, const mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass &>' requested here
return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this));
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowPassBase<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::clonePass' requested here
delete __ptr;
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::operator()' requested here
get_deleter()(__ptr);
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass> >::~unique_ptr' requested here
return std::make_unique<LegalizeControlFlowPass>();
^
4 errors generated.
[2910/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/IR/CMakeFiles/obj.MhloDialect.dir/hlo_ops.cc.o
ninja: build stopped: subcommand failed.
``` | 1.0 | [build] cant build iree - Hello,
I have some problems and errors while building IREE
I just followed ```getting started``` to build
https://google.github.io/iree/building-from-source/getting-started/
Could you give some hints or instructions for this problems?
Thanks!
## Trials
```
$ sudo apt-get install clang lld
$ sudo apt-get install clang++
$ sudo apt install python-clang-4.0
$ sudo apt install python-clang-5.0
$ sudo apt install python-clang-6.0
$ sudo apt install python-clang-7
$ sudo apt install python-clang-8
$ sudo apt install python-clang-9
$ sudo apt install python3-clang-10
$ sudo apt update
$ sudo apt upgrade
$ sudo apt autoremove
# g++ to clang++
$ sudo update-alternatives --config c++
# gcc to clang
$ sudo update-alternatives --config cc
```
```
$ export CC=clang
$ export CXX=clang++
$ sudo rm -r ../iree-build
$ cmake -B ../iree-build/ -DCMAKE_BUILD_TYPE=RelWithDebInfo . -GNinja
$ cmake --build ../iree-build/ -j6
```
## Machine Spec
- OS : Ubuntu 18.04.5 LTS
- CPU : Intel(R) Core(TM) i7-10700K CPU @ 3.80GHz
- RAM : 16GB
- SSD : nvme, samsung
## Verbose log
### Results of checking ```$CC```, ```$CXX``` after failed with errors

```
$ cmake --build ../iree-build/ -j6
[0/2] Re-checking globbed directories...
[2905/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o
FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o
/usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:47:24: error: only virtual member functions can be marked 'override'
void runOnFunction() override {
^~~~~~~~~
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:48:5: error: use of undeclared identifier 'getFunction'
getFunction().walk([&](WhileOp whileOp) { MatchAndRewrite(whileOp); });
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>' requested here
return std::make_unique<ControlFlowToScfPass>();
^
/home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'ControlFlowToScfPass'
virtual void runOnOperation() = 0;
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:125:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, const mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass &>' requested here
return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this));
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowToScfPassBase<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::clonePass' requested here
delete __ptr;
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::operator()' requested here
get_deleter()(__ptr);
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass> >::~unique_ptr' requested here
return std::make_unique<ControlFlowToScfPass>();
^
4 errors generated.
[2906/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o
FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o
/usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:187:24: error: only virtual member functions can be marked 'override'
void runOnFunction() override;
^~~~~~~~
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:205:38: error: use of undeclared identifier 'getFunction'
(void)applyPatternsAndFoldGreedily(getFunction(), std::move(patterns));
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>' requested here
return std::make_unique<LegalizeToStandardPass>();
^
/home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeToStandardPass'
virtual void runOnOperation() = 0;
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:229:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, const mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass &>' requested here
return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this));
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeToStandardPassBase<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::clonePass' requested here
delete __ptr;
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::operator()' requested here
get_deleter()(__ptr);
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass> >::~unique_ptr' requested here
return std::make_unique<LegalizeToStandardPass>();
^
4 errors generated.
[2907/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o
FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o
/usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:43:24: error: only virtual member functions can be marked 'override'
void runOnFunction() override;
^~~~~~~~
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:216:15: error: use of undeclared identifier 'getFunction'
auto func = getFunction();
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>' requested here
return std::make_unique<LegalizeControlFlowPass>();
^
/home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeControlFlowPass'
virtual void runOnOperation() = 0;
^
In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19:
In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24:
In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80:
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass'
{ return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); }
^
/home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:99:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, const mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass &>' requested here
return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this));
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowPassBase<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::clonePass' requested here
delete __ptr;
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::operator()' requested here
get_deleter()(__ptr);
^
/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass> >::~unique_ptr' requested here
return std::make_unique<LegalizeControlFlowPass>();
^
4 errors generated.
[2910/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/IR/CMakeFiles/obj.MhloDialect.dir/hlo_ops.cc.o
ninja: build stopped: subcommand failed.
``` | infrastructure | cant build iree hello i have some problems and errors while building iree i just followed getting started to build could you give some hints or instructions for this problems thanks trials sudo apt get install clang lld sudo apt get install clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install clang sudo apt update sudo apt upgrade sudo apt autoremove g to clang sudo update alternatives config c gcc to clang sudo update alternatives config cc export cc clang export cxx clang sudo rm r iree build cmake b iree build dcmake build type relwithdebinfo gninja cmake build iree build machine spec os ubuntu lts cpu intel r core tm cpu ram ssd nvme samsung verbose log results of checking cc cxx after failed with errors cmake build iree build re checking globbed directories building cxx object third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o failed third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o usr bin clang dgtest has rtti d stdc constant macros d stdc format macros d stdc limit macros i home kkimj iree third party mlir hlo lib dialect mhlo transforms i home kkimj iree build third party mlir hlo lib dialect mhlo transforms i home kkimj iree third party llvm project llvm include i home kkimj iree build third party llvm project llvm include i home kkimj iree third party llvm project mlir include i home kkimj iree build third party llvm project llvm tools mlir include i home kkimj iree third party mlir hlo include i home kkimj iree build third party mlir hlo include i home kkimj iree build third party mlir hlo fpic fvisibility inlines hidden werror date time werror unguarded availability new w fdiagnostics color ffunction sections fdata sections g dndebug fpic fno exceptions fno rtti std gnu md mt third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o mf third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o d o third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o c home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc error only virtual member functions can be marked override void runonfunction override home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc error use of undeclared identifier getfunction getfunction walk whileop whileop matchandrewrite whileop in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc in file included from home kkimj iree third party llvm project llvm include llvm support casting h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace controlflowtoscfpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc note in instantiation of function template specialization std make unique requested here return std make unique home kkimj iree third party llvm project mlir include mlir pass pass h note unimplemented pure virtual method runonoperation in controlflowtoscfpass virtual void runonoperation in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc in file included from home kkimj iree third party llvm project llvm include llvm support casting h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace controlflowtoscfpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo include mlir hlo dialect mhlo transforms mhlo passes h inc note in instantiation of function template specialization std make unique requested here return std make unique static cast this usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function mlir mhlo legalizecontrolflowtoscfpassbase clonepass requested here delete ptr usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function std default delete operator requested here get deleter ptr home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc note in instantiation of member function std unique ptr unique ptr requested here return std make unique errors generated building cxx object third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o failed third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o usr bin clang dgtest has rtti d stdc constant macros d stdc format macros d stdc limit macros i home kkimj iree third party mlir hlo lib dialect mhlo transforms i home kkimj iree build third party mlir hlo lib dialect mhlo transforms i home kkimj iree third party llvm project llvm include i home kkimj iree build third party llvm project llvm include i home kkimj iree third party llvm project mlir include i home kkimj iree build third party llvm project llvm tools mlir include i home kkimj iree third party mlir hlo include i home kkimj iree build third party mlir hlo include i home kkimj iree build third party mlir hlo fpic fvisibility inlines hidden werror date time werror unguarded availability new w fdiagnostics color ffunction sections fdata sections g dndebug fpic fno exceptions fno rtti std gnu md mt third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o mf third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o d o third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o c home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc error only virtual member functions can be marked override void runonfunction override home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc error use of undeclared identifier getfunction void applypatternsandfoldgreedily getfunction std move patterns in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc in file included from home kkimj iree third party llvm project llvm include llvm adt stringswitch h in file included from home kkimj iree third party llvm project llvm include llvm adt stringref h in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizetostandardpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc note in instantiation of function template specialization std make unique requested here return std make unique home kkimj iree third party llvm project mlir include mlir pass pass h note unimplemented pure virtual method runonoperation in legalizetostandardpass virtual void runonoperation in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc in file included from home kkimj iree third party llvm project llvm include llvm adt stringswitch h in file included from home kkimj iree third party llvm project llvm include llvm adt stringref h in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizetostandardpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo include mlir hlo dialect mhlo transforms mhlo passes h inc note in instantiation of function template specialization std make unique requested here return std make unique static cast this usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function mlir mhlo legalizetostandardpassbase clonepass requested here delete ptr usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function std default delete operator requested here get deleter ptr home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc note in instantiation of member function std unique ptr unique ptr requested here return std make unique errors generated building cxx object third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o failed third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o usr bin clang dgtest has rtti d stdc constant macros d stdc format macros d stdc limit macros i home kkimj iree third party mlir hlo lib dialect mhlo transforms i home kkimj iree build third party mlir hlo lib dialect mhlo transforms i home kkimj iree third party llvm project llvm include i home kkimj iree build third party llvm project llvm include i home kkimj iree third party llvm project mlir include i home kkimj iree build third party llvm project llvm tools mlir include i home kkimj iree third party mlir hlo include i home kkimj iree build third party mlir hlo include i home kkimj iree build third party mlir hlo fpic fvisibility inlines hidden werror date time werror unguarded availability new w fdiagnostics color ffunction sections fdata sections g dndebug fpic fno exceptions fno rtti std gnu md mt third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o mf third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o d o third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o c home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc error only virtual member functions can be marked override void runonfunction override home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc error use of undeclared identifier getfunction auto func getfunction in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizecontrolflowpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc note in instantiation of function template specialization std make unique requested here return std make unique home kkimj iree third party llvm project mlir include mlir pass pass h note unimplemented pure virtual method runonoperation in legalizecontrolflowpass virtual void runonoperation in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizecontrolflowpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo include mlir hlo dialect mhlo transforms mhlo passes h inc note in instantiation of function template specialization std make unique requested here return std make unique static cast this usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function mlir mhlo legalizecontrolflowpassbase clonepass requested here delete ptr usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function std default delete operator requested here get deleter ptr home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc note in instantiation of member function std unique ptr unique ptr requested here return std make unique errors generated building cxx object third party mlir hlo lib dialect mhlo ir cmakefiles obj mhlodialect dir hlo ops cc o ninja build stopped subcommand failed | 1 |
6,125 | 22,240,708,398 | IssuesEvent | 2022-06-09 04:46:08 | uselagoon/lagoon | https://api.github.com/repos/uselagoon/lagoon | opened | Ability to have a different backup schedule for standby production environments | 8-automation-helpers | We are seeing excess database load that is at least partially driven by the fact that both production environments are backing up the database every 6 hours.
There is no business need for this particular customer to backup the standby database, only the active database.
It would be handy to have the ability to define a backup schedule for active, and another for standby. | 1.0 | Ability to have a different backup schedule for standby production environments - We are seeing excess database load that is at least partially driven by the fact that both production environments are backing up the database every 6 hours.
There is no business need for this particular customer to backup the standby database, only the active database.
It would be handy to have the ability to define a backup schedule for active, and another for standby. | non_infrastructure | ability to have a different backup schedule for standby production environments we are seeing excess database load that is at least partially driven by the fact that both production environments are backing up the database every hours there is no business need for this particular customer to backup the standby database only the active database it would be handy to have the ability to define a backup schedule for active and another for standby | 0 |
10,791 | 8,722,706,220 | IssuesEvent | 2018-12-09 15:09:09 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | NetCore2.1->2.2 Upgrade Issue; TargetFramework dropped/not recognized | area-Infrastructure | I'm trying to upgrade my web application from netCore 2.1 to 2.2. I have updated the TargetFramework to netcoreapp2.2 and I receive the following errors upon restore/build/clean:
```
MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
```
I have also done a manual clean (deleting obj/bin directories).
I **can** build if I change from using `<TargetFramework>` to `<TargetFrameworks>`, however I'd rather not do this and attempt to locate the underlying issue. I have two test projects in the same solution that reference the main project. I was able to migrate _them_ to netcoreapp2.2 without switching to TargetFramework**s**, I would like to do the same thing for the main web app.
When I use TargetFramework (singular), Visual Studio updates the dependencies graph and removes the SDK node:

I've tried adding a global.json specifying the SDK version directly as 2.2.100 but that doesn't make any difference.
Here is my .csproj file:
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<AssemblyName>MyProduct.Web</AssemblyName>
<OutputType>Exe</OutputType>
<TargetFramework>netcoreapp2.2</TargetFramework>
<PreserveCompilationContext>true</PreserveCompilationContext>
<LangVersion>latest</LangVersion>
</PropertyGroup>
<PropertyGroup>
<GenerateAssemblyInfo>false</GenerateAssemblyInfo>
<PackageId>$(AssemblyName)</PackageId>
<Authors>MyCompany</Authors>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)' == 'Debug'">
<RazorCompileOnBuild>false</RazorCompileOnBuild>
<RazorCompileOnPublish>false</RazorCompileOnPublish>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)'=='Debug'">
<DefineConstants>DEBUG;TRACE</DefineConstants>
</PropertyGroup>
<ItemGroup>
<None Update="wwwroot\**\*">
<CopyToPublishDirectory>PreserveNewest</CopyToPublishDirectory>
</None>
</ItemGroup>
<ItemGroup>
<ProjectReference Include="..\MyProduct.Web.Core\MyProduct.Web.Core.csproj" />
<ProjectReference Include="..\MyProduct.Web.Api.Contracts\MyProduct.Web.Api.Contracts.csproj" />
<ProjectReference Include="..\MyProduct.Web.ViewModels\MyProduct.Web.ViewModels.csproj" />
</ItemGroup>
<ItemGroup>
<PackageReference Include="Microsoft.AspNetCore.App" />
<PackageReference Include="BeginCollectionItemCore" Version="1.0.4" />
<PackageReference Include="Novell.Directory.Ldap.NETStandard" Version="2.3.5" />
<PackageReference Include="Microsoft.AspNetCore.Authentication.Cookies" Version="2.2.0" />
<PackageReference Include="Microsoft.AspNetCore.Authentication.JwtBearer" Version="2.2.0" />
</ItemGroup>
<ItemGroup>
<PackageReference Include="Microsoft.AspNetCore.Razor.Design" Version="2.2.0" PrivateAssets="All" />
<PackageReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Design" Version="2.2.0" PrivateAssets="All" />
<DotNetCliToolReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Tools" Version="2.0.4" />
</ItemGroup>
<ItemGroup>
<Content Update="Views\_ViewImports.cshtml">
<Pack>$(IncludeRazorContentInPack)</Pack>
</Content>
</ItemGroup>
</Project>
``` | 1.0 | NetCore2.1->2.2 Upgrade Issue; TargetFramework dropped/not recognized - I'm trying to upgrade my web application from netCore 2.1 to 2.2. I have updated the TargetFramework to netcoreapp2.2 and I receive the following errors upon restore/build/clean:
```
MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2)
```
I have also done a manual clean (deleting obj/bin directories).
I **can** build if I change from using `<TargetFramework>` to `<TargetFrameworks>`, however I'd rather not do this and attempt to locate the underlying issue. I have two test projects in the same solution that reference the main project. I was able to migrate _them_ to netcoreapp2.2 without switching to TargetFramework**s**, I would like to do the same thing for the main web app.
When I use TargetFramework (singular), Visual Studio updates the dependencies graph and removes the SDK node:

I've tried adding a global.json specifying the SDK version directly as 2.2.100 but that doesn't make any difference.
Here is my .csproj file:
```xml
<Project Sdk="Microsoft.NET.Sdk.Web">
<PropertyGroup>
<AssemblyName>MyProduct.Web</AssemblyName>
<OutputType>Exe</OutputType>
<TargetFramework>netcoreapp2.2</TargetFramework>
<PreserveCompilationContext>true</PreserveCompilationContext>
<LangVersion>latest</LangVersion>
</PropertyGroup>
<PropertyGroup>
<GenerateAssemblyInfo>false</GenerateAssemblyInfo>
<PackageId>$(AssemblyName)</PackageId>
<Authors>MyCompany</Authors>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)' == 'Debug'">
<RazorCompileOnBuild>false</RazorCompileOnBuild>
<RazorCompileOnPublish>false</RazorCompileOnPublish>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)'=='Debug'">
<DefineConstants>DEBUG;TRACE</DefineConstants>
</PropertyGroup>
<ItemGroup>
<None Update="wwwroot\**\*">
<CopyToPublishDirectory>PreserveNewest</CopyToPublishDirectory>
</None>
</ItemGroup>
<ItemGroup>
<ProjectReference Include="..\MyProduct.Web.Core\MyProduct.Web.Core.csproj" />
<ProjectReference Include="..\MyProduct.Web.Api.Contracts\MyProduct.Web.Api.Contracts.csproj" />
<ProjectReference Include="..\MyProduct.Web.ViewModels\MyProduct.Web.ViewModels.csproj" />
</ItemGroup>
<ItemGroup>
<PackageReference Include="Microsoft.AspNetCore.App" />
<PackageReference Include="BeginCollectionItemCore" Version="1.0.4" />
<PackageReference Include="Novell.Directory.Ldap.NETStandard" Version="2.3.5" />
<PackageReference Include="Microsoft.AspNetCore.Authentication.Cookies" Version="2.2.0" />
<PackageReference Include="Microsoft.AspNetCore.Authentication.JwtBearer" Version="2.2.0" />
</ItemGroup>
<ItemGroup>
<PackageReference Include="Microsoft.AspNetCore.Razor.Design" Version="2.2.0" PrivateAssets="All" />
<PackageReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Design" Version="2.2.0" PrivateAssets="All" />
<DotNetCliToolReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Tools" Version="2.0.4" />
</ItemGroup>
<ItemGroup>
<Content Update="Views\_ViewImports.cshtml">
<Pack>$(IncludeRazorContentInPack)</Pack>
</Content>
</ItemGroup>
</Project>
``` | infrastructure | upgrade issue targetframework dropped not recognized i m trying to upgrade my web application from netcore to i have updated the targetframework to and i receive the following errors upon restore build clean myproduct web csproj error package microsoft aspnetcore app is not compatible with netcoreapp version package microsoft aspnetcore app supports netcoreapp version myproduct web csproj error package microsoft netcore app is not compatible with netcoreapp version package microsoft netcore app supports netcoreapp version myproduct web csproj error package microsoft aspnetcore app is not compatible with netcoreapp version win package microsoft aspnetcore app supports netcoreapp version myproduct web csproj error package microsoft netcore app is not compatible with netcoreapp version win package microsoft netcore app supports netcoreapp version i have also done a manual clean deleting obj bin directories i can build if i change from using to however i d rather not do this and attempt to locate the underlying issue i have two test projects in the same solution that reference the main project i was able to migrate them to without switching to targetframework s i would like to do the same thing for the main web app when i use targetframework singular visual studio updates the dependencies graph and removes the sdk node i ve tried adding a global json specifying the sdk version directly as but that doesn t make any difference here is my csproj file xml myproduct web exe true latest false assemblyname mycompany false false debug trace preservenewest includerazorcontentinpack | 1 |
33,054 | 27,176,315,470 | IssuesEvent | 2023-02-18 02:54:47 | NASA-IMPACT/csdap-cumulus | https://api.github.com/repos/NASA-IMPACT/csdap-cumulus | closed | Standup Cumulus development deployment in CBA sandbox account for Kris | good first issue infrastructure | Standup Cumulus 13.4.0 deployment to `csda-app-sbx-7894` CBA AWS account for Kris to use for development, by following the instructions in this repo's `README.md` file.
Acceptance criteria:
- [x] Successful Terraform deployment per the `README.md` instructions, with the exception of running a successful E2E test (i.e., all resources managed by Terraform are successfully created)
- [x] Stretch: successfully run sample E2E discovery/ingestion test, per the last part of the instructions in the `README.md` file (if this cannot be completed in the same sprint as the item above, this issue should still be closed, and new issues for any problems blocking this criterion should be created and added to the subsequent sprint) | 1.0 | Standup Cumulus development deployment in CBA sandbox account for Kris - Standup Cumulus 13.4.0 deployment to `csda-app-sbx-7894` CBA AWS account for Kris to use for development, by following the instructions in this repo's `README.md` file.
Acceptance criteria:
- [x] Successful Terraform deployment per the `README.md` instructions, with the exception of running a successful E2E test (i.e., all resources managed by Terraform are successfully created)
- [x] Stretch: successfully run sample E2E discovery/ingestion test, per the last part of the instructions in the `README.md` file (if this cannot be completed in the same sprint as the item above, this issue should still be closed, and new issues for any problems blocking this criterion should be created and added to the subsequent sprint) | infrastructure | standup cumulus development deployment in cba sandbox account for kris standup cumulus deployment to csda app sbx cba aws account for kris to use for development by following the instructions in this repo s readme md file acceptance criteria successful terraform deployment per the readme md instructions with the exception of running a successful test i e all resources managed by terraform are successfully created stretch successfully run sample discovery ingestion test per the last part of the instructions in the readme md file if this cannot be completed in the same sprint as the item above this issue should still be closed and new issues for any problems blocking this criterion should be created and added to the subsequent sprint | 1 |
237,712 | 7,763,480,595 | IssuesEvent | 2018-06-01 16:42:31 | AnalyticalGraphicsInc/cesium | https://api.github.com/repos/AnalyticalGraphicsInc/cesium | opened | Sandcastle gallery missing scrollbar at certain sizes | priority - high | 
I assume this is the fact that the labels wrapped to a second line and it's pushing everything down (and we have overflow hidden so the scrollbar is underneath everything).
@emackey any ideas for a quick fix? | 1.0 | Sandcastle gallery missing scrollbar at certain sizes - 
I assume this is the fact that the labels wrapped to a second line and it's pushing everything down (and we have overflow hidden so the scrollbar is underneath everything).
@emackey any ideas for a quick fix? | non_infrastructure | sandcastle gallery missing scrollbar at certain sizes i assume this is the fact that the labels wrapped to a second line and it s pushing everything down and we have overflow hidden so the scrollbar is underneath everything emackey any ideas for a quick fix | 0 |
33,772 | 27,801,037,986 | IssuesEvent | 2023-03-17 15:51:26 | pixiebrix/pixiebrix-extension | https://api.github.com/repos/pixiebrix/pixiebrix-extension | opened | Rainforest: Create a new "PixieBrix Extension" Site and Environments for builds of the extension | infrastructure developer experience | This is slice 1 of the Test-Independent Rainforest project, see full implementation sketch here:
https://www.notion.so/pixiebrix/Test-Independent-Rainforest-Environments-67a031a2298547e2835dbccf5d35a5a8?pvs=4#091a70e60fc94aae8eed123e5a329041
In a nutshell:
- Add a new "PixieBrix Extension" Site to Rainforest
- Add 3 new Environments to the "PixieBrix Extension" site - one for `main`, `release` and `cws` builds of the extension
- Modify `upload_extension.sh` to build & upload the extension with consistent filenames that will be overridden on s3
- Modify our rainforest github workflow to run tests without a custom start url
- Update existing tests to use the "PixieBrix Extension" site and start at the Environment `start_url` | 1.0 | Rainforest: Create a new "PixieBrix Extension" Site and Environments for builds of the extension - This is slice 1 of the Test-Independent Rainforest project, see full implementation sketch here:
https://www.notion.so/pixiebrix/Test-Independent-Rainforest-Environments-67a031a2298547e2835dbccf5d35a5a8?pvs=4#091a70e60fc94aae8eed123e5a329041
In a nutshell:
- Add a new "PixieBrix Extension" Site to Rainforest
- Add 3 new Environments to the "PixieBrix Extension" site - one for `main`, `release` and `cws` builds of the extension
- Modify `upload_extension.sh` to build & upload the extension with consistent filenames that will be overridden on s3
- Modify our rainforest github workflow to run tests without a custom start url
- Update existing tests to use the "PixieBrix Extension" site and start at the Environment `start_url` | infrastructure | rainforest create a new pixiebrix extension site and environments for builds of the extension this is slice of the test independent rainforest project see full implementation sketch here in a nutshell add a new pixiebrix extension site to rainforest add new environments to the pixiebrix extension site one for main release and cws builds of the extension modify upload extension sh to build upload the extension with consistent filenames that will be overridden on modify our rainforest github workflow to run tests without a custom start url update existing tests to use the pixiebrix extension site and start at the environment start url | 1 |
6,884 | 6,654,524,516 | IssuesEvent | 2017-09-29 13:11:19 | Teemperor/ClangModulesCMake | https://api.github.com/repos/Teemperor/ClangModulesCMake | closed | Test libraries in a user-specified order, not in the order of the include directories. | infrastructure | E.g. we should always test/mount STL before third-party libraries. Currently we do it in whatever way the includes are coming in. | 1.0 | Test libraries in a user-specified order, not in the order of the include directories. - E.g. we should always test/mount STL before third-party libraries. Currently we do it in whatever way the includes are coming in. | infrastructure | test libraries in a user specified order not in the order of the include directories e g we should always test mount stl before third party libraries currently we do it in whatever way the includes are coming in | 1 |
18,255 | 12,875,880,953 | IssuesEvent | 2020-07-11 01:14:35 | wix/yoshi | https://api.github.com/repos/wix/yoshi | closed | indeterministic error in legacy tests: port 3200 is in used | 💤 Stale 🛠 Infrastructure | - non-production issue
# 🐛 Bug Report
run on your local machine (not in CI), `yarn run test:legacy:test` at the root of the project.
I'm sorry that i don't have logs or anything else to share. it sometimes happen.
### 🔦 Context <!-- optional -->
while debugging/working on the project locally.
### 💁 Possible Solution <!-- optional -->
a quick fix is to manually close the process that the port belongs to.
I didn't spend the time to investigate what is the goal/responsibilities of the service with the taken port so i don't have a good permanent solution.
but if it helps, these are the ways i'm fixing this problem in my projects:
- if it's a child process: try to set the port to `0` and extract all the listening ports from the child-process (https://www.npmjs.com/package/find-process)
- if it's on the main process (test-process): if it's possible in your scenario, create a mock-server that will be able to give you the port.
### 🌍 Your Environment
yoshi: `v4.61.2`
node: `v12.16.1` | 1.0 | indeterministic error in legacy tests: port 3200 is in used - - non-production issue
# 🐛 Bug Report
run on your local machine (not in CI), `yarn run test:legacy:test` at the root of the project.
I'm sorry that i don't have logs or anything else to share. it sometimes happen.
### 🔦 Context <!-- optional -->
while debugging/working on the project locally.
### 💁 Possible Solution <!-- optional -->
a quick fix is to manually close the process that the port belongs to.
I didn't spend the time to investigate what is the goal/responsibilities of the service with the taken port so i don't have a good permanent solution.
but if it helps, these are the ways i'm fixing this problem in my projects:
- if it's a child process: try to set the port to `0` and extract all the listening ports from the child-process (https://www.npmjs.com/package/find-process)
- if it's on the main process (test-process): if it's possible in your scenario, create a mock-server that will be able to give you the port.
### 🌍 Your Environment
yoshi: `v4.61.2`
node: `v12.16.1` | infrastructure | indeterministic error in legacy tests port is in used non production issue 🐛 bug report run on your local machine not in ci yarn run test legacy test at the root of the project i m sorry that i don t have logs or anything else to share it sometimes happen 🔦 context while debugging working on the project locally 💁 possible solution a quick fix is to manually close the process that the port belongs to i didn t spend the time to investigate what is the goal responsibilities of the service with the taken port so i don t have a good permanent solution but if it helps these are the ways i m fixing this problem in my projects if it s a child process try to set the port to and extract all the listening ports from the child process if it s on the main process test process if it s possible in your scenario create a mock server that will be able to give you the port 🌍 your environment yoshi node | 1 |
15,230 | 11,423,462,439 | IssuesEvent | 2020-02-03 15:56:58 | SteeltoeOSS/steeltoe | https://api.github.com/repos/SteeltoeOSS/steeltoe | closed | CI Builds should pass for ALL PRs | Infrastructure | Due to the way GemFire is currently distributed, the CI build process has to use secrets, which results in failures on all CI builds from non-team members. PR builds should be made to pass for _all_ PRs, not just core Steeltoe team members | 1.0 | CI Builds should pass for ALL PRs - Due to the way GemFire is currently distributed, the CI build process has to use secrets, which results in failures on all CI builds from non-team members. PR builds should be made to pass for _all_ PRs, not just core Steeltoe team members | infrastructure | ci builds should pass for all prs due to the way gemfire is currently distributed the ci build process has to use secrets which results in failures on all ci builds from non team members pr builds should be made to pass for all prs not just core steeltoe team members | 1 |
852 | 2,964,247,202 | IssuesEvent | 2015-07-10 15:35:33 | RIOT-OS/RIOT | https://api.github.com/repos/RIOT-OS/RIOT | opened | Proposal for directory structure for automated tests | CI-Infrastructure Request for Comments tests | Define a directory under `$RIOTBASE/tests/` and outside `tests/unittests` where sources for automated tests can be placed. It has turned out in discussions [#3296](https://github.com/RIOT-OS/RIOT/pull/3296#issuecomment-118338031), [#3307](https://github.com/RIOT-OS/RIOT/pull/3307#discussion_r33860255) that the established *unittests* becomes insufficent for extended testing.
Motivation
----------
The agreement of a seperate directory structure should face the following difficulties.
### tests/unittests ###
The tests in this directory are build together in one blob. This is implicating some weaknesses.
+ With the increasing number of test cases the memory ressources of some boards are exhausted.
+ It prevents the desired isolation of the running test cases.
+ Necessary control of preconditions can not be archieved for test sequences that would require reinitializing (e.g `ng_ipv6_init()`)
### tests ###
The directory is, beside of *unittests* home of several more or less freely named subdirectories for different test scenarios. It is not adviseable to mix the automated tests inbetween.
+ Increases naming collisions or encourages creative naming to prevent them.
+ Makes it difficult to keep track of locations.
Proposal
--------
The idea is to name name subdirectory of tests(lets say *'automated'*) and put the tests in subdirectories. These may reflect the location of either the headers or the units source files to follow the "principle of least suprise".
Every *unit* or *test object* has further subdirs for sepeated test which require reflashing, etc...
### Example for resulting structure ###
(Reflecting loctions of C files)
tests/automated/core/atomic/
tests/automated/core/...
tests/automated/core/thread/
tests/automated/...
tests/automated/sys/auto_init/auto_init/
tests/automated/sys/...
tests/automated/sys/net/application_layer/ng_zep/ng_zep/
tests/automated/sys/net/...
tests/automated/sys/net/transport_layer/ng_udp/ng_udp/
| 1.0 | Proposal for directory structure for automated tests - Define a directory under `$RIOTBASE/tests/` and outside `tests/unittests` where sources for automated tests can be placed. It has turned out in discussions [#3296](https://github.com/RIOT-OS/RIOT/pull/3296#issuecomment-118338031), [#3307](https://github.com/RIOT-OS/RIOT/pull/3307#discussion_r33860255) that the established *unittests* becomes insufficent for extended testing.
Motivation
----------
The agreement of a seperate directory structure should face the following difficulties.
### tests/unittests ###
The tests in this directory are build together in one blob. This is implicating some weaknesses.
+ With the increasing number of test cases the memory ressources of some boards are exhausted.
+ It prevents the desired isolation of the running test cases.
+ Necessary control of preconditions can not be archieved for test sequences that would require reinitializing (e.g `ng_ipv6_init()`)
### tests ###
The directory is, beside of *unittests* home of several more or less freely named subdirectories for different test scenarios. It is not adviseable to mix the automated tests inbetween.
+ Increases naming collisions or encourages creative naming to prevent them.
+ Makes it difficult to keep track of locations.
Proposal
--------
The idea is to name name subdirectory of tests(lets say *'automated'*) and put the tests in subdirectories. These may reflect the location of either the headers or the units source files to follow the "principle of least suprise".
Every *unit* or *test object* has further subdirs for sepeated test which require reflashing, etc...
### Example for resulting structure ###
(Reflecting loctions of C files)
tests/automated/core/atomic/
tests/automated/core/...
tests/automated/core/thread/
tests/automated/...
tests/automated/sys/auto_init/auto_init/
tests/automated/sys/...
tests/automated/sys/net/application_layer/ng_zep/ng_zep/
tests/automated/sys/net/...
tests/automated/sys/net/transport_layer/ng_udp/ng_udp/
| infrastructure | proposal for directory structure for automated tests define a directory under riotbase tests and outside tests unittests where sources for automated tests can be placed it has turned out in discussions that the established unittests becomes insufficent for extended testing motivation the agreement of a seperate directory structure should face the following difficulties tests unittests the tests in this directory are build together in one blob this is implicating some weaknesses with the increasing number of test cases the memory ressources of some boards are exhausted it prevents the desired isolation of the running test cases necessary control of preconditions can not be archieved for test sequences that would require reinitializing e g ng init tests the directory is beside of unittests home of several more or less freely named subdirectories for different test scenarios it is not adviseable to mix the automated tests inbetween increases naming collisions or encourages creative naming to prevent them makes it difficult to keep track of locations proposal the idea is to name name subdirectory of tests lets say automated and put the tests in subdirectories these may reflect the location of either the headers or the units source files to follow the principle of least suprise every unit or test object has further subdirs for sepeated test which require reflashing etc example for resulting structure reflecting loctions of c files tests automated core atomic tests automated core tests automated core thread tests automated tests automated sys auto init auto init tests automated sys tests automated sys net application layer ng zep ng zep tests automated sys net tests automated sys net transport layer ng udp ng udp | 1 |
90,956 | 26,227,724,748 | IssuesEvent | 2023-01-04 20:23:54 | chaotic-aur/packages | https://api.github.com/repos/chaotic-aur/packages | closed | [Request] KCM modules | request:new-pkg interfere required priority:low bug:PKGBUILD | ### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/systemd-kcm
https://aur.archlinux.org/packages/kcm-polkit-kde-git
https://aur.archlinux.org/packages/fancontrol-kcm
https://aur.archlinux.org/packages/kcm-uefi
https://aur.archlinux.org/packages/kcm-grub2-git
https://aur.archlinux.org/packages/kcm-colorful-git
https://aur.archlinux.org/packages/fancontrol-kcm-git
https://aur.archlinux.org/packages/jackman-kcm-git
### Utility this package has for you
Managing more setting at the same place.
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a few.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_ | 1.0 | [Request] KCM modules - ### Link to the package(s) in the AUR
https://aur.archlinux.org/packages/systemd-kcm
https://aur.archlinux.org/packages/kcm-polkit-kde-git
https://aur.archlinux.org/packages/fancontrol-kcm
https://aur.archlinux.org/packages/kcm-uefi
https://aur.archlinux.org/packages/kcm-grub2-git
https://aur.archlinux.org/packages/kcm-colorful-git
https://aur.archlinux.org/packages/fancontrol-kcm-git
https://aur.archlinux.org/packages/jackman-kcm-git
### Utility this package has for you
Managing more setting at the same place.
### Do you consider the package(s) to be useful for every Chaotic-AUR user?
No, but for a few.
### Do you consider the package to be useful for feature testing/preview?
- [ ] Yes
### Have you tested if the package builds in a clean chroot?
- [ ] Yes
### Does the package's license allow redistributing it?
YES!
### Have you searched the issues to ensure this request is unique?
- [X] YES!
### Have you read the README to ensure this package is not banned?
- [X] YES!
### More information
_No response_ | non_infrastructure | kcm modules link to the package s in the aur utility this package has for you managing more setting at the same place do you consider the package s to be useful for every chaotic aur user no but for a few do you consider the package to be useful for feature testing preview yes have you tested if the package builds in a clean chroot yes does the package s license allow redistributing it yes have you searched the issues to ensure this request is unique yes have you read the readme to ensure this package is not banned yes more information no response | 0 |
417,298 | 12,158,198,798 | IssuesEvent | 2020-04-26 02:29:06 | rubyforgood/casa | https://api.github.com/repos/rubyforgood/casa | closed | Remove `other` from case_contact contact_type dropdown | :raised_hands: Volunteer Priority: Medium Status: Available help wanted | Part of epic #3 (volunteer dashboard)
`other` should be removed from the dropdown menu under `contact_type` when creating a new `case_contact`
<a href="https://casa-r4g-staging.herokuapp.com/case_contacts/new">see here for reference on staging</a> | 1.0 | Remove `other` from case_contact contact_type dropdown - Part of epic #3 (volunteer dashboard)
`other` should be removed from the dropdown menu under `contact_type` when creating a new `case_contact`
<a href="https://casa-r4g-staging.herokuapp.com/case_contacts/new">see here for reference on staging</a> | non_infrastructure | remove other from case contact contact type dropdown part of epic volunteer dashboard other should be removed from the dropdown menu under contact type when creating a new case contact | 0 |
108,578 | 23,631,659,708 | IssuesEvent | 2022-08-25 09:49:52 | arduino/arduino-ide | https://api.github.com/repos/arduino/arduino-ide | closed | Only the IP address is shown for unrecognized network boards | topic: code type: imperfection topic: CLI | ### Describe the problem
This is the follow-up of #1327.
Arduino IDE 2.x shows only the IP address of connected but unrecognized boards. Before #1327, `network` boards not recognized by the Arduino CLI were not listed in the "Board Selector" menu. See the changes [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-fb37f20bea00881acee3aafddb1ecefcecf41ce59845ca1510da79e918ee0837R332-R342) and [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-e42c82bb67e277cfa4598239952afd65db44dba55dc7d68df619dfccfa648279R409-R423).
#1327 will remove the above-mentioned filtering, and the Arduino IDE will show all `network` boards. Unfortunately, the IDE receives insufficient information from the CLI, so the IDE has no chance to show anything else but the IP.
<img width="313" alt="Screen Shot 2022-08-17 at 14 41 58" src="https://user-images.githubusercontent.com/1405703/185129514-5d1d6c10-7bcf-4530-b42e-4420738b8979.png">
<img width="637" alt="Screen Shot 2022-08-17 at 14 41 40" src="https://user-images.githubusercontent.com/1405703/185129587-c3a22d16-1e2d-47c5-936f-834c0ce53f56.png">
<img width="281" alt="Screen Shot 2022-08-17 at 14 41 23" src="https://user-images.githubusercontent.com/1405703/185129661-a4a7a903-5e92-431c-af7b-b9b1e210b872.png">
The `mdns-discovery` is aware of these additional things; hence the Arduino IDE 1.x can enhance the label of an unrecognized `network` board:
```
HELLO "ciao" 1
{
"eventType": "hello",
"message": "Invalid HELLO command",
"error": true
}
HELLO 1 "ciao"
{
"eventType": "hello",
"message": "OK",
"protocolVersion": 1
}
START_SYNC
{
"eventType": "start_sync",
"message": "OK"
}
{
"eventType": "add",
"port": {
"address": "10.0.1.25",
"label": "esp32s2 at 10.0.1.25",
"protocol": "network",
"protocolLabel": "Network Port",
"properties": {
".": "esp32s2usb",
"auth_upload": "yes",
"board": "esp32s2usb",
"hostname": "esp32s2.local.",
"port": "3232",
"ssh_upload": "no",
"tcp_check": "no"
}
}
}
```
The same information is missing from the Arduino CLI board watch (`board list -w --format json`):
```
{
"type": "add",
"address": "/dev/cu.usbmodem03",
"protocol": "serial",
"protocol_label": "Serial Port (USB)"
}
{
"type": "add",
"address": "10.0.1.25",
"protocol": "network",
"protocol_label": "Network Port"
}
```
So the IDE fails to show `esp32s2 at 10.0.1.25` and falls back to `Unknown 10.0.1.25`.
### To reproduce
The simplest way is to try it with Arduino CLI:
```
./arduino-cli board list -w --format json
```
### Expected behavior
Arduino IDE should receive the properties from the CLI discovered by the `mdns-discovery`.
```json
{
"properties": {
".": "esp32s2usb",
"auth_upload": "yes",
"board": "esp32s2usb",
"hostname": "esp32s2.local.",
"port": "3232",
"ssh_upload": "no",
"tcp_check": "no"
}
}
```
Or the IDE should receive the `"label": "esp32s2 at 10.0.1.25",` from the CLI.
### Arduino IDE version
9e2b73a
### Operating system
macOS
### Operating system version
12.3.1
### Additional context
Related (if not the same): https://github.com/arduino/arduino-ide/issues/567
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=)
- [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds)
- [X] My report contains all necessary details | 1.0 | Only the IP address is shown for unrecognized network boards - ### Describe the problem
This is the follow-up of #1327.
Arduino IDE 2.x shows only the IP address of connected but unrecognized boards. Before #1327, `network` boards not recognized by the Arduino CLI were not listed in the "Board Selector" menu. See the changes [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-fb37f20bea00881acee3aafddb1ecefcecf41ce59845ca1510da79e918ee0837R332-R342) and [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-e42c82bb67e277cfa4598239952afd65db44dba55dc7d68df619dfccfa648279R409-R423).
#1327 will remove the above-mentioned filtering, and the Arduino IDE will show all `network` boards. Unfortunately, the IDE receives insufficient information from the CLI, so the IDE has no chance to show anything else but the IP.
<img width="313" alt="Screen Shot 2022-08-17 at 14 41 58" src="https://user-images.githubusercontent.com/1405703/185129514-5d1d6c10-7bcf-4530-b42e-4420738b8979.png">
<img width="637" alt="Screen Shot 2022-08-17 at 14 41 40" src="https://user-images.githubusercontent.com/1405703/185129587-c3a22d16-1e2d-47c5-936f-834c0ce53f56.png">
<img width="281" alt="Screen Shot 2022-08-17 at 14 41 23" src="https://user-images.githubusercontent.com/1405703/185129661-a4a7a903-5e92-431c-af7b-b9b1e210b872.png">
The `mdns-discovery` is aware of these additional things; hence the Arduino IDE 1.x can enhance the label of an unrecognized `network` board:
```
HELLO "ciao" 1
{
"eventType": "hello",
"message": "Invalid HELLO command",
"error": true
}
HELLO 1 "ciao"
{
"eventType": "hello",
"message": "OK",
"protocolVersion": 1
}
START_SYNC
{
"eventType": "start_sync",
"message": "OK"
}
{
"eventType": "add",
"port": {
"address": "10.0.1.25",
"label": "esp32s2 at 10.0.1.25",
"protocol": "network",
"protocolLabel": "Network Port",
"properties": {
".": "esp32s2usb",
"auth_upload": "yes",
"board": "esp32s2usb",
"hostname": "esp32s2.local.",
"port": "3232",
"ssh_upload": "no",
"tcp_check": "no"
}
}
}
```
The same information is missing from the Arduino CLI board watch (`board list -w --format json`):
```
{
"type": "add",
"address": "/dev/cu.usbmodem03",
"protocol": "serial",
"protocol_label": "Serial Port (USB)"
}
{
"type": "add",
"address": "10.0.1.25",
"protocol": "network",
"protocol_label": "Network Port"
}
```
So the IDE fails to show `esp32s2 at 10.0.1.25` and falls back to `Unknown 10.0.1.25`.
### To reproduce
The simplest way is to try it with Arduino CLI:
```
./arduino-cli board list -w --format json
```
### Expected behavior
Arduino IDE should receive the properties from the CLI discovered by the `mdns-discovery`.
```json
{
"properties": {
".": "esp32s2usb",
"auth_upload": "yes",
"board": "esp32s2usb",
"hostname": "esp32s2.local.",
"port": "3232",
"ssh_upload": "no",
"tcp_check": "no"
}
}
```
Or the IDE should receive the `"label": "esp32s2 at 10.0.1.25",` from the CLI.
### Arduino IDE version
9e2b73a
### Operating system
macOS
### Operating system version
12.3.1
### Additional context
Related (if not the same): https://github.com/arduino/arduino-ide/issues/567
### Issue checklist
- [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=)
- [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds)
- [X] My report contains all necessary details | non_infrastructure | only the ip address is shown for unrecognized network boards describe the problem this is the follow up of arduino ide x shows only the ip address of connected but unrecognized boards before network boards not recognized by the arduino cli were not listed in the board selector menu see the changes and will remove the above mentioned filtering and the arduino ide will show all network boards unfortunately the ide receives insufficient information from the cli so the ide has no chance to show anything else but the ip img width alt screen shot at src img width alt screen shot at src img width alt screen shot at src the mdns discovery is aware of these additional things hence the arduino ide x can enhance the label of an unrecognized network board hello ciao eventtype hello message invalid hello command error true hello ciao eventtype hello message ok protocolversion start sync eventtype start sync message ok eventtype add port address label at protocol network protocollabel network port properties auth upload yes board hostname local port ssh upload no tcp check no the same information is missing from the arduino cli board watch board list w format json type add address dev cu protocol serial protocol label serial port usb type add address protocol network protocol label network port so the ide fails to show at and falls back to unknown to reproduce the simplest way is to try it with arduino cli arduino cli board list w format json expected behavior arduino ide should receive the properties from the cli discovered by the mdns discovery json properties auth upload yes board hostname local port ssh upload no tcp check no or the ide should receive the label at from the cli arduino ide version operating system macos operating system version additional context related if not the same issue checklist i searched for previous reports in i verified the problem still occurs when using the latest my report contains all necessary details | 0 |
431,144 | 12,475,699,316 | IssuesEvent | 2020-05-29 12:08:05 | borgbase/vorta | https://api.github.com/repos/borgbase/vorta | closed | Pass through borg error message for all commands | priority:medium type:enhancement | **How to reproduce:**
* Remove your rights on your borg cache files (this stems from running borg with sudo)
* Try to add existing repo with vorta
**Result:**
An error message is displayed extremely short but is immediately replaced with a generic error message. Here a popup of the original log would be helpful. (Example: "Borg cli failed and output the following: ...")
**Workaround:**
Look inside the log-file, here you can clearly see the error | 1.0 | Pass through borg error message for all commands - **How to reproduce:**
* Remove your rights on your borg cache files (this stems from running borg with sudo)
* Try to add existing repo with vorta
**Result:**
An error message is displayed extremely short but is immediately replaced with a generic error message. Here a popup of the original log would be helpful. (Example: "Borg cli failed and output the following: ...")
**Workaround:**
Look inside the log-file, here you can clearly see the error | non_infrastructure | pass through borg error message for all commands how to reproduce remove your rights on your borg cache files this stems from running borg with sudo try to add existing repo with vorta result an error message is displayed extremely short but is immediately replaced with a generic error message here a popup of the original log would be helpful example borg cli failed and output the following workaround look inside the log file here you can clearly see the error | 0 |
302,434 | 26,143,864,441 | IssuesEvent | 2022-12-29 23:37:18 | fuatakgun/eufy_security | https://api.github.com/repos/fuatakgun/eufy_security | closed | Front Door Lock not updating | entity: lock type: Bug next step: waiting for response next step: retest with alpha | When I unlock it, it unlocks but doesnt update in home assistant. | 1.0 | Front Door Lock not updating - When I unlock it, it unlocks but doesnt update in home assistant. | non_infrastructure | front door lock not updating when i unlock it it unlocks but doesnt update in home assistant | 0 |
34,504 | 30,030,538,341 | IssuesEvent | 2023-06-27 09:17:45 | SonarSource/sonar-scanner-msbuild | https://api.github.com/repos/SonarSource/sonar-scanner-msbuild | closed | Reuse mend.yml template for downloading and running mend | Type: Improvement Area: Infrastructure | There is a template that already handles the download and the execution of mend. We should reuse it to avoid duplication and be able to do changes across repositories.
It can be found here: https://dev.azure.com/sonarsource/DotNetTeam%20Project/_git/pipelines-yaml-templates?path=/mend.yml&version=GTv2.0.0
Example of usage:
https://github.com/SonarSource/sonarlint-omnisharp/blob/7ce958a943c2bbe6fe38136a4529360e26292bd1/azure-pipelines.yml#L93-L102
We will have to update the `Mend-Scan.ps1` script or maybe inline it in the main yaml file. | 1.0 | Reuse mend.yml template for downloading and running mend - There is a template that already handles the download and the execution of mend. We should reuse it to avoid duplication and be able to do changes across repositories.
It can be found here: https://dev.azure.com/sonarsource/DotNetTeam%20Project/_git/pipelines-yaml-templates?path=/mend.yml&version=GTv2.0.0
Example of usage:
https://github.com/SonarSource/sonarlint-omnisharp/blob/7ce958a943c2bbe6fe38136a4529360e26292bd1/azure-pipelines.yml#L93-L102
We will have to update the `Mend-Scan.ps1` script or maybe inline it in the main yaml file. | infrastructure | reuse mend yml template for downloading and running mend there is a template that already handles the download and the execution of mend we should reuse it to avoid duplication and be able to do changes across repositories it can be found here example of usage we will have to update the mend scan script or maybe inline it in the main yaml file | 1 |
282,531 | 30,889,357,958 | IssuesEvent | 2023-08-04 02:36:20 | madhans23/linux-4.1.15 | https://api.github.com/repos/madhans23/linux-4.1.15 | reopened | CVE-2017-7482 (High) detected in linux-stable-rtv4.1.33 | Mend: dependency security vulnerability | ## CVE-2017-7482 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before version 4.12, Kerberos 5 tickets decoded when using the RXRPC keys incorrectly assumes the size of a field. This could lead to the size-remaining variable wrapping and the data pointer going over the end of the buffer. This could possibly lead to memory corruption and possible privilege escalation.
<p>Publish Date: 2018-07-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7482>CVE-2017-7482</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-7482">https://nvd.nist.gov/vuln/detail/CVE-2017-7482</a></p>
<p>Release Date: 2018-07-30</p>
<p>Fix Resolution: 4.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-7482 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-7482 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
In the Linux kernel before version 4.12, Kerberos 5 tickets decoded when using the RXRPC keys incorrectly assumes the size of a field. This could lead to the size-remaining variable wrapping and the data pointer going over the end of the buffer. This could possibly lead to memory corruption and possible privilege escalation.
<p>Publish Date: 2018-07-30
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7482>CVE-2017-7482</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-7482">https://nvd.nist.gov/vuln/detail/CVE-2017-7482</a></p>
<p>Release Date: 2018-07-30</p>
<p>Fix Resolution: 4.12</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files net rxrpc ar key c net rxrpc ar key c vulnerability details in the linux kernel before version kerberos tickets decoded when using the rxrpc keys incorrectly assumes the size of a field this could lead to the size remaining variable wrapping and the data pointer going over the end of the buffer this could possibly lead to memory corruption and possible privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
1,681 | 3,332,737,864 | IssuesEvent | 2015-11-11 21:31:08 | codeforamerica/communities | https://api.github.com/repos/codeforamerica/communities | closed | Meetup Pro | brigade donation infrastructure network growth | Meetup Pro looks like it will have everything we want. The main requests we had were:
* API access to member growth over time.
* Better attendance or email data
The costs looks to be about $10k a year. Its worth it.
Here is part of an email from them.
```
I also wanted to send over the list of data points we're building out, which will be available through our API within the next couple of months. Let me know if there are any data points you'd be interested to see that aren't included on the list!
You'll be able to view the following information about your MUGs:
- name
- description
- category
- start date
- total members
- average age
- gender ratio
- topics
- # past events
- # past RSVPs
- avg. # RSVPs/event
- # repeat RSVPers
- date of last scheduled MUP
- date of next scheduled MUP
- date joined Meetup Pro
``` | 1.0 | Meetup Pro - Meetup Pro looks like it will have everything we want. The main requests we had were:
* API access to member growth over time.
* Better attendance or email data
The costs looks to be about $10k a year. Its worth it.
Here is part of an email from them.
```
I also wanted to send over the list of data points we're building out, which will be available through our API within the next couple of months. Let me know if there are any data points you'd be interested to see that aren't included on the list!
You'll be able to view the following information about your MUGs:
- name
- description
- category
- start date
- total members
- average age
- gender ratio
- topics
- # past events
- # past RSVPs
- avg. # RSVPs/event
- # repeat RSVPers
- date of last scheduled MUP
- date of next scheduled MUP
- date joined Meetup Pro
``` | infrastructure | meetup pro meetup pro looks like it will have everything we want the main requests we had were api access to member growth over time better attendance or email data the costs looks to be about a year its worth it here is part of an email from them i also wanted to send over the list of data points we re building out which will be available through our api within the next couple of months let me know if there are any data points you d be interested to see that aren t included on the list you ll be able to view the following information about your mugs name description category start date total members average age gender ratio topics past events past rsvps avg rsvps event repeat rsvpers date of last scheduled mup date of next scheduled mup date joined meetup pro | 1 |
1,058 | 3,024,884,396 | IssuesEvent | 2015-08-03 01:53:47 | catapult-project/catapult | https://api.github.com/repos/catapult-project/catapult | opened | python tests fail on windows | Infrastructure | <a href="https://github.com/anniesullie"><img src="https://avatars.githubusercontent.com/u/4331931?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [anniesullie](https://github.com/anniesullie)**
_Friday Jul 10, 2015 at 20:06 GMT_
_Originally opened as https://github.com/google/trace-viewer/issues/1093_
----
After https://codereview.appspot.com/251100043/, there are still 3 failing tests. I'm not sure how to handle c:\ getting stuck on the absolute paths in Windows. Output:
======================================================================
ERROR: testInlineStylesheetURLs (tvcm.html_module_unittest.HTMLModuleTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module_unittest.py", line 359, in testInlineStylesheetURLs
my_component.AppendDirectlyDependentFilenamesTo(computed_deps)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 89, in AppendDirectlyDependentFilenamesTo
self.loader, module_dirname, contents)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__
self._Load(containing_dirname)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load
self._images = [resolve_url(x) for x in matches]
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url
image = self.loader.LoadImage(abs_path)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage
raise module.DepsException("url('%s') did not exist" % abs_path)
DepsException: While loading:
a.b.my_component
Got: url('c:\tmp\a\something.jpg') did not exist
======================================================================
ERROR: testGetAllDependentFilenamesRecursive (tvcm.module_unittest.ModuleIntegrationTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module_unittest.py", line 118, in testGetAllDependentFilenamesRecursive
my_module = loader.LoadModule(module_name='z.foo')
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 140, in LoadModule
m.Load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 38, in Load
super(HTMLModule, self).Load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module.py", line 187, in Load
style_sheet = self.loader.LoadStyleSheet(name)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet
style_sheet.load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load
self._InitParsedStyleSheetIfNeeded()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded
self.loader, module_dirname, self.contents)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__
self._Load(containing_dirname)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load
self._images = [resolve_url(x) for x in matches]
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url
image = self.loader.LoadImage(abs_path)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage
raise module.DepsException("url('%s') did not exist" % abs_path)
DepsException: While loading:
z.foo
z.foo.css
Got: url('c:\x\y\z\foo.jpeg') did not exist
======================================================================
ERROR: testImages (tvcm.style_sheet_unittest.StyleSheetUnittest)
----------------------------------------------------------------------
Traceback (most recent call last):
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet_unittest.py", line 29, in testImages
foo_x = loader.LoadStyleSheet('foo.x')
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet
style_sheet.load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load
self._InitParsedStyleSheetIfNeeded()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded
self.loader, module_dirname, self.contents)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__
self._Load(containing_dirname)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load
self._images = [resolve_url(x) for x in matches]
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url
image = self.loader.LoadImage(abs_path)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage
raise module.DepsException("url('%s') did not exist" % abs_path)
DepsException: While loading:
foo.x.css
Got: url('c:\src\images\bar.jpeg') did not exist
----------------------------------------------------------------------
Ran 87 tests in 11.067s
FAILED (errors=3, skipped=22)
| 1.0 | python tests fail on windows - <a href="https://github.com/anniesullie"><img src="https://avatars.githubusercontent.com/u/4331931?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [anniesullie](https://github.com/anniesullie)**
_Friday Jul 10, 2015 at 20:06 GMT_
_Originally opened as https://github.com/google/trace-viewer/issues/1093_
----
After https://codereview.appspot.com/251100043/, there are still 3 failing tests. I'm not sure how to handle c:\ getting stuck on the absolute paths in Windows. Output:
======================================================================
ERROR: testInlineStylesheetURLs (tvcm.html_module_unittest.HTMLModuleTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module_unittest.py", line 359, in testInlineStylesheetURLs
my_component.AppendDirectlyDependentFilenamesTo(computed_deps)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 89, in AppendDirectlyDependentFilenamesTo
self.loader, module_dirname, contents)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__
self._Load(containing_dirname)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load
self._images = [resolve_url(x) for x in matches]
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url
image = self.loader.LoadImage(abs_path)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage
raise module.DepsException("url('%s') did not exist" % abs_path)
DepsException: While loading:
a.b.my_component
Got: url('c:\tmp\a\something.jpg') did not exist
======================================================================
ERROR: testGetAllDependentFilenamesRecursive (tvcm.module_unittest.ModuleIntegrationTests)
----------------------------------------------------------------------
Traceback (most recent call last):
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module_unittest.py", line 118, in testGetAllDependentFilenamesRecursive
my_module = loader.LoadModule(module_name='z.foo')
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 140, in LoadModule
m.Load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 38, in Load
super(HTMLModule, self).Load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module.py", line 187, in Load
style_sheet = self.loader.LoadStyleSheet(name)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet
style_sheet.load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load
self._InitParsedStyleSheetIfNeeded()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded
self.loader, module_dirname, self.contents)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__
self._Load(containing_dirname)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load
self._images = [resolve_url(x) for x in matches]
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url
image = self.loader.LoadImage(abs_path)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage
raise module.DepsException("url('%s') did not exist" % abs_path)
DepsException: While loading:
z.foo
z.foo.css
Got: url('c:\x\y\z\foo.jpeg') did not exist
======================================================================
ERROR: testImages (tvcm.style_sheet_unittest.StyleSheetUnittest)
----------------------------------------------------------------------
Traceback (most recent call last):
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet_unittest.py", line 29, in testImages
foo_x = loader.LoadStyleSheet('foo.x')
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet
style_sheet.load()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load
self._InitParsedStyleSheetIfNeeded()
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded
self.loader, module_dirname, self.contents)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__
self._Load(containing_dirname)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load
self._images = [resolve_url(x) for x in matches]
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url
image = self.loader.LoadImage(abs_path)
File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage
raise module.DepsException("url('%s') did not exist" % abs_path)
DepsException: While loading:
foo.x.css
Got: url('c:\src\images\bar.jpeg') did not exist
----------------------------------------------------------------------
Ran 87 tests in 11.067s
FAILED (errors=3, skipped=22)
| infrastructure | python tests fail on windows issue by friday jul at gmt originally opened as after there are still failing tests i m not sure how to handle c getting stuck on the absolute paths in windows output error testinlinestylesheeturls tvcm html module unittest htmlmoduletests traceback most recent call last file c github trace viewer tracing third party tvcm tvcm html module unittest py line in testinlinestylesheeturls my component appenddirectlydependentfilenamesto computed deps file c github trace viewer tracing third party tvcm tvcm html module py line in appenddirectlydependentfilenamesto self loader module dirname contents file c github trace viewer tracing third party tvcm tvcm style sheet py line in init self load containing dirname file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self images file c github trace viewer tracing third party tvcm tvcm style sheet py line in resolve url image self loader loadimage abs path file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadimage raise module depsexception url s did not exist abs path depsexception while loading a b my component got url c tmp a something jpg did not exist error testgetalldependentfilenamesrecursive tvcm module unittest moduleintegrationtests traceback most recent call last file c github trace viewer tracing third party tvcm tvcm module unittest py line in testgetalldependentfilenamesrecursive my module loader loadmodule module name z foo file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadmodule m load file c github trace viewer tracing third party tvcm tvcm html module py line in load super htmlmodule self load file c github trace viewer tracing third party tvcm tvcm module py line in load style sheet self loader loadstylesheet name file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadstylesheet style sheet load file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self initparsedstylesheetifneeded file c github trace viewer tracing third party tvcm tvcm style sheet py line in initparsedstylesheetifneeded self loader module dirname self contents file c github trace viewer tracing third party tvcm tvcm style sheet py line in init self load containing dirname file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self images file c github trace viewer tracing third party tvcm tvcm style sheet py line in resolve url image self loader loadimage abs path file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadimage raise module depsexception url s did not exist abs path depsexception while loading z foo z foo css got url c x y z foo jpeg did not exist error testimages tvcm style sheet unittest stylesheetunittest traceback most recent call last file c github trace viewer tracing third party tvcm tvcm style sheet unittest py line in testimages foo x loader loadstylesheet foo x file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadstylesheet style sheet load file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self initparsedstylesheetifneeded file c github trace viewer tracing third party tvcm tvcm style sheet py line in initparsedstylesheetifneeded self loader module dirname self contents file c github trace viewer tracing third party tvcm tvcm style sheet py line in init self load containing dirname file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self images file c github trace viewer tracing third party tvcm tvcm style sheet py line in resolve url image self loader loadimage abs path file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadimage raise module depsexception url s did not exist abs path depsexception while loading foo x css got url c src images bar jpeg did not exist ran tests in failed errors skipped | 1 |
18,370 | 12,960,368,299 | IssuesEvent | 2020-07-20 14:16:33 | libero/reviewer | https://api.github.com/repos/libero/reviewer | closed | Integrate MECA to chart and umbrella tests | Infrastructure | Chart and staging env:
- [x] ingress in chart for /meca-result
- [x] add sftp and bucket envvar to chart
- [x] add sftp secret to staging deploy (ejp test env)
Umbrella repo:
- [x] add sftp container so that submission can handle when 'submit' button is clicked in browsertests
- [x] add envvars for meca stuff in umbrella compose file(s)
| 1.0 | Integrate MECA to chart and umbrella tests - Chart and staging env:
- [x] ingress in chart for /meca-result
- [x] add sftp and bucket envvar to chart
- [x] add sftp secret to staging deploy (ejp test env)
Umbrella repo:
- [x] add sftp container so that submission can handle when 'submit' button is clicked in browsertests
- [x] add envvars for meca stuff in umbrella compose file(s)
| infrastructure | integrate meca to chart and umbrella tests chart and staging env ingress in chart for meca result add sftp and bucket envvar to chart add sftp secret to staging deploy ejp test env umbrella repo add sftp container so that submission can handle when submit button is clicked in browsertests add envvars for meca stuff in umbrella compose file s | 1 |
91,701 | 18,676,614,623 | IssuesEvent | 2021-10-31 17:11:33 | CiviWiki/OpenCiviWiki | https://api.github.com/repos/CiviWiki/OpenCiviWiki | closed | [BUG]: Major bug fixes through the UI | bug community code quality documentation triage | ### Description
There are multiple routing issues and bugs which I am spotting through testing, this is a self-assigned issue and I'll elaborate on the issues I resolve here soon.
[Update 1]:
The following bugs have been spotted:
- [x] Reverse name maps for urls
- [x] Profile doesn't load from the header after login
- [x] Profile creation keeps loading infinitely
- [x] HTTP 403 for POST on profile setup
- [ ] The footer needs to be added to the bottom
| 1.0 | [BUG]: Major bug fixes through the UI - ### Description
There are multiple routing issues and bugs which I am spotting through testing, this is a self-assigned issue and I'll elaborate on the issues I resolve here soon.
[Update 1]:
The following bugs have been spotted:
- [x] Reverse name maps for urls
- [x] Profile doesn't load from the header after login
- [x] Profile creation keeps loading infinitely
- [x] HTTP 403 for POST on profile setup
- [ ] The footer needs to be added to the bottom
| non_infrastructure | major bug fixes through the ui description there are multiple routing issues and bugs which i am spotting through testing this is a self assigned issue and i ll elaborate on the issues i resolve here soon the following bugs have been spotted reverse name maps for urls profile doesn t load from the header after login profile creation keeps loading infinitely http for post on profile setup the footer needs to be added to the bottom | 0 |
18,047 | 4,230,354,727 | IssuesEvent | 2016-07-04 11:41:54 | centreon/centreon | https://api.github.com/repos/centreon/centreon | closed | Temps de disponibilité | Kind/Documentation | Bonjour,
J'ai quelque chose qui me parait bizarre au niveau des rapports sur le temps de disponibilité.
J'ai un groupe de 3 hôtes, j'ai eu une coupure sur l'un de ces 3 hôtes, le temps sur cette journée apparait comme ceci pour le groupe :

Cependant quand je vais sur mon hôte qui a eu la coupure le temps apparait comme ceci sur la même journée :

J'ai l'impression que le temps est divisé, ça ressemble à un bug ?
Si je ne me trompe pas, on aimerait en théorie avoir le temps d'indisponibilité du groupe, en additionnant les temps des hôtes ?
Cordialement,
Bastien
| 1.0 | Temps de disponibilité - Bonjour,
J'ai quelque chose qui me parait bizarre au niveau des rapports sur le temps de disponibilité.
J'ai un groupe de 3 hôtes, j'ai eu une coupure sur l'un de ces 3 hôtes, le temps sur cette journée apparait comme ceci pour le groupe :

Cependant quand je vais sur mon hôte qui a eu la coupure le temps apparait comme ceci sur la même journée :

J'ai l'impression que le temps est divisé, ça ressemble à un bug ?
Si je ne me trompe pas, on aimerait en théorie avoir le temps d'indisponibilité du groupe, en additionnant les temps des hôtes ?
Cordialement,
Bastien
| non_infrastructure | temps de disponibilité bonjour j ai quelque chose qui me parait bizarre au niveau des rapports sur le temps de disponibilité j ai un groupe de hôtes j ai eu une coupure sur l un de ces hôtes le temps sur cette journée apparait comme ceci pour le groupe cependant quand je vais sur mon hôte qui a eu la coupure le temps apparait comme ceci sur la même journée j ai l impression que le temps est divisé ça ressemble à un bug si je ne me trompe pas on aimerait en théorie avoir le temps d indisponibilité du groupe en additionnant les temps des hôtes cordialement bastien | 0 |
701 | 9,961,019,916 | IssuesEvent | 2019-07-06 22:45:13 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | HTTP2: Pending bytes from a previous frame may not be flushed if another pending write is canceled | area-System.Net.Http tenet-reliability | Consider two frames both trying to acquire the write lock concurrently. One of them gets it, the other has to wait and registers itself as a pending writer. The first frame then doesn't require itself to be flushed if there are pending writers, and so writes itself to the buffer but doesn't explicitly flush. Between the time that it checks for pending writers and does the write, however, the other frame is canceled, and the pending writers count is decremented back to 0. There's now data in the buffer that needed to be flushed, but there's no pending write that's going to do it.
We could "easily" address this by only counting a non-cancelable write as one that can be pending, but from HttpClient every request will use a cancelable token, and thus all header/data frame writes associated with that request will be cancelable.
A more robust solution is probably to queue a forced flush in the case where a pending waiter is canceled.
cc: @geoffkizer | True | HTTP2: Pending bytes from a previous frame may not be flushed if another pending write is canceled - Consider two frames both trying to acquire the write lock concurrently. One of them gets it, the other has to wait and registers itself as a pending writer. The first frame then doesn't require itself to be flushed if there are pending writers, and so writes itself to the buffer but doesn't explicitly flush. Between the time that it checks for pending writers and does the write, however, the other frame is canceled, and the pending writers count is decremented back to 0. There's now data in the buffer that needed to be flushed, but there's no pending write that's going to do it.
We could "easily" address this by only counting a non-cancelable write as one that can be pending, but from HttpClient every request will use a cancelable token, and thus all header/data frame writes associated with that request will be cancelable.
A more robust solution is probably to queue a forced flush in the case where a pending waiter is canceled.
cc: @geoffkizer | non_infrastructure | pending bytes from a previous frame may not be flushed if another pending write is canceled consider two frames both trying to acquire the write lock concurrently one of them gets it the other has to wait and registers itself as a pending writer the first frame then doesn t require itself to be flushed if there are pending writers and so writes itself to the buffer but doesn t explicitly flush between the time that it checks for pending writers and does the write however the other frame is canceled and the pending writers count is decremented back to there s now data in the buffer that needed to be flushed but there s no pending write that s going to do it we could easily address this by only counting a non cancelable write as one that can be pending but from httpclient every request will use a cancelable token and thus all header data frame writes associated with that request will be cancelable a more robust solution is probably to queue a forced flush in the case where a pending waiter is canceled cc geoffkizer | 0 |
2,204 | 3,560,165,307 | IssuesEvent | 2016-01-23 00:09:04 | Unidata/siphon | https://api.github.com/repos/Unidata/siphon | closed | nose -> py.test | Infrastructure | Nose is dying (https://nose.readthedocs.org/en/latest/index.html) and py.test seems to be what all the cool kids are using. Shouldn't take much to move over, but will be easier to do sooner rather than later. | 1.0 | nose -> py.test - Nose is dying (https://nose.readthedocs.org/en/latest/index.html) and py.test seems to be what all the cool kids are using. Shouldn't take much to move over, but will be easier to do sooner rather than later. | infrastructure | nose py test nose is dying and py test seems to be what all the cool kids are using shouldn t take much to move over but will be easier to do sooner rather than later | 1 |
11,736 | 9,403,696,453 | IssuesEvent | 2019-04-09 02:35:30 | hotosm/tasking-manager | https://api.github.com/repos/hotosm/tasking-manager | reopened | Host javascript resources ourselves | Difficulty: Easy Priority: High Status: Needs help Type: Infrastructure | Feature request from #886 discussion: host javascript resources ourselves instead of CDN.
| 1.0 | Host javascript resources ourselves - Feature request from #886 discussion: host javascript resources ourselves instead of CDN.
| infrastructure | host javascript resources ourselves feature request from discussion host javascript resources ourselves instead of cdn | 1 |
55,728 | 14,020,206,752 | IssuesEvent | 2020-10-29 19:19:24 | srivatsamarichi/ContosoAir | https://api.github.com/repos/srivatsamarichi/ContosoAir | opened | CVE-2018-19827 (High) detected in node-sass-4.14.1.tgz, opennmsopennms-source-25.1.0-1 | security vulnerability | ## CVE-2018-19827 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b>, <b>opennmsopennms-source-25.1.0-1</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: ContosoAir/package.json</p>
<p>Path to vulnerable library: ContosoAir/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/ContosoAir/commit/7e3d160bd69713f60688f97955fd688a3fe91b8f">7e3d160bd69713f60688f97955fd688a3fe91b8f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a use-after-free vulnerability exists in the SharedPtr class in SharedPtr.cpp (or SharedPtr.hpp) that may cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19827>CVE-2018-19827</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/pull/2784">https://github.com/sass/libsass/pull/2784</a></p>
<p>Release Date: 2019-08-29</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-19827 (High) detected in node-sass-4.14.1.tgz, opennmsopennms-source-25.1.0-1 - ## CVE-2018-19827 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b>, <b>opennmsopennms-source-25.1.0-1</b></p></summary>
<p>
<details><summary><b>node-sass-4.14.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p>
<p>Path to dependency file: ContosoAir/package.json</p>
<p>Path to vulnerable library: ContosoAir/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/ContosoAir/commit/7e3d160bd69713f60688f97955fd688a3fe91b8f">7e3d160bd69713f60688f97955fd688a3fe91b8f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a use-after-free vulnerability exists in the SharedPtr class in SharedPtr.cpp (or SharedPtr.hpp) that may cause a denial of service (application crash) or possibly have unspecified other impact.
<p>Publish Date: 2018-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19827>CVE-2018-19827</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/sass/libsass/pull/2784">https://github.com/sass/libsass/pull/2784</a></p>
<p>Release Date: 2019-08-29</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in node sass tgz opennmsopennms source cve high severity vulnerability vulnerable libraries node sass tgz opennmsopennms source node sass tgz wrapper around libsass library home page a href path to dependency file contosoair package json path to vulnerable library contosoair node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass a use after free vulnerability exists in the sharedptr class in sharedptr cpp or sharedptr hpp that may cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource | 0 |
5,064 | 5,415,194,997 | IssuesEvent | 2017-03-01 20:58:31 | Flatline4/Flatline4 | https://api.github.com/repos/Flatline4/Flatline4 | closed | Set up TSLint | infrastructure | It would be nice to have a linter in place to enforce certain rules are followed in the code base. The rules will need to be configured to our liking, and we will need to go through the existing code and make sure these rules are followed. Also, it would be nice to add running the linter as an npm script and make it run as part of the production build (the production build should not run if TSLint fails). | 1.0 | Set up TSLint - It would be nice to have a linter in place to enforce certain rules are followed in the code base. The rules will need to be configured to our liking, and we will need to go through the existing code and make sure these rules are followed. Also, it would be nice to add running the linter as an npm script and make it run as part of the production build (the production build should not run if TSLint fails). | infrastructure | set up tslint it would be nice to have a linter in place to enforce certain rules are followed in the code base the rules will need to be configured to our liking and we will need to go through the existing code and make sure these rules are followed also it would be nice to add running the linter as an npm script and make it run as part of the production build the production build should not run if tslint fails | 1 |
26,742 | 20,623,075,392 | IssuesEvent | 2022-03-07 19:25:42 | arduino/arduino-ide | https://api.github.com/repos/arduino/arduino-ide | closed | Redundant test build version indicators | topic: infrastructure type: imperfection | ## Describe the bug
Test builds are generated for every PR creation and push as well as pushes to the `main` branch. These make it easy for anyone to contribute to the project through beta testing.
It is important that the specific test build in use be easily identifiable. The Git commit hash is used for this purpose. For example, the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9ecff86bbefe3a34549e0e546d95ee1ff3439f97 is `2.0.0-rc3-snapshot-9ecff86`
🐛 The version indicator is now added twice to the test build versions. For example, the the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9b1f15def893f1f3df58e459e66a87c3fcdb930f is `2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d`
## To Reproduce
1. Open the list of `push` or `pull_request` event-triggered runs of the `Arduino IDE` workflow:
https://github.com/arduino/arduino-ide/actions/workflows/build.yml?query=event%3Apull_request+event%3Apush+is%3Asuccess
1. Click on any recent item in the list of runs.
For example: https://github.com/arduino/arduino-ide/actions/runs/1858133816
1. Under the "**Artifacts**" section of the page, click on the download link of the artifact for your operating system.
1. Wait for the download to finish.
1. Extract or install the downloaded file.
🐛 The extracted folder contains redundant version indicators (e.g., `arduino-ide_2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d_Windows_64bit`)
1. Start the Arduino IDE test build.
🐛 The window title contains redundant version indicators (e.g., "**sketch_feb17a | Arduino IDE 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d**")
1. Select **Help > About Arduino IDE** from the Arduino IDE menus.
1. Click the <kbd>Copy</kbd> button.
1. Paste the contents of the clipboard into any text editor.
🐛 The pasted text contains redundant version indicators.
For example:
```text
Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d
Date: 2022-02-17T10:40:50.003Z
CLI Version: 0.21.0 [10107d24]
Copyright © 2022 Arduino SA
```
## Expected behavior
The tester builds don't have redundant version indicators (e.g., `2.0.0-rc4-snapshot.9b1f15d`).
## Desktop
- OS: Windows 10, Ubuntu 20.04
- Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d
Date: 2022-02-17T10:40:50.003Z
CLI Version: 0.21.0 [10107d24]
## Additional context
The change was introduced at https://github.com/arduino/arduino-ide/commit/f660058c7550aaeeab81a6bc59f060d1ca2b901a
There was a previous change from the `snapshot.<hash>` (e.g., `2.0.0-rc3-snapshot.69ac1f4`) to `snapshot-<hash>` format (e.g., `2.0.0-rc3-snapshot-112153f`), which was introduced in https://github.com/arduino/arduino-ide/commit/112153fb965f63d952d126c8244cd3f84f0a1a1b
---
The additional characters use up some of [the precious allowance of path length doled out by Microsoft](https://github.com/arduino/arduino-ide/issues/106) for the beta testers on Windows.
---
Seeing this identifier in the filename redundantly, but with two different formats (note the `.` in one and `-` in the other) makes me nervous that it is an indication of something fragile. | 1.0 | Redundant test build version indicators - ## Describe the bug
Test builds are generated for every PR creation and push as well as pushes to the `main` branch. These make it easy for anyone to contribute to the project through beta testing.
It is important that the specific test build in use be easily identifiable. The Git commit hash is used for this purpose. For example, the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9ecff86bbefe3a34549e0e546d95ee1ff3439f97 is `2.0.0-rc3-snapshot-9ecff86`
🐛 The version indicator is now added twice to the test build versions. For example, the the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9b1f15def893f1f3df58e459e66a87c3fcdb930f is `2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d`
## To Reproduce
1. Open the list of `push` or `pull_request` event-triggered runs of the `Arduino IDE` workflow:
https://github.com/arduino/arduino-ide/actions/workflows/build.yml?query=event%3Apull_request+event%3Apush+is%3Asuccess
1. Click on any recent item in the list of runs.
For example: https://github.com/arduino/arduino-ide/actions/runs/1858133816
1. Under the "**Artifacts**" section of the page, click on the download link of the artifact for your operating system.
1. Wait for the download to finish.
1. Extract or install the downloaded file.
🐛 The extracted folder contains redundant version indicators (e.g., `arduino-ide_2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d_Windows_64bit`)
1. Start the Arduino IDE test build.
🐛 The window title contains redundant version indicators (e.g., "**sketch_feb17a | Arduino IDE 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d**")
1. Select **Help > About Arduino IDE** from the Arduino IDE menus.
1. Click the <kbd>Copy</kbd> button.
1. Paste the contents of the clipboard into any text editor.
🐛 The pasted text contains redundant version indicators.
For example:
```text
Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d
Date: 2022-02-17T10:40:50.003Z
CLI Version: 0.21.0 [10107d24]
Copyright © 2022 Arduino SA
```
## Expected behavior
The tester builds don't have redundant version indicators (e.g., `2.0.0-rc4-snapshot.9b1f15d`).
## Desktop
- OS: Windows 10, Ubuntu 20.04
- Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d
Date: 2022-02-17T10:40:50.003Z
CLI Version: 0.21.0 [10107d24]
## Additional context
The change was introduced at https://github.com/arduino/arduino-ide/commit/f660058c7550aaeeab81a6bc59f060d1ca2b901a
There was a previous change from the `snapshot.<hash>` (e.g., `2.0.0-rc3-snapshot.69ac1f4`) to `snapshot-<hash>` format (e.g., `2.0.0-rc3-snapshot-112153f`), which was introduced in https://github.com/arduino/arduino-ide/commit/112153fb965f63d952d126c8244cd3f84f0a1a1b
---
The additional characters use up some of [the precious allowance of path length doled out by Microsoft](https://github.com/arduino/arduino-ide/issues/106) for the beta testers on Windows.
---
Seeing this identifier in the filename redundantly, but with two different formats (note the `.` in one and `-` in the other) makes me nervous that it is an indication of something fragile. | infrastructure | redundant test build version indicators describe the bug test builds are generated for every pr creation and push as well as pushes to the main branch these make it easy for anyone to contribute to the project through beta testing it is important that the specific test build in use be easily identifiable the git commit hash is used for this purpose for example the version assigned to this commit is snapshot 🐛 the version indicator is now added twice to the test build versions for example the the version assigned to this commit is snapshot snapshot to reproduce open the list of push or pull request event triggered runs of the arduino ide workflow click on any recent item in the list of runs for example under the artifacts section of the page click on the download link of the artifact for your operating system wait for the download to finish extract or install the downloaded file 🐛 the extracted folder contains redundant version indicators e g arduino ide snapshot snapshot windows start the arduino ide test build 🐛 the window title contains redundant version indicators e g sketch arduino ide snapshot snapshot select help about arduino ide from the arduino ide menus click the copy button paste the contents of the clipboard into any text editor 🐛 the pasted text contains redundant version indicators for example text version snapshot snapshot date cli version copyright © arduino sa expected behavior the tester builds don t have redundant version indicators e g snapshot desktop os windows ubuntu version snapshot snapshot date cli version additional context the change was introduced at there was a previous change from the snapshot e g snapshot to snapshot format e g snapshot which was introduced in the additional characters use up some of for the beta testers on windows seeing this identifier in the filename redundantly but with two different formats note the in one and in the other makes me nervous that it is an indication of something fragile | 1 |
386,499 | 11,440,086,363 | IssuesEvent | 2020-02-05 08:56:33 | cheminfo/nmr-displayer | https://api.github.com/repos/cheminfo/nmr-displayer | closed | simplify data struture | Priority | Currently in the JSON and internally we have data1d and data2d.

This makes things quite complex in order to display the list of spectra but also in order to parse jcamp that may contain 1D and 2D (in the same jcamp).
We should have only ONE array containing all the data called `spectra`
@hamed-musallam Please could you do this refactoring and the change of all the examples ? (actually it will just be replace 'data1d' by 'spectra'
I will then improve the loading of spectra to deal with 'linked' jcamp. | 1.0 | simplify data struture - Currently in the JSON and internally we have data1d and data2d.

This makes things quite complex in order to display the list of spectra but also in order to parse jcamp that may contain 1D and 2D (in the same jcamp).
We should have only ONE array containing all the data called `spectra`
@hamed-musallam Please could you do this refactoring and the change of all the examples ? (actually it will just be replace 'data1d' by 'spectra'
I will then improve the loading of spectra to deal with 'linked' jcamp. | non_infrastructure | simplify data struture currently in the json and internally we have and this makes things quite complex in order to display the list of spectra but also in order to parse jcamp that may contain and in the same jcamp we should have only one array containing all the data called spectra hamed musallam please could you do this refactoring and the change of all the examples actually it will just be replace by spectra i will then improve the loading of spectra to deal with linked jcamp | 0 |
359,106 | 10,660,190,992 | IssuesEvent | 2019-10-18 09:24:09 | KamilSJaron/genomic-features-of-asexual-animals | https://api.github.com/repos/KamilSJaron/genomic-features-of-asexual-animals | closed | heterozygosity structure section | PRIORITY TODO help wanted | This section is a real pain in my butt. I find it less and less trustworthy, but at the same time, it's one of the very novel sections that is at least trying something new.
The original idea was to
1. get the heterozygosity structure of everything, look at those known (Meloidogyne hybrids; rotifers etc) and guess something about those that are unknown
We first had a problem with estimates in rotifers, as they were very off the reality. This is solved now thanks to @reubwn. The second problem was with controversy regarding the genome structure of root-knot nematodes (one of the reviewers did not like that we took for granted the 'one diverged hybrid copy present'). That does not affect the visualization, but it makes the interpretation of the plot harder.
However, with a fixed scale for rotifers, the others became unreadable - all squashed together.
<img width="851" alt="Screenshot 2019-10-10 at 11 00 30" src="https://user-images.githubusercontent.com/8181573/66559366-31154f00-eb4d-11e9-966b-d2670c8d9c3e.png">
Then we figured, that one needs to focus on relative sizes between different heterozygosity topologies anyway, so we divided each bar by it's size, getting
<img width="522" alt="Screenshot 2019-10-10 at 11 03 28" src="https://user-images.githubusercontent.com/8181573/66559641-9ec17b00-eb4d-11e9-97c2-b0f984de2510.png">
First I thought it's perfect, showing clearly both structure in triploids and tetraploids, while the absolute values are already shown in Figure 2. But then I realized that even the relative counts are absolute value dependent. We made the argument, that more equidistant genomic copies are in triploids, higher fraction of ABC loci there will be. However, even this expectation is dependent on the absolute value. I.e. how low heterozygosity organism, there will be very small fraction of ABC loci even if the genomic copies will be equidistant, contrasting to highly heterozygous species where even small divergence between the two more closely related haplotypes cause a relatively high proportion of ABC loci. I made some naive calculations for our species and it really does matter.
So, back to the original plot. However, now to make it more readable for triploids, I made it a two-panel plot for triploid and tetraploids. What do you think?

Aesthetics can be polished, the question is if the different scale axes are confusing or not. To make it more intuitive we can scale triploids from 0 to 10, so they appear visually smaller or something like that. Just to improve intuitive reading of the plot.
This is linked to, I am not sure anymore how to phrase this section. AGAIN! Suggestions welcomed. | 1.0 | heterozygosity structure section - This section is a real pain in my butt. I find it less and less trustworthy, but at the same time, it's one of the very novel sections that is at least trying something new.
The original idea was to
1. get the heterozygosity structure of everything, look at those known (Meloidogyne hybrids; rotifers etc) and guess something about those that are unknown
We first had a problem with estimates in rotifers, as they were very off the reality. This is solved now thanks to @reubwn. The second problem was with controversy regarding the genome structure of root-knot nematodes (one of the reviewers did not like that we took for granted the 'one diverged hybrid copy present'). That does not affect the visualization, but it makes the interpretation of the plot harder.
However, with a fixed scale for rotifers, the others became unreadable - all squashed together.
<img width="851" alt="Screenshot 2019-10-10 at 11 00 30" src="https://user-images.githubusercontent.com/8181573/66559366-31154f00-eb4d-11e9-966b-d2670c8d9c3e.png">
Then we figured, that one needs to focus on relative sizes between different heterozygosity topologies anyway, so we divided each bar by it's size, getting
<img width="522" alt="Screenshot 2019-10-10 at 11 03 28" src="https://user-images.githubusercontent.com/8181573/66559641-9ec17b00-eb4d-11e9-97c2-b0f984de2510.png">
First I thought it's perfect, showing clearly both structure in triploids and tetraploids, while the absolute values are already shown in Figure 2. But then I realized that even the relative counts are absolute value dependent. We made the argument, that more equidistant genomic copies are in triploids, higher fraction of ABC loci there will be. However, even this expectation is dependent on the absolute value. I.e. how low heterozygosity organism, there will be very small fraction of ABC loci even if the genomic copies will be equidistant, contrasting to highly heterozygous species where even small divergence between the two more closely related haplotypes cause a relatively high proportion of ABC loci. I made some naive calculations for our species and it really does matter.
So, back to the original plot. However, now to make it more readable for triploids, I made it a two-panel plot for triploid and tetraploids. What do you think?

Aesthetics can be polished, the question is if the different scale axes are confusing or not. To make it more intuitive we can scale triploids from 0 to 10, so they appear visually smaller or something like that. Just to improve intuitive reading of the plot.
This is linked to, I am not sure anymore how to phrase this section. AGAIN! Suggestions welcomed. | non_infrastructure | heterozygosity structure section this section is a real pain in my butt i find it less and less trustworthy but at the same time it s one of the very novel sections that is at least trying something new the original idea was to get the heterozygosity structure of everything look at those known meloidogyne hybrids rotifers etc and guess something about those that are unknown we first had a problem with estimates in rotifers as they were very off the reality this is solved now thanks to reubwn the second problem was with controversy regarding the genome structure of root knot nematodes one of the reviewers did not like that we took for granted the one diverged hybrid copy present that does not affect the visualization but it makes the interpretation of the plot harder however with a fixed scale for rotifers the others became unreadable all squashed together img width alt screenshot at src then we figured that one needs to focus on relative sizes between different heterozygosity topologies anyway so we divided each bar by it s size getting img width alt screenshot at src first i thought it s perfect showing clearly both structure in triploids and tetraploids while the absolute values are already shown in figure but then i realized that even the relative counts are absolute value dependent we made the argument that more equidistant genomic copies are in triploids higher fraction of abc loci there will be however even this expectation is dependent on the absolute value i e how low heterozygosity organism there will be very small fraction of abc loci even if the genomic copies will be equidistant contrasting to highly heterozygous species where even small divergence between the two more closely related haplotypes cause a relatively high proportion of abc loci i made some naive calculations for our species and it really does matter so back to the original plot however now to make it more readable for triploids i made it a two panel plot for triploid and tetraploids what do you think aesthetics can be polished the question is if the different scale axes are confusing or not to make it more intuitive we can scale triploids from to so they appear visually smaller or something like that just to improve intuitive reading of the plot this is linked to i am not sure anymore how to phrase this section again suggestions welcomed | 0 |
353,620 | 25,127,602,730 | IssuesEvent | 2022-11-09 12:58:06 | mhostetter/galois | https://api.github.com/repos/mhostetter/galois | closed | Project Structure | documentation | Hi @mhostetter, I'm opening this issue to better understand the structure of the `galois` package.
In case this should be redirected / moved to another issue, please let me know.
### Unused classes
1. Are the base classes in `galois/array.py` such as [`{Field|Group|Ring}ArrayBase`](https://github.com/mhostetter/galois/blob/49f8f1d8ef2ae5c2dc6aa92c8b0c18f6c8bc3829/galois/array.py#L241) being used? | 1.0 | Project Structure - Hi @mhostetter, I'm opening this issue to better understand the structure of the `galois` package.
In case this should be redirected / moved to another issue, please let me know.
### Unused classes
1. Are the base classes in `galois/array.py` such as [`{Field|Group|Ring}ArrayBase`](https://github.com/mhostetter/galois/blob/49f8f1d8ef2ae5c2dc6aa92c8b0c18f6c8bc3829/galois/array.py#L241) being used? | non_infrastructure | project structure hi mhostetter i m opening this issue to better understand the structure of the galois package in case this should be redirected moved to another issue please let me know unused classes are the base classes in galois array py such as being used | 0 |
417,157 | 12,156,252,800 | IssuesEvent | 2020-04-25 16:32:19 | rajan123456/uPresent | https://api.github.com/repos/rajan123456/uPresent | opened | Management: Allow setting Timezone for a module | enhancement low-priority | - [ ] : Management service should allow timezones in modules
- [ ] : Allow admin to set the timezone of the module from AdminUI.
- [ ] : Attendance to consider timezone of the module during module check. | 1.0 | Management: Allow setting Timezone for a module - - [ ] : Management service should allow timezones in modules
- [ ] : Allow admin to set the timezone of the module from AdminUI.
- [ ] : Attendance to consider timezone of the module during module check. | non_infrastructure | management allow setting timezone for a module management service should allow timezones in modules allow admin to set the timezone of the module from adminui attendance to consider timezone of the module during module check | 0 |
29,185 | 23,780,603,425 | IssuesEvent | 2022-09-02 04:03:41 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | ReflectionUtilities.Clone() fails with "SerializationException: Unable to find assembly" when using non-default AssemblyLoadContext | bug interface/infrastructure | ### Summary
When ApsimX's ReflectionUtilities.Clone() method is called on an instance of a type that belongs to an assembly that was not loaded in the default AssemblyLoadContext, it fails with `SerializationException: Unable to find assembly '...'.`
(Another possibility is that the assembly will also be found in the default load context, and Clone() will succeed, but then there will be an InvalidCastException when the caller of Clone() tries to cast the result.)
A fix is proposed in #7465.
### Background and analysis
ApsimX's [ReflectionUtilities.Clone()](https://github.com/APSIMInitiative/ApsimX/blob/265679cee30b0355e222f6a5832272ecc16f98da/APSIM.Shared/Utilities/ReflectionUtilities.cs#L544) method is implemented by serializing and deserializing using .NET's [BinaryFormatter](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.serialization.formatters.binary.binaryformatter).
When BinaryFormatter performs deserialization, it needs to be able to load assemblies.
.NET Core has something called "AssemblyLoadContext" ([documentation](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext)), which allows assemblies to be loaded in isolated contexts (e.g. in order to load multiple versions of an assembly).
By default, BinaryFormatter.Deserialize loads assemblies using [Assembly.Load(AssemblyName)](https://docs.microsoft.com/en-us/dotnet/api/system.reflection.assembly.load#system-reflection-assembly-load(system-reflection-assemblyname)) (source code references: [[1]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/Formatters/Binary/BinaryObjectReader.cs#L1109), [[2]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/FormatterServices.cs#L319)). In the context of BinaryFormatter, this uses the [default AssemblyLoadContext](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.default).
If the type that is being cloned belongs to an assembly from the default context, this works fine.
But if the type belongs to an assembly from a *different* context, this is problematic: the assembly might not be found in the default context, and if it *is* found it may later [cause an InvalidCastException](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext#type-conversion-issues) (because loading the assembly in the default context results in a new Assembly instance and its types are distinct).
One situation where non-default contexts arise is when writing a COM (Component Object Model) server.
.NET Core loads the COM server in an isolated AssemblyLoadContext (reference: [COM hosting design doc](https://github.com/dotnet/runtime/blob/main/docs/design/features/native-hosting.md#loading-managed-code) and [notes in the .NET Core COM server docs](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com#additional-notes)).
The dependencies of the COM server (e.g. Models.dll) are automatically loaded in this isolated context when the COM server starts.
So if the COM server creates a Simulation instance that is later passed to Clone(), the call will fail:
If Models.dll is not found in the default context, there will be a SerializationException "Unable to find assembly 'Models ...'."
If Models.dll *is* found, there will be an InvalidCastException when the caller of Clone() tries to cast the result to Simulation, because the result's type will be from a different Assembly instance.
### Steps to reproduce
#### Example 1 (minimal, contrived)
This is a minimal (but contrived) example that manually loads an assembly in a non-default AssemblyLoadContext and calls Clone() on an instance of a type from that assembly.
1. Create a class library project "DemoLibrary":
```
dotnet new classlib -f netcoreapp3.1 -o DemoLibrary
```
Create a file "DemoSimulation.cs" in this project with the following contents:
```c#
using System;
namespace DemoLibrary
{
[Serializable]
public class DemoSimulation
{
}
}
```
Run `dotnet build` in this project (this generates bin\Debug\netcoreapp3.1\DemoLibrary.dll).
2. Create a console application project "DemoApp" that loads the library and calls Clone():
```
dotnet new console -f netcoreapp3.1 -o DemoApp
cd DemoApp
dotnet add reference ...\ApsimX\APSIM.Shared\APSIM.Shared.csproj # replace "..." with the actual path
```
Modify Program.cs to have the following contents:
```c#
using APSIM.Shared.Utilities;
using System;
using System.IO;
using System.Reflection;
using System.Runtime.Loader;
namespace DemoApp
{
class DummyAssemblyLoadContext : AssemblyLoadContext
{
}
class Program
{
static void Main(string[] args)
{
AssemblyLoadContext assemblyLoadContext = new DummyAssemblyLoadContext();
Assembly demoLibrary = assemblyLoadContext.LoadFromAssemblyPath(Directory.GetCurrentDirectory() + @"\..\DemoLibrary\bin\Debug\netcoreapp3.1\DemoLibrary.dll");
Type demoSimulationClass = demoLibrary.GetType("DemoLibrary.DemoSimulation", throwOnError: true);
object sim = Activator.CreateInstance(demoSimulationClass); // dynamic equivalent of "DemoSimulation sim = new DemoSimulation();"
object clone = ReflectionUtilities.Clone(sim); // fails with "SerializationException: Unable to find assembly 'DemoLibrary, ...'."
Console.WriteLine("Clone() succeeded. IsInstance: {0}", demoSimulationClass.IsInstanceOfType(clone));
}
}
}
```
Run using `dotnet run`.
Output (on ApsimX at commit 4cc851718f8b):
```
Unhandled exception. System.Runtime.Serialization.SerializationException: Unable to find assembly 'DemoLibrary, Version=1.0.0.0, Culture=neutral, PublicKeyToken=null'.
at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name)
at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream)
at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 344
at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 551
at DemoApp.Program.Main(String[] args) in C:\Users\User\DemoApp\Program.cs:line 21
```
(If a reference to DemoLibrary is added to DemoApp, Clone() will succeed and the output will be `Clone() succeeded. IsInstance: false`.)
#### Example 2: COM Server
This is a larger example that uses a COM (Component Object Model) server. It is the real-life use case that led to the discovery of the issue.
1. Download the repository [dotnet/samples](https://github.com/dotnet/samples) to get the COM Server Demo ([documentation](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com), [source subdirectory](https://github.com/dotnet/samples/tree/main/core/extensions/COMServerDemo)).
2. Modify the COMServer project to use ApsimX:
```
cd samples\core\extensions\COMServerDemo\COMServer
dotnet add reference ...\ApsimX\Models\Models.csproj # replace "..." with the actual path
```
Modify Server.cs to have the following contents:
```c#
using Models.Core;
using Models.Core.Run;
using System;
using System.Collections.Generic;
using System.Runtime.InteropServices;
namespace COMServer
{
[ComVisible(true)]
[Guid(ContractGuids.ServerClass)]
public class Server : IServer
{
double IServer.ComputePi()
{
Simulations sims = Simulations.Create(new Simulation[] { new Simulation() });
// based on APSIM.Cli/Program.cs
Runner runner = new Runner(sims);
List<Exception> errors = runner.Run();
if (errors != null && errors.Count > 0)
{
throw new AggregateException("Simulation ran with errors", errors);
}
return 3.14;
}
}
}
```
Build the server:
```
dotnet build -p:RegFree=True
```
(The above command also copies the COMServer output files to the COMClient project. I'm using registration-free mode because it does not require administrative privileges.)
3. Build and run the COMClient project:
```
cd ..\COMClient
dotnet build -p:RegFree=True
dotnet run --no-build
```
(Note that `dotnet run` must be invoked with `--no-build`, otherwise it might try to automatically build without `-p:RegFree=True`. If that happens, the fix is to run `dotnet clean; dotnet build -p:RegFree=True`. Cleaning is necessary because the build system does not recognise the property change as a reason for a full rebuild.)
Output:
```
Unhandled exception. System.AggregateException: Simulation ran with errors (Error in file: Simulation: Simulation)
---> System.Exception: Error in file: Simulation: Simulation
---> System.Runtime.Serialization.SerializationException: Unable to find assembly 'Models, Version=0.0.0.0, Culture=neutral, PublicKeyToken=null'.
at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name)
at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream)
at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream)
at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj)
at Models.Core.Apsim.Clone[T](T model)
at Models.Core.Run.SimulationDescription.ToSimulation()
--- End of inner exception stack trace ---
at Models.Core.Run.SimulationDescription.ToSimulation()
at Models.Core.Run.SimulationDescription.Prepare()
at APSIM.Shared.JobRunning.JobRunner.Prepare(IRunnable job)
at APSIM.Shared.JobRunning.JobRunner.RunActualJob(IRunnable job, IJobManager jobManager)
--- End of inner exception stack trace ---
at COMServer.Server.IServer.ComputePi()
at IServer.ComputePi()
at COMClient.Program.Main(String[] args) in C:\Users\User\dotnet-samples\core\extensions\COMServerDemo\COMClient\Program.cs:line 12
```
### Proposed fix
#7465
### Alternative workaround
There is a simple (but ugly) workaround which can be implemented in the user's application code without modifying ApsimX. The idea is to add a handler to the event [AssemblyLoadContext.Resolving](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.resolving) of the default context (or the event [AppDomain.AssemblyResolve](https://docs.microsoft.com/en-us/dotnet/api/system.appdomain.assemblyresolve)), and when it occurs load the assembly from the desired context instead. Note that this event only occurs when assembly resolution fails, so this workaround only works if the assembly is *not* found in the default context.
```c#
...
public class Server : IServer
{
public Server()
{
// [Somewhere at the start of the program]
AssemblyLoadContext.Default.Resolving += OnResolving;
}
public static Assembly OnResolving(AssemblyLoadContext originalContext, AssemblyName assemblyName)
{
// [Replace "Server" with the name of some class from the desired load context]
AssemblyLoadContext customContext = AssemblyLoadContext.GetLoadContext(typeof(Server).Assembly);
return customContext.LoadFromAssemblyName(assemblyName);
}
...
}
```
**Discovered by:** @HemdaLevyNZ.
| 1.0 | ReflectionUtilities.Clone() fails with "SerializationException: Unable to find assembly" when using non-default AssemblyLoadContext - ### Summary
When ApsimX's ReflectionUtilities.Clone() method is called on an instance of a type that belongs to an assembly that was not loaded in the default AssemblyLoadContext, it fails with `SerializationException: Unable to find assembly '...'.`
(Another possibility is that the assembly will also be found in the default load context, and Clone() will succeed, but then there will be an InvalidCastException when the caller of Clone() tries to cast the result.)
A fix is proposed in #7465.
### Background and analysis
ApsimX's [ReflectionUtilities.Clone()](https://github.com/APSIMInitiative/ApsimX/blob/265679cee30b0355e222f6a5832272ecc16f98da/APSIM.Shared/Utilities/ReflectionUtilities.cs#L544) method is implemented by serializing and deserializing using .NET's [BinaryFormatter](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.serialization.formatters.binary.binaryformatter).
When BinaryFormatter performs deserialization, it needs to be able to load assemblies.
.NET Core has something called "AssemblyLoadContext" ([documentation](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext)), which allows assemblies to be loaded in isolated contexts (e.g. in order to load multiple versions of an assembly).
By default, BinaryFormatter.Deserialize loads assemblies using [Assembly.Load(AssemblyName)](https://docs.microsoft.com/en-us/dotnet/api/system.reflection.assembly.load#system-reflection-assembly-load(system-reflection-assemblyname)) (source code references: [[1]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/Formatters/Binary/BinaryObjectReader.cs#L1109), [[2]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/FormatterServices.cs#L319)). In the context of BinaryFormatter, this uses the [default AssemblyLoadContext](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.default).
If the type that is being cloned belongs to an assembly from the default context, this works fine.
But if the type belongs to an assembly from a *different* context, this is problematic: the assembly might not be found in the default context, and if it *is* found it may later [cause an InvalidCastException](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext#type-conversion-issues) (because loading the assembly in the default context results in a new Assembly instance and its types are distinct).
One situation where non-default contexts arise is when writing a COM (Component Object Model) server.
.NET Core loads the COM server in an isolated AssemblyLoadContext (reference: [COM hosting design doc](https://github.com/dotnet/runtime/blob/main/docs/design/features/native-hosting.md#loading-managed-code) and [notes in the .NET Core COM server docs](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com#additional-notes)).
The dependencies of the COM server (e.g. Models.dll) are automatically loaded in this isolated context when the COM server starts.
So if the COM server creates a Simulation instance that is later passed to Clone(), the call will fail:
If Models.dll is not found in the default context, there will be a SerializationException "Unable to find assembly 'Models ...'."
If Models.dll *is* found, there will be an InvalidCastException when the caller of Clone() tries to cast the result to Simulation, because the result's type will be from a different Assembly instance.
### Steps to reproduce
#### Example 1 (minimal, contrived)
This is a minimal (but contrived) example that manually loads an assembly in a non-default AssemblyLoadContext and calls Clone() on an instance of a type from that assembly.
1. Create a class library project "DemoLibrary":
```
dotnet new classlib -f netcoreapp3.1 -o DemoLibrary
```
Create a file "DemoSimulation.cs" in this project with the following contents:
```c#
using System;
namespace DemoLibrary
{
[Serializable]
public class DemoSimulation
{
}
}
```
Run `dotnet build` in this project (this generates bin\Debug\netcoreapp3.1\DemoLibrary.dll).
2. Create a console application project "DemoApp" that loads the library and calls Clone():
```
dotnet new console -f netcoreapp3.1 -o DemoApp
cd DemoApp
dotnet add reference ...\ApsimX\APSIM.Shared\APSIM.Shared.csproj # replace "..." with the actual path
```
Modify Program.cs to have the following contents:
```c#
using APSIM.Shared.Utilities;
using System;
using System.IO;
using System.Reflection;
using System.Runtime.Loader;
namespace DemoApp
{
class DummyAssemblyLoadContext : AssemblyLoadContext
{
}
class Program
{
static void Main(string[] args)
{
AssemblyLoadContext assemblyLoadContext = new DummyAssemblyLoadContext();
Assembly demoLibrary = assemblyLoadContext.LoadFromAssemblyPath(Directory.GetCurrentDirectory() + @"\..\DemoLibrary\bin\Debug\netcoreapp3.1\DemoLibrary.dll");
Type demoSimulationClass = demoLibrary.GetType("DemoLibrary.DemoSimulation", throwOnError: true);
object sim = Activator.CreateInstance(demoSimulationClass); // dynamic equivalent of "DemoSimulation sim = new DemoSimulation();"
object clone = ReflectionUtilities.Clone(sim); // fails with "SerializationException: Unable to find assembly 'DemoLibrary, ...'."
Console.WriteLine("Clone() succeeded. IsInstance: {0}", demoSimulationClass.IsInstanceOfType(clone));
}
}
}
```
Run using `dotnet run`.
Output (on ApsimX at commit 4cc851718f8b):
```
Unhandled exception. System.Runtime.Serialization.SerializationException: Unable to find assembly 'DemoLibrary, Version=1.0.0.0, Culture=neutral, PublicKeyToken=null'.
at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name)
at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream)
at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 344
at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 551
at DemoApp.Program.Main(String[] args) in C:\Users\User\DemoApp\Program.cs:line 21
```
(If a reference to DemoLibrary is added to DemoApp, Clone() will succeed and the output will be `Clone() succeeded. IsInstance: false`.)
#### Example 2: COM Server
This is a larger example that uses a COM (Component Object Model) server. It is the real-life use case that led to the discovery of the issue.
1. Download the repository [dotnet/samples](https://github.com/dotnet/samples) to get the COM Server Demo ([documentation](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com), [source subdirectory](https://github.com/dotnet/samples/tree/main/core/extensions/COMServerDemo)).
2. Modify the COMServer project to use ApsimX:
```
cd samples\core\extensions\COMServerDemo\COMServer
dotnet add reference ...\ApsimX\Models\Models.csproj # replace "..." with the actual path
```
Modify Server.cs to have the following contents:
```c#
using Models.Core;
using Models.Core.Run;
using System;
using System.Collections.Generic;
using System.Runtime.InteropServices;
namespace COMServer
{
[ComVisible(true)]
[Guid(ContractGuids.ServerClass)]
public class Server : IServer
{
double IServer.ComputePi()
{
Simulations sims = Simulations.Create(new Simulation[] { new Simulation() });
// based on APSIM.Cli/Program.cs
Runner runner = new Runner(sims);
List<Exception> errors = runner.Run();
if (errors != null && errors.Count > 0)
{
throw new AggregateException("Simulation ran with errors", errors);
}
return 3.14;
}
}
}
```
Build the server:
```
dotnet build -p:RegFree=True
```
(The above command also copies the COMServer output files to the COMClient project. I'm using registration-free mode because it does not require administrative privileges.)
3. Build and run the COMClient project:
```
cd ..\COMClient
dotnet build -p:RegFree=True
dotnet run --no-build
```
(Note that `dotnet run` must be invoked with `--no-build`, otherwise it might try to automatically build without `-p:RegFree=True`. If that happens, the fix is to run `dotnet clean; dotnet build -p:RegFree=True`. Cleaning is necessary because the build system does not recognise the property change as a reason for a full rebuild.)
Output:
```
Unhandled exception. System.AggregateException: Simulation ran with errors (Error in file: Simulation: Simulation)
---> System.Exception: Error in file: Simulation: Simulation
---> System.Runtime.Serialization.SerializationException: Unable to find assembly 'Models, Version=0.0.0.0, Culture=neutral, PublicKeyToken=null'.
at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name)
at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum)
at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run()
at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check)
at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream)
at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream)
at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj)
at Models.Core.Apsim.Clone[T](T model)
at Models.Core.Run.SimulationDescription.ToSimulation()
--- End of inner exception stack trace ---
at Models.Core.Run.SimulationDescription.ToSimulation()
at Models.Core.Run.SimulationDescription.Prepare()
at APSIM.Shared.JobRunning.JobRunner.Prepare(IRunnable job)
at APSIM.Shared.JobRunning.JobRunner.RunActualJob(IRunnable job, IJobManager jobManager)
--- End of inner exception stack trace ---
at COMServer.Server.IServer.ComputePi()
at IServer.ComputePi()
at COMClient.Program.Main(String[] args) in C:\Users\User\dotnet-samples\core\extensions\COMServerDemo\COMClient\Program.cs:line 12
```
### Proposed fix
#7465
### Alternative workaround
There is a simple (but ugly) workaround which can be implemented in the user's application code without modifying ApsimX. The idea is to add a handler to the event [AssemblyLoadContext.Resolving](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.resolving) of the default context (or the event [AppDomain.AssemblyResolve](https://docs.microsoft.com/en-us/dotnet/api/system.appdomain.assemblyresolve)), and when it occurs load the assembly from the desired context instead. Note that this event only occurs when assembly resolution fails, so this workaround only works if the assembly is *not* found in the default context.
```c#
...
public class Server : IServer
{
public Server()
{
// [Somewhere at the start of the program]
AssemblyLoadContext.Default.Resolving += OnResolving;
}
public static Assembly OnResolving(AssemblyLoadContext originalContext, AssemblyName assemblyName)
{
// [Replace "Server" with the name of some class from the desired load context]
AssemblyLoadContext customContext = AssemblyLoadContext.GetLoadContext(typeof(Server).Assembly);
return customContext.LoadFromAssemblyName(assemblyName);
}
...
}
```
**Discovered by:** @HemdaLevyNZ.
| infrastructure | reflectionutilities clone fails with serializationexception unable to find assembly when using non default assemblyloadcontext summary when apsimx s reflectionutilities clone method is called on an instance of a type that belongs to an assembly that was not loaded in the default assemblyloadcontext it fails with serializationexception unable to find assembly another possibility is that the assembly will also be found in the default load context and clone will succeed but then there will be an invalidcastexception when the caller of clone tries to cast the result a fix is proposed in background and analysis apsimx s method is implemented by serializing and deserializing using net s when binaryformatter performs deserialization it needs to be able to load assemblies net core has something called assemblyloadcontext which allows assemblies to be loaded in isolated contexts e g in order to load multiple versions of an assembly by default binaryformatter deserialize loads assemblies using source code references in the context of binaryformatter this uses the if the type that is being cloned belongs to an assembly from the default context this works fine but if the type belongs to an assembly from a different context this is problematic the assembly might not be found in the default context and if it is found it may later because loading the assembly in the default context results in a new assembly instance and its types are distinct one situation where non default contexts arise is when writing a com component object model server net core loads the com server in an isolated assemblyloadcontext reference and the dependencies of the com server e g models dll are automatically loaded in this isolated context when the com server starts so if the com server creates a simulation instance that is later passed to clone the call will fail if models dll is not found in the default context there will be a serializationexception unable to find assembly models if models dll is found there will be an invalidcastexception when the caller of clone tries to cast the result to simulation because the result s type will be from a different assembly instance steps to reproduce example minimal contrived this is a minimal but contrived example that manually loads an assembly in a non default assemblyloadcontext and calls clone on an instance of a type from that assembly create a class library project demolibrary dotnet new classlib f o demolibrary create a file demosimulation cs in this project with the following contents c using system namespace demolibrary public class demosimulation run dotnet build in this project this generates bin debug demolibrary dll create a console application project demoapp that loads the library and calls clone dotnet new console f o demoapp cd demoapp dotnet add reference apsimx apsim shared apsim shared csproj replace with the actual path modify program cs to have the following contents c using apsim shared utilities using system using system io using system reflection using system runtime loader namespace demoapp class dummyassemblyloadcontext assemblyloadcontext class program static void main string args assemblyloadcontext assemblyloadcontext new dummyassemblyloadcontext assembly demolibrary assemblyloadcontext loadfromassemblypath directory getcurrentdirectory demolibrary bin debug demolibrary dll type demosimulationclass demolibrary gettype demolibrary demosimulation throwonerror true object sim activator createinstance demosimulationclass dynamic equivalent of demosimulation sim new demosimulation object clone reflectionutilities clone sim fails with serializationexception unable to find assembly demolibrary console writeline clone succeeded isinstance demosimulationclass isinstanceoftype clone run using dotnet run output on apsimx at commit unhandled exception system runtime serialization serializationexception unable to find assembly demolibrary version culture neutral publickeytoken null at system runtime serialization formatters binary binaryassemblyinfo getassembly at system runtime serialization formatters binary objectreader gettype binaryassemblyinfo assemblyinfo string name at system runtime serialization formatters binary objectmap ctor string objectname string membernames binarytypeenum binarytypeenuma object typeinformationa memberassemids objectreader objectreader objectid binaryassemblyinfo assemblyinfo sizedarray assemidtoassemblytable at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryobjectwithmaptyped record at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryheaderenum binaryheaderenum at system runtime serialization formatters binary binaryparser run at system runtime serialization formatters binary objectreader deserialize binaryparser serparser boolean fcheck at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream boolean check at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream at apsim shared utilities reflectionutilities binarydeserialise stream stream in c users user apsimx apsim shared utilities reflectionutilities cs line at apsim shared utilities reflectionutilities clone object sourceobj in c users user apsimx apsim shared utilities reflectionutilities cs line at demoapp program main string args in c users user demoapp program cs line if a reference to demolibrary is added to demoapp clone will succeed and the output will be clone succeeded isinstance false example com server this is a larger example that uses a com component object model server it is the real life use case that led to the discovery of the issue download the repository to get the com server demo modify the comserver project to use apsimx cd samples core extensions comserverdemo comserver dotnet add reference apsimx models models csproj replace with the actual path modify server cs to have the following contents c using models core using models core run using system using system collections generic using system runtime interopservices namespace comserver public class server iserver double iserver computepi simulations sims simulations create new simulation new simulation based on apsim cli program cs runner runner new runner sims list errors runner run if errors null errors count throw new aggregateexception simulation ran with errors errors return build the server dotnet build p regfree true the above command also copies the comserver output files to the comclient project i m using registration free mode because it does not require administrative privileges build and run the comclient project cd comclient dotnet build p regfree true dotnet run no build note that dotnet run must be invoked with no build otherwise it might try to automatically build without p regfree true if that happens the fix is to run dotnet clean dotnet build p regfree true cleaning is necessary because the build system does not recognise the property change as a reason for a full rebuild output unhandled exception system aggregateexception simulation ran with errors error in file simulation simulation system exception error in file simulation simulation system runtime serialization serializationexception unable to find assembly models version culture neutral publickeytoken null at system runtime serialization formatters binary binaryassemblyinfo getassembly at system runtime serialization formatters binary objectreader gettype binaryassemblyinfo assemblyinfo string name at system runtime serialization formatters binary objectmap ctor string objectname string membernames binarytypeenum binarytypeenuma object typeinformationa memberassemids objectreader objectreader objectid binaryassemblyinfo assemblyinfo sizedarray assemidtoassemblytable at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryobjectwithmaptyped record at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryheaderenum binaryheaderenum at system runtime serialization formatters binary binaryparser run at system runtime serialization formatters binary objectreader deserialize binaryparser serparser boolean fcheck at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream boolean check at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream at apsim shared utilities reflectionutilities binarydeserialise stream stream at apsim shared utilities reflectionutilities clone object sourceobj at models core apsim clone t model at models core run simulationdescription tosimulation end of inner exception stack trace at models core run simulationdescription tosimulation at models core run simulationdescription prepare at apsim shared jobrunning jobrunner prepare irunnable job at apsim shared jobrunning jobrunner runactualjob irunnable job ijobmanager jobmanager end of inner exception stack trace at comserver server iserver computepi at iserver computepi at comclient program main string args in c users user dotnet samples core extensions comserverdemo comclient program cs line proposed fix alternative workaround there is a simple but ugly workaround which can be implemented in the user s application code without modifying apsimx the idea is to add a handler to the event of the default context or the event and when it occurs load the assembly from the desired context instead note that this event only occurs when assembly resolution fails so this workaround only works if the assembly is not found in the default context c public class server iserver public server assemblyloadcontext default resolving onresolving public static assembly onresolving assemblyloadcontext originalcontext assemblyname assemblyname assemblyloadcontext customcontext assemblyloadcontext getloadcontext typeof server assembly return customcontext loadfromassemblyname assemblyname discovered by hemdalevynz | 1 |
7,760 | 7,084,097,406 | IssuesEvent | 2018-01-11 04:32:54 | comses/core.comses.net | https://api.github.com/repos/comses/core.comses.net | closed | set up production webpack deploy | frontend infrastructure | `npm run build` is currently failing - we need to fix this for prod deployment.
https://vuejs.org/v2/guide/deployment.html
| 1.0 | set up production webpack deploy - `npm run build` is currently failing - we need to fix this for prod deployment.
https://vuejs.org/v2/guide/deployment.html
| infrastructure | set up production webpack deploy npm run build is currently failing we need to fix this for prod deployment | 1 |
23,688 | 6,474,614,232 | IssuesEvent | 2017-08-17 18:29:52 | Microsoft/TypeScript | https://api.github.com/repos/Microsoft/TypeScript | reopened | API to tell editor to trigger rename after applying refactoring | Bug VS Code Tracked | **Problem**
Using the new extract method refactoring in TS 2.5, extracted methods currently are always named `newFunction*`. Renaming the extracted method requires a second step after the refactoring as been applied
A better flow would automatically trigger the rename on the new function after the refactoring has been applied.
**Proposal**
In the `getEditsForRefactor` response, return an additional field indicating that an editor should perform another action, such as rename, after the refactor has been applied. | 1.0 | API to tell editor to trigger rename after applying refactoring - **Problem**
Using the new extract method refactoring in TS 2.5, extracted methods currently are always named `newFunction*`. Renaming the extracted method requires a second step after the refactoring as been applied
A better flow would automatically trigger the rename on the new function after the refactoring has been applied.
**Proposal**
In the `getEditsForRefactor` response, return an additional field indicating that an editor should perform another action, such as rename, after the refactor has been applied. | non_infrastructure | api to tell editor to trigger rename after applying refactoring problem using the new extract method refactoring in ts extracted methods currently are always named newfunction renaming the extracted method requires a second step after the refactoring as been applied a better flow would automatically trigger the rename on the new function after the refactoring has been applied proposal in the geteditsforrefactor response return an additional field indicating that an editor should perform another action such as rename after the refactor has been applied | 0 |
11,520 | 9,220,677,820 | IssuesEvent | 2019-03-11 18:03:38 | HumanCellAtlas/secondary-analysis | https://api.github.com/repos/HumanCellAtlas/secondary-analysis | closed | Update Optimus Adapters for Primetime | Q1-2019 infrastructure moved optimus 🚫blocked | It has been a long time since we have run Optimus in the DCP. We think there is work to do to update the adapters. After we update Optimus to have zarr outputs, we should update it so we can run the pipeline in the DCP.
~Consider #355 when doing this ticket, they can be kept separate if the work is too much for one ticket.~
AC:
- [ ] Update adapters to support latest tagged version of Optimus
- [ ] Update lira config
- [ ] Zarr output is created, and is captured in adapters
- [ ] Run with current existing dataset to help determine next steps for testing. | 1.0 | Update Optimus Adapters for Primetime - It has been a long time since we have run Optimus in the DCP. We think there is work to do to update the adapters. After we update Optimus to have zarr outputs, we should update it so we can run the pipeline in the DCP.
~Consider #355 when doing this ticket, they can be kept separate if the work is too much for one ticket.~
AC:
- [ ] Update adapters to support latest tagged version of Optimus
- [ ] Update lira config
- [ ] Zarr output is created, and is captured in adapters
- [ ] Run with current existing dataset to help determine next steps for testing. | infrastructure | update optimus adapters for primetime it has been a long time since we have run optimus in the dcp we think there is work to do to update the adapters after we update optimus to have zarr outputs we should update it so we can run the pipeline in the dcp consider when doing this ticket they can be kept separate if the work is too much for one ticket ac update adapters to support latest tagged version of optimus update lira config zarr output is created and is captured in adapters run with current existing dataset to help determine next steps for testing | 1 |
270,338 | 28,960,505,825 | IssuesEvent | 2023-05-10 01:47:46 | praneethpanasala/linux | https://api.github.com/repos/praneethpanasala/linux | reopened | WS-2021-0554 (High) detected in linuxlinux-4.19.6 | Mend: dependency security vulnerability | ## WS-2021-0554 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
nfc: fix potential NULL pointer deref in nfc_genl_dump_ses_done
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.8 by commit fae9705d281091254d4a81fa2da9d22346097dca, it was introduced in version v3.12 by commit ac22ac466a659f1b2e02a2e2ee23fc5c42da2c95. For more details please see the references link.
<p>Publish Date: 2021-12-19
<p>URL: <a href=https://github.com/gregkh/linux/commit/fae9705d281091254d4a81fa2da9d22346097dca>WS-2021-0554</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002588">https://osv.dev/vulnerability/GSD-2021-1002588</a></p>
<p>Release Date: 2021-12-19</p>
<p>Fix Resolution: v5.15.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2021-0554 (High) detected in linuxlinux-4.19.6 - ## WS-2021-0554 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
nfc: fix potential NULL pointer deref in nfc_genl_dump_ses_done
This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven.
This ID is fixed in Linux Kernel version v5.15.8 by commit fae9705d281091254d4a81fa2da9d22346097dca, it was introduced in version v3.12 by commit ac22ac466a659f1b2e02a2e2ee23fc5c42da2c95. For more details please see the references link.
<p>Publish Date: 2021-12-19
<p>URL: <a href=https://github.com/gregkh/linux/commit/fae9705d281091254d4a81fa2da9d22346097dca>WS-2021-0554</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002588">https://osv.dev/vulnerability/GSD-2021-1002588</a></p>
<p>Release Date: 2021-12-19</p>
<p>Fix Resolution: v5.15.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | ws high detected in linuxlinux ws high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details nfc fix potential null pointer deref in nfc genl dump ses done this is an automated id intended to aid in discovery of potential security vulnerabilities the actual impact and attack plausibility have not yet been proven this id is fixed in linux kernel version by commit it was introduced in version by commit for more details please see the references link publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
27,610 | 21,992,288,544 | IssuesEvent | 2022-05-26 00:02:04 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Remove un-needed target when upgrading to v2 BuildTools | enhancement area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate | We have done some change in the PR dotnet/coreclr#11482 to change the license Url for packaging. because coreclr repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the src/.nuget/dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct
| 1.0 | Remove un-needed target when upgrading to v2 BuildTools - We have done some change in the PR dotnet/coreclr#11482 to change the license Url for packaging. because coreclr repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the src/.nuget/dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct
| infrastructure | remove un needed target when upgrading to buildtools we have done some change in the pr dotnet coreclr to change the license url for packaging because coreclr repo not using of the buildtools we had to hack this by adding the temporary target overridelicenseurl in the src nuget dir targets files this target should be removed after upgrading to of the buildtools and need to ensure when removing it the license url in the packages nuspec files is correct | 1 |
18,349 | 12,950,473,198 | IssuesEvent | 2020-07-19 13:20:30 | kopia/kopia | https://api.github.com/repos/kopia/kopia | opened | compression performance regression at commit 272b657508329ac4035c656c2503f6f476d42e7d | bug performance project-infrastructure | This happened back in March between 0.5.2 and HEAD, looks like we lost >31% of compression efficiency of `s2-default` after upgrading compression libraries.
On 200 GB sample files the compressed repository size before the change was 89.98 GB, after it jumped to 118.67 GB.
The suspect change was on 2020-03-15 and upgraded `github.com/klauspost/compress` from `v1.9.7` to `v1.10.3`
Thanks @ntolia for the investigation! | 1.0 | compression performance regression at commit 272b657508329ac4035c656c2503f6f476d42e7d - This happened back in March between 0.5.2 and HEAD, looks like we lost >31% of compression efficiency of `s2-default` after upgrading compression libraries.
On 200 GB sample files the compressed repository size before the change was 89.98 GB, after it jumped to 118.67 GB.
The suspect change was on 2020-03-15 and upgraded `github.com/klauspost/compress` from `v1.9.7` to `v1.10.3`
Thanks @ntolia for the investigation! | infrastructure | compression performance regression at commit this happened back in march between and head looks like we lost of compression efficiency of default after upgrading compression libraries on gb sample files the compressed repository size before the change was gb after it jumped to gb the suspect change was on and upgraded github com klauspost compress from to thanks ntolia for the investigation | 1 |
28,600 | 23,383,631,981 | IssuesEvent | 2022-08-11 11:54:46 | dotnet/sdk | https://api.github.com/repos/dotnet/sdk | closed | Changes to bring templates into SDK introduces source-build prebuilts | Area-Infrastructure untriaged | The installer's references to Microsoft.DotNet.Common.ItemTemplates and Microsoft.DotNet.Common.ProjectTemplates are being detected as prebuilts in source-build. This appears to be a result of https://github.com/dotnet/sdk/pull/26521.
I am pretty sure this is because the templating projects weren't added to the sdk repo's [source-build.slnf](https://github.com/dotnet/sdk/blob/main/source-build.slnf) therefore they are not getting built as part of source-build.
cc @vlada-shubina, @GangWang01 | 1.0 | Changes to bring templates into SDK introduces source-build prebuilts - The installer's references to Microsoft.DotNet.Common.ItemTemplates and Microsoft.DotNet.Common.ProjectTemplates are being detected as prebuilts in source-build. This appears to be a result of https://github.com/dotnet/sdk/pull/26521.
I am pretty sure this is because the templating projects weren't added to the sdk repo's [source-build.slnf](https://github.com/dotnet/sdk/blob/main/source-build.slnf) therefore they are not getting built as part of source-build.
cc @vlada-shubina, @GangWang01 | infrastructure | changes to bring templates into sdk introduces source build prebuilts the installer s references to microsoft dotnet common itemtemplates and microsoft dotnet common projecttemplates are being detected as prebuilts in source build this appears to be a result of i am pretty sure this is because the templating projects weren t added to the sdk repo s therefore they are not getting built as part of source build cc vlada shubina | 1 |
20,439 | 10,519,742,633 | IssuesEvent | 2019-09-29 20:06:08 | DashboardHub/PipelineDashboard | https://api.github.com/repos/DashboardHub/PipelineDashboard | closed | CVE-2018-20822 (Medium) detected in node-sass-v4.12.0 | security vulnerability | ## CVE-2018-20822 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/DashboardHub/PipelineDashboard/commits/bb91378eadab1837766aef02a00d31b46095a1c4">bb91378eadab1837766aef02a00d31b46095a1c4</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_node.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/base.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operation.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/paths.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/values.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/json.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/string.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass2scss.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/factory.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/value.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/callback_bridge.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/functions.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/custom_function_bridge.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/bind.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debugger.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cencode.c
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/number.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/c99func.c
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/values.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/null.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/context.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/b64/encode.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/binding.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822>CVE-2018-20822</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20822 (Medium) detected in node-sass-v4.12.0 - ## CVE-2018-20822 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary>
<p>
<p>:rainbow: Node.js bindings to libsass</p>
<p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/DashboardHub/PipelineDashboard/commits/bb91378eadab1837766aef02a00d31b46095a1c4">bb91378eadab1837766aef02a00d31b46095a1c4</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/unchecked.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_node.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/base.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operation.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/contrib/plugin.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_superselector.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/paths.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_unification.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/values.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/json.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/checked.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/string.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass2scss.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/factory.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/value.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/callback_bridge.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/functions.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/custom_function_bridge.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/bind.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/backtrace.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/sass_value_wrapper.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debugger.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cencode.c
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/base64vlq.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/number.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/c99func.c
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/values.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_subset_map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass2scss.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/null.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/context.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/script/test-leaks.pl
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/b64/encode.h
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/sass_types/map.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.hpp
- /PipelineDashboard/web/node_modules/node-sass/src/binding.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.cpp
- /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debug.hpp
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822>CVE-2018-20822</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries pipelinedashboard web node modules node sass src libsass src expand hpp pipelinedashboard web node modules node sass src libsass src color maps cpp pipelinedashboard web node modules node sass src libsass src sass util hpp pipelinedashboard web node modules node sass src libsass src unchecked h pipelinedashboard web node modules node sass src libsass src output hpp pipelinedashboard web node modules node sass src libsass src sass values hpp pipelinedashboard web node modules node sass src libsass src util hpp pipelinedashboard web node modules node sass src libsass src emitter hpp pipelinedashboard web node modules node sass src libsass src lexer cpp pipelinedashboard web node modules node sass src libsass test test node cpp pipelinedashboard web node modules node sass src libsass src plugins cpp pipelinedashboard web node modules node sass src libsass include sass base h pipelinedashboard web node modules node sass src libsass src position hpp pipelinedashboard web node modules node sass src libsass src subset map hpp pipelinedashboard web node modules node sass src libsass src operation hpp pipelinedashboard web node modules node sass src libsass src remove placeholders cpp pipelinedashboard web node modules node sass src libsass src error handling hpp pipelinedashboard web node modules node sass src custom importer bridge cpp pipelinedashboard web node modules node sass src libsass contrib plugin cpp pipelinedashboard web node modules node sass src libsass src functions hpp pipelinedashboard web node modules node sass src libsass test test superselector cpp pipelinedashboard web node modules node sass src libsass src eval hpp pipelinedashboard web node modules node sass src libsass src string hpp pipelinedashboard web node modules node sass src libsass src error handling cpp pipelinedashboard web node modules node sass src sass context wrapper h pipelinedashboard web node modules node sass src libsass src node cpp pipelinedashboard web node modules node sass src libsass src parser cpp pipelinedashboard web node modules node sass src libsass src subset map cpp pipelinedashboard web node modules node sass src libsass src emitter cpp pipelinedashboard web node modules node sass src libsass src listize cpp pipelinedashboard web node modules node sass src libsass src ast hpp pipelinedashboard web node modules node sass src libsass src sass functions hpp pipelinedashboard web node modules node sass src libsass src memory sharedptr cpp pipelinedashboard web node modules node sass src libsass src output cpp pipelinedashboard web node modules node sass src libsass src check nesting cpp pipelinedashboard web node modules node sass src libsass src ast def macros hpp pipelinedashboard web node modules node sass src libsass src cssize hpp pipelinedashboard web node modules node sass src libsass src functions cpp pipelinedashboard web node modules node sass src libsass src paths hpp pipelinedashboard web node modules node sass src libsass src prelexer cpp pipelinedashboard web node modules node sass src libsass src ast fwd decl hpp pipelinedashboard web node modules node sass src sass types color cpp pipelinedashboard web node modules node sass src libsass test test unification cpp pipelinedashboard web node modules node sass src libsass src inspect hpp pipelinedashboard web node modules node sass src libsass src values cpp pipelinedashboard web node modules node sass src libsass src sass util cpp pipelinedashboard web node modules node sass src libsass src source map hpp pipelinedashboard web node modules node sass src sass types list h pipelinedashboard web node modules node sass src libsass src json cpp pipelinedashboard web node modules node sass src libsass src check nesting hpp pipelinedashboard web node modules node sass src libsass src units cpp pipelinedashboard web node modules node sass src libsass src units hpp pipelinedashboard web node modules node sass src libsass src context cpp pipelinedashboard web node modules node sass src libsass src checked h pipelinedashboard web node modules node sass src libsass src listize hpp pipelinedashboard web node modules node sass src sass types string cpp pipelinedashboard web node modules node sass src libsass src context hpp pipelinedashboard web node modules node sass src libsass src prelexer hpp pipelinedashboard web node modules node sass src sass types boolean h pipelinedashboard web node modules node sass src libsass include h pipelinedashboard web node modules node sass src libsass src eval cpp pipelinedashboard web node modules node sass src libsass src expand cpp pipelinedashboard web node modules node sass src sass types factory cpp pipelinedashboard web node modules node sass src libsass src operators cpp pipelinedashboard web node modules node sass src sass types boolean cpp pipelinedashboard web node modules node sass src libsass src source map cpp pipelinedashboard web node modules node sass src sass types value h pipelinedashboard web node modules node sass src libsass src string cpp pipelinedashboard web node modules node sass src callback bridge h pipelinedashboard web node modules node sass src libsass src file cpp pipelinedashboard web node modules node sass src libsass src sass cpp pipelinedashboard web node modules node sass src libsass src node hpp pipelinedashboard web node modules node sass src libsass src environment cpp pipelinedashboard web node modules node sass src libsass src extend hpp pipelinedashboard web node modules node sass src libsass src sass context hpp pipelinedashboard web node modules node sass src libsass src operators hpp pipelinedashboard web node modules node sass src libsass src constants hpp pipelinedashboard web node modules node sass src libsass src sass hpp pipelinedashboard web node modules node sass src libsass src ast fwd decl cpp pipelinedashboard web node modules node sass src libsass src parser hpp pipelinedashboard web node modules node sass src libsass src constants cpp pipelinedashboard web node modules node sass src sass types list cpp pipelinedashboard web node modules node sass src libsass src cssize cpp pipelinedashboard web node modules node sass src libsass include sass functions h pipelinedashboard web node modules node sass src libsass src util cpp pipelinedashboard web node modules node sass src custom function bridge cpp pipelinedashboard web node modules node sass src custom importer bridge h pipelinedashboard web node modules node sass src libsass src bind cpp pipelinedashboard web node modules node sass src libsass src inspect cpp pipelinedashboard web node modules node sass src libsass src sass functions cpp pipelinedashboard web node modules node sass src libsass src backtrace cpp pipelinedashboard web node modules node sass src libsass src extend cpp pipelinedashboard web node modules node sass src sass types sass value wrapper h pipelinedashboard web node modules node sass src libsass src debugger hpp pipelinedashboard web node modules node sass src libsass src cencode c pipelinedashboard web node modules node sass src libsass src cpp pipelinedashboard web node modules node sass src sass types number cpp pipelinedashboard web node modules node sass src sass types color h pipelinedashboard web node modules node sass src libsass src c pipelinedashboard web node modules node sass src libsass src position cpp pipelinedashboard web node modules node sass src libsass src remove placeholders hpp pipelinedashboard web node modules node sass src libsass src sass values cpp pipelinedashboard web node modules node sass src libsass include sass values h pipelinedashboard web node modules node sass src libsass test test subset map cpp pipelinedashboard web node modules node sass src libsass src cpp pipelinedashboard web node modules node sass src sass types null cpp pipelinedashboard web node modules node sass src libsass include sass context h pipelinedashboard web node modules node sass src libsass src ast cpp pipelinedashboard web node modules node sass src libsass src to c cpp pipelinedashboard web node modules node sass src libsass src to value hpp pipelinedashboard web node modules node sass src libsass src color maps hpp pipelinedashboard web node modules node sass src sass context wrapper cpp pipelinedashboard web node modules node sass src libsass script test leaks pl pipelinedashboard web node modules node sass src libsass src memory sharedptr hpp pipelinedashboard web node modules node sass src libsass src lexer hpp pipelinedashboard web node modules node sass src libsass src to c hpp pipelinedashboard web node modules node sass src libsass src to value cpp pipelinedashboard web node modules node sass src libsass src encode h pipelinedashboard web node modules node sass src libsass src file hpp pipelinedashboard web node modules node sass src sass types map cpp pipelinedashboard web node modules node sass src libsass src environment hpp pipelinedashboard web node modules node sass src libsass src plugins hpp pipelinedashboard web node modules node sass src binding cpp pipelinedashboard web node modules node sass src libsass src sass context cpp pipelinedashboard web node modules node sass src libsass src debug hpp vulnerability details libsass allows attackers to cause a denial of service uncontrolled recursion in sass complex selector perform in ast hpp and sass inspect operator in inspect cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
72,445 | 3,385,863,189 | IssuesEvent | 2015-11-27 14:09:24 | thesgc/chembiohub_helpdesk | https://api.github.com/repos/thesgc/chembiohub_helpdesk | opened | data overview page has buttons for Add single project and Add multiple project- these should refer t | app: AssayReg name: Karen priority: Low status: New | data overview page has buttons for Add single project and Add multiple project- these should refer to activities. Also the Activities label currently reads Projects. | 1.0 | data overview page has buttons for Add single project and Add multiple project- these should refer t - data overview page has buttons for Add single project and Add multiple project- these should refer to activities. Also the Activities label currently reads Projects. | non_infrastructure | data overview page has buttons for add single project and add multiple project these should refer t data overview page has buttons for add single project and add multiple project these should refer to activities also the activities label currently reads projects | 0 |
212,671 | 7,241,345,285 | IssuesEvent | 2018-02-14 00:42:21 | MRN-Code/coinstac | https://api.github.com/repos/MRN-Code/coinstac | opened | Notification Badges on Throughout Collections Nav for Unmapped Vars | low priority | # Problem
Notification badges should be used throughout the application to guide users to areas needing there attention. One such area is Collections when unmapped vars have been found on a pending run.
# Solution
- Badge on Collection nav item
- Badge on corresponding collection list item.
- Badge on Consortia tab of collection | 1.0 | Notification Badges on Throughout Collections Nav for Unmapped Vars - # Problem
Notification badges should be used throughout the application to guide users to areas needing there attention. One such area is Collections when unmapped vars have been found on a pending run.
# Solution
- Badge on Collection nav item
- Badge on corresponding collection list item.
- Badge on Consortia tab of collection | non_infrastructure | notification badges on throughout collections nav for unmapped vars problem notification badges should be used throughout the application to guide users to areas needing there attention one such area is collections when unmapped vars have been found on a pending run solution badge on collection nav item badge on corresponding collection list item badge on consortia tab of collection | 0 |
42,508 | 9,222,617,720 | IssuesEvent | 2019-03-11 23:37:39 | publiclab/plots2 | https://api.github.com/repos/publiclab/plots2 | opened | Welcome, Outreachy aspirants! (Summer- 2019) | outreachy summer-of-code | Hello folks from Outreachy - We realized we don't have a main place for you all to ask questions and get support, so we wanted to say WELCOME and create this issue as a discussion zone. Please leave a comment to say hello, in part so we know who's interested in the program! @publiclab/reviewers @publiclab/plots2-reviewers @publiclab/mentors please keep an eye open for opportunities to welcome and support our Outreachy applicants! 👍 🎉
## First contribution
It's so great to see that several of you have made your first contribution. 🙌 This should really help as all our contributions follow this workflow. If you haven't yet, please see this page to get started:
https://code.publiclab.org
## What next?
If you're looking for a next project, we might recommend one which starts to involve multiple files in the application code, so you can start to get a better understanding of the application's workings. Are you familiar with how a Rails application works, in terms of how the model accesses the database, the controller performs the logic, and the views display the actual web pages? If not, we can help you start to pick apart how our system works one piece at a time.
You can find a lot of relatively small but not completely documented issues here:
https://github.com/publiclab/plots2/labels/fto-candidate
If you choose one, please leave a comment and we can do our best to help find any missing information or help guide you through the next steps. Don't hesitate to ask questions!
Larger collections of tasks can often be found under the `break-me-up` label, which you can go through to find smaller projects to "break off" and solve:
https://github.com/publiclab/plots2/labels/break-me-up
## Supporting tasks
Another area you might be interested in contributing is in supporting other new people to the project. This is an often unrecognized part of a healthy collaborative project, so we've made these issues to encourage people to take part in this crucial activity. We'd love your help, and as a newcomer yourself, you know exactly what it feels like to be new, so you know how important it is to be welcoming and supportive!
https://github.com/publiclab/plots2/labels/support
You can help new contributors in getting started by creating new FTO issues - https://github.com/publiclab/plots2/issues/4963
Thanks and we're happy to answer questions!!
| 1.0 | Welcome, Outreachy aspirants! (Summer- 2019) - Hello folks from Outreachy - We realized we don't have a main place for you all to ask questions and get support, so we wanted to say WELCOME and create this issue as a discussion zone. Please leave a comment to say hello, in part so we know who's interested in the program! @publiclab/reviewers @publiclab/plots2-reviewers @publiclab/mentors please keep an eye open for opportunities to welcome and support our Outreachy applicants! 👍 🎉
## First contribution
It's so great to see that several of you have made your first contribution. 🙌 This should really help as all our contributions follow this workflow. If you haven't yet, please see this page to get started:
https://code.publiclab.org
## What next?
If you're looking for a next project, we might recommend one which starts to involve multiple files in the application code, so you can start to get a better understanding of the application's workings. Are you familiar with how a Rails application works, in terms of how the model accesses the database, the controller performs the logic, and the views display the actual web pages? If not, we can help you start to pick apart how our system works one piece at a time.
You can find a lot of relatively small but not completely documented issues here:
https://github.com/publiclab/plots2/labels/fto-candidate
If you choose one, please leave a comment and we can do our best to help find any missing information or help guide you through the next steps. Don't hesitate to ask questions!
Larger collections of tasks can often be found under the `break-me-up` label, which you can go through to find smaller projects to "break off" and solve:
https://github.com/publiclab/plots2/labels/break-me-up
## Supporting tasks
Another area you might be interested in contributing is in supporting other new people to the project. This is an often unrecognized part of a healthy collaborative project, so we've made these issues to encourage people to take part in this crucial activity. We'd love your help, and as a newcomer yourself, you know exactly what it feels like to be new, so you know how important it is to be welcoming and supportive!
https://github.com/publiclab/plots2/labels/support
You can help new contributors in getting started by creating new FTO issues - https://github.com/publiclab/plots2/issues/4963
Thanks and we're happy to answer questions!!
| non_infrastructure | welcome outreachy aspirants summer hello folks from outreachy we realized we don t have a main place for you all to ask questions and get support so we wanted to say welcome and create this issue as a discussion zone please leave a comment to say hello in part so we know who s interested in the program publiclab reviewers publiclab reviewers publiclab mentors please keep an eye open for opportunities to welcome and support our outreachy applicants 👍 🎉 first contribution it s so great to see that several of you have made your first contribution 🙌 this should really help as all our contributions follow this workflow if you haven t yet please see this page to get started what next if you re looking for a next project we might recommend one which starts to involve multiple files in the application code so you can start to get a better understanding of the application s workings are you familiar with how a rails application works in terms of how the model accesses the database the controller performs the logic and the views display the actual web pages if not we can help you start to pick apart how our system works one piece at a time you can find a lot of relatively small but not completely documented issues here if you choose one please leave a comment and we can do our best to help find any missing information or help guide you through the next steps don t hesitate to ask questions larger collections of tasks can often be found under the break me up label which you can go through to find smaller projects to break off and solve supporting tasks another area you might be interested in contributing is in supporting other new people to the project this is an often unrecognized part of a healthy collaborative project so we ve made these issues to encourage people to take part in this crucial activity we d love your help and as a newcomer yourself you know exactly what it feels like to be new so you know how important it is to be welcoming and supportive you can help new contributors in getting started by creating new fto issues thanks and we re happy to answer questions | 0 |
33,722 | 27,752,654,754 | IssuesEvent | 2023-03-15 22:15:00 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Can't build in an ARM64 Windows machine, fails in crosscompontents step | area-Infrastructure-coreclr untriaged | Machine: Surface Pro X (ARM64) with Windows 11
Command: `.\build.cmd clr+libs -c release -arch arm64` (repros with or without the `arch` argument)
This is the error that shows up without any code changes:
<details><summary>Original error output (expand)</summary>
```
[vcvarsall.bat] Environment initialized for: 'x64'
BUILD: Regenerating the Visual Studio solution
Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64"
vs2022 x64 linux "-DCMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE
_OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" "-DCLR_CROSS_COMPONENTS_BUILD=1"
Not searching for unused variables given on the command line.
-- The C compiler identification is MSVC 19.33.31631.0
-- The CXX compiler identification is MSVC 19.33.31631.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - failed
-- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe
-- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe - broken
CMake Error at C:/Program Files/CMake/share/cmake-3.26/Modules/CMakeTestCCompiler.cmake:67 (message):
The C compiler
"C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe"
is not able to compile a simple test program.
It fails with the following output:
Change Dir: C:/repos/runtime/artifacts/obj/coreclr/linux.arm64.Release/x64/CMakeFiles/CMakeScratch/TryCompile-cyevcr
Run Build Command(s):C:/PROGRA~1/ninja/ninja.exe -v cmTC_f9773 && [1/2] C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\
cl.exe /nologo /DWIN32 /D_WINDOWS /W3 /Zi /Ob0 /Od /RTC1 -MDd /showIncludes /FoCMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /FdCMakeFiles\cmTC_f97
73.dir\ /FS -c C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64\CMakeFiles\CMakeScratch\TryCompile-cyevcr\testCCompiler.c
[2/2] cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin
\100226~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.31
6\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /versi
on:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib
comdlg32.lib advapi32.lib && cd ."
FAILED: cmTC_f9773.exe
cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\10022
6~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\
HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0
/machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdl
g32.lib advapi32.lib && cd ."
LINK Pass 1: command "C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCo
mpiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.l
ib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MANIFESTFILE:CMakeFiles\cmTC_f977
3.dir/intermediate.manifest CMakeFiles\cmTC_f9773.dir/manifest.res" failed (exit code 1120) with the following output:
testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_InitBase
testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_Shutdown
LINK : error LNK2001: unresolved external symbol mainCRTStartup
C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\MSVCRTD.lib : warning LNK4272: library machine type 'ARM6
4' conflicts with target machine type 'x64'
C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\OLDNAMES.lib : warning LNK4272: library machine type 'ARM
64' conflicts with target machine type 'x64'
cmTC_f9773.exe : fatal error LNK1120: 3 unresolved externals
ninja: build stopped: subcommand failed.
CMake will not be able to correctly generate this project.
Call Stack (most recent call first):
CMakeLists.txt:16 (project)
-- Configuring incomplete, errors occurred!
BUILD: Error: failed to generate native component build project
C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR
OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1.
Build FAILED.
C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR
OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1.
0 Warning(s)
1 Error(s)
Time Elapsed 00:03:49.01
Build failed with exit code 1. Check errors above.
Some builds failed:
Configuration: release, Architecture: arm64
```
</details>
As you can see, there's some places where x64 is being used instead of arm64. I tried adding this line in Subsets.props after line 280:
```diff
<PropertyGroup>
<CrossDacHostArch>x64</CrossDacHostArch>
+ <CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm64'">arm64</CrossDacHostArch>
<CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm'">x86</CrossDacHostArch>
</PropertyGroup>
```
But I am still seeing an error. Here is the result:
<details><summary>Error output with modification (expand)</summary>
```powershell
❯ .\build.cmd clr+libs -c release -arch arm64
Determining projects to restore...
Tool 'coverlet.console' (version '3.1.2') was restored. Available commands: coverlet
Tool 'dotnet-reportgenerator-globaltool' (version '5.0.2') was restored. Available commands: reportgenerator
Tool 'microsoft.dotnet.xharness.cli' (version '1.0.0-prerelease.23080.1') was restored. Available commands: xharness
Tool 'microsoft.visualstudio.slngen.tool' (version '8.5.17') was restored. Available commands: slngen
Tool 'dotnet-format' (version '6.0.240501') was restored. Available commands: dotnet-format
Restore was successful.
All projects are up-to-date for restore.
Determining projects to restore...
All projects are up-to-date for restore.
System.Private.CoreLib.Generators -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib.Generators\Release\netstandard2.0\System.Private.CoreLib.Generators.dll
Microsoft.Interop.SourceGeneration -> C:\repos\runtime\artifacts\bin\Microsoft.Interop.SourceGeneration\Release\netstandard2.0\Microsoft.Interop.SourceGeneration.dll
LibraryImportGenerator -> C:\repos\runtime\artifacts\bin\LibraryImportGenerator\Release\netstandard2.0\Microsoft.Interop.LibraryImportGenerator.dll
System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib\ref\Release\net8.0\System.Private.CoreLib.dll
System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release\IL\System.Private.CoreLib.dll
Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -release -os windows -pgodatapath "C:\Users\calope\.nuget\packages\optimization.windows_nt-arm64.pgo.corecl
r\1.0.0-prerelease.23068.4"
BUILD: Starting Build at 11:53:06.31
BUILD: Commencing CoreCLR product build
BUILD: Checking prerequisites
BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe
BUILD: Number of processor cores 8
BUILD: Commencing build of native components for windows.arm64.Release
BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64
**********************************************************************
** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0
** Copyright (c) 2022 Microsoft Corporation
**********************************************************************
[vcvarsall.bat] Environment initialized for: 'x86_arm64'
BUILD: Regenerating the Visual Studio solution
Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release" vs2022 arm64 windows "-D
CMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=windows" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH="C:\Users\calope\.nuget\pa
ckages\optimization.windows_nt-arm64.pgo.coreclr\1.0.0-prerelease.23068.4"" "-DCLR_CMAKE_PGO_OPTIMIZE=1"
The CMake command line is the same as the last run. Skipping running CMake.
running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release --target install --config Release --
[1/4] Linking RC shared library dlls\mscorrc\mscorrc.dll
[2/4] Linking RC shared library dlls\clretwrc\clretwrc.dll
[3/4] cmd.exe /C "cd /D C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release\dlls\mscoree\coreclr && powershell -NoProfile -ExecutionPolicy ByPass -File C:/repos/run
time/src/coreclr/dlls/mscoree/coreclr/GenClrDebugResource.ps1 -dac C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscordac/mscordaccore.dll -dbi C:/repos/
runtime/artifacts/obj/coreclr/windows.arm64.Releas-- Install configuration: "Release"
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./clretwrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/clretwrc.pdb
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/clretwrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/clretwrc.pdb
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./mscorrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/mscorrc.pdb
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/mscorrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/mscorrc.pdb
e/dlls/mscordbi/mscordbi.dll -out C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscoree/coreclr/clr_debug_resource.bin"
[3/4] Install the project...
BUILD: Build succeeded. Finished at 11:53:16.15
BUILD: Product binaries are available at C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release
Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1" -release -os linux -hostarch arm64 -component crosscomponents
BUILD: Starting Build at 11:53:16.49
BUILD: Commencing CoreCLR product build
BUILD: Checking prerequisites
BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe
BUILD: Number of processor cores 8
BUILD: Commencing build of native components for linux.arm64.Release
BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64
**********************************************************************
** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0
** Copyright (c) 2022 Microsoft Corporation
**********************************************************************
[vcvarsall.bat] Environment initialized for: 'x86_arm64'
BUILD: Regenerating the Visual Studio solution
Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release" vs2022 arm64 linux "-DCMAK
E_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" "
-DCLR_CROSS_COMPONENTS_BUILD=1"
The CMake command line is the same as the last run. Skipping running CMake.
running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release --target crosscomponents --config Release --
ninja: error: loading 'build.ninja': The system cannot find the file specified.
BUILD: Error: native component build failed. Refer to the build log files for details.
"C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.log"
"C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.wrn"
"C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.err"
C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1
" -release -os linux -hostarch arm64 -component crosscomponents" exited with code 1.
```
</details>
Am I missing something or executing the command wrong? I made sure to install all the required dependencies as described in our workflow instructions.
I attached a couple of binlogs in case they are useful:
- Without any code changes: [original_msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10970321/original_msbuild.binlog.zip)
- After my modification: [msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10960289/msbuild.binlog.zip)
@jkoritzinsky @hoyosjs | 1.0 | Can't build in an ARM64 Windows machine, fails in crosscompontents step - Machine: Surface Pro X (ARM64) with Windows 11
Command: `.\build.cmd clr+libs -c release -arch arm64` (repros with or without the `arch` argument)
This is the error that shows up without any code changes:
<details><summary>Original error output (expand)</summary>
```
[vcvarsall.bat] Environment initialized for: 'x64'
BUILD: Regenerating the Visual Studio solution
Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64"
vs2022 x64 linux "-DCMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE
_OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" "-DCLR_CROSS_COMPONENTS_BUILD=1"
Not searching for unused variables given on the command line.
-- The C compiler identification is MSVC 19.33.31631.0
-- The CXX compiler identification is MSVC 19.33.31631.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - failed
-- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe
-- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe - broken
CMake Error at C:/Program Files/CMake/share/cmake-3.26/Modules/CMakeTestCCompiler.cmake:67 (message):
The C compiler
"C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe"
is not able to compile a simple test program.
It fails with the following output:
Change Dir: C:/repos/runtime/artifacts/obj/coreclr/linux.arm64.Release/x64/CMakeFiles/CMakeScratch/TryCompile-cyevcr
Run Build Command(s):C:/PROGRA~1/ninja/ninja.exe -v cmTC_f9773 && [1/2] C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\
cl.exe /nologo /DWIN32 /D_WINDOWS /W3 /Zi /Ob0 /Od /RTC1 -MDd /showIncludes /FoCMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /FdCMakeFiles\cmTC_f97
73.dir\ /FS -c C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64\CMakeFiles\CMakeScratch\TryCompile-cyevcr\testCCompiler.c
[2/2] cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin
\100226~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.31
6\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /versi
on:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib
comdlg32.lib advapi32.lib && cd ."
FAILED: cmTC_f9773.exe
cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\10022
6~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\
HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0
/machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdl
g32.lib advapi32.lib && cd ."
LINK Pass 1: command "C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCo
mpiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.l
ib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MANIFESTFILE:CMakeFiles\cmTC_f977
3.dir/intermediate.manifest CMakeFiles\cmTC_f9773.dir/manifest.res" failed (exit code 1120) with the following output:
testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_InitBase
testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_Shutdown
LINK : error LNK2001: unresolved external symbol mainCRTStartup
C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\MSVCRTD.lib : warning LNK4272: library machine type 'ARM6
4' conflicts with target machine type 'x64'
C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\OLDNAMES.lib : warning LNK4272: library machine type 'ARM
64' conflicts with target machine type 'x64'
cmTC_f9773.exe : fatal error LNK1120: 3 unresolved externals
ninja: build stopped: subcommand failed.
CMake will not be able to correctly generate this project.
Call Stack (most recent call first):
CMakeLists.txt:16 (project)
-- Configuring incomplete, errors occurred!
BUILD: Error: failed to generate native component build project
C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR
OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1.
Build FAILED.
C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR
OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1.
0 Warning(s)
1 Error(s)
Time Elapsed 00:03:49.01
Build failed with exit code 1. Check errors above.
Some builds failed:
Configuration: release, Architecture: arm64
```
</details>
As you can see, there's some places where x64 is being used instead of arm64. I tried adding this line in Subsets.props after line 280:
```diff
<PropertyGroup>
<CrossDacHostArch>x64</CrossDacHostArch>
+ <CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm64'">arm64</CrossDacHostArch>
<CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm'">x86</CrossDacHostArch>
</PropertyGroup>
```
But I am still seeing an error. Here is the result:
<details><summary>Error output with modification (expand)</summary>
```powershell
❯ .\build.cmd clr+libs -c release -arch arm64
Determining projects to restore...
Tool 'coverlet.console' (version '3.1.2') was restored. Available commands: coverlet
Tool 'dotnet-reportgenerator-globaltool' (version '5.0.2') was restored. Available commands: reportgenerator
Tool 'microsoft.dotnet.xharness.cli' (version '1.0.0-prerelease.23080.1') was restored. Available commands: xharness
Tool 'microsoft.visualstudio.slngen.tool' (version '8.5.17') was restored. Available commands: slngen
Tool 'dotnet-format' (version '6.0.240501') was restored. Available commands: dotnet-format
Restore was successful.
All projects are up-to-date for restore.
Determining projects to restore...
All projects are up-to-date for restore.
System.Private.CoreLib.Generators -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib.Generators\Release\netstandard2.0\System.Private.CoreLib.Generators.dll
Microsoft.Interop.SourceGeneration -> C:\repos\runtime\artifacts\bin\Microsoft.Interop.SourceGeneration\Release\netstandard2.0\Microsoft.Interop.SourceGeneration.dll
LibraryImportGenerator -> C:\repos\runtime\artifacts\bin\LibraryImportGenerator\Release\netstandard2.0\Microsoft.Interop.LibraryImportGenerator.dll
System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib\ref\Release\net8.0\System.Private.CoreLib.dll
System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release\IL\System.Private.CoreLib.dll
Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -release -os windows -pgodatapath "C:\Users\calope\.nuget\packages\optimization.windows_nt-arm64.pgo.corecl
r\1.0.0-prerelease.23068.4"
BUILD: Starting Build at 11:53:06.31
BUILD: Commencing CoreCLR product build
BUILD: Checking prerequisites
BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe
BUILD: Number of processor cores 8
BUILD: Commencing build of native components for windows.arm64.Release
BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64
**********************************************************************
** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0
** Copyright (c) 2022 Microsoft Corporation
**********************************************************************
[vcvarsall.bat] Environment initialized for: 'x86_arm64'
BUILD: Regenerating the Visual Studio solution
Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release" vs2022 arm64 windows "-D
CMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=windows" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH="C:\Users\calope\.nuget\pa
ckages\optimization.windows_nt-arm64.pgo.coreclr\1.0.0-prerelease.23068.4"" "-DCLR_CMAKE_PGO_OPTIMIZE=1"
The CMake command line is the same as the last run. Skipping running CMake.
running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release --target install --config Release --
[1/4] Linking RC shared library dlls\mscorrc\mscorrc.dll
[2/4] Linking RC shared library dlls\clretwrc\clretwrc.dll
[3/4] cmd.exe /C "cd /D C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release\dlls\mscoree\coreclr && powershell -NoProfile -ExecutionPolicy ByPass -File C:/repos/run
time/src/coreclr/dlls/mscoree/coreclr/GenClrDebugResource.ps1 -dac C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscordac/mscordaccore.dll -dbi C:/repos/
runtime/artifacts/obj/coreclr/windows.arm64.Releas-- Install configuration: "Release"
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./clretwrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/clretwrc.pdb
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/clretwrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/clretwrc.pdb
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./mscorrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/mscorrc.pdb
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/mscorrc.dll
-- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/mscorrc.pdb
e/dlls/mscordbi/mscordbi.dll -out C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscoree/coreclr/clr_debug_resource.bin"
[3/4] Install the project...
BUILD: Build succeeded. Finished at 11:53:16.15
BUILD: Product binaries are available at C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release
Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1" -release -os linux -hostarch arm64 -component crosscomponents
BUILD: Starting Build at 11:53:16.49
BUILD: Commencing CoreCLR product build
BUILD: Checking prerequisites
BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe
BUILD: Number of processor cores 8
BUILD: Commencing build of native components for linux.arm64.Release
BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64
**********************************************************************
** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0
** Copyright (c) 2022 Microsoft Corporation
**********************************************************************
[vcvarsall.bat] Environment initialized for: 'x86_arm64'
BUILD: Regenerating the Visual Studio solution
Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release" vs2022 arm64 linux "-DCMAK
E_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" "
-DCLR_CROSS_COMPONENTS_BUILD=1"
The CMake command line is the same as the last run. Skipping running CMake.
running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release --target crosscomponents --config Release --
ninja: error: loading 'build.ninja': The system cannot find the file specified.
BUILD: Error: native component build failed. Refer to the build log files for details.
"C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.log"
"C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.wrn"
"C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.err"
C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1
" -release -os linux -hostarch arm64 -component crosscomponents" exited with code 1.
```
</details>
Am I missing something or executing the command wrong? I made sure to install all the required dependencies as described in our workflow instructions.
I attached a couple of binlogs in case they are useful:
- Without any code changes: [original_msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10970321/original_msbuild.binlog.zip)
- After my modification: [msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10960289/msbuild.binlog.zip)
@jkoritzinsky @hoyosjs | infrastructure | can t build in an windows machine fails in crosscompontents step machine surface pro x with windows command build cmd clr libs c release arch repros with or without the arch argument this is the error that shows up without any code changes original error output expand environment initialized for build regenerating the visual studio solution calling c repos runtime eng native gen buildsys cmd c repos runtime src coreclr c repos runtime artifacts obj coreclr linux release linux dcmake build type release dclr cmake target arch dclr cmake target os linux dclr cmake pgo instrument dclr cmake optdata path dclr cmake pgo optimize dclr cross components build not searching for unused variables given on the command line the c compiler identification is msvc the cxx compiler identification is msvc detecting c compiler abi info detecting c compiler abi info failed check for working c compiler c program files microsoft visual studio preview vc tools msvc bin cl exe check for working c compiler c program files microsoft visual studio preview vc tools msvc bin cl exe broken cmake error at c program files cmake share cmake modules cmaketestccompiler cmake message the c compiler c program files microsoft visual studio preview vc tools msvc bin cl exe is not able to compile a simple test program it fails with the following output change dir c repos runtime artifacts obj coreclr linux release cmakefiles cmakescratch trycompile cyevcr run build command s c progra ninja ninja exe v cmtc c progra micros preview vc tools msvc bin cl exe nologo d windows zi od mdd showincludes focmakefiles cmtc dir testccompiler c obj fdcmakefiles cmtc dir fs c c repos runtime artifacts obj coreclr linux release cmakefiles cmakescratch trycompile cyevcr testccompiler c cmd exe c cd c program files cmake bin cmake exe e vs link exe intdir cmakefiles cmtc dir rc c progra bin rc exe mt c progra bin mt exe manifests c progra micros preview vc tools msvc bin link exe nologo cmakefiles cmtc dir testccompiler c obj out cmtc exe implib cmtc lib pdb cmtc pdb versi on machine debug incremental subsystem console lib lib lib winspool lib lib lib lib uuid lib lib lib cd failed cmtc exe cmd exe c cd c program files cmake bin cmake exe e vs link exe intdir cmakefiles cmtc dir rc c progra bin rc exe mt c progra bin mt exe manifests c progra micros preview vc tools msvc bin link exe nologo cmakefiles cmtc dir testccompiler c obj out cmtc exe implib cmtc lib pdb cmtc pdb version machine debug incremental subsystem console lib lib lib winspool lib lib lib lib uuid lib comdl lib lib cd link pass command c progra micros preview vc tools msvc bin link exe nologo cmakefiles cmtc dir testcco mpiler c obj out cmtc exe implib cmtc lib pdb cmtc pdb version machine debug incremental subsystem console l ib lib lib winspool lib lib lib lib uuid lib lib lib manifest manifestfile cmakefiles cmtc dir intermediate manifest cmakefiles cmtc dir manifest res failed exit code with the following output testccompiler c obj error unresolved external symbol rtc initbase testccompiler c obj error unresolved external symbol rtc shutdown link error unresolved external symbol maincrtstartup c program files microsoft visual studio preview vc tools msvc lib msvcrtd lib warning library machine type conflicts with target machine type c program files microsoft visual studio preview vc tools msvc lib oldnames lib warning library machine type arm conflicts with target machine type cmtc exe fatal error unresolved externals ninja build stopped subcommand failed cmake will not be able to correctly generate this project call stack most recent call first cmakelists txt project configuring incomplete errors occurred build error failed to generate native component build project c repos runtime src coreclr runtime proj error the command c repos runtime src coreclr build runtime cmd cmakeargs dclr cr oss components build release os linux hostarch component crosscomponents exited with code build failed c repos runtime src coreclr runtime proj error the command c repos runtime src coreclr build runtime cmd cmakeargs dclr cr oss components build release os linux hostarch component crosscomponents exited with code warning s error s time elapsed build failed with exit code check errors above some builds failed configuration release architecture as you can see there s some places where is being used instead of i tried adding this line in subsets props after line diff but i am still seeing an error here is the result error output with modification expand powershell ❯ build cmd clr libs c release arch determining projects to restore tool coverlet console version was restored available commands coverlet tool dotnet reportgenerator globaltool version was restored available commands reportgenerator tool microsoft dotnet xharness cli version prerelease was restored available commands xharness tool microsoft visualstudio slngen tool version was restored available commands slngen tool dotnet format version was restored available commands dotnet format restore was successful all projects are up to date for restore determining projects to restore all projects are up to date for restore system private corelib generators c repos runtime artifacts bin system private corelib generators release system private corelib generators dll microsoft interop sourcegeneration c repos runtime artifacts bin microsoft interop sourcegeneration release microsoft interop sourcegeneration dll libraryimportgenerator c repos runtime artifacts bin libraryimportgenerator release microsoft interop libraryimportgenerator dll system private corelib c repos runtime artifacts bin system private corelib ref release system private corelib dll system private corelib c repos runtime artifacts bin coreclr windows release il system private corelib dll executing c repos runtime src coreclr build runtime cmd release os windows pgodatapath c users calope nuget packages optimization windows nt pgo corecl r prerelease build starting build at build commencing coreclr product build build checking prerequisites build using cmake from c program files cmake bin cmake exe build number of processor cores build commencing build of native components for windows release build using environment c program files microsoft visual studio preview vc auxiliary build vcvarsall bat visual studio developer command prompt pre copyright c microsoft corporation environment initialized for build regenerating the visual studio solution calling c repos runtime eng native gen buildsys cmd c repos runtime src coreclr c repos runtime artifacts obj coreclr windows release windows d cmake build type release dclr cmake target arch dclr cmake target os windows dclr cmake pgo instrument dclr cmake optdata path c users calope nuget pa ckages optimization windows nt pgo coreclr prerelease dclr cmake pgo optimize the cmake command line is the same as the last run skipping running cmake running c program files cmake bin cmake exe build c repos runtime artifacts obj coreclr windows release target install config release linking rc shared library dlls mscorrc mscorrc dll linking rc shared library dlls clretwrc clretwrc dll cmd exe c cd d c repos runtime artifacts obj coreclr windows release dlls mscoree coreclr powershell noprofile executionpolicy bypass file c repos run time src coreclr dlls mscoree coreclr genclrdebugresource dac c repos runtime artifacts obj coreclr windows release dlls mscordac mscordaccore dll dbi c repos runtime artifacts obj coreclr windows releas install configuration release installing c repos runtime artifacts bin coreclr windows release clretwrc dll installing c repos runtime artifacts bin coreclr windows release pdb clretwrc pdb installing c repos runtime artifacts bin coreclr windows release sharedframework clretwrc dll installing c repos runtime artifacts bin coreclr windows release sharedframework pdb clretwrc pdb installing c repos runtime artifacts bin coreclr windows release mscorrc dll installing c repos runtime artifacts bin coreclr windows release pdb mscorrc pdb installing c repos runtime artifacts bin coreclr windows release sharedframework mscorrc dll installing c repos runtime artifacts bin coreclr windows release sharedframework pdb mscorrc pdb e dlls mscordbi mscordbi dll out c repos runtime artifacts obj coreclr windows release dlls mscoree coreclr clr debug resource bin install the project build build succeeded finished at build product binaries are available at c repos runtime artifacts bin coreclr windows release executing c repos runtime src coreclr build runtime cmd cmakeargs dclr cross components build release os linux hostarch component crosscomponents build starting build at build commencing coreclr product build build checking prerequisites build using cmake from c program files cmake bin cmake exe build number of processor cores build commencing build of native components for linux release build using environment c program files microsoft visual studio preview vc auxiliary build vcvarsall bat visual studio developer command prompt pre copyright c microsoft corporation environment initialized for build regenerating the visual studio solution calling c repos runtime eng native gen buildsys cmd c repos runtime src coreclr c repos runtime artifacts obj coreclr linux release linux dcmak e build type release dclr cmake target arch dclr cmake target os linux dclr cmake pgo instrument dclr cmake optdata path dclr cmake pgo optimize dclr cross components build the cmake command line is the same as the last run skipping running cmake running c program files cmake bin cmake exe build c repos runtime artifacts obj coreclr linux release target crosscomponents config release ninja error loading build ninja the system cannot find the file specified build error native component build failed refer to the build log files for details c repos runtime artifacts log release coreclr linux release log c repos runtime artifacts log release coreclr linux release wrn c repos runtime artifacts log release coreclr linux release err c repos runtime src coreclr runtime proj error the command c repos runtime src coreclr build runtime cmd cmakeargs dclr cross components build release os linux hostarch component crosscomponents exited with code am i missing something or executing the command wrong i made sure to install all the required dependencies as described in our workflow instructions i attached a couple of binlogs in case they are useful without any code changes after my modification jkoritzinsky hoyosjs | 1 |
8,919 | 7,735,243,683 | IssuesEvent | 2018-05-27 13:02:27 | Integreat/cms | https://api.github.com/repos/Integreat/cms | closed | Remove custom ig-mpdf sql tables | area:infrastructure ease:easy prio:low | As of #719, the custom sql tables for the ig-mpdf plugin are not longer needed for storing the cached pdfs. Hence, they should be removed for all cities. | 1.0 | Remove custom ig-mpdf sql tables - As of #719, the custom sql tables for the ig-mpdf plugin are not longer needed for storing the cached pdfs. Hence, they should be removed for all cities. | infrastructure | remove custom ig mpdf sql tables as of the custom sql tables for the ig mpdf plugin are not longer needed for storing the cached pdfs hence they should be removed for all cities | 1 |
35,161 | 30,803,757,407 | IssuesEvent | 2023-08-01 05:05:58 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Report view 2.0 | interface/infrastructure refactor | There are a few future improvements for Report View (+Grid View) that would be great, especially for working with large tables. Some of these may relate to the Speed Up changes being made and may be due to unneeded background work only apparent when working with lots of data.
1. Display empty grid when no table found (empty as never wrote data) for selected report or the report is disabled rather than the data from the first table in the list (even though the table drop down selector is blank)
2. Lazy loading of report and grid view. Is is possible to only load the grid when the user is on or selects the Data tab (maybe with a loading spinner), or use async loading of data?
3. Is it possible to have lazy loading of data to display in the grid view based on scrollbar position, or maybe only load first 50 lines until the user touches scrollbar or down arrow or page down. This may be a builtin feature of future UI controls.
These will significantly improve the UI responsiveness especially with large data tables.
The same may be true for working with graphs and large data tables. | 1.0 | Report view 2.0 - There are a few future improvements for Report View (+Grid View) that would be great, especially for working with large tables. Some of these may relate to the Speed Up changes being made and may be due to unneeded background work only apparent when working with lots of data.
1. Display empty grid when no table found (empty as never wrote data) for selected report or the report is disabled rather than the data from the first table in the list (even though the table drop down selector is blank)
2. Lazy loading of report and grid view. Is is possible to only load the grid when the user is on or selects the Data tab (maybe with a loading spinner), or use async loading of data?
3. Is it possible to have lazy loading of data to display in the grid view based on scrollbar position, or maybe only load first 50 lines until the user touches scrollbar or down arrow or page down. This may be a builtin feature of future UI controls.
These will significantly improve the UI responsiveness especially with large data tables.
The same may be true for working with graphs and large data tables. | infrastructure | report view there are a few future improvements for report view grid view that would be great especially for working with large tables some of these may relate to the speed up changes being made and may be due to unneeded background work only apparent when working with lots of data display empty grid when no table found empty as never wrote data for selected report or the report is disabled rather than the data from the first table in the list even though the table drop down selector is blank lazy loading of report and grid view is is possible to only load the grid when the user is on or selects the data tab maybe with a loading spinner or use async loading of data is it possible to have lazy loading of data to display in the grid view based on scrollbar position or maybe only load first lines until the user touches scrollbar or down arrow or page down this may be a builtin feature of future ui controls these will significantly improve the ui responsiveness especially with large data tables the same may be true for working with graphs and large data tables | 1 |
30,193 | 24,632,123,094 | IssuesEvent | 2022-10-17 03:43:26 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Weather downloader can't be closed while it is downloading | bug interface/infrastructure | If SILO is down the user can't cancel or close the dialog box. | 1.0 | Weather downloader can't be closed while it is downloading - If SILO is down the user can't cancel or close the dialog box. | infrastructure | weather downloader can t be closed while it is downloading if silo is down the user can t cancel or close the dialog box | 1 |
53,532 | 22,838,850,117 | IssuesEvent | 2022-07-12 19:30:45 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Update CTN documentation | Workgroup: ATSD Service: Geo Product: CTN | Re-organized cross sections and created a direct link here for ease of navigation | 1.0 | Update CTN documentation - Re-organized cross sections and created a direct link here for ease of navigation | non_infrastructure | update ctn documentation re organized cross sections and created a direct link here for ease of navigation | 0 |
241,457 | 20,142,751,866 | IssuesEvent | 2022-02-09 02:07:50 | kevinlizh1992/rtterror | https://api.github.com/repos/kevinlizh1992/rtterror | opened | Service Frontend Small Bugs Summary Report | bug Medium Priority acceptance-test | # Bug Report
---
## 1.Bug URL:
`/service`
## 2.Date Created:
During release 2.
## 3.Defect Description:
During user testing, a number of issues were found in the Service module interface. All issues only affect a small functionality; therefore, they may require creating a separate issue report or not.
- [ ] Service code is not shown on the card.
- [ ] Service price is not shown on the card.
- [ ] When creating/editing a service, there is no text (e.g. "hours") around the time slider.
- [ ] Buttons' styles are not consistent.
- [ ] Combos creation requires using checkboxes to select Service cards, whereas there are individual buttons for each Service card. The resulting interface is confusing because of the mixture.
- [ ] Adding employee to a service button texts needs better/consistent wording.
- [ ] When editing a combo, the Service list layout is misaligned.
- [ ] It is impossible to edit the combo name or price because they are auto generated from the service list. Yet there are text fields for them.
- [ ] The sidebar menu should be either floating or able to hide. Scrolling the page will now show blank area in the page.
## 4.Resolution:
A fix is needed for each of the issues listed.
## 5.Resolution Date:
Not yet resolved.
## 6.Screenshot Bug:
To be added.
---
| 1.0 | Service Frontend Small Bugs Summary Report - # Bug Report
---
## 1.Bug URL:
`/service`
## 2.Date Created:
During release 2.
## 3.Defect Description:
During user testing, a number of issues were found in the Service module interface. All issues only affect a small functionality; therefore, they may require creating a separate issue report or not.
- [ ] Service code is not shown on the card.
- [ ] Service price is not shown on the card.
- [ ] When creating/editing a service, there is no text (e.g. "hours") around the time slider.
- [ ] Buttons' styles are not consistent.
- [ ] Combos creation requires using checkboxes to select Service cards, whereas there are individual buttons for each Service card. The resulting interface is confusing because of the mixture.
- [ ] Adding employee to a service button texts needs better/consistent wording.
- [ ] When editing a combo, the Service list layout is misaligned.
- [ ] It is impossible to edit the combo name or price because they are auto generated from the service list. Yet there are text fields for them.
- [ ] The sidebar menu should be either floating or able to hide. Scrolling the page will now show blank area in the page.
## 4.Resolution:
A fix is needed for each of the issues listed.
## 5.Resolution Date:
Not yet resolved.
## 6.Screenshot Bug:
To be added.
---
| non_infrastructure | service frontend small bugs summary report bug report bug url service date created during release defect description during user testing a number of issues were found in the service module interface all issues only affect a small functionality therefore they may require creating a separate issue report or not service code is not shown on the card service price is not shown on the card when creating editing a service there is no text e g hours around the time slider buttons styles are not consistent combos creation requires using checkboxes to select service cards whereas there are individual buttons for each service card the resulting interface is confusing because of the mixture adding employee to a service button texts needs better consistent wording when editing a combo the service list layout is misaligned it is impossible to edit the combo name or price because they are auto generated from the service list yet there are text fields for them the sidebar menu should be either floating or able to hide scrolling the page will now show blank area in the page resolution a fix is needed for each of the issues listed resolution date not yet resolved screenshot bug to be added | 0 |
26,479 | 20,151,830,769 | IssuesEvent | 2022-02-09 13:09:45 | openforis/fra-platform | https://api.github.com/repos/openforis/fra-platform | closed | Data migration | infrastructure | - [x] Basic tables
- [x] add variable name and column names
- Custom tables
- [x] degraded forest
- [x] extentOfForest
- [x] forestCharacteristics
- [x] growingStock
- [x] sustainable development (s)
| 1.0 | Data migration - - [x] Basic tables
- [x] add variable name and column names
- Custom tables
- [x] degraded forest
- [x] extentOfForest
- [x] forestCharacteristics
- [x] growingStock
- [x] sustainable development (s)
| infrastructure | data migration basic tables add variable name and column names custom tables degraded forest extentofforest forestcharacteristics growingstock sustainable development s | 1 |
6,668 | 6,550,759,874 | IssuesEvent | 2017-09-05 12:28:10 | opencv/opencv | https://api.github.com/repos/opencv/opencv | closed | Sporadic hangs of iOS / MacOSX builds | bug category: infrastructure category: ios/osx | Usually during CMake stage,
... in `try_compile()`
... the last process is xcodebuild (no child processes)
lldb backtrace dump (Xcode 8.3.3):
```
* thread #1, queue = 'com.apple.main-thread', stop reason = signal SIGSTOP
* frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10
frame #1: 0x00007fffa5f0d848 libsystem_kernel.dylib`mach_msg + 232
frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212
frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361
frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420
frame #5: 0x00007fff906a6b91 CoreFoundation`CFRunLoopRun + 97
frame #6: 0x0000000106ccbc59 Xcode3Core`-[Xcode3CommandLineBuildTool _buildWithTimingSection:] + 15332
frame #7: 0x0000000106cd6762 Xcode3Core`-[Xcode3CommandLineBuildTool run] + 4807
frame #8: 0x00000001064cc2af xcodebuild`___lldb_unnamed_symbol1$$xcodebuild + 1311
frame #9: 0x00007fffa5de7235 libdyld.dylib`start + 1
frame #10: 0x00007fffa5de7235 libdyld.dylib`start + 1
thread #2, queue = '<IDEBuildOperation:0x7f8589e47cd0:REfc>-builder-queue :: NSOperation 0x7f8589cae3d0 (QOS: DEFAULT)'
frame #0: 0x00007fffa5f15c22 libsystem_kernel.dylib`__psynch_mutexwait + 10
frame #1: 0x00007fffa6000dfa libsystem_pthread.dylib`_pthread_mutex_lock_wait + 100
frame #2: 0x00007fffa5ffe519 libsystem_pthread.dylib`_pthread_mutex_lock_slow + 285
frame #3: 0x00007fff92069ee7 Foundation`-[NSRecursiveLock lock] + 22
frame #4: 0x000000010d9aa02d DevToolsCore`-[PBXTargetBuildContext lockDependencyGraph] + 52
frame #5: 0x000000010d9c859f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 126
frame #6: 0x000000010d95e15a DevToolsCore`-[PBXTarget(XCBuildables) buildWillStartForBuilder:buildContext:buildParameters:] + 804
frame #7: 0x000000010db7f492 DevToolsCore`-[Xcode3TargetBuildableSnapshot buildForBuilderWillStart:] + 1144
frame #8: 0x00000001075fc4b7 IDEFoundation`-[IDEBuildableSnapshot performBuildForBuilder:buildCommand:buildOnlyTheseFiles:] + 112
frame #9: 0x000000010762ee06 IDEFoundation`-[IDEBuilder primitiveMain] + 1253
frame #10: 0x000000010762e849 IDEFoundation`-[IDEBuilder main] + 264
frame #11: 0x00007fff9209a324 Foundation`-[__NSOperationInternal _start:] + 672
frame #12: 0x00007fff920961db Foundation`__NSOQSchedule_f + 201
frame #13: 0x00007fffa5db18fc libdispatch.dylib`_dispatch_client_callout + 8
frame #14: 0x00007fffa5dc79a0 libdispatch.dylib`_dispatch_queue_serial_drain + 896
frame #15: 0x00007fffa5dba306 libdispatch.dylib`_dispatch_queue_invoke + 1046
frame #16: 0x00007fffa5db36b5 libdispatch.dylib`_dispatch_root_queue_drain + 476
frame #17: 0x00007fffa5db348c libdispatch.dylib`_dispatch_worker_thread3 + 99
frame #18: 0x00007fffa60005a2 libsystem_pthread.dylib`_pthread_wqthread + 1299
frame #19: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13
thread #3
frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10
frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55
frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212
frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361
frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420
frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277
frame #6: 0x00000001108b5143 DTDeviceKitBase`+[DTDKRemoteDeviceConnection startServiceBrowsers] + 217
frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #4
frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10
frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55
frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212
frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361
frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420
frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277
frame #6: 0x00000001108c3df1 DTDeviceKitBase`-[DTDKRemoteDeviceDataListener listenerThreadImplementation] + 1454
frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #5, name = 'com.apple.CFSocket.private'
frame #0: 0x00007fffa5f15eb6 libsystem_kernel.dylib`__select + 10
frame #1: 0x00007fff906a366a CoreFoundation`__CFSocketManager + 682
frame #2: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #3: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #4: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #6
frame #0: 0x00007fffa5f1644e libsystem_kernel.dylib`__workq_kernreturn + 10
frame #1: 0x00007fffa600048e libsystem_pthread.dylib`_pthread_wqthread + 1023
frame #2: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13
thread #7
frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10
frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712
frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594
frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #8
frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10
frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712
frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594
frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #9
frame #0: 0x00007fffa5f0e386 libsystem_kernel.dylib`semaphore_wait_trap + 10
frame #1: 0x00007fffa5dc9a77 libdispatch.dylib`_os_semaphore_wait + 16
frame #2: 0x00007fffa5db77e9 libdispatch.dylib`_dispatch_group_wait_slow + 154
frame #3: 0x000000010669bf14 DVTFoundation`-[DVTFuture _waitUntilFinished] + 137
frame #4: 0x000000010669c1e8 DVTFoundation`-[DVTFuture result] + 22
frame #5: 0x000000010dadab34 DevToolsCore`+[XCCodeSignContext provisioningCommandInputsFromMacroExpansionScope:] + 456
frame #6: 0x000000010dadc8f1 DevToolsCore`+[XCCodeSignContext prepareForCodeSigningWithMacroExpansionScope:] + 172
frame #7: 0x000000010da37aee DevToolsCore`-[XCProductTypeSpecification computeDependenciesWithMacroExpansionScope:] + 1683
frame #8: 0x000000010daed9d3 DevToolsCore`-[XCNativeTargetDGSnapshot(DependencyGraphCreation) computeDependenciesWithMacroExpansionScope:] + 4140
frame #9: 0x000000010d9ad0aa DevToolsCore`-[PBXTargetBuildContext createDependencyGraphWithTargetDGSnapshot:] + 4297
frame #10: 0x000000010d9c8f9b DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) dg_setTargetSnapshot:] + 514
frame #11: 0x000000010d9c863f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 286
frame #12: 0x000000010da2e355 DevToolsCore`-[XCBlockQueue _processNextBlockInThreadSlotNumber:] + 463
frame #13: 0x000000010da2e6fa DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 298
frame #14: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #15: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #16: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #17: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
``` | 1.0 | Sporadic hangs of iOS / MacOSX builds - Usually during CMake stage,
... in `try_compile()`
... the last process is xcodebuild (no child processes)
lldb backtrace dump (Xcode 8.3.3):
```
* thread #1, queue = 'com.apple.main-thread', stop reason = signal SIGSTOP
* frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10
frame #1: 0x00007fffa5f0d848 libsystem_kernel.dylib`mach_msg + 232
frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212
frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361
frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420
frame #5: 0x00007fff906a6b91 CoreFoundation`CFRunLoopRun + 97
frame #6: 0x0000000106ccbc59 Xcode3Core`-[Xcode3CommandLineBuildTool _buildWithTimingSection:] + 15332
frame #7: 0x0000000106cd6762 Xcode3Core`-[Xcode3CommandLineBuildTool run] + 4807
frame #8: 0x00000001064cc2af xcodebuild`___lldb_unnamed_symbol1$$xcodebuild + 1311
frame #9: 0x00007fffa5de7235 libdyld.dylib`start + 1
frame #10: 0x00007fffa5de7235 libdyld.dylib`start + 1
thread #2, queue = '<IDEBuildOperation:0x7f8589e47cd0:REfc>-builder-queue :: NSOperation 0x7f8589cae3d0 (QOS: DEFAULT)'
frame #0: 0x00007fffa5f15c22 libsystem_kernel.dylib`__psynch_mutexwait + 10
frame #1: 0x00007fffa6000dfa libsystem_pthread.dylib`_pthread_mutex_lock_wait + 100
frame #2: 0x00007fffa5ffe519 libsystem_pthread.dylib`_pthread_mutex_lock_slow + 285
frame #3: 0x00007fff92069ee7 Foundation`-[NSRecursiveLock lock] + 22
frame #4: 0x000000010d9aa02d DevToolsCore`-[PBXTargetBuildContext lockDependencyGraph] + 52
frame #5: 0x000000010d9c859f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 126
frame #6: 0x000000010d95e15a DevToolsCore`-[PBXTarget(XCBuildables) buildWillStartForBuilder:buildContext:buildParameters:] + 804
frame #7: 0x000000010db7f492 DevToolsCore`-[Xcode3TargetBuildableSnapshot buildForBuilderWillStart:] + 1144
frame #8: 0x00000001075fc4b7 IDEFoundation`-[IDEBuildableSnapshot performBuildForBuilder:buildCommand:buildOnlyTheseFiles:] + 112
frame #9: 0x000000010762ee06 IDEFoundation`-[IDEBuilder primitiveMain] + 1253
frame #10: 0x000000010762e849 IDEFoundation`-[IDEBuilder main] + 264
frame #11: 0x00007fff9209a324 Foundation`-[__NSOperationInternal _start:] + 672
frame #12: 0x00007fff920961db Foundation`__NSOQSchedule_f + 201
frame #13: 0x00007fffa5db18fc libdispatch.dylib`_dispatch_client_callout + 8
frame #14: 0x00007fffa5dc79a0 libdispatch.dylib`_dispatch_queue_serial_drain + 896
frame #15: 0x00007fffa5dba306 libdispatch.dylib`_dispatch_queue_invoke + 1046
frame #16: 0x00007fffa5db36b5 libdispatch.dylib`_dispatch_root_queue_drain + 476
frame #17: 0x00007fffa5db348c libdispatch.dylib`_dispatch_worker_thread3 + 99
frame #18: 0x00007fffa60005a2 libsystem_pthread.dylib`_pthread_wqthread + 1299
frame #19: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13
thread #3
frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10
frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55
frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212
frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361
frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420
frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277
frame #6: 0x00000001108b5143 DTDeviceKitBase`+[DTDKRemoteDeviceConnection startServiceBrowsers] + 217
frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #4
frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10
frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55
frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212
frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361
frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420
frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277
frame #6: 0x00000001108c3df1 DTDeviceKitBase`-[DTDKRemoteDeviceDataListener listenerThreadImplementation] + 1454
frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #5, name = 'com.apple.CFSocket.private'
frame #0: 0x00007fffa5f15eb6 libsystem_kernel.dylib`__select + 10
frame #1: 0x00007fff906a366a CoreFoundation`__CFSocketManager + 682
frame #2: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #3: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #4: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #6
frame #0: 0x00007fffa5f1644e libsystem_kernel.dylib`__workq_kernreturn + 10
frame #1: 0x00007fffa600048e libsystem_pthread.dylib`_pthread_wqthread + 1023
frame #2: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13
thread #7
frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10
frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712
frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594
frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #8
frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10
frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712
frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594
frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
thread #9
frame #0: 0x00007fffa5f0e386 libsystem_kernel.dylib`semaphore_wait_trap + 10
frame #1: 0x00007fffa5dc9a77 libdispatch.dylib`_os_semaphore_wait + 16
frame #2: 0x00007fffa5db77e9 libdispatch.dylib`_dispatch_group_wait_slow + 154
frame #3: 0x000000010669bf14 DVTFoundation`-[DVTFuture _waitUntilFinished] + 137
frame #4: 0x000000010669c1e8 DVTFoundation`-[DVTFuture result] + 22
frame #5: 0x000000010dadab34 DevToolsCore`+[XCCodeSignContext provisioningCommandInputsFromMacroExpansionScope:] + 456
frame #6: 0x000000010dadc8f1 DevToolsCore`+[XCCodeSignContext prepareForCodeSigningWithMacroExpansionScope:] + 172
frame #7: 0x000000010da37aee DevToolsCore`-[XCProductTypeSpecification computeDependenciesWithMacroExpansionScope:] + 1683
frame #8: 0x000000010daed9d3 DevToolsCore`-[XCNativeTargetDGSnapshot(DependencyGraphCreation) computeDependenciesWithMacroExpansionScope:] + 4140
frame #9: 0x000000010d9ad0aa DevToolsCore`-[PBXTargetBuildContext createDependencyGraphWithTargetDGSnapshot:] + 4297
frame #10: 0x000000010d9c8f9b DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) dg_setTargetSnapshot:] + 514
frame #11: 0x000000010d9c863f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 286
frame #12: 0x000000010da2e355 DevToolsCore`-[XCBlockQueue _processNextBlockInThreadSlotNumber:] + 463
frame #13: 0x000000010da2e6fa DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 298
frame #14: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243
frame #15: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180
frame #16: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286
frame #17: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13
``` | infrastructure | sporadic hangs of ios macosx builds usually during cmake stage in try compile the last process is xcodebuild no child processes lldb backtrace dump xcode thread queue com apple main thread stop reason signal sigstop frame libsystem kernel dylib mach msg trap frame libsystem kernel dylib mach msg frame corefoundation cfrunloopservicemachport frame corefoundation cfrunlooprun frame corefoundation cfrunlooprunspecific frame corefoundation cfrunlooprun frame frame frame xcodebuild lldb unnamed xcodebuild frame libdyld dylib start frame libdyld dylib start thread queue builder queue nsoperation qos default frame libsystem kernel dylib psynch mutexwait frame libsystem pthread dylib pthread mutex lock wait frame libsystem pthread dylib pthread mutex lock slow frame foundation frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame idefoundation frame idefoundation frame idefoundation frame foundation frame foundation nsoqschedule f frame libdispatch dylib dispatch client callout frame libdispatch dylib dispatch queue serial drain frame libdispatch dylib dispatch queue invoke frame libdispatch dylib dispatch root queue drain frame libdispatch dylib dispatch worker frame libsystem pthread dylib pthread wqthread frame libsystem pthread dylib start wqthread thread frame libsystem kernel dylib mach msg trap frame libsystem kernel dylib mach msg frame corefoundation cfrunloopservicemachport frame corefoundation cfrunlooprun frame corefoundation cfrunlooprunspecific frame foundation frame dtdevicekitbase frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib mach msg trap frame libsystem kernel dylib mach msg frame corefoundation cfrunloopservicemachport frame corefoundation cfrunlooprun frame corefoundation cfrunlooprunspecific frame foundation frame dtdevicekitbase frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread name com apple cfsocket private frame libsystem kernel dylib select frame corefoundation cfsocketmanager frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib workq kernreturn frame libsystem pthread dylib pthread wqthread frame libsystem pthread dylib start wqthread thread frame libsystem kernel dylib psynch cvwait frame libsystem pthread dylib pthread cond wait frame devtoolscore frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib psynch cvwait frame libsystem pthread dylib pthread cond wait frame devtoolscore frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib semaphore wait trap frame libdispatch dylib os semaphore wait frame libdispatch dylib dispatch group wait slow frame dvtfoundation frame dvtfoundation frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start | 1 |
15,408 | 11,498,650,197 | IssuesEvent | 2020-02-12 12:26:48 | eclipse/antenna | https://api.github.com/repos/eclipse/antenna | opened | Re-design API of sw360-client library | infrastructure | ### Summary of the Improvement
Since the sw360-client has been established as an own library new requirements exist, since the library should also be usable in a non-antenna context. Meaning, functionalities antenna does not need should be provided in the client library.
e.g. most REST calls to SW360 should be possible with this library
This could also have some non-functional aspects like
- blocking vs non-blocking rest calls.
- making use of a custom or specific Http-library
- configuration of the client could be improved / simplified
- having a central entry point into the library
### Acceptance Criteria
- [ ] Proposal for new API criteria should exist
- [ ] all current functionalities are still possible
### Definition of Done
- Acceptance criteria fulfilled
- The ticket describes what has been done and describes how the results can be verified (e.g. new Jenkins step is inserted, the report can be viewed at X)
- All known limitations are listed and it is described how they will be addressed (e.g., a ticket for the remaining stuff)
- Documentation is updated
- The result has been reviewed: Everything works as described and can be verified reading only the documentation
| 1.0 | Re-design API of sw360-client library - ### Summary of the Improvement
Since the sw360-client has been established as an own library new requirements exist, since the library should also be usable in a non-antenna context. Meaning, functionalities antenna does not need should be provided in the client library.
e.g. most REST calls to SW360 should be possible with this library
This could also have some non-functional aspects like
- blocking vs non-blocking rest calls.
- making use of a custom or specific Http-library
- configuration of the client could be improved / simplified
- having a central entry point into the library
### Acceptance Criteria
- [ ] Proposal for new API criteria should exist
- [ ] all current functionalities are still possible
### Definition of Done
- Acceptance criteria fulfilled
- The ticket describes what has been done and describes how the results can be verified (e.g. new Jenkins step is inserted, the report can be viewed at X)
- All known limitations are listed and it is described how they will be addressed (e.g., a ticket for the remaining stuff)
- Documentation is updated
- The result has been reviewed: Everything works as described and can be verified reading only the documentation
| infrastructure | re design api of client library summary of the improvement since the client has been established as an own library new requirements exist since the library should also be usable in a non antenna context meaning functionalities antenna does not need should be provided in the client library e g most rest calls to should be possible with this library this could also have some non functional aspects like blocking vs non blocking rest calls making use of a custom or specific http library configuration of the client could be improved simplified having a central entry point into the library acceptance criteria proposal for new api criteria should exist all current functionalities are still possible definition of done acceptance criteria fulfilled the ticket describes what has been done and describes how the results can be verified e g new jenkins step is inserted the report can be viewed at x all known limitations are listed and it is described how they will be addressed e g a ticket for the remaining stuff documentation is updated the result has been reviewed everything works as described and can be verified reading only the documentation | 1 |
14,357 | 10,763,074,230 | IssuesEvent | 2019-11-01 02:23:48 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Add a "Fertilised" event to Fertiliser | bug interface/infrastructure newfeature | I am writing some documentation for Report and have realised that there are no events for Fertiliser. Can a "Fertlised" or similar event be added. This would make it more consistent with Irrigation and easier to report on a fertilisation. | 1.0 | Add a "Fertilised" event to Fertiliser - I am writing some documentation for Report and have realised that there are no events for Fertiliser. Can a "Fertlised" or similar event be added. This would make it more consistent with Irrigation and easier to report on a fertilisation. | infrastructure | add a fertilised event to fertiliser i am writing some documentation for report and have realised that there are no events for fertiliser can a fertlised or similar event be added this would make it more consistent with irrigation and easier to report on a fertilisation | 1 |
34,757 | 30,350,911,946 | IssuesEvent | 2023-07-11 18:53:08 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | opened | Consolidate E2E Workflows | tool: monorepo infrastructure | There is a wide array of different workflows that run the E2E tests that we should review and centralize. Some of these may not be in use anymore while others may be redundant and able to be consolidated. This task involves both looking into all of these workflows as well as communicating with any stakeholders to gather feedback on how to proceed with handling the workflow. During this work we should also take steps to clean up any remaining E2E workflows and make sure they are consistent with our current GitHub Action best-practices.
* `cot-build-and-e2e-tests-daily.yml`
* `cot-pr-build-and-e2e-tests.yml`
* `pr-build-and-e2e-tests.yml`
* `pr-smoke-tests.yml`
* `smoke-test-daily-site-check.yml`
* `smoke-test-daily.yml`
* `smoke-test-pr-merge.yml`
* `smoke-test-release.yml`
## Acceptance Criteria
* Review all of our workflows and make sure that the list above is exhaustive in the inclusion of E2E test execution.
* Identify whether a workflow is in use, and if it isn't, ask stakeholders whether it can just be deleted.
* For workflows that could be consolidated, such as the COT ones, combine them into single workflows.
* With the workflows that may be redundant, such as all of the smoke test ones, reach out to stakeholders for possible guidance on how we might consolidate them and do so.
| 1.0 | Consolidate E2E Workflows - There is a wide array of different workflows that run the E2E tests that we should review and centralize. Some of these may not be in use anymore while others may be redundant and able to be consolidated. This task involves both looking into all of these workflows as well as communicating with any stakeholders to gather feedback on how to proceed with handling the workflow. During this work we should also take steps to clean up any remaining E2E workflows and make sure they are consistent with our current GitHub Action best-practices.
* `cot-build-and-e2e-tests-daily.yml`
* `cot-pr-build-and-e2e-tests.yml`
* `pr-build-and-e2e-tests.yml`
* `pr-smoke-tests.yml`
* `smoke-test-daily-site-check.yml`
* `smoke-test-daily.yml`
* `smoke-test-pr-merge.yml`
* `smoke-test-release.yml`
## Acceptance Criteria
* Review all of our workflows and make sure that the list above is exhaustive in the inclusion of E2E test execution.
* Identify whether a workflow is in use, and if it isn't, ask stakeholders whether it can just be deleted.
* For workflows that could be consolidated, such as the COT ones, combine them into single workflows.
* With the workflows that may be redundant, such as all of the smoke test ones, reach out to stakeholders for possible guidance on how we might consolidate them and do so.
| infrastructure | consolidate workflows there is a wide array of different workflows that run the tests that we should review and centralize some of these may not be in use anymore while others may be redundant and able to be consolidated this task involves both looking into all of these workflows as well as communicating with any stakeholders to gather feedback on how to proceed with handling the workflow during this work we should also take steps to clean up any remaining workflows and make sure they are consistent with our current github action best practices cot build and tests daily yml cot pr build and tests yml pr build and tests yml pr smoke tests yml smoke test daily site check yml smoke test daily yml smoke test pr merge yml smoke test release yml acceptance criteria review all of our workflows and make sure that the list above is exhaustive in the inclusion of test execution identify whether a workflow is in use and if it isn t ask stakeholders whether it can just be deleted for workflows that could be consolidated such as the cot ones combine them into single workflows with the workflows that may be redundant such as all of the smoke test ones reach out to stakeholders for possible guidance on how we might consolidate them and do so | 1 |
32,091 | 26,420,215,276 | IssuesEvent | 2023-01-13 19:39:39 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | closed | Test failure in TopicPracticeFragmentTest | Priority: Essential good first issue issue_type_infrastructure issue_user_impact_low issue_temp_ben_triaged issue_user_developer | <img width="1542" alt="Screenshot 2021-06-23 at 6 33 21 PM" src="https://user-images.githubusercontent.com/9396084/123101322-8e505180-d451-11eb-8f2b-d5513968cd5c.png">
| 1.0 | Test failure in TopicPracticeFragmentTest - <img width="1542" alt="Screenshot 2021-06-23 at 6 33 21 PM" src="https://user-images.githubusercontent.com/9396084/123101322-8e505180-d451-11eb-8f2b-d5513968cd5c.png">
| infrastructure | test failure in topicpracticefragmenttest img width alt screenshot at pm src | 1 |
18,491 | 10,235,676,805 | IssuesEvent | 2019-08-19 09:42:39 | jpmorganchase/tessera | https://api.github.com/repos/jpmorganchase/tessera | closed | com.fasterxml.jackson.core:jackson-databind:2.9.9.1 marked vulnerable | security | The dependency `com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1` is marked as having vulnerabilities by the Enforcer plugin.
A newer version should be used which does not contain they flagged issues.
---
Failure message:
```
[WARNING] Rule 3: org.sonatype.ossindex.maven.enforcer.BanVulnerableDependencies failed with message:
Detected 1 vulnerable components:
com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1:compile; https://ossindex.sonatype.org/component/pkg:maven/com.fasterxml.jackson.core/jackson-databind@2.9.9.1
* [CVE-2019-14439] A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x befo... (7.5); https://ossindex.sonatype.org/vuln/ac9dce23-7b35-4691-b05e-a68f58d48b8c
* [CVE-2019-14379] SubTypeValidator.java in FasterXML jackson-databind before 2.9.9.2 mishandles de... (9.8); https://ossindex.sonatype.org/vuln/e5794172-1257-4372-9baf-7b87307a3cc9
``` | True | com.fasterxml.jackson.core:jackson-databind:2.9.9.1 marked vulnerable - The dependency `com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1` is marked as having vulnerabilities by the Enforcer plugin.
A newer version should be used which does not contain they flagged issues.
---
Failure message:
```
[WARNING] Rule 3: org.sonatype.ossindex.maven.enforcer.BanVulnerableDependencies failed with message:
Detected 1 vulnerable components:
com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1:compile; https://ossindex.sonatype.org/component/pkg:maven/com.fasterxml.jackson.core/jackson-databind@2.9.9.1
* [CVE-2019-14439] A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x befo... (7.5); https://ossindex.sonatype.org/vuln/ac9dce23-7b35-4691-b05e-a68f58d48b8c
* [CVE-2019-14379] SubTypeValidator.java in FasterXML jackson-databind before 2.9.9.2 mishandles de... (9.8); https://ossindex.sonatype.org/vuln/e5794172-1257-4372-9baf-7b87307a3cc9
``` | non_infrastructure | com fasterxml jackson core jackson databind marked vulnerable the dependency com fasterxml jackson core jackson databind jar is marked as having vulnerabilities by the enforcer plugin a newer version should be used which does not contain they flagged issues failure message rule org sonatype ossindex maven enforcer banvulnerabledependencies failed with message detected vulnerable components com fasterxml jackson core jackson databind jar compile a polymorphic typing issue was discovered in fasterxml jackson databind x befo subtypevalidator java in fasterxml jackson databind before mishandles de | 0 |
19,697 | 13,395,488,914 | IssuesEvent | 2020-09-03 08:30:31 | onivim/oni2 | https://api.github.com/repos/onivim/oni2 | closed | Can't build the frontend | A-infrastructure platform-linux | Following the instructions from https://onivim.github.io/docs/for-developers/building
```bash
git clone https://github.com/onivim/oni2
cd oni2
esy install
```
Last step fails with some problem about `@esy-ocaml/substs`:
```
error unable to fetch @esy-ocaml/substs@0.0.1:
source: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46
error: error: run ['shasum' '--algorithm' '1'
'/tmp/oni2/_esy/default/tmp/esy-f92046/archive']: exited with 127
fetching dist: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46
error: installation error
fetching @esy-ocaml/substs@0.0.1
esy: exiting due to errors above
```
I checked and I can install `@esy-ocaml/substs@0.0.1` with npm successfully.
If I try to install it with `esy`, it fails as well (with the same error as above).
The OS is Fedora 32. | 1.0 | Can't build the frontend - Following the instructions from https://onivim.github.io/docs/for-developers/building
```bash
git clone https://github.com/onivim/oni2
cd oni2
esy install
```
Last step fails with some problem about `@esy-ocaml/substs`:
```
error unable to fetch @esy-ocaml/substs@0.0.1:
source: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46
error: error: run ['shasum' '--algorithm' '1'
'/tmp/oni2/_esy/default/tmp/esy-f92046/archive']: exited with 127
fetching dist: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46
error: installation error
fetching @esy-ocaml/substs@0.0.1
esy: exiting due to errors above
```
I checked and I can install `@esy-ocaml/substs@0.0.1` with npm successfully.
If I try to install it with `esy`, it fails as well (with the same error as above).
The OS is Fedora 32. | infrastructure | can t build the frontend following the instructions from bash git clone cd esy install last step fails with some problem about esy ocaml substs error unable to fetch esy ocaml substs source archive error error run shasum algorithm tmp esy default tmp esy archive exited with fetching dist archive error installation error fetching esy ocaml substs esy exiting due to errors above i checked and i can install esy ocaml substs with npm successfully if i try to install it with esy it fails as well with the same error as above the os is fedora | 1 |
269,400 | 8,435,411,584 | IssuesEvent | 2018-10-17 13:06:07 | kubeapps/kubeapps | https://api.github.com/repos/kubeapps/kubeapps | closed | [Design/Proposal] Global runtime error handling | component/dashboard priority/important-soon size/M | Currently, when a runtime error happens in the dashboard, the UI stops rendering but "unless specifically handled" the user does not see a message that something went wrong. This was the case https://github.com/kubeapps/kubeapps/issues/635 for example:
React 16 has the concept of errorBoundary that we should explore https://reactjs.org/docs/error-boundaries.html, potentially making it available as a wrapper at the root component.
This component could eventually contain the efforts made here https://github.com/kubeapps/kubeapps/issues/654
The goal being:
* Step 1: If a unexpected runtime error happens, we show a message to the user that `something went wrong` (I believe that we have already some markup for this that suggest to look at the console and so on)
* Step 2: This component might be able to capture some specific custom errors, showing different error types and messages https://github.com/kubeapps/kubeapps/issues/654. This currently is happening at every component level, so making it generic could help.
wdyt @andresmgot @prydonius ? | 1.0 | [Design/Proposal] Global runtime error handling - Currently, when a runtime error happens in the dashboard, the UI stops rendering but "unless specifically handled" the user does not see a message that something went wrong. This was the case https://github.com/kubeapps/kubeapps/issues/635 for example:
React 16 has the concept of errorBoundary that we should explore https://reactjs.org/docs/error-boundaries.html, potentially making it available as a wrapper at the root component.
This component could eventually contain the efforts made here https://github.com/kubeapps/kubeapps/issues/654
The goal being:
* Step 1: If a unexpected runtime error happens, we show a message to the user that `something went wrong` (I believe that we have already some markup for this that suggest to look at the console and so on)
* Step 2: This component might be able to capture some specific custom errors, showing different error types and messages https://github.com/kubeapps/kubeapps/issues/654. This currently is happening at every component level, so making it generic could help.
wdyt @andresmgot @prydonius ? | non_infrastructure | global runtime error handling currently when a runtime error happens in the dashboard the ui stops rendering but unless specifically handled the user does not see a message that something went wrong this was the case for example react has the concept of errorboundary that we should explore potentially making it available as a wrapper at the root component this component could eventually contain the efforts made here the goal being step if a unexpected runtime error happens we show a message to the user that something went wrong i believe that we have already some markup for this that suggest to look at the console and so on step this component might be able to capture some specific custom errors showing different error types and messages this currently is happening at every component level so making it generic could help wdyt andresmgot prydonius | 0 |
3,119 | 4,065,549,130 | IssuesEvent | 2016-05-26 11:44:36 | pjuu/pjuu | https://api.github.com/repos/pjuu/pjuu | closed | Jessie for docker | in progress infrastructure | Convert the docker image to Jessie rather than Wheezy.
There are some issues with the required libraries that need to be tested. | 1.0 | Jessie for docker - Convert the docker image to Jessie rather than Wheezy.
There are some issues with the required libraries that need to be tested. | infrastructure | jessie for docker convert the docker image to jessie rather than wheezy there are some issues with the required libraries that need to be tested | 1 |
13,704 | 10,428,005,741 | IssuesEvent | 2019-09-16 21:18:27 | dotnet/core-setup | https://api.github.com/repos/dotnet/core-setup | closed | The tag v2.1.13-dependencies is missing | area-Infrastructure | ## Steps to reproduce
Go to https://github.com/dotnet/core-setup/tags
## Expected behavior
The tag v2.1.13-dependencies should be present, as it was for 2.1.12:
https://github.com/dotnet/core-setup/releases/tag/v2.1.12-dependencies
## Actual behavior
The tag is not present.
## Environment data
n/a | 1.0 | The tag v2.1.13-dependencies is missing - ## Steps to reproduce
Go to https://github.com/dotnet/core-setup/tags
## Expected behavior
The tag v2.1.13-dependencies should be present, as it was for 2.1.12:
https://github.com/dotnet/core-setup/releases/tag/v2.1.12-dependencies
## Actual behavior
The tag is not present.
## Environment data
n/a | infrastructure | the tag dependencies is missing steps to reproduce go to expected behavior the tag dependencies should be present as it was for actual behavior the tag is not present environment data n a | 1 |
6,655 | 6,547,208,340 | IssuesEvent | 2017-09-04 13:49:27 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | opened | io/secure_server_closing_test times out on CQ on Mac OS | area-infrastructure os-osx | FAILED: none-vm release_x64 standalone/io/secure_server_closing_test
Expected: Pass
Actual: Timeout
https://luci-logdog.appspot.com/v/?s=dart%2Fled%2Fathom_google.com%2Fa807c9b6e6e940045dc7676a153e11294f2a9dfd32c15e063ffe80a92badc95e%2F%2B%2Fsteps%2Fvm_tests%2F0%2Fstdout
| 1.0 | io/secure_server_closing_test times out on CQ on Mac OS - FAILED: none-vm release_x64 standalone/io/secure_server_closing_test
Expected: Pass
Actual: Timeout
https://luci-logdog.appspot.com/v/?s=dart%2Fled%2Fathom_google.com%2Fa807c9b6e6e940045dc7676a153e11294f2a9dfd32c15e063ffe80a92badc95e%2F%2B%2Fsteps%2Fvm_tests%2F0%2Fstdout
| infrastructure | io secure server closing test times out on cq on mac os failed none vm release standalone io secure server closing test expected pass actual timeout | 1 |
92,376 | 8,361,666,776 | IssuesEvent | 2018-10-03 14:52:26 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: jepsen/3/register/majority-ring failed | A-core-kv C-test-failure O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/0f5a289d49105cd317a2713b71546d84c25547f6
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stressrace instead of stress and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
make stress TESTS=jepsen/3/register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m'
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=937459&tab=buildLog
```
The test failed on master:
test.go:500,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: /home/agent/work/.go/bin/roachprod run teamcity-937459-jepsen-3:1-6 -- sh -c "sudo apt-get -qqy upgrade -o Dpkg::Options::='--force-confold' > /dev/null 2>&1" returned:
stderr:
stdout:
.............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................
1:
exit status 100
2:
exit status 100
3:
exit status 100
4:
exit status 100
5:
exit status 100
6:
exit status 100
Error: exit status 100
: exit status 1
test.go:500,cluster.go:890,jepsen.go:74,asm_amd64.s:573,panic.go:377,test.go:501,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: test already failed
``` | 1.0 | roachtest: jepsen/3/register/majority-ring failed - SHA: https://github.com/cockroachdb/cockroach/commits/0f5a289d49105cd317a2713b71546d84c25547f6
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stressrace instead of stress and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
make stress TESTS=jepsen/3/register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m'
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=937459&tab=buildLog
```
The test failed on master:
test.go:500,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: /home/agent/work/.go/bin/roachprod run teamcity-937459-jepsen-3:1-6 -- sh -c "sudo apt-get -qqy upgrade -o Dpkg::Options::='--force-confold' > /dev/null 2>&1" returned:
stderr:
stdout:
.............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................
1:
exit status 100
2:
exit status 100
3:
exit status 100
4:
exit status 100
5:
exit status 100
6:
exit status 100
Error: exit status 100
: exit status 1
test.go:500,cluster.go:890,jepsen.go:74,asm_amd64.s:573,panic.go:377,test.go:501,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: test already failed
``` | non_infrastructure | roachtest jepsen register majority ring failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stressrace instead of stress and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach make stress tests jepsen register majority ring pkg roachtest testtimeout stressflags stderr false maxtime timeout failed test the test failed on master test go cluster go jepsen go jepsen go jepsen go home agent work go bin roachprod run teamcity jepsen sh c sudo apt get qqy upgrade o dpkg options force confold dev null returned stderr stdout exit status exit status exit status exit status exit status exit status error exit status exit status test go cluster go jepsen go asm s panic go test go cluster go jepsen go jepsen go jepsen go test already failed | 0 |
15,190 | 11,402,509,526 | IssuesEvent | 2020-01-31 03:30:41 | MindLeaps/tracker | https://api.github.com/repos/MindLeaps/tracker | closed | Update Ansible Playbooks | infrastructure | We are using an old ansible configuration a lot of it is invalid with the new ansible.
examples include
`sudo: true` should be `become: yes`
apt get `state=installed` should be `state=present` | 1.0 | Update Ansible Playbooks - We are using an old ansible configuration a lot of it is invalid with the new ansible.
examples include
`sudo: true` should be `become: yes`
apt get `state=installed` should be `state=present` | infrastructure | update ansible playbooks we are using an old ansible configuration a lot of it is invalid with the new ansible examples include sudo true should be become yes apt get state installed should be state present | 1 |
17,573 | 24,229,548,653 | IssuesEvent | 2022-09-26 16:59:58 | opensrp/web | https://api.github.com/repos/opensrp/web | closed | [FHIR Road Map] Create new packages for FHIR support | Discussion Ready for Engineering FHIR compatibility | **Work plan**
- Create FHIR focused packages for the following features
- [ ] Team Management --- https://github.com/OpenSRP/web/issues/552
- [ ] User Management --- https://github.com/OpenSRP/web/issues/585
- [ ] Location Management --- https://github.com/OpenSRP/web/issues/553 | True | [FHIR Road Map] Create new packages for FHIR support - **Work plan**
- Create FHIR focused packages for the following features
- [ ] Team Management --- https://github.com/OpenSRP/web/issues/552
- [ ] User Management --- https://github.com/OpenSRP/web/issues/585
- [ ] Location Management --- https://github.com/OpenSRP/web/issues/553 | non_infrastructure | create new packages for fhir support work plan create fhir focused packages for the following features team management user management location management | 0 |
390,913 | 11,565,690,507 | IssuesEvent | 2020-02-20 10:58:49 | luna/enso | https://api.github.com/repos/luna/enso | closed | Clean Up `AstExpression` | Category: Compiler Change: Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement | ### Summary
The file defining the old AST format is a bit of a mess. As it's going to be used as the basis for the IR in the stopgap compiler, it needs a bit of cleaning up.
### Value
My sanity stays slightly more intact.
### Specification
- [x] Clean up the structure of the code.
- [x] Slightly alter the hierarchy of the nodes.
- [x] Document the AST nodes.
- [x] Add additional nodes to represent errors.
- [x] Add a node to allow for tagging arbitrary nodes with additional data (as an `HList`).
### Acceptance Criteria & Test Cases
- All existing tests pass.
| 1.0 | Clean Up `AstExpression` - ### Summary
The file defining the old AST format is a bit of a mess. As it's going to be used as the basis for the IR in the stopgap compiler, it needs a bit of cleaning up.
### Value
My sanity stays slightly more intact.
### Specification
- [x] Clean up the structure of the code.
- [x] Slightly alter the hierarchy of the nodes.
- [x] Document the AST nodes.
- [x] Add additional nodes to represent errors.
- [x] Add a node to allow for tagging arbitrary nodes with additional data (as an `HList`).
### Acceptance Criteria & Test Cases
- All existing tests pass.
| non_infrastructure | clean up astexpression summary the file defining the old ast format is a bit of a mess as it s going to be used as the basis for the ir in the stopgap compiler it needs a bit of cleaning up value my sanity stays slightly more intact specification clean up the structure of the code slightly alter the hierarchy of the nodes document the ast nodes add additional nodes to represent errors add a node to allow for tagging arbitrary nodes with additional data as an hlist acceptance criteria test cases all existing tests pass | 0 |
28,847 | 8,214,348,773 | IssuesEvent | 2018-09-04 22:54:27 | angular/devkit | https://api.github.com/repos/angular/devkit | closed | Purify doesn't work with NgModuleFactory in non-webpack environment | comp: angular-devkit/build-optimizer type: feature | Relates to https://github.com/angular/material2/issues/4137 .
Currently purify plugin only supports hard-coded webpack regex in https://github.com/angular/devkit/blob/master/packages/angular_devkit/build_optimizer/src/purify/purify.ts#L52-L60, and it cannot work with any non-webpack tools.
Since it's too obvious I don't think there're any needs for a repo of reproduction.
Example input:
```javascript
/**
* @fileoverview This file is generated by the Angular template compiler.
* Do not edit.
* @suppress {suspiciousCode,uselessCode,missingProperties,missingOverride}
*/
/* tslint:disable */
import * as i0 from '@angular/core';
import * as i1 from '@angular/material';
import * as i2 from '@angular/cdk';
import * as i3 from '@angular/platform-browser';
import * as i4 from '@angular/common';
import * as i5 from '@angular/forms';
import * as i6 from '@angular/http';
export var MdCoreModuleNgFactory = i0.ɵcmf(i1.MdCoreModule, [], function (_l) {
return i0.ɵmod([i0.ɵmpd(512, i0.ComponentFactoryResolver, i0.ɵCodegenComponentFactoryResolver, [[8, []], [3, i0.ComponentFactoryResolver], i0.NgModuleRef]), i0.ɵmpd(6144, i2.DIR_DOCUMENT, null, [i3.DOCUMENT]), i0.ɵmpd(4608, i2.Directionality, i2.Directionality, [[2, i2.DIR_DOCUMENT]]), i0.ɵmpd(4608, i2.Platform, i2.Platform, []), i0.ɵmpd(5120, i1.ScrollDispatcher, i1.ɵe, [[3, i1.ScrollDispatcher],
i0.NgZone, i2.Platform]), i0.ɵmpd(5120, i1.ViewportRuler, i1.ɵc, [[3, i1.ViewportRuler],
i1.ScrollDispatcher]), i0.ɵmpd(4608, i1.ScrollStrategyOptions, i1.ScrollStrategyOptions, [i1.ScrollDispatcher, i1.ViewportRuler]), i0.ɵmpd(4608, i2.MdMutationObserverFactory, i2.MdMutationObserverFactory, []), i0.ɵmpd(5120, i1.OverlayContainer, i1.ɵa, [[3, i1.OverlayContainer]]), i0.ɵmpd(4608, i1.ɵt, i1.ɵt, [i1.ViewportRuler]),
i0.ɵmpd(4608, i1.Overlay, i1.Overlay, [i1.ScrollStrategyOptions, i1.OverlayContainer,
i0.ComponentFactoryResolver, i1.ɵt, i0.ApplicationRef, i0.Injector, i0.NgZone]),
i0.ɵmpd(4608, i4.NgLocalization, i4.NgLocaleLocalization, [i0.LOCALE_ID]), i0.ɵmpd(4608, i2.InteractivityChecker, i2.InteractivityChecker, [i2.Platform]), i0.ɵmpd(4608, i2.FocusTrapFactory, i2.FocusTrapFactory, [i2.InteractivityChecker, i2.Platform,
i0.NgZone]), i0.ɵmpd(5120, i2.LiveAnnouncer, i2.LIVE_ANNOUNCER_PROVIDER_FACTORY, [[3, i2.LiveAnnouncer], [2, i2.LIVE_ANNOUNCER_ELEMENT_TOKEN], i2.Platform]),
i0.ɵmpd(512, i1.CompatibilityModule, i1.CompatibilityModule, []),
i0.ɵmpd(512, i2.BidiModule, i2.BidiModule, []), i0.ɵmpd(256, i1.MATERIAL_SANITY_CHECKS, true, []), i0.ɵmpd(512, i1.MdCommonModule, i1.MdCommonModule, [[2, i3.DOCUMENT], [2, i1.MATERIAL_SANITY_CHECKS]]), i0.ɵmpd(512, i1.MdLineModule, i1.MdLineModule, []), i0.ɵmpd(512, i2.PlatformModule, i2.PlatformModule, []), i0.ɵmpd(512, i1.ScrollDispatchModule, i1.ScrollDispatchModule, []), i0.ɵmpd(512, i1.MdRippleModule, i1.MdRippleModule, []),
i0.ɵmpd(512, i2.ObserveContentModule, i2.ObserveContentModule, []),
i0.ɵmpd(512, i2.PortalModule, i2.PortalModule, []), i0.ɵmpd(512, i1.OverlayModule, i1.OverlayModule, []), i0.ɵmpd(512, i4.CommonModule, i4.CommonModule, []), i0.ɵmpd(512, i2.A11yModule, i2.A11yModule, []),
i0.ɵmpd(512, i1.MdSelectionModule, i1.MdSelectionModule, []), i0.ɵmpd(512, i1.MdOptionModule, i1.MdOptionModule, []), i0.ɵmpd(512, i1.MdCoreModule, i1.MdCoreModule, [])]);
});
``` | 1.0 | Purify doesn't work with NgModuleFactory in non-webpack environment - Relates to https://github.com/angular/material2/issues/4137 .
Currently purify plugin only supports hard-coded webpack regex in https://github.com/angular/devkit/blob/master/packages/angular_devkit/build_optimizer/src/purify/purify.ts#L52-L60, and it cannot work with any non-webpack tools.
Since it's too obvious I don't think there're any needs for a repo of reproduction.
Example input:
```javascript
/**
* @fileoverview This file is generated by the Angular template compiler.
* Do not edit.
* @suppress {suspiciousCode,uselessCode,missingProperties,missingOverride}
*/
/* tslint:disable */
import * as i0 from '@angular/core';
import * as i1 from '@angular/material';
import * as i2 from '@angular/cdk';
import * as i3 from '@angular/platform-browser';
import * as i4 from '@angular/common';
import * as i5 from '@angular/forms';
import * as i6 from '@angular/http';
export var MdCoreModuleNgFactory = i0.ɵcmf(i1.MdCoreModule, [], function (_l) {
return i0.ɵmod([i0.ɵmpd(512, i0.ComponentFactoryResolver, i0.ɵCodegenComponentFactoryResolver, [[8, []], [3, i0.ComponentFactoryResolver], i0.NgModuleRef]), i0.ɵmpd(6144, i2.DIR_DOCUMENT, null, [i3.DOCUMENT]), i0.ɵmpd(4608, i2.Directionality, i2.Directionality, [[2, i2.DIR_DOCUMENT]]), i0.ɵmpd(4608, i2.Platform, i2.Platform, []), i0.ɵmpd(5120, i1.ScrollDispatcher, i1.ɵe, [[3, i1.ScrollDispatcher],
i0.NgZone, i2.Platform]), i0.ɵmpd(5120, i1.ViewportRuler, i1.ɵc, [[3, i1.ViewportRuler],
i1.ScrollDispatcher]), i0.ɵmpd(4608, i1.ScrollStrategyOptions, i1.ScrollStrategyOptions, [i1.ScrollDispatcher, i1.ViewportRuler]), i0.ɵmpd(4608, i2.MdMutationObserverFactory, i2.MdMutationObserverFactory, []), i0.ɵmpd(5120, i1.OverlayContainer, i1.ɵa, [[3, i1.OverlayContainer]]), i0.ɵmpd(4608, i1.ɵt, i1.ɵt, [i1.ViewportRuler]),
i0.ɵmpd(4608, i1.Overlay, i1.Overlay, [i1.ScrollStrategyOptions, i1.OverlayContainer,
i0.ComponentFactoryResolver, i1.ɵt, i0.ApplicationRef, i0.Injector, i0.NgZone]),
i0.ɵmpd(4608, i4.NgLocalization, i4.NgLocaleLocalization, [i0.LOCALE_ID]), i0.ɵmpd(4608, i2.InteractivityChecker, i2.InteractivityChecker, [i2.Platform]), i0.ɵmpd(4608, i2.FocusTrapFactory, i2.FocusTrapFactory, [i2.InteractivityChecker, i2.Platform,
i0.NgZone]), i0.ɵmpd(5120, i2.LiveAnnouncer, i2.LIVE_ANNOUNCER_PROVIDER_FACTORY, [[3, i2.LiveAnnouncer], [2, i2.LIVE_ANNOUNCER_ELEMENT_TOKEN], i2.Platform]),
i0.ɵmpd(512, i1.CompatibilityModule, i1.CompatibilityModule, []),
i0.ɵmpd(512, i2.BidiModule, i2.BidiModule, []), i0.ɵmpd(256, i1.MATERIAL_SANITY_CHECKS, true, []), i0.ɵmpd(512, i1.MdCommonModule, i1.MdCommonModule, [[2, i3.DOCUMENT], [2, i1.MATERIAL_SANITY_CHECKS]]), i0.ɵmpd(512, i1.MdLineModule, i1.MdLineModule, []), i0.ɵmpd(512, i2.PlatformModule, i2.PlatformModule, []), i0.ɵmpd(512, i1.ScrollDispatchModule, i1.ScrollDispatchModule, []), i0.ɵmpd(512, i1.MdRippleModule, i1.MdRippleModule, []),
i0.ɵmpd(512, i2.ObserveContentModule, i2.ObserveContentModule, []),
i0.ɵmpd(512, i2.PortalModule, i2.PortalModule, []), i0.ɵmpd(512, i1.OverlayModule, i1.OverlayModule, []), i0.ɵmpd(512, i4.CommonModule, i4.CommonModule, []), i0.ɵmpd(512, i2.A11yModule, i2.A11yModule, []),
i0.ɵmpd(512, i1.MdSelectionModule, i1.MdSelectionModule, []), i0.ɵmpd(512, i1.MdOptionModule, i1.MdOptionModule, []), i0.ɵmpd(512, i1.MdCoreModule, i1.MdCoreModule, [])]);
});
``` | non_infrastructure | purify doesn t work with ngmodulefactory in non webpack environment relates to currently purify plugin only supports hard coded webpack regex in and it cannot work with any non webpack tools since it s too obvious i don t think there re any needs for a repo of reproduction example input javascript fileoverview this file is generated by the angular template compiler do not edit suppress suspiciouscode uselesscode missingproperties missingoverride tslint disable import as from angular core import as from angular material import as from angular cdk import as from angular platform browser import as from angular common import as from angular forms import as from angular http export var mdcoremodulengfactory ɵcmf mdcoremodule function l return ɵmod ngmoduleref ɵmpd dir document null ɵmpd directionality directionality ɵmpd platform platform ɵmpd scrolldispatcher ɵe ngzone platform ɵmpd viewportruler ɵc scrolldispatcher ɵmpd scrollstrategyoptions scrollstrategyoptions ɵmpd mdmutationobserverfactory mdmutationobserverfactory ɵmpd overlaycontainer ɵa ɵmpd ɵt ɵt ɵmpd overlay overlay scrollstrategyoptions overlaycontainer componentfactoryresolver ɵt applicationref injector ngzone ɵmpd nglocalization nglocalelocalization ɵmpd interactivitychecker interactivitychecker ɵmpd focustrapfactory focustrapfactory interactivitychecker platform ngzone ɵmpd liveannouncer live announcer provider factory platform ɵmpd compatibilitymodule compatibilitymodule ɵmpd bidimodule bidimodule ɵmpd material sanity checks true ɵmpd mdcommonmodule mdcommonmodule ɵmpd mdlinemodule mdlinemodule ɵmpd platformmodule platformmodule ɵmpd scrolldispatchmodule scrolldispatchmodule ɵmpd mdripplemodule mdripplemodule ɵmpd observecontentmodule observecontentmodule ɵmpd portalmodule portalmodule ɵmpd overlaymodule overlaymodule ɵmpd commonmodule commonmodule ɵmpd ɵmpd mdselectionmodule mdselectionmodule ɵmpd mdoptionmodule mdoptionmodule ɵmpd mdcoremodule mdcoremodule | 0 |
23,838 | 16,616,680,803 | IssuesEvent | 2021-06-02 17:37:10 | google/iree | https://api.github.com/repos/google/iree | closed | Add an IWYU (include-what-you-use) presubmit check | enhancement ➕ infrastructure 🛠️ quality of life 😊 | Just found that some defines doesn't match file path, and I fixed them in #3574. It would be good if we can have a check for this. File an issue for tracking this, feel free to close the issue if it's not reasonable. | 1.0 | Add an IWYU (include-what-you-use) presubmit check - Just found that some defines doesn't match file path, and I fixed them in #3574. It would be good if we can have a check for this. File an issue for tracking this, feel free to close the issue if it's not reasonable. | infrastructure | add an iwyu include what you use presubmit check just found that some defines doesn t match file path and i fixed them in it would be good if we can have a check for this file an issue for tracking this feel free to close the issue if it s not reasonable | 1 |
29,749 | 24,236,104,345 | IssuesEvent | 2022-09-26 23:31:36 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | ASP.NET Core v6.0.8 source build sucessed and installed ,but enter f11 can't view source code | area-infrastructure | 1, installed " aspnetcore-runtime-6.0.8-dev-win-x64.msi " package
**2, already add nuget.config to the project**
`<?xml version="1.0" encoding="utf-8"?>
<configuration>
<packageSources>
`<clear` />
<add key="MyBuildOfAspNetCore" value="E:\core3\AspNetCore\artifacts\packages\Debug\Shipping\" />
<add key="NuGet.org" value="https://api.nuget.org/v3/index.json" />
</packageSources>
`</configuration>``
**3 use vs 2022 create a mvc project ,enter f11 view souce code, but failed**
[https://images.cnblogs.com/cnblogs_com/blogs/729358/galleries/2211331/o_220901004234_2.png](url)
remark(operate system is win10)
who can help me ,thanks very much !
| 1.0 | ASP.NET Core v6.0.8 source build sucessed and installed ,but enter f11 can't view source code - 1, installed " aspnetcore-runtime-6.0.8-dev-win-x64.msi " package
**2, already add nuget.config to the project**
`<?xml version="1.0" encoding="utf-8"?>
<configuration>
<packageSources>
`<clear` />
<add key="MyBuildOfAspNetCore" value="E:\core3\AspNetCore\artifacts\packages\Debug\Shipping\" />
<add key="NuGet.org" value="https://api.nuget.org/v3/index.json" />
</packageSources>
`</configuration>``
**3 use vs 2022 create a mvc project ,enter f11 view souce code, but failed**
[https://images.cnblogs.com/cnblogs_com/blogs/729358/galleries/2211331/o_220901004234_2.png](url)
remark(operate system is win10)
who can help me ,thanks very much !
| infrastructure | asp net core source build sucessed and installed but enter can t view source code installed aspnetcore runtime dev win msi package already add nuget config to the project use vs create a mvc project enter view souce code but failed url remark operate system is who can help me thanks very much | 1 |
16,478 | 11,986,205,054 | IssuesEvent | 2020-04-07 18:52:56 | enarx/enarx | https://api.github.com/repos/enarx/enarx | closed | Enable SGX revision 28 on the SGX kernels | infrastructure intel sgx | Right now they're using the latest tag of SGX, revision 26. Enable revision 28 so we can provide testing feedback to the upstreaming effort; upload these kernel source packages into the Enarx copr. | 1.0 | Enable SGX revision 28 on the SGX kernels - Right now they're using the latest tag of SGX, revision 26. Enable revision 28 so we can provide testing feedback to the upstreaming effort; upload these kernel source packages into the Enarx copr. | infrastructure | enable sgx revision on the sgx kernels right now they re using the latest tag of sgx revision enable revision so we can provide testing feedback to the upstreaming effort upload these kernel source packages into the enarx copr | 1 |
4,318 | 4,973,764,270 | IssuesEvent | 2016-12-06 02:36:37 | archco/cosmos-css | https://api.github.com/repos/archco/cosmos-css | opened | Update 편의를 위해 src 폴더구조를 변경 | enhancement infrastructure | `src/`안에 폴더를
- base source : `sass/` `js/`
- custom : `custom/` 으로 나눈다.
| 1.0 | Update 편의를 위해 src 폴더구조를 변경 - `src/`안에 폴더를
- base source : `sass/` `js/`
- custom : `custom/` 으로 나눈다.
| infrastructure | update 편의를 위해 src 폴더구조를 변경 src 안에 폴더를 base source sass js custom custom 으로 나눈다 | 1 |
128 | 2,533,509,844 | IssuesEvent | 2015-01-24 00:11:23 | RobDixonIII/Bloom | https://api.github.com/repos/RobDixonIII/Bloom | closed | Implement Dynamic Tabs with Prism Modules | feature infrastructure | Prism uses an regioned injection approach for modules that loads the UI upon application start. This can make dynamic tabs problematic since they are not present at that time.
Create a way to dynamically inject Prism modules into tabs at runtime. | 1.0 | Implement Dynamic Tabs with Prism Modules - Prism uses an regioned injection approach for modules that loads the UI upon application start. This can make dynamic tabs problematic since they are not present at that time.
Create a way to dynamically inject Prism modules into tabs at runtime. | infrastructure | implement dynamic tabs with prism modules prism uses an regioned injection approach for modules that loads the ui upon application start this can make dynamic tabs problematic since they are not present at that time create a way to dynamically inject prism modules into tabs at runtime | 1 |
26,250 | 19,826,329,157 | IssuesEvent | 2022-01-20 07:04:58 | google/site-kit-wp | https://api.github.com/repos/google/site-kit-wp | opened | Fix VRT tests | P0 Type: Infrastructure | ## Bug Description
Our VRT tests recently started failing for what seems to be external factors. The last update to our reference images was 9 days ago in https://github.com/google/site-kit-wp/pull/4587 which only updated 6 images and we currently have 42 failures, so this wouldn't have been responsible.
The most likely cause it seems is that the Google Sans web font has changed as the diffs are highlighting changes in primary headings which are the main places this font is used. Roboto use does not raise any differences so it can't be a general web font issue.
## Steps to reproduce
* Run VRT tests on develop
* Run VRT tests on main
## Screenshots


## Additional Context
- Seems to have started on 2022-01-19
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* Visual regression references should be updated
## Implementation Brief
* Run VRT tests, review and approve the changes
### Test Coverage
* N/A
## QA Brief
* <!-- One or more bullet points for how to test that the issue has been resolved. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| 1.0 | Fix VRT tests - ## Bug Description
Our VRT tests recently started failing for what seems to be external factors. The last update to our reference images was 9 days ago in https://github.com/google/site-kit-wp/pull/4587 which only updated 6 images and we currently have 42 failures, so this wouldn't have been responsible.
The most likely cause it seems is that the Google Sans web font has changed as the diffs are highlighting changes in primary headings which are the main places this font is used. Roboto use does not raise any differences so it can't be a general web font issue.
## Steps to reproduce
* Run VRT tests on develop
* Run VRT tests on main
## Screenshots


## Additional Context
- Seems to have started on 2022-01-19
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* Visual regression references should be updated
## Implementation Brief
* Run VRT tests, review and approve the changes
### Test Coverage
* N/A
## QA Brief
* <!-- One or more bullet points for how to test that the issue has been resolved. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
| infrastructure | fix vrt tests bug description our vrt tests recently started failing for what seems to be external factors the last update to our reference images was days ago in which only updated images and we currently have failures so this wouldn t have been responsible the most likely cause it seems is that the google sans web font has changed as the diffs are highlighting changes in primary headings which are the main places this font is used roboto use does not raise any differences so it can t be a general web font issue steps to reproduce run vrt tests on develop run vrt tests on main screenshots additional context seems to have started on do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria visual regression references should be updated implementation brief run vrt tests review and approve the changes test coverage n a qa brief changelog entry | 1 |
121,072 | 10,149,315,419 | IssuesEvent | 2019-08-05 14:58:29 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | opened | Extended system test failure: SharedClassesAPI_0 | test failure | Failure link
------------
https://ci.adoptopenjdk.net/job/Test_openjdk11_j9_extended.system_x86-64_mac/7
Failing thread reference can be found in the results tar.gz, here:
openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\execute\execute.stdout
and the error message itself can be found in the output for that thread, here:
C:\Users\ADAMFarley\Downloads\systemtest_test_output(1).tar\systemtest_test_output(1)\openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\results\8.WL3.stderr
Optional info
-------------
- intermittent failure: no
- regression or new test: regression
- if regression, what are the last passing / first failing public SHAs (OpenJ9, OMR, JCL) :
Unknown.
Failure output
--------------
JVMSHRC659E An error has occurred while opening shared memory
JVMSHRC336E Port layer error code = -174
JVMSHRC337E Platform error message: Invalid argument
JVMSHRC026E Cannot create cache of requested size: Please check your SHMMAX and SHMMIN settings
JVMSHRC663I Error recovery: destroyed semaphore set with id=54067200 associated with shared class cache.
JVMJ9VM015W Initialization error for library j9shr29(11): JVMJ9VM009E J9VMDllMain failed
Error: Could not create the Java Virtual Machine.
Error: A fatal exception has occurred. Program will exit.
| 1.0 | Extended system test failure: SharedClassesAPI_0 - Failure link
------------
https://ci.adoptopenjdk.net/job/Test_openjdk11_j9_extended.system_x86-64_mac/7
Failing thread reference can be found in the results tar.gz, here:
openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\execute\execute.stdout
and the error message itself can be found in the output for that thread, here:
C:\Users\ADAMFarley\Downloads\systemtest_test_output(1).tar\systemtest_test_output(1)\openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\results\8.WL3.stderr
Optional info
-------------
- intermittent failure: no
- regression or new test: regression
- if regression, what are the last passing / first failing public SHAs (OpenJ9, OMR, JCL) :
Unknown.
Failure output
--------------
JVMSHRC659E An error has occurred while opening shared memory
JVMSHRC336E Port layer error code = -174
JVMSHRC337E Platform error message: Invalid argument
JVMSHRC026E Cannot create cache of requested size: Please check your SHMMAX and SHMMIN settings
JVMSHRC663I Error recovery: destroyed semaphore set with id=54067200 associated with shared class cache.
JVMJ9VM015W Initialization error for library j9shr29(11): JVMJ9VM009E J9VMDllMain failed
Error: Could not create the Java Virtual Machine.
Error: A fatal exception has occurred. Program will exit.
| non_infrastructure | extended system test failure sharedclassesapi failure link failing thread reference can be found in the results tar gz here openjdk tests testconfig test output sharedclassesapi sharedclassesapi execute execute stdout and the error message itself can be found in the output for that thread here c users adamfarley downloads systemtest test output tar systemtest test output openjdk tests testconfig test output sharedclassesapi sharedclassesapi results stderr optional info intermittent failure no regression or new test regression if regression what are the last passing first failing public shas omr jcl unknown failure output an error has occurred while opening shared memory port layer error code platform error message invalid argument cannot create cache of requested size please check your shmmax and shmmin settings error recovery destroyed semaphore set with id associated with shared class cache initialization error for library failed error could not create the java virtual machine error a fatal exception has occurred program will exit | 0 |
17,349 | 12,305,447,557 | IssuesEvent | 2020-05-11 22:33:22 | 18F/tts-tech-portfolio | https://api.github.com/repos/18F/tts-tech-portfolio | closed | Develop Process to Investigate new tools and technologies | epic: acquisition procurement and budgeting epic: policy and guidance epic: software and infrastructure groomed: accepted | ## Background information
There are new technologies, tools, etc that come to my attention either through emails, conferences, other people's suggestions, etc -- and I think they might have value in our inventory now or later -- but I am unsure if looking into them would be valuable or if I do consider them valuable -- what to do with that evaluation.
As of right now, the only guidance there is for when someone wants to use a new tool or technology is to submit a request to IT Standards. This isn't the best route because it bypasses the Tech Portfolio and we lose sight on what is being requested, if something is similar to another thing or if there is budget for the request.
This differs from https://github.com/18F/tts-tech-portfolio/issues/339 because that is more about analyzing a requirement once one exist; where this is proposing an idea for something to become a requirement in the first place.
## User stories
- As someone in the Tech Portfolio who has a good idea, I would like to be able to explore that good idea in a structured way.
- As someone in the Tech Portfolio who thinks a tool would be useful TTS wide, I don't have a way of knowing if that tool has already been proposed and what the result of that proposal was.
- As someone in the Tech Portfolio who plans on looking into a new tool, I would like a way to communicate the activity with other team members.
- As someone in the Tech Portfolio who thinks of a way of improving the organization through the use of a tool, I would like to know where that falls within the priorities of the portfolio.
## Implementation
- [ ] Develop framework (thinking concept, discovery, proposal (https://github.com/18F/tts-tech-portfolio/issues/339), decision, prototype, deploy)
- [ ] determine how the Emerging Technology office in OGP fits in
- [ ] consider vendor pitch days
## Acceptance criteria:
- There is a structured way for members of the Tech Portfolio to present ideas and for them to get prioritized within our workload and a record is kept for historical reference
| 1.0 | Develop Process to Investigate new tools and technologies - ## Background information
There are new technologies, tools, etc that come to my attention either through emails, conferences, other people's suggestions, etc -- and I think they might have value in our inventory now or later -- but I am unsure if looking into them would be valuable or if I do consider them valuable -- what to do with that evaluation.
As of right now, the only guidance there is for when someone wants to use a new tool or technology is to submit a request to IT Standards. This isn't the best route because it bypasses the Tech Portfolio and we lose sight on what is being requested, if something is similar to another thing or if there is budget for the request.
This differs from https://github.com/18F/tts-tech-portfolio/issues/339 because that is more about analyzing a requirement once one exist; where this is proposing an idea for something to become a requirement in the first place.
## User stories
- As someone in the Tech Portfolio who has a good idea, I would like to be able to explore that good idea in a structured way.
- As someone in the Tech Portfolio who thinks a tool would be useful TTS wide, I don't have a way of knowing if that tool has already been proposed and what the result of that proposal was.
- As someone in the Tech Portfolio who plans on looking into a new tool, I would like a way to communicate the activity with other team members.
- As someone in the Tech Portfolio who thinks of a way of improving the organization through the use of a tool, I would like to know where that falls within the priorities of the portfolio.
## Implementation
- [ ] Develop framework (thinking concept, discovery, proposal (https://github.com/18F/tts-tech-portfolio/issues/339), decision, prototype, deploy)
- [ ] determine how the Emerging Technology office in OGP fits in
- [ ] consider vendor pitch days
## Acceptance criteria:
- There is a structured way for members of the Tech Portfolio to present ideas and for them to get prioritized within our workload and a record is kept for historical reference
| infrastructure | develop process to investigate new tools and technologies background information there are new technologies tools etc that come to my attention either through emails conferences other people s suggestions etc and i think they might have value in our inventory now or later but i am unsure if looking into them would be valuable or if i do consider them valuable what to do with that evaluation as of right now the only guidance there is for when someone wants to use a new tool or technology is to submit a request to it standards this isn t the best route because it bypasses the tech portfolio and we lose sight on what is being requested if something is similar to another thing or if there is budget for the request this differs from because that is more about analyzing a requirement once one exist where this is proposing an idea for something to become a requirement in the first place user stories as someone in the tech portfolio who has a good idea i would like to be able to explore that good idea in a structured way as someone in the tech portfolio who thinks a tool would be useful tts wide i don t have a way of knowing if that tool has already been proposed and what the result of that proposal was as someone in the tech portfolio who plans on looking into a new tool i would like a way to communicate the activity with other team members as someone in the tech portfolio who thinks of a way of improving the organization through the use of a tool i would like to know where that falls within the priorities of the portfolio implementation develop framework thinking concept discovery proposal decision prototype deploy determine how the emerging technology office in ogp fits in consider vendor pitch days acceptance criteria there is a structured way for members of the tech portfolio to present ideas and for them to get prioritized within our workload and a record is kept for historical reference | 1 |
5,379 | 5,627,511,639 | IssuesEvent | 2017-04-05 01:50:24 | amnh-library/API-Portal | https://api.github.com/repos/amnh-library/API-Portal | opened | Change user directories in dev | infrastructure | Move all apps & code from mcaprio user in dev to apiproject user | 1.0 | Change user directories in dev - Move all apps & code from mcaprio user in dev to apiproject user | infrastructure | change user directories in dev move all apps code from mcaprio user in dev to apiproject user | 1 |
149,800 | 5,726,210,990 | IssuesEvent | 2017-04-20 18:24:41 | screwdriver-cd/screwdriver | https://api.github.com/repos/screwdriver-cd/screwdriver | opened | Publish template should show the version | change/patch component/UX priority/medium | The [template-main](https://github.com/screwdriver-cd/template-main) works fine to publish template, but it does not say what version it just publishes.
https://cd.screwdriver.cd/pipelines/152/builds/3558
We should at least show the VERSION in the log. | 1.0 | Publish template should show the version - The [template-main](https://github.com/screwdriver-cd/template-main) works fine to publish template, but it does not say what version it just publishes.
https://cd.screwdriver.cd/pipelines/152/builds/3558
We should at least show the VERSION in the log. | non_infrastructure | publish template should show the version the works fine to publish template but it does not say what version it just publishes we should at least show the version in the log | 0 |
28,097 | 22,959,455,363 | IssuesEvent | 2022-07-19 14:17:24 | IBM-Cloud/terraform-provider-ibm | https://api.github.com/repos/IBM-Cloud/terraform-provider-ibm | closed | problems with IPs reservations | service/VPC Infrastructure service/Resource Management | <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform IBM Provider Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
$ terraform version
Terraform v1.2.2
on linux_amd64
+ provider registry.terraform.io/ibm-cloud/ibm v1.42.0
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* ibm_is_subnet_reserved_ip
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation.
```hcl
# Copy-paste your Terraform configurations here - for large Terraform configs,
# please share a link to the ZIP file.
provider "ibm" {
region = "eu-de"
}
data "ibm_is_image" "example" {
name = "ibm-debian-11-3-minimal-amd64-1"
}
resource "ibm_is_vpc" "example" {
name = "example-vpc"
}
resource "ibm_is_vpc_address_prefix" "example" {
name = "example-address-prefix"
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
cidr = "10.240.0.0/24"
}
resource "ibm_is_subnet" "example" {
name = "example-subnet"
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
ipv4_cidr_block = "10.240.0.0/24"
depends_on = [ibm_is_vpc_address_prefix.example]
}
resource "ibm_is_subnet_reserved_ip" "example" {
name = "example-reserved-ip"
subnet = ibm_is_subnet.example.id
}
resource "ibm_is_ssh_key" "example" {
name = "example-ssh"
public_key = "ssh-rsa ...."
}
resource "ibm_is_instance" "example1" {
image = data.ibm_is_image.example.id
name = "example-instance-1-reserved-ip"
profile = "bx2-2x8"
primary_network_interface {
name = "eth0"
subnet = ibm_is_subnet.example.id
primary_ip {
reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip
}
}
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
keys = [ibm_is_ssh_key.example.id]
}
resource "ibm_is_instance" "example2" {
image = data.ibm_is_image.example.id
name = "example-instance-2-reserved-ip"
profile = "bx2-2x8"
primary_network_interface {
name = "eth0"
subnet = ibm_is_subnet.example.id
primary_ip {
reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip
}
}
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
keys = [ibm_is_ssh_key.example.id]
}
```
### Debug Output
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Panic Output
<!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. --->
### Expected Behavior
<!--- What should have happened? --->
It should have failed since the IP should be unique for each VSI.
### Actual Behavior
<!--- What actually happened? --->

### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform apply`
$ terraform apply
data.ibm_is_image.example: Reading...
ibm_is_ssh_key.example: Refreshing state... [id=r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6]
ibm_is_vpc.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78]
data.ibm_is_image.example: Read complete after 1s [id=r010-b6c658a8-901c-4a6c-802f-f3488a114e28]
ibm_is_vpc_address_prefix.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78/r010-c9f48a74-16d7-447f-af80-718e7b8778dd]
ibm_is_subnet.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622]
ibm_is_subnet_reserved_ip.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622/02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c]
Terraform used the selected providers to generate the following execution plan. Resource actions are indicated with the following symbols:
+ create
Terraform will perform the following actions:
# ibm_is_instance.example1 will be created
+ resource "ibm_is_instance" "example1" {
+ availability_policy_host_failure = (known after apply)
+ bandwidth = (known after apply)
+ crn = (known after apply)
+ default_trusted_profile_auto_link = (known after apply)
+ disks = (known after apply)
+ force_action = false
+ gpu = (known after apply)
+ id = (known after apply)
+ image = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"
+ keys = [
+ "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",
]
+ memory = (known after apply)
+ metadata_service_enabled = (known after apply)
+ name = "example-instance-1-reserved-ip"
+ placement_target = (known after apply)
+ profile = "bx2-2x8"
+ resource_controller_url = (known after apply)
+ resource_crn = (known after apply)
+ resource_group = (known after apply)
+ resource_group_name = (known after apply)
+ resource_name = (known after apply)
+ resource_status = (known after apply)
+ status = (known after apply)
+ status_reasons = (known after apply)
+ tags = (known after apply)
+ total_network_bandwidth = (known after apply)
+ total_volume_bandwidth = (known after apply)
+ vcpu = (known after apply)
+ volume_attachments = (known after apply)
+ vpc = "r010-da0ba964-30b9-4505-b783-ed248b065f78"
+ wait_before_delete = true
+ zone = "eu-de-1"
+ boot_volume {
+ encryption = (known after apply)
+ iops = (known after apply)
+ name = (known after apply)
+ profile = (known after apply)
+ size = (known after apply)
+ snapshot = (known after apply)
+ volume_id = (known after apply)
}
+ network_interfaces {
+ allow_ip_spoofing = (known after apply)
+ id = (known after apply)
+ name = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = (known after apply)
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = (known after apply)
+ resource_type = (known after apply)
}
}
+ primary_network_interface {
+ allow_ip_spoofing = false
+ id = (known after apply)
+ name = "eth0"
+ port_speed = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = "02b7-461d844c-4701-463c-b9fe-92240a753622"
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"
+ resource_type = (known after apply)
}
}
}
# ibm_is_instance.example2 will be created
+ resource "ibm_is_instance" "example2" {
+ availability_policy_host_failure = (known after apply)
+ bandwidth = (known after apply)
+ crn = (known after apply)
+ default_trusted_profile_auto_link = (known after apply)
+ disks = (known after apply)
+ force_action = false
+ gpu = (known after apply)
+ id = (known after apply)
+ image = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"
+ keys = [
+ "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",
]
+ memory = (known after apply)
+ metadata_service_enabled = (known after apply)
+ name = "example-instance-2-reserved-ip"
+ placement_target = (known after apply)
+ profile = "bx2-2x8"
+ resource_controller_url = (known after apply)
+ resource_crn = (known after apply)
+ resource_group = (known after apply)
+ resource_group_name = (known after apply)
+ resource_name = (known after apply)
+ resource_status = (known after apply)
+ status = (known after apply)
+ status_reasons = (known after apply)
+ tags = (known after apply)
+ total_network_bandwidth = (known after apply)
+ total_volume_bandwidth = (known after apply)
+ vcpu = (known after apply)
+ volume_attachments = (known after apply)
+ vpc = "r010-da0ba964-30b9-4505-b783-ed248b065f78"
+ wait_before_delete = true
+ zone = "eu-de-1"
+ boot_volume {
+ encryption = (known after apply)
+ iops = (known after apply)
+ name = (known after apply)
+ profile = (known after apply)
+ size = (known after apply)
+ snapshot = (known after apply)
+ volume_id = (known after apply)
}
+ network_interfaces {
+ allow_ip_spoofing = (known after apply)
+ id = (known after apply)
+ name = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = (known after apply)
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = (known after apply)
+ resource_type = (known after apply)
}
}
+ primary_network_interface {
+ allow_ip_spoofing = false
+ id = (known after apply)
+ name = "eth0"
+ port_speed = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = "02b7-461d844c-4701-463c-b9fe-92240a753622"
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"
+ resource_type = (known after apply)
}
}
}
Plan: 2 to add, 0 to change, 0 to destroy.
Do you want to perform these actions?
Terraform will perform the actions described above.
Only 'yes' will be accepted to approve.
Enter a value: yes
ibm_is_instance.example1: Creating...
ibm_is_instance.example2: Creating...
ibm_is_instance.example1: Still creating... [10s elapsed]
ibm_is_instance.example2: Still creating... [10s elapsed]
ibm_is_instance.example1: Still creating... [20s elapsed]
ibm_is_instance.example2: Still creating... [20s elapsed]
ibm_is_instance.example2: Creation complete after 27s [id=02b7_cba53a91-2924-40fd-993e-af42fda2b98b]
ibm_is_instance.example1: Creation complete after 29s [id=02b7_7b407a4f-d951-4ac6-84ed-609bdb36f2e9]
Apply complete! Resources: 2 added, 0 changed, 0 destroyed.
moby@f2898e894286:~/cloud-practice/internal/extra/vm-ibm-exam
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
* #0000
| 1.0 | problems with IPs reservations - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform IBM Provider Version
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
$ terraform version
Terraform v1.2.2
on linux_amd64
+ provider registry.terraform.io/ibm-cloud/ibm v1.42.0
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* ibm_is_subnet_reserved_ip
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation.
```hcl
# Copy-paste your Terraform configurations here - for large Terraform configs,
# please share a link to the ZIP file.
provider "ibm" {
region = "eu-de"
}
data "ibm_is_image" "example" {
name = "ibm-debian-11-3-minimal-amd64-1"
}
resource "ibm_is_vpc" "example" {
name = "example-vpc"
}
resource "ibm_is_vpc_address_prefix" "example" {
name = "example-address-prefix"
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
cidr = "10.240.0.0/24"
}
resource "ibm_is_subnet" "example" {
name = "example-subnet"
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
ipv4_cidr_block = "10.240.0.0/24"
depends_on = [ibm_is_vpc_address_prefix.example]
}
resource "ibm_is_subnet_reserved_ip" "example" {
name = "example-reserved-ip"
subnet = ibm_is_subnet.example.id
}
resource "ibm_is_ssh_key" "example" {
name = "example-ssh"
public_key = "ssh-rsa ...."
}
resource "ibm_is_instance" "example1" {
image = data.ibm_is_image.example.id
name = "example-instance-1-reserved-ip"
profile = "bx2-2x8"
primary_network_interface {
name = "eth0"
subnet = ibm_is_subnet.example.id
primary_ip {
reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip
}
}
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
keys = [ibm_is_ssh_key.example.id]
}
resource "ibm_is_instance" "example2" {
image = data.ibm_is_image.example.id
name = "example-instance-2-reserved-ip"
profile = "bx2-2x8"
primary_network_interface {
name = "eth0"
subnet = ibm_is_subnet.example.id
primary_ip {
reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip
}
}
vpc = ibm_is_vpc.example.id
zone = "eu-de-1"
keys = [ibm_is_ssh_key.example.id]
}
```
### Debug Output
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Panic Output
<!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. --->
### Expected Behavior
<!--- What should have happened? --->
It should have failed since the IP should be unique for each VSI.
### Actual Behavior
<!--- What actually happened? --->

### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform apply`
$ terraform apply
data.ibm_is_image.example: Reading...
ibm_is_ssh_key.example: Refreshing state... [id=r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6]
ibm_is_vpc.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78]
data.ibm_is_image.example: Read complete after 1s [id=r010-b6c658a8-901c-4a6c-802f-f3488a114e28]
ibm_is_vpc_address_prefix.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78/r010-c9f48a74-16d7-447f-af80-718e7b8778dd]
ibm_is_subnet.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622]
ibm_is_subnet_reserved_ip.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622/02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c]
Terraform used the selected providers to generate the following execution plan. Resource actions are indicated with the following symbols:
+ create
Terraform will perform the following actions:
# ibm_is_instance.example1 will be created
+ resource "ibm_is_instance" "example1" {
+ availability_policy_host_failure = (known after apply)
+ bandwidth = (known after apply)
+ crn = (known after apply)
+ default_trusted_profile_auto_link = (known after apply)
+ disks = (known after apply)
+ force_action = false
+ gpu = (known after apply)
+ id = (known after apply)
+ image = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"
+ keys = [
+ "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",
]
+ memory = (known after apply)
+ metadata_service_enabled = (known after apply)
+ name = "example-instance-1-reserved-ip"
+ placement_target = (known after apply)
+ profile = "bx2-2x8"
+ resource_controller_url = (known after apply)
+ resource_crn = (known after apply)
+ resource_group = (known after apply)
+ resource_group_name = (known after apply)
+ resource_name = (known after apply)
+ resource_status = (known after apply)
+ status = (known after apply)
+ status_reasons = (known after apply)
+ tags = (known after apply)
+ total_network_bandwidth = (known after apply)
+ total_volume_bandwidth = (known after apply)
+ vcpu = (known after apply)
+ volume_attachments = (known after apply)
+ vpc = "r010-da0ba964-30b9-4505-b783-ed248b065f78"
+ wait_before_delete = true
+ zone = "eu-de-1"
+ boot_volume {
+ encryption = (known after apply)
+ iops = (known after apply)
+ name = (known after apply)
+ profile = (known after apply)
+ size = (known after apply)
+ snapshot = (known after apply)
+ volume_id = (known after apply)
}
+ network_interfaces {
+ allow_ip_spoofing = (known after apply)
+ id = (known after apply)
+ name = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = (known after apply)
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = (known after apply)
+ resource_type = (known after apply)
}
}
+ primary_network_interface {
+ allow_ip_spoofing = false
+ id = (known after apply)
+ name = "eth0"
+ port_speed = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = "02b7-461d844c-4701-463c-b9fe-92240a753622"
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"
+ resource_type = (known after apply)
}
}
}
# ibm_is_instance.example2 will be created
+ resource "ibm_is_instance" "example2" {
+ availability_policy_host_failure = (known after apply)
+ bandwidth = (known after apply)
+ crn = (known after apply)
+ default_trusted_profile_auto_link = (known after apply)
+ disks = (known after apply)
+ force_action = false
+ gpu = (known after apply)
+ id = (known after apply)
+ image = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"
+ keys = [
+ "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",
]
+ memory = (known after apply)
+ metadata_service_enabled = (known after apply)
+ name = "example-instance-2-reserved-ip"
+ placement_target = (known after apply)
+ profile = "bx2-2x8"
+ resource_controller_url = (known after apply)
+ resource_crn = (known after apply)
+ resource_group = (known after apply)
+ resource_group_name = (known after apply)
+ resource_name = (known after apply)
+ resource_status = (known after apply)
+ status = (known after apply)
+ status_reasons = (known after apply)
+ tags = (known after apply)
+ total_network_bandwidth = (known after apply)
+ total_volume_bandwidth = (known after apply)
+ vcpu = (known after apply)
+ volume_attachments = (known after apply)
+ vpc = "r010-da0ba964-30b9-4505-b783-ed248b065f78"
+ wait_before_delete = true
+ zone = "eu-de-1"
+ boot_volume {
+ encryption = (known after apply)
+ iops = (known after apply)
+ name = (known after apply)
+ profile = (known after apply)
+ size = (known after apply)
+ snapshot = (known after apply)
+ volume_id = (known after apply)
}
+ network_interfaces {
+ allow_ip_spoofing = (known after apply)
+ id = (known after apply)
+ name = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = (known after apply)
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = (known after apply)
+ resource_type = (known after apply)
}
}
+ primary_network_interface {
+ allow_ip_spoofing = false
+ id = (known after apply)
+ name = "eth0"
+ port_speed = (known after apply)
+ primary_ipv4_address = (known after apply)
+ security_groups = (known after apply)
+ subnet = "02b7-461d844c-4701-463c-b9fe-92240a753622"
+ primary_ip {
+ address = (known after apply)
+ auto_delete = (known after apply)
+ href = (known after apply)
+ name = (known after apply)
+ reserved_ip = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"
+ resource_type = (known after apply)
}
}
}
Plan: 2 to add, 0 to change, 0 to destroy.
Do you want to perform these actions?
Terraform will perform the actions described above.
Only 'yes' will be accepted to approve.
Enter a value: yes
ibm_is_instance.example1: Creating...
ibm_is_instance.example2: Creating...
ibm_is_instance.example1: Still creating... [10s elapsed]
ibm_is_instance.example2: Still creating... [10s elapsed]
ibm_is_instance.example1: Still creating... [20s elapsed]
ibm_is_instance.example2: Still creating... [20s elapsed]
ibm_is_instance.example2: Creation complete after 27s [id=02b7_cba53a91-2924-40fd-993e-af42fda2b98b]
ibm_is_instance.example1: Creation complete after 29s [id=02b7_7b407a4f-d951-4ac6-84ed-609bdb36f2e9]
Apply complete! Resources: 2 added, 0 changed, 0 destroyed.
moby@f2898e894286:~/cloud-practice/internal/extra/vm-ibm-exam
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
* #0000
| infrastructure | problems with ips reservations please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform ibm provider version terraform version terraform on linux provider registry terraform io ibm cloud ibm affected resource s ibm is subnet reserved ip terraform configuration files please include all terraform configurations required to reproduce the bug bug reports without a functional reproduction may be closed without investigation hcl copy paste your terraform configurations here for large terraform configs please share a link to the zip file provider ibm region eu de data ibm is image example name ibm debian minimal resource ibm is vpc example name example vpc resource ibm is vpc address prefix example name example address prefix vpc ibm is vpc example id zone eu de cidr resource ibm is subnet example name example subnet vpc ibm is vpc example id zone eu de cidr block depends on resource ibm is subnet reserved ip example name example reserved ip subnet ibm is subnet example id resource ibm is ssh key example name example ssh public key ssh rsa resource ibm is instance image data ibm is image example id name example instance reserved ip profile primary network interface name subnet ibm is subnet example id primary ip reserved ip ibm is subnet reserved ip example reserved ip vpc ibm is vpc example id zone eu de keys resource ibm is instance image data ibm is image example id name example instance reserved ip profile primary network interface name subnet ibm is subnet example id primary ip reserved ip ibm is subnet reserved ip example reserved ip vpc ibm is vpc example id zone eu de keys debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the panic output expected behavior it should have failed since the ip should be unique for each vsi actual behavior steps to reproduce terraform apply terraform apply data ibm is image example reading ibm is ssh key example refreshing state ibm is vpc example refreshing state data ibm is image example read complete after ibm is vpc address prefix example refreshing state ibm is subnet example refreshing state ibm is subnet reserved ip example refreshing state terraform used the selected providers to generate the following execution plan resource actions are indicated with the following symbols create terraform will perform the following actions ibm is instance will be created resource ibm is instance availability policy host failure known after apply bandwidth known after apply crn known after apply default trusted profile auto link known after apply disks known after apply force action false gpu known after apply id known after apply image keys memory known after apply metadata service enabled known after apply name example instance reserved ip placement target known after apply profile resource controller url known after apply resource crn known after apply resource group known after apply resource group name known after apply resource name known after apply resource status known after apply status known after apply status reasons known after apply tags known after apply total network bandwidth known after apply total volume bandwidth known after apply vcpu known after apply volume attachments known after apply vpc wait before delete true zone eu de boot volume encryption known after apply iops known after apply name known after apply profile known after apply size known after apply snapshot known after apply volume id known after apply network interfaces allow ip spoofing known after apply id known after apply name known after apply primary address known after apply security groups known after apply subnet known after apply primary ip address known after apply auto delete known after apply href known after apply name known after apply reserved ip known after apply resource type known after apply primary network interface allow ip spoofing false id known after apply name port speed known after apply primary address known after apply security groups known after apply subnet primary ip address known after apply auto delete known after apply href known after apply name known after apply reserved ip resource type known after apply ibm is instance will be created resource ibm is instance availability policy host failure known after apply bandwidth known after apply crn known after apply default trusted profile auto link known after apply disks known after apply force action false gpu known after apply id known after apply image keys memory known after apply metadata service enabled known after apply name example instance reserved ip placement target known after apply profile resource controller url known after apply resource crn known after apply resource group known after apply resource group name known after apply resource name known after apply resource status known after apply status known after apply status reasons known after apply tags known after apply total network bandwidth known after apply total volume bandwidth known after apply vcpu known after apply volume attachments known after apply vpc wait before delete true zone eu de boot volume encryption known after apply iops known after apply name known after apply profile known after apply size known after apply snapshot known after apply volume id known after apply network interfaces allow ip spoofing known after apply id known after apply name known after apply primary address known after apply security groups known after apply subnet known after apply primary ip address known after apply auto delete known after apply href known after apply name known after apply reserved ip known after apply resource type known after apply primary network interface allow ip spoofing false id known after apply name port speed known after apply primary address known after apply security groups known after apply subnet primary ip address known after apply auto delete known after apply href known after apply name known after apply reserved ip resource type known after apply plan to add to change to destroy do you want to perform these actions terraform will perform the actions described above only yes will be accepted to approve enter a value yes ibm is instance creating ibm is instance creating ibm is instance still creating ibm is instance still creating ibm is instance still creating ibm is instance still creating ibm is instance creation complete after ibm is instance creation complete after apply complete resources added changed destroyed moby cloud practice internal extra vm ibm exam important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example | 1 |
26,418 | 20,103,162,550 | IssuesEvent | 2022-02-07 07:44:42 | Greenstand/treetracker-web-map-client | https://api.github.com/repos/Greenstand/treetracker-web-map-client | opened | A solution to share/reuse components crossing repositories. | good first issue infrastructure | For example, the share components, they are the same between wallet app and web map app, so we need to find a good solution to share the components crossing repos.
---
Some hints:
- Please read our [readme](https://github.com/Greenstand/treetracker-web-map-client#treetracker-web) for more information/guide/tutorial.
- Here is [an engineering book](https://greenstand.gitbook.io/engineering/) in Greenstand.
- To know more about our organization, visit our [website](https://greenstand.org).
- If you want to join the slack community (some resources need the community member's permission), please leave your email address.
- More issues in this project: https://github.com/orgs/Greenstand/projects/26
| 1.0 | A solution to share/reuse components crossing repositories. - For example, the share components, they are the same between wallet app and web map app, so we need to find a good solution to share the components crossing repos.
---
Some hints:
- Please read our [readme](https://github.com/Greenstand/treetracker-web-map-client#treetracker-web) for more information/guide/tutorial.
- Here is [an engineering book](https://greenstand.gitbook.io/engineering/) in Greenstand.
- To know more about our organization, visit our [website](https://greenstand.org).
- If you want to join the slack community (some resources need the community member's permission), please leave your email address.
- More issues in this project: https://github.com/orgs/Greenstand/projects/26
| infrastructure | a solution to share reuse components crossing repositories for example the share components they are the same between wallet app and web map app so we need to find a good solution to share the components crossing repos some hints please read our for more information guide tutorial here is in greenstand to know more about our organization visit our if you want to join the slack community some resources need the community member s permission please leave your email address more issues in this project | 1 |
12,239 | 9,659,947,627 | IssuesEvent | 2019-05-20 14:30:28 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Input docker behaviour with Openshift 3.10 | :infrastructure Filebeat Metricbeat [zube]: In Progress containers enhancement libbeat v7.2.0 | Filebeat Version : 6.4.3-1
Openshift version : 3.10
Since Openshift version 3.10, docker has been replaced by cri-o. Log format changed from json to cri format and we need to use input docker to parse file as describe in this issue https://github.com/elastic/beats/issues/7418
Nethertheless, the path of docker logs in Openshift 3.10 is /var/log/containers/<pod_name>_<namespace>_<container_name>-<container_id>.log. Docker input consider this is this /var/lib/docker/containers/<container_id>/*.log.
I used this workaround for all docker
<pre>filebeat:
inputs:
- type: docker
containers:
path: "/var/log/containers"
ids:
- ''
symlinks: true
multiline.pattern: '^\['
multiline.match: after
multiline.negate: true
fields:
type: openshift-pods
fields_under_root: true
exclude_lines: ["DEBUG","TRACE"]
exclude_files: ["rabbitmq","nginx","redis"]</pre>
It will considered this path -> /var/log/containers//*.log and it's ok if I wanted to process all log on the same way. In the other side, I need to process rabbitmq, nginx and redis pods on another ways but I'm blocked;
For example
<pre>
filebeat:
inputs:
- type: docker
containers:
path: "/var/log/containers"
ids:
- ''
symlinks: true
multiline.pattern: '^[0-9]{4}-[0-9]{2}-[0-9]{2}'
multiline.match: after
multiline.negate: true
fields:
type: rabbit-pods
fields_under_root: true
## Negative pattern not supported
exclude_lines: [".*(?!rabbitmq).*"]
##
exclude_lines: ["DEBUG","TRACE"]
</pre>
This example failed. I can't use exclude_line with negative pattern. Include line is not supported too. I think it should be good to be able to modify the "/*.log" automatically added by filebeat with input docker.
PS : directory /var/log/pods should be the good directory for input but it contain only ID of dockers and I'll lost all <pod_name>_<namespace>_<container_name>-<container_id> I mentionned before. | 1.0 | Input docker behaviour with Openshift 3.10 - Filebeat Version : 6.4.3-1
Openshift version : 3.10
Since Openshift version 3.10, docker has been replaced by cri-o. Log format changed from json to cri format and we need to use input docker to parse file as describe in this issue https://github.com/elastic/beats/issues/7418
Nethertheless, the path of docker logs in Openshift 3.10 is /var/log/containers/<pod_name>_<namespace>_<container_name>-<container_id>.log. Docker input consider this is this /var/lib/docker/containers/<container_id>/*.log.
I used this workaround for all docker
<pre>filebeat:
inputs:
- type: docker
containers:
path: "/var/log/containers"
ids:
- ''
symlinks: true
multiline.pattern: '^\['
multiline.match: after
multiline.negate: true
fields:
type: openshift-pods
fields_under_root: true
exclude_lines: ["DEBUG","TRACE"]
exclude_files: ["rabbitmq","nginx","redis"]</pre>
It will considered this path -> /var/log/containers//*.log and it's ok if I wanted to process all log on the same way. In the other side, I need to process rabbitmq, nginx and redis pods on another ways but I'm blocked;
For example
<pre>
filebeat:
inputs:
- type: docker
containers:
path: "/var/log/containers"
ids:
- ''
symlinks: true
multiline.pattern: '^[0-9]{4}-[0-9]{2}-[0-9]{2}'
multiline.match: after
multiline.negate: true
fields:
type: rabbit-pods
fields_under_root: true
## Negative pattern not supported
exclude_lines: [".*(?!rabbitmq).*"]
##
exclude_lines: ["DEBUG","TRACE"]
</pre>
This example failed. I can't use exclude_line with negative pattern. Include line is not supported too. I think it should be good to be able to modify the "/*.log" automatically added by filebeat with input docker.
PS : directory /var/log/pods should be the good directory for input but it contain only ID of dockers and I'll lost all <pod_name>_<namespace>_<container_name>-<container_id> I mentionned before. | infrastructure | input docker behaviour with openshift filebeat version openshift version since openshift version docker has been replaced by cri o log format changed from json to cri format and we need to use input docker to parse file as describe in this issue nethertheless the path of docker logs in openshift is var log containers log docker input consider this is this var lib docker containers log i used this workaround for all docker filebeat inputs type docker containers path var log containers ids symlinks true multiline pattern multiline match after multiline negate true fields type openshift pods fields under root true exclude lines exclude files it will considered this path var log containers log and it s ok if i wanted to process all log on the same way in the other side i need to process rabbitmq nginx and redis pods on another ways but i m blocked for example filebeat inputs type docker containers path var log containers ids symlinks true multiline pattern multiline match after multiline negate true fields type rabbit pods fields under root true negative pattern not supported exclude lines exclude lines this example failed i can t use exclude line with negative pattern include line is not supported too i think it should be good to be able to modify the log automatically added by filebeat with input docker ps directory var log pods should be the good directory for input but it contain only id of dockers and i ll lost all i mentionned before | 1 |
35,588 | 31,849,543,709 | IssuesEvent | 2023-09-14 23:34:56 | casangi/astrohack | https://api.github.com/repos/casangi/astrohack | opened | Fix matplotlib close issue when multiple plots are created | Area:Infrastructure visualization | This is made simpler with a few steps:
- [ ] Factorize figure and axes creation with the function now located in _locit_commons
- [ ] Factorize figure closing with the function now located in _locit_commons
After this factorization several solutions can be tested as the code will need to be changed in a single place | 1.0 | Fix matplotlib close issue when multiple plots are created - This is made simpler with a few steps:
- [ ] Factorize figure and axes creation with the function now located in _locit_commons
- [ ] Factorize figure closing with the function now located in _locit_commons
After this factorization several solutions can be tested as the code will need to be changed in a single place | infrastructure | fix matplotlib close issue when multiple plots are created this is made simpler with a few steps factorize figure and axes creation with the function now located in locit commons factorize figure closing with the function now located in locit commons after this factorization several solutions can be tested as the code will need to be changed in a single place | 1 |
645,686 | 21,012,132,522 | IssuesEvent | 2022-03-30 07:43:45 | MaibornWolff/codecharta | https://api.github.com/repos/MaibornWolff/codecharta | closed | File path in side bar not entirely visible | bug priority:high | # Bug
## Expected Behavior
When I click on a building, the entire file path should be visible.
## Actual Behavior
Long file paths are cut and not entirely visible.
## Screenshots

## Steps to Reproduce the Problem
1. Open Codecharta
2. Click on a building with a long filename
## Specifications
- is released in [online-demo](https://maibornwolff.github.io/codecharta/visualization/app/index.html?file=codecharta.cc.json&file=codecharta_analysis.cc.json): yes
- CodeCharta Version: 1.95.0
- OS: Windows 10
- Browser: Chrome
| 1.0 | File path in side bar not entirely visible - # Bug
## Expected Behavior
When I click on a building, the entire file path should be visible.
## Actual Behavior
Long file paths are cut and not entirely visible.
## Screenshots

## Steps to Reproduce the Problem
1. Open Codecharta
2. Click on a building with a long filename
## Specifications
- is released in [online-demo](https://maibornwolff.github.io/codecharta/visualization/app/index.html?file=codecharta.cc.json&file=codecharta_analysis.cc.json): yes
- CodeCharta Version: 1.95.0
- OS: Windows 10
- Browser: Chrome
| non_infrastructure | file path in side bar not entirely visible bug expected behavior when i click on a building the entire file path should be visible actual behavior long file paths are cut and not entirely visible screenshots steps to reproduce the problem open codecharta click on a building with a long filename specifications is released in yes codecharta version os windows browser chrome | 0 |
18,598 | 13,055,987,064 | IssuesEvent | 2020-07-30 03:18:52 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | [documentation] 403 forbidden (Trac #1980) | Incomplete Migration Migrated from Trac infrastructure task | Migrated from https://code.icecube.wisc.edu/ticket/1980
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:44",
"description": "Whenever the nightly doc rebuild happens, we get 403 forbidden for quite a while (many minutes to hours). While the US users may not care as much, since it happens late at night, Europe/Asia is awake and working. So, let's make sure this doesn't happen.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067284370534",
"component": "infrastructure",
"summary": "[documentation] 403 forbidden",
"priority": "normal",
"keywords": "",
"time": "2017-04-09T13:25:11",
"milestone": "",
"owner": "nega",
"type": "task"
}
```
| 1.0 | [documentation] 403 forbidden (Trac #1980) - Migrated from https://code.icecube.wisc.edu/ticket/1980
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:44",
"description": "Whenever the nightly doc rebuild happens, we get 403 forbidden for quite a while (many minutes to hours). While the US users may not care as much, since it happens late at night, Europe/Asia is awake and working. So, let's make sure this doesn't happen.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067284370534",
"component": "infrastructure",
"summary": "[documentation] 403 forbidden",
"priority": "normal",
"keywords": "",
"time": "2017-04-09T13:25:11",
"milestone": "",
"owner": "nega",
"type": "task"
}
```
| infrastructure | forbidden trac migrated from json status closed changetime description whenever the nightly doc rebuild happens we get forbidden for quite a while many minutes to hours while the us users may not care as much since it happens late at night europe asia is awake and working so let s make sure this doesn t happen reporter david schultz cc resolution fixed ts component infrastructure summary forbidden priority normal keywords time milestone owner nega type task | 1 |
11,909 | 9,516,166,931 | IssuesEvent | 2019-04-26 08:09:10 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | [Metricbeat] Flaky test mssql | :infrastructure [zube]: In Progress flaky-test | ## Flaky Test
* **Test Name:** x-pack/metricbeat/tests/system/test_mssql.py
* **Link:** https://travis-ci.org/elastic/beats/jobs/522183265#L683
### Stack Trace
```
etricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 ...
======================================================================
ERROR: MSSQL module outputs an event.
----------------------------------------------------------------------
Traceback (most recent call last):
File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 75, in test_performance
self.assert_fields_are_documented(evt)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented
expected_fields, dict_fields, aliases = self.load_fields()
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields
subfields, subdictfields, subaliases = extract_fields(item["fields"], "")
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
KeyError: 'fields'
-------------------- >> begin captured stdout << ---------------------
render config
--------------------- >> end captured stdout << ----------------------
-------------------- >> begin captured logging << --------------------
compose.config.config: DEBUG: Using configuration files: ./docker-compose.yml
compose.config.environment: WARNING: The PWD variable is not set. Defaulting to a blank string.
docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg']
docker.utils.config: DEBUG: No config file found
docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg']
docker.utils.config: DEBUG: No config file found
compose.service: INFO: Building mssql
docker.api.build: DEBUG: Looking for auth config
docker.api.build: DEBUG: No auth config in memory - loading from filesystem
docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg']
docker.utils.config: DEBUG: No config file found
docker.api.build: DEBUG: Sending auth config ()
compose.parallel: DEBUG: Pending: set([<Service: mssql>])
compose.parallel: DEBUG: Starting producer thread for <Service: mssql>
compose.parallel: DEBUG: Pending: set([<Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>])
compose.parallel: DEBUG: Starting producer thread for <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.service: DEBUG: Added config hash: 3634631a392fe7d5944094282be0b8e946cb6a3c8f386ada7e1e488316879ad7
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Finished processing: <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Finished processing: <Service: mssql>
compose.parallel: DEBUG: Pending: set([])
--------------------- >> end captured logging << ---------------------
======================================================================
ERROR: MSSQL module outputs an event.
----------------------------------------------------------------------
Traceback (most recent call last):
File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 46, in test_status
self.assert_fields_are_documented(evt)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented
expected_fields, dict_fields, aliases = self.load_fields()
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields
subfields, subdictfields, subaliases = extract_fields(item["fields"], "")
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
KeyError: 'fields'
-------------------- >> begin captured stdout << ---------------------
render config
--------------------- >> end captured stdout << ----------------------
```
| 1.0 | [Metricbeat] Flaky test mssql - ## Flaky Test
* **Test Name:** x-pack/metricbeat/tests/system/test_mssql.py
* **Link:** https://travis-ci.org/elastic/beats/jobs/522183265#L683
### Stack Trace
```
etricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 ...
======================================================================
ERROR: MSSQL module outputs an event.
----------------------------------------------------------------------
Traceback (most recent call last):
File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 75, in test_performance
self.assert_fields_are_documented(evt)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented
expected_fields, dict_fields, aliases = self.load_fields()
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields
subfields, subdictfields, subaliases = extract_fields(item["fields"], "")
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
KeyError: 'fields'
-------------------- >> begin captured stdout << ---------------------
render config
--------------------- >> end captured stdout << ----------------------
-------------------- >> begin captured logging << --------------------
compose.config.config: DEBUG: Using configuration files: ./docker-compose.yml
compose.config.environment: WARNING: The PWD variable is not set. Defaulting to a blank string.
docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg']
docker.utils.config: DEBUG: No config file found
docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg']
docker.utils.config: DEBUG: No config file found
compose.service: INFO: Building mssql
docker.api.build: DEBUG: Looking for auth config
docker.api.build: DEBUG: No auth config in memory - loading from filesystem
docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg']
docker.utils.config: DEBUG: No config file found
docker.api.build: DEBUG: Sending auth config ()
compose.parallel: DEBUG: Pending: set([<Service: mssql>])
compose.parallel: DEBUG: Starting producer thread for <Service: mssql>
compose.parallel: DEBUG: Pending: set([<Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>])
compose.parallel: DEBUG: Starting producer thread for <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.service: DEBUG: Added config hash: 3634631a392fe7d5944094282be0b8e946cb6a3c8f386ada7e1e488316879ad7
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Finished processing: <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>
compose.parallel: DEBUG: Pending: set([])
compose.parallel: DEBUG: Finished processing: <Service: mssql>
compose.parallel: DEBUG: Pending: set([])
--------------------- >> end captured logging << ---------------------
======================================================================
ERROR: MSSQL module outputs an event.
----------------------------------------------------------------------
Traceback (most recent call last):
File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 46, in test_status
self.assert_fields_are_documented(evt)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented
expected_fields, dict_fields, aliases = self.load_fields()
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields
subfields, subdictfields, subaliases = extract_fields(item["fields"], "")
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields
subfields, subdictfields, subaliases = extract_fields(field["fields"], newName)
KeyError: 'fields'
-------------------- >> begin captured stdout << ---------------------
render config
--------------------- >> end captured stdout << ----------------------
```
| infrastructure | flaky test mssql flaky test test name x pack metricbeat tests system test mssql py link stack trace etricbeat snapshot mssql error mssql module outputs an event traceback most recent call last file go src github com elastic beats x pack metricbeat tests system test mssql py line in test performance self assert fields are documented evt file go src github com elastic beats libbeat tests system beat beat py line in assert fields are documented expected fields dict fields aliases self load fields file go src github com elastic beats libbeat tests system beat beat py line in load fields subfields subdictfields subaliases extract fields item file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname keyerror fields begin captured stdout render config end captured stdout begin captured logging compose config config debug using configuration files docker compose yml compose config environment warning the pwd variable is not set defaulting to a blank string docker utils config debug trying paths docker utils config debug no config file found docker utils config debug trying paths docker utils config debug no config file found compose service info building mssql docker api build debug looking for auth config docker api build debug no auth config in memory loading from filesystem docker utils config debug trying paths docker utils config debug no config file found docker api build debug sending auth config compose parallel debug pending set compose parallel debug starting producer thread for compose parallel debug pending set compose parallel debug starting producer thread for compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose service debug added config hash compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug finished processing compose parallel debug pending set compose parallel debug finished processing compose parallel debug pending set end captured logging error mssql module outputs an event traceback most recent call last file go src github com elastic beats x pack metricbeat tests system test mssql py line in test status self assert fields are documented evt file go src github com elastic beats libbeat tests system beat beat py line in assert fields are documented expected fields dict fields aliases self load fields file go src github com elastic beats libbeat tests system beat beat py line in load fields subfields subdictfields subaliases extract fields item file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname keyerror fields begin captured stdout render config end captured stdout | 1 |
7,358 | 6,918,829,551 | IssuesEvent | 2017-11-29 13:38:52 | Graylog2/graylog2-server | https://api.github.com/repos/Graylog2/graylog2-server | opened | Refactor Message class | infrastructure | The `Message` class' origins date very far back and the resulting code is not overly clean. There is a lot of overlapping API surface, which can be difficult to use and understand the consequences of.
Worse, we cannot inject anything into newly created Message objects, which means that any additional behavior either needs to be passed in as extra parameters to its various methods, or has to be added to the class itself which really is not be best place to have.
The idea is to create a `MessageProvider` that can be injected wherever new `Message` objects are required. The provider can then make sure that message objects are "injected" with the necessary references to other systems (such as stats recording, message weighers, etc). By creating a special provider we can work around the overhead of injecting directly. | 1.0 | Refactor Message class - The `Message` class' origins date very far back and the resulting code is not overly clean. There is a lot of overlapping API surface, which can be difficult to use and understand the consequences of.
Worse, we cannot inject anything into newly created Message objects, which means that any additional behavior either needs to be passed in as extra parameters to its various methods, or has to be added to the class itself which really is not be best place to have.
The idea is to create a `MessageProvider` that can be injected wherever new `Message` objects are required. The provider can then make sure that message objects are "injected" with the necessary references to other systems (such as stats recording, message weighers, etc). By creating a special provider we can work around the overhead of injecting directly. | infrastructure | refactor message class the message class origins date very far back and the resulting code is not overly clean there is a lot of overlapping api surface which can be difficult to use and understand the consequences of worse we cannot inject anything into newly created message objects which means that any additional behavior either needs to be passed in as extra parameters to its various methods or has to be added to the class itself which really is not be best place to have the idea is to create a messageprovider that can be injected wherever new message objects are required the provider can then make sure that message objects are injected with the necessary references to other systems such as stats recording message weighers etc by creating a special provider we can work around the overhead of injecting directly | 1 |
50,008 | 6,291,989,371 | IssuesEvent | 2017-07-20 03:18:04 | PaddlePaddle/Paddle | https://api.github.com/repos/PaddlePaddle/Paddle | closed | Paddle 应该是什么形态 | design_doc | 如果要允许用户在iPython和Jupiter之类的界面里写Paddle程序,那么Paddle得是一个library(提供本地函数调用)或者一个RPC server(提供远程函数调用),而不能是目前的executable command line tool的形式。
在library和RPC server之间的选择是个问题。欢迎大家讨论。 | 1.0 | Paddle 应该是什么形态 - 如果要允许用户在iPython和Jupiter之类的界面里写Paddle程序,那么Paddle得是一个library(提供本地函数调用)或者一个RPC server(提供远程函数调用),而不能是目前的executable command line tool的形式。
在library和RPC server之间的选择是个问题。欢迎大家讨论。 | non_infrastructure | paddle 应该是什么形态 如果要允许用户在ipython和jupiter之类的界面里写paddle程序,那么paddle得是一个library(提供本地函数调用)或者一个rpc server(提供远程函数调用),而不能是目前的executable command line tool的形式。 在library和rpc server之间的选择是个问题。欢迎大家讨论。 | 0 |
13,258 | 10,170,321,815 | IssuesEvent | 2019-08-08 04:46:55 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | closed | Sign NuGet packages prior to publish | area-Infrastructure | In PR https://github.com/dotnet/coreclr/pull/26020 sign checking has been deferred. This is being deferred because the coreclr doesn't presently sign `nupkg` files. Once that is done, sign validation can be re-enabled.
cc @dotnet/coreclr-infra | 1.0 | Sign NuGet packages prior to publish - In PR https://github.com/dotnet/coreclr/pull/26020 sign checking has been deferred. This is being deferred because the coreclr doesn't presently sign `nupkg` files. Once that is done, sign validation can be re-enabled.
cc @dotnet/coreclr-infra | infrastructure | sign nuget packages prior to publish in pr sign checking has been deferred this is being deferred because the coreclr doesn t presently sign nupkg files once that is done sign validation can be re enabled cc dotnet coreclr infra | 1 |
2,256 | 2,524,982,973 | IssuesEvent | 2015-01-20 21:24:34 | graybeal/ont | https://api.github.com/repos/graybeal/ont | opened | Add conditions field for usage metadata | 1 star enhancement imported Priority-Medium | _From [cmt1...@gmail.com](https://code.google.com/u/111118716825497141417/) on October 22, 2009 10:17:19_
What capability do you want added or improved? Add a field where ontology providers specify the "conditions" under which
author credit is required. Where do you want this capability to be accessible? Usage/permissions/license metadata tab What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=198_ | 1.0 | Add conditions field for usage metadata - _From [cmt1...@gmail.com](https://code.google.com/u/111118716825497141417/) on October 22, 2009 10:17:19_
What capability do you want added or improved? Add a field where ontology providers specify the "conditions" under which
author credit is required. Where do you want this capability to be accessible? Usage/permissions/license metadata tab What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=198_ | non_infrastructure | add conditions field for usage metadata from on october what capability do you want added or improved add a field where ontology providers specify the conditions under which author credit is required where do you want this capability to be accessible usage permissions license metadata tab what sort of input command mechanism do you want what is the desired output content format location other details of your desired capability what version of the product are you using please provide any additional information below particular ontology ies text contents of vocabulary operating system browser version firefox safari chrome ie etc screenshot etc original issue | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.