Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
4,485
5,110,358,838
IssuesEvent
2017-01-05 23:56:37
sketch-city/project-ideas
https://api.github.com/repos/sketch-city/project-ideas
closed
Regional Server Network for Regional Apps
infrastructure
Please describe your project, the problem you're solving, and why it's important. Keep it brief! Link to further reading if necessary. Local code irrespective of device downloadable from local servers / as close to backbone as possible.
1.0
Regional Server Network for Regional Apps - Please describe your project, the problem you're solving, and why it's important. Keep it brief! Link to further reading if necessary. Local code irrespective of device downloadable from local servers / as close to backbone as possible.
infrastructure
regional server network for regional apps please describe your project the problem you re solving and why it s important keep it brief link to further reading if necessary local code irrespective of device downloadable from local servers as close to backbone as possible
1
21,754
14,786,431,571
IssuesEvent
2021-01-12 05:29:43
pol-is/polis
https://api.github.com/repos/pol-is/polis
opened
Use GitHub Container Registry to store pre-built containers for test workflows
⚒️ infrastructure
Right now, we're pushing nightly builds to docker hub. In theory, this makes deploying quicker for new people, since can pull instead of building. In practice, I don't think it's used much. Having said that, we're also building containers in order to run cypress tests. If we start building them for cross-browser testing of old browsers on BrowserStack, then we'll be building the containers at least twice per commit. This expends twice as many build minutes as we need, and each docker build takes about 8 minutes. We could instead build the containers in one workflow, and push them to GitHub Container Registry. These would be set to private, so just for internal tests. We could then pull them in the workflows that needs to spin up an instance, without rebuilding them each time (e.g. cypress tests, browserstack tests, etc). GitHub Container Registry vs GitHub Docker Registry: https://docs.github.com/en/free-pro-team@latest/packages/guides/migrating-to-github-container-registry-for-docker-images (main thing is that GHCR has more fine-grained permissions) Enabling: https://docs.github.com/en/free-pro-team@latest/packages/guides/enabling-improved-container-support
1.0
Use GitHub Container Registry to store pre-built containers for test workflows - Right now, we're pushing nightly builds to docker hub. In theory, this makes deploying quicker for new people, since can pull instead of building. In practice, I don't think it's used much. Having said that, we're also building containers in order to run cypress tests. If we start building them for cross-browser testing of old browsers on BrowserStack, then we'll be building the containers at least twice per commit. This expends twice as many build minutes as we need, and each docker build takes about 8 minutes. We could instead build the containers in one workflow, and push them to GitHub Container Registry. These would be set to private, so just for internal tests. We could then pull them in the workflows that needs to spin up an instance, without rebuilding them each time (e.g. cypress tests, browserstack tests, etc). GitHub Container Registry vs GitHub Docker Registry: https://docs.github.com/en/free-pro-team@latest/packages/guides/migrating-to-github-container-registry-for-docker-images (main thing is that GHCR has more fine-grained permissions) Enabling: https://docs.github.com/en/free-pro-team@latest/packages/guides/enabling-improved-container-support
infrastructure
use github container registry to store pre built containers for test workflows right now we re pushing nightly builds to docker hub in theory this makes deploying quicker for new people since can pull instead of building in practice i don t think it s used much having said that we re also building containers in order to run cypress tests if we start building them for cross browser testing of old browsers on browserstack then we ll be building the containers at least twice per commit this expends twice as many build minutes as we need and each docker build takes about minutes we could instead build the containers in one workflow and push them to github container registry these would be set to private so just for internal tests we could then pull them in the workflows that needs to spin up an instance without rebuilding them each time e g cypress tests browserstack tests etc github container registry vs github docker registry main thing is that ghcr has more fine grained permissions enabling
1
14,160
10,678,268,586
IssuesEvent
2019-10-21 16:56:24
aspnet/AspNetCore
https://api.github.com/repos/aspnet/AspNetCore
closed
Unable to .\restore.cmd - TagBuilderWebSite.csproj
area-infrastructure
### Describe the bug Cannot restore from from clean master ### To Reproduce Steps to reproduce the behavior: 1. Using this version of ASP.NET Core '418e35c4396682aea6ee6e9833ff3bd3bfe624eb' 2. Run this code '\.restore.cmd' 3. See error ### Expected behavior Successful restore ### Screenshots If applicable, add screenshots to help explain your problem. ### Additional context dotnet --info: ``` C:\dev\AspNetCore>dotnet --info A compatible installed .NET Core SDK for global.json version [5.0.100-alpha1-014696] from [C:\dev\AspNetCore\global.json] was not found Install the [5.0.100-alpha1-014696] .NET Core SDK or update [C:\dev\AspNetCore\global.json] with an installed .NET Core SDK: 1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk] 1.0.0 [C:\Program Files\dotnet\sdk] 1.0.4 [C:\Program Files\dotnet\sdk] 2.1.4 [C:\Program Files\dotnet\sdk] 2.1.103 [C:\Program Files\dotnet\sdk] 2.1.104 [C:\Program Files\dotnet\sdk] 2.1.201 [C:\Program Files\dotnet\sdk] 2.1.202 [C:\Program Files\dotnet\sdk] 2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk] 2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk] 2.1.300 [C:\Program Files\dotnet\sdk] 2.1.401 [C:\Program Files\dotnet\sdk] 2.1.402 [C:\Program Files\dotnet\sdk] 2.1.403 [C:\Program Files\dotnet\sdk] 2.1.502 [C:\Program Files\dotnet\sdk] 2.1.503 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009426 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009472 [C:\Program Files\dotnet\sdk] 2.1.602 [C:\Program Files\dotnet\sdk] 2.1.800 [C:\Program Files\dotnet\sdk] 2.2.101 [C:\Program Files\dotnet\sdk] 3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk] 3.0.100 [C:\Program Files\dotnet\sdk] 3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk] Host (useful for support): Version: 5.0.0-alpha1.19514.1 Commit: 4ace84dbf9 .NET Core SDKs installed: 1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk] 1.0.0 [C:\Program Files\dotnet\sdk] 1.0.4 [C:\Program Files\dotnet\sdk] 2.1.4 [C:\Program Files\dotnet\sdk] 2.1.103 [C:\Program Files\dotnet\sdk] 2.1.104 [C:\Program Files\dotnet\sdk] 2.1.201 [C:\Program Files\dotnet\sdk] 2.1.202 [C:\Program Files\dotnet\sdk] 2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk] 2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk] 2.1.300 [C:\Program Files\dotnet\sdk] 2.1.401 [C:\Program Files\dotnet\sdk] 2.1.402 [C:\Program Files\dotnet\sdk] 2.1.403 [C:\Program Files\dotnet\sdk] 2.1.502 [C:\Program Files\dotnet\sdk] 2.1.503 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009426 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009472 [C:\Program Files\dotnet\sdk] 2.1.602 [C:\Program Files\dotnet\sdk] 2.1.800 [C:\Program Files\dotnet\sdk] 2.2.101 [C:\Program Files\dotnet\sdk] 3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk] 3.0.100 [C:\Program Files\dotnet\sdk] 3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk] .NET Core runtimes installed: Microsoft.AspNetCore.All 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.App 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.0.0-preview6.19307.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.0-preview1.19508.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.0-dev [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 1.0.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.0.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.1.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.1.2 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.0-preview1-26216-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.0-rc1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.3-servicing-26724-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.0-alpha1.19514.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] To install additional .NET Core runtimes or SDKs: https://aka.ms/dotnet-download ``` Console output: ``` C:\dev\AspNetCore>git status On branch master Your branch is up to date with 'origin/master'. nothing to commit, working tree clean C:\dev\AspNetCore>.\restore.cmd Building of C# project is enabled and has dependencies on NodeJS projects. Building of NodeJS projects is enabled since node is detected in C:\Program Files. Wiederherstellung in "41,94 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade .sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process. Wiederherstellung in "14,54 ms" für "C:\dev\AspNetCore\eng\tools\RepoTasks\RepoTasks.csproj " abgeschlossen. RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\netcoreapp5.0\RepoTasks.dll RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\net472\RepoTasks.dll Wiederherstellung in "41,11 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade .sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process. C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft. Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T agHelpersWebSite\TagHelpersWebSite.csproj] Fehler beim Buildvorgang. C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft. Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T agHelpersWebSite\TagHelpersWebSite.csproj] 0 Warnung(en) 1 Fehler Verstrichene Zeit 00:00:15.58 Build failed. ``` I am new here, and have tried A LOT of things. I've got the restore and build running once, but installed VS 2019 Preview in an attempt to get targeting the locally built artifacts. I've uninstalled already, but am still getting these errors now. How does the build pipeline decide which version of VS to use, and why is it that as soon as I have any other version of VS, it will always break?
1.0
Unable to .\restore.cmd - TagBuilderWebSite.csproj - ### Describe the bug Cannot restore from from clean master ### To Reproduce Steps to reproduce the behavior: 1. Using this version of ASP.NET Core '418e35c4396682aea6ee6e9833ff3bd3bfe624eb' 2. Run this code '\.restore.cmd' 3. See error ### Expected behavior Successful restore ### Screenshots If applicable, add screenshots to help explain your problem. ### Additional context dotnet --info: ``` C:\dev\AspNetCore>dotnet --info A compatible installed .NET Core SDK for global.json version [5.0.100-alpha1-014696] from [C:\dev\AspNetCore\global.json] was not found Install the [5.0.100-alpha1-014696] .NET Core SDK or update [C:\dev\AspNetCore\global.json] with an installed .NET Core SDK: 1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk] 1.0.0 [C:\Program Files\dotnet\sdk] 1.0.4 [C:\Program Files\dotnet\sdk] 2.1.4 [C:\Program Files\dotnet\sdk] 2.1.103 [C:\Program Files\dotnet\sdk] 2.1.104 [C:\Program Files\dotnet\sdk] 2.1.201 [C:\Program Files\dotnet\sdk] 2.1.202 [C:\Program Files\dotnet\sdk] 2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk] 2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk] 2.1.300 [C:\Program Files\dotnet\sdk] 2.1.401 [C:\Program Files\dotnet\sdk] 2.1.402 [C:\Program Files\dotnet\sdk] 2.1.403 [C:\Program Files\dotnet\sdk] 2.1.502 [C:\Program Files\dotnet\sdk] 2.1.503 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009426 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009472 [C:\Program Files\dotnet\sdk] 2.1.602 [C:\Program Files\dotnet\sdk] 2.1.800 [C:\Program Files\dotnet\sdk] 2.2.101 [C:\Program Files\dotnet\sdk] 3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk] 3.0.100 [C:\Program Files\dotnet\sdk] 3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk] Host (useful for support): Version: 5.0.0-alpha1.19514.1 Commit: 4ace84dbf9 .NET Core SDKs installed: 1.0.0-preview4-004233 [C:\Program Files\dotnet\sdk] 1.0.0 [C:\Program Files\dotnet\sdk] 1.0.4 [C:\Program Files\dotnet\sdk] 2.1.4 [C:\Program Files\dotnet\sdk] 2.1.103 [C:\Program Files\dotnet\sdk] 2.1.104 [C:\Program Files\dotnet\sdk] 2.1.201 [C:\Program Files\dotnet\sdk] 2.1.202 [C:\Program Files\dotnet\sdk] 2.1.300-preview1-008174 [C:\Program Files\dotnet\sdk] 2.1.300-rc1-008673 [C:\Program Files\dotnet\sdk] 2.1.300 [C:\Program Files\dotnet\sdk] 2.1.401 [C:\Program Files\dotnet\sdk] 2.1.402 [C:\Program Files\dotnet\sdk] 2.1.403 [C:\Program Files\dotnet\sdk] 2.1.502 [C:\Program Files\dotnet\sdk] 2.1.503 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009426 [C:\Program Files\dotnet\sdk] 2.1.600-preview-009472 [C:\Program Files\dotnet\sdk] 2.1.602 [C:\Program Files\dotnet\sdk] 2.1.800 [C:\Program Files\dotnet\sdk] 2.2.101 [C:\Program Files\dotnet\sdk] 3.0.100-preview6-012264 [C:\Program Files\dotnet\sdk] 3.0.100 [C:\Program Files\dotnet\sdk] 3.1.100-preview1-014459 [C:\Program Files\dotnet\sdk] .NET Core runtimes installed: Microsoft.AspNetCore.All 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.All 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.All] Microsoft.AspNetCore.App 2.1.0-preview1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.0-rc1-final [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.0.0-preview6.19307.2 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 3.1.0-preview1.19508.20 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 5.0.0-dev [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 1.0.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.0.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.1.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 1.1.2 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.0.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.0-preview1-26216-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.0-rc1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.3-servicing-26724-03 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.3 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.4 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.5 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.6 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.7 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.9 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.1.12 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 2.2.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 5.0.0-alpha1.19514.1 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.0.0-preview6-27804-01 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 3.1.0-preview1.19506.1 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] To install additional .NET Core runtimes or SDKs: https://aka.ms/dotnet-download ``` Console output: ``` C:\dev\AspNetCore>git status On branch master Your branch is up to date with 'origin/master'. nothing to commit, working tree clean C:\dev\AspNetCore>.\restore.cmd Building of C# project is enabled and has dependencies on NodeJS projects. Building of NodeJS projects is enabled since node is detected in C:\Program Files. Wiederherstellung in "41,94 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade .sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process. Wiederherstellung in "14,54 ms" für "C:\dev\AspNetCore\eng\tools\RepoTasks\RepoTasks.csproj " abgeschlossen. RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\netcoreapp5.0\RepoTasks.dll RepoTasks -> C:\dev\AspNetCore\artifacts\bin\RepoTasks\Release\net472\RepoTasks.dll Wiederherstellung in "41,11 ms" für "C:\Users\admin\.nuget\packages\microsoft.dotnet.arcade .sdk\1.0.0-beta.19462.4\tools\Tools.proj" abgeschlossen. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\". Note: This change will not be visible if PowerShell was run as a child process. dotnet-install: .NET Core Runtime version 5.0.0-alpha1.19514.1 is already installed. dotnet-install: Adding to current process PATH: "C:\dev\AspNetCore\.dotnet\x86\". Note: This change will not be visible if PowerShell was run as a child process. C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft. Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T agHelpersWebSite\TagHelpersWebSite.csproj] Fehler beim Buildvorgang. C:\Program Files (x86)\Microsoft Visual Studio\2019\Enterprise\MSBuild\Current\Bin\Microsoft. Common.CurrentVersion.targets(31,3): error MSB4024: Die importierte Projektdatei "C:\dev\AspN etCore\src\Mvc\test\WebSites\TagHelpersWebSite\TagHelpersWebSite.csproj.user" konnte nicht ge laden werden Das Stammelement ist nicht vorhanden. [C:\dev\AspNetCore\src\Mvc\test\WebSites\T agHelpersWebSite\TagHelpersWebSite.csproj] 0 Warnung(en) 1 Fehler Verstrichene Zeit 00:00:15.58 Build failed. ``` I am new here, and have tried A LOT of things. I've got the restore and build running once, but installed VS 2019 Preview in an attempt to get targeting the locally built artifacts. I've uninstalled already, but am still getting these errors now. How does the build pipeline decide which version of VS to use, and why is it that as soon as I have any other version of VS, it will always break?
infrastructure
unable to restore cmd tagbuilderwebsite csproj describe the bug cannot restore from from clean master to reproduce steps to reproduce the behavior using this version of asp net core run this code restore cmd see error expected behavior successful restore screenshots if applicable add screenshots to help explain your problem additional context dotnet info c dev aspnetcore dotnet info a compatible installed net core sdk for global json version from was not found install the net core sdk or update with an installed net core sdk preview preview host useful for support version commit net core sdks installed preview preview net core runtimes installed microsoft aspnetcore all final microsoft aspnetcore all final microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore all microsoft aspnetcore app final microsoft aspnetcore app final microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app dev microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app servicing microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app to install additional net core runtimes or sdks console output c dev aspnetcore git status on branch master your branch is up to date with origin master nothing to commit working tree clean c dev aspnetcore restore cmd building of c project is enabled and has dependencies on nodejs projects building of nodejs projects is enabled since node is detected in c program files wiederherstellung in ms für c users admin nuget packages microsoft dotnet arcade sdk beta tools tools proj abgeschlossen dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process wiederherstellung in ms für c dev aspnetcore eng tools repotasks repotasks csproj abgeschlossen repotasks c dev aspnetcore artifacts bin repotasks release repotasks dll repotasks c dev aspnetcore artifacts bin repotasks release repotasks dll wiederherstellung in ms für c users admin nuget packages microsoft dotnet arcade sdk beta tools tools proj abgeschlossen dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process dotnet install net core runtime version is already installed dotnet install adding to current process path c dev aspnetcore dotnet note this change will not be visible if powershell was run as a child process c program files microsoft visual studio enterprise msbuild current bin microsoft common currentversion targets error die importierte projektdatei c dev aspn etcore src mvc test websites taghelperswebsite taghelperswebsite csproj user konnte nicht ge laden werden das stammelement ist nicht vorhanden c dev aspnetcore src mvc test websites t aghelperswebsite taghelperswebsite csproj fehler beim buildvorgang c program files microsoft visual studio enterprise msbuild current bin microsoft common currentversion targets error die importierte projektdatei c dev aspn etcore src mvc test websites taghelperswebsite taghelperswebsite csproj user konnte nicht ge laden werden das stammelement ist nicht vorhanden c dev aspnetcore src mvc test websites t aghelperswebsite taghelperswebsite csproj warnung en fehler verstrichene zeit build failed i am new here and have tried a lot of things i ve got the restore and build running once but installed vs preview in an attempt to get targeting the locally built artifacts i ve uninstalled already but am still getting these errors now how does the build pipeline decide which version of vs to use and why is it that as soon as i have any other version of vs it will always break
1
10,253
8,453,125,979
IssuesEvent
2018-10-20 12:33:09
TeamBravo2018/cloned-rfid-card-detection
https://api.github.com/repos/TeamBravo2018/cloned-rfid-card-detection
opened
Setup MQTT Simulator
backlog item infrastructure messaging test production
### Description ### Simulate messaging between the applications by using MQTT simulator. [https://dzone.com/articles/top-3-online-tools-to-simulate-an-mqtt-client]()
1.0
Setup MQTT Simulator - ### Description ### Simulate messaging between the applications by using MQTT simulator. [https://dzone.com/articles/top-3-online-tools-to-simulate-an-mqtt-client]()
infrastructure
setup mqtt simulator description simulate messaging between the applications by using mqtt simulator
1
813,125
30,446,199,738
IssuesEvent
2023-07-15 17:48:03
ncssar/radiolog
https://api.github.com/repos/ncssar/radiolog
opened
generated log PDFs should be searchable
enhancement Priority:Medium
not so important for generated clue report PDFs, but all others - radio log, team logs, clue log - should be searchable (using e.g. acrobat reader); surprised to see that they are currently not searchable
1.0
generated log PDFs should be searchable - not so important for generated clue report PDFs, but all others - radio log, team logs, clue log - should be searchable (using e.g. acrobat reader); surprised to see that they are currently not searchable
non_infrastructure
generated log pdfs should be searchable not so important for generated clue report pdfs but all others radio log team logs clue log should be searchable using e g acrobat reader surprised to see that they are currently not searchable
0
716,244
24,626,321,471
IssuesEvent
2022-10-16 15:11:39
AY2223S1-CS2103T-T10-1/tp
https://api.github.com/repos/AY2223S1-CS2103T-T10-1/tp
opened
As a forgetful student, I want to know which of my friends take a common module
type.Story priority.High
so that I know who to approach when I need help with that module's work.
1.0
As a forgetful student, I want to know which of my friends take a common module - so that I know who to approach when I need help with that module's work.
non_infrastructure
as a forgetful student i want to know which of my friends take a common module so that i know who to approach when i need help with that module s work
0
293,730
25,318,877,925
IssuesEvent
2022-11-18 00:54:22
dotnet/sdk
https://api.github.com/repos/dotnet/sdk
closed
dotnet test fail to pass MSBuild properties
Area-DotNet Test untriaged
### Describe the bug Until .Net 6.x, we were able to run ` dotnet test App.sln -p Property=Value ` and the property was adequately passed down to MSBuild, interpreted by project files and the like. On upgrading to .Net 7.x, we found the property was no longer passed down. However, we have a workaround ` dotnet test -p Property=Value App.sln ` ### To Reproduce <!-- We ❤ code! Point us to a minimalistic repro project hosted in a GitHub repo, Gist snippet, or other means to see the isolated behavior. We may close this issue if: - the repro project you share with us is complex. We can't investigate custom projects, so don't point us to such, please. - if we will not be able to repro the behavior you're reporting --> Fail - runs only one test ` dotnet test TestProject1.csproj -p:CIBuild=Integration ` Correct - runs only one test ` dotnet test -p:CIBuild=Unit TestProject1.csproj ` Correct - runs the two tests ` dotnet test -p:CIBuild=Integration TestProject1.csproj ` **TestProject1.csproj** ```xml <Project Sdk="Microsoft.NET.Sdk"> <PropertyGroup> <TargetFramework>net7.0</TargetFramework> <ImplicitUsings>enable</ImplicitUsings> <Nullable>enable</Nullable> <IsPackable>false</IsPackable> <ApplicationManifest>app.manifest</ApplicationManifest> </PropertyGroup> <ItemGroup> <PackageReference Include="Microsoft.NET.Test.Sdk" Version="17.3.2" /> <PackageReference Include="MSTest.TestAdapter" Version="2.2.10" /> <PackageReference Include="MSTest.TestFramework" Version="2.2.10" /> <PackageReference Include="coverlet.collector" Version="3.1.2" /> </ItemGroup> </Project> ``` **Directory.Build.props** ```xml <Project> <PropertyGroup> <LangVersion>11.0</LangVersion> <TargetFramework>net7.0</TargetFramework> <RunSettingsFilePath>$(MSBuildThisFileDirectory)Unit.runsettings</RunSettingsFilePath> <RunSettingsFilePath Condition="$(CIBuild) != ''">$(MSBuildThisFileDirectory)$(CIBuild).runsettings</RunSettingsFilePath> </PropertyGroup> </Project> ``` **Unit.runsettings** ```xml <RunSettings> <RunConfiguration> <TestCaseFilter>TestCategory=Unit</TestCaseFilter> </RunConfiguration> </RunSettings> ``` **Integration.runsettings** ```xml <RunSettings> <RunConfiguration> <TestCaseFilter>TestCategory=Unit|TestCategory=Integration</TestCaseFilter> </RunConfiguration> </RunSettings> ``` **UnitTest1.cs** ```xml namespace TestProject1 { [TestClass] public class UnitTest1 { [TestMethod, TestCategory("Unit")] public void TestMethod1() { } [TestMethod, TestCategory("Integration")] public void TestMethod2() { } } } ``` ### Exceptions (if any) No exceptions ### Further technical details - Include the output of `dotnet --info` ` .NET SDK: Version: 7.0.100 Commit: e12b7af219 Runtime Environment: OS Name: Windows OS Version: 10.0.22621 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\7.0.100\ Host: Version: 7.0.0 Architecture: x64 Commit: d099f075e4 .NET SDKs installed: 3.1.425 [C:\Program Files\dotnet\sdk] 6.0.202 [C:\Program Files\dotnet\sdk] 6.0.306 [C:\Program Files\dotnet\sdk] 7.0.100 [C:\Program Files\dotnet\sdk] .NET runtimes installed: Microsoft.AspNetCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Other architectures found: x86 [C:\Program Files (x86)\dotnet] registered at [HKLM\SOFTWARE\dotnet\Setup\InstalledVersions\x86\InstallLocation] Environment variables: Not set global.json file: Not found Learn more: https://aka.ms/dotnet/info Download .NET: https://aka.ms/dotnet/download ` - VS 17.4 ### Related Passing properties seem to have worked since earlier versions, but it is not well documented. #23198
1.0
dotnet test fail to pass MSBuild properties - ### Describe the bug Until .Net 6.x, we were able to run ` dotnet test App.sln -p Property=Value ` and the property was adequately passed down to MSBuild, interpreted by project files and the like. On upgrading to .Net 7.x, we found the property was no longer passed down. However, we have a workaround ` dotnet test -p Property=Value App.sln ` ### To Reproduce <!-- We ❤ code! Point us to a minimalistic repro project hosted in a GitHub repo, Gist snippet, or other means to see the isolated behavior. We may close this issue if: - the repro project you share with us is complex. We can't investigate custom projects, so don't point us to such, please. - if we will not be able to repro the behavior you're reporting --> Fail - runs only one test ` dotnet test TestProject1.csproj -p:CIBuild=Integration ` Correct - runs only one test ` dotnet test -p:CIBuild=Unit TestProject1.csproj ` Correct - runs the two tests ` dotnet test -p:CIBuild=Integration TestProject1.csproj ` **TestProject1.csproj** ```xml <Project Sdk="Microsoft.NET.Sdk"> <PropertyGroup> <TargetFramework>net7.0</TargetFramework> <ImplicitUsings>enable</ImplicitUsings> <Nullable>enable</Nullable> <IsPackable>false</IsPackable> <ApplicationManifest>app.manifest</ApplicationManifest> </PropertyGroup> <ItemGroup> <PackageReference Include="Microsoft.NET.Test.Sdk" Version="17.3.2" /> <PackageReference Include="MSTest.TestAdapter" Version="2.2.10" /> <PackageReference Include="MSTest.TestFramework" Version="2.2.10" /> <PackageReference Include="coverlet.collector" Version="3.1.2" /> </ItemGroup> </Project> ``` **Directory.Build.props** ```xml <Project> <PropertyGroup> <LangVersion>11.0</LangVersion> <TargetFramework>net7.0</TargetFramework> <RunSettingsFilePath>$(MSBuildThisFileDirectory)Unit.runsettings</RunSettingsFilePath> <RunSettingsFilePath Condition="$(CIBuild) != ''">$(MSBuildThisFileDirectory)$(CIBuild).runsettings</RunSettingsFilePath> </PropertyGroup> </Project> ``` **Unit.runsettings** ```xml <RunSettings> <RunConfiguration> <TestCaseFilter>TestCategory=Unit</TestCaseFilter> </RunConfiguration> </RunSettings> ``` **Integration.runsettings** ```xml <RunSettings> <RunConfiguration> <TestCaseFilter>TestCategory=Unit|TestCategory=Integration</TestCaseFilter> </RunConfiguration> </RunSettings> ``` **UnitTest1.cs** ```xml namespace TestProject1 { [TestClass] public class UnitTest1 { [TestMethod, TestCategory("Unit")] public void TestMethod1() { } [TestMethod, TestCategory("Integration")] public void TestMethod2() { } } } ``` ### Exceptions (if any) No exceptions ### Further technical details - Include the output of `dotnet --info` ` .NET SDK: Version: 7.0.100 Commit: e12b7af219 Runtime Environment: OS Name: Windows OS Version: 10.0.22621 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\7.0.100\ Host: Version: 7.0.0 Architecture: x64 Commit: d099f075e4 .NET SDKs installed: 3.1.425 [C:\Program Files\dotnet\sdk] 6.0.202 [C:\Program Files\dotnet\sdk] 6.0.306 [C:\Program Files\dotnet\sdk] 7.0.100 [C:\Program Files\dotnet\sdk] .NET runtimes installed: Microsoft.AspNetCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.AspNetCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.AspNetCore.App] Microsoft.NETCore.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.NETCore.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.NETCore.App] Microsoft.WindowsDesktop.App 3.1.31 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 6.0.11 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Microsoft.WindowsDesktop.App 7.0.0 [C:\Program Files\dotnet\shared\Microsoft.WindowsDesktop.App] Other architectures found: x86 [C:\Program Files (x86)\dotnet] registered at [HKLM\SOFTWARE\dotnet\Setup\InstalledVersions\x86\InstallLocation] Environment variables: Not set global.json file: Not found Learn more: https://aka.ms/dotnet/info Download .NET: https://aka.ms/dotnet/download ` - VS 17.4 ### Related Passing properties seem to have worked since earlier versions, but it is not well documented. #23198
non_infrastructure
dotnet test fail to pass msbuild properties describe the bug until net x we were able to run dotnet test app sln p property value and the property was adequately passed down to msbuild interpreted by project files and the like on upgrading to net x we found the property was no longer passed down however we have a workaround dotnet test p property value app sln to reproduce we ❤ code point us to a minimalistic repro project hosted in a github repo gist snippet or other means to see the isolated behavior we may close this issue if the repro project you share with us is complex we can t investigate custom projects so don t point us to such please if we will not be able to repro the behavior you re reporting fail runs only one test dotnet test csproj p cibuild integration correct runs only one test dotnet test p cibuild unit csproj correct runs the two tests dotnet test p cibuild integration csproj csproj xml enable enable false app manifest directory build props xml msbuildthisfiledirectory unit runsettings msbuildthisfiledirectory cibuild runsettings unit runsettings xml testcategory unit integration runsettings xml testcategory unit testcategory integration cs xml namespace public class public void public void exceptions if any no exceptions further technical details include the output of dotnet info net sdk version commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk host version architecture commit net sdks installed net runtimes installed microsoft aspnetcore app microsoft aspnetcore app microsoft aspnetcore app microsoft netcore app microsoft netcore app microsoft netcore app microsoft windowsdesktop app microsoft windowsdesktop app microsoft windowsdesktop app other architectures found registered at environment variables not set global json file not found learn more download net vs related passing properties seem to have worked since earlier versions but it is not well documented
0
6,370
6,361,319,796
IssuesEvent
2017-07-31 12:37:23
warg-lang/warg
https://api.github.com/repos/warg-lang/warg
opened
Generate accessible static analysis diagnostics
ci enhancement infrastructure
[neovim](https://github.com/neovim/neovim) provides a nice [diagnostics overview](https://neovim.io/doc/reports/clang/) using Clang Static Analysis. While it is not completely transparent how to do that, making similar page would be great.
1.0
Generate accessible static analysis diagnostics - [neovim](https://github.com/neovim/neovim) provides a nice [diagnostics overview](https://neovim.io/doc/reports/clang/) using Clang Static Analysis. While it is not completely transparent how to do that, making similar page would be great.
infrastructure
generate accessible static analysis diagnostics provides a nice using clang static analysis while it is not completely transparent how to do that making similar page would be great
1
6,826
6,657,418,989
IssuesEvent
2017-09-30 05:13:33
nathantspencer/AtLAS-mobile
https://api.github.com/repos/nathantspencer/AtLAS-mobile
closed
UI + Request for Sign In/Out
Client-Server Security UI
Sign in request will require username and password. Sign out request will require username and authentication key.
True
UI + Request for Sign In/Out - Sign in request will require username and password. Sign out request will require username and authentication key.
non_infrastructure
ui request for sign in out sign in request will require username and password sign out request will require username and authentication key
0
57,982
7,110,431,567
IssuesEvent
2018-01-17 10:36:10
Userfeeds/Apps
https://api.github.com/repos/Userfeeds/Apps
closed
Inform that Metamask is disabled
design
Currently whenever my Metamask is not logged I can't see any functional buttons. I'd rather have them disabled and inform the user about this issue. ![metamask](https://user-images.githubusercontent.com/6234032/34741243-442e5cde-f582-11e7-8ba9-4f81d8e7f10f.png)
1.0
Inform that Metamask is disabled - Currently whenever my Metamask is not logged I can't see any functional buttons. I'd rather have them disabled and inform the user about this issue. ![metamask](https://user-images.githubusercontent.com/6234032/34741243-442e5cde-f582-11e7-8ba9-4f81d8e7f10f.png)
non_infrastructure
inform that metamask is disabled currently whenever my metamask is not logged i can t see any functional buttons i d rather have them disabled and inform the user about this issue
0
2,631
2,699,148,226
IssuesEvent
2015-04-03 14:47:24
itgsod-lukas-michanek/Neocache
https://api.github.com/repos/itgsod-lukas-michanek/Neocache
opened
Project cleanup
documentation
We are soon about to leave the stage of database modelling, but before we do, make sure that everything is properly commented, formatted and everything else.
1.0
Project cleanup - We are soon about to leave the stage of database modelling, but before we do, make sure that everything is properly commented, formatted and everything else.
non_infrastructure
project cleanup we are soon about to leave the stage of database modelling but before we do make sure that everything is properly commented formatted and everything else
0
179,745
21,580,319,605
IssuesEvent
2022-05-02 17:59:45
vincenzodistasio97/excel-to-json
https://api.github.com/repos/vincenzodistasio97/excel-to-json
opened
CVE-2020-28498 (Medium) detected in elliptic-6.4.0.tgz
security vulnerability
## CVE-2020-28498 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p> <p>Path to dependency file: /client/package.json</p> <p>Path to vulnerable library: /client/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.1.tgz (Root Library) - webpack-3.8.1.tgz - node-libs-browser-2.1.0.tgz - crypto-browserify-3.12.0.tgz - create-ecdh-4.0.0.tgz - :x: **elliptic-6.4.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/excel-to-json/commit/e367d4db4134dc676344b2b9fb2443300bd3c9c7">e367d4db4134dc676344b2b9fb2443300bd3c9c7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution (elliptic): 6.5.4</p> <p>Direct dependency fix Resolution (react-scripts): 1.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28498 (Medium) detected in elliptic-6.4.0.tgz - ## CVE-2020-28498 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.4.0.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.4.0.tgz</a></p> <p>Path to dependency file: /client/package.json</p> <p>Path to vulnerable library: /client/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.1.tgz (Root Library) - webpack-3.8.1.tgz - node-libs-browser-2.1.0.tgz - crypto-browserify-3.12.0.tgz - create-ecdh-4.0.0.tgz - :x: **elliptic-6.4.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/excel-to-json/commit/e367d4db4134dc676344b2b9fb2443300bd3c9c7">e367d4db4134dc676344b2b9fb2443300bd3c9c7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution (elliptic): 6.5.4</p> <p>Direct dependency fix Resolution (react-scripts): 1.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve medium detected in elliptic tgz cve medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file client package json path to vulnerable library client node modules elliptic package json dependency hierarchy react scripts tgz root library webpack tgz node libs browser tgz crypto browserify tgz create ecdh tgz x elliptic tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution elliptic direct dependency fix resolution react scripts step up your open source security game with whitesource
0
24,255
17,046,813,829
IssuesEvent
2021-07-06 00:59:33
google/iree
https://api.github.com/repos/google/iree
closed
[build] cant build iree
bug 🐞 help wanted infrastructure 🛠️ support 🤗
Hello, I have some problems and errors while building IREE I just followed ```getting started``` to build https://google.github.io/iree/building-from-source/getting-started/ Could you give some hints or instructions for this problems? Thanks! ## Trials ``` $ sudo apt-get install clang lld $ sudo apt-get install clang++ $ sudo apt install python-clang-4.0 $ sudo apt install python-clang-5.0 $ sudo apt install python-clang-6.0 $ sudo apt install python-clang-7 $ sudo apt install python-clang-8 $ sudo apt install python-clang-9 $ sudo apt install python3-clang-10 $ sudo apt update $ sudo apt upgrade $ sudo apt autoremove # g++ to clang++ $ sudo update-alternatives --config c++ # gcc to clang $ sudo update-alternatives --config cc ``` ``` $ export CC=clang $ export CXX=clang++ $ sudo rm -r ../iree-build $ cmake -B ../iree-build/ -DCMAKE_BUILD_TYPE=RelWithDebInfo . -GNinja $ cmake --build ../iree-build/ -j6 ``` ## Machine Spec - OS : Ubuntu 18.04.5 LTS - CPU : Intel(R) Core(TM) i7-10700K CPU @ 3.80GHz - RAM : 16GB - SSD : nvme, samsung ## Verbose log ### Results of checking ```$CC```, ```$CXX``` after failed with errors ![image](https://user-images.githubusercontent.com/40026920/124450483-6c46bf80-ddbf-11eb-9601-4e38690d6330.png) ``` $ cmake --build ../iree-build/ -j6 [0/2] Re-checking globbed directories... [2905/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o /usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:47:24: error: only virtual member functions can be marked 'override' void runOnFunction() override { ^~~~~~~~~ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:48:5: error: use of undeclared identifier 'getFunction' getFunction().walk([&](WhileOp whileOp) { MatchAndRewrite(whileOp); }); ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>' requested here return std::make_unique<ControlFlowToScfPass>(); ^ /home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'ControlFlowToScfPass' virtual void runOnOperation() = 0; ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:125:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, const mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass &>' requested here return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this)); ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowToScfPassBase<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::clonePass' requested here delete __ptr; ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::operator()' requested here get_deleter()(__ptr); ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass> >::~unique_ptr' requested here return std::make_unique<ControlFlowToScfPass>(); ^ 4 errors generated. [2906/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o /usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:187:24: error: only virtual member functions can be marked 'override' void runOnFunction() override; ^~~~~~~~ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:205:38: error: use of undeclared identifier 'getFunction' (void)applyPatternsAndFoldGreedily(getFunction(), std::move(patterns)); ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>' requested here return std::make_unique<LegalizeToStandardPass>(); ^ /home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeToStandardPass' virtual void runOnOperation() = 0; ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:229:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, const mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass &>' requested here return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this)); ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeToStandardPassBase<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::clonePass' requested here delete __ptr; ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::operator()' requested here get_deleter()(__ptr); ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass> >::~unique_ptr' requested here return std::make_unique<LegalizeToStandardPass>(); ^ 4 errors generated. [2907/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o /usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:43:24: error: only virtual member functions can be marked 'override' void runOnFunction() override; ^~~~~~~~ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:216:15: error: use of undeclared identifier 'getFunction' auto func = getFunction(); ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>' requested here return std::make_unique<LegalizeControlFlowPass>(); ^ /home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeControlFlowPass' virtual void runOnOperation() = 0; ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:99:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, const mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass &>' requested here return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this)); ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowPassBase<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::clonePass' requested here delete __ptr; ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::operator()' requested here get_deleter()(__ptr); ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass> >::~unique_ptr' requested here return std::make_unique<LegalizeControlFlowPass>(); ^ 4 errors generated. [2910/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/IR/CMakeFiles/obj.MhloDialect.dir/hlo_ops.cc.o ninja: build stopped: subcommand failed. ```
1.0
[build] cant build iree - Hello, I have some problems and errors while building IREE I just followed ```getting started``` to build https://google.github.io/iree/building-from-source/getting-started/ Could you give some hints or instructions for this problems? Thanks! ## Trials ``` $ sudo apt-get install clang lld $ sudo apt-get install clang++ $ sudo apt install python-clang-4.0 $ sudo apt install python-clang-5.0 $ sudo apt install python-clang-6.0 $ sudo apt install python-clang-7 $ sudo apt install python-clang-8 $ sudo apt install python-clang-9 $ sudo apt install python3-clang-10 $ sudo apt update $ sudo apt upgrade $ sudo apt autoremove # g++ to clang++ $ sudo update-alternatives --config c++ # gcc to clang $ sudo update-alternatives --config cc ``` ``` $ export CC=clang $ export CXX=clang++ $ sudo rm -r ../iree-build $ cmake -B ../iree-build/ -DCMAKE_BUILD_TYPE=RelWithDebInfo . -GNinja $ cmake --build ../iree-build/ -j6 ``` ## Machine Spec - OS : Ubuntu 18.04.5 LTS - CPU : Intel(R) Core(TM) i7-10700K CPU @ 3.80GHz - RAM : 16GB - SSD : nvme, samsung ## Verbose log ### Results of checking ```$CC```, ```$CXX``` after failed with errors ![image](https://user-images.githubusercontent.com/40026920/124450483-6c46bf80-ddbf-11eb-9601-4e38690d6330.png) ``` $ cmake --build ../iree-build/ -j6 [0/2] Re-checking globbed directories... [2905/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o /usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/mhlo_control_flow_to_scf.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:47:24: error: only virtual member functions can be marked 'override' void runOnFunction() override { ^~~~~~~~~ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:48:5: error: use of undeclared identifier 'getFunction' getFunction().walk([&](WhileOp whileOp) { MatchAndRewrite(whileOp); }); ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>' requested here return std::make_unique<ControlFlowToScfPass>(); ^ /home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'ControlFlowToScfPass' virtual void runOnOperation() = 0; ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:16: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/Support/Casting.h:20: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:125:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, const mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass &>' requested here return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this)); ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowToScfPassBase<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::clonePass' requested here delete __ptr; ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass>::operator()' requested here get_deleter()(__ptr); ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/mhlo_control_flow_to_scf.cc:202:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass, std::default_delete<mlir::mhlo::(anonymous namespace)::ControlFlowToScfPass> >::~unique_ptr' requested here return std::make_unique<ControlFlowToScfPass>(); ^ 4 errors generated. [2906/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o /usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_to_standard.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:187:24: error: only virtual member functions can be marked 'override' void runOnFunction() override; ^~~~~~~~ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:205:38: error: use of undeclared identifier 'getFunction' (void)applyPatternsAndFoldGreedily(getFunction(), std::move(patterns)); ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>' requested here return std::make_unique<LegalizeToStandardPass>(); ^ /home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeToStandardPass' virtual void runOnOperation() = 0; ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringSwitch.h:15: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/StringRef.h:12: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:229:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, const mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass &>' requested here return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this)); ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeToStandardPassBase<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::clonePass' requested here delete __ptr; ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass>::operator()' requested here get_deleter()(__ptr); ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_to_standard.cc:192:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeToStandardPass> >::~unique_ptr' requested here return std::make_unique<LegalizeToStandardPass>(); ^ 4 errors generated. [2907/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o FAILED: third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o /usr/bin/clang++ -DGTEST_HAS_RTTI=0 -D__STDC_CONSTANT_MACROS -D__STDC_FORMAT_MACROS -D__STDC_LIMIT_MACROS -I/home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree-build/third_party/mlir-hlo/lib/Dialect/mhlo/transforms -I/home/kkimj/iree/third_party/llvm-project/llvm/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/include -I/home/kkimj/iree/third_party/llvm-project/mlir/include -I/home/kkimj/iree-build/third_party/llvm-project/llvm/tools/mlir/include -I/home/kkimj/iree/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo/include -I/home/kkimj/iree-build/third_party/mlir-hlo -fPIC -fvisibility-inlines-hidden -Werror=date-time -Werror=unguarded-availability-new -w -fdiagnostics-color -ffunction-sections -fdata-sections -O2 -g -DNDEBUG -fPIC -fno-exceptions -fno-rtti -std=gnu++14 -MD -MT third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -MF third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o.d -o third_party/mlir-hlo/lib/Dialect/mhlo/transforms/CMakeFiles/obj.MhloToStandard.dir/legalize_control_flow.cc.o -c /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:43:24: error: only virtual member functions can be marked 'override' void runOnFunction() override; ^~~~~~~~ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:216:15: error: use of undeclared identifier 'getFunction' auto func = getFunction(); ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:15: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>' requested here return std::make_unique<LegalizeControlFlowPass>(); ^ /home/kkimj/iree/third_party/llvm-project/mlir/include/mlir/Pass/Pass.h:177:16: note: unimplemented pure virtual method 'runOnOperation' in 'LegalizeControlFlowPass' virtual void runOnOperation() = 0; ^ In file included from /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:18: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/STLExtras.h:19: In file included from /home/kkimj/iree/third_party/llvm-project/llvm/include/llvm/ADT/Optional.h:24: In file included from /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/memory:80: /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:821:34: error: allocating an object of abstract class type 'mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass' { return unique_ptr<_Tp>(new _Tp(std::forward<_Args>(__args)...)); } ^ /home/kkimj/iree/third_party/mlir-hlo/include/mlir-hlo/Dialect/mhlo/transforms/mhlo_passes.h.inc:99:17: note: in instantiation of function template specialization 'std::make_unique<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, const mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass &>' requested here return std::make_unique<DerivedT>(*static_cast<const DerivedT *>(this)); ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:78:2: note: in instantiation of member function 'mlir::mhlo::LegalizeControlFlowPassBase<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::clonePass' requested here delete __ptr; ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/7.5.0/../../../../include/c++/7.5.0/bits/unique_ptr.h:263:4: note: in instantiation of member function 'std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass>::operator()' requested here get_deleter()(__ptr); ^ /home/kkimj/iree/third_party/mlir-hlo/lib/Dialect/mhlo/transforms/legalize_control_flow.cc:237:10: note: in instantiation of member function 'std::unique_ptr<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass, std::default_delete<mlir::mhlo::(anonymous namespace)::LegalizeControlFlowPass> >::~unique_ptr' requested here return std::make_unique<LegalizeControlFlowPass>(); ^ 4 errors generated. [2910/4241] Building CXX object third_party/mlir-hlo/lib/Dialect/mhlo/IR/CMakeFiles/obj.MhloDialect.dir/hlo_ops.cc.o ninja: build stopped: subcommand failed. ```
infrastructure
cant build iree hello i have some problems and errors while building iree i just followed getting started to build could you give some hints or instructions for this problems thanks trials sudo apt get install clang lld sudo apt get install clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install python clang sudo apt install clang sudo apt update sudo apt upgrade sudo apt autoremove g to clang sudo update alternatives config c gcc to clang sudo update alternatives config cc export cc clang export cxx clang sudo rm r iree build cmake b iree build dcmake build type relwithdebinfo gninja cmake build iree build machine spec os ubuntu lts cpu intel r core tm cpu ram ssd nvme samsung verbose log results of checking cc cxx after failed with errors cmake build iree build re checking globbed directories building cxx object third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o failed third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o usr bin clang dgtest has rtti d stdc constant macros d stdc format macros d stdc limit macros i home kkimj iree third party mlir hlo lib dialect mhlo transforms i home kkimj iree build third party mlir hlo lib dialect mhlo transforms i home kkimj iree third party llvm project llvm include i home kkimj iree build third party llvm project llvm include i home kkimj iree third party llvm project mlir include i home kkimj iree build third party llvm project llvm tools mlir include i home kkimj iree third party mlir hlo include i home kkimj iree build third party mlir hlo include i home kkimj iree build third party mlir hlo fpic fvisibility inlines hidden werror date time werror unguarded availability new w fdiagnostics color ffunction sections fdata sections g dndebug fpic fno exceptions fno rtti std gnu md mt third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o mf third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o d o third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir mhlo control flow to scf cc o c home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc error only virtual member functions can be marked override void runonfunction override home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc error use of undeclared identifier getfunction getfunction walk whileop whileop matchandrewrite whileop in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc in file included from home kkimj iree third party llvm project llvm include llvm support casting h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace controlflowtoscfpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc note in instantiation of function template specialization std make unique requested here return std make unique home kkimj iree third party llvm project mlir include mlir pass pass h note unimplemented pure virtual method runonoperation in controlflowtoscfpass virtual void runonoperation in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc in file included from home kkimj iree third party llvm project llvm include llvm support casting h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace controlflowtoscfpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo include mlir hlo dialect mhlo transforms mhlo passes h inc note in instantiation of function template specialization std make unique requested here return std make unique static cast this usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function mlir mhlo legalizecontrolflowtoscfpassbase clonepass requested here delete ptr usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function std default delete operator requested here get deleter ptr home kkimj iree third party mlir hlo lib dialect mhlo transforms mhlo control flow to scf cc note in instantiation of member function std unique ptr unique ptr requested here return std make unique errors generated building cxx object third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o failed third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o usr bin clang dgtest has rtti d stdc constant macros d stdc format macros d stdc limit macros i home kkimj iree third party mlir hlo lib dialect mhlo transforms i home kkimj iree build third party mlir hlo lib dialect mhlo transforms i home kkimj iree third party llvm project llvm include i home kkimj iree build third party llvm project llvm include i home kkimj iree third party llvm project mlir include i home kkimj iree build third party llvm project llvm tools mlir include i home kkimj iree third party mlir hlo include i home kkimj iree build third party mlir hlo include i home kkimj iree build third party mlir hlo fpic fvisibility inlines hidden werror date time werror unguarded availability new w fdiagnostics color ffunction sections fdata sections g dndebug fpic fno exceptions fno rtti std gnu md mt third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o mf third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o d o third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize to standard cc o c home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc error only virtual member functions can be marked override void runonfunction override home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc error use of undeclared identifier getfunction void applypatternsandfoldgreedily getfunction std move patterns in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc in file included from home kkimj iree third party llvm project llvm include llvm adt stringswitch h in file included from home kkimj iree third party llvm project llvm include llvm adt stringref h in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizetostandardpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc note in instantiation of function template specialization std make unique requested here return std make unique home kkimj iree third party llvm project mlir include mlir pass pass h note unimplemented pure virtual method runonoperation in legalizetostandardpass virtual void runonoperation in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc in file included from home kkimj iree third party llvm project llvm include llvm adt stringswitch h in file included from home kkimj iree third party llvm project llvm include llvm adt stringref h in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizetostandardpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo include mlir hlo dialect mhlo transforms mhlo passes h inc note in instantiation of function template specialization std make unique requested here return std make unique static cast this usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function mlir mhlo legalizetostandardpassbase clonepass requested here delete ptr usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function std default delete operator requested here get deleter ptr home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize to standard cc note in instantiation of member function std unique ptr unique ptr requested here return std make unique errors generated building cxx object third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o failed third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o usr bin clang dgtest has rtti d stdc constant macros d stdc format macros d stdc limit macros i home kkimj iree third party mlir hlo lib dialect mhlo transforms i home kkimj iree build third party mlir hlo lib dialect mhlo transforms i home kkimj iree third party llvm project llvm include i home kkimj iree build third party llvm project llvm include i home kkimj iree third party llvm project mlir include i home kkimj iree build third party llvm project llvm tools mlir include i home kkimj iree third party mlir hlo include i home kkimj iree build third party mlir hlo include i home kkimj iree build third party mlir hlo fpic fvisibility inlines hidden werror date time werror unguarded availability new w fdiagnostics color ffunction sections fdata sections g dndebug fpic fno exceptions fno rtti std gnu md mt third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o mf third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o d o third party mlir hlo lib dialect mhlo transforms cmakefiles obj mhlotostandard dir legalize control flow cc o c home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc error only virtual member functions can be marked override void runonfunction override home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc error use of undeclared identifier getfunction auto func getfunction in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizecontrolflowpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc note in instantiation of function template specialization std make unique requested here return std make unique home kkimj iree third party llvm project mlir include mlir pass pass h note unimplemented pure virtual method runonoperation in legalizecontrolflowpass virtual void runonoperation in file included from home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc in file included from home kkimj iree third party llvm project llvm include llvm adt stlextras h in file included from home kkimj iree third party llvm project llvm include llvm adt optional h in file included from usr bin lib gcc linux gnu include c memory usr bin lib gcc linux gnu include c bits unique ptr h error allocating an object of abstract class type mlir mhlo anonymous namespace legalizecontrolflowpass return unique ptr new tp std forward args home kkimj iree third party mlir hlo include mlir hlo dialect mhlo transforms mhlo passes h inc note in instantiation of function template specialization std make unique requested here return std make unique static cast this usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function mlir mhlo legalizecontrolflowpassbase clonepass requested here delete ptr usr bin lib gcc linux gnu include c bits unique ptr h note in instantiation of member function std default delete operator requested here get deleter ptr home kkimj iree third party mlir hlo lib dialect mhlo transforms legalize control flow cc note in instantiation of member function std unique ptr unique ptr requested here return std make unique errors generated building cxx object third party mlir hlo lib dialect mhlo ir cmakefiles obj mhlodialect dir hlo ops cc o ninja build stopped subcommand failed
1
6,125
22,240,708,398
IssuesEvent
2022-06-09 04:46:08
uselagoon/lagoon
https://api.github.com/repos/uselagoon/lagoon
opened
Ability to have a different backup schedule for standby production environments
8-automation-helpers
We are seeing excess database load that is at least partially driven by the fact that both production environments are backing up the database every 6 hours. There is no business need for this particular customer to backup the standby database, only the active database. It would be handy to have the ability to define a backup schedule for active, and another for standby.
1.0
Ability to have a different backup schedule for standby production environments - We are seeing excess database load that is at least partially driven by the fact that both production environments are backing up the database every 6 hours. There is no business need for this particular customer to backup the standby database, only the active database. It would be handy to have the ability to define a backup schedule for active, and another for standby.
non_infrastructure
ability to have a different backup schedule for standby production environments we are seeing excess database load that is at least partially driven by the fact that both production environments are backing up the database every hours there is no business need for this particular customer to backup the standby database only the active database it would be handy to have the ability to define a backup schedule for active and another for standby
0
10,791
8,722,706,220
IssuesEvent
2018-12-09 15:09:09
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
NetCore2.1->2.2 Upgrade Issue; TargetFramework dropped/not recognized
area-Infrastructure
I'm trying to upgrade my web application from netCore 2.1 to 2.2. I have updated the TargetFramework to netcoreapp2.2 and I receive the following errors upon restore/build/clean: ``` MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) ``` I have also done a manual clean (deleting obj/bin directories). I **can** build if I change from using `<TargetFramework>` to `<TargetFrameworks>`, however I'd rather not do this and attempt to locate the underlying issue. I have two test projects in the same solution that reference the main project. I was able to migrate _them_ to netcoreapp2.2 without switching to TargetFramework**s**, I would like to do the same thing for the main web app. When I use TargetFramework (singular), Visual Studio updates the dependencies graph and removes the SDK node: ![image](https://user-images.githubusercontent.com/395615/49516771-63d52380-f868-11e8-9f5c-49ac2676585f.png) I've tried adding a global.json specifying the SDK version directly as 2.2.100 but that doesn't make any difference. Here is my .csproj file: ```xml <Project Sdk="Microsoft.NET.Sdk.Web"> <PropertyGroup> <AssemblyName>MyProduct.Web</AssemblyName> <OutputType>Exe</OutputType> <TargetFramework>netcoreapp2.2</TargetFramework> <PreserveCompilationContext>true</PreserveCompilationContext> <LangVersion>latest</LangVersion> </PropertyGroup> <PropertyGroup> <GenerateAssemblyInfo>false</GenerateAssemblyInfo> <PackageId>$(AssemblyName)</PackageId> <Authors>MyCompany</Authors> </PropertyGroup> <PropertyGroup Condition="'$(Configuration)' == 'Debug'"> <RazorCompileOnBuild>false</RazorCompileOnBuild> <RazorCompileOnPublish>false</RazorCompileOnPublish> </PropertyGroup> <PropertyGroup Condition="'$(Configuration)'=='Debug'"> <DefineConstants>DEBUG;TRACE</DefineConstants> </PropertyGroup> <ItemGroup> <None Update="wwwroot\**\*"> <CopyToPublishDirectory>PreserveNewest</CopyToPublishDirectory> </None> </ItemGroup> <ItemGroup> <ProjectReference Include="..\MyProduct.Web.Core\MyProduct.Web.Core.csproj" /> <ProjectReference Include="..\MyProduct.Web.Api.Contracts\MyProduct.Web.Api.Contracts.csproj" /> <ProjectReference Include="..\MyProduct.Web.ViewModels\MyProduct.Web.ViewModels.csproj" /> </ItemGroup> <ItemGroup> <PackageReference Include="Microsoft.AspNetCore.App" /> <PackageReference Include="BeginCollectionItemCore" Version="1.0.4" /> <PackageReference Include="Novell.Directory.Ldap.NETStandard" Version="2.3.5" /> <PackageReference Include="Microsoft.AspNetCore.Authentication.Cookies" Version="2.2.0" /> <PackageReference Include="Microsoft.AspNetCore.Authentication.JwtBearer" Version="2.2.0" /> </ItemGroup> <ItemGroup> <PackageReference Include="Microsoft.AspNetCore.Razor.Design" Version="2.2.0" PrivateAssets="All" /> <PackageReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Design" Version="2.2.0" PrivateAssets="All" /> <DotNetCliToolReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Tools" Version="2.0.4" /> </ItemGroup> <ItemGroup> <Content Update="Views\_ViewImports.cshtml"> <Pack>$(IncludeRazorContentInPack)</Pack> </Content> </ItemGroup> </Project> ```
1.0
NetCore2.1->2.2 Upgrade Issue; TargetFramework dropped/not recognized - I'm trying to upgrade my web application from netCore 2.1 to 2.2. I have updated the TargetFramework to netcoreapp2.2 and I receive the following errors upon restore/build/clean: ``` MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1). Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) MyProduct.Web.csproj : error NU1202: Package Microsoft.AspNetCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.AspNetCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) MyProduct.Web.csproj : error NU1202: Package Microsoft.NETCore.App 2.2.0 is not compatible with netcoreapp2.1 (.NETCoreApp,Version=v2.1) / win-x64. Package Microsoft.NETCore.App 2.2.0 supports: netcoreapp2.2 (.NETCoreApp,Version=v2.2) ``` I have also done a manual clean (deleting obj/bin directories). I **can** build if I change from using `<TargetFramework>` to `<TargetFrameworks>`, however I'd rather not do this and attempt to locate the underlying issue. I have two test projects in the same solution that reference the main project. I was able to migrate _them_ to netcoreapp2.2 without switching to TargetFramework**s**, I would like to do the same thing for the main web app. When I use TargetFramework (singular), Visual Studio updates the dependencies graph and removes the SDK node: ![image](https://user-images.githubusercontent.com/395615/49516771-63d52380-f868-11e8-9f5c-49ac2676585f.png) I've tried adding a global.json specifying the SDK version directly as 2.2.100 but that doesn't make any difference. Here is my .csproj file: ```xml <Project Sdk="Microsoft.NET.Sdk.Web"> <PropertyGroup> <AssemblyName>MyProduct.Web</AssemblyName> <OutputType>Exe</OutputType> <TargetFramework>netcoreapp2.2</TargetFramework> <PreserveCompilationContext>true</PreserveCompilationContext> <LangVersion>latest</LangVersion> </PropertyGroup> <PropertyGroup> <GenerateAssemblyInfo>false</GenerateAssemblyInfo> <PackageId>$(AssemblyName)</PackageId> <Authors>MyCompany</Authors> </PropertyGroup> <PropertyGroup Condition="'$(Configuration)' == 'Debug'"> <RazorCompileOnBuild>false</RazorCompileOnBuild> <RazorCompileOnPublish>false</RazorCompileOnPublish> </PropertyGroup> <PropertyGroup Condition="'$(Configuration)'=='Debug'"> <DefineConstants>DEBUG;TRACE</DefineConstants> </PropertyGroup> <ItemGroup> <None Update="wwwroot\**\*"> <CopyToPublishDirectory>PreserveNewest</CopyToPublishDirectory> </None> </ItemGroup> <ItemGroup> <ProjectReference Include="..\MyProduct.Web.Core\MyProduct.Web.Core.csproj" /> <ProjectReference Include="..\MyProduct.Web.Api.Contracts\MyProduct.Web.Api.Contracts.csproj" /> <ProjectReference Include="..\MyProduct.Web.ViewModels\MyProduct.Web.ViewModels.csproj" /> </ItemGroup> <ItemGroup> <PackageReference Include="Microsoft.AspNetCore.App" /> <PackageReference Include="BeginCollectionItemCore" Version="1.0.4" /> <PackageReference Include="Novell.Directory.Ldap.NETStandard" Version="2.3.5" /> <PackageReference Include="Microsoft.AspNetCore.Authentication.Cookies" Version="2.2.0" /> <PackageReference Include="Microsoft.AspNetCore.Authentication.JwtBearer" Version="2.2.0" /> </ItemGroup> <ItemGroup> <PackageReference Include="Microsoft.AspNetCore.Razor.Design" Version="2.2.0" PrivateAssets="All" /> <PackageReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Design" Version="2.2.0" PrivateAssets="All" /> <DotNetCliToolReference Include="Microsoft.VisualStudio.Web.CodeGeneration.Tools" Version="2.0.4" /> </ItemGroup> <ItemGroup> <Content Update="Views\_ViewImports.cshtml"> <Pack>$(IncludeRazorContentInPack)</Pack> </Content> </ItemGroup> </Project> ```
infrastructure
upgrade issue targetframework dropped not recognized i m trying to upgrade my web application from netcore to i have updated the targetframework to and i receive the following errors upon restore build clean myproduct web csproj error package microsoft aspnetcore app is not compatible with netcoreapp version package microsoft aspnetcore app supports netcoreapp version myproduct web csproj error package microsoft netcore app is not compatible with netcoreapp version package microsoft netcore app supports netcoreapp version myproduct web csproj error package microsoft aspnetcore app is not compatible with netcoreapp version win package microsoft aspnetcore app supports netcoreapp version myproduct web csproj error package microsoft netcore app is not compatible with netcoreapp version win package microsoft netcore app supports netcoreapp version i have also done a manual clean deleting obj bin directories i can build if i change from using to however i d rather not do this and attempt to locate the underlying issue i have two test projects in the same solution that reference the main project i was able to migrate them to without switching to targetframework s i would like to do the same thing for the main web app when i use targetframework singular visual studio updates the dependencies graph and removes the sdk node i ve tried adding a global json specifying the sdk version directly as but that doesn t make any difference here is my csproj file xml myproduct web exe true latest false assemblyname mycompany false false debug trace preservenewest includerazorcontentinpack
1
33,054
27,176,315,470
IssuesEvent
2023-02-18 02:54:47
NASA-IMPACT/csdap-cumulus
https://api.github.com/repos/NASA-IMPACT/csdap-cumulus
closed
Standup Cumulus development deployment in CBA sandbox account for Kris
good first issue infrastructure
Standup Cumulus 13.4.0 deployment to `csda-app-sbx-7894` CBA AWS account for Kris to use for development, by following the instructions in this repo's `README.md` file. Acceptance criteria: - [x] Successful Terraform deployment per the `README.md` instructions, with the exception of running a successful E2E test (i.e., all resources managed by Terraform are successfully created) - [x] Stretch: successfully run sample E2E discovery/ingestion test, per the last part of the instructions in the `README.md` file (if this cannot be completed in the same sprint as the item above, this issue should still be closed, and new issues for any problems blocking this criterion should be created and added to the subsequent sprint)
1.0
Standup Cumulus development deployment in CBA sandbox account for Kris - Standup Cumulus 13.4.0 deployment to `csda-app-sbx-7894` CBA AWS account for Kris to use for development, by following the instructions in this repo's `README.md` file. Acceptance criteria: - [x] Successful Terraform deployment per the `README.md` instructions, with the exception of running a successful E2E test (i.e., all resources managed by Terraform are successfully created) - [x] Stretch: successfully run sample E2E discovery/ingestion test, per the last part of the instructions in the `README.md` file (if this cannot be completed in the same sprint as the item above, this issue should still be closed, and new issues for any problems blocking this criterion should be created and added to the subsequent sprint)
infrastructure
standup cumulus development deployment in cba sandbox account for kris standup cumulus deployment to csda app sbx cba aws account for kris to use for development by following the instructions in this repo s readme md file acceptance criteria successful terraform deployment per the readme md instructions with the exception of running a successful test i e all resources managed by terraform are successfully created stretch successfully run sample discovery ingestion test per the last part of the instructions in the readme md file if this cannot be completed in the same sprint as the item above this issue should still be closed and new issues for any problems blocking this criterion should be created and added to the subsequent sprint
1
237,712
7,763,480,595
IssuesEvent
2018-06-01 16:42:31
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
opened
Sandcastle gallery missing scrollbar at certain sizes
priority - high
![image](https://user-images.githubusercontent.com/1495196/40852643-193ab342-6599-11e8-841b-4001631510a8.png) I assume this is the fact that the labels wrapped to a second line and it's pushing everything down (and we have overflow hidden so the scrollbar is underneath everything). @emackey any ideas for a quick fix?
1.0
Sandcastle gallery missing scrollbar at certain sizes - ![image](https://user-images.githubusercontent.com/1495196/40852643-193ab342-6599-11e8-841b-4001631510a8.png) I assume this is the fact that the labels wrapped to a second line and it's pushing everything down (and we have overflow hidden so the scrollbar is underneath everything). @emackey any ideas for a quick fix?
non_infrastructure
sandcastle gallery missing scrollbar at certain sizes i assume this is the fact that the labels wrapped to a second line and it s pushing everything down and we have overflow hidden so the scrollbar is underneath everything emackey any ideas for a quick fix
0
33,772
27,801,037,986
IssuesEvent
2023-03-17 15:51:26
pixiebrix/pixiebrix-extension
https://api.github.com/repos/pixiebrix/pixiebrix-extension
opened
Rainforest: Create a new "PixieBrix Extension" Site and Environments for builds of the extension
infrastructure developer experience
This is slice 1 of the Test-Independent Rainforest project, see full implementation sketch here: https://www.notion.so/pixiebrix/Test-Independent-Rainforest-Environments-67a031a2298547e2835dbccf5d35a5a8?pvs=4#091a70e60fc94aae8eed123e5a329041 In a nutshell: - Add a new "PixieBrix Extension" Site to Rainforest - Add 3 new Environments to the "PixieBrix Extension" site - one for `main`, `release` and `cws` builds of the extension - Modify `upload_extension.sh` to build & upload the extension with consistent filenames that will be overridden on s3 - Modify our rainforest github workflow to run tests without a custom start url - Update existing tests to use the "PixieBrix Extension" site and start at the Environment `start_url`
1.0
Rainforest: Create a new "PixieBrix Extension" Site and Environments for builds of the extension - This is slice 1 of the Test-Independent Rainforest project, see full implementation sketch here: https://www.notion.so/pixiebrix/Test-Independent-Rainforest-Environments-67a031a2298547e2835dbccf5d35a5a8?pvs=4#091a70e60fc94aae8eed123e5a329041 In a nutshell: - Add a new "PixieBrix Extension" Site to Rainforest - Add 3 new Environments to the "PixieBrix Extension" site - one for `main`, `release` and `cws` builds of the extension - Modify `upload_extension.sh` to build & upload the extension with consistent filenames that will be overridden on s3 - Modify our rainforest github workflow to run tests without a custom start url - Update existing tests to use the "PixieBrix Extension" site and start at the Environment `start_url`
infrastructure
rainforest create a new pixiebrix extension site and environments for builds of the extension this is slice of the test independent rainforest project see full implementation sketch here in a nutshell add a new pixiebrix extension site to rainforest add new environments to the pixiebrix extension site one for main release and cws builds of the extension modify upload extension sh to build upload the extension with consistent filenames that will be overridden on modify our rainforest github workflow to run tests without a custom start url update existing tests to use the pixiebrix extension site and start at the environment start url
1
6,884
6,654,524,516
IssuesEvent
2017-09-29 13:11:19
Teemperor/ClangModulesCMake
https://api.github.com/repos/Teemperor/ClangModulesCMake
closed
Test libraries in a user-specified order, not in the order of the include directories.
infrastructure
E.g. we should always test/mount STL before third-party libraries. Currently we do it in whatever way the includes are coming in.
1.0
Test libraries in a user-specified order, not in the order of the include directories. - E.g. we should always test/mount STL before third-party libraries. Currently we do it in whatever way the includes are coming in.
infrastructure
test libraries in a user specified order not in the order of the include directories e g we should always test mount stl before third party libraries currently we do it in whatever way the includes are coming in
1
18,255
12,875,880,953
IssuesEvent
2020-07-11 01:14:35
wix/yoshi
https://api.github.com/repos/wix/yoshi
closed
indeterministic error in legacy tests: port 3200 is in used
💤 Stale 🛠 Infrastructure
- non-production issue # 🐛 Bug Report run on your local machine (not in CI), `yarn run test:legacy:test` at the root of the project. I'm sorry that i don't have logs or anything else to share. it sometimes happen. ### 🔦 Context <!-- optional --> while debugging/working on the project locally. ### 💁 Possible Solution <!-- optional --> a quick fix is to manually close the process that the port belongs to. I didn't spend the time to investigate what is the goal/responsibilities of the service with the taken port so i don't have a good permanent solution. but if it helps, these are the ways i'm fixing this problem in my projects: - if it's a child process: try to set the port to `0` and extract all the listening ports from the child-process (https://www.npmjs.com/package/find-process) - if it's on the main process (test-process): if it's possible in your scenario, create a mock-server that will be able to give you the port. ### 🌍 Your Environment yoshi: `v4.61.2` node: `v12.16.1`
1.0
indeterministic error in legacy tests: port 3200 is in used - - non-production issue # 🐛 Bug Report run on your local machine (not in CI), `yarn run test:legacy:test` at the root of the project. I'm sorry that i don't have logs or anything else to share. it sometimes happen. ### 🔦 Context <!-- optional --> while debugging/working on the project locally. ### 💁 Possible Solution <!-- optional --> a quick fix is to manually close the process that the port belongs to. I didn't spend the time to investigate what is the goal/responsibilities of the service with the taken port so i don't have a good permanent solution. but if it helps, these are the ways i'm fixing this problem in my projects: - if it's a child process: try to set the port to `0` and extract all the listening ports from the child-process (https://www.npmjs.com/package/find-process) - if it's on the main process (test-process): if it's possible in your scenario, create a mock-server that will be able to give you the port. ### 🌍 Your Environment yoshi: `v4.61.2` node: `v12.16.1`
infrastructure
indeterministic error in legacy tests port is in used non production issue 🐛 bug report run on your local machine not in ci yarn run test legacy test at the root of the project i m sorry that i don t have logs or anything else to share it sometimes happen 🔦 context while debugging working on the project locally 💁 possible solution a quick fix is to manually close the process that the port belongs to i didn t spend the time to investigate what is the goal responsibilities of the service with the taken port so i don t have a good permanent solution but if it helps these are the ways i m fixing this problem in my projects if it s a child process try to set the port to and extract all the listening ports from the child process if it s on the main process test process if it s possible in your scenario create a mock server that will be able to give you the port 🌍 your environment yoshi node
1
15,230
11,423,462,439
IssuesEvent
2020-02-03 15:56:58
SteeltoeOSS/steeltoe
https://api.github.com/repos/SteeltoeOSS/steeltoe
closed
CI Builds should pass for ALL PRs
Infrastructure
Due to the way GemFire is currently distributed, the CI build process has to use secrets, which results in failures on all CI builds from non-team members. PR builds should be made to pass for _all_ PRs, not just core Steeltoe team members
1.0
CI Builds should pass for ALL PRs - Due to the way GemFire is currently distributed, the CI build process has to use secrets, which results in failures on all CI builds from non-team members. PR builds should be made to pass for _all_ PRs, not just core Steeltoe team members
infrastructure
ci builds should pass for all prs due to the way gemfire is currently distributed the ci build process has to use secrets which results in failures on all ci builds from non team members pr builds should be made to pass for all prs not just core steeltoe team members
1
852
2,964,247,202
IssuesEvent
2015-07-10 15:35:33
RIOT-OS/RIOT
https://api.github.com/repos/RIOT-OS/RIOT
opened
Proposal for directory structure for automated tests
CI-Infrastructure Request for Comments tests
Define a directory under `$RIOTBASE/tests/` and outside `tests/unittests` where sources for automated tests can be placed. It has turned out in discussions [#3296](https://github.com/RIOT-OS/RIOT/pull/3296#issuecomment-118338031), [#3307](https://github.com/RIOT-OS/RIOT/pull/3307#discussion_r33860255) that the established *unittests* becomes insufficent for extended testing. Motivation ---------- The agreement of a seperate directory structure should face the following difficulties. ### tests/unittests ### The tests in this directory are build together in one blob. This is implicating some weaknesses. + With the increasing number of test cases the memory ressources of some boards are exhausted. + It prevents the desired isolation of the running test cases. + Necessary control of preconditions can not be archieved for test sequences that would require reinitializing (e.g `ng_ipv6_init()`) ### tests ### The directory is, beside of *unittests* home of several more or less freely named subdirectories for different test scenarios. It is not adviseable to mix the automated tests inbetween. + Increases naming collisions or encourages creative naming to prevent them. + Makes it difficult to keep track of locations. Proposal -------- The idea is to name name subdirectory of tests(lets say *'automated'*) and put the tests in subdirectories. These may reflect the location of either the headers or the units source files to follow the "principle of least suprise". Every *unit* or *test object* has further subdirs for sepeated test which require reflashing, etc... ### Example for resulting structure ### (Reflecting loctions of C files) tests/automated/core/atomic/ tests/automated/core/... tests/automated/core/thread/ tests/automated/... tests/automated/sys/auto_init/auto_init/ tests/automated/sys/... tests/automated/sys/net/application_layer/ng_zep/ng_zep/ tests/automated/sys/net/... tests/automated/sys/net/transport_layer/ng_udp/ng_udp/
1.0
Proposal for directory structure for automated tests - Define a directory under `$RIOTBASE/tests/` and outside `tests/unittests` where sources for automated tests can be placed. It has turned out in discussions [#3296](https://github.com/RIOT-OS/RIOT/pull/3296#issuecomment-118338031), [#3307](https://github.com/RIOT-OS/RIOT/pull/3307#discussion_r33860255) that the established *unittests* becomes insufficent for extended testing. Motivation ---------- The agreement of a seperate directory structure should face the following difficulties. ### tests/unittests ### The tests in this directory are build together in one blob. This is implicating some weaknesses. + With the increasing number of test cases the memory ressources of some boards are exhausted. + It prevents the desired isolation of the running test cases. + Necessary control of preconditions can not be archieved for test sequences that would require reinitializing (e.g `ng_ipv6_init()`) ### tests ### The directory is, beside of *unittests* home of several more or less freely named subdirectories for different test scenarios. It is not adviseable to mix the automated tests inbetween. + Increases naming collisions or encourages creative naming to prevent them. + Makes it difficult to keep track of locations. Proposal -------- The idea is to name name subdirectory of tests(lets say *'automated'*) and put the tests in subdirectories. These may reflect the location of either the headers or the units source files to follow the "principle of least suprise". Every *unit* or *test object* has further subdirs for sepeated test which require reflashing, etc... ### Example for resulting structure ### (Reflecting loctions of C files) tests/automated/core/atomic/ tests/automated/core/... tests/automated/core/thread/ tests/automated/... tests/automated/sys/auto_init/auto_init/ tests/automated/sys/... tests/automated/sys/net/application_layer/ng_zep/ng_zep/ tests/automated/sys/net/... tests/automated/sys/net/transport_layer/ng_udp/ng_udp/
infrastructure
proposal for directory structure for automated tests define a directory under riotbase tests and outside tests unittests where sources for automated tests can be placed it has turned out in discussions that the established unittests becomes insufficent for extended testing motivation the agreement of a seperate directory structure should face the following difficulties tests unittests the tests in this directory are build together in one blob this is implicating some weaknesses with the increasing number of test cases the memory ressources of some boards are exhausted it prevents the desired isolation of the running test cases necessary control of preconditions can not be archieved for test sequences that would require reinitializing e g ng init tests the directory is beside of unittests home of several more or less freely named subdirectories for different test scenarios it is not adviseable to mix the automated tests inbetween increases naming collisions or encourages creative naming to prevent them makes it difficult to keep track of locations proposal the idea is to name name subdirectory of tests lets say automated and put the tests in subdirectories these may reflect the location of either the headers or the units source files to follow the principle of least suprise every unit or test object has further subdirs for sepeated test which require reflashing etc example for resulting structure reflecting loctions of c files tests automated core atomic tests automated core tests automated core thread tests automated tests automated sys auto init auto init tests automated sys tests automated sys net application layer ng zep ng zep tests automated sys net tests automated sys net transport layer ng udp ng udp
1
90,956
26,227,724,748
IssuesEvent
2023-01-04 20:23:54
chaotic-aur/packages
https://api.github.com/repos/chaotic-aur/packages
closed
[Request] KCM modules
request:new-pkg interfere required priority:low bug:PKGBUILD
### Link to the package(s) in the AUR https://aur.archlinux.org/packages/systemd-kcm https://aur.archlinux.org/packages/kcm-polkit-kde-git https://aur.archlinux.org/packages/fancontrol-kcm https://aur.archlinux.org/packages/kcm-uefi https://aur.archlinux.org/packages/kcm-grub2-git https://aur.archlinux.org/packages/kcm-colorful-git https://aur.archlinux.org/packages/fancontrol-kcm-git https://aur.archlinux.org/packages/jackman-kcm-git ### Utility this package has for you Managing more setting at the same place. ### Do you consider the package(s) to be useful for every Chaotic-AUR user? No, but for a few. ### Do you consider the package to be useful for feature testing/preview? - [ ] Yes ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### Does the package's license allow redistributing it? YES! ### Have you searched the issues to ensure this request is unique? - [X] YES! ### Have you read the README to ensure this package is not banned? - [X] YES! ### More information _No response_
1.0
[Request] KCM modules - ### Link to the package(s) in the AUR https://aur.archlinux.org/packages/systemd-kcm https://aur.archlinux.org/packages/kcm-polkit-kde-git https://aur.archlinux.org/packages/fancontrol-kcm https://aur.archlinux.org/packages/kcm-uefi https://aur.archlinux.org/packages/kcm-grub2-git https://aur.archlinux.org/packages/kcm-colorful-git https://aur.archlinux.org/packages/fancontrol-kcm-git https://aur.archlinux.org/packages/jackman-kcm-git ### Utility this package has for you Managing more setting at the same place. ### Do you consider the package(s) to be useful for every Chaotic-AUR user? No, but for a few. ### Do you consider the package to be useful for feature testing/preview? - [ ] Yes ### Have you tested if the package builds in a clean chroot? - [ ] Yes ### Does the package's license allow redistributing it? YES! ### Have you searched the issues to ensure this request is unique? - [X] YES! ### Have you read the README to ensure this package is not banned? - [X] YES! ### More information _No response_
non_infrastructure
kcm modules link to the package s in the aur utility this package has for you managing more setting at the same place do you consider the package s to be useful for every chaotic aur user no but for a few do you consider the package to be useful for feature testing preview yes have you tested if the package builds in a clean chroot yes does the package s license allow redistributing it yes have you searched the issues to ensure this request is unique yes have you read the readme to ensure this package is not banned yes more information no response
0
417,298
12,158,198,798
IssuesEvent
2020-04-26 02:29:06
rubyforgood/casa
https://api.github.com/repos/rubyforgood/casa
closed
Remove `other` from case_contact contact_type dropdown
:raised_hands: Volunteer Priority: Medium Status: Available help wanted
Part of epic #3 (volunteer dashboard) `other` should be removed from the dropdown menu under `contact_type` when creating a new `case_contact` <a href="https://casa-r4g-staging.herokuapp.com/case_contacts/new">see here for reference on staging</a>
1.0
Remove `other` from case_contact contact_type dropdown - Part of epic #3 (volunteer dashboard) `other` should be removed from the dropdown menu under `contact_type` when creating a new `case_contact` <a href="https://casa-r4g-staging.herokuapp.com/case_contacts/new">see here for reference on staging</a>
non_infrastructure
remove other from case contact contact type dropdown part of epic volunteer dashboard other should be removed from the dropdown menu under contact type when creating a new case contact
0
108,578
23,631,659,708
IssuesEvent
2022-08-25 09:49:52
arduino/arduino-ide
https://api.github.com/repos/arduino/arduino-ide
closed
Only the IP address is shown for unrecognized network boards
topic: code type: imperfection topic: CLI
### Describe the problem This is the follow-up of #1327. Arduino IDE 2.x shows only the IP address of connected but unrecognized boards. Before #1327, `network` boards not recognized by the Arduino CLI were not listed in the "Board Selector" menu. See the changes [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-fb37f20bea00881acee3aafddb1ecefcecf41ce59845ca1510da79e918ee0837R332-R342) and [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-e42c82bb67e277cfa4598239952afd65db44dba55dc7d68df619dfccfa648279R409-R423). #1327 will remove the above-mentioned filtering, and the Arduino IDE will show all `network` boards. Unfortunately, the IDE receives insufficient information from the CLI, so the IDE has no chance to show anything else but the IP. <img width="313" alt="Screen Shot 2022-08-17 at 14 41 58" src="https://user-images.githubusercontent.com/1405703/185129514-5d1d6c10-7bcf-4530-b42e-4420738b8979.png"> <img width="637" alt="Screen Shot 2022-08-17 at 14 41 40" src="https://user-images.githubusercontent.com/1405703/185129587-c3a22d16-1e2d-47c5-936f-834c0ce53f56.png"> <img width="281" alt="Screen Shot 2022-08-17 at 14 41 23" src="https://user-images.githubusercontent.com/1405703/185129661-a4a7a903-5e92-431c-af7b-b9b1e210b872.png"> The `mdns-discovery` is aware of these additional things; hence the Arduino IDE 1.x can enhance the label of an unrecognized `network` board: ``` HELLO "ciao" 1 { "eventType": "hello", "message": "Invalid HELLO command", "error": true } HELLO 1 "ciao" { "eventType": "hello", "message": "OK", "protocolVersion": 1 } START_SYNC { "eventType": "start_sync", "message": "OK" } { "eventType": "add", "port": { "address": "10.0.1.25", "label": "esp32s2 at 10.0.1.25", "protocol": "network", "protocolLabel": "Network Port", "properties": { ".": "esp32s2usb", "auth_upload": "yes", "board": "esp32s2usb", "hostname": "esp32s2.local.", "port": "3232", "ssh_upload": "no", "tcp_check": "no" } } } ``` The same information is missing from the Arduino CLI board watch (`board list -w --format json`): ``` { "type": "add", "address": "/dev/cu.usbmodem03", "protocol": "serial", "protocol_label": "Serial Port (USB)" } { "type": "add", "address": "10.0.1.25", "protocol": "network", "protocol_label": "Network Port" } ``` So the IDE fails to show `esp32s2 at 10.0.1.25` and falls back to `Unknown 10.0.1.25`. ### To reproduce The simplest way is to try it with Arduino CLI: ``` ./arduino-cli board list -w --format json ``` ### Expected behavior Arduino IDE should receive the properties from the CLI discovered by the `mdns-discovery`. ```json { "properties": { ".": "esp32s2usb", "auth_upload": "yes", "board": "esp32s2usb", "hostname": "esp32s2.local.", "port": "3232", "ssh_upload": "no", "tcp_check": "no" } } ``` Or the IDE should receive the `"label": "esp32s2 at 10.0.1.25",` from the CLI. ### Arduino IDE version 9e2b73a ### Operating system macOS ### Operating system version 12.3.1 ### Additional context Related (if not the same): https://github.com/arduino/arduino-ide/issues/567 ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds) - [X] My report contains all necessary details
1.0
Only the IP address is shown for unrecognized network boards - ### Describe the problem This is the follow-up of #1327. Arduino IDE 2.x shows only the IP address of connected but unrecognized boards. Before #1327, `network` boards not recognized by the Arduino CLI were not listed in the "Board Selector" menu. See the changes [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-fb37f20bea00881acee3aafddb1ecefcecf41ce59845ca1510da79e918ee0837R332-R342) and [here](https://github.com/arduino/arduino-ide/commit/74bfdc4c56d7a1577a4e800a378c21b82c1da5f8#diff-e42c82bb67e277cfa4598239952afd65db44dba55dc7d68df619dfccfa648279R409-R423). #1327 will remove the above-mentioned filtering, and the Arduino IDE will show all `network` boards. Unfortunately, the IDE receives insufficient information from the CLI, so the IDE has no chance to show anything else but the IP. <img width="313" alt="Screen Shot 2022-08-17 at 14 41 58" src="https://user-images.githubusercontent.com/1405703/185129514-5d1d6c10-7bcf-4530-b42e-4420738b8979.png"> <img width="637" alt="Screen Shot 2022-08-17 at 14 41 40" src="https://user-images.githubusercontent.com/1405703/185129587-c3a22d16-1e2d-47c5-936f-834c0ce53f56.png"> <img width="281" alt="Screen Shot 2022-08-17 at 14 41 23" src="https://user-images.githubusercontent.com/1405703/185129661-a4a7a903-5e92-431c-af7b-b9b1e210b872.png"> The `mdns-discovery` is aware of these additional things; hence the Arduino IDE 1.x can enhance the label of an unrecognized `network` board: ``` HELLO "ciao" 1 { "eventType": "hello", "message": "Invalid HELLO command", "error": true } HELLO 1 "ciao" { "eventType": "hello", "message": "OK", "protocolVersion": 1 } START_SYNC { "eventType": "start_sync", "message": "OK" } { "eventType": "add", "port": { "address": "10.0.1.25", "label": "esp32s2 at 10.0.1.25", "protocol": "network", "protocolLabel": "Network Port", "properties": { ".": "esp32s2usb", "auth_upload": "yes", "board": "esp32s2usb", "hostname": "esp32s2.local.", "port": "3232", "ssh_upload": "no", "tcp_check": "no" } } } ``` The same information is missing from the Arduino CLI board watch (`board list -w --format json`): ``` { "type": "add", "address": "/dev/cu.usbmodem03", "protocol": "serial", "protocol_label": "Serial Port (USB)" } { "type": "add", "address": "10.0.1.25", "protocol": "network", "protocol_label": "Network Port" } ``` So the IDE fails to show `esp32s2 at 10.0.1.25` and falls back to `Unknown 10.0.1.25`. ### To reproduce The simplest way is to try it with Arduino CLI: ``` ./arduino-cli board list -w --format json ``` ### Expected behavior Arduino IDE should receive the properties from the CLI discovered by the `mdns-discovery`. ```json { "properties": { ".": "esp32s2usb", "auth_upload": "yes", "board": "esp32s2usb", "hostname": "esp32s2.local.", "port": "3232", "ssh_upload": "no", "tcp_check": "no" } } ``` Or the IDE should receive the `"label": "esp32s2 at 10.0.1.25",` from the CLI. ### Arduino IDE version 9e2b73a ### Operating system macOS ### Operating system version 12.3.1 ### Additional context Related (if not the same): https://github.com/arduino/arduino-ide/issues/567 ### Issue checklist - [X] I searched for previous reports in [the issue tracker](https://github.com/arduino/arduino-ide/issues?q=) - [X] I verified the problem still occurs when using the latest [nightly build](https://github.com/arduino/arduino-ide#nightly-builds) - [X] My report contains all necessary details
non_infrastructure
only the ip address is shown for unrecognized network boards describe the problem this is the follow up of arduino ide x shows only the ip address of connected but unrecognized boards before network boards not recognized by the arduino cli were not listed in the board selector menu see the changes and will remove the above mentioned filtering and the arduino ide will show all network boards unfortunately the ide receives insufficient information from the cli so the ide has no chance to show anything else but the ip img width alt screen shot at src img width alt screen shot at src img width alt screen shot at src the mdns discovery is aware of these additional things hence the arduino ide x can enhance the label of an unrecognized network board hello ciao eventtype hello message invalid hello command error true hello ciao eventtype hello message ok protocolversion start sync eventtype start sync message ok eventtype add port address label at protocol network protocollabel network port properties auth upload yes board hostname local port ssh upload no tcp check no the same information is missing from the arduino cli board watch board list w format json type add address dev cu protocol serial protocol label serial port usb type add address protocol network protocol label network port so the ide fails to show at and falls back to unknown to reproduce the simplest way is to try it with arduino cli arduino cli board list w format json expected behavior arduino ide should receive the properties from the cli discovered by the mdns discovery json properties auth upload yes board hostname local port ssh upload no tcp check no or the ide should receive the label at from the cli arduino ide version operating system macos operating system version additional context related if not the same issue checklist i searched for previous reports in i verified the problem still occurs when using the latest my report contains all necessary details
0
431,144
12,475,699,316
IssuesEvent
2020-05-29 12:08:05
borgbase/vorta
https://api.github.com/repos/borgbase/vorta
closed
Pass through borg error message for all commands
priority:medium type:enhancement
**How to reproduce:** * Remove your rights on your borg cache files (this stems from running borg with sudo) * Try to add existing repo with vorta **Result:** An error message is displayed extremely short but is immediately replaced with a generic error message. Here a popup of the original log would be helpful. (Example: "Borg cli failed and output the following: ...") **Workaround:** Look inside the log-file, here you can clearly see the error
1.0
Pass through borg error message for all commands - **How to reproduce:** * Remove your rights on your borg cache files (this stems from running borg with sudo) * Try to add existing repo with vorta **Result:** An error message is displayed extremely short but is immediately replaced with a generic error message. Here a popup of the original log would be helpful. (Example: "Borg cli failed and output the following: ...") **Workaround:** Look inside the log-file, here you can clearly see the error
non_infrastructure
pass through borg error message for all commands how to reproduce remove your rights on your borg cache files this stems from running borg with sudo try to add existing repo with vorta result an error message is displayed extremely short but is immediately replaced with a generic error message here a popup of the original log would be helpful example borg cli failed and output the following workaround look inside the log file here you can clearly see the error
0
302,434
26,143,864,441
IssuesEvent
2022-12-29 23:37:18
fuatakgun/eufy_security
https://api.github.com/repos/fuatakgun/eufy_security
closed
Front Door Lock not updating
entity: lock type: Bug next step: waiting for response next step: retest with alpha
When I unlock it, it unlocks but doesnt update in home assistant.
1.0
Front Door Lock not updating - When I unlock it, it unlocks but doesnt update in home assistant.
non_infrastructure
front door lock not updating when i unlock it it unlocks but doesnt update in home assistant
0
34,504
30,030,538,341
IssuesEvent
2023-06-27 09:17:45
SonarSource/sonar-scanner-msbuild
https://api.github.com/repos/SonarSource/sonar-scanner-msbuild
closed
Reuse mend.yml template for downloading and running mend
Type: Improvement Area: Infrastructure
There is a template that already handles the download and the execution of mend. We should reuse it to avoid duplication and be able to do changes across repositories. It can be found here: https://dev.azure.com/sonarsource/DotNetTeam%20Project/_git/pipelines-yaml-templates?path=/mend.yml&version=GTv2.0.0 Example of usage: https://github.com/SonarSource/sonarlint-omnisharp/blob/7ce958a943c2bbe6fe38136a4529360e26292bd1/azure-pipelines.yml#L93-L102 We will have to update the `Mend-Scan.ps1` script or maybe inline it in the main yaml file.
1.0
Reuse mend.yml template for downloading and running mend - There is a template that already handles the download and the execution of mend. We should reuse it to avoid duplication and be able to do changes across repositories. It can be found here: https://dev.azure.com/sonarsource/DotNetTeam%20Project/_git/pipelines-yaml-templates?path=/mend.yml&version=GTv2.0.0 Example of usage: https://github.com/SonarSource/sonarlint-omnisharp/blob/7ce958a943c2bbe6fe38136a4529360e26292bd1/azure-pipelines.yml#L93-L102 We will have to update the `Mend-Scan.ps1` script or maybe inline it in the main yaml file.
infrastructure
reuse mend yml template for downloading and running mend there is a template that already handles the download and the execution of mend we should reuse it to avoid duplication and be able to do changes across repositories it can be found here example of usage we will have to update the mend scan script or maybe inline it in the main yaml file
1
282,531
30,889,357,958
IssuesEvent
2023-08-04 02:36:20
madhans23/linux-4.1.15
https://api.github.com/repos/madhans23/linux-4.1.15
reopened
CVE-2017-7482 (High) detected in linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2017-7482 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before version 4.12, Kerberos 5 tickets decoded when using the RXRPC keys incorrectly assumes the size of a field. This could lead to the size-remaining variable wrapping and the data pointer going over the end of the buffer. This could possibly lead to memory corruption and possible privilege escalation. <p>Publish Date: 2018-07-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7482>CVE-2017-7482</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-7482">https://nvd.nist.gov/vuln/detail/CVE-2017-7482</a></p> <p>Release Date: 2018-07-30</p> <p>Fix Resolution: 4.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-7482 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2017-7482 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/rxrpc/ar-key.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In the Linux kernel before version 4.12, Kerberos 5 tickets decoded when using the RXRPC keys incorrectly assumes the size of a field. This could lead to the size-remaining variable wrapping and the data pointer going over the end of the buffer. This could possibly lead to memory corruption and possible privilege escalation. <p>Publish Date: 2018-07-30 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7482>CVE-2017-7482</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-7482">https://nvd.nist.gov/vuln/detail/CVE-2017-7482</a></p> <p>Release Date: 2018-07-30</p> <p>Fix Resolution: 4.12</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in base branch master vulnerable source files net rxrpc ar key c net rxrpc ar key c vulnerability details in the linux kernel before version kerberos tickets decoded when using the rxrpc keys incorrectly assumes the size of a field this could lead to the size remaining variable wrapping and the data pointer going over the end of the buffer this could possibly lead to memory corruption and possible privilege escalation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
1,681
3,332,737,864
IssuesEvent
2015-11-11 21:31:08
codeforamerica/communities
https://api.github.com/repos/codeforamerica/communities
closed
Meetup Pro
brigade donation infrastructure network growth
Meetup Pro looks like it will have everything we want. The main requests we had were: * API access to member growth over time. * Better attendance or email data The costs looks to be about $10k a year. Its worth it. Here is part of an email from them. ``` I also wanted to send over the list of data points we're building out, which will be available through our API within the next couple of months. Let me know if there are any data points you'd be interested to see that aren't included on the list! You'll be able to view the following information about your MUGs: - name - description - category - start date - total members - average age - gender ratio - topics - # past events - # past RSVPs - avg. # RSVPs/event - # repeat RSVPers - date of last scheduled MUP - date of next scheduled MUP - date joined Meetup Pro ```
1.0
Meetup Pro - Meetup Pro looks like it will have everything we want. The main requests we had were: * API access to member growth over time. * Better attendance or email data The costs looks to be about $10k a year. Its worth it. Here is part of an email from them. ``` I also wanted to send over the list of data points we're building out, which will be available through our API within the next couple of months. Let me know if there are any data points you'd be interested to see that aren't included on the list! You'll be able to view the following information about your MUGs: - name - description - category - start date - total members - average age - gender ratio - topics - # past events - # past RSVPs - avg. # RSVPs/event - # repeat RSVPers - date of last scheduled MUP - date of next scheduled MUP - date joined Meetup Pro ```
infrastructure
meetup pro meetup pro looks like it will have everything we want the main requests we had were api access to member growth over time better attendance or email data the costs looks to be about a year its worth it here is part of an email from them i also wanted to send over the list of data points we re building out which will be available through our api within the next couple of months let me know if there are any data points you d be interested to see that aren t included on the list you ll be able to view the following information about your mugs name description category start date total members average age gender ratio topics past events past rsvps avg rsvps event repeat rsvpers date of last scheduled mup date of next scheduled mup date joined meetup pro
1
1,058
3,024,884,396
IssuesEvent
2015-08-03 01:53:47
catapult-project/catapult
https://api.github.com/repos/catapult-project/catapult
opened
python tests fail on windows
Infrastructure
<a href="https://github.com/anniesullie"><img src="https://avatars.githubusercontent.com/u/4331931?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [anniesullie](https://github.com/anniesullie)** _Friday Jul 10, 2015 at 20:06 GMT_ _Originally opened as https://github.com/google/trace-viewer/issues/1093_ ---- After https://codereview.appspot.com/251100043/, there are still 3 failing tests. I'm not sure how to handle c:\ getting stuck on the absolute paths in Windows. Output: ====================================================================== ERROR: testInlineStylesheetURLs (tvcm.html_module_unittest.HTMLModuleTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module_unittest.py", line 359, in testInlineStylesheetURLs my_component.AppendDirectlyDependentFilenamesTo(computed_deps) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 89, in AppendDirectlyDependentFilenamesTo self.loader, module_dirname, contents) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__ self._Load(containing_dirname) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load self._images = [resolve_url(x) for x in matches] File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url image = self.loader.LoadImage(abs_path) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage raise module.DepsException("url('%s') did not exist" % abs_path) DepsException: While loading: a.b.my_component Got: url('c:\tmp\a\something.jpg') did not exist ====================================================================== ERROR: testGetAllDependentFilenamesRecursive (tvcm.module_unittest.ModuleIntegrationTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module_unittest.py", line 118, in testGetAllDependentFilenamesRecursive my_module = loader.LoadModule(module_name='z.foo') File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 140, in LoadModule m.Load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 38, in Load super(HTMLModule, self).Load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module.py", line 187, in Load style_sheet = self.loader.LoadStyleSheet(name) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet style_sheet.load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load self._InitParsedStyleSheetIfNeeded() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded self.loader, module_dirname, self.contents) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__ self._Load(containing_dirname) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load self._images = [resolve_url(x) for x in matches] File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url image = self.loader.LoadImage(abs_path) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage raise module.DepsException("url('%s') did not exist" % abs_path) DepsException: While loading: z.foo z.foo.css Got: url('c:\x\y\z\foo.jpeg') did not exist ====================================================================== ERROR: testImages (tvcm.style_sheet_unittest.StyleSheetUnittest) ---------------------------------------------------------------------- Traceback (most recent call last): File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet_unittest.py", line 29, in testImages foo_x = loader.LoadStyleSheet('foo.x') File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet style_sheet.load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load self._InitParsedStyleSheetIfNeeded() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded self.loader, module_dirname, self.contents) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__ self._Load(containing_dirname) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load self._images = [resolve_url(x) for x in matches] File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url image = self.loader.LoadImage(abs_path) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage raise module.DepsException("url('%s') did not exist" % abs_path) DepsException: While loading: foo.x.css Got: url('c:\src\images\bar.jpeg') did not exist ---------------------------------------------------------------------- Ran 87 tests in 11.067s FAILED (errors=3, skipped=22)
1.0
python tests fail on windows - <a href="https://github.com/anniesullie"><img src="https://avatars.githubusercontent.com/u/4331931?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [anniesullie](https://github.com/anniesullie)** _Friday Jul 10, 2015 at 20:06 GMT_ _Originally opened as https://github.com/google/trace-viewer/issues/1093_ ---- After https://codereview.appspot.com/251100043/, there are still 3 failing tests. I'm not sure how to handle c:\ getting stuck on the absolute paths in Windows. Output: ====================================================================== ERROR: testInlineStylesheetURLs (tvcm.html_module_unittest.HTMLModuleTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module_unittest.py", line 359, in testInlineStylesheetURLs my_component.AppendDirectlyDependentFilenamesTo(computed_deps) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 89, in AppendDirectlyDependentFilenamesTo self.loader, module_dirname, contents) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__ self._Load(containing_dirname) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load self._images = [resolve_url(x) for x in matches] File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url image = self.loader.LoadImage(abs_path) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage raise module.DepsException("url('%s') did not exist" % abs_path) DepsException: While loading: a.b.my_component Got: url('c:\tmp\a\something.jpg') did not exist ====================================================================== ERROR: testGetAllDependentFilenamesRecursive (tvcm.module_unittest.ModuleIntegrationTests) ---------------------------------------------------------------------- Traceback (most recent call last): File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module_unittest.py", line 118, in testGetAllDependentFilenamesRecursive my_module = loader.LoadModule(module_name='z.foo') File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 140, in LoadModule m.Load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\html_module.py", line 38, in Load super(HTMLModule, self).Load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\module.py", line 187, in Load style_sheet = self.loader.LoadStyleSheet(name) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet style_sheet.load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load self._InitParsedStyleSheetIfNeeded() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded self.loader, module_dirname, self.contents) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__ self._Load(containing_dirname) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load self._images = [resolve_url(x) for x in matches] File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url image = self.loader.LoadImage(abs_path) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage raise module.DepsException("url('%s') did not exist" % abs_path) DepsException: While loading: z.foo z.foo.css Got: url('c:\x\y\z\foo.jpeg') did not exist ====================================================================== ERROR: testImages (tvcm.style_sheet_unittest.StyleSheetUnittest) ---------------------------------------------------------------------- Traceback (most recent call last): File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet_unittest.py", line 29, in testImages foo_x = loader.LoadStyleSheet('foo.x') File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 177, in LoadStyleSheet style_sheet.load() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 129, in load self._InitParsedStyleSheetIfNeeded() File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 136, in _InitParsedStyleSheetIfNeeded self.loader, module_dirname, self.contents) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 34, in __init__ self._Load(containing_dirname) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 88, in _Load self._images = [resolve_url(x) for x in matches] File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\style_sheet.py", line 84, in resolve_url image = self.loader.LoadImage(abs_path) File "c:\github\trace-viewer\tracing\third_party\tvcm\tvcm\resource_loader.py", line 186, in LoadImage raise module.DepsException("url('%s') did not exist" % abs_path) DepsException: While loading: foo.x.css Got: url('c:\src\images\bar.jpeg') did not exist ---------------------------------------------------------------------- Ran 87 tests in 11.067s FAILED (errors=3, skipped=22)
infrastructure
python tests fail on windows issue by friday jul at gmt originally opened as after there are still failing tests i m not sure how to handle c getting stuck on the absolute paths in windows output error testinlinestylesheeturls tvcm html module unittest htmlmoduletests traceback most recent call last file c github trace viewer tracing third party tvcm tvcm html module unittest py line in testinlinestylesheeturls my component appenddirectlydependentfilenamesto computed deps file c github trace viewer tracing third party tvcm tvcm html module py line in appenddirectlydependentfilenamesto self loader module dirname contents file c github trace viewer tracing third party tvcm tvcm style sheet py line in init self load containing dirname file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self images file c github trace viewer tracing third party tvcm tvcm style sheet py line in resolve url image self loader loadimage abs path file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadimage raise module depsexception url s did not exist abs path depsexception while loading a b my component got url c tmp a something jpg did not exist error testgetalldependentfilenamesrecursive tvcm module unittest moduleintegrationtests traceback most recent call last file c github trace viewer tracing third party tvcm tvcm module unittest py line in testgetalldependentfilenamesrecursive my module loader loadmodule module name z foo file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadmodule m load file c github trace viewer tracing third party tvcm tvcm html module py line in load super htmlmodule self load file c github trace viewer tracing third party tvcm tvcm module py line in load style sheet self loader loadstylesheet name file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadstylesheet style sheet load file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self initparsedstylesheetifneeded file c github trace viewer tracing third party tvcm tvcm style sheet py line in initparsedstylesheetifneeded self loader module dirname self contents file c github trace viewer tracing third party tvcm tvcm style sheet py line in init self load containing dirname file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self images file c github trace viewer tracing third party tvcm tvcm style sheet py line in resolve url image self loader loadimage abs path file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadimage raise module depsexception url s did not exist abs path depsexception while loading z foo z foo css got url c x y z foo jpeg did not exist error testimages tvcm style sheet unittest stylesheetunittest traceback most recent call last file c github trace viewer tracing third party tvcm tvcm style sheet unittest py line in testimages foo x loader loadstylesheet foo x file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadstylesheet style sheet load file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self initparsedstylesheetifneeded file c github trace viewer tracing third party tvcm tvcm style sheet py line in initparsedstylesheetifneeded self loader module dirname self contents file c github trace viewer tracing third party tvcm tvcm style sheet py line in init self load containing dirname file c github trace viewer tracing third party tvcm tvcm style sheet py line in load self images file c github trace viewer tracing third party tvcm tvcm style sheet py line in resolve url image self loader loadimage abs path file c github trace viewer tracing third party tvcm tvcm resource loader py line in loadimage raise module depsexception url s did not exist abs path depsexception while loading foo x css got url c src images bar jpeg did not exist ran tests in failed errors skipped
1
18,370
12,960,368,299
IssuesEvent
2020-07-20 14:16:33
libero/reviewer
https://api.github.com/repos/libero/reviewer
closed
Integrate MECA to chart and umbrella tests
Infrastructure
Chart and staging env: - [x] ingress in chart for /meca-result - [x] add sftp and bucket envvar to chart - [x] add sftp secret to staging deploy (ejp test env) Umbrella repo: - [x] add sftp container so that submission can handle when 'submit' button is clicked in browsertests - [x] add envvars for meca stuff in umbrella compose file(s)
1.0
Integrate MECA to chart and umbrella tests - Chart and staging env: - [x] ingress in chart for /meca-result - [x] add sftp and bucket envvar to chart - [x] add sftp secret to staging deploy (ejp test env) Umbrella repo: - [x] add sftp container so that submission can handle when 'submit' button is clicked in browsertests - [x] add envvars for meca stuff in umbrella compose file(s)
infrastructure
integrate meca to chart and umbrella tests chart and staging env ingress in chart for meca result add sftp and bucket envvar to chart add sftp secret to staging deploy ejp test env umbrella repo add sftp container so that submission can handle when submit button is clicked in browsertests add envvars for meca stuff in umbrella compose file s
1
91,701
18,676,614,623
IssuesEvent
2021-10-31 17:11:33
CiviWiki/OpenCiviWiki
https://api.github.com/repos/CiviWiki/OpenCiviWiki
closed
[BUG]: Major bug fixes through the UI
bug community code quality documentation triage
### Description There are multiple routing issues and bugs which I am spotting through testing, this is a self-assigned issue and I'll elaborate on the issues I resolve here soon. [Update 1]: The following bugs have been spotted: - [x] Reverse name maps for urls - [x] Profile doesn't load from the header after login - [x] Profile creation keeps loading infinitely - [x] HTTP 403 for POST on profile setup - [ ] The footer needs to be added to the bottom
1.0
[BUG]: Major bug fixes through the UI - ### Description There are multiple routing issues and bugs which I am spotting through testing, this is a self-assigned issue and I'll elaborate on the issues I resolve here soon. [Update 1]: The following bugs have been spotted: - [x] Reverse name maps for urls - [x] Profile doesn't load from the header after login - [x] Profile creation keeps loading infinitely - [x] HTTP 403 for POST on profile setup - [ ] The footer needs to be added to the bottom
non_infrastructure
major bug fixes through the ui description there are multiple routing issues and bugs which i am spotting through testing this is a self assigned issue and i ll elaborate on the issues i resolve here soon the following bugs have been spotted reverse name maps for urls profile doesn t load from the header after login profile creation keeps loading infinitely http for post on profile setup the footer needs to be added to the bottom
0
18,047
4,230,354,727
IssuesEvent
2016-07-04 11:41:54
centreon/centreon
https://api.github.com/repos/centreon/centreon
closed
Temps de disponibilité
Kind/Documentation
Bonjour, J'ai quelque chose qui me parait bizarre au niveau des rapports sur le temps de disponibilité. J'ai un groupe de 3 hôtes, j'ai eu une coupure sur l'un de ces 3 hôtes, le temps sur cette journée apparait comme ceci pour le groupe : ![centreon_bug_time](https://cloud.githubusercontent.com/assets/2530686/14736838/4b1d223e-0879-11e6-91d8-8f03fd4d421d.jpg) Cependant quand je vais sur mon hôte qui a eu la coupure le temps apparait comme ceci sur la même journée : ![centreon_bug_time2](https://cloud.githubusercontent.com/assets/2530686/14736867/6baf7c68-0879-11e6-80f8-629369c9a3f4.jpg) J'ai l'impression que le temps est divisé, ça ressemble à un bug ? Si je ne me trompe pas, on aimerait en théorie avoir le temps d'indisponibilité du groupe, en additionnant les temps des hôtes ? Cordialement, Bastien
1.0
Temps de disponibilité - Bonjour, J'ai quelque chose qui me parait bizarre au niveau des rapports sur le temps de disponibilité. J'ai un groupe de 3 hôtes, j'ai eu une coupure sur l'un de ces 3 hôtes, le temps sur cette journée apparait comme ceci pour le groupe : ![centreon_bug_time](https://cloud.githubusercontent.com/assets/2530686/14736838/4b1d223e-0879-11e6-91d8-8f03fd4d421d.jpg) Cependant quand je vais sur mon hôte qui a eu la coupure le temps apparait comme ceci sur la même journée : ![centreon_bug_time2](https://cloud.githubusercontent.com/assets/2530686/14736867/6baf7c68-0879-11e6-80f8-629369c9a3f4.jpg) J'ai l'impression que le temps est divisé, ça ressemble à un bug ? Si je ne me trompe pas, on aimerait en théorie avoir le temps d'indisponibilité du groupe, en additionnant les temps des hôtes ? Cordialement, Bastien
non_infrastructure
temps de disponibilité bonjour j ai quelque chose qui me parait bizarre au niveau des rapports sur le temps de disponibilité j ai un groupe de hôtes j ai eu une coupure sur l un de ces hôtes le temps sur cette journée apparait comme ceci pour le groupe cependant quand je vais sur mon hôte qui a eu la coupure le temps apparait comme ceci sur la même journée j ai l impression que le temps est divisé ça ressemble à un bug si je ne me trompe pas on aimerait en théorie avoir le temps d indisponibilité du groupe en additionnant les temps des hôtes cordialement bastien
0
701
9,961,019,916
IssuesEvent
2019-07-06 22:45:13
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
HTTP2: Pending bytes from a previous frame may not be flushed if another pending write is canceled
area-System.Net.Http tenet-reliability
Consider two frames both trying to acquire the write lock concurrently. One of them gets it, the other has to wait and registers itself as a pending writer. The first frame then doesn't require itself to be flushed if there are pending writers, and so writes itself to the buffer but doesn't explicitly flush. Between the time that it checks for pending writers and does the write, however, the other frame is canceled, and the pending writers count is decremented back to 0. There's now data in the buffer that needed to be flushed, but there's no pending write that's going to do it. We could "easily" address this by only counting a non-cancelable write as one that can be pending, but from HttpClient every request will use a cancelable token, and thus all header/data frame writes associated with that request will be cancelable. A more robust solution is probably to queue a forced flush in the case where a pending waiter is canceled. cc: @geoffkizer
True
HTTP2: Pending bytes from a previous frame may not be flushed if another pending write is canceled - Consider two frames both trying to acquire the write lock concurrently. One of them gets it, the other has to wait and registers itself as a pending writer. The first frame then doesn't require itself to be flushed if there are pending writers, and so writes itself to the buffer but doesn't explicitly flush. Between the time that it checks for pending writers and does the write, however, the other frame is canceled, and the pending writers count is decremented back to 0. There's now data in the buffer that needed to be flushed, but there's no pending write that's going to do it. We could "easily" address this by only counting a non-cancelable write as one that can be pending, but from HttpClient every request will use a cancelable token, and thus all header/data frame writes associated with that request will be cancelable. A more robust solution is probably to queue a forced flush in the case where a pending waiter is canceled. cc: @geoffkizer
non_infrastructure
pending bytes from a previous frame may not be flushed if another pending write is canceled consider two frames both trying to acquire the write lock concurrently one of them gets it the other has to wait and registers itself as a pending writer the first frame then doesn t require itself to be flushed if there are pending writers and so writes itself to the buffer but doesn t explicitly flush between the time that it checks for pending writers and does the write however the other frame is canceled and the pending writers count is decremented back to there s now data in the buffer that needed to be flushed but there s no pending write that s going to do it we could easily address this by only counting a non cancelable write as one that can be pending but from httpclient every request will use a cancelable token and thus all header data frame writes associated with that request will be cancelable a more robust solution is probably to queue a forced flush in the case where a pending waiter is canceled cc geoffkizer
0
2,204
3,560,165,307
IssuesEvent
2016-01-23 00:09:04
Unidata/siphon
https://api.github.com/repos/Unidata/siphon
closed
nose -> py.test
Infrastructure
Nose is dying (https://nose.readthedocs.org/en/latest/index.html) and py.test seems to be what all the cool kids are using. Shouldn't take much to move over, but will be easier to do sooner rather than later.
1.0
nose -> py.test - Nose is dying (https://nose.readthedocs.org/en/latest/index.html) and py.test seems to be what all the cool kids are using. Shouldn't take much to move over, but will be easier to do sooner rather than later.
infrastructure
nose py test nose is dying and py test seems to be what all the cool kids are using shouldn t take much to move over but will be easier to do sooner rather than later
1
11,736
9,403,696,453
IssuesEvent
2019-04-09 02:35:30
hotosm/tasking-manager
https://api.github.com/repos/hotosm/tasking-manager
reopened
Host javascript resources ourselves
Difficulty: Easy Priority: High Status: Needs help Type: Infrastructure
Feature request from #886 discussion: host javascript resources ourselves instead of CDN.
1.0
Host javascript resources ourselves - Feature request from #886 discussion: host javascript resources ourselves instead of CDN.
infrastructure
host javascript resources ourselves feature request from discussion host javascript resources ourselves instead of cdn
1
55,728
14,020,206,752
IssuesEvent
2020-10-29 19:19:24
srivatsamarichi/ContosoAir
https://api.github.com/repos/srivatsamarichi/ContosoAir
opened
CVE-2018-19827 (High) detected in node-sass-4.14.1.tgz, opennmsopennms-source-25.1.0-1
security vulnerability
## CVE-2018-19827 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b>, <b>opennmsopennms-source-25.1.0-1</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: ContosoAir/package.json</p> <p>Path to vulnerable library: ContosoAir/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/ContosoAir/commit/7e3d160bd69713f60688f97955fd688a3fe91b8f">7e3d160bd69713f60688f97955fd688a3fe91b8f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a use-after-free vulnerability exists in the SharedPtr class in SharedPtr.cpp (or SharedPtr.hpp) that may cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19827>CVE-2018-19827</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/pull/2784">https://github.com/sass/libsass/pull/2784</a></p> <p>Release Date: 2019-08-29</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-19827 (High) detected in node-sass-4.14.1.tgz, opennmsopennms-source-25.1.0-1 - ## CVE-2018-19827 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.14.1.tgz</b>, <b>opennmsopennms-source-25.1.0-1</b></p></summary> <p> <details><summary><b>node-sass-4.14.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.14.1.tgz</a></p> <p>Path to dependency file: ContosoAir/package.json</p> <p>Path to vulnerable library: ContosoAir/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.14.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/ContosoAir/commit/7e3d160bd69713f60688f97955fd688a3fe91b8f">7e3d160bd69713f60688f97955fd688a3fe91b8f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a use-after-free vulnerability exists in the SharedPtr class in SharedPtr.cpp (or SharedPtr.hpp) that may cause a denial of service (application crash) or possibly have unspecified other impact. <p>Publish Date: 2018-12-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-19827>CVE-2018-19827</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/sass/libsass/pull/2784">https://github.com/sass/libsass/pull/2784</a></p> <p>Release Date: 2019-08-29</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in node sass tgz opennmsopennms source cve high severity vulnerability vulnerable libraries node sass tgz opennmsopennms source node sass tgz wrapper around libsass library home page a href path to dependency file contosoair package json path to vulnerable library contosoair node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href found in base branch master vulnerability details in libsass a use after free vulnerability exists in the sharedptr class in sharedptr cpp or sharedptr hpp that may cause a denial of service application crash or possibly have unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
5,064
5,415,194,997
IssuesEvent
2017-03-01 20:58:31
Flatline4/Flatline4
https://api.github.com/repos/Flatline4/Flatline4
closed
Set up TSLint
infrastructure
It would be nice to have a linter in place to enforce certain rules are followed in the code base. The rules will need to be configured to our liking, and we will need to go through the existing code and make sure these rules are followed. Also, it would be nice to add running the linter as an npm script and make it run as part of the production build (the production build should not run if TSLint fails).
1.0
Set up TSLint - It would be nice to have a linter in place to enforce certain rules are followed in the code base. The rules will need to be configured to our liking, and we will need to go through the existing code and make sure these rules are followed. Also, it would be nice to add running the linter as an npm script and make it run as part of the production build (the production build should not run if TSLint fails).
infrastructure
set up tslint it would be nice to have a linter in place to enforce certain rules are followed in the code base the rules will need to be configured to our liking and we will need to go through the existing code and make sure these rules are followed also it would be nice to add running the linter as an npm script and make it run as part of the production build the production build should not run if tslint fails
1
26,742
20,623,075,392
IssuesEvent
2022-03-07 19:25:42
arduino/arduino-ide
https://api.github.com/repos/arduino/arduino-ide
closed
Redundant test build version indicators
topic: infrastructure type: imperfection
## Describe the bug Test builds are generated for every PR creation and push as well as pushes to the `main` branch. These make it easy for anyone to contribute to the project through beta testing. It is important that the specific test build in use be easily identifiable. The Git commit hash is used for this purpose. For example, the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9ecff86bbefe3a34549e0e546d95ee1ff3439f97 is `2.0.0-rc3-snapshot-9ecff86` 🐛 The version indicator is now added twice to the test build versions. For example, the the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9b1f15def893f1f3df58e459e66a87c3fcdb930f is `2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d` ## To Reproduce 1. Open the list of `push` or `pull_request` event-triggered runs of the `Arduino IDE` workflow: https://github.com/arduino/arduino-ide/actions/workflows/build.yml?query=event%3Apull_request+event%3Apush+is%3Asuccess 1. Click on any recent item in the list of runs. For example: https://github.com/arduino/arduino-ide/actions/runs/1858133816 1. Under the "**Artifacts**" section of the page, click on the download link of the artifact for your operating system. 1. Wait for the download to finish. 1. Extract or install the downloaded file. 🐛 The extracted folder contains redundant version indicators (e.g., `arduino-ide_2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d_Windows_64bit`) 1. Start the Arduino IDE test build. 🐛 The window title contains redundant version indicators (e.g., "**sketch_feb17a | Arduino IDE 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d**") 1. Select **Help > About Arduino IDE** from the Arduino IDE menus. 1. Click the <kbd>Copy</kbd> button. 1. Paste the contents of the clipboard into any text editor. 🐛 The pasted text contains redundant version indicators. For example: ```text Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d Date: 2022-02-17T10:40:50.003Z CLI Version: 0.21.0 [10107d24] Copyright © 2022 Arduino SA ``` ## Expected behavior The tester builds don't have redundant version indicators (e.g., `2.0.0-rc4-snapshot.9b1f15d`). ## Desktop - OS: Windows 10, Ubuntu 20.04 - Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d Date: 2022-02-17T10:40:50.003Z CLI Version: 0.21.0 [10107d24] ## Additional context The change was introduced at https://github.com/arduino/arduino-ide/commit/f660058c7550aaeeab81a6bc59f060d1ca2b901a There was a previous change from the `snapshot.<hash>` (e.g., `2.0.0-rc3-snapshot.69ac1f4`) to `snapshot-<hash>` format (e.g., `2.0.0-rc3-snapshot-112153f`), which was introduced in https://github.com/arduino/arduino-ide/commit/112153fb965f63d952d126c8244cd3f84f0a1a1b --- The additional characters use up some of [the precious allowance of path length doled out by Microsoft](https://github.com/arduino/arduino-ide/issues/106) for the beta testers on Windows. --- Seeing this identifier in the filename redundantly, but with two different formats (note the `.` in one and `-` in the other) makes me nervous that it is an indication of something fragile.
1.0
Redundant test build version indicators - ## Describe the bug Test builds are generated for every PR creation and push as well as pushes to the `main` branch. These make it easy for anyone to contribute to the project through beta testing. It is important that the specific test build in use be easily identifiable. The Git commit hash is used for this purpose. For example, the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9ecff86bbefe3a34549e0e546d95ee1ff3439f97 is `2.0.0-rc3-snapshot-9ecff86` 🐛 The version indicator is now added twice to the test build versions. For example, the the version assigned to this commit: https://github.com/arduino/arduino-ide/commit/9b1f15def893f1f3df58e459e66a87c3fcdb930f is `2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d` ## To Reproduce 1. Open the list of `push` or `pull_request` event-triggered runs of the `Arduino IDE` workflow: https://github.com/arduino/arduino-ide/actions/workflows/build.yml?query=event%3Apull_request+event%3Apush+is%3Asuccess 1. Click on any recent item in the list of runs. For example: https://github.com/arduino/arduino-ide/actions/runs/1858133816 1. Under the "**Artifacts**" section of the page, click on the download link of the artifact for your operating system. 1. Wait for the download to finish. 1. Extract or install the downloaded file. 🐛 The extracted folder contains redundant version indicators (e.g., `arduino-ide_2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d_Windows_64bit`) 1. Start the Arduino IDE test build. 🐛 The window title contains redundant version indicators (e.g., "**sketch_feb17a | Arduino IDE 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d**") 1. Select **Help > About Arduino IDE** from the Arduino IDE menus. 1. Click the <kbd>Copy</kbd> button. 1. Paste the contents of the clipboard into any text editor. 🐛 The pasted text contains redundant version indicators. For example: ```text Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d Date: 2022-02-17T10:40:50.003Z CLI Version: 0.21.0 [10107d24] Copyright © 2022 Arduino SA ``` ## Expected behavior The tester builds don't have redundant version indicators (e.g., `2.0.0-rc4-snapshot.9b1f15d`). ## Desktop - OS: Windows 10, Ubuntu 20.04 - Version: 2.0.0-rc4-snapshot.9b1f15d-snapshot-9b1f15d Date: 2022-02-17T10:40:50.003Z CLI Version: 0.21.0 [10107d24] ## Additional context The change was introduced at https://github.com/arduino/arduino-ide/commit/f660058c7550aaeeab81a6bc59f060d1ca2b901a There was a previous change from the `snapshot.<hash>` (e.g., `2.0.0-rc3-snapshot.69ac1f4`) to `snapshot-<hash>` format (e.g., `2.0.0-rc3-snapshot-112153f`), which was introduced in https://github.com/arduino/arduino-ide/commit/112153fb965f63d952d126c8244cd3f84f0a1a1b --- The additional characters use up some of [the precious allowance of path length doled out by Microsoft](https://github.com/arduino/arduino-ide/issues/106) for the beta testers on Windows. --- Seeing this identifier in the filename redundantly, but with two different formats (note the `.` in one and `-` in the other) makes me nervous that it is an indication of something fragile.
infrastructure
redundant test build version indicators describe the bug test builds are generated for every pr creation and push as well as pushes to the main branch these make it easy for anyone to contribute to the project through beta testing it is important that the specific test build in use be easily identifiable the git commit hash is used for this purpose for example the version assigned to this commit is snapshot 🐛 the version indicator is now added twice to the test build versions for example the the version assigned to this commit is snapshot snapshot to reproduce open the list of push or pull request event triggered runs of the arduino ide workflow click on any recent item in the list of runs for example under the artifacts section of the page click on the download link of the artifact for your operating system wait for the download to finish extract or install the downloaded file 🐛 the extracted folder contains redundant version indicators e g arduino ide snapshot snapshot windows start the arduino ide test build 🐛 the window title contains redundant version indicators e g sketch arduino ide snapshot snapshot select help about arduino ide from the arduino ide menus click the copy button paste the contents of the clipboard into any text editor 🐛 the pasted text contains redundant version indicators for example text version snapshot snapshot date cli version copyright © arduino sa expected behavior the tester builds don t have redundant version indicators e g snapshot desktop os windows ubuntu version snapshot snapshot date cli version additional context the change was introduced at there was a previous change from the snapshot e g snapshot to snapshot format e g snapshot which was introduced in the additional characters use up some of for the beta testers on windows seeing this identifier in the filename redundantly but with two different formats note the in one and in the other makes me nervous that it is an indication of something fragile
1
386,499
11,440,086,363
IssuesEvent
2020-02-05 08:56:33
cheminfo/nmr-displayer
https://api.github.com/repos/cheminfo/nmr-displayer
closed
simplify data struture
Priority
Currently in the JSON and internally we have data1d and data2d. ![image](https://user-images.githubusercontent.com/1484241/73596399-2dee0680-4522-11ea-9325-cef88e984bcf.png) This makes things quite complex in order to display the list of spectra but also in order to parse jcamp that may contain 1D and 2D (in the same jcamp). We should have only ONE array containing all the data called `spectra` @hamed-musallam Please could you do this refactoring and the change of all the examples ? (actually it will just be replace 'data1d' by 'spectra' I will then improve the loading of spectra to deal with 'linked' jcamp.
1.0
simplify data struture - Currently in the JSON and internally we have data1d and data2d. ![image](https://user-images.githubusercontent.com/1484241/73596399-2dee0680-4522-11ea-9325-cef88e984bcf.png) This makes things quite complex in order to display the list of spectra but also in order to parse jcamp that may contain 1D and 2D (in the same jcamp). We should have only ONE array containing all the data called `spectra` @hamed-musallam Please could you do this refactoring and the change of all the examples ? (actually it will just be replace 'data1d' by 'spectra' I will then improve the loading of spectra to deal with 'linked' jcamp.
non_infrastructure
simplify data struture currently in the json and internally we have and this makes things quite complex in order to display the list of spectra but also in order to parse jcamp that may contain and in the same jcamp we should have only one array containing all the data called spectra hamed musallam please could you do this refactoring and the change of all the examples actually it will just be replace by spectra i will then improve the loading of spectra to deal with linked jcamp
0
359,106
10,660,190,992
IssuesEvent
2019-10-18 09:24:09
KamilSJaron/genomic-features-of-asexual-animals
https://api.github.com/repos/KamilSJaron/genomic-features-of-asexual-animals
closed
heterozygosity structure section
PRIORITY TODO help wanted
This section is a real pain in my butt. I find it less and less trustworthy, but at the same time, it's one of the very novel sections that is at least trying something new. The original idea was to 1. get the heterozygosity structure of everything, look at those known (Meloidogyne hybrids; rotifers etc) and guess something about those that are unknown We first had a problem with estimates in rotifers, as they were very off the reality. This is solved now thanks to @reubwn. The second problem was with controversy regarding the genome structure of root-knot nematodes (one of the reviewers did not like that we took for granted the 'one diverged hybrid copy present'). That does not affect the visualization, but it makes the interpretation of the plot harder. However, with a fixed scale for rotifers, the others became unreadable - all squashed together. <img width="851" alt="Screenshot 2019-10-10 at 11 00 30" src="https://user-images.githubusercontent.com/8181573/66559366-31154f00-eb4d-11e9-966b-d2670c8d9c3e.png"> Then we figured, that one needs to focus on relative sizes between different heterozygosity topologies anyway, so we divided each bar by it's size, getting <img width="522" alt="Screenshot 2019-10-10 at 11 03 28" src="https://user-images.githubusercontent.com/8181573/66559641-9ec17b00-eb4d-11e9-97c2-b0f984de2510.png"> First I thought it's perfect, showing clearly both structure in triploids and tetraploids, while the absolute values are already shown in Figure 2. But then I realized that even the relative counts are absolute value dependent. We made the argument, that more equidistant genomic copies are in triploids, higher fraction of ABC loci there will be. However, even this expectation is dependent on the absolute value. I.e. how low heterozygosity organism, there will be very small fraction of ABC loci even if the genomic copies will be equidistant, contrasting to highly heterozygous species where even small divergence between the two more closely related haplotypes cause a relatively high proportion of ABC loci. I made some naive calculations for our species and it really does matter. So, back to the original plot. However, now to make it more readable for triploids, I made it a two-panel plot for triploid and tetraploids. What do you think? ![Figure_3_tetraploid_heterozygosity_two_plots](https://user-images.githubusercontent.com/8181573/66560051-7dad5a00-eb4e-11e9-9504-018ad019cc6a.png) Aesthetics can be polished, the question is if the different scale axes are confusing or not. To make it more intuitive we can scale triploids from 0 to 10, so they appear visually smaller or something like that. Just to improve intuitive reading of the plot. This is linked to, I am not sure anymore how to phrase this section. AGAIN! Suggestions welcomed.
1.0
heterozygosity structure section - This section is a real pain in my butt. I find it less and less trustworthy, but at the same time, it's one of the very novel sections that is at least trying something new. The original idea was to 1. get the heterozygosity structure of everything, look at those known (Meloidogyne hybrids; rotifers etc) and guess something about those that are unknown We first had a problem with estimates in rotifers, as they were very off the reality. This is solved now thanks to @reubwn. The second problem was with controversy regarding the genome structure of root-knot nematodes (one of the reviewers did not like that we took for granted the 'one diverged hybrid copy present'). That does not affect the visualization, but it makes the interpretation of the plot harder. However, with a fixed scale for rotifers, the others became unreadable - all squashed together. <img width="851" alt="Screenshot 2019-10-10 at 11 00 30" src="https://user-images.githubusercontent.com/8181573/66559366-31154f00-eb4d-11e9-966b-d2670c8d9c3e.png"> Then we figured, that one needs to focus on relative sizes between different heterozygosity topologies anyway, so we divided each bar by it's size, getting <img width="522" alt="Screenshot 2019-10-10 at 11 03 28" src="https://user-images.githubusercontent.com/8181573/66559641-9ec17b00-eb4d-11e9-97c2-b0f984de2510.png"> First I thought it's perfect, showing clearly both structure in triploids and tetraploids, while the absolute values are already shown in Figure 2. But then I realized that even the relative counts are absolute value dependent. We made the argument, that more equidistant genomic copies are in triploids, higher fraction of ABC loci there will be. However, even this expectation is dependent on the absolute value. I.e. how low heterozygosity organism, there will be very small fraction of ABC loci even if the genomic copies will be equidistant, contrasting to highly heterozygous species where even small divergence between the two more closely related haplotypes cause a relatively high proportion of ABC loci. I made some naive calculations for our species and it really does matter. So, back to the original plot. However, now to make it more readable for triploids, I made it a two-panel plot for triploid and tetraploids. What do you think? ![Figure_3_tetraploid_heterozygosity_two_plots](https://user-images.githubusercontent.com/8181573/66560051-7dad5a00-eb4e-11e9-9504-018ad019cc6a.png) Aesthetics can be polished, the question is if the different scale axes are confusing or not. To make it more intuitive we can scale triploids from 0 to 10, so they appear visually smaller or something like that. Just to improve intuitive reading of the plot. This is linked to, I am not sure anymore how to phrase this section. AGAIN! Suggestions welcomed.
non_infrastructure
heterozygosity structure section this section is a real pain in my butt i find it less and less trustworthy but at the same time it s one of the very novel sections that is at least trying something new the original idea was to get the heterozygosity structure of everything look at those known meloidogyne hybrids rotifers etc and guess something about those that are unknown we first had a problem with estimates in rotifers as they were very off the reality this is solved now thanks to reubwn the second problem was with controversy regarding the genome structure of root knot nematodes one of the reviewers did not like that we took for granted the one diverged hybrid copy present that does not affect the visualization but it makes the interpretation of the plot harder however with a fixed scale for rotifers the others became unreadable all squashed together img width alt screenshot at src then we figured that one needs to focus on relative sizes between different heterozygosity topologies anyway so we divided each bar by it s size getting img width alt screenshot at src first i thought it s perfect showing clearly both structure in triploids and tetraploids while the absolute values are already shown in figure but then i realized that even the relative counts are absolute value dependent we made the argument that more equidistant genomic copies are in triploids higher fraction of abc loci there will be however even this expectation is dependent on the absolute value i e how low heterozygosity organism there will be very small fraction of abc loci even if the genomic copies will be equidistant contrasting to highly heterozygous species where even small divergence between the two more closely related haplotypes cause a relatively high proportion of abc loci i made some naive calculations for our species and it really does matter so back to the original plot however now to make it more readable for triploids i made it a two panel plot for triploid and tetraploids what do you think aesthetics can be polished the question is if the different scale axes are confusing or not to make it more intuitive we can scale triploids from to so they appear visually smaller or something like that just to improve intuitive reading of the plot this is linked to i am not sure anymore how to phrase this section again suggestions welcomed
0
353,620
25,127,602,730
IssuesEvent
2022-11-09 12:58:06
mhostetter/galois
https://api.github.com/repos/mhostetter/galois
closed
Project Structure
documentation
Hi @mhostetter, I'm opening this issue to better understand the structure of the `galois` package. In case this should be redirected / moved to another issue, please let me know. ### Unused classes 1. Are the base classes in `galois/array.py` such as [`{Field|Group|Ring}ArrayBase`](https://github.com/mhostetter/galois/blob/49f8f1d8ef2ae5c2dc6aa92c8b0c18f6c8bc3829/galois/array.py#L241) being used?
1.0
Project Structure - Hi @mhostetter, I'm opening this issue to better understand the structure of the `galois` package. In case this should be redirected / moved to another issue, please let me know. ### Unused classes 1. Are the base classes in `galois/array.py` such as [`{Field|Group|Ring}ArrayBase`](https://github.com/mhostetter/galois/blob/49f8f1d8ef2ae5c2dc6aa92c8b0c18f6c8bc3829/galois/array.py#L241) being used?
non_infrastructure
project structure hi mhostetter i m opening this issue to better understand the structure of the galois package in case this should be redirected moved to another issue please let me know unused classes are the base classes in galois array py such as being used
0
417,157
12,156,252,800
IssuesEvent
2020-04-25 16:32:19
rajan123456/uPresent
https://api.github.com/repos/rajan123456/uPresent
opened
Management: Allow setting Timezone for a module
enhancement low-priority
- [ ] : Management service should allow timezones in modules - [ ] : Allow admin to set the timezone of the module from AdminUI. - [ ] : Attendance to consider timezone of the module during module check.
1.0
Management: Allow setting Timezone for a module - - [ ] : Management service should allow timezones in modules - [ ] : Allow admin to set the timezone of the module from AdminUI. - [ ] : Attendance to consider timezone of the module during module check.
non_infrastructure
management allow setting timezone for a module management service should allow timezones in modules allow admin to set the timezone of the module from adminui attendance to consider timezone of the module during module check
0
29,185
23,780,603,425
IssuesEvent
2022-09-02 04:03:41
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
ReflectionUtilities.Clone() fails with "SerializationException: Unable to find assembly" when using non-default AssemblyLoadContext
bug interface/infrastructure
### Summary When ApsimX's ReflectionUtilities.Clone() method is called on an instance of a type that belongs to an assembly that was not loaded in the default AssemblyLoadContext, it fails with `SerializationException: Unable to find assembly '...'.` (Another possibility is that the assembly will also be found in the default load context, and Clone() will succeed, but then there will be an InvalidCastException when the caller of Clone() tries to cast the result.) A fix is proposed in #7465. ### Background and analysis ApsimX's [ReflectionUtilities.Clone()](https://github.com/APSIMInitiative/ApsimX/blob/265679cee30b0355e222f6a5832272ecc16f98da/APSIM.Shared/Utilities/ReflectionUtilities.cs#L544) method is implemented by serializing and deserializing using .NET's [BinaryFormatter](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.serialization.formatters.binary.binaryformatter). When BinaryFormatter performs deserialization, it needs to be able to load assemblies. .NET Core has something called "AssemblyLoadContext" ([documentation](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext)), which allows assemblies to be loaded in isolated contexts (e.g. in order to load multiple versions of an assembly). By default, BinaryFormatter.Deserialize loads assemblies using [Assembly.Load(AssemblyName)](https://docs.microsoft.com/en-us/dotnet/api/system.reflection.assembly.load#system-reflection-assembly-load(system-reflection-assemblyname)) (source code references: [[1]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/Formatters/Binary/BinaryObjectReader.cs#L1109), [[2]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/FormatterServices.cs#L319)). In the context of BinaryFormatter, this uses the [default AssemblyLoadContext](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.default). If the type that is being cloned belongs to an assembly from the default context, this works fine. But if the type belongs to an assembly from a *different* context, this is problematic: the assembly might not be found in the default context, and if it *is* found it may later [cause an InvalidCastException](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext#type-conversion-issues) (because loading the assembly in the default context results in a new Assembly instance and its types are distinct). One situation where non-default contexts arise is when writing a COM (Component Object Model) server. .NET Core loads the COM server in an isolated AssemblyLoadContext (reference: [COM hosting design doc](https://github.com/dotnet/runtime/blob/main/docs/design/features/native-hosting.md#loading-managed-code) and [notes in the .NET Core COM server docs](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com#additional-notes)). The dependencies of the COM server (e.g. Models.dll) are automatically loaded in this isolated context when the COM server starts. So if the COM server creates a Simulation instance that is later passed to Clone(), the call will fail: If Models.dll is not found in the default context, there will be a SerializationException "Unable to find assembly 'Models ...'." If Models.dll *is* found, there will be an InvalidCastException when the caller of Clone() tries to cast the result to Simulation, because the result's type will be from a different Assembly instance. ### Steps to reproduce #### Example 1 (minimal, contrived) This is a minimal (but contrived) example that manually loads an assembly in a non-default AssemblyLoadContext and calls Clone() on an instance of a type from that assembly. 1. Create a class library project "DemoLibrary": ``` dotnet new classlib -f netcoreapp3.1 -o DemoLibrary ``` Create a file "DemoSimulation.cs" in this project with the following contents: ```c# using System; namespace DemoLibrary { [Serializable] public class DemoSimulation { } } ``` Run `dotnet build` in this project (this generates bin\Debug\netcoreapp3.1\DemoLibrary.dll). 2. Create a console application project "DemoApp" that loads the library and calls Clone(): ``` dotnet new console -f netcoreapp3.1 -o DemoApp cd DemoApp dotnet add reference ...\ApsimX\APSIM.Shared\APSIM.Shared.csproj # replace "..." with the actual path ``` Modify Program.cs to have the following contents: ```c# using APSIM.Shared.Utilities; using System; using System.IO; using System.Reflection; using System.Runtime.Loader; namespace DemoApp { class DummyAssemblyLoadContext : AssemblyLoadContext { } class Program { static void Main(string[] args) { AssemblyLoadContext assemblyLoadContext = new DummyAssemblyLoadContext(); Assembly demoLibrary = assemblyLoadContext.LoadFromAssemblyPath(Directory.GetCurrentDirectory() + @"\..\DemoLibrary\bin\Debug\netcoreapp3.1\DemoLibrary.dll"); Type demoSimulationClass = demoLibrary.GetType("DemoLibrary.DemoSimulation", throwOnError: true); object sim = Activator.CreateInstance(demoSimulationClass); // dynamic equivalent of "DemoSimulation sim = new DemoSimulation();" object clone = ReflectionUtilities.Clone(sim); // fails with "SerializationException: Unable to find assembly 'DemoLibrary, ...'." Console.WriteLine("Clone() succeeded. IsInstance: {0}", demoSimulationClass.IsInstanceOfType(clone)); } } } ``` Run using `dotnet run`. Output (on ApsimX at commit 4cc851718f8b): ``` Unhandled exception. System.Runtime.Serialization.SerializationException: Unable to find assembly 'DemoLibrary, Version=1.0.0.0, Culture=neutral, PublicKeyToken=null'. at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name) at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream) at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 344 at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 551 at DemoApp.Program.Main(String[] args) in C:\Users\User\DemoApp\Program.cs:line 21 ``` (If a reference to DemoLibrary is added to DemoApp, Clone() will succeed and the output will be `Clone() succeeded. IsInstance: false`.) #### Example 2: COM Server This is a larger example that uses a COM (Component Object Model) server. It is the real-life use case that led to the discovery of the issue. 1. Download the repository [dotnet/samples](https://github.com/dotnet/samples) to get the COM Server Demo ([documentation](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com), [source subdirectory](https://github.com/dotnet/samples/tree/main/core/extensions/COMServerDemo)). 2. Modify the COMServer project to use ApsimX: ``` cd samples\core\extensions\COMServerDemo\COMServer dotnet add reference ...\ApsimX\Models\Models.csproj # replace "..." with the actual path ``` Modify Server.cs to have the following contents: ```c# using Models.Core; using Models.Core.Run; using System; using System.Collections.Generic; using System.Runtime.InteropServices; namespace COMServer { [ComVisible(true)] [Guid(ContractGuids.ServerClass)] public class Server : IServer { double IServer.ComputePi() { Simulations sims = Simulations.Create(new Simulation[] { new Simulation() }); // based on APSIM.Cli/Program.cs Runner runner = new Runner(sims); List<Exception> errors = runner.Run(); if (errors != null && errors.Count > 0) { throw new AggregateException("Simulation ran with errors", errors); } return 3.14; } } } ``` Build the server: ``` dotnet build -p:RegFree=True ``` (The above command also copies the COMServer output files to the COMClient project. I'm using registration-free mode because it does not require administrative privileges.) 3. Build and run the COMClient project: ``` cd ..\COMClient dotnet build -p:RegFree=True dotnet run --no-build ``` (Note that `dotnet run` must be invoked with `--no-build`, otherwise it might try to automatically build without `-p:RegFree=True`. If that happens, the fix is to run `dotnet clean; dotnet build -p:RegFree=True`. Cleaning is necessary because the build system does not recognise the property change as a reason for a full rebuild.) Output: ``` Unhandled exception. System.AggregateException: Simulation ran with errors (Error in file: Simulation: Simulation) ---> System.Exception: Error in file: Simulation: Simulation ---> System.Runtime.Serialization.SerializationException: Unable to find assembly 'Models, Version=0.0.0.0, Culture=neutral, PublicKeyToken=null'. at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name) at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream) at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream) at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj) at Models.Core.Apsim.Clone[T](T model) at Models.Core.Run.SimulationDescription.ToSimulation() --- End of inner exception stack trace --- at Models.Core.Run.SimulationDescription.ToSimulation() at Models.Core.Run.SimulationDescription.Prepare() at APSIM.Shared.JobRunning.JobRunner.Prepare(IRunnable job) at APSIM.Shared.JobRunning.JobRunner.RunActualJob(IRunnable job, IJobManager jobManager) --- End of inner exception stack trace --- at COMServer.Server.IServer.ComputePi() at IServer.ComputePi() at COMClient.Program.Main(String[] args) in C:\Users\User\dotnet-samples\core\extensions\COMServerDemo\COMClient\Program.cs:line 12 ``` ### Proposed fix #7465 ### Alternative workaround There is a simple (but ugly) workaround which can be implemented in the user's application code without modifying ApsimX. The idea is to add a handler to the event [AssemblyLoadContext.Resolving](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.resolving) of the default context (or the event [AppDomain.AssemblyResolve](https://docs.microsoft.com/en-us/dotnet/api/system.appdomain.assemblyresolve)), and when it occurs load the assembly from the desired context instead. Note that this event only occurs when assembly resolution fails, so this workaround only works if the assembly is *not* found in the default context. ```c# ... public class Server : IServer { public Server() { // [Somewhere at the start of the program] AssemblyLoadContext.Default.Resolving += OnResolving; } public static Assembly OnResolving(AssemblyLoadContext originalContext, AssemblyName assemblyName) { // [Replace "Server" with the name of some class from the desired load context] AssemblyLoadContext customContext = AssemblyLoadContext.GetLoadContext(typeof(Server).Assembly); return customContext.LoadFromAssemblyName(assemblyName); } ... } ``` **Discovered by:** @HemdaLevyNZ.
1.0
ReflectionUtilities.Clone() fails with "SerializationException: Unable to find assembly" when using non-default AssemblyLoadContext - ### Summary When ApsimX's ReflectionUtilities.Clone() method is called on an instance of a type that belongs to an assembly that was not loaded in the default AssemblyLoadContext, it fails with `SerializationException: Unable to find assembly '...'.` (Another possibility is that the assembly will also be found in the default load context, and Clone() will succeed, but then there will be an InvalidCastException when the caller of Clone() tries to cast the result.) A fix is proposed in #7465. ### Background and analysis ApsimX's [ReflectionUtilities.Clone()](https://github.com/APSIMInitiative/ApsimX/blob/265679cee30b0355e222f6a5832272ecc16f98da/APSIM.Shared/Utilities/ReflectionUtilities.cs#L544) method is implemented by serializing and deserializing using .NET's [BinaryFormatter](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.serialization.formatters.binary.binaryformatter). When BinaryFormatter performs deserialization, it needs to be able to load assemblies. .NET Core has something called "AssemblyLoadContext" ([documentation](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext)), which allows assemblies to be loaded in isolated contexts (e.g. in order to load multiple versions of an assembly). By default, BinaryFormatter.Deserialize loads assemblies using [Assembly.Load(AssemblyName)](https://docs.microsoft.com/en-us/dotnet/api/system.reflection.assembly.load#system-reflection-assembly-load(system-reflection-assemblyname)) (source code references: [[1]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/Formatters/Binary/BinaryObjectReader.cs#L1109), [[2]](https://github.com/dotnet/runtime/blob/v6.0.7/src/libraries/System.Runtime.Serialization.Formatters/src/System/Runtime/Serialization/FormatterServices.cs#L319)). In the context of BinaryFormatter, this uses the [default AssemblyLoadContext](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.default). If the type that is being cloned belongs to an assembly from the default context, this works fine. But if the type belongs to an assembly from a *different* context, this is problematic: the assembly might not be found in the default context, and if it *is* found it may later [cause an InvalidCastException](https://docs.microsoft.com/en-us/dotnet/core/dependency-loading/understanding-assemblyloadcontext#type-conversion-issues) (because loading the assembly in the default context results in a new Assembly instance and its types are distinct). One situation where non-default contexts arise is when writing a COM (Component Object Model) server. .NET Core loads the COM server in an isolated AssemblyLoadContext (reference: [COM hosting design doc](https://github.com/dotnet/runtime/blob/main/docs/design/features/native-hosting.md#loading-managed-code) and [notes in the .NET Core COM server docs](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com#additional-notes)). The dependencies of the COM server (e.g. Models.dll) are automatically loaded in this isolated context when the COM server starts. So if the COM server creates a Simulation instance that is later passed to Clone(), the call will fail: If Models.dll is not found in the default context, there will be a SerializationException "Unable to find assembly 'Models ...'." If Models.dll *is* found, there will be an InvalidCastException when the caller of Clone() tries to cast the result to Simulation, because the result's type will be from a different Assembly instance. ### Steps to reproduce #### Example 1 (minimal, contrived) This is a minimal (but contrived) example that manually loads an assembly in a non-default AssemblyLoadContext and calls Clone() on an instance of a type from that assembly. 1. Create a class library project "DemoLibrary": ``` dotnet new classlib -f netcoreapp3.1 -o DemoLibrary ``` Create a file "DemoSimulation.cs" in this project with the following contents: ```c# using System; namespace DemoLibrary { [Serializable] public class DemoSimulation { } } ``` Run `dotnet build` in this project (this generates bin\Debug\netcoreapp3.1\DemoLibrary.dll). 2. Create a console application project "DemoApp" that loads the library and calls Clone(): ``` dotnet new console -f netcoreapp3.1 -o DemoApp cd DemoApp dotnet add reference ...\ApsimX\APSIM.Shared\APSIM.Shared.csproj # replace "..." with the actual path ``` Modify Program.cs to have the following contents: ```c# using APSIM.Shared.Utilities; using System; using System.IO; using System.Reflection; using System.Runtime.Loader; namespace DemoApp { class DummyAssemblyLoadContext : AssemblyLoadContext { } class Program { static void Main(string[] args) { AssemblyLoadContext assemblyLoadContext = new DummyAssemblyLoadContext(); Assembly demoLibrary = assemblyLoadContext.LoadFromAssemblyPath(Directory.GetCurrentDirectory() + @"\..\DemoLibrary\bin\Debug\netcoreapp3.1\DemoLibrary.dll"); Type demoSimulationClass = demoLibrary.GetType("DemoLibrary.DemoSimulation", throwOnError: true); object sim = Activator.CreateInstance(demoSimulationClass); // dynamic equivalent of "DemoSimulation sim = new DemoSimulation();" object clone = ReflectionUtilities.Clone(sim); // fails with "SerializationException: Unable to find assembly 'DemoLibrary, ...'." Console.WriteLine("Clone() succeeded. IsInstance: {0}", demoSimulationClass.IsInstanceOfType(clone)); } } } ``` Run using `dotnet run`. Output (on ApsimX at commit 4cc851718f8b): ``` Unhandled exception. System.Runtime.Serialization.SerializationException: Unable to find assembly 'DemoLibrary, Version=1.0.0.0, Culture=neutral, PublicKeyToken=null'. at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name) at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream) at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 344 at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj) in C:\Users\User\ApsimX\APSIM.Shared\Utilities\ReflectionUtilities.cs:line 551 at DemoApp.Program.Main(String[] args) in C:\Users\User\DemoApp\Program.cs:line 21 ``` (If a reference to DemoLibrary is added to DemoApp, Clone() will succeed and the output will be `Clone() succeeded. IsInstance: false`.) #### Example 2: COM Server This is a larger example that uses a COM (Component Object Model) server. It is the real-life use case that led to the discovery of the issue. 1. Download the repository [dotnet/samples](https://github.com/dotnet/samples) to get the COM Server Demo ([documentation](https://docs.microsoft.com/en-us/dotnet/core/native-interop/expose-components-to-com), [source subdirectory](https://github.com/dotnet/samples/tree/main/core/extensions/COMServerDemo)). 2. Modify the COMServer project to use ApsimX: ``` cd samples\core\extensions\COMServerDemo\COMServer dotnet add reference ...\ApsimX\Models\Models.csproj # replace "..." with the actual path ``` Modify Server.cs to have the following contents: ```c# using Models.Core; using Models.Core.Run; using System; using System.Collections.Generic; using System.Runtime.InteropServices; namespace COMServer { [ComVisible(true)] [Guid(ContractGuids.ServerClass)] public class Server : IServer { double IServer.ComputePi() { Simulations sims = Simulations.Create(new Simulation[] { new Simulation() }); // based on APSIM.Cli/Program.cs Runner runner = new Runner(sims); List<Exception> errors = runner.Run(); if (errors != null && errors.Count > 0) { throw new AggregateException("Simulation ran with errors", errors); } return 3.14; } } } ``` Build the server: ``` dotnet build -p:RegFree=True ``` (The above command also copies the COMServer output files to the COMClient project. I'm using registration-free mode because it does not require administrative privileges.) 3. Build and run the COMClient project: ``` cd ..\COMClient dotnet build -p:RegFree=True dotnet run --no-build ``` (Note that `dotnet run` must be invoked with `--no-build`, otherwise it might try to automatically build without `-p:RegFree=True`. If that happens, the fix is to run `dotnet clean; dotnet build -p:RegFree=True`. Cleaning is necessary because the build system does not recognise the property change as a reason for a full rebuild.) Output: ``` Unhandled exception. System.AggregateException: Simulation ran with errors (Error in file: Simulation: Simulation) ---> System.Exception: Error in file: Simulation: Simulation ---> System.Runtime.Serialization.SerializationException: Unable to find assembly 'Models, Version=0.0.0.0, Culture=neutral, PublicKeyToken=null'. at System.Runtime.Serialization.Formatters.Binary.BinaryAssemblyInfo.GetAssembly() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.GetType(BinaryAssemblyInfo assemblyInfo, String name) at System.Runtime.Serialization.Formatters.Binary.ObjectMap..ctor(String objectName, String[] memberNames, BinaryTypeEnum[] binaryTypeEnumA, Object[] typeInformationA, Int32[] memberAssemIds, ObjectReader objectReader, Int32 objectId, BinaryAssemblyInfo assemblyInfo, SizedArray assemIdToAssemblyTable) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryObjectWithMapTyped record) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.ReadObjectWithMapTyped(BinaryHeaderEnum binaryHeaderEnum) at System.Runtime.Serialization.Formatters.Binary.BinaryParser.Run() at System.Runtime.Serialization.Formatters.Binary.ObjectReader.Deserialize(BinaryParser serParser, Boolean fCheck) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream, Boolean check) at System.Runtime.Serialization.Formatters.Binary.BinaryFormatter.Deserialize(Stream serializationStream) at APSIM.Shared.Utilities.ReflectionUtilities.BinaryDeserialise(Stream stream) at APSIM.Shared.Utilities.ReflectionUtilities.Clone(Object sourceObj) at Models.Core.Apsim.Clone[T](T model) at Models.Core.Run.SimulationDescription.ToSimulation() --- End of inner exception stack trace --- at Models.Core.Run.SimulationDescription.ToSimulation() at Models.Core.Run.SimulationDescription.Prepare() at APSIM.Shared.JobRunning.JobRunner.Prepare(IRunnable job) at APSIM.Shared.JobRunning.JobRunner.RunActualJob(IRunnable job, IJobManager jobManager) --- End of inner exception stack trace --- at COMServer.Server.IServer.ComputePi() at IServer.ComputePi() at COMClient.Program.Main(String[] args) in C:\Users\User\dotnet-samples\core\extensions\COMServerDemo\COMClient\Program.cs:line 12 ``` ### Proposed fix #7465 ### Alternative workaround There is a simple (but ugly) workaround which can be implemented in the user's application code without modifying ApsimX. The idea is to add a handler to the event [AssemblyLoadContext.Resolving](https://docs.microsoft.com/en-us/dotnet/api/system.runtime.loader.assemblyloadcontext.resolving) of the default context (or the event [AppDomain.AssemblyResolve](https://docs.microsoft.com/en-us/dotnet/api/system.appdomain.assemblyresolve)), and when it occurs load the assembly from the desired context instead. Note that this event only occurs when assembly resolution fails, so this workaround only works if the assembly is *not* found in the default context. ```c# ... public class Server : IServer { public Server() { // [Somewhere at the start of the program] AssemblyLoadContext.Default.Resolving += OnResolving; } public static Assembly OnResolving(AssemblyLoadContext originalContext, AssemblyName assemblyName) { // [Replace "Server" with the name of some class from the desired load context] AssemblyLoadContext customContext = AssemblyLoadContext.GetLoadContext(typeof(Server).Assembly); return customContext.LoadFromAssemblyName(assemblyName); } ... } ``` **Discovered by:** @HemdaLevyNZ.
infrastructure
reflectionutilities clone fails with serializationexception unable to find assembly when using non default assemblyloadcontext summary when apsimx s reflectionutilities clone method is called on an instance of a type that belongs to an assembly that was not loaded in the default assemblyloadcontext it fails with serializationexception unable to find assembly another possibility is that the assembly will also be found in the default load context and clone will succeed but then there will be an invalidcastexception when the caller of clone tries to cast the result a fix is proposed in background and analysis apsimx s method is implemented by serializing and deserializing using net s when binaryformatter performs deserialization it needs to be able to load assemblies net core has something called assemblyloadcontext which allows assemblies to be loaded in isolated contexts e g in order to load multiple versions of an assembly by default binaryformatter deserialize loads assemblies using source code references in the context of binaryformatter this uses the if the type that is being cloned belongs to an assembly from the default context this works fine but if the type belongs to an assembly from a different context this is problematic the assembly might not be found in the default context and if it is found it may later because loading the assembly in the default context results in a new assembly instance and its types are distinct one situation where non default contexts arise is when writing a com component object model server net core loads the com server in an isolated assemblyloadcontext reference and the dependencies of the com server e g models dll are automatically loaded in this isolated context when the com server starts so if the com server creates a simulation instance that is later passed to clone the call will fail if models dll is not found in the default context there will be a serializationexception unable to find assembly models if models dll is found there will be an invalidcastexception when the caller of clone tries to cast the result to simulation because the result s type will be from a different assembly instance steps to reproduce example minimal contrived this is a minimal but contrived example that manually loads an assembly in a non default assemblyloadcontext and calls clone on an instance of a type from that assembly create a class library project demolibrary dotnet new classlib f o demolibrary create a file demosimulation cs in this project with the following contents c using system namespace demolibrary public class demosimulation run dotnet build in this project this generates bin debug demolibrary dll create a console application project demoapp that loads the library and calls clone dotnet new console f o demoapp cd demoapp dotnet add reference apsimx apsim shared apsim shared csproj replace with the actual path modify program cs to have the following contents c using apsim shared utilities using system using system io using system reflection using system runtime loader namespace demoapp class dummyassemblyloadcontext assemblyloadcontext class program static void main string args assemblyloadcontext assemblyloadcontext new dummyassemblyloadcontext assembly demolibrary assemblyloadcontext loadfromassemblypath directory getcurrentdirectory demolibrary bin debug demolibrary dll type demosimulationclass demolibrary gettype demolibrary demosimulation throwonerror true object sim activator createinstance demosimulationclass dynamic equivalent of demosimulation sim new demosimulation object clone reflectionutilities clone sim fails with serializationexception unable to find assembly demolibrary console writeline clone succeeded isinstance demosimulationclass isinstanceoftype clone run using dotnet run output on apsimx at commit unhandled exception system runtime serialization serializationexception unable to find assembly demolibrary version culture neutral publickeytoken null at system runtime serialization formatters binary binaryassemblyinfo getassembly at system runtime serialization formatters binary objectreader gettype binaryassemblyinfo assemblyinfo string name at system runtime serialization formatters binary objectmap ctor string objectname string membernames binarytypeenum binarytypeenuma object typeinformationa memberassemids objectreader objectreader objectid binaryassemblyinfo assemblyinfo sizedarray assemidtoassemblytable at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryobjectwithmaptyped record at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryheaderenum binaryheaderenum at system runtime serialization formatters binary binaryparser run at system runtime serialization formatters binary objectreader deserialize binaryparser serparser boolean fcheck at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream boolean check at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream at apsim shared utilities reflectionutilities binarydeserialise stream stream in c users user apsimx apsim shared utilities reflectionutilities cs line at apsim shared utilities reflectionutilities clone object sourceobj in c users user apsimx apsim shared utilities reflectionutilities cs line at demoapp program main string args in c users user demoapp program cs line if a reference to demolibrary is added to demoapp clone will succeed and the output will be clone succeeded isinstance false example com server this is a larger example that uses a com component object model server it is the real life use case that led to the discovery of the issue download the repository to get the com server demo modify the comserver project to use apsimx cd samples core extensions comserverdemo comserver dotnet add reference apsimx models models csproj replace with the actual path modify server cs to have the following contents c using models core using models core run using system using system collections generic using system runtime interopservices namespace comserver public class server iserver double iserver computepi simulations sims simulations create new simulation new simulation based on apsim cli program cs runner runner new runner sims list errors runner run if errors null errors count throw new aggregateexception simulation ran with errors errors return build the server dotnet build p regfree true the above command also copies the comserver output files to the comclient project i m using registration free mode because it does not require administrative privileges build and run the comclient project cd comclient dotnet build p regfree true dotnet run no build note that dotnet run must be invoked with no build otherwise it might try to automatically build without p regfree true if that happens the fix is to run dotnet clean dotnet build p regfree true cleaning is necessary because the build system does not recognise the property change as a reason for a full rebuild output unhandled exception system aggregateexception simulation ran with errors error in file simulation simulation system exception error in file simulation simulation system runtime serialization serializationexception unable to find assembly models version culture neutral publickeytoken null at system runtime serialization formatters binary binaryassemblyinfo getassembly at system runtime serialization formatters binary objectreader gettype binaryassemblyinfo assemblyinfo string name at system runtime serialization formatters binary objectmap ctor string objectname string membernames binarytypeenum binarytypeenuma object typeinformationa memberassemids objectreader objectreader objectid binaryassemblyinfo assemblyinfo sizedarray assemidtoassemblytable at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryobjectwithmaptyped record at system runtime serialization formatters binary binaryparser readobjectwithmaptyped binaryheaderenum binaryheaderenum at system runtime serialization formatters binary binaryparser run at system runtime serialization formatters binary objectreader deserialize binaryparser serparser boolean fcheck at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream boolean check at system runtime serialization formatters binary binaryformatter deserialize stream serializationstream at apsim shared utilities reflectionutilities binarydeserialise stream stream at apsim shared utilities reflectionutilities clone object sourceobj at models core apsim clone t model at models core run simulationdescription tosimulation end of inner exception stack trace at models core run simulationdescription tosimulation at models core run simulationdescription prepare at apsim shared jobrunning jobrunner prepare irunnable job at apsim shared jobrunning jobrunner runactualjob irunnable job ijobmanager jobmanager end of inner exception stack trace at comserver server iserver computepi at iserver computepi at comclient program main string args in c users user dotnet samples core extensions comserverdemo comclient program cs line proposed fix alternative workaround there is a simple but ugly workaround which can be implemented in the user s application code without modifying apsimx the idea is to add a handler to the event of the default context or the event and when it occurs load the assembly from the desired context instead note that this event only occurs when assembly resolution fails so this workaround only works if the assembly is not found in the default context c public class server iserver public server assemblyloadcontext default resolving onresolving public static assembly onresolving assemblyloadcontext originalcontext assemblyname assemblyname assemblyloadcontext customcontext assemblyloadcontext getloadcontext typeof server assembly return customcontext loadfromassemblyname assemblyname discovered by hemdalevynz
1
7,760
7,084,097,406
IssuesEvent
2018-01-11 04:32:54
comses/core.comses.net
https://api.github.com/repos/comses/core.comses.net
closed
set up production webpack deploy
frontend infrastructure
`npm run build` is currently failing - we need to fix this for prod deployment. https://vuejs.org/v2/guide/deployment.html
1.0
set up production webpack deploy - `npm run build` is currently failing - we need to fix this for prod deployment. https://vuejs.org/v2/guide/deployment.html
infrastructure
set up production webpack deploy npm run build is currently failing we need to fix this for prod deployment
1
23,688
6,474,614,232
IssuesEvent
2017-08-17 18:29:52
Microsoft/TypeScript
https://api.github.com/repos/Microsoft/TypeScript
reopened
API to tell editor to trigger rename after applying refactoring
Bug VS Code Tracked
**Problem** Using the new extract method refactoring in TS 2.5, extracted methods currently are always named `newFunction*`. Renaming the extracted method requires a second step after the refactoring as been applied A better flow would automatically trigger the rename on the new function after the refactoring has been applied. **Proposal** In the `getEditsForRefactor` response, return an additional field indicating that an editor should perform another action, such as rename, after the refactor has been applied.
1.0
API to tell editor to trigger rename after applying refactoring - **Problem** Using the new extract method refactoring in TS 2.5, extracted methods currently are always named `newFunction*`. Renaming the extracted method requires a second step after the refactoring as been applied A better flow would automatically trigger the rename on the new function after the refactoring has been applied. **Proposal** In the `getEditsForRefactor` response, return an additional field indicating that an editor should perform another action, such as rename, after the refactor has been applied.
non_infrastructure
api to tell editor to trigger rename after applying refactoring problem using the new extract method refactoring in ts extracted methods currently are always named newfunction renaming the extracted method requires a second step after the refactoring as been applied a better flow would automatically trigger the rename on the new function after the refactoring has been applied proposal in the geteditsforrefactor response return an additional field indicating that an editor should perform another action such as rename after the refactor has been applied
0
11,520
9,220,677,820
IssuesEvent
2019-03-11 18:03:38
HumanCellAtlas/secondary-analysis
https://api.github.com/repos/HumanCellAtlas/secondary-analysis
closed
Update Optimus Adapters for Primetime
Q1-2019 infrastructure moved optimus 🚫blocked
It has been a long time since we have run Optimus in the DCP. We think there is work to do to update the adapters. After we update Optimus to have zarr outputs, we should update it so we can run the pipeline in the DCP. ~Consider #355 when doing this ticket, they can be kept separate if the work is too much for one ticket.~ AC: - [ ] Update adapters to support latest tagged version of Optimus - [ ] Update lira config - [ ] Zarr output is created, and is captured in adapters - [ ] Run with current existing dataset to help determine next steps for testing.
1.0
Update Optimus Adapters for Primetime - It has been a long time since we have run Optimus in the DCP. We think there is work to do to update the adapters. After we update Optimus to have zarr outputs, we should update it so we can run the pipeline in the DCP. ~Consider #355 when doing this ticket, they can be kept separate if the work is too much for one ticket.~ AC: - [ ] Update adapters to support latest tagged version of Optimus - [ ] Update lira config - [ ] Zarr output is created, and is captured in adapters - [ ] Run with current existing dataset to help determine next steps for testing.
infrastructure
update optimus adapters for primetime it has been a long time since we have run optimus in the dcp we think there is work to do to update the adapters after we update optimus to have zarr outputs we should update it so we can run the pipeline in the dcp consider when doing this ticket they can be kept separate if the work is too much for one ticket ac update adapters to support latest tagged version of optimus update lira config zarr output is created and is captured in adapters run with current existing dataset to help determine next steps for testing
1
270,338
28,960,505,825
IssuesEvent
2023-05-10 01:47:46
praneethpanasala/linux
https://api.github.com/repos/praneethpanasala/linux
reopened
WS-2021-0554 (High) detected in linuxlinux-4.19.6
Mend: dependency security vulnerability
## WS-2021-0554 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> nfc: fix potential NULL pointer deref in nfc_genl_dump_ses_done This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven. This ID is fixed in Linux Kernel version v5.15.8 by commit fae9705d281091254d4a81fa2da9d22346097dca, it was introduced in version v3.12 by commit ac22ac466a659f1b2e02a2e2ee23fc5c42da2c95. For more details please see the references link. <p>Publish Date: 2021-12-19 <p>URL: <a href=https://github.com/gregkh/linux/commit/fae9705d281091254d4a81fa2da9d22346097dca>WS-2021-0554</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002588">https://osv.dev/vulnerability/GSD-2021-1002588</a></p> <p>Release Date: 2021-12-19</p> <p>Fix Resolution: v5.15.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0554 (High) detected in linuxlinux-4.19.6 - ## WS-2021-0554 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/praneethpanasala/linux/commits/d80c4f847c91020292cb280132b15e2ea147f1a3">d80c4f847c91020292cb280132b15e2ea147f1a3</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> nfc: fix potential NULL pointer deref in nfc_genl_dump_ses_done This is an automated ID intended to aid in discovery of potential security vulnerabilities. The actual impact and attack plausibility have not yet been proven. This ID is fixed in Linux Kernel version v5.15.8 by commit fae9705d281091254d4a81fa2da9d22346097dca, it was introduced in version v3.12 by commit ac22ac466a659f1b2e02a2e2ee23fc5c42da2c95. For more details please see the references link. <p>Publish Date: 2021-12-19 <p>URL: <a href=https://github.com/gregkh/linux/commit/fae9705d281091254d4a81fa2da9d22346097dca>WS-2021-0554</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://osv.dev/vulnerability/GSD-2021-1002588">https://osv.dev/vulnerability/GSD-2021-1002588</a></p> <p>Release Date: 2021-12-19</p> <p>Fix Resolution: v5.15.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
ws high detected in linuxlinux ws high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details nfc fix potential null pointer deref in nfc genl dump ses done this is an automated id intended to aid in discovery of potential security vulnerabilities the actual impact and attack plausibility have not yet been proven this id is fixed in linux kernel version by commit it was introduced in version by commit for more details please see the references link publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
27,610
21,992,288,544
IssuesEvent
2022-05-26 00:02:04
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Remove un-needed target when upgrading to v2 BuildTools
enhancement area-Infrastructure-coreclr no-recent-activity backlog-cleanup-candidate
We have done some change in the PR dotnet/coreclr#11482 to change the license Url for packaging. because coreclr repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the src/.nuget/dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct
1.0
Remove un-needed target when upgrading to v2 BuildTools - We have done some change in the PR dotnet/coreclr#11482 to change the license Url for packaging. because coreclr repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the src/.nuget/dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct
infrastructure
remove un needed target when upgrading to buildtools we have done some change in the pr dotnet coreclr to change the license url for packaging because coreclr repo not using of the buildtools we had to hack this by adding the temporary target overridelicenseurl in the src nuget dir targets files this target should be removed after upgrading to of the buildtools and need to ensure when removing it the license url in the packages nuspec files is correct
1
18,349
12,950,473,198
IssuesEvent
2020-07-19 13:20:30
kopia/kopia
https://api.github.com/repos/kopia/kopia
opened
compression performance regression at commit 272b657508329ac4035c656c2503f6f476d42e7d
bug performance project-infrastructure
This happened back in March between 0.5.2 and HEAD, looks like we lost >31% of compression efficiency of `s2-default` after upgrading compression libraries. On 200 GB sample files the compressed repository size before the change was 89.98 GB, after it jumped to 118.67 GB. The suspect change was on 2020-03-15 and upgraded `github.com/klauspost/compress` from `v1.9.7` to `v1.10.3` Thanks @ntolia for the investigation!
1.0
compression performance regression at commit 272b657508329ac4035c656c2503f6f476d42e7d - This happened back in March between 0.5.2 and HEAD, looks like we lost >31% of compression efficiency of `s2-default` after upgrading compression libraries. On 200 GB sample files the compressed repository size before the change was 89.98 GB, after it jumped to 118.67 GB. The suspect change was on 2020-03-15 and upgraded `github.com/klauspost/compress` from `v1.9.7` to `v1.10.3` Thanks @ntolia for the investigation!
infrastructure
compression performance regression at commit this happened back in march between and head looks like we lost of compression efficiency of default after upgrading compression libraries on gb sample files the compressed repository size before the change was gb after it jumped to gb the suspect change was on and upgraded github com klauspost compress from to thanks ntolia for the investigation
1
28,600
23,383,631,981
IssuesEvent
2022-08-11 11:54:46
dotnet/sdk
https://api.github.com/repos/dotnet/sdk
closed
Changes to bring templates into SDK introduces source-build prebuilts
Area-Infrastructure untriaged
The installer's references to Microsoft.DotNet.Common.ItemTemplates and Microsoft.DotNet.Common.ProjectTemplates are being detected as prebuilts in source-build. This appears to be a result of https://github.com/dotnet/sdk/pull/26521. I am pretty sure this is because the templating projects weren't added to the sdk repo's [source-build.slnf](https://github.com/dotnet/sdk/blob/main/source-build.slnf) therefore they are not getting built as part of source-build. cc @vlada-shubina, @GangWang01
1.0
Changes to bring templates into SDK introduces source-build prebuilts - The installer's references to Microsoft.DotNet.Common.ItemTemplates and Microsoft.DotNet.Common.ProjectTemplates are being detected as prebuilts in source-build. This appears to be a result of https://github.com/dotnet/sdk/pull/26521. I am pretty sure this is because the templating projects weren't added to the sdk repo's [source-build.slnf](https://github.com/dotnet/sdk/blob/main/source-build.slnf) therefore they are not getting built as part of source-build. cc @vlada-shubina, @GangWang01
infrastructure
changes to bring templates into sdk introduces source build prebuilts the installer s references to microsoft dotnet common itemtemplates and microsoft dotnet common projecttemplates are being detected as prebuilts in source build this appears to be a result of i am pretty sure this is because the templating projects weren t added to the sdk repo s therefore they are not getting built as part of source build cc vlada shubina
1
20,439
10,519,742,633
IssuesEvent
2019-09-29 20:06:08
DashboardHub/PipelineDashboard
https://api.github.com/repos/DashboardHub/PipelineDashboard
closed
CVE-2018-20822 (Medium) detected in node-sass-v4.12.0
security vulnerability
## CVE-2018-20822 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/DashboardHub/PipelineDashboard/commits/bb91378eadab1837766aef02a00d31b46095a1c4">bb91378eadab1837766aef02a00d31b46095a1c4</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/unchecked.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_node.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/base.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operation.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.hpp - /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/contrib/plugin.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_superselector.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/paths.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_unification.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/values.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/json.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/checked.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/string.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass2scss.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/factory.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/value.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.cpp - /PipelineDashboard/web/node_modules/node-sass/src/callback_bridge.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/functions.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.cpp - /PipelineDashboard/web/node_modules/node-sass/src/custom_function_bridge.cpp - /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/bind.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/backtrace.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debugger.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cencode.c - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/base64vlq.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/number.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/c99func.c - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/values.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_subset_map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass2scss.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/null.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/context.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/script/test-leaks.pl - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/b64/encode.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.hpp - /PipelineDashboard/web/node_modules/node-sass/src/binding.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debug.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822>CVE-2018-20822</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-20822 (Medium) detected in node-sass-v4.12.0 - ## CVE-2018-20822 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.12.0</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://api.github.com/repos/DashboardHub/PipelineDashboard/commits/bb91378eadab1837766aef02a00d31b46095a1c4">bb91378eadab1837766aef02a00d31b46095a1c4</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (125)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/unchecked.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_node.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/base.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operation.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.hpp - /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/contrib/plugin.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_superselector.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/error_handling.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/subset_map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/emitter.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/output.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_def_macros.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/functions.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/paths.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_unification.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/values.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_util.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/json.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/check_nesting.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/units.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8/checked.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/listize.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/string.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/context.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/prelexer.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass2scss.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/eval.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/expand.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/factory.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/boolean.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/source_map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/value.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/utf8_string.cpp - /PipelineDashboard/web/node_modules/node-sass/src/callback_bridge.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/node.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/operators.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast_fwd_decl.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/parser.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/constants.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/list.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cssize.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/functions.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/util.cpp - /PipelineDashboard/web/node_modules/node-sass/src/custom_function_bridge.cpp - /PipelineDashboard/web/node_modules/node-sass/src/custom_importer_bridge.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/bind.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/inspect.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_functions.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/backtrace.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/extend.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/sass_value_wrapper.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debugger.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/cencode.c - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/base64vlq.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/number.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/color.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/c99func.c - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/position.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/remove_placeholders.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_values.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/values.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/test/test_subset_map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass2scss.cpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/null.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/include/sass/context.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/ast.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/color_maps.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_context_wrapper.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/script/test-leaks.pl - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/memory/SharedPtr.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/lexer.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_c.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/to_value.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/b64/encode.h - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/file.hpp - /PipelineDashboard/web/node_modules/node-sass/src/sass_types/map.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/environment.hpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/plugins.hpp - /PipelineDashboard/web/node_modules/node-sass/src/binding.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/sass_context.cpp - /PipelineDashboard/web/node_modules/node-sass/src/libsass/src/debug.hpp </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp). <p>Publish Date: 2019-04-23 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822>CVE-2018-20822</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve medium detected in node sass cve medium severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries pipelinedashboard web node modules node sass src libsass src expand hpp pipelinedashboard web node modules node sass src libsass src color maps cpp pipelinedashboard web node modules node sass src libsass src sass util hpp pipelinedashboard web node modules node sass src libsass src unchecked h pipelinedashboard web node modules node sass src libsass src output hpp pipelinedashboard web node modules node sass src libsass src sass values hpp pipelinedashboard web node modules node sass src libsass src util hpp pipelinedashboard web node modules node sass src libsass src emitter hpp pipelinedashboard web node modules node sass src libsass src lexer cpp pipelinedashboard web node modules node sass src libsass test test node cpp pipelinedashboard web node modules node sass src libsass src plugins cpp pipelinedashboard web node modules node sass src libsass include sass base h pipelinedashboard web node modules node sass src libsass src position hpp pipelinedashboard web node modules node sass src libsass src subset map hpp pipelinedashboard web node modules node sass src libsass src operation hpp pipelinedashboard web node modules node sass src libsass src remove placeholders cpp pipelinedashboard web node modules node sass src libsass src error handling hpp pipelinedashboard web node modules node sass src custom importer bridge cpp pipelinedashboard web node modules node sass src libsass contrib plugin cpp pipelinedashboard web node modules node sass src libsass src functions hpp pipelinedashboard web node modules node sass src libsass test test superselector cpp pipelinedashboard web node modules node sass src libsass src eval hpp pipelinedashboard web node modules node sass src libsass src string hpp pipelinedashboard web node modules node sass src libsass src error handling cpp pipelinedashboard web node modules node sass src sass context wrapper h pipelinedashboard web node modules node sass src libsass src node cpp pipelinedashboard web node modules node sass src libsass src parser cpp pipelinedashboard web node modules node sass src libsass src subset map cpp pipelinedashboard web node modules node sass src libsass src emitter cpp pipelinedashboard web node modules node sass src libsass src listize cpp pipelinedashboard web node modules node sass src libsass src ast hpp pipelinedashboard web node modules node sass src libsass src sass functions hpp pipelinedashboard web node modules node sass src libsass src memory sharedptr cpp pipelinedashboard web node modules node sass src libsass src output cpp pipelinedashboard web node modules node sass src libsass src check nesting cpp pipelinedashboard web node modules node sass src libsass src ast def macros hpp pipelinedashboard web node modules node sass src libsass src cssize hpp pipelinedashboard web node modules node sass src libsass src functions cpp pipelinedashboard web node modules node sass src libsass src paths hpp pipelinedashboard web node modules node sass src libsass src prelexer cpp pipelinedashboard web node modules node sass src libsass src ast fwd decl hpp pipelinedashboard web node modules node sass src sass types color cpp pipelinedashboard web node modules node sass src libsass test test unification cpp pipelinedashboard web node modules node sass src libsass src inspect hpp pipelinedashboard web node modules node sass src libsass src values cpp pipelinedashboard web node modules node sass src libsass src sass util cpp pipelinedashboard web node modules node sass src libsass src source map hpp pipelinedashboard web node modules node sass src sass types list h pipelinedashboard web node modules node sass src libsass src json cpp pipelinedashboard web node modules node sass src libsass src check nesting hpp pipelinedashboard web node modules node sass src libsass src units cpp pipelinedashboard web node modules node sass src libsass src units hpp pipelinedashboard web node modules node sass src libsass src context cpp pipelinedashboard web node modules node sass src libsass src checked h pipelinedashboard web node modules node sass src libsass src listize hpp pipelinedashboard web node modules node sass src sass types string cpp pipelinedashboard web node modules node sass src libsass src context hpp pipelinedashboard web node modules node sass src libsass src prelexer hpp pipelinedashboard web node modules node sass src sass types boolean h pipelinedashboard web node modules node sass src libsass include h pipelinedashboard web node modules node sass src libsass src eval cpp pipelinedashboard web node modules node sass src libsass src expand cpp pipelinedashboard web node modules node sass src sass types factory cpp pipelinedashboard web node modules node sass src libsass src operators cpp pipelinedashboard web node modules node sass src sass types boolean cpp pipelinedashboard web node modules node sass src libsass src source map cpp pipelinedashboard web node modules node sass src sass types value h pipelinedashboard web node modules node sass src libsass src string cpp pipelinedashboard web node modules node sass src callback bridge h pipelinedashboard web node modules node sass src libsass src file cpp pipelinedashboard web node modules node sass src libsass src sass cpp pipelinedashboard web node modules node sass src libsass src node hpp pipelinedashboard web node modules node sass src libsass src environment cpp pipelinedashboard web node modules node sass src libsass src extend hpp pipelinedashboard web node modules node sass src libsass src sass context hpp pipelinedashboard web node modules node sass src libsass src operators hpp pipelinedashboard web node modules node sass src libsass src constants hpp pipelinedashboard web node modules node sass src libsass src sass hpp pipelinedashboard web node modules node sass src libsass src ast fwd decl cpp pipelinedashboard web node modules node sass src libsass src parser hpp pipelinedashboard web node modules node sass src libsass src constants cpp pipelinedashboard web node modules node sass src sass types list cpp pipelinedashboard web node modules node sass src libsass src cssize cpp pipelinedashboard web node modules node sass src libsass include sass functions h pipelinedashboard web node modules node sass src libsass src util cpp pipelinedashboard web node modules node sass src custom function bridge cpp pipelinedashboard web node modules node sass src custom importer bridge h pipelinedashboard web node modules node sass src libsass src bind cpp pipelinedashboard web node modules node sass src libsass src inspect cpp pipelinedashboard web node modules node sass src libsass src sass functions cpp pipelinedashboard web node modules node sass src libsass src backtrace cpp pipelinedashboard web node modules node sass src libsass src extend cpp pipelinedashboard web node modules node sass src sass types sass value wrapper h pipelinedashboard web node modules node sass src libsass src debugger hpp pipelinedashboard web node modules node sass src libsass src cencode c pipelinedashboard web node modules node sass src libsass src cpp pipelinedashboard web node modules node sass src sass types number cpp pipelinedashboard web node modules node sass src sass types color h pipelinedashboard web node modules node sass src libsass src c pipelinedashboard web node modules node sass src libsass src position cpp pipelinedashboard web node modules node sass src libsass src remove placeholders hpp pipelinedashboard web node modules node sass src libsass src sass values cpp pipelinedashboard web node modules node sass src libsass include sass values h pipelinedashboard web node modules node sass src libsass test test subset map cpp pipelinedashboard web node modules node sass src libsass src cpp pipelinedashboard web node modules node sass src sass types null cpp pipelinedashboard web node modules node sass src libsass include sass context h pipelinedashboard web node modules node sass src libsass src ast cpp pipelinedashboard web node modules node sass src libsass src to c cpp pipelinedashboard web node modules node sass src libsass src to value hpp pipelinedashboard web node modules node sass src libsass src color maps hpp pipelinedashboard web node modules node sass src sass context wrapper cpp pipelinedashboard web node modules node sass src libsass script test leaks pl pipelinedashboard web node modules node sass src libsass src memory sharedptr hpp pipelinedashboard web node modules node sass src libsass src lexer hpp pipelinedashboard web node modules node sass src libsass src to c hpp pipelinedashboard web node modules node sass src libsass src to value cpp pipelinedashboard web node modules node sass src libsass src encode h pipelinedashboard web node modules node sass src libsass src file hpp pipelinedashboard web node modules node sass src sass types map cpp pipelinedashboard web node modules node sass src libsass src environment hpp pipelinedashboard web node modules node sass src libsass src plugins hpp pipelinedashboard web node modules node sass src binding cpp pipelinedashboard web node modules node sass src libsass src sass context cpp pipelinedashboard web node modules node sass src libsass src debug hpp vulnerability details libsass allows attackers to cause a denial of service uncontrolled recursion in sass complex selector perform in ast hpp and sass inspect operator in inspect cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
72,445
3,385,863,189
IssuesEvent
2015-11-27 14:09:24
thesgc/chembiohub_helpdesk
https://api.github.com/repos/thesgc/chembiohub_helpdesk
opened
data overview page has buttons for Add single project and Add multiple project- these should refer t
app: AssayReg name: Karen priority: Low status: New
data overview page has buttons for Add single project and Add multiple project- these should refer to activities. Also the Activities label currently reads Projects.
1.0
data overview page has buttons for Add single project and Add multiple project- these should refer t - data overview page has buttons for Add single project and Add multiple project- these should refer to activities. Also the Activities label currently reads Projects.
non_infrastructure
data overview page has buttons for add single project and add multiple project these should refer t data overview page has buttons for add single project and add multiple project these should refer to activities also the activities label currently reads projects
0
212,671
7,241,345,285
IssuesEvent
2018-02-14 00:42:21
MRN-Code/coinstac
https://api.github.com/repos/MRN-Code/coinstac
opened
Notification Badges on Throughout Collections Nav for Unmapped Vars
low priority
# Problem Notification badges should be used throughout the application to guide users to areas needing there attention. One such area is Collections when unmapped vars have been found on a pending run. # Solution - Badge on Collection nav item - Badge on corresponding collection list item. - Badge on Consortia tab of collection
1.0
Notification Badges on Throughout Collections Nav for Unmapped Vars - # Problem Notification badges should be used throughout the application to guide users to areas needing there attention. One such area is Collections when unmapped vars have been found on a pending run. # Solution - Badge on Collection nav item - Badge on corresponding collection list item. - Badge on Consortia tab of collection
non_infrastructure
notification badges on throughout collections nav for unmapped vars problem notification badges should be used throughout the application to guide users to areas needing there attention one such area is collections when unmapped vars have been found on a pending run solution badge on collection nav item badge on corresponding collection list item badge on consortia tab of collection
0
42,508
9,222,617,720
IssuesEvent
2019-03-11 23:37:39
publiclab/plots2
https://api.github.com/repos/publiclab/plots2
opened
Welcome, Outreachy aspirants! (Summer- 2019)
outreachy summer-of-code
Hello folks from Outreachy - We realized we don't have a main place for you all to ask questions and get support, so we wanted to say WELCOME and create this issue as a discussion zone. Please leave a comment to say hello, in part so we know who's interested in the program! @publiclab/reviewers @publiclab/plots2-reviewers @publiclab/mentors please keep an eye open for opportunities to welcome and support our Outreachy applicants! 👍 🎉 ## First contribution It's so great to see that several of you have made your first contribution. 🙌 This should really help as all our contributions follow this workflow. If you haven't yet, please see this page to get started: https://code.publiclab.org ## What next? If you're looking for a next project, we might recommend one which starts to involve multiple files in the application code, so you can start to get a better understanding of the application's workings. Are you familiar with how a Rails application works, in terms of how the model accesses the database, the controller performs the logic, and the views display the actual web pages? If not, we can help you start to pick apart how our system works one piece at a time. You can find a lot of relatively small but not completely documented issues here: https://github.com/publiclab/plots2/labels/fto-candidate If you choose one, please leave a comment and we can do our best to help find any missing information or help guide you through the next steps. Don't hesitate to ask questions! Larger collections of tasks can often be found under the `break-me-up` label, which you can go through to find smaller projects to "break off" and solve: https://github.com/publiclab/plots2/labels/break-me-up ## Supporting tasks Another area you might be interested in contributing is in supporting other new people to the project. This is an often unrecognized part of a healthy collaborative project, so we've made these issues to encourage people to take part in this crucial activity. We'd love your help, and as a newcomer yourself, you know exactly what it feels like to be new, so you know how important it is to be welcoming and supportive! https://github.com/publiclab/plots2/labels/support You can help new contributors in getting started by creating new FTO issues - https://github.com/publiclab/plots2/issues/4963 Thanks and we're happy to answer questions!!
1.0
Welcome, Outreachy aspirants! (Summer- 2019) - Hello folks from Outreachy - We realized we don't have a main place for you all to ask questions and get support, so we wanted to say WELCOME and create this issue as a discussion zone. Please leave a comment to say hello, in part so we know who's interested in the program! @publiclab/reviewers @publiclab/plots2-reviewers @publiclab/mentors please keep an eye open for opportunities to welcome and support our Outreachy applicants! 👍 🎉 ## First contribution It's so great to see that several of you have made your first contribution. 🙌 This should really help as all our contributions follow this workflow. If you haven't yet, please see this page to get started: https://code.publiclab.org ## What next? If you're looking for a next project, we might recommend one which starts to involve multiple files in the application code, so you can start to get a better understanding of the application's workings. Are you familiar with how a Rails application works, in terms of how the model accesses the database, the controller performs the logic, and the views display the actual web pages? If not, we can help you start to pick apart how our system works one piece at a time. You can find a lot of relatively small but not completely documented issues here: https://github.com/publiclab/plots2/labels/fto-candidate If you choose one, please leave a comment and we can do our best to help find any missing information or help guide you through the next steps. Don't hesitate to ask questions! Larger collections of tasks can often be found under the `break-me-up` label, which you can go through to find smaller projects to "break off" and solve: https://github.com/publiclab/plots2/labels/break-me-up ## Supporting tasks Another area you might be interested in contributing is in supporting other new people to the project. This is an often unrecognized part of a healthy collaborative project, so we've made these issues to encourage people to take part in this crucial activity. We'd love your help, and as a newcomer yourself, you know exactly what it feels like to be new, so you know how important it is to be welcoming and supportive! https://github.com/publiclab/plots2/labels/support You can help new contributors in getting started by creating new FTO issues - https://github.com/publiclab/plots2/issues/4963 Thanks and we're happy to answer questions!!
non_infrastructure
welcome outreachy aspirants summer hello folks from outreachy we realized we don t have a main place for you all to ask questions and get support so we wanted to say welcome and create this issue as a discussion zone please leave a comment to say hello in part so we know who s interested in the program publiclab reviewers publiclab reviewers publiclab mentors please keep an eye open for opportunities to welcome and support our outreachy applicants 👍 🎉 first contribution it s so great to see that several of you have made your first contribution 🙌 this should really help as all our contributions follow this workflow if you haven t yet please see this page to get started what next if you re looking for a next project we might recommend one which starts to involve multiple files in the application code so you can start to get a better understanding of the application s workings are you familiar with how a rails application works in terms of how the model accesses the database the controller performs the logic and the views display the actual web pages if not we can help you start to pick apart how our system works one piece at a time you can find a lot of relatively small but not completely documented issues here if you choose one please leave a comment and we can do our best to help find any missing information or help guide you through the next steps don t hesitate to ask questions larger collections of tasks can often be found under the break me up label which you can go through to find smaller projects to break off and solve supporting tasks another area you might be interested in contributing is in supporting other new people to the project this is an often unrecognized part of a healthy collaborative project so we ve made these issues to encourage people to take part in this crucial activity we d love your help and as a newcomer yourself you know exactly what it feels like to be new so you know how important it is to be welcoming and supportive you can help new contributors in getting started by creating new fto issues thanks and we re happy to answer questions
0
33,722
27,752,654,754
IssuesEvent
2023-03-15 22:15:00
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Can't build in an ARM64 Windows machine, fails in crosscompontents step
area-Infrastructure-coreclr untriaged
Machine: Surface Pro X (ARM64) with Windows 11 Command: `.\build.cmd clr+libs -c release -arch arm64` (repros with or without the `arch` argument) This is the error that shows up without any code changes: <details><summary>Original error output (expand)</summary> ``` [vcvarsall.bat] Environment initialized for: 'x64' BUILD: Regenerating the Visual Studio solution Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64" vs2022 x64 linux "-DCMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE _OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" "-DCLR_CROSS_COMPONENTS_BUILD=1" Not searching for unused variables given on the command line. -- The C compiler identification is MSVC 19.33.31631.0 -- The CXX compiler identification is MSVC 19.33.31631.0 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - failed -- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe -- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe - broken CMake Error at C:/Program Files/CMake/share/cmake-3.26/Modules/CMakeTestCCompiler.cmake:67 (message): The C compiler "C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe" is not able to compile a simple test program. It fails with the following output: Change Dir: C:/repos/runtime/artifacts/obj/coreclr/linux.arm64.Release/x64/CMakeFiles/CMakeScratch/TryCompile-cyevcr Run Build Command(s):C:/PROGRA~1/ninja/ninja.exe -v cmTC_f9773 && [1/2] C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\ cl.exe /nologo /DWIN32 /D_WINDOWS /W3 /Zi /Ob0 /Od /RTC1 -MDd /showIncludes /FoCMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /FdCMakeFiles\cmTC_f97 73.dir\ /FS -c C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64\CMakeFiles\CMakeScratch\TryCompile-cyevcr\testCCompiler.c [2/2] cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin \100226~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.31 6\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /versi on:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib && cd ." FAILED: cmTC_f9773.exe cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\10022 6~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\ HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdl g32.lib advapi32.lib && cd ." LINK Pass 1: command "C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCo mpiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.l ib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MANIFESTFILE:CMakeFiles\cmTC_f977 3.dir/intermediate.manifest CMakeFiles\cmTC_f9773.dir/manifest.res" failed (exit code 1120) with the following output: testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_InitBase testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_Shutdown LINK : error LNK2001: unresolved external symbol mainCRTStartup C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\MSVCRTD.lib : warning LNK4272: library machine type 'ARM6 4' conflicts with target machine type 'x64' C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\OLDNAMES.lib : warning LNK4272: library machine type 'ARM 64' conflicts with target machine type 'x64' cmTC_f9773.exe : fatal error LNK1120: 3 unresolved externals ninja: build stopped: subcommand failed. CMake will not be able to correctly generate this project. Call Stack (most recent call first): CMakeLists.txt:16 (project) -- Configuring incomplete, errors occurred! BUILD: Error: failed to generate native component build project C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1. Build FAILED. C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1. 0 Warning(s) 1 Error(s) Time Elapsed 00:03:49.01 Build failed with exit code 1. Check errors above. Some builds failed: Configuration: release, Architecture: arm64 ``` </details> As you can see, there's some places where x64 is being used instead of arm64. I tried adding this line in Subsets.props after line 280: ```diff <PropertyGroup> <CrossDacHostArch>x64</CrossDacHostArch> + <CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm64'">arm64</CrossDacHostArch> <CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm'">x86</CrossDacHostArch> </PropertyGroup> ``` But I am still seeing an error. Here is the result: <details><summary>Error output with modification (expand)</summary> ```powershell ❯ .\build.cmd clr+libs -c release -arch arm64 Determining projects to restore... Tool 'coverlet.console' (version '3.1.2') was restored. Available commands: coverlet Tool 'dotnet-reportgenerator-globaltool' (version '5.0.2') was restored. Available commands: reportgenerator Tool 'microsoft.dotnet.xharness.cli' (version '1.0.0-prerelease.23080.1') was restored. Available commands: xharness Tool 'microsoft.visualstudio.slngen.tool' (version '8.5.17') was restored. Available commands: slngen Tool 'dotnet-format' (version '6.0.240501') was restored. Available commands: dotnet-format Restore was successful. All projects are up-to-date for restore. Determining projects to restore... All projects are up-to-date for restore. System.Private.CoreLib.Generators -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib.Generators\Release\netstandard2.0\System.Private.CoreLib.Generators.dll Microsoft.Interop.SourceGeneration -> C:\repos\runtime\artifacts\bin\Microsoft.Interop.SourceGeneration\Release\netstandard2.0\Microsoft.Interop.SourceGeneration.dll LibraryImportGenerator -> C:\repos\runtime\artifacts\bin\LibraryImportGenerator\Release\netstandard2.0\Microsoft.Interop.LibraryImportGenerator.dll System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib\ref\Release\net8.0\System.Private.CoreLib.dll System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release\IL\System.Private.CoreLib.dll Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -release -os windows -pgodatapath "C:\Users\calope\.nuget\packages\optimization.windows_nt-arm64.pgo.corecl r\1.0.0-prerelease.23068.4" BUILD: Starting Build at 11:53:06.31 BUILD: Commencing CoreCLR product build BUILD: Checking prerequisites BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe BUILD: Number of processor cores 8 BUILD: Commencing build of native components for windows.arm64.Release BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64 ********************************************************************** ** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0 ** Copyright (c) 2022 Microsoft Corporation ********************************************************************** [vcvarsall.bat] Environment initialized for: 'x86_arm64' BUILD: Regenerating the Visual Studio solution Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release" vs2022 arm64 windows "-D CMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=windows" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH="C:\Users\calope\.nuget\pa ckages\optimization.windows_nt-arm64.pgo.coreclr\1.0.0-prerelease.23068.4"" "-DCLR_CMAKE_PGO_OPTIMIZE=1" The CMake command line is the same as the last run. Skipping running CMake. running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release --target install --config Release -- [1/4] Linking RC shared library dlls\mscorrc\mscorrc.dll [2/4] Linking RC shared library dlls\clretwrc\clretwrc.dll [3/4] cmd.exe /C "cd /D C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release\dlls\mscoree\coreclr && powershell -NoProfile -ExecutionPolicy ByPass -File C:/repos/run time/src/coreclr/dlls/mscoree/coreclr/GenClrDebugResource.ps1 -dac C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscordac/mscordaccore.dll -dbi C:/repos/ runtime/artifacts/obj/coreclr/windows.arm64.Releas-- Install configuration: "Release" -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./clretwrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/clretwrc.pdb -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/clretwrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/clretwrc.pdb -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./mscorrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/mscorrc.pdb -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/mscorrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/mscorrc.pdb e/dlls/mscordbi/mscordbi.dll -out C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscoree/coreclr/clr_debug_resource.bin" [3/4] Install the project... BUILD: Build succeeded. Finished at 11:53:16.15 BUILD: Product binaries are available at C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1" -release -os linux -hostarch arm64 -component crosscomponents BUILD: Starting Build at 11:53:16.49 BUILD: Commencing CoreCLR product build BUILD: Checking prerequisites BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe BUILD: Number of processor cores 8 BUILD: Commencing build of native components for linux.arm64.Release BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64 ********************************************************************** ** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0 ** Copyright (c) 2022 Microsoft Corporation ********************************************************************** [vcvarsall.bat] Environment initialized for: 'x86_arm64' BUILD: Regenerating the Visual Studio solution Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release" vs2022 arm64 linux "-DCMAK E_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" " -DCLR_CROSS_COMPONENTS_BUILD=1" The CMake command line is the same as the last run. Skipping running CMake. running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release --target crosscomponents --config Release -- ninja: error: loading 'build.ninja': The system cannot find the file specified. BUILD: Error: native component build failed. Refer to the build log files for details. "C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.log" "C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.wrn" "C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.err" C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1 " -release -os linux -hostarch arm64 -component crosscomponents" exited with code 1. ``` </details> Am I missing something or executing the command wrong? I made sure to install all the required dependencies as described in our workflow instructions. I attached a couple of binlogs in case they are useful: - Without any code changes: [original_msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10970321/original_msbuild.binlog.zip) - After my modification: [msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10960289/msbuild.binlog.zip) @jkoritzinsky @hoyosjs
1.0
Can't build in an ARM64 Windows machine, fails in crosscompontents step - Machine: Surface Pro X (ARM64) with Windows 11 Command: `.\build.cmd clr+libs -c release -arch arm64` (repros with or without the `arch` argument) This is the error that shows up without any code changes: <details><summary>Original error output (expand)</summary> ``` [vcvarsall.bat] Environment initialized for: 'x64' BUILD: Regenerating the Visual Studio solution Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64" vs2022 x64 linux "-DCMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE _OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" "-DCLR_CROSS_COMPONENTS_BUILD=1" Not searching for unused variables given on the command line. -- The C compiler identification is MSVC 19.33.31631.0 -- The CXX compiler identification is MSVC 19.33.31631.0 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - failed -- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe -- Check for working C compiler: C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe - broken CMake Error at C:/Program Files/CMake/share/cmake-3.26/Modules/CMakeTestCCompiler.cmake:67 (message): The C compiler "C:/Program Files/Microsoft Visual Studio/2022/Preview/VC/Tools/MSVC/14.33.31629/bin/HostX64/x64/cl.exe" is not able to compile a simple test program. It fails with the following output: Change Dir: C:/repos/runtime/artifacts/obj/coreclr/linux.arm64.Release/x64/CMakeFiles/CMakeScratch/TryCompile-cyevcr Run Build Command(s):C:/PROGRA~1/ninja/ninja.exe -v cmTC_f9773 && [1/2] C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\ cl.exe /nologo /DWIN32 /D_WINDOWS /W3 /Zi /Ob0 /Od /RTC1 -MDd /showIncludes /FoCMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /FdCMakeFiles\cmTC_f97 73.dir\ /FS -c C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release\x64\CMakeFiles\CMakeScratch\TryCompile-cyevcr\testCCompiler.c [2/2] cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin \100226~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.31 6\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /versi on:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib && cd ." FAILED: cmTC_f9773.exe cmd.exe /C "cd . && "C:\Program Files\CMake\bin\cmake.exe" -E vs_link_exe --intdir=CMakeFiles\cmTC_f9773.dir --rc=C:\PROGRA~2\WI3CF2~1\10\bin\10022 6~1.0\x64\rc.exe --mt=C:\PROGRA~2\WI3CF2~1\10\bin\100226~1.0\x64\mt.exe --manifests -- C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\ HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCompiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.lib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdl g32.lib advapi32.lib && cd ." LINK Pass 1: command "C:\PROGRA~1\MICROS~1\2022\Preview\VC\Tools\MSVC\1433~1.316\bin\HostX64\x64\link.exe /nologo CMakeFiles\cmTC_f9773.dir\testCCo mpiler.c.obj /out:cmTC_f9773.exe /implib:cmTC_f9773.lib /pdb:cmTC_f9773.pdb /version:0.0 /machine:x64 /debug /INCREMENTAL /subsystem:console kernel32.l ib user32.lib gdi32.lib winspool.lib shell32.lib ole32.lib oleaut32.lib uuid.lib comdlg32.lib advapi32.lib /MANIFEST /MANIFESTFILE:CMakeFiles\cmTC_f977 3.dir/intermediate.manifest CMakeFiles\cmTC_f9773.dir/manifest.res" failed (exit code 1120) with the following output: testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_InitBase testCCompiler.c.obj : error LNK2001: unresolved external symbol _RTC_Shutdown LINK : error LNK2001: unresolved external symbol mainCRTStartup C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\MSVCRTD.lib : warning LNK4272: library machine type 'ARM6 4' conflicts with target machine type 'x64' C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Tools\MSVC\14.33.31629\lib\ARM64\OLDNAMES.lib : warning LNK4272: library machine type 'ARM 64' conflicts with target machine type 'x64' cmTC_f9773.exe : fatal error LNK1120: 3 unresolved externals ninja: build stopped: subcommand failed. CMake will not be able to correctly generate this project. Call Stack (most recent call first): CMakeLists.txt:16 (project) -- Configuring incomplete, errors occurred! BUILD: Error: failed to generate native component build project C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1. Build FAILED. C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CR OSS_COMPONENTS_BUILD=1" -release -os linux -hostarch x64 -component crosscomponents" exited with code 1. 0 Warning(s) 1 Error(s) Time Elapsed 00:03:49.01 Build failed with exit code 1. Check errors above. Some builds failed: Configuration: release, Architecture: arm64 ``` </details> As you can see, there's some places where x64 is being used instead of arm64. I tried adding this line in Subsets.props after line 280: ```diff <PropertyGroup> <CrossDacHostArch>x64</CrossDacHostArch> + <CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm64'">arm64</CrossDacHostArch> <CrossDacHostArch Condition="'$(TargetArchitecture)' == 'arm'">x86</CrossDacHostArch> </PropertyGroup> ``` But I am still seeing an error. Here is the result: <details><summary>Error output with modification (expand)</summary> ```powershell ❯ .\build.cmd clr+libs -c release -arch arm64 Determining projects to restore... Tool 'coverlet.console' (version '3.1.2') was restored. Available commands: coverlet Tool 'dotnet-reportgenerator-globaltool' (version '5.0.2') was restored. Available commands: reportgenerator Tool 'microsoft.dotnet.xharness.cli' (version '1.0.0-prerelease.23080.1') was restored. Available commands: xharness Tool 'microsoft.visualstudio.slngen.tool' (version '8.5.17') was restored. Available commands: slngen Tool 'dotnet-format' (version '6.0.240501') was restored. Available commands: dotnet-format Restore was successful. All projects are up-to-date for restore. Determining projects to restore... All projects are up-to-date for restore. System.Private.CoreLib.Generators -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib.Generators\Release\netstandard2.0\System.Private.CoreLib.Generators.dll Microsoft.Interop.SourceGeneration -> C:\repos\runtime\artifacts\bin\Microsoft.Interop.SourceGeneration\Release\netstandard2.0\Microsoft.Interop.SourceGeneration.dll LibraryImportGenerator -> C:\repos\runtime\artifacts\bin\LibraryImportGenerator\Release\netstandard2.0\Microsoft.Interop.LibraryImportGenerator.dll System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\System.Private.CoreLib\ref\Release\net8.0\System.Private.CoreLib.dll System.Private.CoreLib -> C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release\IL\System.Private.CoreLib.dll Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -release -os windows -pgodatapath "C:\Users\calope\.nuget\packages\optimization.windows_nt-arm64.pgo.corecl r\1.0.0-prerelease.23068.4" BUILD: Starting Build at 11:53:06.31 BUILD: Commencing CoreCLR product build BUILD: Checking prerequisites BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe BUILD: Number of processor cores 8 BUILD: Commencing build of native components for windows.arm64.Release BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64 ********************************************************************** ** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0 ** Copyright (c) 2022 Microsoft Corporation ********************************************************************** [vcvarsall.bat] Environment initialized for: 'x86_arm64' BUILD: Regenerating the Visual Studio solution Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release" vs2022 arm64 windows "-D CMAKE_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=windows" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH="C:\Users\calope\.nuget\pa ckages\optimization.windows_nt-arm64.pgo.coreclr\1.0.0-prerelease.23068.4"" "-DCLR_CMAKE_PGO_OPTIMIZE=1" The CMake command line is the same as the last run. Skipping running CMake. running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release --target install --config Release -- [1/4] Linking RC shared library dlls\mscorrc\mscorrc.dll [2/4] Linking RC shared library dlls\clretwrc\clretwrc.dll [3/4] cmd.exe /C "cd /D C:\repos\runtime\artifacts\obj\coreclr\windows.arm64.Release\dlls\mscoree\coreclr && powershell -NoProfile -ExecutionPolicy ByPass -File C:/repos/run time/src/coreclr/dlls/mscoree/coreclr/GenClrDebugResource.ps1 -dac C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscordac/mscordaccore.dll -dbi C:/repos/ runtime/artifacts/obj/coreclr/windows.arm64.Releas-- Install configuration: "Release" -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./clretwrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/clretwrc.pdb -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/clretwrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/clretwrc.pdb -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./mscorrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/./PDB/mscorrc.pdb -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/mscorrc.dll -- Installing: C:/repos/runtime/artifacts/bin/coreclr/windows.arm64.Release/sharedFramework/PDB/mscorrc.pdb e/dlls/mscordbi/mscordbi.dll -out C:/repos/runtime/artifacts/obj/coreclr/windows.arm64.Release/dlls/mscoree/coreclr/clr_debug_resource.bin" [3/4] Install the project... BUILD: Build succeeded. Finished at 11:53:16.15 BUILD: Product binaries are available at C:\repos\runtime\artifacts\bin\coreclr\windows.arm64.Release Executing "C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1" -release -os linux -hostarch arm64 -component crosscomponents BUILD: Starting Build at 11:53:16.49 BUILD: Commencing CoreCLR product build BUILD: Checking prerequisites BUILD: Using CMake from C:\Program Files\CMake\bin\cmake.exe BUILD: Number of processor cores 8 BUILD: Commencing build of native components for linux.arm64.Release BUILD: Using environment: "C:\Program Files\Microsoft Visual Studio\2022\Preview\VC\Auxiliary\Build\vcvarsall.bat" x86_arm64 ********************************************************************** ** Visual Studio 2022 Developer Command Prompt v17.6.0-pre.1.0 ** Copyright (c) 2022 Microsoft Corporation ********************************************************************** [vcvarsall.bat] Environment initialized for: 'x86_arm64' BUILD: Regenerating the Visual Studio solution Calling "C:\repos\runtime\eng\native\gen-buildsys.cmd" "C:\repos\runtime\src\coreclr" "C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release" vs2022 arm64 linux "-DCMAK E_BUILD_TYPE=Release" "-DCLR_CMAKE_TARGET_ARCH=arm64" "-DCLR_CMAKE_TARGET_OS=linux" "-DCLR_CMAKE_PGO_INSTRUMENT=0" "-DCLR_CMAKE_OPTDATA_PATH=" "-DCLR_CMAKE_PGO_OPTIMIZE=0" " -DCLR_CROSS_COMPONENTS_BUILD=1" The CMake command line is the same as the last run. Skipping running CMake. running "C:\Program Files\CMake\bin\cmake.exe" --build C:\repos\runtime\artifacts\obj\coreclr\linux.arm64.Release --target crosscomponents --config Release -- ninja: error: loading 'build.ninja': The system cannot find the file specified. BUILD: Error: native component build failed. Refer to the build log files for details. "C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.log" "C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.wrn" "C:\repos\runtime\artifacts\log\Release\CoreCLR_linux__arm64__Release__arm64.err" C:\repos\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: The command ""C:\repos\runtime\src\coreclr\build-runtime.cmd" -arm64 -cmakeargs "-DCLR_CROSS_COMPONENTS_BUILD=1 " -release -os linux -hostarch arm64 -component crosscomponents" exited with code 1. ``` </details> Am I missing something or executing the command wrong? I made sure to install all the required dependencies as described in our workflow instructions. I attached a couple of binlogs in case they are useful: - Without any code changes: [original_msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10970321/original_msbuild.binlog.zip) - After my modification: [msbuild.binlog.zip](https://github.com/dotnet/runtime/files/10960289/msbuild.binlog.zip) @jkoritzinsky @hoyosjs
infrastructure
can t build in an windows machine fails in crosscompontents step machine surface pro x with windows command build cmd clr libs c release arch repros with or without the arch argument this is the error that shows up without any code changes original error output expand environment initialized for build regenerating the visual studio solution calling c repos runtime eng native gen buildsys cmd c repos runtime src coreclr c repos runtime artifacts obj coreclr linux release linux dcmake build type release dclr cmake target arch dclr cmake target os linux dclr cmake pgo instrument dclr cmake optdata path dclr cmake pgo optimize dclr cross components build not searching for unused variables given on the command line the c compiler identification is msvc the cxx compiler identification is msvc detecting c compiler abi info detecting c compiler abi info failed check for working c compiler c program files microsoft visual studio preview vc tools msvc bin cl exe check for working c compiler c program files microsoft visual studio preview vc tools msvc bin cl exe broken cmake error at c program files cmake share cmake modules cmaketestccompiler cmake message the c compiler c program files microsoft visual studio preview vc tools msvc bin cl exe is not able to compile a simple test program it fails with the following output change dir c repos runtime artifacts obj coreclr linux release cmakefiles cmakescratch trycompile cyevcr run build command s c progra ninja ninja exe v cmtc c progra micros preview vc tools msvc bin cl exe nologo d windows zi od mdd showincludes focmakefiles cmtc dir testccompiler c obj fdcmakefiles cmtc dir fs c c repos runtime artifacts obj coreclr linux release cmakefiles cmakescratch trycompile cyevcr testccompiler c cmd exe c cd c program files cmake bin cmake exe e vs link exe intdir cmakefiles cmtc dir rc c progra bin rc exe mt c progra bin mt exe manifests c progra micros preview vc tools msvc bin link exe nologo cmakefiles cmtc dir testccompiler c obj out cmtc exe implib cmtc lib pdb cmtc pdb versi on machine debug incremental subsystem console lib lib lib winspool lib lib lib lib uuid lib lib lib cd failed cmtc exe cmd exe c cd c program files cmake bin cmake exe e vs link exe intdir cmakefiles cmtc dir rc c progra bin rc exe mt c progra bin mt exe manifests c progra micros preview vc tools msvc bin link exe nologo cmakefiles cmtc dir testccompiler c obj out cmtc exe implib cmtc lib pdb cmtc pdb version machine debug incremental subsystem console lib lib lib winspool lib lib lib lib uuid lib comdl lib lib cd link pass command c progra micros preview vc tools msvc bin link exe nologo cmakefiles cmtc dir testcco mpiler c obj out cmtc exe implib cmtc lib pdb cmtc pdb version machine debug incremental subsystem console l ib lib lib winspool lib lib lib lib uuid lib lib lib manifest manifestfile cmakefiles cmtc dir intermediate manifest cmakefiles cmtc dir manifest res failed exit code with the following output testccompiler c obj error unresolved external symbol rtc initbase testccompiler c obj error unresolved external symbol rtc shutdown link error unresolved external symbol maincrtstartup c program files microsoft visual studio preview vc tools msvc lib msvcrtd lib warning library machine type conflicts with target machine type c program files microsoft visual studio preview vc tools msvc lib oldnames lib warning library machine type arm conflicts with target machine type cmtc exe fatal error unresolved externals ninja build stopped subcommand failed cmake will not be able to correctly generate this project call stack most recent call first cmakelists txt project configuring incomplete errors occurred build error failed to generate native component build project c repos runtime src coreclr runtime proj error the command c repos runtime src coreclr build runtime cmd cmakeargs dclr cr oss components build release os linux hostarch component crosscomponents exited with code build failed c repos runtime src coreclr runtime proj error the command c repos runtime src coreclr build runtime cmd cmakeargs dclr cr oss components build release os linux hostarch component crosscomponents exited with code warning s error s time elapsed build failed with exit code check errors above some builds failed configuration release architecture as you can see there s some places where is being used instead of i tried adding this line in subsets props after line diff but i am still seeing an error here is the result error output with modification expand powershell ❯ build cmd clr libs c release arch determining projects to restore tool coverlet console version was restored available commands coverlet tool dotnet reportgenerator globaltool version was restored available commands reportgenerator tool microsoft dotnet xharness cli version prerelease was restored available commands xharness tool microsoft visualstudio slngen tool version was restored available commands slngen tool dotnet format version was restored available commands dotnet format restore was successful all projects are up to date for restore determining projects to restore all projects are up to date for restore system private corelib generators c repos runtime artifacts bin system private corelib generators release system private corelib generators dll microsoft interop sourcegeneration c repos runtime artifacts bin microsoft interop sourcegeneration release microsoft interop sourcegeneration dll libraryimportgenerator c repos runtime artifacts bin libraryimportgenerator release microsoft interop libraryimportgenerator dll system private corelib c repos runtime artifacts bin system private corelib ref release system private corelib dll system private corelib c repos runtime artifacts bin coreclr windows release il system private corelib dll executing c repos runtime src coreclr build runtime cmd release os windows pgodatapath c users calope nuget packages optimization windows nt pgo corecl r prerelease build starting build at build commencing coreclr product build build checking prerequisites build using cmake from c program files cmake bin cmake exe build number of processor cores build commencing build of native components for windows release build using environment c program files microsoft visual studio preview vc auxiliary build vcvarsall bat visual studio developer command prompt pre copyright c microsoft corporation environment initialized for build regenerating the visual studio solution calling c repos runtime eng native gen buildsys cmd c repos runtime src coreclr c repos runtime artifacts obj coreclr windows release windows d cmake build type release dclr cmake target arch dclr cmake target os windows dclr cmake pgo instrument dclr cmake optdata path c users calope nuget pa ckages optimization windows nt pgo coreclr prerelease dclr cmake pgo optimize the cmake command line is the same as the last run skipping running cmake running c program files cmake bin cmake exe build c repos runtime artifacts obj coreclr windows release target install config release linking rc shared library dlls mscorrc mscorrc dll linking rc shared library dlls clretwrc clretwrc dll cmd exe c cd d c repos runtime artifacts obj coreclr windows release dlls mscoree coreclr powershell noprofile executionpolicy bypass file c repos run time src coreclr dlls mscoree coreclr genclrdebugresource dac c repos runtime artifacts obj coreclr windows release dlls mscordac mscordaccore dll dbi c repos runtime artifacts obj coreclr windows releas install configuration release installing c repos runtime artifacts bin coreclr windows release clretwrc dll installing c repos runtime artifacts bin coreclr windows release pdb clretwrc pdb installing c repos runtime artifacts bin coreclr windows release sharedframework clretwrc dll installing c repos runtime artifacts bin coreclr windows release sharedframework pdb clretwrc pdb installing c repos runtime artifacts bin coreclr windows release mscorrc dll installing c repos runtime artifacts bin coreclr windows release pdb mscorrc pdb installing c repos runtime artifacts bin coreclr windows release sharedframework mscorrc dll installing c repos runtime artifacts bin coreclr windows release sharedframework pdb mscorrc pdb e dlls mscordbi mscordbi dll out c repos runtime artifacts obj coreclr windows release dlls mscoree coreclr clr debug resource bin install the project build build succeeded finished at build product binaries are available at c repos runtime artifacts bin coreclr windows release executing c repos runtime src coreclr build runtime cmd cmakeargs dclr cross components build release os linux hostarch component crosscomponents build starting build at build commencing coreclr product build build checking prerequisites build using cmake from c program files cmake bin cmake exe build number of processor cores build commencing build of native components for linux release build using environment c program files microsoft visual studio preview vc auxiliary build vcvarsall bat visual studio developer command prompt pre copyright c microsoft corporation environment initialized for build regenerating the visual studio solution calling c repos runtime eng native gen buildsys cmd c repos runtime src coreclr c repos runtime artifacts obj coreclr linux release linux dcmak e build type release dclr cmake target arch dclr cmake target os linux dclr cmake pgo instrument dclr cmake optdata path dclr cmake pgo optimize dclr cross components build the cmake command line is the same as the last run skipping running cmake running c program files cmake bin cmake exe build c repos runtime artifacts obj coreclr linux release target crosscomponents config release ninja error loading build ninja the system cannot find the file specified build error native component build failed refer to the build log files for details c repos runtime artifacts log release coreclr linux release log c repos runtime artifacts log release coreclr linux release wrn c repos runtime artifacts log release coreclr linux release err c repos runtime src coreclr runtime proj error the command c repos runtime src coreclr build runtime cmd cmakeargs dclr cross components build release os linux hostarch component crosscomponents exited with code am i missing something or executing the command wrong i made sure to install all the required dependencies as described in our workflow instructions i attached a couple of binlogs in case they are useful without any code changes after my modification jkoritzinsky hoyosjs
1
8,919
7,735,243,683
IssuesEvent
2018-05-27 13:02:27
Integreat/cms
https://api.github.com/repos/Integreat/cms
closed
Remove custom ig-mpdf sql tables
area:infrastructure ease:easy prio:low
As of #719, the custom sql tables for the ig-mpdf plugin are not longer needed for storing the cached pdfs. Hence, they should be removed for all cities.
1.0
Remove custom ig-mpdf sql tables - As of #719, the custom sql tables for the ig-mpdf plugin are not longer needed for storing the cached pdfs. Hence, they should be removed for all cities.
infrastructure
remove custom ig mpdf sql tables as of the custom sql tables for the ig mpdf plugin are not longer needed for storing the cached pdfs hence they should be removed for all cities
1
35,161
30,803,757,407
IssuesEvent
2023-08-01 05:05:58
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Report view 2.0
interface/infrastructure refactor
There are a few future improvements for Report View (+Grid View) that would be great, especially for working with large tables. Some of these may relate to the Speed Up changes being made and may be due to unneeded background work only apparent when working with lots of data. 1. Display empty grid when no table found (empty as never wrote data) for selected report or the report is disabled rather than the data from the first table in the list (even though the table drop down selector is blank) 2. Lazy loading of report and grid view. Is is possible to only load the grid when the user is on or selects the Data tab (maybe with a loading spinner), or use async loading of data? 3. Is it possible to have lazy loading of data to display in the grid view based on scrollbar position, or maybe only load first 50 lines until the user touches scrollbar or down arrow or page down. This may be a builtin feature of future UI controls. These will significantly improve the UI responsiveness especially with large data tables. The same may be true for working with graphs and large data tables.
1.0
Report view 2.0 - There are a few future improvements for Report View (+Grid View) that would be great, especially for working with large tables. Some of these may relate to the Speed Up changes being made and may be due to unneeded background work only apparent when working with lots of data. 1. Display empty grid when no table found (empty as never wrote data) for selected report or the report is disabled rather than the data from the first table in the list (even though the table drop down selector is blank) 2. Lazy loading of report and grid view. Is is possible to only load the grid when the user is on or selects the Data tab (maybe with a loading spinner), or use async loading of data? 3. Is it possible to have lazy loading of data to display in the grid view based on scrollbar position, or maybe only load first 50 lines until the user touches scrollbar or down arrow or page down. This may be a builtin feature of future UI controls. These will significantly improve the UI responsiveness especially with large data tables. The same may be true for working with graphs and large data tables.
infrastructure
report view there are a few future improvements for report view grid view that would be great especially for working with large tables some of these may relate to the speed up changes being made and may be due to unneeded background work only apparent when working with lots of data display empty grid when no table found empty as never wrote data for selected report or the report is disabled rather than the data from the first table in the list even though the table drop down selector is blank lazy loading of report and grid view is is possible to only load the grid when the user is on or selects the data tab maybe with a loading spinner or use async loading of data is it possible to have lazy loading of data to display in the grid view based on scrollbar position or maybe only load first lines until the user touches scrollbar or down arrow or page down this may be a builtin feature of future ui controls these will significantly improve the ui responsiveness especially with large data tables the same may be true for working with graphs and large data tables
1
30,193
24,632,123,094
IssuesEvent
2022-10-17 03:43:26
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Weather downloader can't be closed while it is downloading
bug interface/infrastructure
If SILO is down the user can't cancel or close the dialog box.
1.0
Weather downloader can't be closed while it is downloading - If SILO is down the user can't cancel or close the dialog box.
infrastructure
weather downloader can t be closed while it is downloading if silo is down the user can t cancel or close the dialog box
1
53,532
22,838,850,117
IssuesEvent
2022-07-12 19:30:45
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Update CTN documentation
Workgroup: ATSD Service: Geo Product: CTN
Re-organized cross sections and created a direct link here for ease of navigation
1.0
Update CTN documentation - Re-organized cross sections and created a direct link here for ease of navigation
non_infrastructure
update ctn documentation re organized cross sections and created a direct link here for ease of navigation
0
241,457
20,142,751,866
IssuesEvent
2022-02-09 02:07:50
kevinlizh1992/rtterror
https://api.github.com/repos/kevinlizh1992/rtterror
opened
Service Frontend Small Bugs Summary Report
bug Medium Priority acceptance-test
# Bug Report --- ## 1.Bug URL: `/service` ## 2.Date Created: During release 2. ## 3.Defect Description: During user testing, a number of issues were found in the Service module interface. All issues only affect a small functionality; therefore, they may require creating a separate issue report or not. - [ ] Service code is not shown on the card. - [ ] Service price is not shown on the card. - [ ] When creating/editing a service, there is no text (e.g. "hours") around the time slider. - [ ] Buttons' styles are not consistent. - [ ] Combos creation requires using checkboxes to select Service cards, whereas there are individual buttons for each Service card. The resulting interface is confusing because of the mixture. - [ ] Adding employee to a service button texts needs better/consistent wording. - [ ] When editing a combo, the Service list layout is misaligned. - [ ] It is impossible to edit the combo name or price because they are auto generated from the service list. Yet there are text fields for them. - [ ] The sidebar menu should be either floating or able to hide. Scrolling the page will now show blank area in the page. ## 4.Resolution: A fix is needed for each of the issues listed. ## 5.Resolution Date: Not yet resolved. ## 6.Screenshot Bug: To be added. ---
1.0
Service Frontend Small Bugs Summary Report - # Bug Report --- ## 1.Bug URL: `/service` ## 2.Date Created: During release 2. ## 3.Defect Description: During user testing, a number of issues were found in the Service module interface. All issues only affect a small functionality; therefore, they may require creating a separate issue report or not. - [ ] Service code is not shown on the card. - [ ] Service price is not shown on the card. - [ ] When creating/editing a service, there is no text (e.g. "hours") around the time slider. - [ ] Buttons' styles are not consistent. - [ ] Combos creation requires using checkboxes to select Service cards, whereas there are individual buttons for each Service card. The resulting interface is confusing because of the mixture. - [ ] Adding employee to a service button texts needs better/consistent wording. - [ ] When editing a combo, the Service list layout is misaligned. - [ ] It is impossible to edit the combo name or price because they are auto generated from the service list. Yet there are text fields for them. - [ ] The sidebar menu should be either floating or able to hide. Scrolling the page will now show blank area in the page. ## 4.Resolution: A fix is needed for each of the issues listed. ## 5.Resolution Date: Not yet resolved. ## 6.Screenshot Bug: To be added. ---
non_infrastructure
service frontend small bugs summary report bug report bug url service date created during release defect description during user testing a number of issues were found in the service module interface all issues only affect a small functionality therefore they may require creating a separate issue report or not service code is not shown on the card service price is not shown on the card when creating editing a service there is no text e g hours around the time slider buttons styles are not consistent combos creation requires using checkboxes to select service cards whereas there are individual buttons for each service card the resulting interface is confusing because of the mixture adding employee to a service button texts needs better consistent wording when editing a combo the service list layout is misaligned it is impossible to edit the combo name or price because they are auto generated from the service list yet there are text fields for them the sidebar menu should be either floating or able to hide scrolling the page will now show blank area in the page resolution a fix is needed for each of the issues listed resolution date not yet resolved screenshot bug to be added
0
26,479
20,151,830,769
IssuesEvent
2022-02-09 13:09:45
openforis/fra-platform
https://api.github.com/repos/openforis/fra-platform
closed
Data migration
infrastructure
- [x] Basic tables - [x] add variable name and column names - Custom tables - [x] degraded forest - [x] extentOfForest - [x] forestCharacteristics - [x] growingStock - [x] sustainable development (s)
1.0
Data migration - - [x] Basic tables - [x] add variable name and column names - Custom tables - [x] degraded forest - [x] extentOfForest - [x] forestCharacteristics - [x] growingStock - [x] sustainable development (s)
infrastructure
data migration basic tables add variable name and column names custom tables degraded forest extentofforest forestcharacteristics growingstock sustainable development s
1
6,668
6,550,759,874
IssuesEvent
2017-09-05 12:28:10
opencv/opencv
https://api.github.com/repos/opencv/opencv
closed
Sporadic hangs of iOS / MacOSX builds
bug category: infrastructure category: ios/osx
Usually during CMake stage, ... in `try_compile()` ... the last process is xcodebuild (no child processes) lldb backtrace dump (Xcode 8.3.3): ``` * thread #1, queue = 'com.apple.main-thread', stop reason = signal SIGSTOP * frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10 frame #1: 0x00007fffa5f0d848 libsystem_kernel.dylib`mach_msg + 232 frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212 frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361 frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420 frame #5: 0x00007fff906a6b91 CoreFoundation`CFRunLoopRun + 97 frame #6: 0x0000000106ccbc59 Xcode3Core`-[Xcode3CommandLineBuildTool _buildWithTimingSection:] + 15332 frame #7: 0x0000000106cd6762 Xcode3Core`-[Xcode3CommandLineBuildTool run] + 4807 frame #8: 0x00000001064cc2af xcodebuild`___lldb_unnamed_symbol1$$xcodebuild + 1311 frame #9: 0x00007fffa5de7235 libdyld.dylib`start + 1 frame #10: 0x00007fffa5de7235 libdyld.dylib`start + 1 thread #2, queue = '<IDEBuildOperation:0x7f8589e47cd0:REfc>-builder-queue :: NSOperation 0x7f8589cae3d0 (QOS: DEFAULT)' frame #0: 0x00007fffa5f15c22 libsystem_kernel.dylib`__psynch_mutexwait + 10 frame #1: 0x00007fffa6000dfa libsystem_pthread.dylib`_pthread_mutex_lock_wait + 100 frame #2: 0x00007fffa5ffe519 libsystem_pthread.dylib`_pthread_mutex_lock_slow + 285 frame #3: 0x00007fff92069ee7 Foundation`-[NSRecursiveLock lock] + 22 frame #4: 0x000000010d9aa02d DevToolsCore`-[PBXTargetBuildContext lockDependencyGraph] + 52 frame #5: 0x000000010d9c859f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 126 frame #6: 0x000000010d95e15a DevToolsCore`-[PBXTarget(XCBuildables) buildWillStartForBuilder:buildContext:buildParameters:] + 804 frame #7: 0x000000010db7f492 DevToolsCore`-[Xcode3TargetBuildableSnapshot buildForBuilderWillStart:] + 1144 frame #8: 0x00000001075fc4b7 IDEFoundation`-[IDEBuildableSnapshot performBuildForBuilder:buildCommand:buildOnlyTheseFiles:] + 112 frame #9: 0x000000010762ee06 IDEFoundation`-[IDEBuilder primitiveMain] + 1253 frame #10: 0x000000010762e849 IDEFoundation`-[IDEBuilder main] + 264 frame #11: 0x00007fff9209a324 Foundation`-[__NSOperationInternal _start:] + 672 frame #12: 0x00007fff920961db Foundation`__NSOQSchedule_f + 201 frame #13: 0x00007fffa5db18fc libdispatch.dylib`_dispatch_client_callout + 8 frame #14: 0x00007fffa5dc79a0 libdispatch.dylib`_dispatch_queue_serial_drain + 896 frame #15: 0x00007fffa5dba306 libdispatch.dylib`_dispatch_queue_invoke + 1046 frame #16: 0x00007fffa5db36b5 libdispatch.dylib`_dispatch_root_queue_drain + 476 frame #17: 0x00007fffa5db348c libdispatch.dylib`_dispatch_worker_thread3 + 99 frame #18: 0x00007fffa60005a2 libsystem_pthread.dylib`_pthread_wqthread + 1299 frame #19: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13 thread #3 frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10 frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55 frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212 frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361 frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420 frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277 frame #6: 0x00000001108b5143 DTDeviceKitBase`+[DTDKRemoteDeviceConnection startServiceBrowsers] + 217 frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #4 frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10 frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55 frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212 frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361 frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420 frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277 frame #6: 0x00000001108c3df1 DTDeviceKitBase`-[DTDKRemoteDeviceDataListener listenerThreadImplementation] + 1454 frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #5, name = 'com.apple.CFSocket.private' frame #0: 0x00007fffa5f15eb6 libsystem_kernel.dylib`__select + 10 frame #1: 0x00007fff906a366a CoreFoundation`__CFSocketManager + 682 frame #2: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #3: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #4: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #6 frame #0: 0x00007fffa5f1644e libsystem_kernel.dylib`__workq_kernreturn + 10 frame #1: 0x00007fffa600048e libsystem_pthread.dylib`_pthread_wqthread + 1023 frame #2: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13 thread #7 frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10 frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712 frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594 frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #8 frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10 frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712 frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594 frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #9 frame #0: 0x00007fffa5f0e386 libsystem_kernel.dylib`semaphore_wait_trap + 10 frame #1: 0x00007fffa5dc9a77 libdispatch.dylib`_os_semaphore_wait + 16 frame #2: 0x00007fffa5db77e9 libdispatch.dylib`_dispatch_group_wait_slow + 154 frame #3: 0x000000010669bf14 DVTFoundation`-[DVTFuture _waitUntilFinished] + 137 frame #4: 0x000000010669c1e8 DVTFoundation`-[DVTFuture result] + 22 frame #5: 0x000000010dadab34 DevToolsCore`+[XCCodeSignContext provisioningCommandInputsFromMacroExpansionScope:] + 456 frame #6: 0x000000010dadc8f1 DevToolsCore`+[XCCodeSignContext prepareForCodeSigningWithMacroExpansionScope:] + 172 frame #7: 0x000000010da37aee DevToolsCore`-[XCProductTypeSpecification computeDependenciesWithMacroExpansionScope:] + 1683 frame #8: 0x000000010daed9d3 DevToolsCore`-[XCNativeTargetDGSnapshot(DependencyGraphCreation) computeDependenciesWithMacroExpansionScope:] + 4140 frame #9: 0x000000010d9ad0aa DevToolsCore`-[PBXTargetBuildContext createDependencyGraphWithTargetDGSnapshot:] + 4297 frame #10: 0x000000010d9c8f9b DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) dg_setTargetSnapshot:] + 514 frame #11: 0x000000010d9c863f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 286 frame #12: 0x000000010da2e355 DevToolsCore`-[XCBlockQueue _processNextBlockInThreadSlotNumber:] + 463 frame #13: 0x000000010da2e6fa DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 298 frame #14: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #15: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #16: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #17: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 ```
1.0
Sporadic hangs of iOS / MacOSX builds - Usually during CMake stage, ... in `try_compile()` ... the last process is xcodebuild (no child processes) lldb backtrace dump (Xcode 8.3.3): ``` * thread #1, queue = 'com.apple.main-thread', stop reason = signal SIGSTOP * frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10 frame #1: 0x00007fffa5f0d848 libsystem_kernel.dylib`mach_msg + 232 frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212 frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361 frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420 frame #5: 0x00007fff906a6b91 CoreFoundation`CFRunLoopRun + 97 frame #6: 0x0000000106ccbc59 Xcode3Core`-[Xcode3CommandLineBuildTool _buildWithTimingSection:] + 15332 frame #7: 0x0000000106cd6762 Xcode3Core`-[Xcode3CommandLineBuildTool run] + 4807 frame #8: 0x00000001064cc2af xcodebuild`___lldb_unnamed_symbol1$$xcodebuild + 1311 frame #9: 0x00007fffa5de7235 libdyld.dylib`start + 1 frame #10: 0x00007fffa5de7235 libdyld.dylib`start + 1 thread #2, queue = '<IDEBuildOperation:0x7f8589e47cd0:REfc>-builder-queue :: NSOperation 0x7f8589cae3d0 (QOS: DEFAULT)' frame #0: 0x00007fffa5f15c22 libsystem_kernel.dylib`__psynch_mutexwait + 10 frame #1: 0x00007fffa6000dfa libsystem_pthread.dylib`_pthread_mutex_lock_wait + 100 frame #2: 0x00007fffa5ffe519 libsystem_pthread.dylib`_pthread_mutex_lock_slow + 285 frame #3: 0x00007fff92069ee7 Foundation`-[NSRecursiveLock lock] + 22 frame #4: 0x000000010d9aa02d DevToolsCore`-[PBXTargetBuildContext lockDependencyGraph] + 52 frame #5: 0x000000010d9c859f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 126 frame #6: 0x000000010d95e15a DevToolsCore`-[PBXTarget(XCBuildables) buildWillStartForBuilder:buildContext:buildParameters:] + 804 frame #7: 0x000000010db7f492 DevToolsCore`-[Xcode3TargetBuildableSnapshot buildForBuilderWillStart:] + 1144 frame #8: 0x00000001075fc4b7 IDEFoundation`-[IDEBuildableSnapshot performBuildForBuilder:buildCommand:buildOnlyTheseFiles:] + 112 frame #9: 0x000000010762ee06 IDEFoundation`-[IDEBuilder primitiveMain] + 1253 frame #10: 0x000000010762e849 IDEFoundation`-[IDEBuilder main] + 264 frame #11: 0x00007fff9209a324 Foundation`-[__NSOperationInternal _start:] + 672 frame #12: 0x00007fff920961db Foundation`__NSOQSchedule_f + 201 frame #13: 0x00007fffa5db18fc libdispatch.dylib`_dispatch_client_callout + 8 frame #14: 0x00007fffa5dc79a0 libdispatch.dylib`_dispatch_queue_serial_drain + 896 frame #15: 0x00007fffa5dba306 libdispatch.dylib`_dispatch_queue_invoke + 1046 frame #16: 0x00007fffa5db36b5 libdispatch.dylib`_dispatch_root_queue_drain + 476 frame #17: 0x00007fffa5db348c libdispatch.dylib`_dispatch_worker_thread3 + 99 frame #18: 0x00007fffa60005a2 libsystem_pthread.dylib`_pthread_wqthread + 1299 frame #19: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13 thread #3 frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10 frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55 frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212 frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361 frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420 frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277 frame #6: 0x00000001108b5143 DTDeviceKitBase`+[DTDKRemoteDeviceConnection startServiceBrowsers] + 217 frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #4 frame #0: 0x00007fffa5f0e34a libsystem_kernel.dylib`mach_msg_trap + 10 frame #1: 0x00007fffa5f0d797 libsystem_kernel.dylib`mach_msg + 55 frame #2: 0x00007fff90668434 CoreFoundation`__CFRunLoopServiceMachPort + 212 frame #3: 0x00007fff906678c1 CoreFoundation`__CFRunLoopRun + 1361 frame #4: 0x00007fff90667114 CoreFoundation`CFRunLoopRunSpecific + 420 frame #5: 0x00007fff9207a252 Foundation`-[NSRunLoop(NSRunLoop) runMode:beforeDate:] + 277 frame #6: 0x00000001108c3df1 DTDeviceKitBase`-[DTDKRemoteDeviceDataListener listenerThreadImplementation] + 1454 frame #7: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #8: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #9: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #10: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #5, name = 'com.apple.CFSocket.private' frame #0: 0x00007fffa5f15eb6 libsystem_kernel.dylib`__select + 10 frame #1: 0x00007fff906a366a CoreFoundation`__CFSocketManager + 682 frame #2: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #3: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #4: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #6 frame #0: 0x00007fffa5f1644e libsystem_kernel.dylib`__workq_kernreturn + 10 frame #1: 0x00007fffa600048e libsystem_pthread.dylib`_pthread_wqthread + 1023 frame #2: 0x00007fffa600007d libsystem_pthread.dylib`start_wqthread + 13 thread #7 frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10 frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712 frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594 frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #8 frame #0: 0x00007fffa5f15bf2 libsystem_kernel.dylib`__psynch_cvwait + 10 frame #1: 0x00007fffa60017fa libsystem_pthread.dylib`_pthread_cond_wait + 712 frame #2: 0x000000010da2e822 DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 594 frame #3: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #4: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #5: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #6: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 thread #9 frame #0: 0x00007fffa5f0e386 libsystem_kernel.dylib`semaphore_wait_trap + 10 frame #1: 0x00007fffa5dc9a77 libdispatch.dylib`_os_semaphore_wait + 16 frame #2: 0x00007fffa5db77e9 libdispatch.dylib`_dispatch_group_wait_slow + 154 frame #3: 0x000000010669bf14 DVTFoundation`-[DVTFuture _waitUntilFinished] + 137 frame #4: 0x000000010669c1e8 DVTFoundation`-[DVTFuture result] + 22 frame #5: 0x000000010dadab34 DevToolsCore`+[XCCodeSignContext provisioningCommandInputsFromMacroExpansionScope:] + 456 frame #6: 0x000000010dadc8f1 DevToolsCore`+[XCCodeSignContext prepareForCodeSigningWithMacroExpansionScope:] + 172 frame #7: 0x000000010da37aee DevToolsCore`-[XCProductTypeSpecification computeDependenciesWithMacroExpansionScope:] + 1683 frame #8: 0x000000010daed9d3 DevToolsCore`-[XCNativeTargetDGSnapshot(DependencyGraphCreation) computeDependenciesWithMacroExpansionScope:] + 4140 frame #9: 0x000000010d9ad0aa DevToolsCore`-[PBXTargetBuildContext createDependencyGraphWithTargetDGSnapshot:] + 4297 frame #10: 0x000000010d9c8f9b DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) dg_setTargetSnapshot:] + 514 frame #11: 0x000000010d9c863f DevToolsCore`-[PBXTargetBuildContext(DependencyGraphEvents) processDependencyGraphEvents] + 286 frame #12: 0x000000010da2e355 DevToolsCore`-[XCBlockQueue _processNextBlockInThreadSlotNumber:] + 463 frame #13: 0x000000010da2e6fa DevToolsCore`-[XCBlockQueue _processBlocksInThreadSlotNumber:] + 298 frame #14: 0x00007fff9208a8ad Foundation`__NSThread__start__ + 1243 frame #15: 0x00007fffa600093b libsystem_pthread.dylib`_pthread_body + 180 frame #16: 0x00007fffa6000887 libsystem_pthread.dylib`_pthread_start + 286 frame #17: 0x00007fffa600008d libsystem_pthread.dylib`thread_start + 13 ```
infrastructure
sporadic hangs of ios macosx builds usually during cmake stage in try compile the last process is xcodebuild no child processes lldb backtrace dump xcode thread queue com apple main thread stop reason signal sigstop frame libsystem kernel dylib mach msg trap frame libsystem kernel dylib mach msg frame corefoundation cfrunloopservicemachport frame corefoundation cfrunlooprun frame corefoundation cfrunlooprunspecific frame corefoundation cfrunlooprun frame frame frame xcodebuild lldb unnamed xcodebuild frame libdyld dylib start frame libdyld dylib start thread queue builder queue nsoperation qos default frame libsystem kernel dylib psynch mutexwait frame libsystem pthread dylib pthread mutex lock wait frame libsystem pthread dylib pthread mutex lock slow frame foundation frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame idefoundation frame idefoundation frame idefoundation frame foundation frame foundation nsoqschedule f frame libdispatch dylib dispatch client callout frame libdispatch dylib dispatch queue serial drain frame libdispatch dylib dispatch queue invoke frame libdispatch dylib dispatch root queue drain frame libdispatch dylib dispatch worker frame libsystem pthread dylib pthread wqthread frame libsystem pthread dylib start wqthread thread frame libsystem kernel dylib mach msg trap frame libsystem kernel dylib mach msg frame corefoundation cfrunloopservicemachport frame corefoundation cfrunlooprun frame corefoundation cfrunlooprunspecific frame foundation frame dtdevicekitbase frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib mach msg trap frame libsystem kernel dylib mach msg frame corefoundation cfrunloopservicemachport frame corefoundation cfrunlooprun frame corefoundation cfrunlooprunspecific frame foundation frame dtdevicekitbase frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread name com apple cfsocket private frame libsystem kernel dylib select frame corefoundation cfsocketmanager frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib workq kernreturn frame libsystem pthread dylib pthread wqthread frame libsystem pthread dylib start wqthread thread frame libsystem kernel dylib psynch cvwait frame libsystem pthread dylib pthread cond wait frame devtoolscore frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib psynch cvwait frame libsystem pthread dylib pthread cond wait frame devtoolscore frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start thread frame libsystem kernel dylib semaphore wait trap frame libdispatch dylib os semaphore wait frame libdispatch dylib dispatch group wait slow frame dvtfoundation frame dvtfoundation frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame devtoolscore frame foundation nsthread start frame libsystem pthread dylib pthread body frame libsystem pthread dylib pthread start frame libsystem pthread dylib thread start
1
15,408
11,498,650,197
IssuesEvent
2020-02-12 12:26:48
eclipse/antenna
https://api.github.com/repos/eclipse/antenna
opened
Re-design API of sw360-client library
infrastructure
### Summary of the Improvement Since the sw360-client has been established as an own library new requirements exist, since the library should also be usable in a non-antenna context. Meaning, functionalities antenna does not need should be provided in the client library. e.g. most REST calls to SW360 should be possible with this library This could also have some non-functional aspects like - blocking vs non-blocking rest calls. - making use of a custom or specific Http-library - configuration of the client could be improved / simplified - having a central entry point into the library ### Acceptance Criteria - [ ] Proposal for new API criteria should exist - [ ] all current functionalities are still possible ### Definition of Done - Acceptance criteria fulfilled - The ticket describes what has been done and describes how the results can be verified (e.g. new Jenkins step is inserted, the report can be viewed at X) - All known limitations are listed and it is described how they will be addressed (e.g., a ticket for the remaining stuff) - Documentation is updated - The result has been reviewed: Everything works as described and can be verified reading only the documentation
1.0
Re-design API of sw360-client library - ### Summary of the Improvement Since the sw360-client has been established as an own library new requirements exist, since the library should also be usable in a non-antenna context. Meaning, functionalities antenna does not need should be provided in the client library. e.g. most REST calls to SW360 should be possible with this library This could also have some non-functional aspects like - blocking vs non-blocking rest calls. - making use of a custom or specific Http-library - configuration of the client could be improved / simplified - having a central entry point into the library ### Acceptance Criteria - [ ] Proposal for new API criteria should exist - [ ] all current functionalities are still possible ### Definition of Done - Acceptance criteria fulfilled - The ticket describes what has been done and describes how the results can be verified (e.g. new Jenkins step is inserted, the report can be viewed at X) - All known limitations are listed and it is described how they will be addressed (e.g., a ticket for the remaining stuff) - Documentation is updated - The result has been reviewed: Everything works as described and can be verified reading only the documentation
infrastructure
re design api of client library summary of the improvement since the client has been established as an own library new requirements exist since the library should also be usable in a non antenna context meaning functionalities antenna does not need should be provided in the client library e g most rest calls to should be possible with this library this could also have some non functional aspects like blocking vs non blocking rest calls making use of a custom or specific http library configuration of the client could be improved simplified having a central entry point into the library acceptance criteria proposal for new api criteria should exist all current functionalities are still possible definition of done acceptance criteria fulfilled the ticket describes what has been done and describes how the results can be verified e g new jenkins step is inserted the report can be viewed at x all known limitations are listed and it is described how they will be addressed e g a ticket for the remaining stuff documentation is updated the result has been reviewed everything works as described and can be verified reading only the documentation
1
14,357
10,763,074,230
IssuesEvent
2019-11-01 02:23:48
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Add a "Fertilised" event to Fertiliser
bug interface/infrastructure newfeature
I am writing some documentation for Report and have realised that there are no events for Fertiliser. Can a "Fertlised" or similar event be added. This would make it more consistent with Irrigation and easier to report on a fertilisation.
1.0
Add a "Fertilised" event to Fertiliser - I am writing some documentation for Report and have realised that there are no events for Fertiliser. Can a "Fertlised" or similar event be added. This would make it more consistent with Irrigation and easier to report on a fertilisation.
infrastructure
add a fertilised event to fertiliser i am writing some documentation for report and have realised that there are no events for fertiliser can a fertlised or similar event be added this would make it more consistent with irrigation and easier to report on a fertilisation
1
34,757
30,350,911,946
IssuesEvent
2023-07-11 18:53:08
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
opened
Consolidate E2E Workflows
tool: monorepo infrastructure
There is a wide array of different workflows that run the E2E tests that we should review and centralize. Some of these may not be in use anymore while others may be redundant and able to be consolidated. This task involves both looking into all of these workflows as well as communicating with any stakeholders to gather feedback on how to proceed with handling the workflow. During this work we should also take steps to clean up any remaining E2E workflows and make sure they are consistent with our current GitHub Action best-practices. * `cot-build-and-e2e-tests-daily.yml` * `cot-pr-build-and-e2e-tests.yml` * `pr-build-and-e2e-tests.yml` * `pr-smoke-tests.yml` * `smoke-test-daily-site-check.yml` * `smoke-test-daily.yml` * `smoke-test-pr-merge.yml` * `smoke-test-release.yml` ## Acceptance Criteria * Review all of our workflows and make sure that the list above is exhaustive in the inclusion of E2E test execution. * Identify whether a workflow is in use, and if it isn't, ask stakeholders whether it can just be deleted. * For workflows that could be consolidated, such as the COT ones, combine them into single workflows. * With the workflows that may be redundant, such as all of the smoke test ones, reach out to stakeholders for possible guidance on how we might consolidate them and do so.
1.0
Consolidate E2E Workflows - There is a wide array of different workflows that run the E2E tests that we should review and centralize. Some of these may not be in use anymore while others may be redundant and able to be consolidated. This task involves both looking into all of these workflows as well as communicating with any stakeholders to gather feedback on how to proceed with handling the workflow. During this work we should also take steps to clean up any remaining E2E workflows and make sure they are consistent with our current GitHub Action best-practices. * `cot-build-and-e2e-tests-daily.yml` * `cot-pr-build-and-e2e-tests.yml` * `pr-build-and-e2e-tests.yml` * `pr-smoke-tests.yml` * `smoke-test-daily-site-check.yml` * `smoke-test-daily.yml` * `smoke-test-pr-merge.yml` * `smoke-test-release.yml` ## Acceptance Criteria * Review all of our workflows and make sure that the list above is exhaustive in the inclusion of E2E test execution. * Identify whether a workflow is in use, and if it isn't, ask stakeholders whether it can just be deleted. * For workflows that could be consolidated, such as the COT ones, combine them into single workflows. * With the workflows that may be redundant, such as all of the smoke test ones, reach out to stakeholders for possible guidance on how we might consolidate them and do so.
infrastructure
consolidate workflows there is a wide array of different workflows that run the tests that we should review and centralize some of these may not be in use anymore while others may be redundant and able to be consolidated this task involves both looking into all of these workflows as well as communicating with any stakeholders to gather feedback on how to proceed with handling the workflow during this work we should also take steps to clean up any remaining workflows and make sure they are consistent with our current github action best practices cot build and tests daily yml cot pr build and tests yml pr build and tests yml pr smoke tests yml smoke test daily site check yml smoke test daily yml smoke test pr merge yml smoke test release yml acceptance criteria review all of our workflows and make sure that the list above is exhaustive in the inclusion of test execution identify whether a workflow is in use and if it isn t ask stakeholders whether it can just be deleted for workflows that could be consolidated such as the cot ones combine them into single workflows with the workflows that may be redundant such as all of the smoke test ones reach out to stakeholders for possible guidance on how we might consolidate them and do so
1
32,091
26,420,215,276
IssuesEvent
2023-01-13 19:39:39
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
closed
Test failure in TopicPracticeFragmentTest
Priority: Essential good first issue issue_type_infrastructure issue_user_impact_low issue_temp_ben_triaged issue_user_developer
<img width="1542" alt="Screenshot 2021-06-23 at 6 33 21 PM" src="https://user-images.githubusercontent.com/9396084/123101322-8e505180-d451-11eb-8f2b-d5513968cd5c.png">
1.0
Test failure in TopicPracticeFragmentTest - <img width="1542" alt="Screenshot 2021-06-23 at 6 33 21 PM" src="https://user-images.githubusercontent.com/9396084/123101322-8e505180-d451-11eb-8f2b-d5513968cd5c.png">
infrastructure
test failure in topicpracticefragmenttest img width alt screenshot at pm src
1
18,491
10,235,676,805
IssuesEvent
2019-08-19 09:42:39
jpmorganchase/tessera
https://api.github.com/repos/jpmorganchase/tessera
closed
com.fasterxml.jackson.core:jackson-databind:2.9.9.1 marked vulnerable
security
The dependency `com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1` is marked as having vulnerabilities by the Enforcer plugin. A newer version should be used which does not contain they flagged issues. --- Failure message: ``` [WARNING] Rule 3: org.sonatype.ossindex.maven.enforcer.BanVulnerableDependencies failed with message: Detected 1 vulnerable components: com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1:compile; https://ossindex.sonatype.org/component/pkg:maven/com.fasterxml.jackson.core/jackson-databind@2.9.9.1 * [CVE-2019-14439] A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x befo... (7.5); https://ossindex.sonatype.org/vuln/ac9dce23-7b35-4691-b05e-a68f58d48b8c * [CVE-2019-14379] SubTypeValidator.java in FasterXML jackson-databind before 2.9.9.2 mishandles de... (9.8); https://ossindex.sonatype.org/vuln/e5794172-1257-4372-9baf-7b87307a3cc9 ```
True
com.fasterxml.jackson.core:jackson-databind:2.9.9.1 marked vulnerable - The dependency `com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1` is marked as having vulnerabilities by the Enforcer plugin. A newer version should be used which does not contain they flagged issues. --- Failure message: ``` [WARNING] Rule 3: org.sonatype.ossindex.maven.enforcer.BanVulnerableDependencies failed with message: Detected 1 vulnerable components: com.fasterxml.jackson.core:jackson-databind:jar:2.9.9.1:compile; https://ossindex.sonatype.org/component/pkg:maven/com.fasterxml.jackson.core/jackson-databind@2.9.9.1 * [CVE-2019-14439] A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x befo... (7.5); https://ossindex.sonatype.org/vuln/ac9dce23-7b35-4691-b05e-a68f58d48b8c * [CVE-2019-14379] SubTypeValidator.java in FasterXML jackson-databind before 2.9.9.2 mishandles de... (9.8); https://ossindex.sonatype.org/vuln/e5794172-1257-4372-9baf-7b87307a3cc9 ```
non_infrastructure
com fasterxml jackson core jackson databind marked vulnerable the dependency com fasterxml jackson core jackson databind jar is marked as having vulnerabilities by the enforcer plugin a newer version should be used which does not contain they flagged issues failure message rule org sonatype ossindex maven enforcer banvulnerabledependencies failed with message detected vulnerable components com fasterxml jackson core jackson databind jar compile a polymorphic typing issue was discovered in fasterxml jackson databind x befo subtypevalidator java in fasterxml jackson databind before mishandles de
0
19,697
13,395,488,914
IssuesEvent
2020-09-03 08:30:31
onivim/oni2
https://api.github.com/repos/onivim/oni2
closed
Can't build the frontend
A-infrastructure platform-linux
Following the instructions from https://onivim.github.io/docs/for-developers/building ```bash git clone https://github.com/onivim/oni2 cd oni2 esy install ``` Last step fails with some problem about `@esy-ocaml/substs`: ``` error unable to fetch @esy-ocaml/substs@0.0.1: source: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46 error: error: run ['shasum' '--algorithm' '1' '/tmp/oni2/_esy/default/tmp/esy-f92046/archive']: exited with 127 fetching dist: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46 error: installation error fetching @esy-ocaml/substs@0.0.1 esy: exiting due to errors above ``` I checked and I can install `@esy-ocaml/substs@0.0.1` with npm successfully. If I try to install it with `esy`, it fails as well (with the same error as above). The OS is Fedora 32.
1.0
Can't build the frontend - Following the instructions from https://onivim.github.io/docs/for-developers/building ```bash git clone https://github.com/onivim/oni2 cd oni2 esy install ``` Last step fails with some problem about `@esy-ocaml/substs`: ``` error unable to fetch @esy-ocaml/substs@0.0.1: source: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46 error: error: run ['shasum' '--algorithm' '1' '/tmp/oni2/_esy/default/tmp/esy-f92046/archive']: exited with 127 fetching dist: archive:https://registry.npmjs.org/@esy-ocaml/substs/-/substs-0.0.1.tgz#sha1:59ebdbbaedcda123fc7ed8fb2b302b7d819e9a46 error: installation error fetching @esy-ocaml/substs@0.0.1 esy: exiting due to errors above ``` I checked and I can install `@esy-ocaml/substs@0.0.1` with npm successfully. If I try to install it with `esy`, it fails as well (with the same error as above). The OS is Fedora 32.
infrastructure
can t build the frontend following the instructions from bash git clone cd esy install last step fails with some problem about esy ocaml substs error unable to fetch esy ocaml substs source archive error error run shasum algorithm tmp esy default tmp esy archive exited with fetching dist archive error installation error fetching esy ocaml substs esy exiting due to errors above i checked and i can install esy ocaml substs with npm successfully if i try to install it with esy it fails as well with the same error as above the os is fedora
1
269,400
8,435,411,584
IssuesEvent
2018-10-17 13:06:07
kubeapps/kubeapps
https://api.github.com/repos/kubeapps/kubeapps
closed
[Design/Proposal] Global runtime error handling
component/dashboard priority/important-soon size/M
Currently, when a runtime error happens in the dashboard, the UI stops rendering but "unless specifically handled" the user does not see a message that something went wrong. This was the case https://github.com/kubeapps/kubeapps/issues/635 for example: React 16 has the concept of errorBoundary that we should explore https://reactjs.org/docs/error-boundaries.html, potentially making it available as a wrapper at the root component. This component could eventually contain the efforts made here https://github.com/kubeapps/kubeapps/issues/654 The goal being: * Step 1: If a unexpected runtime error happens, we show a message to the user that `something went wrong` (I believe that we have already some markup for this that suggest to look at the console and so on) * Step 2: This component might be able to capture some specific custom errors, showing different error types and messages https://github.com/kubeapps/kubeapps/issues/654. This currently is happening at every component level, so making it generic could help. wdyt @andresmgot @prydonius ?
1.0
[Design/Proposal] Global runtime error handling - Currently, when a runtime error happens in the dashboard, the UI stops rendering but "unless specifically handled" the user does not see a message that something went wrong. This was the case https://github.com/kubeapps/kubeapps/issues/635 for example: React 16 has the concept of errorBoundary that we should explore https://reactjs.org/docs/error-boundaries.html, potentially making it available as a wrapper at the root component. This component could eventually contain the efforts made here https://github.com/kubeapps/kubeapps/issues/654 The goal being: * Step 1: If a unexpected runtime error happens, we show a message to the user that `something went wrong` (I believe that we have already some markup for this that suggest to look at the console and so on) * Step 2: This component might be able to capture some specific custom errors, showing different error types and messages https://github.com/kubeapps/kubeapps/issues/654. This currently is happening at every component level, so making it generic could help. wdyt @andresmgot @prydonius ?
non_infrastructure
global runtime error handling currently when a runtime error happens in the dashboard the ui stops rendering but unless specifically handled the user does not see a message that something went wrong this was the case for example react has the concept of errorboundary that we should explore potentially making it available as a wrapper at the root component this component could eventually contain the efforts made here the goal being step if a unexpected runtime error happens we show a message to the user that something went wrong i believe that we have already some markup for this that suggest to look at the console and so on step this component might be able to capture some specific custom errors showing different error types and messages this currently is happening at every component level so making it generic could help wdyt andresmgot prydonius
0
3,119
4,065,549,130
IssuesEvent
2016-05-26 11:44:36
pjuu/pjuu
https://api.github.com/repos/pjuu/pjuu
closed
Jessie for docker
in progress infrastructure
Convert the docker image to Jessie rather than Wheezy. There are some issues with the required libraries that need to be tested.
1.0
Jessie for docker - Convert the docker image to Jessie rather than Wheezy. There are some issues with the required libraries that need to be tested.
infrastructure
jessie for docker convert the docker image to jessie rather than wheezy there are some issues with the required libraries that need to be tested
1
13,704
10,428,005,741
IssuesEvent
2019-09-16 21:18:27
dotnet/core-setup
https://api.github.com/repos/dotnet/core-setup
closed
The tag v2.1.13-dependencies is missing
area-Infrastructure
## Steps to reproduce Go to https://github.com/dotnet/core-setup/tags ## Expected behavior The tag v2.1.13-dependencies should be present, as it was for 2.1.12: https://github.com/dotnet/core-setup/releases/tag/v2.1.12-dependencies ## Actual behavior The tag is not present. ## Environment data n/a
1.0
The tag v2.1.13-dependencies is missing - ## Steps to reproduce Go to https://github.com/dotnet/core-setup/tags ## Expected behavior The tag v2.1.13-dependencies should be present, as it was for 2.1.12: https://github.com/dotnet/core-setup/releases/tag/v2.1.12-dependencies ## Actual behavior The tag is not present. ## Environment data n/a
infrastructure
the tag dependencies is missing steps to reproduce go to expected behavior the tag dependencies should be present as it was for actual behavior the tag is not present environment data n a
1
6,655
6,547,208,340
IssuesEvent
2017-09-04 13:49:27
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
opened
io/secure_server_closing_test times out on CQ on Mac OS
area-infrastructure os-osx
FAILED: none-vm release_x64 standalone/io/secure_server_closing_test Expected: Pass Actual: Timeout https://luci-logdog.appspot.com/v/?s=dart%2Fled%2Fathom_google.com%2Fa807c9b6e6e940045dc7676a153e11294f2a9dfd32c15e063ffe80a92badc95e%2F%2B%2Fsteps%2Fvm_tests%2F0%2Fstdout
1.0
io/secure_server_closing_test times out on CQ on Mac OS - FAILED: none-vm release_x64 standalone/io/secure_server_closing_test Expected: Pass Actual: Timeout https://luci-logdog.appspot.com/v/?s=dart%2Fled%2Fathom_google.com%2Fa807c9b6e6e940045dc7676a153e11294f2a9dfd32c15e063ffe80a92badc95e%2F%2B%2Fsteps%2Fvm_tests%2F0%2Fstdout
infrastructure
io secure server closing test times out on cq on mac os failed none vm release standalone io secure server closing test expected pass actual timeout
1
92,376
8,361,666,776
IssuesEvent
2018-10-03 14:52:26
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: jepsen/3/register/majority-ring failed
A-core-kv C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/0f5a289d49105cd317a2713b71546d84c25547f6 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stressrace instead of stress and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ make stress TESTS=jepsen/3/register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m' ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=937459&tab=buildLog ``` The test failed on master: test.go:500,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: /home/agent/work/.go/bin/roachprod run teamcity-937459-jepsen-3:1-6 -- sh -c "sudo apt-get -qqy upgrade -o Dpkg::Options::='--force-confold' > /dev/null 2>&1" returned: stderr: stdout: ............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................. 1: exit status 100 2: exit status 100 3: exit status 100 4: exit status 100 5: exit status 100 6: exit status 100 Error: exit status 100 : exit status 1 test.go:500,cluster.go:890,jepsen.go:74,asm_amd64.s:573,panic.go:377,test.go:501,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: test already failed ```
1.0
roachtest: jepsen/3/register/majority-ring failed - SHA: https://github.com/cockroachdb/cockroach/commits/0f5a289d49105cd317a2713b71546d84c25547f6 Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stressrace instead of stress and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ make stress TESTS=jepsen/3/register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-stderr=false -maxtime 20m -timeout 10m' ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=937459&tab=buildLog ``` The test failed on master: test.go:500,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: /home/agent/work/.go/bin/roachprod run teamcity-937459-jepsen-3:1-6 -- sh -c "sudo apt-get -qqy upgrade -o Dpkg::Options::='--force-confold' > /dev/null 2>&1" returned: stderr: stdout: ............................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................................. 1: exit status 100 2: exit status 100 3: exit status 100 4: exit status 100 5: exit status 100 6: exit status 100 Error: exit status 100 : exit status 1 test.go:500,cluster.go:890,jepsen.go:74,asm_amd64.s:573,panic.go:377,test.go:501,cluster.go:890,jepsen.go:86,jepsen.go:125,jepsen.go:285: test already failed ```
non_infrastructure
roachtest jepsen register majority ring failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stressrace instead of stress and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach make stress tests jepsen register majority ring pkg roachtest testtimeout stressflags stderr false maxtime timeout failed test the test failed on master test go cluster go jepsen go jepsen go jepsen go home agent work go bin roachprod run teamcity jepsen sh c sudo apt get qqy upgrade o dpkg options force confold dev null returned stderr stdout exit status exit status exit status exit status exit status exit status error exit status exit status test go cluster go jepsen go asm s panic go test go cluster go jepsen go jepsen go jepsen go test already failed
0
15,190
11,402,509,526
IssuesEvent
2020-01-31 03:30:41
MindLeaps/tracker
https://api.github.com/repos/MindLeaps/tracker
closed
Update Ansible Playbooks
infrastructure
We are using an old ansible configuration a lot of it is invalid with the new ansible. examples include `sudo: true` should be `become: yes` apt get `state=installed` should be `state=present`
1.0
Update Ansible Playbooks - We are using an old ansible configuration a lot of it is invalid with the new ansible. examples include `sudo: true` should be `become: yes` apt get `state=installed` should be `state=present`
infrastructure
update ansible playbooks we are using an old ansible configuration a lot of it is invalid with the new ansible examples include sudo true should be become yes apt get state installed should be state present
1
17,573
24,229,548,653
IssuesEvent
2022-09-26 16:59:58
opensrp/web
https://api.github.com/repos/opensrp/web
closed
[FHIR Road Map] Create new packages for FHIR support
Discussion Ready for Engineering FHIR compatibility
**Work plan** - Create FHIR focused packages for the following features - [ ] Team Management --- https://github.com/OpenSRP/web/issues/552 - [ ] User Management --- https://github.com/OpenSRP/web/issues/585 - [ ] Location Management --- https://github.com/OpenSRP/web/issues/553
True
[FHIR Road Map] Create new packages for FHIR support - **Work plan** - Create FHIR focused packages for the following features - [ ] Team Management --- https://github.com/OpenSRP/web/issues/552 - [ ] User Management --- https://github.com/OpenSRP/web/issues/585 - [ ] Location Management --- https://github.com/OpenSRP/web/issues/553
non_infrastructure
create new packages for fhir support work plan create fhir focused packages for the following features team management user management location management
0
390,913
11,565,690,507
IssuesEvent
2020-02-20 10:58:49
luna/enso
https://api.github.com/repos/luna/enso
closed
Clean Up `AstExpression`
Category: Compiler Change: Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement
### Summary The file defining the old AST format is a bit of a mess. As it's going to be used as the basis for the IR in the stopgap compiler, it needs a bit of cleaning up. ### Value My sanity stays slightly more intact. ### Specification - [x] Clean up the structure of the code. - [x] Slightly alter the hierarchy of the nodes. - [x] Document the AST nodes. - [x] Add additional nodes to represent errors. - [x] Add a node to allow for tagging arbitrary nodes with additional data (as an `HList`). ### Acceptance Criteria & Test Cases - All existing tests pass.
1.0
Clean Up `AstExpression` - ### Summary The file defining the old AST format is a bit of a mess. As it's going to be used as the basis for the IR in the stopgap compiler, it needs a bit of cleaning up. ### Value My sanity stays slightly more intact. ### Specification - [x] Clean up the structure of the code. - [x] Slightly alter the hierarchy of the nodes. - [x] Document the AST nodes. - [x] Add additional nodes to represent errors. - [x] Add a node to allow for tagging arbitrary nodes with additional data (as an `HList`). ### Acceptance Criteria & Test Cases - All existing tests pass.
non_infrastructure
clean up astexpression summary the file defining the old ast format is a bit of a mess as it s going to be used as the basis for the ir in the stopgap compiler it needs a bit of cleaning up value my sanity stays slightly more intact specification clean up the structure of the code slightly alter the hierarchy of the nodes document the ast nodes add additional nodes to represent errors add a node to allow for tagging arbitrary nodes with additional data as an hlist acceptance criteria test cases all existing tests pass
0
28,847
8,214,348,773
IssuesEvent
2018-09-04 22:54:27
angular/devkit
https://api.github.com/repos/angular/devkit
closed
Purify doesn't work with NgModuleFactory in non-webpack environment
comp: angular-devkit/build-optimizer type: feature
Relates to https://github.com/angular/material2/issues/4137 . Currently purify plugin only supports hard-coded webpack regex in https://github.com/angular/devkit/blob/master/packages/angular_devkit/build_optimizer/src/purify/purify.ts#L52-L60, and it cannot work with any non-webpack tools. Since it's too obvious I don't think there're any needs for a repo of reproduction. Example input: ```javascript /** * @fileoverview This file is generated by the Angular template compiler. * Do not edit. * @suppress {suspiciousCode,uselessCode,missingProperties,missingOverride} */ /* tslint:disable */ import * as i0 from '@angular/core'; import * as i1 from '@angular/material'; import * as i2 from '@angular/cdk'; import * as i3 from '@angular/platform-browser'; import * as i4 from '@angular/common'; import * as i5 from '@angular/forms'; import * as i6 from '@angular/http'; export var MdCoreModuleNgFactory = i0.ɵcmf(i1.MdCoreModule, [], function (_l) { return i0.ɵmod([i0.ɵmpd(512, i0.ComponentFactoryResolver, i0.ɵCodegenComponentFactoryResolver, [[8, []], [3, i0.ComponentFactoryResolver], i0.NgModuleRef]), i0.ɵmpd(6144, i2.DIR_DOCUMENT, null, [i3.DOCUMENT]), i0.ɵmpd(4608, i2.Directionality, i2.Directionality, [[2, i2.DIR_DOCUMENT]]), i0.ɵmpd(4608, i2.Platform, i2.Platform, []), i0.ɵmpd(5120, i1.ScrollDispatcher, i1.ɵe, [[3, i1.ScrollDispatcher], i0.NgZone, i2.Platform]), i0.ɵmpd(5120, i1.ViewportRuler, i1.ɵc, [[3, i1.ViewportRuler], i1.ScrollDispatcher]), i0.ɵmpd(4608, i1.ScrollStrategyOptions, i1.ScrollStrategyOptions, [i1.ScrollDispatcher, i1.ViewportRuler]), i0.ɵmpd(4608, i2.MdMutationObserverFactory, i2.MdMutationObserverFactory, []), i0.ɵmpd(5120, i1.OverlayContainer, i1.ɵa, [[3, i1.OverlayContainer]]), i0.ɵmpd(4608, i1.ɵt, i1.ɵt, [i1.ViewportRuler]), i0.ɵmpd(4608, i1.Overlay, i1.Overlay, [i1.ScrollStrategyOptions, i1.OverlayContainer, i0.ComponentFactoryResolver, i1.ɵt, i0.ApplicationRef, i0.Injector, i0.NgZone]), i0.ɵmpd(4608, i4.NgLocalization, i4.NgLocaleLocalization, [i0.LOCALE_ID]), i0.ɵmpd(4608, i2.InteractivityChecker, i2.InteractivityChecker, [i2.Platform]), i0.ɵmpd(4608, i2.FocusTrapFactory, i2.FocusTrapFactory, [i2.InteractivityChecker, i2.Platform, i0.NgZone]), i0.ɵmpd(5120, i2.LiveAnnouncer, i2.LIVE_ANNOUNCER_PROVIDER_FACTORY, [[3, i2.LiveAnnouncer], [2, i2.LIVE_ANNOUNCER_ELEMENT_TOKEN], i2.Platform]), i0.ɵmpd(512, i1.CompatibilityModule, i1.CompatibilityModule, []), i0.ɵmpd(512, i2.BidiModule, i2.BidiModule, []), i0.ɵmpd(256, i1.MATERIAL_SANITY_CHECKS, true, []), i0.ɵmpd(512, i1.MdCommonModule, i1.MdCommonModule, [[2, i3.DOCUMENT], [2, i1.MATERIAL_SANITY_CHECKS]]), i0.ɵmpd(512, i1.MdLineModule, i1.MdLineModule, []), i0.ɵmpd(512, i2.PlatformModule, i2.PlatformModule, []), i0.ɵmpd(512, i1.ScrollDispatchModule, i1.ScrollDispatchModule, []), i0.ɵmpd(512, i1.MdRippleModule, i1.MdRippleModule, []), i0.ɵmpd(512, i2.ObserveContentModule, i2.ObserveContentModule, []), i0.ɵmpd(512, i2.PortalModule, i2.PortalModule, []), i0.ɵmpd(512, i1.OverlayModule, i1.OverlayModule, []), i0.ɵmpd(512, i4.CommonModule, i4.CommonModule, []), i0.ɵmpd(512, i2.A11yModule, i2.A11yModule, []), i0.ɵmpd(512, i1.MdSelectionModule, i1.MdSelectionModule, []), i0.ɵmpd(512, i1.MdOptionModule, i1.MdOptionModule, []), i0.ɵmpd(512, i1.MdCoreModule, i1.MdCoreModule, [])]); }); ```
1.0
Purify doesn't work with NgModuleFactory in non-webpack environment - Relates to https://github.com/angular/material2/issues/4137 . Currently purify plugin only supports hard-coded webpack regex in https://github.com/angular/devkit/blob/master/packages/angular_devkit/build_optimizer/src/purify/purify.ts#L52-L60, and it cannot work with any non-webpack tools. Since it's too obvious I don't think there're any needs for a repo of reproduction. Example input: ```javascript /** * @fileoverview This file is generated by the Angular template compiler. * Do not edit. * @suppress {suspiciousCode,uselessCode,missingProperties,missingOverride} */ /* tslint:disable */ import * as i0 from '@angular/core'; import * as i1 from '@angular/material'; import * as i2 from '@angular/cdk'; import * as i3 from '@angular/platform-browser'; import * as i4 from '@angular/common'; import * as i5 from '@angular/forms'; import * as i6 from '@angular/http'; export var MdCoreModuleNgFactory = i0.ɵcmf(i1.MdCoreModule, [], function (_l) { return i0.ɵmod([i0.ɵmpd(512, i0.ComponentFactoryResolver, i0.ɵCodegenComponentFactoryResolver, [[8, []], [3, i0.ComponentFactoryResolver], i0.NgModuleRef]), i0.ɵmpd(6144, i2.DIR_DOCUMENT, null, [i3.DOCUMENT]), i0.ɵmpd(4608, i2.Directionality, i2.Directionality, [[2, i2.DIR_DOCUMENT]]), i0.ɵmpd(4608, i2.Platform, i2.Platform, []), i0.ɵmpd(5120, i1.ScrollDispatcher, i1.ɵe, [[3, i1.ScrollDispatcher], i0.NgZone, i2.Platform]), i0.ɵmpd(5120, i1.ViewportRuler, i1.ɵc, [[3, i1.ViewportRuler], i1.ScrollDispatcher]), i0.ɵmpd(4608, i1.ScrollStrategyOptions, i1.ScrollStrategyOptions, [i1.ScrollDispatcher, i1.ViewportRuler]), i0.ɵmpd(4608, i2.MdMutationObserverFactory, i2.MdMutationObserverFactory, []), i0.ɵmpd(5120, i1.OverlayContainer, i1.ɵa, [[3, i1.OverlayContainer]]), i0.ɵmpd(4608, i1.ɵt, i1.ɵt, [i1.ViewportRuler]), i0.ɵmpd(4608, i1.Overlay, i1.Overlay, [i1.ScrollStrategyOptions, i1.OverlayContainer, i0.ComponentFactoryResolver, i1.ɵt, i0.ApplicationRef, i0.Injector, i0.NgZone]), i0.ɵmpd(4608, i4.NgLocalization, i4.NgLocaleLocalization, [i0.LOCALE_ID]), i0.ɵmpd(4608, i2.InteractivityChecker, i2.InteractivityChecker, [i2.Platform]), i0.ɵmpd(4608, i2.FocusTrapFactory, i2.FocusTrapFactory, [i2.InteractivityChecker, i2.Platform, i0.NgZone]), i0.ɵmpd(5120, i2.LiveAnnouncer, i2.LIVE_ANNOUNCER_PROVIDER_FACTORY, [[3, i2.LiveAnnouncer], [2, i2.LIVE_ANNOUNCER_ELEMENT_TOKEN], i2.Platform]), i0.ɵmpd(512, i1.CompatibilityModule, i1.CompatibilityModule, []), i0.ɵmpd(512, i2.BidiModule, i2.BidiModule, []), i0.ɵmpd(256, i1.MATERIAL_SANITY_CHECKS, true, []), i0.ɵmpd(512, i1.MdCommonModule, i1.MdCommonModule, [[2, i3.DOCUMENT], [2, i1.MATERIAL_SANITY_CHECKS]]), i0.ɵmpd(512, i1.MdLineModule, i1.MdLineModule, []), i0.ɵmpd(512, i2.PlatformModule, i2.PlatformModule, []), i0.ɵmpd(512, i1.ScrollDispatchModule, i1.ScrollDispatchModule, []), i0.ɵmpd(512, i1.MdRippleModule, i1.MdRippleModule, []), i0.ɵmpd(512, i2.ObserveContentModule, i2.ObserveContentModule, []), i0.ɵmpd(512, i2.PortalModule, i2.PortalModule, []), i0.ɵmpd(512, i1.OverlayModule, i1.OverlayModule, []), i0.ɵmpd(512, i4.CommonModule, i4.CommonModule, []), i0.ɵmpd(512, i2.A11yModule, i2.A11yModule, []), i0.ɵmpd(512, i1.MdSelectionModule, i1.MdSelectionModule, []), i0.ɵmpd(512, i1.MdOptionModule, i1.MdOptionModule, []), i0.ɵmpd(512, i1.MdCoreModule, i1.MdCoreModule, [])]); }); ```
non_infrastructure
purify doesn t work with ngmodulefactory in non webpack environment relates to currently purify plugin only supports hard coded webpack regex in and it cannot work with any non webpack tools since it s too obvious i don t think there re any needs for a repo of reproduction example input javascript fileoverview this file is generated by the angular template compiler do not edit suppress suspiciouscode uselesscode missingproperties missingoverride tslint disable import as from angular core import as from angular material import as from angular cdk import as from angular platform browser import as from angular common import as from angular forms import as from angular http export var mdcoremodulengfactory ɵcmf mdcoremodule function l return ɵmod ngmoduleref ɵmpd dir document null ɵmpd directionality directionality ɵmpd platform platform ɵmpd scrolldispatcher ɵe ngzone platform ɵmpd viewportruler ɵc scrolldispatcher ɵmpd scrollstrategyoptions scrollstrategyoptions ɵmpd mdmutationobserverfactory mdmutationobserverfactory ɵmpd overlaycontainer ɵa ɵmpd ɵt ɵt ɵmpd overlay overlay scrollstrategyoptions overlaycontainer componentfactoryresolver ɵt applicationref injector ngzone ɵmpd nglocalization nglocalelocalization ɵmpd interactivitychecker interactivitychecker ɵmpd focustrapfactory focustrapfactory interactivitychecker platform ngzone ɵmpd liveannouncer live announcer provider factory platform ɵmpd compatibilitymodule compatibilitymodule ɵmpd bidimodule bidimodule ɵmpd material sanity checks true ɵmpd mdcommonmodule mdcommonmodule ɵmpd mdlinemodule mdlinemodule ɵmpd platformmodule platformmodule ɵmpd scrolldispatchmodule scrolldispatchmodule ɵmpd mdripplemodule mdripplemodule ɵmpd observecontentmodule observecontentmodule ɵmpd portalmodule portalmodule ɵmpd overlaymodule overlaymodule ɵmpd commonmodule commonmodule ɵmpd ɵmpd mdselectionmodule mdselectionmodule ɵmpd mdoptionmodule mdoptionmodule ɵmpd mdcoremodule mdcoremodule
0
23,838
16,616,680,803
IssuesEvent
2021-06-02 17:37:10
google/iree
https://api.github.com/repos/google/iree
closed
Add an IWYU (include-what-you-use) presubmit check
enhancement ➕ infrastructure 🛠️ quality of life 😊
Just found that some defines doesn't match file path, and I fixed them in #3574. It would be good if we can have a check for this. File an issue for tracking this, feel free to close the issue if it's not reasonable.
1.0
Add an IWYU (include-what-you-use) presubmit check - Just found that some defines doesn't match file path, and I fixed them in #3574. It would be good if we can have a check for this. File an issue for tracking this, feel free to close the issue if it's not reasonable.
infrastructure
add an iwyu include what you use presubmit check just found that some defines doesn t match file path and i fixed them in it would be good if we can have a check for this file an issue for tracking this feel free to close the issue if it s not reasonable
1
29,749
24,236,104,345
IssuesEvent
2022-09-26 23:31:36
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
ASP.NET Core v6.0.8 source build sucessed and installed ,but enter f11 can't view source code
area-infrastructure
1, installed " aspnetcore-runtime-6.0.8-dev-win-x64.msi " package **2, already add nuget.config to the project** `<?xml version="1.0" encoding="utf-8"?> <configuration> <packageSources> `<clear` /> <add key="MyBuildOfAspNetCore" value="E:\core3\AspNetCore\artifacts\packages\Debug\Shipping\" /> <add key="NuGet.org" value="https://api.nuget.org/v3/index.json" /> </packageSources> `</configuration>`` **3 use vs 2022 create a mvc project ,enter f11 view souce code, but failed** [https://images.cnblogs.com/cnblogs_com/blogs/729358/galleries/2211331/o_220901004234_2.png](url) remark(operate system is win10) who can help me ,thanks very much !
1.0
ASP.NET Core v6.0.8 source build sucessed and installed ,but enter f11 can't view source code - 1, installed " aspnetcore-runtime-6.0.8-dev-win-x64.msi " package **2, already add nuget.config to the project** `<?xml version="1.0" encoding="utf-8"?> <configuration> <packageSources> `<clear` /> <add key="MyBuildOfAspNetCore" value="E:\core3\AspNetCore\artifacts\packages\Debug\Shipping\" /> <add key="NuGet.org" value="https://api.nuget.org/v3/index.json" /> </packageSources> `</configuration>`` **3 use vs 2022 create a mvc project ,enter f11 view souce code, but failed** [https://images.cnblogs.com/cnblogs_com/blogs/729358/galleries/2211331/o_220901004234_2.png](url) remark(operate system is win10) who can help me ,thanks very much !
infrastructure
asp net core source build sucessed and installed but enter can t view source code installed aspnetcore runtime dev win msi package already add nuget config to the project use vs create a mvc project enter view souce code but failed url remark operate system is who can help me thanks very much
1
16,478
11,986,205,054
IssuesEvent
2020-04-07 18:52:56
enarx/enarx
https://api.github.com/repos/enarx/enarx
closed
Enable SGX revision 28 on the SGX kernels
infrastructure intel sgx
Right now they're using the latest tag of SGX, revision 26. Enable revision 28 so we can provide testing feedback to the upstreaming effort; upload these kernel source packages into the Enarx copr.
1.0
Enable SGX revision 28 on the SGX kernels - Right now they're using the latest tag of SGX, revision 26. Enable revision 28 so we can provide testing feedback to the upstreaming effort; upload these kernel source packages into the Enarx copr.
infrastructure
enable sgx revision on the sgx kernels right now they re using the latest tag of sgx revision enable revision so we can provide testing feedback to the upstreaming effort upload these kernel source packages into the enarx copr
1
4,318
4,973,764,270
IssuesEvent
2016-12-06 02:36:37
archco/cosmos-css
https://api.github.com/repos/archco/cosmos-css
opened
Update 편의를 위해 src 폴더구조를 변경
enhancement infrastructure
`src/`안에 폴더를 - base source : `sass/` `js/` - custom : `custom/` 으로 나눈다.
1.0
Update 편의를 위해 src 폴더구조를 변경 - `src/`안에 폴더를 - base source : `sass/` `js/` - custom : `custom/` 으로 나눈다.
infrastructure
update 편의를 위해 src 폴더구조를 변경 src 안에 폴더를 base source sass js custom custom 으로 나눈다
1
128
2,533,509,844
IssuesEvent
2015-01-24 00:11:23
RobDixonIII/Bloom
https://api.github.com/repos/RobDixonIII/Bloom
closed
Implement Dynamic Tabs with Prism Modules
feature infrastructure
Prism uses an regioned injection approach for modules that loads the UI upon application start. This can make dynamic tabs problematic since they are not present at that time. Create a way to dynamically inject Prism modules into tabs at runtime.
1.0
Implement Dynamic Tabs with Prism Modules - Prism uses an regioned injection approach for modules that loads the UI upon application start. This can make dynamic tabs problematic since they are not present at that time. Create a way to dynamically inject Prism modules into tabs at runtime.
infrastructure
implement dynamic tabs with prism modules prism uses an regioned injection approach for modules that loads the ui upon application start this can make dynamic tabs problematic since they are not present at that time create a way to dynamically inject prism modules into tabs at runtime
1
26,250
19,826,329,157
IssuesEvent
2022-01-20 07:04:58
google/site-kit-wp
https://api.github.com/repos/google/site-kit-wp
opened
Fix VRT tests
P0 Type: Infrastructure
## Bug Description Our VRT tests recently started failing for what seems to be external factors. The last update to our reference images was 9 days ago in https://github.com/google/site-kit-wp/pull/4587 which only updated 6 images and we currently have 42 failures, so this wouldn't have been responsible. The most likely cause it seems is that the Google Sans web font has changed as the diffs are highlighting changes in primary headings which are the main places this font is used. Roboto use does not raise any differences so it can't be a general web font issue. ## Steps to reproduce * Run VRT tests on develop * Run VRT tests on main ## Screenshots ![image](https://user-images.githubusercontent.com/1621608/150288499-3e665cdf-6bfa-4ea9-89c2-55af6c8b3774.png) ![image](https://user-images.githubusercontent.com/1621608/150289406-7bab1ffb-35ff-4b03-b0c1-d8718e09ca67.png) ## Additional Context - Seems to have started on 2022-01-19 --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * Visual regression references should be updated ## Implementation Brief * Run VRT tests, review and approve the changes ### Test Coverage * N/A ## QA Brief * <!-- One or more bullet points for how to test that the issue has been resolved. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
1.0
Fix VRT tests - ## Bug Description Our VRT tests recently started failing for what seems to be external factors. The last update to our reference images was 9 days ago in https://github.com/google/site-kit-wp/pull/4587 which only updated 6 images and we currently have 42 failures, so this wouldn't have been responsible. The most likely cause it seems is that the Google Sans web font has changed as the diffs are highlighting changes in primary headings which are the main places this font is used. Roboto use does not raise any differences so it can't be a general web font issue. ## Steps to reproduce * Run VRT tests on develop * Run VRT tests on main ## Screenshots ![image](https://user-images.githubusercontent.com/1621608/150288499-3e665cdf-6bfa-4ea9-89c2-55af6c8b3774.png) ![image](https://user-images.githubusercontent.com/1621608/150289406-7bab1ffb-35ff-4b03-b0c1-d8718e09ca67.png) ## Additional Context - Seems to have started on 2022-01-19 --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * Visual regression references should be updated ## Implementation Brief * Run VRT tests, review and approve the changes ### Test Coverage * N/A ## QA Brief * <!-- One or more bullet points for how to test that the issue has been resolved. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
infrastructure
fix vrt tests bug description our vrt tests recently started failing for what seems to be external factors the last update to our reference images was days ago in which only updated images and we currently have failures so this wouldn t have been responsible the most likely cause it seems is that the google sans web font has changed as the diffs are highlighting changes in primary headings which are the main places this font is used roboto use does not raise any differences so it can t be a general web font issue steps to reproduce run vrt tests on develop run vrt tests on main screenshots additional context seems to have started on do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria visual regression references should be updated implementation brief run vrt tests review and approve the changes test coverage n a qa brief changelog entry
1
121,072
10,149,315,419
IssuesEvent
2019-08-05 14:58:29
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
opened
Extended system test failure: SharedClassesAPI_0
test failure
Failure link ------------ https://ci.adoptopenjdk.net/job/Test_openjdk11_j9_extended.system_x86-64_mac/7 Failing thread reference can be found in the results tar.gz, here: openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\execute\execute.stdout and the error message itself can be found in the output for that thread, here: C:\Users\ADAMFarley\Downloads\systemtest_test_output(1).tar\systemtest_test_output(1)\openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\results\8.WL3.stderr Optional info ------------- - intermittent failure: no - regression or new test: regression - if regression, what are the last passing / first failing public SHAs (OpenJ9, OMR, JCL) : Unknown. Failure output -------------- JVMSHRC659E An error has occurred while opening shared memory JVMSHRC336E Port layer error code = -174 JVMSHRC337E Platform error message: Invalid argument JVMSHRC026E Cannot create cache of requested size: Please check your SHMMAX and SHMMIN settings JVMSHRC663I Error recovery: destroyed semaphore set with id=54067200 associated with shared class cache. JVMJ9VM015W Initialization error for library j9shr29(11): JVMJ9VM009E J9VMDllMain failed Error: Could not create the Java Virtual Machine. Error: A fatal exception has occurred. Program will exit.
1.0
Extended system test failure: SharedClassesAPI_0 - Failure link ------------ https://ci.adoptopenjdk.net/job/Test_openjdk11_j9_extended.system_x86-64_mac/7 Failing thread reference can be found in the results tar.gz, here: openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\execute\execute.stdout and the error message itself can be found in the output for that thread, here: C:\Users\ADAMFarley\Downloads\systemtest_test_output(1).tar\systemtest_test_output(1)\openjdk-tests\TestConfig\test_output_15648593132736\SharedClassesAPI_0\20190803-205835-SharedClassesAPI\results\8.WL3.stderr Optional info ------------- - intermittent failure: no - regression or new test: regression - if regression, what are the last passing / first failing public SHAs (OpenJ9, OMR, JCL) : Unknown. Failure output -------------- JVMSHRC659E An error has occurred while opening shared memory JVMSHRC336E Port layer error code = -174 JVMSHRC337E Platform error message: Invalid argument JVMSHRC026E Cannot create cache of requested size: Please check your SHMMAX and SHMMIN settings JVMSHRC663I Error recovery: destroyed semaphore set with id=54067200 associated with shared class cache. JVMJ9VM015W Initialization error for library j9shr29(11): JVMJ9VM009E J9VMDllMain failed Error: Could not create the Java Virtual Machine. Error: A fatal exception has occurred. Program will exit.
non_infrastructure
extended system test failure sharedclassesapi failure link failing thread reference can be found in the results tar gz here openjdk tests testconfig test output sharedclassesapi sharedclassesapi execute execute stdout and the error message itself can be found in the output for that thread here c users adamfarley downloads systemtest test output tar systemtest test output openjdk tests testconfig test output sharedclassesapi sharedclassesapi results stderr optional info intermittent failure no regression or new test regression if regression what are the last passing first failing public shas omr jcl unknown failure output an error has occurred while opening shared memory port layer error code platform error message invalid argument cannot create cache of requested size please check your shmmax and shmmin settings error recovery destroyed semaphore set with id associated with shared class cache initialization error for library failed error could not create the java virtual machine error a fatal exception has occurred program will exit
0
17,349
12,305,447,557
IssuesEvent
2020-05-11 22:33:22
18F/tts-tech-portfolio
https://api.github.com/repos/18F/tts-tech-portfolio
closed
Develop Process to Investigate new tools and technologies
epic: acquisition procurement and budgeting epic: policy and guidance epic: software and infrastructure groomed: accepted
## Background information There are new technologies, tools, etc that come to my attention either through emails, conferences, other people's suggestions, etc -- and I think they might have value in our inventory now or later -- but I am unsure if looking into them would be valuable or if I do consider them valuable -- what to do with that evaluation. As of right now, the only guidance there is for when someone wants to use a new tool or technology is to submit a request to IT Standards. This isn't the best route because it bypasses the Tech Portfolio and we lose sight on what is being requested, if something is similar to another thing or if there is budget for the request. This differs from https://github.com/18F/tts-tech-portfolio/issues/339 because that is more about analyzing a requirement once one exist; where this is proposing an idea for something to become a requirement in the first place. ## User stories - As someone in the Tech Portfolio who has a good idea, I would like to be able to explore that good idea in a structured way. - As someone in the Tech Portfolio who thinks a tool would be useful TTS wide, I don't have a way of knowing if that tool has already been proposed and what the result of that proposal was. - As someone in the Tech Portfolio who plans on looking into a new tool, I would like a way to communicate the activity with other team members. - As someone in the Tech Portfolio who thinks of a way of improving the organization through the use of a tool, I would like to know where that falls within the priorities of the portfolio. ## Implementation - [ ] Develop framework (thinking concept, discovery, proposal (https://github.com/18F/tts-tech-portfolio/issues/339), decision, prototype, deploy) - [ ] determine how the Emerging Technology office in OGP fits in - [ ] consider vendor pitch days ## Acceptance criteria: - There is a structured way for members of the Tech Portfolio to present ideas and for them to get prioritized within our workload and a record is kept for historical reference
1.0
Develop Process to Investigate new tools and technologies - ## Background information There are new technologies, tools, etc that come to my attention either through emails, conferences, other people's suggestions, etc -- and I think they might have value in our inventory now or later -- but I am unsure if looking into them would be valuable or if I do consider them valuable -- what to do with that evaluation. As of right now, the only guidance there is for when someone wants to use a new tool or technology is to submit a request to IT Standards. This isn't the best route because it bypasses the Tech Portfolio and we lose sight on what is being requested, if something is similar to another thing or if there is budget for the request. This differs from https://github.com/18F/tts-tech-portfolio/issues/339 because that is more about analyzing a requirement once one exist; where this is proposing an idea for something to become a requirement in the first place. ## User stories - As someone in the Tech Portfolio who has a good idea, I would like to be able to explore that good idea in a structured way. - As someone in the Tech Portfolio who thinks a tool would be useful TTS wide, I don't have a way of knowing if that tool has already been proposed and what the result of that proposal was. - As someone in the Tech Portfolio who plans on looking into a new tool, I would like a way to communicate the activity with other team members. - As someone in the Tech Portfolio who thinks of a way of improving the organization through the use of a tool, I would like to know where that falls within the priorities of the portfolio. ## Implementation - [ ] Develop framework (thinking concept, discovery, proposal (https://github.com/18F/tts-tech-portfolio/issues/339), decision, prototype, deploy) - [ ] determine how the Emerging Technology office in OGP fits in - [ ] consider vendor pitch days ## Acceptance criteria: - There is a structured way for members of the Tech Portfolio to present ideas and for them to get prioritized within our workload and a record is kept for historical reference
infrastructure
develop process to investigate new tools and technologies background information there are new technologies tools etc that come to my attention either through emails conferences other people s suggestions etc and i think they might have value in our inventory now or later but i am unsure if looking into them would be valuable or if i do consider them valuable what to do with that evaluation as of right now the only guidance there is for when someone wants to use a new tool or technology is to submit a request to it standards this isn t the best route because it bypasses the tech portfolio and we lose sight on what is being requested if something is similar to another thing or if there is budget for the request this differs from because that is more about analyzing a requirement once one exist where this is proposing an idea for something to become a requirement in the first place user stories as someone in the tech portfolio who has a good idea i would like to be able to explore that good idea in a structured way as someone in the tech portfolio who thinks a tool would be useful tts wide i don t have a way of knowing if that tool has already been proposed and what the result of that proposal was as someone in the tech portfolio who plans on looking into a new tool i would like a way to communicate the activity with other team members as someone in the tech portfolio who thinks of a way of improving the organization through the use of a tool i would like to know where that falls within the priorities of the portfolio implementation develop framework thinking concept discovery proposal decision prototype deploy determine how the emerging technology office in ogp fits in consider vendor pitch days acceptance criteria there is a structured way for members of the tech portfolio to present ideas and for them to get prioritized within our workload and a record is kept for historical reference
1
5,379
5,627,511,639
IssuesEvent
2017-04-05 01:50:24
amnh-library/API-Portal
https://api.github.com/repos/amnh-library/API-Portal
opened
Change user directories in dev
infrastructure
Move all apps & code from mcaprio user in dev to apiproject user
1.0
Change user directories in dev - Move all apps & code from mcaprio user in dev to apiproject user
infrastructure
change user directories in dev move all apps code from mcaprio user in dev to apiproject user
1
149,800
5,726,210,990
IssuesEvent
2017-04-20 18:24:41
screwdriver-cd/screwdriver
https://api.github.com/repos/screwdriver-cd/screwdriver
opened
Publish template should show the version
change/patch component/UX priority/medium
The [template-main](https://github.com/screwdriver-cd/template-main) works fine to publish template, but it does not say what version it just publishes. https://cd.screwdriver.cd/pipelines/152/builds/3558 We should at least show the VERSION in the log.
1.0
Publish template should show the version - The [template-main](https://github.com/screwdriver-cd/template-main) works fine to publish template, but it does not say what version it just publishes. https://cd.screwdriver.cd/pipelines/152/builds/3558 We should at least show the VERSION in the log.
non_infrastructure
publish template should show the version the works fine to publish template but it does not say what version it just publishes we should at least show the version in the log
0
28,097
22,959,455,363
IssuesEvent
2022-07-19 14:17:24
IBM-Cloud/terraform-provider-ibm
https://api.github.com/repos/IBM-Cloud/terraform-provider-ibm
closed
problems with IPs reservations
service/VPC Infrastructure service/Resource Management
<!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform IBM Provider Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> $ terraform version Terraform v1.2.2 on linux_amd64 + provider registry.terraform.io/ibm-cloud/ibm v1.42.0 ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * ibm_is_subnet_reserved_ip ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation. ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please share a link to the ZIP file. provider "ibm" { region = "eu-de" } data "ibm_is_image" "example" { name = "ibm-debian-11-3-minimal-amd64-1" } resource "ibm_is_vpc" "example" { name = "example-vpc" } resource "ibm_is_vpc_address_prefix" "example" { name = "example-address-prefix" vpc = ibm_is_vpc.example.id zone = "eu-de-1" cidr = "10.240.0.0/24" } resource "ibm_is_subnet" "example" { name = "example-subnet" vpc = ibm_is_vpc.example.id zone = "eu-de-1" ipv4_cidr_block = "10.240.0.0/24" depends_on = [ibm_is_vpc_address_prefix.example] } resource "ibm_is_subnet_reserved_ip" "example" { name = "example-reserved-ip" subnet = ibm_is_subnet.example.id } resource "ibm_is_ssh_key" "example" { name = "example-ssh" public_key = "ssh-rsa ...." } resource "ibm_is_instance" "example1" { image = data.ibm_is_image.example.id name = "example-instance-1-reserved-ip" profile = "bx2-2x8" primary_network_interface { name = "eth0" subnet = ibm_is_subnet.example.id primary_ip { reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip } } vpc = ibm_is_vpc.example.id zone = "eu-de-1" keys = [ibm_is_ssh_key.example.id] } resource "ibm_is_instance" "example2" { image = data.ibm_is_image.example.id name = "example-instance-2-reserved-ip" profile = "bx2-2x8" primary_network_interface { name = "eth0" subnet = ibm_is_subnet.example.id primary_ip { reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip } } vpc = ibm_is_vpc.example.id zone = "eu-de-1" keys = [ibm_is_ssh_key.example.id] } ``` ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html). ---> ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior <!--- What should have happened? ---> It should have failed since the IP should be unique for each VSI. ### Actual Behavior <!--- What actually happened? ---> ![image](https://user-images.githubusercontent.com/108330420/176244496-9341156e-35c1-4bfc-babd-b02f87272cc5.png) ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` $ terraform apply data.ibm_is_image.example: Reading... ibm_is_ssh_key.example: Refreshing state... [id=r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6] ibm_is_vpc.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78] data.ibm_is_image.example: Read complete after 1s [id=r010-b6c658a8-901c-4a6c-802f-f3488a114e28] ibm_is_vpc_address_prefix.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78/r010-c9f48a74-16d7-447f-af80-718e7b8778dd] ibm_is_subnet.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622] ibm_is_subnet_reserved_ip.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622/02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c] Terraform used the selected providers to generate the following execution plan. Resource actions are indicated with the following symbols:   + create Terraform will perform the following actions:   # ibm_is_instance.example1 will be created   + resource "ibm_is_instance" "example1" {       + availability_policy_host_failure  = (known after apply)       + bandwidth                         = (known after apply)       + crn                               = (known after apply)       + default_trusted_profile_auto_link = (known after apply)       + disks                             = (known after apply)       + force_action                      = false       + gpu                               = (known after apply)       + id                                = (known after apply)       + image                             = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"       + keys                              = [           + "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",         ]       + memory                            = (known after apply)       + metadata_service_enabled          = (known after apply)       + name                              = "example-instance-1-reserved-ip"       + placement_target                  = (known after apply)       + profile                           = "bx2-2x8"       + resource_controller_url           = (known after apply)       + resource_crn                      = (known after apply)       + resource_group                    = (known after apply)       + resource_group_name               = (known after apply)       + resource_name                     = (known after apply)       + resource_status                   = (known after apply)       + status                            = (known after apply)       + status_reasons                    = (known after apply)       + tags                              = (known after apply)       + total_network_bandwidth           = (known after apply)       + total_volume_bandwidth            = (known after apply)       + vcpu                              = (known after apply)       + volume_attachments                = (known after apply)       + vpc                               = "r010-da0ba964-30b9-4505-b783-ed248b065f78"       + wait_before_delete                = true       + zone                              = "eu-de-1"       + boot_volume {           + encryption = (known after apply)           + iops       = (known after apply)           + name       = (known after apply)           + profile    = (known after apply)           + size       = (known after apply)           + snapshot   = (known after apply)           + volume_id  = (known after apply)         }       + network_interfaces {           + allow_ip_spoofing    = (known after apply)           + id                   = (known after apply)           + name                 = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = (known after apply)           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = (known after apply)               + resource_type = (known after apply)             }         }       + primary_network_interface {           + allow_ip_spoofing    = false           + id                   = (known after apply)           + name                 = "eth0"           + port_speed           = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = "02b7-461d844c-4701-463c-b9fe-92240a753622"           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"               + resource_type = (known after apply)             }         }     }   # ibm_is_instance.example2 will be created   + resource "ibm_is_instance" "example2" {       + availability_policy_host_failure  = (known after apply)       + bandwidth                         = (known after apply)       + crn                               = (known after apply)       + default_trusted_profile_auto_link = (known after apply)       + disks                             = (known after apply)       + force_action                      = false       + gpu                               = (known after apply)       + id                                = (known after apply)       + image                             = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"       + keys                              = [           + "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",         ]       + memory                            = (known after apply)       + metadata_service_enabled          = (known after apply)       + name                              = "example-instance-2-reserved-ip"       + placement_target                  = (known after apply)       + profile                           = "bx2-2x8"       + resource_controller_url           = (known after apply)       + resource_crn                      = (known after apply)       + resource_group                    = (known after apply)       + resource_group_name               = (known after apply)       + resource_name                     = (known after apply)       + resource_status                   = (known after apply)       + status                            = (known after apply)       + status_reasons                    = (known after apply)       + tags                              = (known after apply)       + total_network_bandwidth           = (known after apply)       + total_volume_bandwidth            = (known after apply)       + vcpu                              = (known after apply)       + volume_attachments                = (known after apply)       + vpc                               = "r010-da0ba964-30b9-4505-b783-ed248b065f78"       + wait_before_delete                = true       + zone                              = "eu-de-1"       + boot_volume {           + encryption = (known after apply)           + iops       = (known after apply)           + name       = (known after apply)           + profile    = (known after apply)           + size       = (known after apply)           + snapshot   = (known after apply)           + volume_id  = (known after apply)         }       + network_interfaces {           + allow_ip_spoofing    = (known after apply)           + id                   = (known after apply)           + name                 = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = (known after apply)           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = (known after apply)               + resource_type = (known after apply)             }         }       + primary_network_interface {           + allow_ip_spoofing    = false           + id                   = (known after apply)           + name                 = "eth0"           + port_speed           = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = "02b7-461d844c-4701-463c-b9fe-92240a753622"           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"               + resource_type = (known after apply)             }         }     } Plan: 2 to add, 0 to change, 0 to destroy. Do you want to perform these actions?   Terraform will perform the actions described above.   Only 'yes' will be accepted to approve.   Enter a value: yes ibm_is_instance.example1: Creating... ibm_is_instance.example2: Creating... ibm_is_instance.example1: Still creating... [10s elapsed] ibm_is_instance.example2: Still creating... [10s elapsed] ibm_is_instance.example1: Still creating... [20s elapsed] ibm_is_instance.example2: Still creating... [20s elapsed] ibm_is_instance.example2: Creation complete after 27s [id=02b7_cba53a91-2924-40fd-993e-af42fda2b98b] ibm_is_instance.example1: Creation complete after 29s [id=02b7_7b407a4f-d951-4ac6-84ed-609bdb36f2e9] Apply complete! Resources: 2 added, 0 changed, 0 destroyed. moby@f2898e894286:~/cloud-practice/internal/extra/vm-ibm-exam ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: ---> * #0000
1.0
problems with IPs reservations - <!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform CLI and Terraform IBM Provider Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> $ terraform version Terraform v1.2.2 on linux_amd64 + provider registry.terraform.io/ibm-cloud/ibm v1.42.0 ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * ibm_is_subnet_reserved_ip ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> Please include all Terraform configurations required to reproduce the bug. Bug reports without a functional reproduction may be closed without investigation. ```hcl # Copy-paste your Terraform configurations here - for large Terraform configs, # please share a link to the ZIP file. provider "ibm" { region = "eu-de" } data "ibm_is_image" "example" { name = "ibm-debian-11-3-minimal-amd64-1" } resource "ibm_is_vpc" "example" { name = "example-vpc" } resource "ibm_is_vpc_address_prefix" "example" { name = "example-address-prefix" vpc = ibm_is_vpc.example.id zone = "eu-de-1" cidr = "10.240.0.0/24" } resource "ibm_is_subnet" "example" { name = "example-subnet" vpc = ibm_is_vpc.example.id zone = "eu-de-1" ipv4_cidr_block = "10.240.0.0/24" depends_on = [ibm_is_vpc_address_prefix.example] } resource "ibm_is_subnet_reserved_ip" "example" { name = "example-reserved-ip" subnet = ibm_is_subnet.example.id } resource "ibm_is_ssh_key" "example" { name = "example-ssh" public_key = "ssh-rsa ...." } resource "ibm_is_instance" "example1" { image = data.ibm_is_image.example.id name = "example-instance-1-reserved-ip" profile = "bx2-2x8" primary_network_interface { name = "eth0" subnet = ibm_is_subnet.example.id primary_ip { reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip } } vpc = ibm_is_vpc.example.id zone = "eu-de-1" keys = [ibm_is_ssh_key.example.id] } resource "ibm_is_instance" "example2" { image = data.ibm_is_image.example.id name = "example-instance-2-reserved-ip" profile = "bx2-2x8" primary_network_interface { name = "eth0" subnet = ibm_is_subnet.example.id primary_ip { reserved_ip = ibm_is_subnet_reserved_ip.example.reserved_ip } } vpc = ibm_is_vpc.example.id zone = "eu-de-1" keys = [ibm_is_ssh_key.example.id] } ``` ### Debug Output <!--- Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist. To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html). ---> ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior <!--- What should have happened? ---> It should have failed since the IP should be unique for each VSI. ### Actual Behavior <!--- What actually happened? ---> ![image](https://user-images.githubusercontent.com/108330420/176244496-9341156e-35c1-4bfc-babd-b02f87272cc5.png) ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` $ terraform apply data.ibm_is_image.example: Reading... ibm_is_ssh_key.example: Refreshing state... [id=r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6] ibm_is_vpc.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78] data.ibm_is_image.example: Read complete after 1s [id=r010-b6c658a8-901c-4a6c-802f-f3488a114e28] ibm_is_vpc_address_prefix.example: Refreshing state... [id=r010-da0ba964-30b9-4505-b783-ed248b065f78/r010-c9f48a74-16d7-447f-af80-718e7b8778dd] ibm_is_subnet.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622] ibm_is_subnet_reserved_ip.example: Refreshing state... [id=02b7-461d844c-4701-463c-b9fe-92240a753622/02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c] Terraform used the selected providers to generate the following execution plan. Resource actions are indicated with the following symbols:   + create Terraform will perform the following actions:   # ibm_is_instance.example1 will be created   + resource "ibm_is_instance" "example1" {       + availability_policy_host_failure  = (known after apply)       + bandwidth                         = (known after apply)       + crn                               = (known after apply)       + default_trusted_profile_auto_link = (known after apply)       + disks                             = (known after apply)       + force_action                      = false       + gpu                               = (known after apply)       + id                                = (known after apply)       + image                             = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"       + keys                              = [           + "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",         ]       + memory                            = (known after apply)       + metadata_service_enabled          = (known after apply)       + name                              = "example-instance-1-reserved-ip"       + placement_target                  = (known after apply)       + profile                           = "bx2-2x8"       + resource_controller_url           = (known after apply)       + resource_crn                      = (known after apply)       + resource_group                    = (known after apply)       + resource_group_name               = (known after apply)       + resource_name                     = (known after apply)       + resource_status                   = (known after apply)       + status                            = (known after apply)       + status_reasons                    = (known after apply)       + tags                              = (known after apply)       + total_network_bandwidth           = (known after apply)       + total_volume_bandwidth            = (known after apply)       + vcpu                              = (known after apply)       + volume_attachments                = (known after apply)       + vpc                               = "r010-da0ba964-30b9-4505-b783-ed248b065f78"       + wait_before_delete                = true       + zone                              = "eu-de-1"       + boot_volume {           + encryption = (known after apply)           + iops       = (known after apply)           + name       = (known after apply)           + profile    = (known after apply)           + size       = (known after apply)           + snapshot   = (known after apply)           + volume_id  = (known after apply)         }       + network_interfaces {           + allow_ip_spoofing    = (known after apply)           + id                   = (known after apply)           + name                 = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = (known after apply)           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = (known after apply)               + resource_type = (known after apply)             }         }       + primary_network_interface {           + allow_ip_spoofing    = false           + id                   = (known after apply)           + name                 = "eth0"           + port_speed           = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = "02b7-461d844c-4701-463c-b9fe-92240a753622"           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"               + resource_type = (known after apply)             }         }     }   # ibm_is_instance.example2 will be created   + resource "ibm_is_instance" "example2" {       + availability_policy_host_failure  = (known after apply)       + bandwidth                         = (known after apply)       + crn                               = (known after apply)       + default_trusted_profile_auto_link = (known after apply)       + disks                             = (known after apply)       + force_action                      = false       + gpu                               = (known after apply)       + id                                = (known after apply)       + image                             = "r010-b6c658a8-901c-4a6c-802f-f3488a114e28"       + keys                              = [           + "r010-03ee12ab-eda5-4e0c-901e-61121d7df6a6",         ]       + memory                            = (known after apply)       + metadata_service_enabled          = (known after apply)       + name                              = "example-instance-2-reserved-ip"       + placement_target                  = (known after apply)       + profile                           = "bx2-2x8"       + resource_controller_url           = (known after apply)       + resource_crn                      = (known after apply)       + resource_group                    = (known after apply)       + resource_group_name               = (known after apply)       + resource_name                     = (known after apply)       + resource_status                   = (known after apply)       + status                            = (known after apply)       + status_reasons                    = (known after apply)       + tags                              = (known after apply)       + total_network_bandwidth           = (known after apply)       + total_volume_bandwidth            = (known after apply)       + vcpu                              = (known after apply)       + volume_attachments                = (known after apply)       + vpc                               = "r010-da0ba964-30b9-4505-b783-ed248b065f78"       + wait_before_delete                = true       + zone                              = "eu-de-1"       + boot_volume {           + encryption = (known after apply)           + iops       = (known after apply)           + name       = (known after apply)           + profile    = (known after apply)           + size       = (known after apply)           + snapshot   = (known after apply)           + volume_id  = (known after apply)         }       + network_interfaces {           + allow_ip_spoofing    = (known after apply)           + id                   = (known after apply)           + name                 = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = (known after apply)           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = (known after apply)               + resource_type = (known after apply)             }         }       + primary_network_interface {           + allow_ip_spoofing    = false           + id                   = (known after apply)           + name                 = "eth0"           + port_speed           = (known after apply)           + primary_ipv4_address = (known after apply)           + security_groups      = (known after apply)           + subnet               = "02b7-461d844c-4701-463c-b9fe-92240a753622"           + primary_ip {               + address       = (known after apply)               + auto_delete   = (known after apply)               + href          = (known after apply)               + name          = (known after apply)               + reserved_ip   = "02b7-31e1fc79-3ee2-40af-8d45-bb91dd255a9c"               + resource_type = (known after apply)             }         }     } Plan: 2 to add, 0 to change, 0 to destroy. Do you want to perform these actions?   Terraform will perform the actions described above.   Only 'yes' will be accepted to approve.   Enter a value: yes ibm_is_instance.example1: Creating... ibm_is_instance.example2: Creating... ibm_is_instance.example1: Still creating... [10s elapsed] ibm_is_instance.example2: Still creating... [10s elapsed] ibm_is_instance.example1: Still creating... [20s elapsed] ibm_is_instance.example2: Still creating... [20s elapsed] ibm_is_instance.example2: Creation complete after 27s [id=02b7_cba53a91-2924-40fd-993e-af42fda2b98b] ibm_is_instance.example1: Creation complete after 29s [id=02b7_7b407a4f-d951-4ac6-84ed-609bdb36f2e9] Apply complete! Resources: 2 added, 0 changed, 0 destroyed. moby@f2898e894286:~/cloud-practice/internal/extra/vm-ibm-exam ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: ---> * #0000
infrastructure
problems with ips reservations please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform ibm provider version terraform version terraform on linux provider registry terraform io ibm cloud ibm affected resource s ibm is subnet reserved ip terraform configuration files please include all terraform configurations required to reproduce the bug bug reports without a functional reproduction may be closed without investigation hcl copy paste your terraform configurations here for large terraform configs please share a link to the zip file provider ibm region eu de data ibm is image example name ibm debian minimal resource ibm is vpc example name example vpc resource ibm is vpc address prefix example name example address prefix vpc ibm is vpc example id zone eu de cidr resource ibm is subnet example name example subnet vpc ibm is vpc example id zone eu de cidr block depends on resource ibm is subnet reserved ip example name example reserved ip subnet ibm is subnet example id resource ibm is ssh key example name example ssh public key ssh rsa resource ibm is instance image data ibm is image example id name example instance reserved ip profile primary network interface name subnet ibm is subnet example id primary ip reserved ip ibm is subnet reserved ip example reserved ip vpc ibm is vpc example id zone eu de keys resource ibm is instance image data ibm is image example id name example instance reserved ip profile primary network interface name subnet ibm is subnet example id primary ip reserved ip ibm is subnet reserved ip example reserved ip vpc ibm is vpc example id zone eu de keys debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the panic output expected behavior it should have failed since the ip should be unique for each vsi actual behavior steps to reproduce terraform apply terraform apply data ibm is image example reading ibm is ssh key example refreshing state ibm is vpc example refreshing state data ibm is image example read complete after ibm is vpc address prefix example refreshing state ibm is subnet example refreshing state ibm is subnet reserved ip example refreshing state terraform used the selected providers to generate the following execution plan resource actions are indicated with the following symbols   create terraform will perform the following actions   ibm is instance will be created   resource ibm is instance       availability policy host failure   known after apply       bandwidth                         known after apply       crn                               known after apply       default trusted profile auto link known after apply       disks                             known after apply       force action                       false       gpu                               known after apply       id                                 known after apply       image                                   keys                                                       memory                             known after apply       metadata service enabled           known after apply       name                               example instance reserved ip       placement target                   known after apply       profile                                 resource controller url           known after apply       resource crn                       known after apply       resource group                     known after apply       resource group name               known after apply       resource name                     known after apply       resource status                   known after apply       status                             known after apply       status reasons                     known after apply       tags                               known after apply       total network bandwidth           known after apply       total volume bandwidth             known after apply       vcpu                               known after apply       volume attachments                 known after apply       vpc                                     wait before delete                 true       zone                               eu de       boot volume           encryption known after apply           iops       known after apply           name       known after apply           profile     known after apply           size       known after apply           snapshot   known after apply           volume id   known after apply               network interfaces           allow ip spoofing     known after apply           id                   known after apply           name                 known after apply           primary address known after apply           security groups       known after apply           subnet               known after apply           primary ip               address       known after apply               auto delete   known after apply               href           known after apply               name           known after apply               reserved ip   known after apply               resource type known after apply                           primary network interface           allow ip spoofing     false           id                   known after apply           name                           port speed           known after apply           primary address known after apply           security groups       known after apply           subnet                         primary ip               address       known after apply               auto delete   known after apply               href           known after apply               name           known after apply               reserved ip                 resource type known after apply                           ibm is instance will be created   resource ibm is instance       availability policy host failure   known after apply       bandwidth                         known after apply       crn                               known after apply       default trusted profile auto link known after apply       disks                             known after apply       force action                       false       gpu                               known after apply       id                                 known after apply       image                                   keys                                                       memory                             known after apply       metadata service enabled           known after apply       name                               example instance reserved ip       placement target                   known after apply       profile                                 resource controller url           known after apply       resource crn                       known after apply       resource group                     known after apply       resource group name               known after apply       resource name                     known after apply       resource status                   known after apply       status                             known after apply       status reasons                     known after apply       tags                               known after apply       total network bandwidth           known after apply       total volume bandwidth             known after apply       vcpu                               known after apply       volume attachments                 known after apply       vpc                                     wait before delete                 true       zone                               eu de       boot volume           encryption known after apply           iops       known after apply           name       known after apply           profile     known after apply           size       known after apply           snapshot   known after apply           volume id   known after apply               network interfaces           allow ip spoofing     known after apply           id                   known after apply           name                 known after apply           primary address known after apply           security groups       known after apply           subnet               known after apply           primary ip               address       known after apply               auto delete   known after apply               href           known after apply               name           known after apply               reserved ip   known after apply               resource type known after apply                           primary network interface           allow ip spoofing     false           id                   known after apply           name                           port speed           known after apply           primary address known after apply           security groups       known after apply           subnet                         primary ip               address       known after apply               auto delete   known after apply               href           known after apply               name           known after apply               reserved ip                 resource type known after apply                         plan to add to change to destroy do you want to perform these actions   terraform will perform the actions described above   only yes will be accepted to approve   enter a value yes ibm is instance creating ibm is instance creating ibm is instance still creating ibm is instance still creating ibm is instance still creating ibm is instance still creating ibm is instance creation complete after ibm is instance creation complete after apply complete resources added changed destroyed moby cloud practice internal extra vm ibm exam important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example
1
26,418
20,103,162,550
IssuesEvent
2022-02-07 07:44:42
Greenstand/treetracker-web-map-client
https://api.github.com/repos/Greenstand/treetracker-web-map-client
opened
A solution to share/reuse components crossing repositories.
good first issue infrastructure
For example, the share components, they are the same between wallet app and web map app, so we need to find a good solution to share the components crossing repos. --- Some hints: - Please read our [readme](https://github.com/Greenstand/treetracker-web-map-client#treetracker-web) for more information/guide/tutorial. - Here is [an engineering book](https://greenstand.gitbook.io/engineering/) in Greenstand. - To know more about our organization, visit our [website](https://greenstand.org). - If you want to join the slack community (some resources need the community member's permission), please leave your email address. - More issues in this project: https://github.com/orgs/Greenstand/projects/26
1.0
A solution to share/reuse components crossing repositories. - For example, the share components, they are the same between wallet app and web map app, so we need to find a good solution to share the components crossing repos. --- Some hints: - Please read our [readme](https://github.com/Greenstand/treetracker-web-map-client#treetracker-web) for more information/guide/tutorial. - Here is [an engineering book](https://greenstand.gitbook.io/engineering/) in Greenstand. - To know more about our organization, visit our [website](https://greenstand.org). - If you want to join the slack community (some resources need the community member's permission), please leave your email address. - More issues in this project: https://github.com/orgs/Greenstand/projects/26
infrastructure
a solution to share reuse components crossing repositories for example the share components they are the same between wallet app and web map app so we need to find a good solution to share the components crossing repos some hints please read our for more information guide tutorial here is in greenstand to know more about our organization visit our if you want to join the slack community some resources need the community member s permission please leave your email address more issues in this project
1
12,239
9,659,947,627
IssuesEvent
2019-05-20 14:30:28
elastic/beats
https://api.github.com/repos/elastic/beats
closed
Input docker behaviour with Openshift 3.10
:infrastructure Filebeat Metricbeat [zube]: In Progress containers enhancement libbeat v7.2.0
Filebeat Version : 6.4.3-1 Openshift version : 3.10 Since Openshift version 3.10, docker has been replaced by cri-o. Log format changed from json to cri format and we need to use input docker to parse file as describe in this issue https://github.com/elastic/beats/issues/7418 Nethertheless, the path of docker logs in Openshift 3.10 is /var/log/containers/<pod_name>_<namespace>_<container_name>-<container_id>.log. Docker input consider this is this /var/lib/docker/containers/<container_id>/*.log. I used this workaround for all docker <pre>filebeat: inputs: - type: docker containers: path: "/var/log/containers" ids: - '' symlinks: true multiline.pattern: '^\[' multiline.match: after multiline.negate: true fields: type: openshift-pods fields_under_root: true exclude_lines: ["DEBUG","TRACE"] exclude_files: ["rabbitmq","nginx","redis"]</pre> It will considered this path -> /var/log/containers//*.log and it's ok if I wanted to process all log on the same way. In the other side, I need to process rabbitmq, nginx and redis pods on another ways but I'm blocked; For example <pre> filebeat: inputs: - type: docker containers: path: "/var/log/containers" ids: - '' symlinks: true multiline.pattern: '^[0-9]{4}-[0-9]{2}-[0-9]{2}' multiline.match: after multiline.negate: true fields: type: rabbit-pods fields_under_root: true ## Negative pattern not supported exclude_lines: [".*(?!rabbitmq).*"] ## exclude_lines: ["DEBUG","TRACE"] </pre> This example failed. I can't use exclude_line with negative pattern. Include line is not supported too. I think it should be good to be able to modify the "/*.log" automatically added by filebeat with input docker. PS : directory /var/log/pods should be the good directory for input but it contain only ID of dockers and I'll lost all <pod_name>_<namespace>_<container_name>-<container_id> I mentionned before.
1.0
Input docker behaviour with Openshift 3.10 - Filebeat Version : 6.4.3-1 Openshift version : 3.10 Since Openshift version 3.10, docker has been replaced by cri-o. Log format changed from json to cri format and we need to use input docker to parse file as describe in this issue https://github.com/elastic/beats/issues/7418 Nethertheless, the path of docker logs in Openshift 3.10 is /var/log/containers/<pod_name>_<namespace>_<container_name>-<container_id>.log. Docker input consider this is this /var/lib/docker/containers/<container_id>/*.log. I used this workaround for all docker <pre>filebeat: inputs: - type: docker containers: path: "/var/log/containers" ids: - '' symlinks: true multiline.pattern: '^\[' multiline.match: after multiline.negate: true fields: type: openshift-pods fields_under_root: true exclude_lines: ["DEBUG","TRACE"] exclude_files: ["rabbitmq","nginx","redis"]</pre> It will considered this path -> /var/log/containers//*.log and it's ok if I wanted to process all log on the same way. In the other side, I need to process rabbitmq, nginx and redis pods on another ways but I'm blocked; For example <pre> filebeat: inputs: - type: docker containers: path: "/var/log/containers" ids: - '' symlinks: true multiline.pattern: '^[0-9]{4}-[0-9]{2}-[0-9]{2}' multiline.match: after multiline.negate: true fields: type: rabbit-pods fields_under_root: true ## Negative pattern not supported exclude_lines: [".*(?!rabbitmq).*"] ## exclude_lines: ["DEBUG","TRACE"] </pre> This example failed. I can't use exclude_line with negative pattern. Include line is not supported too. I think it should be good to be able to modify the "/*.log" automatically added by filebeat with input docker. PS : directory /var/log/pods should be the good directory for input but it contain only ID of dockers and I'll lost all <pod_name>_<namespace>_<container_name>-<container_id> I mentionned before.
infrastructure
input docker behaviour with openshift filebeat version openshift version since openshift version docker has been replaced by cri o log format changed from json to cri format and we need to use input docker to parse file as describe in this issue nethertheless the path of docker logs in openshift is var log containers log docker input consider this is this var lib docker containers log i used this workaround for all docker filebeat inputs type docker containers path var log containers ids symlinks true multiline pattern multiline match after multiline negate true fields type openshift pods fields under root true exclude lines exclude files it will considered this path var log containers log and it s ok if i wanted to process all log on the same way in the other side i need to process rabbitmq nginx and redis pods on another ways but i m blocked for example filebeat inputs type docker containers path var log containers ids symlinks true multiline pattern multiline match after multiline negate true fields type rabbit pods fields under root true negative pattern not supported exclude lines exclude lines this example failed i can t use exclude line with negative pattern include line is not supported too i think it should be good to be able to modify the log automatically added by filebeat with input docker ps directory var log pods should be the good directory for input but it contain only id of dockers and i ll lost all i mentionned before
1
35,588
31,849,543,709
IssuesEvent
2023-09-14 23:34:56
casangi/astrohack
https://api.github.com/repos/casangi/astrohack
opened
Fix matplotlib close issue when multiple plots are created
Area:Infrastructure visualization
This is made simpler with a few steps: - [ ] Factorize figure and axes creation with the function now located in _locit_commons - [ ] Factorize figure closing with the function now located in _locit_commons After this factorization several solutions can be tested as the code will need to be changed in a single place
1.0
Fix matplotlib close issue when multiple plots are created - This is made simpler with a few steps: - [ ] Factorize figure and axes creation with the function now located in _locit_commons - [ ] Factorize figure closing with the function now located in _locit_commons After this factorization several solutions can be tested as the code will need to be changed in a single place
infrastructure
fix matplotlib close issue when multiple plots are created this is made simpler with a few steps factorize figure and axes creation with the function now located in locit commons factorize figure closing with the function now located in locit commons after this factorization several solutions can be tested as the code will need to be changed in a single place
1
645,686
21,012,132,522
IssuesEvent
2022-03-30 07:43:45
MaibornWolff/codecharta
https://api.github.com/repos/MaibornWolff/codecharta
closed
File path in side bar not entirely visible
bug priority:high
# Bug ## Expected Behavior When I click on a building, the entire file path should be visible. ## Actual Behavior Long file paths are cut and not entirely visible. ## Screenshots ![image](https://user-images.githubusercontent.com/48218172/160423443-30309182-f538-47b5-8da7-1cf9e5c28336.png) ## Steps to Reproduce the Problem 1. Open Codecharta 2. Click on a building with a long filename ## Specifications - is released in [online-demo](https://maibornwolff.github.io/codecharta/visualization/app/index.html?file=codecharta.cc.json&file=codecharta_analysis.cc.json): yes - CodeCharta Version: 1.95.0 - OS: Windows 10 - Browser: Chrome
1.0
File path in side bar not entirely visible - # Bug ## Expected Behavior When I click on a building, the entire file path should be visible. ## Actual Behavior Long file paths are cut and not entirely visible. ## Screenshots ![image](https://user-images.githubusercontent.com/48218172/160423443-30309182-f538-47b5-8da7-1cf9e5c28336.png) ## Steps to Reproduce the Problem 1. Open Codecharta 2. Click on a building with a long filename ## Specifications - is released in [online-demo](https://maibornwolff.github.io/codecharta/visualization/app/index.html?file=codecharta.cc.json&file=codecharta_analysis.cc.json): yes - CodeCharta Version: 1.95.0 - OS: Windows 10 - Browser: Chrome
non_infrastructure
file path in side bar not entirely visible bug expected behavior when i click on a building the entire file path should be visible actual behavior long file paths are cut and not entirely visible screenshots steps to reproduce the problem open codecharta click on a building with a long filename specifications is released in yes codecharta version os windows browser chrome
0
18,598
13,055,987,064
IssuesEvent
2020-07-30 03:18:52
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[documentation] 403 forbidden (Trac #1980)
Incomplete Migration Migrated from Trac infrastructure task
Migrated from https://code.icecube.wisc.edu/ticket/1980 ```json { "status": "closed", "changetime": "2019-02-13T14:14:44", "description": "Whenever the nightly doc rebuild happens, we get 403 forbidden for quite a while (many minutes to hours). While the US users may not care as much, since it happens late at night, Europe/Asia is awake and working. So, let's make sure this doesn't happen.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067284370534", "component": "infrastructure", "summary": "[documentation] 403 forbidden", "priority": "normal", "keywords": "", "time": "2017-04-09T13:25:11", "milestone": "", "owner": "nega", "type": "task" } ```
1.0
[documentation] 403 forbidden (Trac #1980) - Migrated from https://code.icecube.wisc.edu/ticket/1980 ```json { "status": "closed", "changetime": "2019-02-13T14:14:44", "description": "Whenever the nightly doc rebuild happens, we get 403 forbidden for quite a while (many minutes to hours). While the US users may not care as much, since it happens late at night, Europe/Asia is awake and working. So, let's make sure this doesn't happen.", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067284370534", "component": "infrastructure", "summary": "[documentation] 403 forbidden", "priority": "normal", "keywords": "", "time": "2017-04-09T13:25:11", "milestone": "", "owner": "nega", "type": "task" } ```
infrastructure
forbidden trac migrated from json status closed changetime description whenever the nightly doc rebuild happens we get forbidden for quite a while many minutes to hours while the us users may not care as much since it happens late at night europe asia is awake and working so let s make sure this doesn t happen reporter david schultz cc resolution fixed ts component infrastructure summary forbidden priority normal keywords time milestone owner nega type task
1
11,909
9,516,166,931
IssuesEvent
2019-04-26 08:09:10
elastic/beats
https://api.github.com/repos/elastic/beats
closed
[Metricbeat] Flaky test mssql
:infrastructure [zube]: In Progress flaky-test
## Flaky Test * **Test Name:** x-pack/metricbeat/tests/system/test_mssql.py * **Link:** https://travis-ci.org/elastic/beats/jobs/522183265#L683 ### Stack Trace ``` etricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 ... ====================================================================== ERROR: MSSQL module outputs an event. ---------------------------------------------------------------------- Traceback (most recent call last): File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 75, in test_performance self.assert_fields_are_documented(evt) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented expected_fields, dict_fields, aliases = self.load_fields() File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields subfields, subdictfields, subaliases = extract_fields(item["fields"], "") File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) KeyError: 'fields' -------------------- >> begin captured stdout << --------------------- render config --------------------- >> end captured stdout << ---------------------- -------------------- >> begin captured logging << -------------------- compose.config.config: DEBUG: Using configuration files: ./docker-compose.yml compose.config.environment: WARNING: The PWD variable is not set. Defaulting to a blank string. docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] docker.utils.config: DEBUG: No config file found docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] docker.utils.config: DEBUG: No config file found compose.service: INFO: Building mssql docker.api.build: DEBUG: Looking for auth config docker.api.build: DEBUG: No auth config in memory - loading from filesystem docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] docker.utils.config: DEBUG: No config file found docker.api.build: DEBUG: Sending auth config () compose.parallel: DEBUG: Pending: set([<Service: mssql>]) compose.parallel: DEBUG: Starting producer thread for <Service: mssql> compose.parallel: DEBUG: Pending: set([<Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>]) compose.parallel: DEBUG: Starting producer thread for <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)> compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.service: DEBUG: Added config hash: 3634631a392fe7d5944094282be0b8e946cb6a3c8f386ada7e1e488316879ad7 compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Finished processing: <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)> compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Finished processing: <Service: mssql> compose.parallel: DEBUG: Pending: set([]) --------------------- >> end captured logging << --------------------- ====================================================================== ERROR: MSSQL module outputs an event. ---------------------------------------------------------------------- Traceback (most recent call last): File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 46, in test_status self.assert_fields_are_documented(evt) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented expected_fields, dict_fields, aliases = self.load_fields() File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields subfields, subdictfields, subaliases = extract_fields(item["fields"], "") File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) KeyError: 'fields' -------------------- >> begin captured stdout << --------------------- render config --------------------- >> end captured stdout << ---------------------- ```
1.0
[Metricbeat] Flaky test mssql - ## Flaky Test * **Test Name:** x-pack/metricbeat/tests/system/test_mssql.py * **Link:** https://travis-ci.org/elastic/beats/jobs/522183265#L683 ### Stack Trace ``` etricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 ... ====================================================================== ERROR: MSSQL module outputs an event. ---------------------------------------------------------------------- Traceback (most recent call last): File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 75, in test_performance self.assert_fields_are_documented(evt) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented expected_fields, dict_fields, aliases = self.load_fields() File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields subfields, subdictfields, subaliases = extract_fields(item["fields"], "") File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) KeyError: 'fields' -------------------- >> begin captured stdout << --------------------- render config --------------------- >> end captured stdout << ---------------------- -------------------- >> begin captured logging << -------------------- compose.config.config: DEBUG: Using configuration files: ./docker-compose.yml compose.config.environment: WARNING: The PWD variable is not set. Defaulting to a blank string. docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] docker.utils.config: DEBUG: No config file found docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] docker.utils.config: DEBUG: No config file found compose.service: INFO: Building mssql docker.api.build: DEBUG: Looking for auth config docker.api.build: DEBUG: No auth config in memory - loading from filesystem docker.utils.config: DEBUG: Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] docker.utils.config: DEBUG: No config file found docker.api.build: DEBUG: Sending auth config () compose.parallel: DEBUG: Pending: set([<Service: mssql>]) compose.parallel: DEBUG: Starting producer thread for <Service: mssql> compose.parallel: DEBUG: Pending: set([<Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)>]) compose.parallel: DEBUG: Starting producer thread for <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)> compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.service: DEBUG: Added config hash: 3634631a392fe7d5944094282be0b8e946cb6a3c8f386ada7e1e488316879ad7 compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Finished processing: <Container: metricbeat_8_0_0_bba0daf1ae-snapshot_mssql_1 (12114c)> compose.parallel: DEBUG: Pending: set([]) compose.parallel: DEBUG: Finished processing: <Service: mssql> compose.parallel: DEBUG: Pending: set([]) --------------------- >> end captured logging << --------------------- ====================================================================== ERROR: MSSQL module outputs an event. ---------------------------------------------------------------------- Traceback (most recent call last): File "/go/src/github.com/elastic/beats/x-pack/metricbeat/tests/system/test_mssql.py", line 46, in test_status self.assert_fields_are_documented(evt) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 645, in assert_fields_are_documented expected_fields, dict_fields, aliases = self.load_fields() File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 578, in load_fields subfields, subdictfields, subaliases = extract_fields(item["fields"], "") File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) File "/go/src/github.com/elastic/beats/libbeat/tests/system/beat/beat.py", line 538, in extract_fields subfields, subdictfields, subaliases = extract_fields(field["fields"], newName) KeyError: 'fields' -------------------- >> begin captured stdout << --------------------- render config --------------------- >> end captured stdout << ---------------------- ```
infrastructure
flaky test mssql flaky test test name x pack metricbeat tests system test mssql py link stack trace etricbeat snapshot mssql error mssql module outputs an event traceback most recent call last file go src github com elastic beats x pack metricbeat tests system test mssql py line in test performance self assert fields are documented evt file go src github com elastic beats libbeat tests system beat beat py line in assert fields are documented expected fields dict fields aliases self load fields file go src github com elastic beats libbeat tests system beat beat py line in load fields subfields subdictfields subaliases extract fields item file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname keyerror fields begin captured stdout render config end captured stdout begin captured logging compose config config debug using configuration files docker compose yml compose config environment warning the pwd variable is not set defaulting to a blank string docker utils config debug trying paths docker utils config debug no config file found docker utils config debug trying paths docker utils config debug no config file found compose service info building mssql docker api build debug looking for auth config docker api build debug no auth config in memory loading from filesystem docker utils config debug trying paths docker utils config debug no config file found docker api build debug sending auth config compose parallel debug pending set compose parallel debug starting producer thread for compose parallel debug pending set compose parallel debug starting producer thread for compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose service debug added config hash compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug pending set compose parallel debug finished processing compose parallel debug pending set compose parallel debug finished processing compose parallel debug pending set end captured logging error mssql module outputs an event traceback most recent call last file go src github com elastic beats x pack metricbeat tests system test mssql py line in test status self assert fields are documented evt file go src github com elastic beats libbeat tests system beat beat py line in assert fields are documented expected fields dict fields aliases self load fields file go src github com elastic beats libbeat tests system beat beat py line in load fields subfields subdictfields subaliases extract fields item file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname file go src github com elastic beats libbeat tests system beat beat py line in extract fields subfields subdictfields subaliases extract fields field newname keyerror fields begin captured stdout render config end captured stdout
1
7,358
6,918,829,551
IssuesEvent
2017-11-29 13:38:52
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
opened
Refactor Message class
infrastructure
The `Message` class' origins date very far back and the resulting code is not overly clean. There is a lot of overlapping API surface, which can be difficult to use and understand the consequences of. Worse, we cannot inject anything into newly created Message objects, which means that any additional behavior either needs to be passed in as extra parameters to its various methods, or has to be added to the class itself which really is not be best place to have. The idea is to create a `MessageProvider` that can be injected wherever new `Message` objects are required. The provider can then make sure that message objects are "injected" with the necessary references to other systems (such as stats recording, message weighers, etc). By creating a special provider we can work around the overhead of injecting directly.
1.0
Refactor Message class - The `Message` class' origins date very far back and the resulting code is not overly clean. There is a lot of overlapping API surface, which can be difficult to use and understand the consequences of. Worse, we cannot inject anything into newly created Message objects, which means that any additional behavior either needs to be passed in as extra parameters to its various methods, or has to be added to the class itself which really is not be best place to have. The idea is to create a `MessageProvider` that can be injected wherever new `Message` objects are required. The provider can then make sure that message objects are "injected" with the necessary references to other systems (such as stats recording, message weighers, etc). By creating a special provider we can work around the overhead of injecting directly.
infrastructure
refactor message class the message class origins date very far back and the resulting code is not overly clean there is a lot of overlapping api surface which can be difficult to use and understand the consequences of worse we cannot inject anything into newly created message objects which means that any additional behavior either needs to be passed in as extra parameters to its various methods or has to be added to the class itself which really is not be best place to have the idea is to create a messageprovider that can be injected wherever new message objects are required the provider can then make sure that message objects are injected with the necessary references to other systems such as stats recording message weighers etc by creating a special provider we can work around the overhead of injecting directly
1
50,008
6,291,989,371
IssuesEvent
2017-07-20 03:18:04
PaddlePaddle/Paddle
https://api.github.com/repos/PaddlePaddle/Paddle
closed
Paddle 应该是什么形态
design_doc
如果要允许用户在iPython和Jupiter之类的界面里写Paddle程序,那么Paddle得是一个library(提供本地函数调用)或者一个RPC server(提供远程函数调用),而不能是目前的executable command line tool的形式。 在library和RPC server之间的选择是个问题。欢迎大家讨论。
1.0
Paddle 应该是什么形态 - 如果要允许用户在iPython和Jupiter之类的界面里写Paddle程序,那么Paddle得是一个library(提供本地函数调用)或者一个RPC server(提供远程函数调用),而不能是目前的executable command line tool的形式。 在library和RPC server之间的选择是个问题。欢迎大家讨论。
non_infrastructure
paddle 应该是什么形态 如果要允许用户在ipython和jupiter之类的界面里写paddle程序,那么paddle得是一个library(提供本地函数调用)或者一个rpc server(提供远程函数调用),而不能是目前的executable command line tool的形式。 在library和rpc server之间的选择是个问题。欢迎大家讨论。
0
13,258
10,170,321,815
IssuesEvent
2019-08-08 04:46:55
dotnet/coreclr
https://api.github.com/repos/dotnet/coreclr
closed
Sign NuGet packages prior to publish
area-Infrastructure
In PR https://github.com/dotnet/coreclr/pull/26020 sign checking has been deferred. This is being deferred because the coreclr doesn't presently sign `nupkg` files. Once that is done, sign validation can be re-enabled. cc @dotnet/coreclr-infra
1.0
Sign NuGet packages prior to publish - In PR https://github.com/dotnet/coreclr/pull/26020 sign checking has been deferred. This is being deferred because the coreclr doesn't presently sign `nupkg` files. Once that is done, sign validation can be re-enabled. cc @dotnet/coreclr-infra
infrastructure
sign nuget packages prior to publish in pr sign checking has been deferred this is being deferred because the coreclr doesn t presently sign nupkg files once that is done sign validation can be re enabled cc dotnet coreclr infra
1
2,256
2,524,982,973
IssuesEvent
2015-01-20 21:24:34
graybeal/ont
https://api.github.com/repos/graybeal/ont
opened
Add conditions field for usage metadata
1 star enhancement imported Priority-Medium
_From [cmt1...@gmail.com](https://code.google.com/u/111118716825497141417/) on October 22, 2009 10:17:19_ What capability do you want added or improved? Add a field where ontology providers specify the "conditions" under which author credit is required. Where do you want this capability to be accessible? Usage/permissions/license metadata tab What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.) _Original issue: http://code.google.com/p/mmisw/issues/detail?id=198_
1.0
Add conditions field for usage metadata - _From [cmt1...@gmail.com](https://code.google.com/u/111118716825497141417/) on October 22, 2009 10:17:19_ What capability do you want added or improved? Add a field where ontology providers specify the "conditions" under which author credit is required. Where do you want this capability to be accessible? Usage/permissions/license metadata tab What sort of input/command mechanism do you want? What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.) _Original issue: http://code.google.com/p/mmisw/issues/detail?id=198_
non_infrastructure
add conditions field for usage metadata from on october what capability do you want added or improved add a field where ontology providers specify the conditions under which author credit is required where do you want this capability to be accessible usage permissions license metadata tab what sort of input command mechanism do you want what is the desired output content format location other details of your desired capability what version of the product are you using please provide any additional information below particular ontology ies text contents of vocabulary operating system browser version firefox safari chrome ie etc screenshot etc original issue
0