Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
232,171 | 7,655,668,505 | IssuesEvent | 2018-05-10 13:59:52 | byu-oit/news-website | https://api.github.com/repos/byu-oit/news-website | opened | Use BYU paragraph types on random pages | priority-low | The Title IX, Terms of Use, and other pages use Bootstrap paragraphs. We need to use our own paragraphs so we can get rid of Bootstrap paragraphs. | 1.0 | Use BYU paragraph types on random pages - The Title IX, Terms of Use, and other pages use Bootstrap paragraphs. We need to use our own paragraphs so we can get rid of Bootstrap paragraphs. | non_infrastructure | use byu paragraph types on random pages the title ix terms of use and other pages use bootstrap paragraphs we need to use our own paragraphs so we can get rid of bootstrap paragraphs | 0 |
25,628 | 18,926,937,373 | IssuesEvent | 2021-11-17 10:31:30 | daviddellarossa/CM3025-virtual-reality | https://api.github.com/repos/daviddellarossa/CM3025-virtual-reality | closed | Creation of a Game state machine for Game manager | enhancement game infrastructure gameplay | This state machine manages the various states during the gameplay.
Some states can be specific of a game mode.
E.g.: Game mode Game can have a different set of states than game mode Learning tool.
The game states are controlled by the game mode manager.
## Done when:
The Game manager is provided with a State Machine that can change the state game based on events.
The specific set of game states is provided by the currently chosen Game mode.
## Dependencies
Requires: #21 | 1.0 | Creation of a Game state machine for Game manager - This state machine manages the various states during the gameplay.
Some states can be specific of a game mode.
E.g.: Game mode Game can have a different set of states than game mode Learning tool.
The game states are controlled by the game mode manager.
## Done when:
The Game manager is provided with a State Machine that can change the state game based on events.
The specific set of game states is provided by the currently chosen Game mode.
## Dependencies
Requires: #21 | infrastructure | creation of a game state machine for game manager this state machine manages the various states during the gameplay some states can be specific of a game mode e g game mode game can have a different set of states than game mode learning tool the game states are controlled by the game mode manager done when the game manager is provided with a state machine that can change the state game based on events the specific set of game states is provided by the currently chosen game mode dependencies requires | 1 |
184,170 | 6,706,319,762 | IssuesEvent | 2017-10-12 06:24:57 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.glassdoor.com - design is broken | browser-firefox priority-important status-needstriage type-stylo | <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.glassdoor.com/member/profile/index.htm
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: Broken layout.
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/10/09a2969d-86f6-44b1-97ab-7c396f3f17f0.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.glassdoor.com - design is broken - <!-- @browser: Firefox 58.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.glassdoor.com/member/profile/index.htm
**Browser / Version**: Firefox 58.0
**Operating System**: Windows 10
**Tested Another Browser**: No
**Problem type**: Design is broken
**Description**: Broken layout.
**Steps to Reproduce**:
layout.css.servo.enabled: true
[](https://webcompat.com/uploads/2017/10/09a2969d-86f6-44b1-97ab-7c396f3f17f0.jpg)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | design is broken url browser version firefox operating system windows tested another browser no problem type design is broken description broken layout steps to reproduce layout css servo enabled true from with ❤️ | 0 |
29,397 | 23,980,336,961 | IssuesEvent | 2022-09-13 14:37:21 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Android.Device_Emulator.Aot_Llvm - APP_CRASH | blocking-clean-ci os-android untriaged area-Infrastructure-mono | Frequency in [last 30 days](https://runfo.azurewebsites.net/search/tests/?q=started%3A%7E30+definition%3Aruntime-extra-platforms+name%3A%22android.device_emulator.aot_llvm.test+work+item%22) in Runfo as of 9/13:
- **First occurrence on 9/11 - Rolling run [11877](https://runfo.azurewebsites.net/view/build/?number=11877)**
- Each Rolling run or PR have 2, 4 or 6 failures in pairs of platforms:
- net7.0-Android-Release-x64-Mono_Release-Ubuntu.1804.Amd64.Android.29.Open
- net7.0-Android-Release-arm64-Mono_Release-Windows.10.Amd64.Android.Open
Looks like **deterministic failure**
### Error
[Console](https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-516b7dd3f8944d569d/Android.Device_Emulator.Aot_Llvm.Test.Attempt.3/1/console.72d92676.log?%3Fhelixlogtype%3Dresult)
```
[04:19:32] dbug: Success!
[04:19:32] info: Starting instrumentation class 'net.dot.MonoRunner' on net.dot.Android.Device_Emulator.Aot_Llvm.Test
[04:19:32] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 shell am instrument -w net.dot.Android.Device_Emulator.Aot_Llvm.Test/net.dot.MonoRunner'
[04:19:34] info: Running instrumentation class net.dot.MonoRunner took 2.5401223 seconds
[04:19:34] dbug: Exit code: 0
Std out:
INSTRUMENTATION_RESULT: shortMsg=Process crashed.
INSTRUMENTATION_CODE: 0
[04:19:34] info: Short message:
Process crashed.
[04:19:34] fail: No value for 'return-code' provided in instrumentation result. This may indicate a crashed test (see log)
[04:19:37] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 logcat -d '
[04:19:38] info: Wrote current ADB log to C:\h\w\AEA309BC\w\A9220975\uploads\adb-logcat-net.dot.Android.Device_Emulator.Aot_Llvm.Test-net.dot.MonoRunner.log
[04:19:41] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 bugreport "C:\\h\\w\\AEA309BC\\w\\A9220975\\uploads\\adb-bugreport-net.dot.Android.Device_Emulator.Aot_Llvm.Test.zip"'
[04:22:06] info: Wrote ADB bugreport to C:\h\w\AEA309BC\w\A9220975\uploads\adb-bugreport-net.dot.Android.Device_Emulator.Aot_Llvm.Test.zip
[04:22:06] info: Attempting to remove apk 'net.dot.Android.Device_Emulator.Aot_Llvm.Test'..
[04:22:06] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 uninstall net.dot.Android.Device_Emulator.Aot_Llvm.Test'
[04:22:07] info: Successfully uninstalled net.dot.Android.Device_Emulator.Aot_Llvm.Test
[04:22:07] dbug: Saving diagnostics data to 'C:\h\w\AEA309BC\w\A9220975\e/diagnostics.json'
XHarness exit code: 80 (APP_CRASH)
``` | 1.0 | Android.Device_Emulator.Aot_Llvm - APP_CRASH - Frequency in [last 30 days](https://runfo.azurewebsites.net/search/tests/?q=started%3A%7E30+definition%3Aruntime-extra-platforms+name%3A%22android.device_emulator.aot_llvm.test+work+item%22) in Runfo as of 9/13:
- **First occurrence on 9/11 - Rolling run [11877](https://runfo.azurewebsites.net/view/build/?number=11877)**
- Each Rolling run or PR have 2, 4 or 6 failures in pairs of platforms:
- net7.0-Android-Release-x64-Mono_Release-Ubuntu.1804.Amd64.Android.29.Open
- net7.0-Android-Release-arm64-Mono_Release-Windows.10.Amd64.Android.Open
Looks like **deterministic failure**
### Error
[Console](https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-heads-main-516b7dd3f8944d569d/Android.Device_Emulator.Aot_Llvm.Test.Attempt.3/1/console.72d92676.log?%3Fhelixlogtype%3Dresult)
```
[04:19:32] dbug: Success!
[04:19:32] info: Starting instrumentation class 'net.dot.MonoRunner' on net.dot.Android.Device_Emulator.Aot_Llvm.Test
[04:19:32] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 shell am instrument -w net.dot.Android.Device_Emulator.Aot_Llvm.Test/net.dot.MonoRunner'
[04:19:34] info: Running instrumentation class net.dot.MonoRunner took 2.5401223 seconds
[04:19:34] dbug: Exit code: 0
Std out:
INSTRUMENTATION_RESULT: shortMsg=Process crashed.
INSTRUMENTATION_CODE: 0
[04:19:34] info: Short message:
Process crashed.
[04:19:34] fail: No value for 'return-code' provided in instrumentation result. This may indicate a crashed test (see log)
[04:19:37] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 logcat -d '
[04:19:38] info: Wrote current ADB log to C:\h\w\AEA309BC\w\A9220975\uploads\adb-logcat-net.dot.Android.Device_Emulator.Aot_Llvm.Test-net.dot.MonoRunner.log
[04:19:41] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 bugreport "C:\\h\\w\\AEA309BC\\w\\A9220975\\uploads\\adb-bugreport-net.dot.Android.Device_Emulator.Aot_Llvm.Test.zip"'
[04:22:06] info: Wrote ADB bugreport to C:\h\w\AEA309BC\w\A9220975\uploads\adb-bugreport-net.dot.Android.Device_Emulator.Aot_Llvm.Test.zip
[04:22:06] info: Attempting to remove apk 'net.dot.Android.Device_Emulator.Aot_Llvm.Test'..
[04:22:06] dbug: Executing command: 'C:\h\w\AEA309BC\p\microsoft.dotnet.xharness.cli\1.0.0-prerelease.22456.1\runtimes\any\native\adb\windows\adb.exe -s 09301JEC225884 uninstall net.dot.Android.Device_Emulator.Aot_Llvm.Test'
[04:22:07] info: Successfully uninstalled net.dot.Android.Device_Emulator.Aot_Llvm.Test
[04:22:07] dbug: Saving diagnostics data to 'C:\h\w\AEA309BC\w\A9220975\e/diagnostics.json'
XHarness exit code: 80 (APP_CRASH)
``` | infrastructure | android device emulator aot llvm app crash frequency in in runfo as of first occurrence on rolling run each rolling run or pr have or failures in pairs of platforms android release mono release ubuntu android open android release mono release windows android open looks like deterministic failure error dbug success info starting instrumentation class net dot monorunner on net dot android device emulator aot llvm test dbug executing command c h w p microsoft dotnet xharness cli prerelease runtimes any native adb windows adb exe s shell am instrument w net dot android device emulator aot llvm test net dot monorunner info running instrumentation class net dot monorunner took seconds dbug exit code std out instrumentation result shortmsg process crashed instrumentation code info short message process crashed fail no value for return code provided in instrumentation result this may indicate a crashed test see log dbug executing command c h w p microsoft dotnet xharness cli prerelease runtimes any native adb windows adb exe s logcat d info wrote current adb log to c h w w uploads adb logcat net dot android device emulator aot llvm test net dot monorunner log dbug executing command c h w p microsoft dotnet xharness cli prerelease runtimes any native adb windows adb exe s bugreport c h w w uploads adb bugreport net dot android device emulator aot llvm test zip info wrote adb bugreport to c h w w uploads adb bugreport net dot android device emulator aot llvm test zip info attempting to remove apk net dot android device emulator aot llvm test dbug executing command c h w p microsoft dotnet xharness cli prerelease runtimes any native adb windows adb exe s uninstall net dot android device emulator aot llvm test info successfully uninstalled net dot android device emulator aot llvm test dbug saving diagnostics data to c h w w e diagnostics json xharness exit code app crash | 1 |
31,740 | 26,054,335,824 | IssuesEvent | 2022-12-22 22:39:18 | eosnetworkfoundation/devhub | https://api.github.com/repos/eosnetworkfoundation/devhub | opened | Backend CICD | Infrastructure CICD | Following devrel [issue 135](https://github.com/eosnetworkfoundation/devrel/issues/135), this issue has a true automated CICD system being implemented for the DevHub backend, and added as a required check on pull requests.
If possible, this system should deploy the backend to the cloud (the "CD" part of "CICD") under specific circumstances, such as tagged builds, but [issue 12](https://github.com/eosnetworkfoundation/devhub/issues/12) is almost certainly a prerequisite to this. If this cannot be done, a new issue should be created for that part so the work does not get lost. | 1.0 | Backend CICD - Following devrel [issue 135](https://github.com/eosnetworkfoundation/devrel/issues/135), this issue has a true automated CICD system being implemented for the DevHub backend, and added as a required check on pull requests.
If possible, this system should deploy the backend to the cloud (the "CD" part of "CICD") under specific circumstances, such as tagged builds, but [issue 12](https://github.com/eosnetworkfoundation/devhub/issues/12) is almost certainly a prerequisite to this. If this cannot be done, a new issue should be created for that part so the work does not get lost. | infrastructure | backend cicd following devrel this issue has a true automated cicd system being implemented for the devhub backend and added as a required check on pull requests if possible this system should deploy the backend to the cloud the cd part of cicd under specific circumstances such as tagged builds but is almost certainly a prerequisite to this if this cannot be done a new issue should be created for that part so the work does not get lost | 1 |
314,792 | 23,536,757,725 | IssuesEvent | 2022-08-19 21:59:56 | BrSTU-PO4-Pavel-Galanin/6sem_practice | https://api.github.com/repos/BrSTU-PO4-Pavel-Galanin/6sem_practice | closed | Backend write use Express, Sequelize and MySQL | documentation enhancement | Написать бэкэнд на Java Script используя Express, Sequelize, MySQL, Swagger, Redoc.
Чтобы не скачивать базу данных, можно запихнуть в docker-compose.
Написать README.md с командами запуска. Для сокращения команд можно добавить команды в Makefile. | 1.0 | Backend write use Express, Sequelize and MySQL - Написать бэкэнд на Java Script используя Express, Sequelize, MySQL, Swagger, Redoc.
Чтобы не скачивать базу данных, можно запихнуть в docker-compose.
Написать README.md с командами запуска. Для сокращения команд можно добавить команды в Makefile. | non_infrastructure | backend write use express sequelize and mysql написать бэкэнд на java script используя express sequelize mysql swagger redoc чтобы не скачивать базу данных можно запихнуть в docker compose написать readme md с командами запуска для сокращения команд можно добавить команды в makefile | 0 |
102,204 | 11,276,851,998 | IssuesEvent | 2020-01-15 00:38:09 | 10up/debug-bar-elasticpress | https://api.github.com/repos/10up/debug-bar-elasticpress | closed | readme updates | documentation | - [x] add support level section and badge
- [x] add release version badge
- [x] add WP tested up to badge
- [x] add license file and badge | 1.0 | readme updates - - [x] add support level section and badge
- [x] add release version badge
- [x] add WP tested up to badge
- [x] add license file and badge | non_infrastructure | readme updates add support level section and badge add release version badge add wp tested up to badge add license file and badge | 0 |
8,383 | 7,374,474,636 | IssuesEvent | 2018-03-13 20:28:46 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Remove CoreXT workaround added due to lack of support for some CoreFX packages | Area-Infrastructure | Once CoreXT updates all branches to NuGet 3.5 remove a workaround in DevDivInsertionFiles that blocks automatic insertion of CoreXT incompatible packages.
In default.config file, remove the `CoreFX.` prefix from the following entries:
```
<package id="CoreFX.System.Diagnostics.Process" version="4.1.0" link="src\ExternalAPIs\System.Diagnostics.Process" tags="exapis" />
<package id="CoreFX.System.AppContext" version="4.1.0" link="src\ExternalAPIs\System.AppContext" tags="exapis" />
<package id="CoreFX.System.Security.Cryptography.X509Certificates" version="4.1.0" link="src\ExternalAPIs\System.Security.Cryptography.X509Certificates" tags="exapis" />
<package id="CoreFX.System.Security.Cryptography.Algorithms" version="4.2.0" link="src\ExternalAPIs\System.Security.Cryptography.Algorithms" tags="exapis" />
<package id="CoreFX.System.Reflection.TypeExtensions" version="4.1.0" link="src\ExternalAPIs\System.Reflection.TypeExtensions" tags="exapis" />
<package id="CoreFX.System.Net.Security" version="4.0.0" link="src\ExternalAPIs\System.Net.Security" tags="exapis" />
<package id="CoreFX.System.IO.Compression" version="4.1.0" link="src\ExternalAPIs\System.IO.Compression" tags="exapis" />
```
The packages were generated from `%internal_share%\public\tomat\corext` and manually inserted.
In `src\Setup\DevDivInsertionFiles\BuildDevDivInsertionFiles.vb` remove `IsCoreXTCompatible` property.
| 1.0 | Remove CoreXT workaround added due to lack of support for some CoreFX packages - Once CoreXT updates all branches to NuGet 3.5 remove a workaround in DevDivInsertionFiles that blocks automatic insertion of CoreXT incompatible packages.
In default.config file, remove the `CoreFX.` prefix from the following entries:
```
<package id="CoreFX.System.Diagnostics.Process" version="4.1.0" link="src\ExternalAPIs\System.Diagnostics.Process" tags="exapis" />
<package id="CoreFX.System.AppContext" version="4.1.0" link="src\ExternalAPIs\System.AppContext" tags="exapis" />
<package id="CoreFX.System.Security.Cryptography.X509Certificates" version="4.1.0" link="src\ExternalAPIs\System.Security.Cryptography.X509Certificates" tags="exapis" />
<package id="CoreFX.System.Security.Cryptography.Algorithms" version="4.2.0" link="src\ExternalAPIs\System.Security.Cryptography.Algorithms" tags="exapis" />
<package id="CoreFX.System.Reflection.TypeExtensions" version="4.1.0" link="src\ExternalAPIs\System.Reflection.TypeExtensions" tags="exapis" />
<package id="CoreFX.System.Net.Security" version="4.0.0" link="src\ExternalAPIs\System.Net.Security" tags="exapis" />
<package id="CoreFX.System.IO.Compression" version="4.1.0" link="src\ExternalAPIs\System.IO.Compression" tags="exapis" />
```
The packages were generated from `%internal_share%\public\tomat\corext` and manually inserted.
In `src\Setup\DevDivInsertionFiles\BuildDevDivInsertionFiles.vb` remove `IsCoreXTCompatible` property.
| infrastructure | remove corext workaround added due to lack of support for some corefx packages once corext updates all branches to nuget remove a workaround in devdivinsertionfiles that blocks automatic insertion of corext incompatible packages in default config file remove the corefx prefix from the following entries the packages were generated from internal share public tomat corext and manually inserted in src setup devdivinsertionfiles builddevdivinsertionfiles vb remove iscorextcompatible property | 1 |
121,197 | 15,867,207,663 | IssuesEvent | 2021-04-08 16:37:06 | dotnet/roslyn-analyzers | https://api.github.com/repos/dotnet/roslyn-analyzers | closed | CA2249 encourages folks to ifdef code for "readability" | Resolution-By Design | ### Analyzer
**Diagnostic ID**: [ca2249](https://docs.microsoft.com/dotnet/fundamentals/code-analysis/quality-rules/ca2249): `Consider using String.Contains instead of String.IndexOf`
**SDK**: [Built-in CA analyzers in .NET 5 SDK or later](https://docs.microsoft.com/dotnet/fundamentals/productivity/code-analysis)
**Version**: [SDK 5.0.100](https://dotnet.microsoft.com/download/dotnet/5.0)
### Describe the bug
The analyzer doesn't consider cross-compilation so it tells you to do something to improve readability and the only change that can be made is less readable (and less performant).
### Steps To Reproduce
Do the following in a project with
```xml
<TargetFrameworks>net48;net50</TargetFrameworks>
```
```csharp
string toParse = "5";
if (toParse.IndexOf('.') == -1)
{
toParse += ".0";
}
```
Observe the error:
```
Error CA2249 Use 'string.Contains' instead of 'string.IndexOf' to improve readability
```
### Expected behavior
No warning since the recommendation can't be made consistently for all cross-targeting configurations.
### Actual behavior
A warning where the only way to remove it is to use a less performant string search, or to ifdef.
```csharp
#if NETCOREAPP
if (!toParse.Contains('.'))
#else
if (toParse.IndexOf('.') == -1)
#endif
```
| 1.0 | CA2249 encourages folks to ifdef code for "readability" - ### Analyzer
**Diagnostic ID**: [ca2249](https://docs.microsoft.com/dotnet/fundamentals/code-analysis/quality-rules/ca2249): `Consider using String.Contains instead of String.IndexOf`
**SDK**: [Built-in CA analyzers in .NET 5 SDK or later](https://docs.microsoft.com/dotnet/fundamentals/productivity/code-analysis)
**Version**: [SDK 5.0.100](https://dotnet.microsoft.com/download/dotnet/5.0)
### Describe the bug
The analyzer doesn't consider cross-compilation so it tells you to do something to improve readability and the only change that can be made is less readable (and less performant).
### Steps To Reproduce
Do the following in a project with
```xml
<TargetFrameworks>net48;net50</TargetFrameworks>
```
```csharp
string toParse = "5";
if (toParse.IndexOf('.') == -1)
{
toParse += ".0";
}
```
Observe the error:
```
Error CA2249 Use 'string.Contains' instead of 'string.IndexOf' to improve readability
```
### Expected behavior
No warning since the recommendation can't be made consistently for all cross-targeting configurations.
### Actual behavior
A warning where the only way to remove it is to use a less performant string search, or to ifdef.
```csharp
#if NETCOREAPP
if (!toParse.Contains('.'))
#else
if (toParse.IndexOf('.') == -1)
#endif
```
| non_infrastructure | encourages folks to ifdef code for readability analyzer diagnostic id consider using string contains instead of string indexof sdk version describe the bug the analyzer doesn t consider cross compilation so it tells you to do something to improve readability and the only change that can be made is less readable and less performant steps to reproduce do the following in a project with xml csharp string toparse if toparse indexof toparse observe the error error use string contains instead of string indexof to improve readability expected behavior no warning since the recommendation can t be made consistently for all cross targeting configurations actual behavior a warning where the only way to remove it is to use a less performant string search or to ifdef csharp if netcoreapp if toparse contains else if toparse indexof endif | 0 |
816,971 | 30,620,097,398 | IssuesEvent | 2023-07-24 07:38:58 | 1OS-DevTeam/TeamPlan | https://api.github.com/repos/1OS-DevTeam/TeamPlan | opened | [feat] 홈화면 : 기능(Back-End) 목록 | For: DB/API For: iOS Dev Priority: High Status: In Progress Type: Feature | ## Description ( todo 설명 )
홈화면 동작을 위해 필요한 Back-End 기능제공
## Task ( todo )
* **회원정보 관련**
- [ ] 사용자문구 조회
* **프로젝트 관련**
- [ ] 홈화면용 정보조회 ( 타이틀 / 물방울 갯수 / 완료되지않은 todo 갯수 / 등록일 / 마감일 )
* **도전과제 관련**
- [ ] 홈화면용 정보조회 ( 아이콘 이미지 / 제목 / 설명 )
## Comments
* **프로젝트 관련 예상 데이터타입**
* 타이틀 : String
* 물방울 개수 : int
* todo 개수 : int
* 등록일 : String ("yyyy-MM-dd HH:mm:ss" 형식)
* 마감일 : String ("yyyy-MM-dd HH:mm:ss" 형식)
* **프로젝트 관련 추가 데이터**
폭탄 Progress Bar 표현을 위해 추가데이터 제공이 필요한지 궁금합니다!
| 1.0 | [feat] 홈화면 : 기능(Back-End) 목록 - ## Description ( todo 설명 )
홈화면 동작을 위해 필요한 Back-End 기능제공
## Task ( todo )
* **회원정보 관련**
- [ ] 사용자문구 조회
* **프로젝트 관련**
- [ ] 홈화면용 정보조회 ( 타이틀 / 물방울 갯수 / 완료되지않은 todo 갯수 / 등록일 / 마감일 )
* **도전과제 관련**
- [ ] 홈화면용 정보조회 ( 아이콘 이미지 / 제목 / 설명 )
## Comments
* **프로젝트 관련 예상 데이터타입**
* 타이틀 : String
* 물방울 개수 : int
* todo 개수 : int
* 등록일 : String ("yyyy-MM-dd HH:mm:ss" 형식)
* 마감일 : String ("yyyy-MM-dd HH:mm:ss" 형식)
* **프로젝트 관련 추가 데이터**
폭탄 Progress Bar 표현을 위해 추가데이터 제공이 필요한지 궁금합니다!
| non_infrastructure | 홈화면 기능 back end 목록 description todo 설명 홈화면 동작을 위해 필요한 back end 기능제공 task todo 회원정보 관련 사용자문구 조회 프로젝트 관련 홈화면용 정보조회 타이틀 물방울 갯수 완료되지않은 todo 갯수 등록일 마감일 도전과제 관련 홈화면용 정보조회 아이콘 이미지 제목 설명 comments 프로젝트 관련 예상 데이터타입 타이틀 string 물방울 개수 int todo 개수 int 등록일 string yyyy mm dd hh mm ss 형식 마감일 string yyyy mm dd hh mm ss 형식 프로젝트 관련 추가 데이터 폭탄 progress bar 표현을 위해 추가데이터 제공이 필요한지 궁금합니다 | 0 |
10,940 | 4,851,899,058 | IssuesEvent | 2016-11-11 08:05:49 | Azure/azure-iot-sdks | https://api.github.com/repos/Azure/azure-iot-sdks | closed | Raspberry Pi blocks while running build.sh for Python | build issue CSS question | I've cloned this project on 2016/10/26 on my Raspberry Pi 2 Model B+ with this command:
```
git clone --recursive https://github.com/Azure/azure-iot-sdks
```
After that I've run this two files:
```
./azure-iot-sdks/python/build_all/linux/setup.sh
./azure-iot-sdks/python/build_all/linux/build.sh
```
With the first file I haven't any problem. The problem occurs when I run the file `build.sh`. This code will block my Raspberry Pi while the build is done for 98%. I've uploaded an image with the output.
[](https://cloud.githubusercontent.com/assets/16222780/19731069/7e8898fe-9b9c-11e6-947a-cac6aa5a1fb6.jpg)
<sub>Couldn't take a print screen because Pi was blocked and had to reboot the Pi.</sub>
Did I miss anything or is this a bug in the code?
Btw, I'm using the SDK for Python.
| 1.0 | Raspberry Pi blocks while running build.sh for Python - I've cloned this project on 2016/10/26 on my Raspberry Pi 2 Model B+ with this command:
```
git clone --recursive https://github.com/Azure/azure-iot-sdks
```
After that I've run this two files:
```
./azure-iot-sdks/python/build_all/linux/setup.sh
./azure-iot-sdks/python/build_all/linux/build.sh
```
With the first file I haven't any problem. The problem occurs when I run the file `build.sh`. This code will block my Raspberry Pi while the build is done for 98%. I've uploaded an image with the output.
[](https://cloud.githubusercontent.com/assets/16222780/19731069/7e8898fe-9b9c-11e6-947a-cac6aa5a1fb6.jpg)
<sub>Couldn't take a print screen because Pi was blocked and had to reboot the Pi.</sub>
Did I miss anything or is this a bug in the code?
Btw, I'm using the SDK for Python.
| non_infrastructure | raspberry pi blocks while running build sh for python i ve cloned this project on on my raspberry pi model b with this command git clone recursive after that i ve run this two files azure iot sdks python build all linux setup sh azure iot sdks python build all linux build sh with the first file i haven t any problem the problem occurs when i run the file build sh this code will block my raspberry pi while the build is done for i ve uploaded an image with the output couldn t take a print screen because pi was blocked and had to reboot the pi did i miss anything or is this a bug in the code btw i m using the sdk for python | 0 |
14,894 | 9,431,116,500 | IssuesEvent | 2019-04-12 10:44:22 | Alfresco/alfresco-repository | https://api.github.com/repos/Alfresco/alfresco-repository | closed | CVE-2018-1272 High Severity Vulnerability detected by WhiteSource | security vulnerability | ## CVE-2018-1272 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-5.0.4.RELEASE.jar</b></p></summary>
<p>Spring Core</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /alfresco-repository/pom.xml</p>
<p>Path to vulnerable library: 2/repository/org/springframework/spring-core/5.0.4.RELEASE/spring-core-5.0.4.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-core-5.0.4.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Alfresco/alfresco-repository/commit/a6b33c7b9104af395c59b8805bd5b78bc5e7e3ef">a6b33c7b9104af395c59b8805bd5b78bc5e7e3ef</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.
<p>Publish Date: 2018-04-06
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1272>CVE-2018-1272</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1272">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1272</a></p>
<p>Release Date: 2018-04-06</p>
<p>Fix Resolution: v4.3.15.RELEASE,v5.0.5.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isOpenPROnNewVersion":false,"isPackageBased":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"5.0.4.RELEASE","isTransitiveDependency":true,"dependencyTree":"org.springframework:spring-core:5.0.4.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.3.15.RELEASE,v5.0.5.RELEASE"}],"vulnerabilityIdentifier":"CVE-2018-1272","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-1272 High Severity Vulnerability detected by WhiteSource - ## CVE-2018-1272 - High Severity Vulnerability
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-core-5.0.4.RELEASE.jar</b></p></summary>
<p>Spring Core</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /alfresco-repository/pom.xml</p>
<p>Path to vulnerable library: 2/repository/org/springframework/spring-core/5.0.4.RELEASE/spring-core-5.0.4.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- :x: **spring-core-5.0.4.RELEASE.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Alfresco/alfresco-repository/commit/a6b33c7b9104af395c59b8805bd5b78bc5e7e3ef">a6b33c7b9104af395c59b8805bd5b78bc5e7e3ef</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.
<p>Publish Date: 2018-04-06
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1272>CVE-2018-1272</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://www.whitesourcesoftware.com/wp-content/uploads/2018/10/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1272">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-1272</a></p>
<p>Release Date: 2018-04-06</p>
<p>Fix Resolution: v4.3.15.RELEASE,v5.0.5.RELEASE</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isOpenPROnNewVersion":false,"isPackageBased":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-core","packageVersion":"5.0.4.RELEASE","isTransitiveDependency":true,"dependencyTree":"org.springframework:spring-core:5.0.4.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v4.3.15.RELEASE,v5.0.5.RELEASE"}],"vulnerabilityIdentifier":"CVE-2018-1272","vulnerabilityDetails":"Spring Framework, versions 5.0 prior to 5.0.5 and versions 4.3 prior to 4.3.15 and older unsupported versions, provide client-side support for multipart requests. When Spring MVC or Spring WebFlux server application (server A) receives input from a remote client, and then uses that input to make a multipart request to another server (server B), it can be exposed to an attack, where an extra multipart is inserted in the content of the request from server A, causing server B to use the wrong value for a part it expects. This could to lead privilege escalation, for example, if the part content represents a username or user roles.","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high severity vulnerability detected by whitesource cve high severity vulnerability vulnerable library spring core release jar spring core library home page a href path to dependency file alfresco repository pom xml path to vulnerable library repository org springframework spring core release spring core release jar dependency hierarchy x spring core release jar vulnerable library found in head commit a href vulnerability details spring framework versions prior to and versions prior to and older unsupported versions provide client side support for multipart requests when spring mvc or spring webflux server application server a receives input from a remote client and then uses that input to make a multipart request to another server server b it can be exposed to an attack where an extra multipart is inserted in the content of the request from server a causing server b to use the wrong value for a part it expects this could to lead privilege escalation for example if the part content represents a username or user roles publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution release release step up your open source security game with whitesource isopenpronvulnerability true isopenpronnewversion false ispackagebased true packages vulnerabilityidentifier cve vulnerabilitydetails spring framework versions prior to and versions prior to and older unsupported versions provide client side support for multipart requests when spring mvc or spring webflux server application server a receives input from a remote client and then uses that input to make a multipart request to another server server b it can be exposed to an attack where an extra multipart is inserted in the content of the request from server a causing server b to use the wrong value for a part it expects this could to lead privilege escalation for example if the part content represents a username or user roles high a high ac high pr low s unchanged c high ui none av network i high extradata | 0 |
295,121 | 25,456,499,048 | IssuesEvent | 2022-11-24 14:36:15 | FSGabrsek/Vessel-WebApp | https://api.github.com/repos/FSGabrsek/Vessel-WebApp | closed | Vessel-22 | test | ### Fix tests for MediaSoul implementation
Update models and add hosts components to existing and pre generated component tests.
- [x] Add soul and user objects to testing models
- [x] Add host components to new Input fields | 1.0 | Vessel-22 - ### Fix tests for MediaSoul implementation
Update models and add hosts components to existing and pre generated component tests.
- [x] Add soul and user objects to testing models
- [x] Add host components to new Input fields | non_infrastructure | vessel fix tests for mediasoul implementation update models and add hosts components to existing and pre generated component tests add soul and user objects to testing models add host components to new input fields | 0 |
25,541 | 18,844,859,324 | IssuesEvent | 2021-11-11 13:54:28 | google/web-stories-wp | https://api.github.com/repos/google/web-stories-wp | closed | Use the filter `safe_style_disallowed_chars` to make KSES class simplier | Type: Enhancement Type: Infrastructure PHP Group: WordPress Pod: WP & Infra P4 | <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
Follow up from https://github.com/google/web-stories-wp/pull/7951. Added in WordPress 5.5 ( https://github.com/WordPress/wordpress-develop/commit/fa96d5931826b936095022013e30ee5934018779 ), the `safe_style_disallowed_chars` filter could be used to make the logic in KSES simpler.
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
| 1.0 | Use the filter `safe_style_disallowed_chars` to make KSES class simplier - <!-- NOTE: For help requests, support questions, or general feedback, please use the WordPress.org forums instead: https://wordpress.org/support/plugin/web-stories/ -->
## Feature Description
Follow up from https://github.com/google/web-stories-wp/pull/7951. Added in WordPress 5.5 ( https://github.com/WordPress/wordpress-develop/commit/fa96d5931826b936095022013e30ee5934018779 ), the `safe_style_disallowed_chars` filter could be used to make the logic in KSES simpler.
## Alternatives Considered
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional Context
<!-- Add any other context or screenshots about the feature request. -->
---
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance Criteria
<!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
<!-- One or more bullet points for how to technically implement the feature. -->
| infrastructure | use the filter safe style disallowed chars to make kses class simplier feature description follow up from added in wordpress the safe style disallowed chars filter could be used to make the logic in kses simpler alternatives considered additional context do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief | 1 |
25,380 | 18,670,664,711 | IssuesEvent | 2021-10-30 16:45:14 | battlecode/battlecode21 | https://api.github.com/repos/battlecode/battlecode21 | closed | Write a more robust tournament runner | good first issue infrastructure | @tmightyquinn @j-mao have thoughts too -- feel free to dump! | 1.0 | Write a more robust tournament runner - @tmightyquinn @j-mao have thoughts too -- feel free to dump! | infrastructure | write a more robust tournament runner tmightyquinn j mao have thoughts too feel free to dump | 1 |
31,742 | 26,072,396,722 | IssuesEvent | 2022-12-24 01:36:09 | fahadahmed/roadmap-personal | https://api.github.com/repos/fahadahmed/roadmap-personal | opened | Code Formatting: Setup Eslint, Prettier and Husky | task infrastructure | # Overview
Set up code formatting tools like Eslint, Prettier and Husky for the project
## Tasks
- [ ] Setup Eslint for `app` and `functions` folders
- [ ] Setup Prettier for `app` and `functions` folders
- [ ] Setup Lint-staged and Husky for pre-commit hook
## Additional context
Add any other context or screenshots about the feature request here.
| 1.0 | Code Formatting: Setup Eslint, Prettier and Husky - # Overview
Set up code formatting tools like Eslint, Prettier and Husky for the project
## Tasks
- [ ] Setup Eslint for `app` and `functions` folders
- [ ] Setup Prettier for `app` and `functions` folders
- [ ] Setup Lint-staged and Husky for pre-commit hook
## Additional context
Add any other context or screenshots about the feature request here.
| infrastructure | code formatting setup eslint prettier and husky overview set up code formatting tools like eslint prettier and husky for the project tasks setup eslint for app and functions folders setup prettier for app and functions folders setup lint staged and husky for pre commit hook additional context add any other context or screenshots about the feature request here | 1 |
10,304 | 8,481,962,172 | IssuesEvent | 2018-10-25 17:09:24 | codeforboston/cliff-effects | https://api.github.com/repos/codeforboston/cliff-effects | opened | pre-commit hooks: remove 'lint-staged' and carats? | discussion needed infrastructure | About the awesome changes merged in #904.
1. The 'lint-staged' package doesn't actually limit npm lint fixes to just staged files. It changes the files, though it doesn't add them. Considering this will be in everyone's code and all files will be linted when committed and there shouldn't be leftover files to lint, it shouldn't be a problem either way, but we might consider getting rid of 'lint-staged'.
2. We'd agreed not to have carats in version numbers to try to stabilize our package-lock.json. Hasn't worked, but carats my confuse further efforts.
On other, less crucial notes, just so they're in the record:
1. Right now it prints console logs when it does its thing, which might make newer folks nervous. There's [currently no way to turn that off](https://github.com/typicode/husky/pull/300). Personally, I do generally go for explicit over implicit, but I'm not sure about this specific situation since it'll be happening right at the start. Depending on the outcome of the discussion, we could make a PR on [husky's repo](https://github.com/typicode/husky).
2. It's strange how it can make a commit even if the linting is the only thing that happened. I ruined some code format and made a commit just for that. After linting, there were no changes and it still committed. I wonder how it does that. | 1.0 | pre-commit hooks: remove 'lint-staged' and carats? - About the awesome changes merged in #904.
1. The 'lint-staged' package doesn't actually limit npm lint fixes to just staged files. It changes the files, though it doesn't add them. Considering this will be in everyone's code and all files will be linted when committed and there shouldn't be leftover files to lint, it shouldn't be a problem either way, but we might consider getting rid of 'lint-staged'.
2. We'd agreed not to have carats in version numbers to try to stabilize our package-lock.json. Hasn't worked, but carats my confuse further efforts.
On other, less crucial notes, just so they're in the record:
1. Right now it prints console logs when it does its thing, which might make newer folks nervous. There's [currently no way to turn that off](https://github.com/typicode/husky/pull/300). Personally, I do generally go for explicit over implicit, but I'm not sure about this specific situation since it'll be happening right at the start. Depending on the outcome of the discussion, we could make a PR on [husky's repo](https://github.com/typicode/husky).
2. It's strange how it can make a commit even if the linting is the only thing that happened. I ruined some code format and made a commit just for that. After linting, there were no changes and it still committed. I wonder how it does that. | infrastructure | pre commit hooks remove lint staged and carats about the awesome changes merged in the lint staged package doesn t actually limit npm lint fixes to just staged files it changes the files though it doesn t add them considering this will be in everyone s code and all files will be linted when committed and there shouldn t be leftover files to lint it shouldn t be a problem either way but we might consider getting rid of lint staged we d agreed not to have carats in version numbers to try to stabilize our package lock json hasn t worked but carats my confuse further efforts on other less crucial notes just so they re in the record right now it prints console logs when it does its thing which might make newer folks nervous there s personally i do generally go for explicit over implicit but i m not sure about this specific situation since it ll be happening right at the start depending on the outcome of the discussion we could make a pr on it s strange how it can make a commit even if the linting is the only thing that happened i ruined some code format and made a commit just for that after linting there were no changes and it still committed i wonder how it does that | 1 |
14,427 | 9,307,395,384 | IssuesEvent | 2019-03-25 12:10:54 | AOSC-Dev/aosc-os-abbs | https://api.github.com/repos/AOSC-Dev/aosc-os-abbs | opened | wireshark: security update to 2.6.7 | security to-stable upgrade | <!-- Please remove items do not apply. -->
**CVE IDs:** CVE-2019-9208, CVE-2019-9209
**Other security advisory IDs:** DSA-4416-1
**Descriptions:**
https://www.wireshark.org/docs/relnotes/wireshark-2.6.7.html
wnpa-sec-2019-06 ASN.1 BER and related dissectors crash. Bug 15447. CVE-2019-9209.
wnpa-sec-2019-07 TCAP dissector crash. Bug 15464. CVE-2019-9208.
wnpa-sec-2019-08 RPCAP dissector crash. Bug 15536.
**Architectural progress:**
<!-- Please remove any architecture to which the security vulnerabilities do not apply. -->
- [ ] AMD64 `amd64`
- [ ] 32-bit Optional Environment `optenv32`
- [ ] AArch64 `arm64`
- [ ] ARMv7 `armel`
- [ ] PowerPC 64-bit BE `ppc64`
- [ ] PowerPC 32-bit BE `powerpc`
- [ ] RISC-V 64-bit `riscv64`
| True | wireshark: security update to 2.6.7 - <!-- Please remove items do not apply. -->
**CVE IDs:** CVE-2019-9208, CVE-2019-9209
**Other security advisory IDs:** DSA-4416-1
**Descriptions:**
https://www.wireshark.org/docs/relnotes/wireshark-2.6.7.html
wnpa-sec-2019-06 ASN.1 BER and related dissectors crash. Bug 15447. CVE-2019-9209.
wnpa-sec-2019-07 TCAP dissector crash. Bug 15464. CVE-2019-9208.
wnpa-sec-2019-08 RPCAP dissector crash. Bug 15536.
**Architectural progress:**
<!-- Please remove any architecture to which the security vulnerabilities do not apply. -->
- [ ] AMD64 `amd64`
- [ ] 32-bit Optional Environment `optenv32`
- [ ] AArch64 `arm64`
- [ ] ARMv7 `armel`
- [ ] PowerPC 64-bit BE `ppc64`
- [ ] PowerPC 32-bit BE `powerpc`
- [ ] RISC-V 64-bit `riscv64`
| non_infrastructure | wireshark security update to cve ids cve cve other security advisory ids dsa descriptions wnpa sec asn ber and related dissectors crash bug cve wnpa sec tcap dissector crash bug cve wnpa sec rpcap dissector crash bug architectural progress bit optional environment armel powerpc bit be powerpc bit be powerpc risc v bit | 0 |
34,058 | 4,889,706,980 | IssuesEvent | 2016-11-18 11:08:58 | geopython/pycsw | https://api.github.com/repos/geopython/pycsw | closed | Remote tests failing | bug harvesting tests | # Description
While working on #428 I'm running into an issue where two tests of the `harvesting` suite are failing. The tests and a diff of expected vs result are:
1. `Harvest-zzz-post-GetRecords-filter-wfs-iso`
```
--- Harvest-zzz-post-GetRecords-filter-wfs-iso/expected.xml 2016-11-16 16:49:08.005763367 +0000
+++ Harvest-zzz-post-GetRecords-filter-wfs-iso/response.xml 2016-11-16 16:51:15.607320617 +0000
@@ -703,26 +703,6 @@
<srv:containsOperations>
<srv:SV_OperationMetadata>
<srv:operationName>
- <gco:CharacterString>DefaultMaxFeatures</gco:CharacterString>
- </srv:operationName>
- <srv:DCP>
- <srv:DCPList codeList="http://www.isotc211.org/2005/resources/Codelist/gmxCodelists.xml#DCPList" codeListValue="HTTPGet">HTTPGet</srv:DCPList>
- </srv:DCP>
- <srv:DCP>
- <srv:DCPList codeList="http://www.isotc211.org/2005/resources/Codelist/gmxCodelists.xml#DCPList" codeListValue="HTTPPost">HTTPPost</srv:DCPList>
- </srv:DCP>
- <srv:connectPoint>
- <gmd:CI_OnlineResource>
- <gmd:linkage>
- <gmd:URL>http://maps.cera.govt.nz/arcgis/services/CERA/CERA_TechClasses_WGS84/MapServer/WFSServer</gmd:URL>
- </gmd:linkage>
- </gmd:CI_OnlineResource>
- </srv:connectPoint>
- </srv:SV_OperationMetadata>
- </srv:containsOperations>
- <srv:containsOperations>
- <srv:SV_OperationMetadata>
- <srv:operationName>
<gco:CharacterString>ImplementsSimpleWFS</gco:CharacterString>
</srv:operationName>
<srv:DCP>
```
2. `Harvest-zzz-post-GetRecords-filter-wms-iso`
```
--- expected.xml 2016-11-16 17:05:23.287145828 +0000
+++ response.xml 2016-11-16 17:06:47.897533902 +0000
@@ -1103,7 +1103,7 @@
</gmd:address>
<gmd:onlineResource>
<gmd:linkage>
- <gmd:URL>http://mesonet.agron.iastate.edu/cgi-bin/wms/nexrad/n0r-t.cgi?</gmd:URL>
+ <gmd:URL>https://mesonet.agron.iastate.edu/cgi-bin/wms/nexrad/n0r-t.cgi?</gmd:URL>
</gmd:linkage>
</gmd:onlineResource>
</gmd:CI_Contact>
```
It seems like the remote servers have made some changes to the records being harvested. If so, perhaps we should update the expected files to reflect this. I am not sure of this assumption though.
# Environment
- operating system:
- Python version:
- pycsw version:
- source/distribution
- [x] git clone
- [ ] DebianGIS/UbuntuGIS
- [ ] PyPI
- [ ] zip/tar.gz
- [ ] other (please specify):
- web server
- [x] Apache/mod_wsgi
- [ ] CGI
- [ ] other (please specify):
# Steps to Reproduce
* Clone the latest pycsw master and install dev dependencies.
* Run the remote tests with
```
paver test -r
```
There should be two failing tests
# Additional Information
I am attaching the full results and responses here:
[failing_harvesting_tests.zip](https://github.com/geopython/pycsw/files/595370/failing_harvesting_tests.zip)
| 1.0 | Remote tests failing - # Description
While working on #428 I'm running into an issue where two tests of the `harvesting` suite are failing. The tests and a diff of expected vs result are:
1. `Harvest-zzz-post-GetRecords-filter-wfs-iso`
```
--- Harvest-zzz-post-GetRecords-filter-wfs-iso/expected.xml 2016-11-16 16:49:08.005763367 +0000
+++ Harvest-zzz-post-GetRecords-filter-wfs-iso/response.xml 2016-11-16 16:51:15.607320617 +0000
@@ -703,26 +703,6 @@
<srv:containsOperations>
<srv:SV_OperationMetadata>
<srv:operationName>
- <gco:CharacterString>DefaultMaxFeatures</gco:CharacterString>
- </srv:operationName>
- <srv:DCP>
- <srv:DCPList codeList="http://www.isotc211.org/2005/resources/Codelist/gmxCodelists.xml#DCPList" codeListValue="HTTPGet">HTTPGet</srv:DCPList>
- </srv:DCP>
- <srv:DCP>
- <srv:DCPList codeList="http://www.isotc211.org/2005/resources/Codelist/gmxCodelists.xml#DCPList" codeListValue="HTTPPost">HTTPPost</srv:DCPList>
- </srv:DCP>
- <srv:connectPoint>
- <gmd:CI_OnlineResource>
- <gmd:linkage>
- <gmd:URL>http://maps.cera.govt.nz/arcgis/services/CERA/CERA_TechClasses_WGS84/MapServer/WFSServer</gmd:URL>
- </gmd:linkage>
- </gmd:CI_OnlineResource>
- </srv:connectPoint>
- </srv:SV_OperationMetadata>
- </srv:containsOperations>
- <srv:containsOperations>
- <srv:SV_OperationMetadata>
- <srv:operationName>
<gco:CharacterString>ImplementsSimpleWFS</gco:CharacterString>
</srv:operationName>
<srv:DCP>
```
2. `Harvest-zzz-post-GetRecords-filter-wms-iso`
```
--- expected.xml 2016-11-16 17:05:23.287145828 +0000
+++ response.xml 2016-11-16 17:06:47.897533902 +0000
@@ -1103,7 +1103,7 @@
</gmd:address>
<gmd:onlineResource>
<gmd:linkage>
- <gmd:URL>http://mesonet.agron.iastate.edu/cgi-bin/wms/nexrad/n0r-t.cgi?</gmd:URL>
+ <gmd:URL>https://mesonet.agron.iastate.edu/cgi-bin/wms/nexrad/n0r-t.cgi?</gmd:URL>
</gmd:linkage>
</gmd:onlineResource>
</gmd:CI_Contact>
```
It seems like the remote servers have made some changes to the records being harvested. If so, perhaps we should update the expected files to reflect this. I am not sure of this assumption though.
# Environment
- operating system:
- Python version:
- pycsw version:
- source/distribution
- [x] git clone
- [ ] DebianGIS/UbuntuGIS
- [ ] PyPI
- [ ] zip/tar.gz
- [ ] other (please specify):
- web server
- [x] Apache/mod_wsgi
- [ ] CGI
- [ ] other (please specify):
# Steps to Reproduce
* Clone the latest pycsw master and install dev dependencies.
* Run the remote tests with
```
paver test -r
```
There should be two failing tests
# Additional Information
I am attaching the full results and responses here:
[failing_harvesting_tests.zip](https://github.com/geopython/pycsw/files/595370/failing_harvesting_tests.zip)
| non_infrastructure | remote tests failing description while working on i m running into an issue where two tests of the harvesting suite are failing the tests and a diff of expected vs result are harvest zzz post getrecords filter wfs iso harvest zzz post getrecords filter wfs iso expected xml harvest zzz post getrecords filter wfs iso response xml defaultmaxfeatures httpget httppost implementssimplewfs harvest zzz post getrecords filter wms iso expected xml response xml it seems like the remote servers have made some changes to the records being harvested if so perhaps we should update the expected files to reflect this i am not sure of this assumption though environment operating system python version pycsw version source distribution git clone debiangis ubuntugis pypi zip tar gz other please specify web server apache mod wsgi cgi other please specify steps to reproduce clone the latest pycsw master and install dev dependencies run the remote tests with paver test r there should be two failing tests additional information i am attaching the full results and responses here | 0 |
105,086 | 4,230,264,035 | IssuesEvent | 2016-07-04 11:16:38 | GoogleCloudPlatform/gcloud-eclipse-tools | https://api.github.com/repos/GoogleCloudPlatform/gcloud-eclipse-tools | opened | Install the correct libraries in non-maven projects | enhancement high priority | I assume these are the same three the AppEngine maven archetype adds:

This should be done in AppEngineSdkClasspathContainer | 1.0 | Install the correct libraries in non-maven projects - I assume these are the same three the AppEngine maven archetype adds:

This should be done in AppEngineSdkClasspathContainer | non_infrastructure | install the correct libraries in non maven projects i assume these are the same three the appengine maven archetype adds this should be done in appenginesdkclasspathcontainer | 0 |
247,079 | 26,677,348,280 | IssuesEvent | 2023-01-26 15:11:25 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Filebeat/auditd - Enable Grok regex for "ENRICHED" log_format with auditd | enhancement Filebeat Stalled Team:Security-External Integrations | **Describe the enhancement:**
Using filebeat 7.6.2, auditd module. Unsure if this qualifies as bug or enhancement.
On Enterprise Linux 7 based systems (all I have ATM, this should probably be tested elsewhere), when setting `log_format = ENRICHED` in the system auditd config, the filebeat auditd module will by default choke on the append point for the additional human readable values. It ends up reporting `auditd.log.res` (or `event.outcome`) as something like `success'UID="root`. This is because when enabling `ENRICHED` logging, auditd places an unprintable char after the single-quote, followed directly by the new values. This breaks whitespace-based splits and causes the above garbage in the output.
After some digging, I've gotten this working on my stack by adjusting the relevant Grok pattern as follows (line 14 of `modules/auditd/log/ingest/pipeline.yml`):
`'%{AUDIT_PREFIX} %{AUDIT_KEY_VALUES:auditd.log.kv} msg=[''"](%{DATA:auditd.log.msg}\s+)?%{AUDIT_KEY_VALUES:auditd.log.sub_kv}[''"]\x1d?%{AUDIT_KEY_VALUES:auditd.log.sub_sub_kv}'`
And adding a new `kv` processor for the additional key:value set:
```
- kv:
field: auditd.log.sub_sub_kv
field_split: "\\s+"
value_split: "="
target_field: auditd.log
ignore_missing: true
```
In the above grok, the `\x1d` maps to the ASCII char auditd injects into the log line.
**Describe a specific use case for the enhancement or feature:**
On (at least) Red Hat and CentOS systems running auditd, it can be useful to enable `ENRICHED` logging to place human-readable user and group names directly into the log stream. This however will break the auditd modules parsing of said log, so out of the box one must choose one or the other.
| True | Filebeat/auditd - Enable Grok regex for "ENRICHED" log_format with auditd - **Describe the enhancement:**
Using filebeat 7.6.2, auditd module. Unsure if this qualifies as bug or enhancement.
On Enterprise Linux 7 based systems (all I have ATM, this should probably be tested elsewhere), when setting `log_format = ENRICHED` in the system auditd config, the filebeat auditd module will by default choke on the append point for the additional human readable values. It ends up reporting `auditd.log.res` (or `event.outcome`) as something like `success'UID="root`. This is because when enabling `ENRICHED` logging, auditd places an unprintable char after the single-quote, followed directly by the new values. This breaks whitespace-based splits and causes the above garbage in the output.
After some digging, I've gotten this working on my stack by adjusting the relevant Grok pattern as follows (line 14 of `modules/auditd/log/ingest/pipeline.yml`):
`'%{AUDIT_PREFIX} %{AUDIT_KEY_VALUES:auditd.log.kv} msg=[''"](%{DATA:auditd.log.msg}\s+)?%{AUDIT_KEY_VALUES:auditd.log.sub_kv}[''"]\x1d?%{AUDIT_KEY_VALUES:auditd.log.sub_sub_kv}'`
And adding a new `kv` processor for the additional key:value set:
```
- kv:
field: auditd.log.sub_sub_kv
field_split: "\\s+"
value_split: "="
target_field: auditd.log
ignore_missing: true
```
In the above grok, the `\x1d` maps to the ASCII char auditd injects into the log line.
**Describe a specific use case for the enhancement or feature:**
On (at least) Red Hat and CentOS systems running auditd, it can be useful to enable `ENRICHED` logging to place human-readable user and group names directly into the log stream. This however will break the auditd modules parsing of said log, so out of the box one must choose one or the other.
| non_infrastructure | filebeat auditd enable grok regex for enriched log format with auditd describe the enhancement using filebeat auditd module unsure if this qualifies as bug or enhancement on enterprise linux based systems all i have atm this should probably be tested elsewhere when setting log format enriched in the system auditd config the filebeat auditd module will by default choke on the append point for the additional human readable values it ends up reporting auditd log res or event outcome as something like success uid root this is because when enabling enriched logging auditd places an unprintable char after the single quote followed directly by the new values this breaks whitespace based splits and causes the above garbage in the output after some digging i ve gotten this working on my stack by adjusting the relevant grok pattern as follows line of modules auditd log ingest pipeline yml audit prefix audit key values auditd log kv msg data auditd log msg s audit key values auditd log sub kv audit key values auditd log sub sub kv and adding a new kv processor for the additional key value set kv field auditd log sub sub kv field split s value split target field auditd log ignore missing true in the above grok the maps to the ascii char auditd injects into the log line describe a specific use case for the enhancement or feature on at least red hat and centos systems running auditd it can be useful to enable enriched logging to place human readable user and group names directly into the log stream this however will break the auditd modules parsing of said log so out of the box one must choose one or the other | 0 |
197,695 | 6,962,837,838 | IssuesEvent | 2017-12-08 15:15:17 | SoylentNews/rehash | https://api.github.com/repos/SoylentNews/rehash | closed | Nagger doesn't update after submission | Bug: Non-Critical Priority: Low | _From @lhsi on August 18, 2014 13:18_
The nagger doesn't update after a submission, you need to go to a new page to see the new total.
Step to reproduce: Submit an article submission when there are less than 20 stories in the queue (so the nagger line appears at the top of the page). You are taken to the "Submission received" page (the one that confirms that the submission was received).
Expected: The "Number of stories in the queue" has gone up by one with a new story
Actual: The number remains unchanged. Going to a new page has a new number
Workaround: Go to a new page to see the updated number.
_Copied from original issue: SoylentNews/slashcode#281_
| 1.0 | Nagger doesn't update after submission - _From @lhsi on August 18, 2014 13:18_
The nagger doesn't update after a submission, you need to go to a new page to see the new total.
Step to reproduce: Submit an article submission when there are less than 20 stories in the queue (so the nagger line appears at the top of the page). You are taken to the "Submission received" page (the one that confirms that the submission was received).
Expected: The "Number of stories in the queue" has gone up by one with a new story
Actual: The number remains unchanged. Going to a new page has a new number
Workaround: Go to a new page to see the updated number.
_Copied from original issue: SoylentNews/slashcode#281_
| non_infrastructure | nagger doesn t update after submission from lhsi on august the nagger doesn t update after a submission you need to go to a new page to see the new total step to reproduce submit an article submission when there are less than stories in the queue so the nagger line appears at the top of the page you are taken to the submission received page the one that confirms that the submission was received expected the number of stories in the queue has gone up by one with a new story actual the number remains unchanged going to a new page has a new number workaround go to a new page to see the updated number copied from original issue soylentnews slashcode | 0 |
39,376 | 8,637,583,836 | IssuesEvent | 2018-11-23 11:47:21 | openhealthcare/elcid-rfh | https://api.github.com/repos/openhealthcare/elcid-rfh | closed | Add Assessment and a reason for interaction | Code Review | Adds an assessment text field and a reason for interaction of "LTBI Assessment" to the inline clinical discussion form. | 1.0 | Add Assessment and a reason for interaction - Adds an assessment text field and a reason for interaction of "LTBI Assessment" to the inline clinical discussion form. | non_infrastructure | add assessment and a reason for interaction adds an assessment text field and a reason for interaction of ltbi assessment to the inline clinical discussion form | 0 |
232,715 | 7,674,159,540 | IssuesEvent | 2018-05-15 02:14:07 | kubeflow/kubeflow | https://api.github.com/repos/kubeflow/kubeflow | closed | Central UI Ambassador Integration | area/front-end priority/p1 | We need to integrate the Central UI with Ambassador.
/cc @swiftdiaries | 1.0 | Central UI Ambassador Integration - We need to integrate the Central UI with Ambassador.
/cc @swiftdiaries | non_infrastructure | central ui ambassador integration we need to integrate the central ui with ambassador cc swiftdiaries | 0 |
36,608 | 12,417,835,236 | IssuesEvent | 2020-05-22 21:49:21 | TIBCOSoftware/labs-air | https://api.github.com/repos/TIBCOSoftware/labs-air | closed | WS-2018-0232 (Medium) detected in underscore.string-2.3.3.tgz | security vulnerability | ## WS-2018-0232 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-2.3.3.tgz</b></p></summary>
<p>String manipulation extensions for Underscore.js javascript library.</p>
<p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-2.3.3.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-2.3.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/labs-air/ui/ProjectAir/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/labs-air/ui/ProjectAir/node_modules/underscore.string/package.json</p>
<p>
Dependency Hierarchy:
- component-1.1.0.tgz (Root Library)
- component-resolver-1.3.0.tgz
- component-downloader-1.2.0.tgz
- decompress-0.2.5.tgz
- ext-name-1.0.1.tgz
- :x: **underscore.string-2.3.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TIBCOSoftware/labs-air/commit/91b4245b987bd63abaeb49223360180947327d49">91b4245b987bd63abaeb49223360180947327d49</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Underscore.string, before 3.3.5, is vulnerable to Regular Expression Denial of Service (ReDoS).
<p>Publish Date: 2018-12-30
<p>URL: <a href=https://github.com/epeli/underscore.string/commit/f486cd684c94c12db48b45d52b1472a1b9661029>WS-2018-0232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/745">https://www.npmjs.com/advisories/745</a></p>
<p>Release Date: 2018-12-30</p>
<p>Fix Resolution: 3.3.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore.string","packageVersion":"2.3.3","isTransitiveDependency":true,"dependencyTree":"component:1.1.0;component-resolver:1.3.0;component-downloader:1.2.0;decompress:0.2.5;ext-name:1.0.1;underscore.string:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.3.5"}],"vulnerabilityIdentifier":"WS-2018-0232","vulnerabilityDetails":"Underscore.string, before 3.3.5, is vulnerable to Regular Expression Denial of Service (ReDoS).","vulnerabilityUrl":"https://github.com/epeli/underscore.string/commit/f486cd684c94c12db48b45d52b1472a1b9661029","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> --> | True | WS-2018-0232 (Medium) detected in underscore.string-2.3.3.tgz - ## WS-2018-0232 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-2.3.3.tgz</b></p></summary>
<p>String manipulation extensions for Underscore.js javascript library.</p>
<p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-2.3.3.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-2.3.3.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/labs-air/ui/ProjectAir/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/labs-air/ui/ProjectAir/node_modules/underscore.string/package.json</p>
<p>
Dependency Hierarchy:
- component-1.1.0.tgz (Root Library)
- component-resolver-1.3.0.tgz
- component-downloader-1.2.0.tgz
- decompress-0.2.5.tgz
- ext-name-1.0.1.tgz
- :x: **underscore.string-2.3.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TIBCOSoftware/labs-air/commit/91b4245b987bd63abaeb49223360180947327d49">91b4245b987bd63abaeb49223360180947327d49</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Underscore.string, before 3.3.5, is vulnerable to Regular Expression Denial of Service (ReDoS).
<p>Publish Date: 2018-12-30
<p>URL: <a href=https://github.com/epeli/underscore.string/commit/f486cd684c94c12db48b45d52b1472a1b9661029>WS-2018-0232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/745">https://www.npmjs.com/advisories/745</a></p>
<p>Release Date: 2018-12-30</p>
<p>Fix Resolution: 3.3.5</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore.string","packageVersion":"2.3.3","isTransitiveDependency":true,"dependencyTree":"component:1.1.0;component-resolver:1.3.0;component-downloader:1.2.0;decompress:0.2.5;ext-name:1.0.1;underscore.string:2.3.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.3.5"}],"vulnerabilityIdentifier":"WS-2018-0232","vulnerabilityDetails":"Underscore.string, before 3.3.5, is vulnerable to Regular Expression Denial of Service (ReDoS).","vulnerabilityUrl":"https://github.com/epeli/underscore.string/commit/f486cd684c94c12db48b45d52b1472a1b9661029","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> --> | non_infrastructure | ws medium detected in underscore string tgz ws medium severity vulnerability vulnerable library underscore string tgz string manipulation extensions for underscore js javascript library library home page a href path to dependency file tmp ws scm labs air ui projectair package json path to vulnerable library tmp ws scm labs air ui projectair node modules underscore string package json dependency hierarchy component tgz root library component resolver tgz component downloader tgz decompress tgz ext name tgz x underscore string tgz vulnerable library found in head commit a href vulnerability details underscore string before is vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails underscore string before is vulnerable to regular expression denial of service redos vulnerabilityurl | 0 |
4,243 | 4,926,866,579 | IssuesEvent | 2016-11-26 12:01:27 | Integreat/cms | https://api.github.com/repos/Integreat/cms | opened | Performance Profiling | area:infrastructure ease:medium prio:medium | Analyse how well we can handle multiple concurrent accesses and where our performance bottlenecks are.
This includes the following access points:
* API access
* WordPress Backend Edits
* Webview access
The [P3 plugin](https://wordpress.org/plugins/p3-profiler/) might be a useful tool.
Ideally, the results tell us:
* how many concurrent accesses on each access point we can handle at the moment
* whether the bottleneck is PHP or MySQL
* where exactly most of the computations happen (which PHP function / which MySQL table) | 1.0 | Performance Profiling - Analyse how well we can handle multiple concurrent accesses and where our performance bottlenecks are.
This includes the following access points:
* API access
* WordPress Backend Edits
* Webview access
The [P3 plugin](https://wordpress.org/plugins/p3-profiler/) might be a useful tool.
Ideally, the results tell us:
* how many concurrent accesses on each access point we can handle at the moment
* whether the bottleneck is PHP or MySQL
* where exactly most of the computations happen (which PHP function / which MySQL table) | infrastructure | performance profiling analyse how well we can handle multiple concurrent accesses and where our performance bottlenecks are this includes the following access points api access wordpress backend edits webview access the might be a useful tool ideally the results tell us how many concurrent accesses on each access point we can handle at the moment whether the bottleneck is php or mysql where exactly most of the computations happen which php function which mysql table | 1 |
16,504 | 11,996,895,644 | IssuesEvent | 2020-04-08 17:34:04 | bootstrapworld/curriculum | https://api.github.com/repos/bootstrapworld/curriculum | opened | Path handling error on pathway narrative page | Infrastructure | The following directive should use a relative filepath (./images/Logo.png):
`@image{images/Logo.png, Bootstrap:Data Science}`
But instead it uses an absolute path (/images.Logo.png), resulting in a broken image | 1.0 | Path handling error on pathway narrative page - The following directive should use a relative filepath (./images/Logo.png):
`@image{images/Logo.png, Bootstrap:Data Science}`
But instead it uses an absolute path (/images.Logo.png), resulting in a broken image | infrastructure | path handling error on pathway narrative page the following directive should use a relative filepath images logo png image images logo png bootstrap data science but instead it uses an absolute path images logo png resulting in a broken image | 1 |
742,078 | 25,835,764,833 | IssuesEvent | 2022-12-12 19:30:21 | aseprite/aseprite | https://api.github.com/repos/aseprite/aseprite | closed | OS open/save file dialog | feature imported medium priority ui windows native macos | _From [davidcapello](https://code.google.com/u/davidcapello/) on January 24, 2014 17:21:34_
Open/save files with native Windows file dialog.
Note: The dialog should be customized to:
- [x] show the preview of each file (Windows and macOS)
- [x] show the list of formats/file types (Windows and macOS)
- [ ] ~~show the list of recent folders~~
_Original issue: http://code.google.com/p/aseprite/issues/detail?id=321_
| 1.0 | OS open/save file dialog - _From [davidcapello](https://code.google.com/u/davidcapello/) on January 24, 2014 17:21:34_
Open/save files with native Windows file dialog.
Note: The dialog should be customized to:
- [x] show the preview of each file (Windows and macOS)
- [x] show the list of formats/file types (Windows and macOS)
- [ ] ~~show the list of recent folders~~
_Original issue: http://code.google.com/p/aseprite/issues/detail?id=321_
| non_infrastructure | os open save file dialog from on january open save files with native windows file dialog note the dialog should be customized to show the preview of each file windows and macos show the list of formats file types windows and macos show the list of recent folders original issue | 0 |
312,347 | 9,546,216,709 | IssuesEvent | 2019-05-01 19:16:55 | mozilla/addons-code-manager | https://api.github.com/repos/mozilla/addons-code-manager | closed | Add a reducer for AccordionMenu | priority: p3 | An AccordionMenu component will be added in https://github.com/mozilla/addons-code-manager/issues/667
This will be connected to Redux state so it needs a reducer. | 1.0 | Add a reducer for AccordionMenu - An AccordionMenu component will be added in https://github.com/mozilla/addons-code-manager/issues/667
This will be connected to Redux state so it needs a reducer. | non_infrastructure | add a reducer for accordionmenu an accordionmenu component will be added in this will be connected to redux state so it needs a reducer | 0 |
48,030 | 19,906,543,726 | IssuesEvent | 2022-01-25 13:23:23 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | Plugin Crash during Terraform plan | bug crash service/ec2 | <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
The plans in one of my workspaces (and only one - even though I have others which use the same modules and look extremely similar) is failing with a message about the aws plugin crashing. I reverted my plan back to a known good plan which ran successfully 3 days ago, but this also fails.
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
Originally I was using terraform 0.14.4 and aws provider 3.27 and having the issue, but see the error in terraform 1.0.0 and aws provider 3.45 as well. The error messages look a little different, but no version seems to run.
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
I can't tell what resource is affected.
### Terraform Configuration Files
To keep things brief I am just including one of several calls to the "ec2" module. This configuration just runs a bunch of very similar calls to this module to create several similar ec2 instances.
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
terraform {
required_providers {
aws = {
source = "hashicorp/aws"
version = "3.45.0"
}
}
backend "remote" {
organization = "PotomacEconomics"
workspaces {
name = "compute-prod-pe-us-east-1"
}
}
}
provider "aws" {
region = "us-east-1"
}
#
# Get information needed for configuration
#
data "aws_security_group" "default_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-default"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "windows_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-windows"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "dc_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-dc"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "commonfileserver_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-commonfileserver"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "securityserver_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-securityserver"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
#
# APP1
#
module "ec2_app1" {
source = "app.terraform.io/PotomacEconomics/ec2/aws"
version = "1.0.30"
environment_name = var.environment_name
environment_type = var.environment_type
instance_suffix = "app1"
availability_zone_number = "1"
instance_type = "t3.2xlarge"
ami_id = "ami-xxx"
root_volume_size = 150
secondary_volume_size = 1
iam_instance_profile = "Windows-Server-Role"
private_ip = "x.x.x.x"
vpc_security_group_ids = [
data.aws_security_group.default_security_group.id,
data.aws_security_group.windows_security_group.id
]
backup_plan = "${var.environment_name}_daily"
}
```
### Debug Output
https://gist.github.com/JTP3XP/deb5db6a0201ff5f68a542c2100ee69f
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Panic Output
I'm running this in Terraform Cloud, so I don't know how to get the crash.log
<!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. --->
### Expected Behavior
Plan runs as it did before
<!--- What should have happened? --->
### Actual Behavior
Terraform crashes and blames the aws plugin
<!--- What actually happened? --->
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform plan`
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
* #0000
| 1.0 | Plugin Crash during Terraform plan - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
The plans in one of my workspaces (and only one - even though I have others which use the same modules and look extremely similar) is failing with a message about the aws plugin crashing. I reverted my plan back to a known good plan which ran successfully 3 days ago, but this also fails.
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform CLI and Terraform AWS Provider Version
Originally I was using terraform 0.14.4 and aws provider 3.27 and having the issue, but see the error in terraform 1.0.0 and aws provider 3.45 as well. The error messages look a little different, but no version seems to run.
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
I can't tell what resource is affected.
### Terraform Configuration Files
To keep things brief I am just including one of several calls to the "ec2" module. This configuration just runs a bunch of very similar calls to this module to create several similar ec2 instances.
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
terraform {
required_providers {
aws = {
source = "hashicorp/aws"
version = "3.45.0"
}
}
backend "remote" {
organization = "PotomacEconomics"
workspaces {
name = "compute-prod-pe-us-east-1"
}
}
}
provider "aws" {
region = "us-east-1"
}
#
# Get information needed for configuration
#
data "aws_security_group" "default_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-default"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "windows_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-windows"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "dc_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-dc"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "commonfileserver_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-commonfileserver"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
data "aws_security_group" "securityserver_security_group" {
tags = {
Name = "sgroup-${var.environment_type}-${var.environment_name}-securityserver"
Terraform = "true"
Environment_Name = var.environment_name
Environment_Type = var.environment_type
}
}
#
# APP1
#
module "ec2_app1" {
source = "app.terraform.io/PotomacEconomics/ec2/aws"
version = "1.0.30"
environment_name = var.environment_name
environment_type = var.environment_type
instance_suffix = "app1"
availability_zone_number = "1"
instance_type = "t3.2xlarge"
ami_id = "ami-xxx"
root_volume_size = 150
secondary_volume_size = 1
iam_instance_profile = "Windows-Server-Role"
private_ip = "x.x.x.x"
vpc_security_group_ids = [
data.aws_security_group.default_security_group.id,
data.aws_security_group.windows_security_group.id
]
backup_plan = "${var.environment_name}_daily"
}
```
### Debug Output
https://gist.github.com/JTP3XP/deb5db6a0201ff5f68a542c2100ee69f
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Panic Output
I'm running this in Terraform Cloud, so I don't know how to get the crash.log
<!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. --->
### Expected Behavior
Plan runs as it did before
<!--- What should have happened? --->
### Actual Behavior
Terraform crashes and blames the aws plugin
<!--- What actually happened? --->
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform plan`
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in EC2 Classic? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example:
--->
* #0000
| non_infrastructure | plugin crash during terraform plan please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead the plans in one of my workspaces and only one even though i have others which use the same modules and look extremely similar is failing with a message about the aws plugin crashing i reverted my plan back to a known good plan which ran successfully days ago but this also fails community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform cli and terraform aws provider version originally i was using terraform and aws provider and having the issue but see the error in terraform and aws provider as well the error messages look a little different but no version seems to run affected resource s i can t tell what resource is affected terraform configuration files to keep things brief i am just including one of several calls to the module this configuration just runs a bunch of very similar calls to this module to create several similar instances hcl terraform required providers aws source hashicorp aws version backend remote organization potomaceconomics workspaces name compute prod pe us east provider aws region us east get information needed for configuration data aws security group default security group tags name sgroup var environment type var environment name default terraform true environment name var environment name environment type var environment type data aws security group windows security group tags name sgroup var environment type var environment name windows terraform true environment name var environment name environment type var environment type data aws security group dc security group tags name sgroup var environment type var environment name dc terraform true environment name var environment name environment type var environment type data aws security group commonfileserver security group tags name sgroup var environment type var environment name commonfileserver terraform true environment name var environment name environment type var environment type data aws security group securityserver security group tags name sgroup var environment type var environment name securityserver terraform true environment name var environment name environment type var environment type module source app terraform io potomaceconomics aws version environment name var environment name environment type var environment type instance suffix availability zone number instance type ami id ami xxx root volume size secondary volume size iam instance profile windows server role private ip x x x x vpc security group ids data aws security group default security group id data aws security group windows security group id backup plan var environment name daily debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the panic output i m running this in terraform cloud so i don t know how to get the crash log expected behavior plan runs as it did before actual behavior terraform crashes and blames the aws plugin steps to reproduce terraform plan important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example | 0 |
27,613 | 21,993,579,662 | IssuesEvent | 2022-05-26 02:19:24 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Outerloop stress legs triggering on PRs they shouldn't | area-Infrastructure untriaged | The GCStress, JITStress, and outerloop legs look to be triggering on PRs they shouldn't.
https://github.com/dotnet/runtime/pull/69825
cc: @dotnet/runtime-infrastructure if there were any changes we made. @dotnet/dnceng if this is an AzDO issue we need to report. | 1.0 | Outerloop stress legs triggering on PRs they shouldn't - The GCStress, JITStress, and outerloop legs look to be triggering on PRs they shouldn't.
https://github.com/dotnet/runtime/pull/69825
cc: @dotnet/runtime-infrastructure if there were any changes we made. @dotnet/dnceng if this is an AzDO issue we need to report. | infrastructure | outerloop stress legs triggering on prs they shouldn t the gcstress jitstress and outerloop legs look to be triggering on prs they shouldn t cc dotnet runtime infrastructure if there were any changes we made dotnet dnceng if this is an azdo issue we need to report | 1 |
95,964 | 16,112,995,639 | IssuesEvent | 2021-04-28 01:19:51 | RG4421/terra-dev-site | https://api.github.com/repos/RG4421/terra-dev-site | opened | CVE-2021-23382 (Medium) detected in postcss-7.0.35.tgz, postcss-8.2.6.tgz | security vulnerability | ## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.35.tgz</b>, <b>postcss-8.2.6.tgz</b></p></summary>
<p>
<details><summary><b>postcss-7.0.35.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz</a></p>
<p>Path to dependency file: terra-dev-site/package.json</p>
<p>Path to vulnerable library: terra-dev-site/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-10.1.0.tgz (Root Library)
- :x: **postcss-7.0.35.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-8.2.6.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-8.2.6.tgz">https://registry.npmjs.org/postcss/-/postcss-8.2.6.tgz</a></p>
<p>Path to dependency file: terra-dev-site/package.json</p>
<p>Path to vulnerable library: terra-dev-site/node_modules/doiuse/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-config-terra-3.7.0.tgz (Root Library)
- stylelint-no-unsupported-browser-features-3.0.2.tgz
- doiuse-4.4.1.tgz
- :x: **postcss-8.2.6.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"7.0.35","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"stylelint:10.1.0;postcss:7.0.35","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"8.2.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"stylelint-config-terra:3.7.0;stylelint-no-unsupported-browser-features:3.0.2;doiuse:4.4.1;postcss:8.2.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-23382 (Medium) detected in postcss-7.0.35.tgz, postcss-8.2.6.tgz - ## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-7.0.35.tgz</b>, <b>postcss-8.2.6.tgz</b></p></summary>
<p>
<details><summary><b>postcss-7.0.35.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz</a></p>
<p>Path to dependency file: terra-dev-site/package.json</p>
<p>Path to vulnerable library: terra-dev-site/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-10.1.0.tgz (Root Library)
- :x: **postcss-7.0.35.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-8.2.6.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-8.2.6.tgz">https://registry.npmjs.org/postcss/-/postcss-8.2.6.tgz</a></p>
<p>Path to dependency file: terra-dev-site/package.json</p>
<p>Path to vulnerable library: terra-dev-site/node_modules/doiuse/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- stylelint-config-terra-3.7.0.tgz (Root Library)
- stylelint-no-unsupported-browser-features-3.0.2.tgz
- doiuse-4.4.1.tgz
- :x: **postcss-8.2.6.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"7.0.35","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"stylelint:10.1.0;postcss:7.0.35","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"8.2.6","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"stylelint-config-terra:3.7.0;stylelint-no-unsupported-browser-features:3.0.2;doiuse:4.4.1;postcss:8.2.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve medium detected in postcss tgz postcss tgz cve medium severity vulnerability vulnerable libraries postcss tgz postcss tgz postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file terra dev site package json path to vulnerable library terra dev site node modules postcss package json dependency hierarchy stylelint tgz root library x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file terra dev site package json path to vulnerable library terra dev site node modules doiuse node modules postcss package json dependency hierarchy stylelint config terra tgz root library stylelint no unsupported browser features tgz doiuse tgz x postcss tgz vulnerable library found in base branch master vulnerability details the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree stylelint postcss isminimumfixversionavailable true minimumfixversion postcss packagetype javascript node js packagename postcss packageversion packagefilepaths istransitivedependency true dependencytree stylelint config terra stylelint no unsupported browser features doiuse postcss isminimumfixversionavailable true minimumfixversion postcss basebranches vulnerabilityidentifier cve vulnerabilitydetails the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl vulnerabilityurl | 0 |
53,953 | 6,353,660,927 | IssuesEvent | 2017-07-29 01:09:41 | katiekeel/little-shop | https://api.github.com/repos/katiekeel/little-shop | closed | Story 12: Viewing past orders | Controller Model Testing View | - Feature tests for: visiting /orders when logged in and seeing orders
- Order model that is populated off of cart/session
- Order/order_users/item_orders migration
- Controller actions for authorized /orders, Order creation | 1.0 | Story 12: Viewing past orders - - Feature tests for: visiting /orders when logged in and seeing orders
- Order model that is populated off of cart/session
- Order/order_users/item_orders migration
- Controller actions for authorized /orders, Order creation | non_infrastructure | story viewing past orders feature tests for visiting orders when logged in and seeing orders order model that is populated off of cart session order order users item orders migration controller actions for authorized orders order creation | 0 |
169,975 | 20,841,996,950 | IssuesEvent | 2022-03-21 02:03:06 | directoryxx/Belajar-Microservice-1 | https://api.github.com/repos/directoryxx/Belajar-Microservice-1 | opened | CVE-2022-24771 (High) detected in node-forge-0.10.0.tgz | security vulnerability | ## CVE-2022-24771 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /front/package.json</p>
<p>Path to vulnerable library: /front/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.0.tgz (Root Library)
- webpack-dev-server-3.10.2.tgz
- selfsigned-1.10.8.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code is lenient in checking the digest algorithm structure. This can allow a crafted structure that steals padding bytes and uses unchecked portion of the PKCS#1 encoded message to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24771>CVE-2022-24771</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-24771 (High) detected in node-forge-0.10.0.tgz - ## CVE-2022-24771 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /front/package.json</p>
<p>Path to vulnerable library: /front/node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.0.tgz (Root Library)
- webpack-dev-server-3.10.2.tgz
- selfsigned-1.10.8.tgz
- :x: **node-forge-0.10.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code is lenient in checking the digest algorithm structure. This can allow a crafted structure that steals padding bytes and uses unchecked portion of the PKCS#1 encoded message to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24771>CVE-2022-24771</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in node forge tgz cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file front package json path to vulnerable library front node modules node forge package json dependency hierarchy react scripts tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library found in base branch master vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code is lenient in checking the digest algorithm structure this can allow a crafted structure that steals padding bytes and uses unchecked portion of the pkcs encoded message to forge a signature when a low public exponent is being used the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge step up your open source security game with whitesource | 0 |
17,025 | 12,195,860,248 | IssuesEvent | 2020-04-29 18:05:04 | dotnet/docker-tools | https://api.github.com/repos/dotnet/docker-tools | closed | NullRefEx thrown when running the build command with the dry run option | area:infrastructure bug triaged | ```
System.NullReferenceException: Object reference not set to an instance of an object.
at Microsoft.DotNet.ImageBuilder.ExecuteHelper.Execute(ProcessStartInfo info, Func`2 executor, Boolean isDryRun, String errorMessage, String executeMessageOverride) in /image-builder/src/ExecuteHelper.cs:line 101
at Microsoft.DotNet.ImageBuilder.ExecuteHelper.ExecuteWithRetry(ProcessStartInfo info, Action`1 processStartedCallback, Boolean isDryRun, String errorMessage, String executeMessageOverride) in /image-builder/src/ExecuteHelper.cs:line 56
at Microsoft.DotNet.ImageBuilder.ExecuteHelper.ExecuteWithRetry(String fileName, String args, Boolean isDryRun, String errorMessage, String executeMessageOverride) in /image-builder/src/ExecuteHelper.cs:line 41
at Microsoft.DotNet.ImageBuilder.DockerHelper.PullImage(String image, Boolean isDryRun) in /image-builder/src/DockerHelper.cs:line 97
at Microsoft.DotNet.ImageBuilder.DockerService.PullImage(String image, Boolean isDryRun) in /image-builder/src/DockerService.cs:line 25
at Microsoft.DotNet.ImageBuilder.DockerServiceExtensions.PullBaseImages(IDockerService dockerService, ManifestInfo manifest, Boolean isDryRun) in /image-builder/src/DockerServiceExtensions.cs:line 21
at Microsoft.DotNet.ImageBuilder.Commands.BuildCommand.PullBaseImages() in /image-builder/src/Commands/BuildCommand.cs:line 314
at Microsoft.DotNet.ImageBuilder.Commands.BuildCommand.ExecuteAsync() in /image-builder/src/Commands/BuildCommand.cs:line 43
at Microsoft.DotNet.ImageBuilder.ImageBuilder.Main(String[] args) in /image-builder/src/ImageBuilder.cs:line 66
```
This looks to be a regression caused by https://github.com/dotnet/docker-tools/pull/430 | 1.0 | NullRefEx thrown when running the build command with the dry run option - ```
System.NullReferenceException: Object reference not set to an instance of an object.
at Microsoft.DotNet.ImageBuilder.ExecuteHelper.Execute(ProcessStartInfo info, Func`2 executor, Boolean isDryRun, String errorMessage, String executeMessageOverride) in /image-builder/src/ExecuteHelper.cs:line 101
at Microsoft.DotNet.ImageBuilder.ExecuteHelper.ExecuteWithRetry(ProcessStartInfo info, Action`1 processStartedCallback, Boolean isDryRun, String errorMessage, String executeMessageOverride) in /image-builder/src/ExecuteHelper.cs:line 56
at Microsoft.DotNet.ImageBuilder.ExecuteHelper.ExecuteWithRetry(String fileName, String args, Boolean isDryRun, String errorMessage, String executeMessageOverride) in /image-builder/src/ExecuteHelper.cs:line 41
at Microsoft.DotNet.ImageBuilder.DockerHelper.PullImage(String image, Boolean isDryRun) in /image-builder/src/DockerHelper.cs:line 97
at Microsoft.DotNet.ImageBuilder.DockerService.PullImage(String image, Boolean isDryRun) in /image-builder/src/DockerService.cs:line 25
at Microsoft.DotNet.ImageBuilder.DockerServiceExtensions.PullBaseImages(IDockerService dockerService, ManifestInfo manifest, Boolean isDryRun) in /image-builder/src/DockerServiceExtensions.cs:line 21
at Microsoft.DotNet.ImageBuilder.Commands.BuildCommand.PullBaseImages() in /image-builder/src/Commands/BuildCommand.cs:line 314
at Microsoft.DotNet.ImageBuilder.Commands.BuildCommand.ExecuteAsync() in /image-builder/src/Commands/BuildCommand.cs:line 43
at Microsoft.DotNet.ImageBuilder.ImageBuilder.Main(String[] args) in /image-builder/src/ImageBuilder.cs:line 66
```
This looks to be a regression caused by https://github.com/dotnet/docker-tools/pull/430 | infrastructure | nullrefex thrown when running the build command with the dry run option system nullreferenceexception object reference not set to an instance of an object at microsoft dotnet imagebuilder executehelper execute processstartinfo info func executor boolean isdryrun string errormessage string executemessageoverride in image builder src executehelper cs line at microsoft dotnet imagebuilder executehelper executewithretry processstartinfo info action processstartedcallback boolean isdryrun string errormessage string executemessageoverride in image builder src executehelper cs line at microsoft dotnet imagebuilder executehelper executewithretry string filename string args boolean isdryrun string errormessage string executemessageoverride in image builder src executehelper cs line at microsoft dotnet imagebuilder dockerhelper pullimage string image boolean isdryrun in image builder src dockerhelper cs line at microsoft dotnet imagebuilder dockerservice pullimage string image boolean isdryrun in image builder src dockerservice cs line at microsoft dotnet imagebuilder dockerserviceextensions pullbaseimages idockerservice dockerservice manifestinfo manifest boolean isdryrun in image builder src dockerserviceextensions cs line at microsoft dotnet imagebuilder commands buildcommand pullbaseimages in image builder src commands buildcommand cs line at microsoft dotnet imagebuilder commands buildcommand executeasync in image builder src commands buildcommand cs line at microsoft dotnet imagebuilder imagebuilder main string args in image builder src imagebuilder cs line this looks to be a regression caused by | 1 |
38,726 | 5,000,377,601 | IssuesEvent | 2016-12-10 09:00:59 | syuilo/misskey-web | https://api.github.com/repos/syuilo/misskey-web | opened | Display message when loading timeline | design site: mobile | <!--
Misskeyへの貢献ありがとうございます。
バグの報告、提案などで、可能なら以下の情報を含めてください。
* お使いのブラウザ
* デスクトップ版Misskeyかモバイル版Misskeyか
-->
| 1.0 | Display message when loading timeline - <!--
Misskeyへの貢献ありがとうございます。
バグの報告、提案などで、可能なら以下の情報を含めてください。
* お使いのブラウザ
* デスクトップ版Misskeyかモバイル版Misskeyか
-->
| non_infrastructure | display message when loading timeline misskeyへの貢献ありがとうございます。 バグの報告、提案などで、可能なら以下の情報を含めてください。 お使いのブラウザ デスクトップ版misskeyかモバイル版misskeyか | 0 |
462,972 | 13,257,405,153 | IssuesEvent | 2020-08-20 14:02:39 | asetalias/Logistics-and-Event-Contributions | https://api.github.com/repos/asetalias/Logistics-and-Event-Contributions | opened | Add all previously used assets to the repo | High Priority Maintenance | All the previously used assets including all writeups, proposals, reports, emails, posters, etc. can be added to the repository to create a handbook which can be useful for folks that might need these resources in the future
**Please make sure the assets are redacted and no personal info is being added** | 1.0 | Add all previously used assets to the repo - All the previously used assets including all writeups, proposals, reports, emails, posters, etc. can be added to the repository to create a handbook which can be useful for folks that might need these resources in the future
**Please make sure the assets are redacted and no personal info is being added** | non_infrastructure | add all previously used assets to the repo all the previously used assets including all writeups proposals reports emails posters etc can be added to the repository to create a handbook which can be useful for folks that might need these resources in the future please make sure the assets are redacted and no personal info is being added | 0 |
6,456 | 2,847,735,352 | IssuesEvent | 2015-05-29 18:39:52 | luchanz/redhat-prode | https://api.github.com/repos/luchanz/redhat-prode | closed | Armar modelo de datos copa america | PRIORIDAD_BAJA test | ##### ARMAR JSON O DB DE FIXUTE DE COPA AMERICA)
Así como está el json de los partidos del mundial, hacer la conversión con los datos actuales de los grupos de la copa america. En el mejor de los casos, que esto sea dinámico, sinó estimarlo modificando el archivo en forma estática como está ahora. | 1.0 | Armar modelo de datos copa america - ##### ARMAR JSON O DB DE FIXUTE DE COPA AMERICA)
Así como está el json de los partidos del mundial, hacer la conversión con los datos actuales de los grupos de la copa america. En el mejor de los casos, que esto sea dinámico, sinó estimarlo modificando el archivo en forma estática como está ahora. | non_infrastructure | armar modelo de datos copa america armar json o db de fixute de copa america así como está el json de los partidos del mundial hacer la conversión con los datos actuales de los grupos de la copa america en el mejor de los casos que esto sea dinámico sinó estimarlo modificando el archivo en forma estática como está ahora | 0 |
28,542 | 23,325,095,614 | IssuesEvent | 2022-08-08 20:19:31 | microsoft/WindowsAppSDK | https://api.github.com/repos/microsoft/WindowsAppSDK | closed | Project Reunion runtime installer shows provisioning result of E_ACCESSDENIED during install | bug area-Infrastructure area-Installer v1.0 servicing-consider-1.0 | **Describe the bug**
When attempting to install the Project Reunion runtime, package deployment succeeds but several provisioning steps appear to fail with E_ACCESSDENIED. It's not clear if the runtime installer is required to be elevated.
**Steps to reproduce**
1. .\ProjectReunion-0.8Preview-Install-x64.exe
2. Observe results
```
Deploying package: Microsoft.ProjectReunion.0.8-preview_8000.146.628.0_x86__8wekyb3d8bbwe
Package deployment result : 0x0
Deploying package: Microsoft.ProjectReunion.0.8-preview_8000.146.628.0_x64__8wekyb3d8bbwe
Package deployment result : 0x0
Deploying package: Microsoft.ProjectReunion.Main.0.8-preview_8000.146.628.0_x64__8wekyb3d8bbwe
Package deployment result : 0x0
Provisioning result : 0x80070005
Deploying package: Microsoft.ProjectReunion.DDLM.8000.146.628.0-x8-p_8000.146.628.0_x86__8wekyb3d8bbwe
Package deployment result : 0x0
Provisioning result : 0x80070005
Deploying package: Microsoft.ProjectReunion.DDLM.8000.146.628.0-x6-p_8000.146.628.0_x64__8wekyb3d8bbwe
Package deployment result : 0x0
Provisioning result : 0x80070005
All packages were installed successfully.
```
ProjectReunion-0.8Preview-Install-x64.exe
Windows 10 21H1 19043.985 | 1.0 | Project Reunion runtime installer shows provisioning result of E_ACCESSDENIED during install - **Describe the bug**
When attempting to install the Project Reunion runtime, package deployment succeeds but several provisioning steps appear to fail with E_ACCESSDENIED. It's not clear if the runtime installer is required to be elevated.
**Steps to reproduce**
1. .\ProjectReunion-0.8Preview-Install-x64.exe
2. Observe results
```
Deploying package: Microsoft.ProjectReunion.0.8-preview_8000.146.628.0_x86__8wekyb3d8bbwe
Package deployment result : 0x0
Deploying package: Microsoft.ProjectReunion.0.8-preview_8000.146.628.0_x64__8wekyb3d8bbwe
Package deployment result : 0x0
Deploying package: Microsoft.ProjectReunion.Main.0.8-preview_8000.146.628.0_x64__8wekyb3d8bbwe
Package deployment result : 0x0
Provisioning result : 0x80070005
Deploying package: Microsoft.ProjectReunion.DDLM.8000.146.628.0-x8-p_8000.146.628.0_x86__8wekyb3d8bbwe
Package deployment result : 0x0
Provisioning result : 0x80070005
Deploying package: Microsoft.ProjectReunion.DDLM.8000.146.628.0-x6-p_8000.146.628.0_x64__8wekyb3d8bbwe
Package deployment result : 0x0
Provisioning result : 0x80070005
All packages were installed successfully.
```
ProjectReunion-0.8Preview-Install-x64.exe
Windows 10 21H1 19043.985 | infrastructure | project reunion runtime installer shows provisioning result of e accessdenied during install describe the bug when attempting to install the project reunion runtime package deployment succeeds but several provisioning steps appear to fail with e accessdenied it s not clear if the runtime installer is required to be elevated steps to reproduce projectreunion install exe observe results deploying package microsoft projectreunion preview package deployment result deploying package microsoft projectreunion preview package deployment result deploying package microsoft projectreunion main preview package deployment result provisioning result deploying package microsoft projectreunion ddlm p package deployment result provisioning result deploying package microsoft projectreunion ddlm p package deployment result provisioning result all packages were installed successfully projectreunion install exe windows | 1 |
1,492 | 3,250,812,307 | IssuesEvent | 2015-10-19 04:56:49 | asciidoctor/asciidoctor | https://api.github.com/repos/asciidoctor/asciidoctor | closed | Remove benchmark folder from gem | infrastructure | No need to package this folder in the gem as it's intended for testing and CI only. | 1.0 | Remove benchmark folder from gem - No need to package this folder in the gem as it's intended for testing and CI only. | infrastructure | remove benchmark folder from gem no need to package this folder in the gem as it s intended for testing and ci only | 1 |
292,546 | 21,996,175,232 | IssuesEvent | 2022-05-26 06:38:13 | JohanFalt/Foundry_WoD20 | https://api.github.com/repos/JohanFalt/Foundry_WoD20 | opened | Changes in items | documentation | "settings": {
"created": false,
"iscreated": false,
},
"weapon": {
"attack": {
"roll": true,
"isRollable": true
},
"damage": {
"roll": true,
"isRollable": true
},
"diff": 6,
"difficulty": 6,
"twohanded": false,
"istwohanded": false
}
"power": {
+"bonus"
"difficulty": "", -> "difficulty": 0,
"rollable": true,
"isRollable": true,
"active": false,
"isactive": false
}
"armor": {
-"active"
-"label": "wod.types.armor"
"forms": {
"homid": true,
"hashomid": true,
"glabro": false,
"hasglabro": false,
"crinos": false,
"hascrinos": false,
"hispo": false,
"hashispo": false,
"lupus": false,
"haslupus": false
}
}
"ranged weapon": {
"mode": {
"reload": false,
"burst": false,
"fullauto": false,
"spray": false,
"hasreload": false,
"hasburst": false,
"hasfullauto": false,
"hasspray": false
},
} | 1.0 | Changes in items - "settings": {
"created": false,
"iscreated": false,
},
"weapon": {
"attack": {
"roll": true,
"isRollable": true
},
"damage": {
"roll": true,
"isRollable": true
},
"diff": 6,
"difficulty": 6,
"twohanded": false,
"istwohanded": false
}
"power": {
+"bonus"
"difficulty": "", -> "difficulty": 0,
"rollable": true,
"isRollable": true,
"active": false,
"isactive": false
}
"armor": {
-"active"
-"label": "wod.types.armor"
"forms": {
"homid": true,
"hashomid": true,
"glabro": false,
"hasglabro": false,
"crinos": false,
"hascrinos": false,
"hispo": false,
"hashispo": false,
"lupus": false,
"haslupus": false
}
}
"ranged weapon": {
"mode": {
"reload": false,
"burst": false,
"fullauto": false,
"spray": false,
"hasreload": false,
"hasburst": false,
"hasfullauto": false,
"hasspray": false
},
} | non_infrastructure | changes in items settings created false iscreated false weapon attack roll true isrollable true damage roll true isrollable true diff difficulty twohanded false istwohanded false power bonus difficulty difficulty rollable true isrollable true active false isactive false armor active label wod types armor forms homid true hashomid true glabro false hasglabro false crinos false hascrinos false hispo false hashispo false lupus false haslupus false ranged weapon mode reload false burst false fullauto false spray false hasreload false hasburst false hasfullauto false hasspray false | 0 |
741,749 | 25,816,312,464 | IssuesEvent | 2022-12-12 05:21:52 | milvus-io/milvus | https://api.github.com/repos/milvus-io/milvus | closed | [Bug]: Search failed with error `fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica` without any chaos | kind/bug priority/critical-urgent severity/critical triage/accepted | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Environment
```markdown
- Milvus version: master-20221206-f8cff798
- Deployment mode(standalone or cluster): cluster
- SDK version(e.g. pymilvus v2.0.0rc2): pymilvus==2.3.0.dev15
- OS(Ubuntu or CentOS):
- CPU/Memory:
- GPU:
- Others:
```
### Current Behavior
```
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:13 - INFO - ci_test]: [test][2022-12-07T02:23:09Z] [4.03896960s] Hello_Milvus load -> None (wrapper.py:30)
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:13 - INFO - ci_test]: assert load: 4.039217472076416 (test_data_persistence.py:89)
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:13 - DEBUG - ci_test]: (api_request) : [Collection.search] args: [[[0.1086702406615037, 0.03836368826610158, 0.08882218314766184, 0.10044394949779682, 0.13572320745375752, 0.07011054020093992, 0.09889132419668602, 0.10155191962304858, 0.11281254282514301, 0.009236124677987909, 0.15303027050056867, 0.05284775407586183, 0.1029843927374862, 0.026665505247487422, 0.0......, kwargs: {} (api_request.py:56)
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:23 - ERROR - pymilvus.decorators]: RPC error: [search], <MilvusException: (code=1, message=fail to search on all shard leaders, err=All attempts results:
[2022-12-07T02:23:24.289Z] attempt #1:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #2:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #3:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_3_437880003361443220v1 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #4:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #5:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #6:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #7:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #8:context deadline exceeded
[2022-12-07T02:23:24.290Z] )>, <Time:{'RPC start': '2022-12-07 02:23:13.990435', 'RPC error': '2022-12-07 02:23:23.993269'}> (decorators.py:108)
[2022-12-07T02:23:24.290Z] [2022-12-07 02:23:23 - ERROR - ci_test]: Traceback (most recent call last):
[2022-12-07T02:23:24.290Z] File "/home/jenkins/agent/workspace/tests/python_client/utils/api_request.py", line 26, in inner_wrapper
[2022-12-07T02:23:24.290Z] res = func(*args, **_kwargs)
[2022-12-07T02:23:24.290Z] File "/home/jenkins/agent/workspace/tests/python_client/utils/api_request.py", line 57, in api_request
[2022-12-07T02:23:24.290Z] return func(*arg, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/orm/collection.py", line 610, in search
[2022-12-07T02:23:24.290Z] res = conn.search(self._name, data, anns_field, param, limit, expr,
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 109, in handler
[2022-12-07T02:23:24.290Z] raise e
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 105, in handler
[2022-12-07T02:23:24.290Z] return func(*args, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 136, in handler
[2022-12-07T02:23:24.290Z] ret = func(self, *args, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 85, in handler
[2022-12-07T02:23:24.290Z] raise e
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 50, in handler
[2022-12-07T02:23:24.290Z] return func(self, *args, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 469, in search
[2022-12-07T02:23:24.290Z] return self._execute_search_requests(requests, timeout, round_decimal=round_decimal, auto_id=auto_id, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 438, in _execute_search_requests
[2022-12-07T02:23:24.290Z] raise pre_err
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 429, in _execute_search_requests
[2022-12-07T02:23:24.290Z] raise MilvusException(response.status.error_code, response.status.reason)
[2022-12-07T02:23:24.290Z] pymilvus.exceptions.MilvusException: <MilvusException: (code=1, message=fail to search on all shard leaders, err=All attempts results:
[2022-12-07T02:23:24.290Z] attempt #1:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #2:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #3:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_3_437880003361443220v1 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #4:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #5:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #6:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #7:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #8:context deadline exceeded
[2022-12-07T02:23:24.290Z] )>
[2022-12-07T02:23:24.290Z] (api_request.py:39)
[2022-12-07T02:23:24.290Z] [2022-12-07 02:23:23 - ERROR - ci_test]: (api_response) : <MilvusException: (code=1, message=fail to search on all shard leaders, err=All attempts results:
[2022-12-07T02:23:24.290Z] attempt #1:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #2:fail to get shard leaders from QueryCoord: channel by...... (api_request.py:40)
```
### Expected Behavior
all test cases passed
### Steps To Reproduce
_No response_
### Milvus Log
failed job: https://qa-jenkins.milvus.io/blue/organizations/jenkins/chaos-test-cron/detail/chaos-test-cron/378/pipeline
log:
[artifacts-indexcoord-pod-failure-378-server-logs.tar.gz](https://github.com/milvus-io/milvus/files/10172365/artifacts-indexcoord-pod-failure-378-server-logs.tar.gz)
[artifacts-indexcoord-pod-failure-378-pytest-logs.tar.gz](https://github.com/milvus-io/milvus/files/10172366/artifacts-indexcoord-pod-failure-378-pytest-logs.tar.gz)
### Anything else?
_No response_ | 1.0 | [Bug]: Search failed with error `fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica` without any chaos - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Environment
```markdown
- Milvus version: master-20221206-f8cff798
- Deployment mode(standalone or cluster): cluster
- SDK version(e.g. pymilvus v2.0.0rc2): pymilvus==2.3.0.dev15
- OS(Ubuntu or CentOS):
- CPU/Memory:
- GPU:
- Others:
```
### Current Behavior
```
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:13 - INFO - ci_test]: [test][2022-12-07T02:23:09Z] [4.03896960s] Hello_Milvus load -> None (wrapper.py:30)
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:13 - INFO - ci_test]: assert load: 4.039217472076416 (test_data_persistence.py:89)
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:13 - DEBUG - ci_test]: (api_request) : [Collection.search] args: [[[0.1086702406615037, 0.03836368826610158, 0.08882218314766184, 0.10044394949779682, 0.13572320745375752, 0.07011054020093992, 0.09889132419668602, 0.10155191962304858, 0.11281254282514301, 0.009236124677987909, 0.15303027050056867, 0.05284775407586183, 0.1029843927374862, 0.026665505247487422, 0.0......, kwargs: {} (api_request.py:56)
[2022-12-07T02:23:24.289Z] [2022-12-07 02:23:23 - ERROR - pymilvus.decorators]: RPC error: [search], <MilvusException: (code=1, message=fail to search on all shard leaders, err=All attempts results:
[2022-12-07T02:23:24.289Z] attempt #1:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #2:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #3:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_3_437880003361443220v1 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #4:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #5:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #6:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #7:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #8:context deadline exceeded
[2022-12-07T02:23:24.290Z] )>, <Time:{'RPC start': '2022-12-07 02:23:13.990435', 'RPC error': '2022-12-07 02:23:23.993269'}> (decorators.py:108)
[2022-12-07T02:23:24.290Z] [2022-12-07 02:23:23 - ERROR - ci_test]: Traceback (most recent call last):
[2022-12-07T02:23:24.290Z] File "/home/jenkins/agent/workspace/tests/python_client/utils/api_request.py", line 26, in inner_wrapper
[2022-12-07T02:23:24.290Z] res = func(*args, **_kwargs)
[2022-12-07T02:23:24.290Z] File "/home/jenkins/agent/workspace/tests/python_client/utils/api_request.py", line 57, in api_request
[2022-12-07T02:23:24.290Z] return func(*arg, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/orm/collection.py", line 610, in search
[2022-12-07T02:23:24.290Z] res = conn.search(self._name, data, anns_field, param, limit, expr,
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 109, in handler
[2022-12-07T02:23:24.290Z] raise e
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 105, in handler
[2022-12-07T02:23:24.290Z] return func(*args, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 136, in handler
[2022-12-07T02:23:24.290Z] ret = func(self, *args, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 85, in handler
[2022-12-07T02:23:24.290Z] raise e
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 50, in handler
[2022-12-07T02:23:24.290Z] return func(self, *args, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 469, in search
[2022-12-07T02:23:24.290Z] return self._execute_search_requests(requests, timeout, round_decimal=round_decimal, auto_id=auto_id, **kwargs)
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 438, in _execute_search_requests
[2022-12-07T02:23:24.290Z] raise pre_err
[2022-12-07T02:23:24.290Z] File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 429, in _execute_search_requests
[2022-12-07T02:23:24.290Z] raise MilvusException(response.status.error_code, response.status.reason)
[2022-12-07T02:23:24.290Z] pymilvus.exceptions.MilvusException: <MilvusException: (code=1, message=fail to search on all shard leaders, err=All attempts results:
[2022-12-07T02:23:24.290Z] attempt #1:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #2:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #3:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_3_437880003361443220v1 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #4:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #5:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #6:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #7:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #8:context deadline exceeded
[2022-12-07T02:23:24.290Z] )>
[2022-12-07T02:23:24.290Z] (api_request.py:39)
[2022-12-07T02:23:24.290Z] [2022-12-07 02:23:23 - ERROR - ci_test]: (api_response) : <MilvusException: (code=1, message=fail to search on all shard leaders, err=All attempts results:
[2022-12-07T02:23:24.290Z] attempt #1:fail to get shard leaders from QueryCoord: channel by-dev-rootcoord-dml_2_437880003361443220v0 is not available in any replica
[2022-12-07T02:23:24.290Z] attempt #2:fail to get shard leaders from QueryCoord: channel by...... (api_request.py:40)
```
### Expected Behavior
all test cases passed
### Steps To Reproduce
_No response_
### Milvus Log
failed job: https://qa-jenkins.milvus.io/blue/organizations/jenkins/chaos-test-cron/detail/chaos-test-cron/378/pipeline
log:
[artifacts-indexcoord-pod-failure-378-server-logs.tar.gz](https://github.com/milvus-io/milvus/files/10172365/artifacts-indexcoord-pod-failure-378-server-logs.tar.gz)
[artifacts-indexcoord-pod-failure-378-pytest-logs.tar.gz](https://github.com/milvus-io/milvus/files/10172366/artifacts-indexcoord-pod-failure-378-pytest-logs.tar.gz)
### Anything else?
_No response_ | non_infrastructure | search failed with error fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica without any chaos is there an existing issue for this i have searched the existing issues environment markdown milvus version master deployment mode standalone or cluster cluster sdk version e g pymilvus pymilvus os ubuntu or centos cpu memory gpu others current behavior hello milvus load none wrapper py assert load test data persistence py api request args kwargs api request py rpc error milvusexception code message fail to search on all shard leaders err all attempts results attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt context deadline exceeded decorators py traceback most recent call last file home jenkins agent workspace tests python client utils api request py line in inner wrapper res func args kwargs file home jenkins agent workspace tests python client utils api request py line in api request return func arg kwargs file usr local lib dist packages pymilvus orm collection py line in search res conn search self name data anns field param limit expr file usr local lib dist packages pymilvus decorators py line in handler raise e file usr local lib dist packages pymilvus decorators py line in handler return func args kwargs file usr local lib dist packages pymilvus decorators py line in handler ret func self args kwargs file usr local lib dist packages pymilvus decorators py line in handler raise e file usr local lib dist packages pymilvus decorators py line in handler return func self args kwargs file usr local lib dist packages pymilvus client grpc handler py line in search return self execute search requests requests timeout round decimal round decimal auto id auto id kwargs file usr local lib dist packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib dist packages pymilvus client grpc handler py line in execute search requests raise milvusexception response status error code response status reason pymilvus exceptions milvusexception milvusexception code message fail to search on all shard leaders err all attempts results attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt context deadline exceeded api request py api response milvusexception code message fail to search on all shard leaders err all attempts results attempt fail to get shard leaders from querycoord channel by dev rootcoord dml is not available in any replica attempt fail to get shard leaders from querycoord channel by api request py expected behavior all test cases passed steps to reproduce no response milvus log failed job log anything else no response | 0 |
123,013 | 12,180,933,025 | IssuesEvent | 2020-04-28 13:16:22 | epi052/recon-pipeline | https://api.github.com/repos/epi052/recon-pipeline | closed | Update docs/readme to show install tested on 18.04/20.04/kali | documentation nitpickery | While checking the newest PR from @GreaterGoodest, I tested installation on a fresh ubuntu 20.04 install. No issues found. | 1.0 | Update docs/readme to show install tested on 18.04/20.04/kali - While checking the newest PR from @GreaterGoodest, I tested installation on a fresh ubuntu 20.04 install. No issues found. | non_infrastructure | update docs readme to show install tested on kali while checking the newest pr from greatergoodest i tested installation on a fresh ubuntu install no issues found | 0 |
169,124 | 26,748,865,204 | IssuesEvent | 2023-01-30 17:56:22 | CDCgov/prime-simplereport | https://api.github.com/repos/CDCgov/prime-simplereport | closed | LAC onsite research deck | Design Research | Synthesize learnings from onsite into a research deck to be shared with internal teams and partners
Thurs 1/26: presentation
Friday 1/27: check-in with LAC | 1.0 | LAC onsite research deck - Synthesize learnings from onsite into a research deck to be shared with internal teams and partners
Thurs 1/26: presentation
Friday 1/27: check-in with LAC | non_infrastructure | lac onsite research deck synthesize learnings from onsite into a research deck to be shared with internal teams and partners thurs presentation friday check in with lac | 0 |
602,024 | 18,445,800,753 | IssuesEvent | 2021-10-15 01:39:57 | AY2122S1-CS2103-T16-4/tp | https://api.github.com/repos/AY2122S1-CS2103-T16-4/tp | closed | Wrap up v1.2 | type.Task priority.Low | Wrap up the milestone using a git tag v1.2. When the milestone deadline is near (e.g., 0.5 days before the deadline), if you think some of the ongoing work intended for the current milestone may not finish in time, reassign them to a future milestone.
Do a release on GitHub. Uploading a JAR file to GitHub is optional.
- [x] Create Tag v1.2
- [x] Release JAR file
- [x] Reassign undone issues
| 1.0 | Wrap up v1.2 - Wrap up the milestone using a git tag v1.2. When the milestone deadline is near (e.g., 0.5 days before the deadline), if you think some of the ongoing work intended for the current milestone may not finish in time, reassign them to a future milestone.
Do a release on GitHub. Uploading a JAR file to GitHub is optional.
- [x] Create Tag v1.2
- [x] Release JAR file
- [x] Reassign undone issues
| non_infrastructure | wrap up wrap up the milestone using a git tag when the milestone deadline is near e g days before the deadline if you think some of the ongoing work intended for the current milestone may not finish in time reassign them to a future milestone do a release on github uploading a jar file to github is optional create tag release jar file reassign undone issues | 0 |
95,528 | 27,533,620,434 | IssuesEvent | 2023-03-07 00:46:37 | CasparCG/server | https://api.github.com/repos/CasparCG/server | closed | Make build-scripts checkout themselves when a build is triggered | type/enhancement build | ### Expected Behaviour
When a build is triggered automatically, the build-server should checkout the build-scripts itself from a repository. By going this approach, it would be enough with a pull request against the build-scripts repository to update the scripts when needed.
### Current Behaviour
When changes are made to CasparCG/server that affect files, folders, and versions of media-scanner that needs to be packaged to the final build (for example), it requires changes to the build-scripts located on the build-server. Those changes are dependent ob those who have access to the actual building server, which is bad. | 1.0 | Make build-scripts checkout themselves when a build is triggered - ### Expected Behaviour
When a build is triggered automatically, the build-server should checkout the build-scripts itself from a repository. By going this approach, it would be enough with a pull request against the build-scripts repository to update the scripts when needed.
### Current Behaviour
When changes are made to CasparCG/server that affect files, folders, and versions of media-scanner that needs to be packaged to the final build (for example), it requires changes to the build-scripts located on the build-server. Those changes are dependent ob those who have access to the actual building server, which is bad. | non_infrastructure | make build scripts checkout themselves when a build is triggered expected behaviour when a build is triggered automatically the build server should checkout the build scripts itself from a repository by going this approach it would be enough with a pull request against the build scripts repository to update the scripts when needed current behaviour when changes are made to casparcg server that affect files folders and versions of media scanner that needs to be packaged to the final build for example it requires changes to the build scripts located on the build server those changes are dependent ob those who have access to the actual building server which is bad | 0 |
18,380 | 12,962,102,474 | IssuesEvent | 2020-07-20 16:37:37 | enarx/enarx | https://api.github.com/repos/enarx/enarx | closed | Allow integration tests to run for SEV | amd sev infrastructure | The `has_sev` compiler attribute must be enabled in [integration-tests/build.rs](https://github.com/enarx/enarx/blob/master/integration-tests/build.rs).
Depends on:
- [ ] Shim support for running payload (currently included in PR #684)
This change is motivated by #736. | 1.0 | Allow integration tests to run for SEV - The `has_sev` compiler attribute must be enabled in [integration-tests/build.rs](https://github.com/enarx/enarx/blob/master/integration-tests/build.rs).
Depends on:
- [ ] Shim support for running payload (currently included in PR #684)
This change is motivated by #736. | infrastructure | allow integration tests to run for sev the has sev compiler attribute must be enabled in depends on shim support for running payload currently included in pr this change is motivated by | 1 |
192,960 | 6,877,593,713 | IssuesEvent | 2017-11-20 08:43:07 | OpenNebula/one | https://api.github.com/repos/OpenNebula/one | opened | Restrict RAW content | Priority: Normal Status: Pending Tracker: Request | ---
Author Name: **EOLE Team** (EOLE Team)
Original Redmine Issue: 3498, https://dev.opennebula.org/issues/3498
Original Date: 2015-01-14
---
Hello,
We use a virtfs for our test beds for communications between a jenkins and VMs.
I test with the following RAW:
```
RAW=[TYPE="kvm",DATA="
<devices>
<filesystem type='mount' accessmode='squash'>
<source dir='/' />
<target dir='root' />
</filesystem>
</devices>
"]
```
Then I can mount this virtfs in my VM and access the root of my hypervisor as user *@oneadmin@*:
```
root@ubuntu:~# mount -t 9p -o trans=virtio root /mnt/ -oversion=9p2000.L
root@ubuntu:~# cat /mnt/etc/hostname
nebula1
root@ubuntu:~# cat: /mnt/etc/shadow: Permission denied
root@ubuntu:~# touch /mnt/var/lib/one/datastores/foo
root@ubuntu:~# rm /mnt/var/lib/one/datastores/foo
```
So, I can run *@rm -rf /mnt/var/lib/one/datastores/@* and destroy my infrastructure.
Is there a way to restrict the content of RAW?
| 1.0 | Restrict RAW content - ---
Author Name: **EOLE Team** (EOLE Team)
Original Redmine Issue: 3498, https://dev.opennebula.org/issues/3498
Original Date: 2015-01-14
---
Hello,
We use a virtfs for our test beds for communications between a jenkins and VMs.
I test with the following RAW:
```
RAW=[TYPE="kvm",DATA="
<devices>
<filesystem type='mount' accessmode='squash'>
<source dir='/' />
<target dir='root' />
</filesystem>
</devices>
"]
```
Then I can mount this virtfs in my VM and access the root of my hypervisor as user *@oneadmin@*:
```
root@ubuntu:~# mount -t 9p -o trans=virtio root /mnt/ -oversion=9p2000.L
root@ubuntu:~# cat /mnt/etc/hostname
nebula1
root@ubuntu:~# cat: /mnt/etc/shadow: Permission denied
root@ubuntu:~# touch /mnt/var/lib/one/datastores/foo
root@ubuntu:~# rm /mnt/var/lib/one/datastores/foo
```
So, I can run *@rm -rf /mnt/var/lib/one/datastores/@* and destroy my infrastructure.
Is there a way to restrict the content of RAW?
| non_infrastructure | restrict raw content author name eole team eole team original redmine issue original date hello we use a virtfs for our test beds for communications between a jenkins and vms i test with the following raw raw type kvm data then i can mount this virtfs in my vm and access the root of my hypervisor as user oneadmin root ubuntu mount t o trans virtio root mnt oversion l root ubuntu cat mnt etc hostname root ubuntu cat mnt etc shadow permission denied root ubuntu touch mnt var lib one datastores foo root ubuntu rm mnt var lib one datastores foo so i can run rm rf mnt var lib one datastores and destroy my infrastructure is there a way to restrict the content of raw | 0 |
29,490 | 24,043,159,109 | IssuesEvent | 2022-09-16 05:18:26 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | reopened | StateFragment(Local)Test sometimes fail with SIGSEGV | issue_type_infrastructure issue_temp_ben_triaged | Both StateFragmentTest and StateFragmentLocalTest occasionally crash the JVM with a SIGSEGV (see #2844 for context). This is specifically worse when running the tests on memory-constrained systems, or simultaneously to many other tests. It's not clear where this crash is coming from, but #3969 introduced a mitigation to work around it such that CI runs don't get hung up on it.
This issue is tracking investigating & fixing the tests long-term such that this failure case completely goes away. | 1.0 | StateFragment(Local)Test sometimes fail with SIGSEGV - Both StateFragmentTest and StateFragmentLocalTest occasionally crash the JVM with a SIGSEGV (see #2844 for context). This is specifically worse when running the tests on memory-constrained systems, or simultaneously to many other tests. It's not clear where this crash is coming from, but #3969 introduced a mitigation to work around it such that CI runs don't get hung up on it.
This issue is tracking investigating & fixing the tests long-term such that this failure case completely goes away. | infrastructure | statefragment local test sometimes fail with sigsegv both statefragmenttest and statefragmentlocaltest occasionally crash the jvm with a sigsegv see for context this is specifically worse when running the tests on memory constrained systems or simultaneously to many other tests it s not clear where this crash is coming from but introduced a mitigation to work around it such that ci runs don t get hung up on it this issue is tracking investigating fixing the tests long term such that this failure case completely goes away | 1 |
104,977 | 9,013,549,151 | IssuesEvent | 2019-02-05 19:48:08 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: cdc/sink-chaos failed | C-test-failure O-roachtest O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/153fcf347cd6a53517f607ec1e6c7c14193640dd
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=cdc/sink-chaos PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1052961&tab=buildLog
```
The test failed on release-2.1:
test.go:630,cluster.go:1139,cdc.go:555,cdc.go:109,cluster.go:1465,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1052961-cdc-sink-chaos:4 -- ./workload run tpcc --warehouses=100 --duration=30m --tolerate-errors {pgurl:1-3} returned:
stderr:
stdout:
l
2m20s 0 2.0 2.2 41.9 48.2 48.2 48.2 delivery
2m20s 0 24.0 21.2 28.3 35.7 37.7 37.7 newOrder
2m20s 0 1.0 2.4 7.1 7.1 7.1 7.1 orderStatus
2m20s 0 14.0 22.7 14.7 16.8 18.9 18.9 payment
2m20s 0 3.0 2.4 18.9 19.9 19.9 19.9 stockLevel
_elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
2m21s 0 6.0 2.2 46.1 50.3 50.3 50.3 delivery
2m21s 0 19.0 21.2 29.4 35.7 35.7 35.7 newOrder
2m21s 0 1.0 2.4 6.6 6.6 6.6 6.6 orderStatus
2m21s 0 22.0 22.7 14.2 17.8 28.3 28.3 payment
2m21s 0 0.0 2.4 0.0 0.0 0.0 0.0 stockLevel
: signal: killed
test.go:630,cluster.go:1486,cdc.go:185,cdc.go:375: unexpected status: failed
``` | 2.0 | roachtest: cdc/sink-chaos failed - SHA: https://github.com/cockroachdb/cockroach/commits/153fcf347cd6a53517f607ec1e6c7c14193640dd
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=cdc/sink-chaos PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1052961&tab=buildLog
```
The test failed on release-2.1:
test.go:630,cluster.go:1139,cdc.go:555,cdc.go:109,cluster.go:1465,errgroup.go:57: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1052961-cdc-sink-chaos:4 -- ./workload run tpcc --warehouses=100 --duration=30m --tolerate-errors {pgurl:1-3} returned:
stderr:
stdout:
l
2m20s 0 2.0 2.2 41.9 48.2 48.2 48.2 delivery
2m20s 0 24.0 21.2 28.3 35.7 37.7 37.7 newOrder
2m20s 0 1.0 2.4 7.1 7.1 7.1 7.1 orderStatus
2m20s 0 14.0 22.7 14.7 16.8 18.9 18.9 payment
2m20s 0 3.0 2.4 18.9 19.9 19.9 19.9 stockLevel
_elapsed___errors__ops/sec(inst)___ops/sec(cum)__p50(ms)__p95(ms)__p99(ms)_pMax(ms)
2m21s 0 6.0 2.2 46.1 50.3 50.3 50.3 delivery
2m21s 0 19.0 21.2 29.4 35.7 35.7 35.7 newOrder
2m21s 0 1.0 2.4 6.6 6.6 6.6 6.6 orderStatus
2m21s 0 22.0 22.7 14.2 17.8 28.3 28.3 payment
2m21s 0 0.0 2.4 0.0 0.0 0.0 0.0 stockLevel
: signal: killed
test.go:630,cluster.go:1486,cdc.go:185,cdc.go:375: unexpected status: failed
``` | non_infrastructure | roachtest cdc sink chaos failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests cdc sink chaos pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on release test go cluster go cdc go cdc go cluster go errgroup go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity cdc sink chaos workload run tpcc warehouses duration tolerate errors pgurl returned stderr stdout l delivery neworder orderstatus payment stocklevel elapsed errors ops sec inst ops sec cum ms ms ms pmax ms delivery neworder orderstatus payment stocklevel signal killed test go cluster go cdc go cdc go unexpected status failed | 0 |
18,819 | 13,128,585,059 | IssuesEvent | 2020-08-06 12:33:38 | reapit/foundations | https://api.github.com/repos/reapit/foundations | closed | Investigate how the data in our warehouse should be structured / architected | blocked infrastructure investigate platform-team | We need to ensure that we can provision access to our customers in a secure, best practice way for their data.
- Need to investigate what the best practice is for managing multi tenancy for any data warehouse storage tech that we adopt. This decision may be lead by https://github.com/reapit/foundations/issues/851
- Need to design the schema of entities that we store (almost certainly replications of our API schema)
- Need to understand how we can effectively index our entities for high performance querying by customers and third party products
| 1.0 | Investigate how the data in our warehouse should be structured / architected - We need to ensure that we can provision access to our customers in a secure, best practice way for their data.
- Need to investigate what the best practice is for managing multi tenancy for any data warehouse storage tech that we adopt. This decision may be lead by https://github.com/reapit/foundations/issues/851
- Need to design the schema of entities that we store (almost certainly replications of our API schema)
- Need to understand how we can effectively index our entities for high performance querying by customers and third party products
| infrastructure | investigate how the data in our warehouse should be structured architected we need to ensure that we can provision access to our customers in a secure best practice way for their data need to investigate what the best practice is for managing multi tenancy for any data warehouse storage tech that we adopt this decision may be lead by need to design the schema of entities that we store almost certainly replications of our api schema need to understand how we can effectively index our entities for high performance querying by customers and third party products | 1 |
29,207 | 11,727,035,251 | IssuesEvent | 2020-03-10 15:22:02 | mixcore/mix.core | https://api.github.com/repos/mixcore/mix.core | opened | CVE-2018-14041 (Medium) detected in bootstrap-4.0.0.min.js | security vulnerability | ## CVE-2018-14041 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-4.0.0.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/mix.core/src/Mix.Cms.Web/content/templates/hai-yen/assets/dist/post.html</p>
<p>Path to vulnerable library: /mix.core/src/Mix.Cms.Web/content/templates/hai-yen/assets/dist/post.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-4.0.0.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mixcore/mix.core/commit/ea0ad82049c541137fe26de30d6831204092f506">ea0ad82049c541137fe26de30d6831204092f506</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-target property of scrollspy.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14041>CVE-2018-14041</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-14041 (Medium) detected in bootstrap-4.0.0.min.js - ## CVE-2018-14041 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-4.0.0.min.js</b></p></summary>
<p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/4.0.0/js/bootstrap.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/mix.core/src/Mix.Cms.Web/content/templates/hai-yen/assets/dist/post.html</p>
<p>Path to vulnerable library: /mix.core/src/Mix.Cms.Web/content/templates/hai-yen/assets/dist/post.html</p>
<p>
Dependency Hierarchy:
- :x: **bootstrap-4.0.0.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mixcore/mix.core/commit/ea0ad82049c541137fe26de30d6831204092f506">ea0ad82049c541137fe26de30d6831204092f506</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Bootstrap before 4.1.2, XSS is possible in the data-target property of scrollspy.
<p>Publish Date: 2018-07-13
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14041>CVE-2018-14041</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/twbs/bootstrap/pull/26630">https://github.com/twbs/bootstrap/pull/26630</a></p>
<p>Release Date: 2018-07-13</p>
<p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:4.1.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in bootstrap min js cve medium severity vulnerability vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file tmp ws scm mix core src mix cms web content templates hai yen assets dist post html path to vulnerable library mix core src mix cms web content templates hai yen assets dist post html dependency hierarchy x bootstrap min js vulnerable library found in head commit a href vulnerability details in bootstrap before xss is possible in the data target property of scrollspy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org webjars npm bootstrap org webjars bootstrap step up your open source security game with whitesource | 0 |
20,019 | 13,631,766,062 | IssuesEvent | 2020-09-24 18:32:19 | E3SM-Project/scream | https://api.github.com/repos/E3SM-Project/scream | opened | Build scorpio in SCREAM standalone mode. | Atmosphere Driver Development Task cmake infrastructure | It will allow to test scorpio stuff in all our PR testings and nightlies, rather than require manual cime-builds runs. | 1.0 | Build scorpio in SCREAM standalone mode. - It will allow to test scorpio stuff in all our PR testings and nightlies, rather than require manual cime-builds runs. | infrastructure | build scorpio in scream standalone mode it will allow to test scorpio stuff in all our pr testings and nightlies rather than require manual cime builds runs | 1 |
13,727 | 3,184,883,268 | IssuesEvent | 2015-09-27 19:42:31 | Glucosio/android | https://api.github.com/repos/Glucosio/android | reopened | Terms of Use. | blocker bug copy design | In Hello we should remove the period behind Terms of Use and maybe consider a Glucosio pink line underneath it.
Lets have @elioqoshi give some design feedback one last time and look at the app for any slight styling advice he might have. | 1.0 | Terms of Use. - In Hello we should remove the period behind Terms of Use and maybe consider a Glucosio pink line underneath it.
Lets have @elioqoshi give some design feedback one last time and look at the app for any slight styling advice he might have. | non_infrastructure | terms of use in hello we should remove the period behind terms of use and maybe consider a glucosio pink line underneath it lets have elioqoshi give some design feedback one last time and look at the app for any slight styling advice he might have | 0 |
225,885 | 7,496,030,808 | IssuesEvent | 2018-04-08 04:29:13 | rancher/rancher | https://api.github.com/repos/rancher/rancher | opened | Deleting projects results in project and namespaces stuck in "removing" state forever. | priority/0 version/2.0 | **Rancher versions: Build from master
Steps to reproduce the problem:
Create a cluster with 1 node control and etcd and 2 worked nodes.
Create a project and namespace.
Deploy Daemonset workload.
Delete project.
Deleting projects results in project and namespaces stuck in "removing" state forever.
Namespace is stuck in "Terminating" state in kube cluster.
```
hellonew1 Terminating 7m```
```
Following errors seen in logs:
```
E0408 04:21:46.151559 1 generic_controller.go:204] ProjectController cluster-87bvv/project-rxflf [mgmt-project-rbac-remove] failed with : failed to remove finalizer on cluster-87bvv:project-rxflf
E0408 04:21:51.424194 1 generic_controller.go:204] ProjectRoleTemplateBindingController project-rxflf/creator [cluster-prtb-sync] failed with : failed to remove finalizer on project-rxflf:creator
``` | 1.0 | Deleting projects results in project and namespaces stuck in "removing" state forever. - **Rancher versions: Build from master
Steps to reproduce the problem:
Create a cluster with 1 node control and etcd and 2 worked nodes.
Create a project and namespace.
Deploy Daemonset workload.
Delete project.
Deleting projects results in project and namespaces stuck in "removing" state forever.
Namespace is stuck in "Terminating" state in kube cluster.
```
hellonew1 Terminating 7m```
```
Following errors seen in logs:
```
E0408 04:21:46.151559 1 generic_controller.go:204] ProjectController cluster-87bvv/project-rxflf [mgmt-project-rbac-remove] failed with : failed to remove finalizer on cluster-87bvv:project-rxflf
E0408 04:21:51.424194 1 generic_controller.go:204] ProjectRoleTemplateBindingController project-rxflf/creator [cluster-prtb-sync] failed with : failed to remove finalizer on project-rxflf:creator
``` | non_infrastructure | deleting projects results in project and namespaces stuck in removing state forever rancher versions build from master steps to reproduce the problem create a cluster with node control and etcd and worked nodes create a project and namespace deploy daemonset workload delete project deleting projects results in project and namespaces stuck in removing state forever namespace is stuck in terminating state in kube cluster terminating following errors seen in logs generic controller go projectcontroller cluster project rxflf failed with failed to remove finalizer on cluster project rxflf generic controller go projectroletemplatebindingcontroller project rxflf creator failed with failed to remove finalizer on project rxflf creator | 0 |
235,839 | 25,962,070,110 | IssuesEvent | 2022-12-19 01:03:13 | michaeldotson/contacts-app | https://api.github.com/repos/michaeldotson/contacts-app | opened | CVE-2022-23518 (Medium) detected in rails-html-sanitizer-1.0.4.gem | security vulnerability | ## CVE-2022-23518 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rails-html-sanitizer-1.0.4.gem</b></p></summary>
<p>HTML sanitization for Rails applications</p>
<p>Library home page: <a href="https://rubygems.org/gems/rails-html-sanitizer-1.0.4.gem">https://rubygems.org/gems/rails-html-sanitizer-1.0.4.gem</a></p>
<p>Path to dependency file: /contacts-app/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/rails-html-sanitizer-1.0.4.gem</p>
<p>
Dependency Hierarchy:
- web-console-3.7.0.gem (Root Library)
- actionview-5.2.2.gem
- :x: **rails-html-sanitizer-1.0.4.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
rails-html-sanitizer is responsible for sanitizing HTML fragments in Rails applications. Versions >= 1.0.3, < 1.4.4 are vulnerable to cross-site scripting via data URIs when used in combination with Loofah >= 2.1.0. This issue is patched in version 1.4.4.
<p>Publish Date: 2022-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23518>CVE-2022-23518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/rails/rails-html-sanitizer/security/advisories/GHSA-mcvf-2q2m-x72m">https://github.com/rails/rails-html-sanitizer/security/advisories/GHSA-mcvf-2q2m-x72m</a></p>
<p>Release Date: 2022-12-14</p>
<p>Fix Resolution: rails-html-sanitizer - 1.4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-23518 (Medium) detected in rails-html-sanitizer-1.0.4.gem - ## CVE-2022-23518 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>rails-html-sanitizer-1.0.4.gem</b></p></summary>
<p>HTML sanitization for Rails applications</p>
<p>Library home page: <a href="https://rubygems.org/gems/rails-html-sanitizer-1.0.4.gem">https://rubygems.org/gems/rails-html-sanitizer-1.0.4.gem</a></p>
<p>Path to dependency file: /contacts-app/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/rails-html-sanitizer-1.0.4.gem</p>
<p>
Dependency Hierarchy:
- web-console-3.7.0.gem (Root Library)
- actionview-5.2.2.gem
- :x: **rails-html-sanitizer-1.0.4.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
rails-html-sanitizer is responsible for sanitizing HTML fragments in Rails applications. Versions >= 1.0.3, < 1.4.4 are vulnerable to cross-site scripting via data URIs when used in combination with Loofah >= 2.1.0. This issue is patched in version 1.4.4.
<p>Publish Date: 2022-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-23518>CVE-2022-23518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/rails/rails-html-sanitizer/security/advisories/GHSA-mcvf-2q2m-x72m">https://github.com/rails/rails-html-sanitizer/security/advisories/GHSA-mcvf-2q2m-x72m</a></p>
<p>Release Date: 2022-12-14</p>
<p>Fix Resolution: rails-html-sanitizer - 1.4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in rails html sanitizer gem cve medium severity vulnerability vulnerable library rails html sanitizer gem html sanitization for rails applications library home page a href path to dependency file contacts app gemfile lock path to vulnerable library var lib gems cache rails html sanitizer gem dependency hierarchy web console gem root library actionview gem x rails html sanitizer gem vulnerable library vulnerability details rails html sanitizer is responsible for sanitizing html fragments in rails applications versions this issue is patched in version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rails html sanitizer step up your open source security game with mend | 0 |
22,699 | 4,828,049,207 | IssuesEvent | 2016-11-07 15:19:41 | facebook/react-native | https://api.github.com/repos/facebook/react-native | closed | broken link from in docs/GettingStarted.md | Documentation | There is a link to an [image](https://github.com/facebook/react-native/blob/master/docs/img/react-native-sorry-not-supported.png) in [docs/GettingStarted.md](https://github.com/facebook/react-native/blob/master/docs/GettingStarted.md) which does not exist at that link or anywhere else in the repository, as far as I can tell. | 1.0 | broken link from in docs/GettingStarted.md - There is a link to an [image](https://github.com/facebook/react-native/blob/master/docs/img/react-native-sorry-not-supported.png) in [docs/GettingStarted.md](https://github.com/facebook/react-native/blob/master/docs/GettingStarted.md) which does not exist at that link or anywhere else in the repository, as far as I can tell. | non_infrastructure | broken link from in docs gettingstarted md there is a link to an in which does not exist at that link or anywhere else in the repository as far as i can tell | 0 |
32,549 | 26,778,777,296 | IssuesEvent | 2023-01-31 19:18:06 | cds-snc/url-shortener | https://api.github.com/repos/cds-snc/url-shortener | closed | Setup Terraform project structure in repo | Infrastructure | ## Summary
Create the repo project structure that will hold the Terraform and Terragrunt configuration. [Scan Files `/terragrunt`](https://github.com/cds-snc/scan-files/tree/main/terragrunt) is a good template to follow for this. | 1.0 | Setup Terraform project structure in repo - ## Summary
Create the repo project structure that will hold the Terraform and Terragrunt configuration. [Scan Files `/terragrunt`](https://github.com/cds-snc/scan-files/tree/main/terragrunt) is a good template to follow for this. | infrastructure | setup terraform project structure in repo summary create the repo project structure that will hold the terraform and terragrunt configuration is a good template to follow for this | 1 |
7,646 | 7,041,816,610 | IssuesEvent | 2017-12-30 01:12:47 | glen3b/CyPatScoreboardBot | https://api.github.com/repos/glen3b/CyPatScoreboardBot | closed | Add user/guild/botwide setting override chain | enhancement infrastructure | Allow the bot's default setting to be overridden by the guild, and allow the guild's default setting to be overridden by an individual user - timezones is a good example. | 1.0 | Add user/guild/botwide setting override chain - Allow the bot's default setting to be overridden by the guild, and allow the guild's default setting to be overridden by an individual user - timezones is a good example. | infrastructure | add user guild botwide setting override chain allow the bot s default setting to be overridden by the guild and allow the guild s default setting to be overridden by an individual user timezones is a good example | 1 |
13,259 | 10,170,682,089 | IssuesEvent | 2019-08-08 06:17:27 | askmench/mench-web-app | https://api.github.com/repos/askmench/mench-web-app | opened | Remove ability to change intent link parent | Bugs/ASAP Fix DB/Server/Infrastructure | As we can easily unlink and link as the standard workflow... | 1.0 | Remove ability to change intent link parent - As we can easily unlink and link as the standard workflow... | infrastructure | remove ability to change intent link parent as we can easily unlink and link as the standard workflow | 1 |
18,753 | 13,190,943,701 | IssuesEvent | 2020-08-13 11:09:48 | keptn/keptn | https://api.github.com/repos/keptn/keptn | closed | Provider better indication and workflow for artifacts waiting for approval | bridge.2.0 type:usability | The Service view shows an artifact to be deployed into a stage even when it is waiting for approval. This is misleading and does not reflect the actual state. There is also not a good workflow for a user to get to approve or reject the artifact at that point. Either allow approval/reject in that view as described in #2135 or provide a link to the environment view.
The following screenshot should highlight the current implementation and how to make it better:

**Definition of Done**
- [ ] When an approval is required, add a blue border (same color as used in environment screen) to the stage that is waiting for approval
- [ ] When an approval is required, use a blue border (same color as used in environment screen) to highlight the event in the event stream
- [ ] Added the accept/decline button to the **Approval.triggered** event - this closes: #2135
| True | Provider better indication and workflow for artifacts waiting for approval - The Service view shows an artifact to be deployed into a stage even when it is waiting for approval. This is misleading and does not reflect the actual state. There is also not a good workflow for a user to get to approve or reject the artifact at that point. Either allow approval/reject in that view as described in #2135 or provide a link to the environment view.
The following screenshot should highlight the current implementation and how to make it better:

**Definition of Done**
- [ ] When an approval is required, add a blue border (same color as used in environment screen) to the stage that is waiting for approval
- [ ] When an approval is required, use a blue border (same color as used in environment screen) to highlight the event in the event stream
- [ ] Added the accept/decline button to the **Approval.triggered** event - this closes: #2135
| non_infrastructure | provider better indication and workflow for artifacts waiting for approval the service view shows an artifact to be deployed into a stage even when it is waiting for approval this is misleading and does not reflect the actual state there is also not a good workflow for a user to get to approve or reject the artifact at that point either allow approval reject in that view as described in or provide a link to the environment view the following screenshot should highlight the current implementation and how to make it better definition of done when an approval is required add a blue border same color as used in environment screen to the stage that is waiting for approval when an approval is required use a blue border same color as used in environment screen to highlight the event in the event stream added the accept decline button to the approval triggered event this closes | 0 |
24,184 | 16,990,398,227 | IssuesEvent | 2021-06-30 19:36:28 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | [Wiki] Multiple commits at the same time break image automation | restricted: maintainers status: ready type: bug work: infrastructure | **Describe the bug**
Back-to-back commits break the image automation github actions, as the local actions branch is out of sync with the remote branch. Cases [1](https://github.com/centerofci/mathesar-wiki/runs/2945791717) and [2](https://github.com/centerofci/mathesar-wiki/runs/2955127547).
**Expected behavior**
At the very least, the action should pull after committing and before pushing. In case of a merge error, we should push nothing and exit the action without throwing an error - the scripts will be run against the new commit. | 1.0 | [Wiki] Multiple commits at the same time break image automation - **Describe the bug**
Back-to-back commits break the image automation github actions, as the local actions branch is out of sync with the remote branch. Cases [1](https://github.com/centerofci/mathesar-wiki/runs/2945791717) and [2](https://github.com/centerofci/mathesar-wiki/runs/2955127547).
**Expected behavior**
At the very least, the action should pull after committing and before pushing. In case of a merge error, we should push nothing and exit the action without throwing an error - the scripts will be run against the new commit. | infrastructure | multiple commits at the same time break image automation describe the bug back to back commits break the image automation github actions as the local actions branch is out of sync with the remote branch cases and expected behavior at the very least the action should pull after committing and before pushing in case of a merge error we should push nothing and exit the action without throwing an error the scripts will be run against the new commit | 1 |
279,454 | 21,161,017,051 | IssuesEvent | 2022-04-07 09:22:10 | Decathlon/tzatziki | https://api.github.com/repos/Decathlon/tzatziki | closed | bug: "after \\d+ms|" checks are not always done. | documentation | It seems that guard which contain "after Xms" are not check.
It seems that when using this keyword, the check are done in async, by using another thread from the main one.
I suppose that as the main thread is over before the Xms duration, then the "sub threads" are over too, and so the checks are not done.

PS: Maybe we should have some "labels" available in order to organize the issues?
| 1.0 | bug: "after \\d+ms|" checks are not always done. - It seems that guard which contain "after Xms" are not check.
It seems that when using this keyword, the check are done in async, by using another thread from the main one.
I suppose that as the main thread is over before the Xms duration, then the "sub threads" are over too, and so the checks are not done.

PS: Maybe we should have some "labels" available in order to organize the issues?
| non_infrastructure | bug after d ms checks are not always done it seems that guard which contain after xms are not check it seems that when using this keyword the check are done in async by using another thread from the main one i suppose that as the main thread is over before the xms duration then the sub threads are over too and so the checks are not done ps maybe we should have some labels available in order to organize the issues | 0 |
395,898 | 27,092,337,066 | IssuesEvent | 2023-02-14 22:17:49 | edgedb/edgedb | https://api.github.com/repos/edgedb/edgedb | closed | Document version pinning behavior for `edgedb instance create` et al | documentation | Currently
```
edgedb instance create --version 2.6 demo2something
```
will install the latest version of 2.x while in order to install 2.6 you need to put an `=` in front, like so:
```
edgedb instance create --version =2.6 demo26
```
There is similar behavior in pyproject.toml.
We should document this.
(I also still think that on the command line, it should install the actual specified version, but I won't insist on relitigating that) | 1.0 | Document version pinning behavior for `edgedb instance create` et al - Currently
```
edgedb instance create --version 2.6 demo2something
```
will install the latest version of 2.x while in order to install 2.6 you need to put an `=` in front, like so:
```
edgedb instance create --version =2.6 demo26
```
There is similar behavior in pyproject.toml.
We should document this.
(I also still think that on the command line, it should install the actual specified version, but I won't insist on relitigating that) | non_infrastructure | document version pinning behavior for edgedb instance create et al currently edgedb instance create version will install the latest version of x while in order to install you need to put an in front like so edgedb instance create version there is similar behavior in pyproject toml we should document this i also still think that on the command line it should install the actual specified version but i won t insist on relitigating that | 0 |
46,191 | 5,791,681,540 | IssuesEvent | 2017-05-02 07:01:40 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | reopened | TopicOnReconnectTest.testListenersNonSmartRoutingMultipleServer() | Team: Client Type: Test-Failure | ```
java.lang.AssertionError: Values should be different. Actual: 0
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failEquals(Assert.java:185)
at org.junit.Assert.assertNotEquals(Assert.java:199)
at org.junit.Assert.assertNotEquals(Assert.java:211)
at com.hazelcast.client.listeners.AbstractListenersOnReconnectTest$1.run(AbstractListenersOnReconnectTest.java:64)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:901)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:915)
at com.hazelcast.client.listeners.AbstractListenersOnReconnectTest.testListenersInternal(AbstractListenersOnReconnectTest.java:61)
at com.hazelcast.client.listeners.AbstractListenersOnReconnectTest.testListenersNonSmartRoutingMultipleServer(AbstractListenersOnReconnectTest.java:131)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:88)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:82)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at java.lang.Thread.run(Thread.java:662)
```
https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/4186/testReport/junit/com.hazelcast.client.listeners/TopicOnReconnectTest/testListenersNonSmartRoutingMultipleServer/
| 1.0 | TopicOnReconnectTest.testListenersNonSmartRoutingMultipleServer() - ```
java.lang.AssertionError: Values should be different. Actual: 0
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failEquals(Assert.java:185)
at org.junit.Assert.assertNotEquals(Assert.java:199)
at org.junit.Assert.assertNotEquals(Assert.java:211)
at com.hazelcast.client.listeners.AbstractListenersOnReconnectTest$1.run(AbstractListenersOnReconnectTest.java:64)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:901)
at com.hazelcast.test.HazelcastTestSupport.assertTrueEventually(HazelcastTestSupport.java:915)
at com.hazelcast.client.listeners.AbstractListenersOnReconnectTest.testListenersInternal(AbstractListenersOnReconnectTest.java:61)
at com.hazelcast.client.listeners.AbstractListenersOnReconnectTest.testListenersNonSmartRoutingMultipleServer(AbstractListenersOnReconnectTest.java:131)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:88)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:82)
at java.util.concurrent.FutureTask$Sync.innerRun(FutureTask.java:303)
at java.util.concurrent.FutureTask.run(FutureTask.java:138)
at java.lang.Thread.run(Thread.java:662)
```
https://hazelcast-l337.ci.cloudbees.com/job/new-lab-fast-pr/4186/testReport/junit/com.hazelcast.client.listeners/TopicOnReconnectTest/testListenersNonSmartRoutingMultipleServer/
| non_infrastructure | topiconreconnecttest testlistenersnonsmartroutingmultipleserver java lang assertionerror values should be different actual at org junit assert fail assert java at org junit assert failequals assert java at org junit assert assertnotequals assert java at org junit assert assertnotequals assert java at com hazelcast client listeners abstractlistenersonreconnecttest run abstractlistenersonreconnecttest java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast test hazelcasttestsupport asserttrueeventually hazelcasttestsupport java at com hazelcast client listeners abstractlistenersonreconnecttest testlistenersinternal abstractlistenersonreconnecttest java at com hazelcast client listeners abstractlistenersonreconnecttest testlistenersnonsmartroutingmultipleserver abstractlistenersonreconnecttest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask sync innerrun futuretask java at java util concurrent futuretask run futuretask java at java lang thread run thread java | 0 |
108,670 | 11,597,919,966 | IssuesEvent | 2020-02-24 21:53:58 | tobiasanker/libKitsunemimiCommon | https://api.github.com/repos/tobiasanker/libKitsunemimiCommon | closed | add benchmark-tests | documentation feature / enhancement | similar to the unit-tests, some simple benchmark-tests should be added for performance-analytic | 1.0 | add benchmark-tests - similar to the unit-tests, some simple benchmark-tests should be added for performance-analytic | non_infrastructure | add benchmark tests similar to the unit tests some simple benchmark tests should be added for performance analytic | 0 |
22,539 | 15,256,464,870 | IssuesEvent | 2021-02-20 20:19:46 | Daniel-Mietchen/ideas | https://api.github.com/repos/Daniel-Mietchen/ideas | opened | Look into Cloud68 | infrastructure open science | as per https://cloud68.co/
> Reliable open source digital infrastructure!
Managing Nextcloud & other #WorkFromHome platforms for you
(so that you don't have to). | 1.0 | Look into Cloud68 - as per https://cloud68.co/
> Reliable open source digital infrastructure!
Managing Nextcloud & other #WorkFromHome platforms for you
(so that you don't have to). | infrastructure | look into as per reliable open source digital infrastructure managing nextcloud other workfromhome platforms for you so that you don t have to | 1 |
273,996 | 29,865,521,345 | IssuesEvent | 2023-06-20 03:14:12 | Chiencc/Express | https://api.github.com/repos/Chiencc/Express | closed | WS-2020-0163 (Medium) detected in marked-0.7.0.tgz - autoclosed | Mend: dependency security vulnerability | ## WS-2020-0163 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.7.0.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.7.0.tgz">https://registry.npmjs.org/marked/-/marked-0.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/marked/package.json</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/Express/commit/0673d1f4df72daf41a574593e8ee445dc0a4766d">0673d1f4df72daf41a574593e8ee445dc0a4766d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 1.1.1 is vulnerable to Regular Expression Denial of Service (REDoS). rules.js have multiple unused capture groups which can lead to a Denial of Service.
<p>Publish Date: 2020-07-02
<p>URL: <a href=https://github.com/markedjs/marked/commit/bd4f8c464befad2b304d51e33e89e567326e62e0>WS-2020-0163</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-07-02</p>
<p>Fix Resolution: 1.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2020-0163 (Medium) detected in marked-0.7.0.tgz - autoclosed - ## WS-2020-0163 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.7.0.tgz</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.7.0.tgz">https://registry.npmjs.org/marked/-/marked-0.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/marked/package.json</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/Express/commit/0673d1f4df72daf41a574593e8ee445dc0a4766d">0673d1f4df72daf41a574593e8ee445dc0a4766d</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
marked before 1.1.1 is vulnerable to Regular Expression Denial of Service (REDoS). rules.js have multiple unused capture groups which can lead to a Denial of Service.
<p>Publish Date: 2020-07-02
<p>URL: <a href=https://github.com/markedjs/marked/commit/bd4f8c464befad2b304d51e33e89e567326e62e0>WS-2020-0163</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-07-02</p>
<p>Fix Resolution: 1.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | ws medium detected in marked tgz autoclosed ws medium severity vulnerability vulnerable library marked tgz a markdown parser built for speed library home page a href path to dependency file package json path to vulnerable library node modules marked package json dependency hierarchy x marked tgz vulnerable library found in head commit a href found in base branch master vulnerability details marked before is vulnerable to regular expression denial of service redos rules js have multiple unused capture groups which can lead to a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution step up your open source security game with mend | 0 |
53,666 | 3,043,687,181 | IssuesEvent | 2015-08-10 00:31:30 | MarchingLizard/iwiorg | https://api.github.com/repos/MarchingLizard/iwiorg | opened | Countdown Bar Versions | low-priority |
*Version One: Registration not yet open*
Pre-register
Should have countdown like it does now
Verbiage
Help secure your spot for a future trip
Pre-register for winter and summer trips
*Version Two: Registration Open*
Countdown is replaced with apply now button
Verbiage:
What are you waiting for?
Apply now for December, January, and February Trips
You can still pre-register for Summer 2016 trips here
Question: Does that make sense?
*Version Three: Registration closing soon*
Time is running out...don't miss your chance
Apply now for December, January, and February TripsYou can still pre-register for Summer 2016 trips here
*Version Four: Registration is now closed*
Clock changes to pre-register now button
Verbiage:
Winter 2015/2016registration is now closed.
Help secure your spot for summer 2016 and winter 2016/17 by pre-registering now. | 1.0 | Countdown Bar Versions -
*Version One: Registration not yet open*
Pre-register
Should have countdown like it does now
Verbiage
Help secure your spot for a future trip
Pre-register for winter and summer trips
*Version Two: Registration Open*
Countdown is replaced with apply now button
Verbiage:
What are you waiting for?
Apply now for December, January, and February Trips
You can still pre-register for Summer 2016 trips here
Question: Does that make sense?
*Version Three: Registration closing soon*
Time is running out...don't miss your chance
Apply now for December, January, and February TripsYou can still pre-register for Summer 2016 trips here
*Version Four: Registration is now closed*
Clock changes to pre-register now button
Verbiage:
Winter 2015/2016registration is now closed.
Help secure your spot for summer 2016 and winter 2016/17 by pre-registering now. | non_infrastructure | countdown bar versions version one registration not yet open pre register should have countdown like it does now verbiage help secure your spot for a future trip pre register for winter and summer trips version two registration open countdown is replaced with apply now button verbiage what are you waiting for apply now for december january and february trips you can still pre register for summer trips here question does that make sense version three registration closing soon time is running out don t miss your chance apply now for december january and february tripsyou can still pre register for summer trips here version four registration is now closed clock changes to pre register now button verbiage winter is now closed help secure your spot for summer and winter by pre registering now | 0 |
4,872 | 5,309,370,341 | IssuesEvent | 2017-02-12 09:38:56 | refu-lang/refu | https://api.github.com/repos/refu-lang/refu | opened | Make compatible with Windows OS | infrastructure | Nobody has ever attempted to compile refu in Windows. I believe it will require some work to make it happen. This issue is to track work towards that goal. | 1.0 | Make compatible with Windows OS - Nobody has ever attempted to compile refu in Windows. I believe it will require some work to make it happen. This issue is to track work towards that goal. | infrastructure | make compatible with windows os nobody has ever attempted to compile refu in windows i believe it will require some work to make it happen this issue is to track work towards that goal | 1 |
1,484 | 3,248,038,796 | IssuesEvent | 2015-10-17 00:39:53 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | opened | [Blocked] dci-win-rbld-12 has run out of disk space | Area-Infrastructure Known Contributor Issue | [dci-win-rbld-12](http://dotnet-ci.cloudapp.net/job/dotnet_roslyn_prtest_win_dbg_eta/1070/console) has run out of disk space, preventing ETA runs. | 1.0 | [Blocked] dci-win-rbld-12 has run out of disk space - [dci-win-rbld-12](http://dotnet-ci.cloudapp.net/job/dotnet_roslyn_prtest_win_dbg_eta/1070/console) has run out of disk space, preventing ETA runs. | infrastructure | dci win rbld has run out of disk space has run out of disk space preventing eta runs | 1 |
274,063 | 29,866,170,203 | IssuesEvent | 2023-06-20 04:12:02 | Chiencc/asuswrt-gt-ac5300 | https://api.github.com/repos/Chiencc/asuswrt-gt-ac5300 | closed | quaggaquagga-0.99.24.1: 6 vulnerabilities (highest severity is: 9.8) - autoclosed | Mend: dependency security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p></summary>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
</p>
</p></p>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/zebra/rtadv.c</b>
</p>
<p></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (quaggaquagga version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2016-1245](https://www.mend.io/vulnerability-database/CVE-2016-1245) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | quaggaquagga-0.99.24.1 | Direct | 1.0.20161017 | ❌ |
| [CVE-2016-2342](https://www.mend.io/vulnerability-database/CVE-2016-2342) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 8.1 | quaggaquagga-0.99.24.1 | Direct | 1.0.20160309 | ❌ |
| [CVE-2017-5495](https://www.mend.io/vulnerability-database/CVE-2017-5495) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | quaggaquagga-0.99.24.1 | Direct | quagga-1.1.1 | ❌ |
| [CVE-2017-16227](https://www.mend.io/vulnerability-database/CVE-2017-16227) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | quaggaquagga-0.99.24.1 | Direct | 1.2.2 | ❌ |
| [CVE-2012-0249](https://www.mend.io/vulnerability-database/CVE-2012-0249) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.3 | quaggaquagga-0.99.24.1 | Direct | 0.99.20.1 | ❌ |
| [CVE-2012-0250](https://www.mend.io/vulnerability-database/CVE-2012-0250) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.3 | quaggaquagga-0.99.24.1 | Direct | 0.99.20.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2016-1245</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/zebra/rtadv.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
It was discovered that the zebra daemon in Quagga before 1.0.20161017 suffered from a stack-based buffer overflow when processing IPv6 Neighbor Discovery messages. The root cause was relying on BUFSIZ to be compatible with a message size; however, BUFSIZ is system-dependent.
<p>Publish Date: 2017-02-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-1245>CVE-2016-1245</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-1245">https://nvd.nist.gov/vuln/detail/CVE-2016-1245</a></p>
<p>Release Date: 2017-02-22</p>
<p>Fix Resolution: 1.0.20161017</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2016-2342</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (3)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_mplsvpn.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_mplsvpn.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_mplsvpn.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
The bgp_nlri_parse_vpnv4 function in bgp_mplsvpn.c in the VPNv4 NLRI parser in bgpd in Quagga before 1.0.20160309, when a certain VPNv4 configuration is used, relies on a Labeled-VPN SAFI routes-data length field during a data copy, which allows remote attackers to execute arbitrary code or cause a denial of service (stack-based buffer overflow) via a crafted packet.
<p>Publish Date: 2016-03-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-2342>CVE-2016-2342</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2342">https://nvd.nist.gov/vuln/detail/CVE-2016-2342</a></p>
<p>Release Date: 2016-03-17</p>
<p>Fix Resolution: 1.0.20160309</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2017-5495</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/lib/vty.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of Quagga, 0.93 through 1.1.0, are vulnerable to an unbounded memory allocation in the telnet 'vty' CLI, leading to a Denial-of-Service of Quagga daemons, or even the entire host. When Quagga daemons are configured with their telnet CLI enabled, anyone who can connect to the TCP ports can trigger this vulnerability, prior to authentication. Most distributions restrict the Quagga telnet interface to local access only by default. The Quagga telnet interface 'vty' input buffer grows automatically, without bound, so long as a newline is not entered. This allows an attacker to cause the Quagga daemon to allocate unbounded memory by sending very long strings without a newline. Eventually the daemon is terminated by the system, or the system itself runs out of memory. This is fixed in Quagga 1.1.1 and Free Range Routing (FRR) Protocol Suite 2017-01-10.
<p>Publish Date: 2017-01-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-5495>CVE-2017-5495</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-5495">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-5495</a></p>
<p>Release Date: 2017-01-24</p>
<p>Fix Resolution: quagga-1.1.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2017-16227</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_aspath.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
The aspath_put function in bgpd/bgp_aspath.c in Quagga before 1.2.2 allows remote attackers to cause a denial of service (session drop) via BGP UPDATE messages, because AS_PATH size calculation for long paths counts certain bytes twice and consequently constructs an invalid message.
<p>Publish Date: 2017-10-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-16227>CVE-2017-16227</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16227">https://nvd.nist.gov/vuln/detail/CVE-2017-16227</a></p>
<p>Release Date: 2017-10-29</p>
<p>Fix Resolution: 1.2.2</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-0249</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (3)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
Buffer overflow in the ospf_ls_upd_list_lsa function in ospf_packet.c in the OSPFv2 implementation in ospfd in Quagga before 0.99.20.1 allows remote attackers to cause a denial of service (assertion failure and daemon exit) via a Link State Update (aka LS Update) packet that is smaller than the length specified in its header.
<p>Publish Date: 2012-04-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-0249>CVE-2012-0249</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249</a></p>
<p>Release Date: 2012-04-05</p>
<p>Fix Resolution: 0.99.20.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-0250</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (3)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
Buffer overflow in the OSPFv2 implementation in ospfd in Quagga before 0.99.20.1 allows remote attackers to cause a denial of service (daemon crash) via a Link State Update (aka LS Update) packet containing a network-LSA link-state advertisement for which the data-structure length is smaller than the value in the Length header field.
<p>Publish Date: 2012-04-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-0250>CVE-2012-0250</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249</a></p>
<p>Release Date: 2012-04-05</p>
<p>Fix Resolution: 0.99.20.1</p>
</p>
<p></p>
</details> | True | quaggaquagga-0.99.24.1: 6 vulnerabilities (highest severity is: 9.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p></summary>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
</p>
</p></p>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/zebra/rtadv.c</b>
</p>
<p></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (quaggaquagga version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2016-1245](https://www.mend.io/vulnerability-database/CVE-2016-1245) | <img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Critical | 9.8 | quaggaquagga-0.99.24.1 | Direct | 1.0.20161017 | ❌ |
| [CVE-2016-2342](https://www.mend.io/vulnerability-database/CVE-2016-2342) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 8.1 | quaggaquagga-0.99.24.1 | Direct | 1.0.20160309 | ❌ |
| [CVE-2017-5495](https://www.mend.io/vulnerability-database/CVE-2017-5495) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | quaggaquagga-0.99.24.1 | Direct | quagga-1.1.1 | ❌ |
| [CVE-2017-16227](https://www.mend.io/vulnerability-database/CVE-2017-16227) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> High | 7.5 | quaggaquagga-0.99.24.1 | Direct | 1.2.2 | ❌ |
| [CVE-2012-0249](https://www.mend.io/vulnerability-database/CVE-2012-0249) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.3 | quaggaquagga-0.99.24.1 | Direct | 0.99.20.1 | ❌ |
| [CVE-2012-0250](https://www.mend.io/vulnerability-database/CVE-2012-0250) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 4.3 | quaggaquagga-0.99.24.1 | Direct | 0.99.20.1 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> CVE-2016-1245</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/zebra/rtadv.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
It was discovered that the zebra daemon in Quagga before 1.0.20161017 suffered from a stack-based buffer overflow when processing IPv6 Neighbor Discovery messages. The root cause was relying on BUFSIZ to be compatible with a message size; however, BUFSIZ is system-dependent.
<p>Publish Date: 2017-02-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-1245>CVE-2016-1245</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-1245">https://nvd.nist.gov/vuln/detail/CVE-2016-1245</a></p>
<p>Release Date: 2017-02-22</p>
<p>Fix Resolution: 1.0.20161017</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2016-2342</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (3)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_mplsvpn.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_mplsvpn.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_mplsvpn.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
The bgp_nlri_parse_vpnv4 function in bgp_mplsvpn.c in the VPNv4 NLRI parser in bgpd in Quagga before 1.0.20160309, when a certain VPNv4 configuration is used, relies on a Labeled-VPN SAFI routes-data length field during a data copy, which allows remote attackers to execute arbitrary code or cause a denial of service (stack-based buffer overflow) via a crafted packet.
<p>Publish Date: 2016-03-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2016-2342>CVE-2016-2342</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2342">https://nvd.nist.gov/vuln/detail/CVE-2016-2342</a></p>
<p>Release Date: 2016-03-17</p>
<p>Fix Resolution: 1.0.20160309</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2017-5495</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/lib/vty.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of Quagga, 0.93 through 1.1.0, are vulnerable to an unbounded memory allocation in the telnet 'vty' CLI, leading to a Denial-of-Service of Quagga daemons, or even the entire host. When Quagga daemons are configured with their telnet CLI enabled, anyone who can connect to the TCP ports can trigger this vulnerability, prior to authentication. Most distributions restrict the Quagga telnet interface to local access only by default. The Quagga telnet interface 'vty' input buffer grows automatically, without bound, so long as a newline is not entered. This allows an attacker to cause the Quagga daemon to allocate unbounded memory by sending very long strings without a newline. Eventually the daemon is terminated by the system, or the system itself runs out of memory. This is fixed in Quagga 1.1.1 and Free Range Routing (FRR) Protocol Suite 2017-01-10.
<p>Publish Date: 2017-01-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-5495>CVE-2017-5495</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-5495">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-5495</a></p>
<p>Release Date: 2017-01-24</p>
<p>Fix Resolution: quagga-1.1.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> CVE-2017-16227</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/bgpd/bgp_aspath.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
The aspath_put function in bgpd/bgp_aspath.c in Quagga before 1.2.2 allows remote attackers to cause a denial of service (session drop) via BGP UPDATE messages, because AS_PATH size calculation for long paths counts certain bytes twice and consequently constructs an invalid message.
<p>Publish Date: 2017-10-29
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-16227>CVE-2017-16227</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2017-16227">https://nvd.nist.gov/vuln/detail/CVE-2017-16227</a></p>
<p>Release Date: 2017-10-29</p>
<p>Fix Resolution: 1.2.2</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-0249</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (3)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
Buffer overflow in the ospf_ls_upd_list_lsa function in ospf_packet.c in the OSPFv2 implementation in ospfd in Quagga before 0.99.20.1 allows remote attackers to cause a denial of service (assertion failure and daemon exit) via a Link State Update (aka LS Update) packet that is smaller than the length specified in its header.
<p>Publish Date: 2012-04-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-0249>CVE-2012-0249</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249</a></p>
<p>Release Date: 2012-04-05</p>
<p>Fix Resolution: 0.99.20.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2012-0250</summary>
### Vulnerable Library - <b>quaggaquagga-0.99.24.1</b></p>
<p>
<p>quagga</p>
<p>Library home page: <a href=https://git.savannah.gnu.org/cgit/quagga.git>https://git.savannah.gnu.org/cgit/quagga.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Chiencc/asuswrt-gt-ac5300/commit/0c45ce909374d16605095db4fce9a89b9b6bafd5">0c45ce909374d16605095db4fce9a89b9b6bafd5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (3)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/release/src/router/quagga/ospfd/ospf_packet.h</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
Buffer overflow in the OSPFv2 implementation in ospfd in Quagga before 0.99.20.1 allows remote attackers to cause a denial of service (daemon crash) via a Link State Update (aka LS Update) packet containing a network-LSA link-state advertisement for which the data-structure length is smaller than the value in the Length header field.
<p>Publish Date: 2012-04-05
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2012-0250>CVE-2012-0250</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Adjacent
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-0249</a></p>
<p>Release Date: 2012-04-05</p>
<p>Fix Resolution: 0.99.20.1</p>
</p>
<p></p>
</details> | non_infrastructure | quaggaquagga vulnerabilities highest severity is autoclosed vulnerable library quaggaquagga quagga library home page a href found in head commit a href vulnerable source files release src router quagga zebra rtadv c vulnerabilities cve severity cvss dependency type fixed in quaggaquagga version remediation available critical quaggaquagga direct high quaggaquagga direct high quaggaquagga direct quagga high quaggaquagga direct medium quaggaquagga direct medium quaggaquagga direct details cve vulnerable library quaggaquagga quagga library home page a href found in head commit a href found in base branch master vulnerable source files release src router quagga zebra rtadv c vulnerability details it was discovered that the zebra daemon in quagga before suffered from a stack based buffer overflow when processing neighbor discovery messages the root cause was relying on bufsiz to be compatible with a message size however bufsiz is system dependent publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cve vulnerable library quaggaquagga quagga library home page a href found in head commit a href found in base branch master vulnerable source files release src router quagga bgpd bgp mplsvpn c release src router quagga bgpd bgp mplsvpn c release src router quagga bgpd bgp mplsvpn c vulnerability details the bgp nlri parse function in bgp mplsvpn c in the nlri parser in bgpd in quagga before when a certain configuration is used relies on a labeled vpn safi routes data length field during a data copy which allows remote attackers to execute arbitrary code or cause a denial of service stack based buffer overflow via a crafted packet publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cve vulnerable library quaggaquagga quagga library home page a href found in head commit a href found in base branch master vulnerable source files release src router quagga lib vty c vulnerability details all versions of quagga through are vulnerable to an unbounded memory allocation in the telnet vty cli leading to a denial of service of quagga daemons or even the entire host when quagga daemons are configured with their telnet cli enabled anyone who can connect to the tcp ports can trigger this vulnerability prior to authentication most distributions restrict the quagga telnet interface to local access only by default the quagga telnet interface vty input buffer grows automatically without bound so long as a newline is not entered this allows an attacker to cause the quagga daemon to allocate unbounded memory by sending very long strings without a newline eventually the daemon is terminated by the system or the system itself runs out of memory this is fixed in quagga and free range routing frr protocol suite publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution quagga cve vulnerable library quaggaquagga quagga library home page a href found in head commit a href found in base branch master vulnerable source files release src router quagga bgpd bgp aspath c vulnerability details the aspath put function in bgpd bgp aspath c in quagga before allows remote attackers to cause a denial of service session drop via bgp update messages because as path size calculation for long paths counts certain bytes twice and consequently constructs an invalid message publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cve vulnerable library quaggaquagga quagga library home page a href found in head commit a href found in base branch master vulnerable source files release src router quagga ospfd ospf packet h release src router quagga ospfd ospf packet h release src router quagga ospfd ospf packet h vulnerability details buffer overflow in the ospf ls upd list lsa function in ospf packet c in the implementation in ospfd in quagga before allows remote attackers to cause a denial of service assertion failure and daemon exit via a link state update aka ls update packet that is smaller than the length specified in its header publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cve vulnerable library quaggaquagga quagga library home page a href found in head commit a href found in base branch master vulnerable source files release src router quagga ospfd ospf packet h release src router quagga ospfd ospf packet h release src router quagga ospfd ospf packet h vulnerability details buffer overflow in the implementation in ospfd in quagga before allows remote attackers to cause a denial of service daemon crash via a link state update aka ls update packet containing a network lsa link state advertisement for which the data structure length is smaller than the value in the length header field publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
26,482 | 20,152,351,046 | IssuesEvent | 2022-02-09 13:37:08 | amosproj/amos2021ws05-fin-prod-port-quick-check | https://api.github.com/repos/amosproj/amos2021ws05-fin-prod-port-quick-check | closed | finish Frontend Docker Container | est. size: 2 type: infrastructure real size: 2 priority: high | ## User story
1. As a developer
2. I want a frontend container
3. So that everything runs in a container
## Acceptance criteria
- [x] Frontend Docker Container exists
- [x] Frontend Docker Container is used
## Definition of done
* Approved by product owner
* Code has been peer reviewed and approved
* No syntax or runtime errors emerged
* Code has to be included in the release candidate
| 1.0 | finish Frontend Docker Container - ## User story
1. As a developer
2. I want a frontend container
3. So that everything runs in a container
## Acceptance criteria
- [x] Frontend Docker Container exists
- [x] Frontend Docker Container is used
## Definition of done
* Approved by product owner
* Code has been peer reviewed and approved
* No syntax or runtime errors emerged
* Code has to be included in the release candidate
| infrastructure | finish frontend docker container user story as a developer i want a frontend container so that everything runs in a container acceptance criteria frontend docker container exists frontend docker container is used definition of done approved by product owner code has been peer reviewed and approved no syntax or runtime errors emerged code has to be included in the release candidate | 1 |
526,960 | 15,305,599,809 | IssuesEvent | 2021-02-24 18:19:55 | celo-org/celo-monorepo | https://api.github.com/repos/celo-org/celo-monorepo | closed | Eng Support for WalletConnect Design Scoping | Priority: P1 devX feature | @AlexBHarley - ToDo: Create separate issues for this
Engineering design
Investigate, then Valora side of things | 1.0 | Eng Support for WalletConnect Design Scoping - @AlexBHarley - ToDo: Create separate issues for this
Engineering design
Investigate, then Valora side of things | non_infrastructure | eng support for walletconnect design scoping alexbharley todo create separate issues for this engineering design investigate then valora side of things | 0 |
29,984 | 24,450,894,129 | IssuesEvent | 2022-10-06 22:50:50 | GaloisInc/cclyzerpp | https://api.github.com/repos/GaloisInc/cclyzerpp | closed | Determine how long Docker images are kept | infrastructure | The CI system now builds and pushes two Docker images for each commit on a PR or to `main`. This is convenient, but this convenience should be balanced against cost of storage. We should determine how long these images are stored, and possibly revise when they are pushed to GHCR. | 1.0 | Determine how long Docker images are kept - The CI system now builds and pushes two Docker images for each commit on a PR or to `main`. This is convenient, but this convenience should be balanced against cost of storage. We should determine how long these images are stored, and possibly revise when they are pushed to GHCR. | infrastructure | determine how long docker images are kept the ci system now builds and pushes two docker images for each commit on a pr or to main this is convenient but this convenience should be balanced against cost of storage we should determine how long these images are stored and possibly revise when they are pushed to ghcr | 1 |
10,672 | 8,676,809,687 | IssuesEvent | 2018-11-30 15:08:31 | trufflesuite/truffle | https://api.github.com/repos/trufflesuite/truffle | closed | Add integration tests for `truffle init` | Test Infrastructure help wanted | ## Issue
`truffle init` currently has no integration tests! this can inadvertently allow breaking of `truffle init` when refactoring or modifying the `truffle-box` package (specifically when playing around with `truffle unbox <options>`).
## Requirements
The scenario tests should at least exercise a basic `truffle init` . | 1.0 | Add integration tests for `truffle init` - ## Issue
`truffle init` currently has no integration tests! this can inadvertently allow breaking of `truffle init` when refactoring or modifying the `truffle-box` package (specifically when playing around with `truffle unbox <options>`).
## Requirements
The scenario tests should at least exercise a basic `truffle init` . | infrastructure | add integration tests for truffle init issue truffle init currently has no integration tests this can inadvertently allow breaking of truffle init when refactoring or modifying the truffle box package specifically when playing around with truffle unbox requirements the scenario tests should at least exercise a basic truffle init | 1 |
16,488 | 21,445,564,491 | IssuesEvent | 2022-04-25 05:45:54 | zotero/zotero | https://api.github.com/repos/zotero/zotero | opened | Classic citation dialog: Add all selected items to Multiple Sources pane | Papercuts Word Processor Integration Bug | https://forums.zotero.org/discussion/96671/citing-multiple-papers
Not that we want to do more to the classic dialog, but this seems like a trivial fix to a clear bug (albeit one that has probably existed for a decade), and it will help people until we have a new citation dialog. | 1.0 | Classic citation dialog: Add all selected items to Multiple Sources pane - https://forums.zotero.org/discussion/96671/citing-multiple-papers
Not that we want to do more to the classic dialog, but this seems like a trivial fix to a clear bug (albeit one that has probably existed for a decade), and it will help people until we have a new citation dialog. | non_infrastructure | classic citation dialog add all selected items to multiple sources pane not that we want to do more to the classic dialog but this seems like a trivial fix to a clear bug albeit one that has probably existed for a decade and it will help people until we have a new citation dialog | 0 |
29,090 | 23,721,243,254 | IssuesEvent | 2022-08-30 15:30:19 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Undo revert of Arcade update to ingest Workload Manifest task | blocking-servicing area-Infrastructure-mono | Basically, undo https://github.com/dotnet/runtime/pull/71110 (but don't undo the SDK update), and fix up the publish task as described below (copy/pasted from the PR):
```
Runtime will have to take an arcade update, a llvm-project update, and Jacques' change next month, plus a fix to the publish stage invoked here: https://github.com/dotnet/runtime/blob/cd19e8bf35d18cef6f30d896313cbeb1be2a06dd/eng/pipelines/official/jobs/prepare-signed-artifacts.yml#L48-L57. The problem seems to be that you're trying to upload all blobs to `blobArtifacts`, but there are now .msi's with duplicate names. From @joeloff:
> [7:52 PM] Jacques Eloff
it organizes builds by feature band for VS insertions, so MSIs can be duplicated in multiple folders. Copying them the same location won't work.
> because in staging, each sets of files for each feature band is published to a separate VSDROP container and there can be overlap, but not all blobbies have the same files.
Wanted to make sure this was on your radar. The failure happens during the AzDO artifact upload task that happens at the end of that step:
> Async Command Start: Upload Artifact
> ...
> Fail to upload 'D:\a\_work\1\s\artifacts\PackageDownload\IntermediateArtifacts\workloads-vs\Insertion\6.0.300\Microsoft.AOT.win-x64.Cross.android-arm.6.0.7.msi' due to 'Blob is incomplete (missing block). Blob: af95b7c0d1f1ec11b656501ac504b8de, Expected Offset: 0, Actual Offset: 4194304'.
```
The root cause of this was the change to the workload manifest tasks in Arcade: https://github.com/dotnet/arcade/pull/8645
This requires reaction in runtime's usage of those tasks: https://github.com/dotnet/runtime/blob/6212eabc5efa507ee0ca2fc21dfdc327abc73c0a/src/workloads/workloads.csproj#L113-L133
We tried to react with https://github.com/dotnet/runtime/pull/71046, but that caused us to try to publish multiple .msi's with the same name to the same place. Basically we need to publish the output of the new task & preserve its output folder structure, rather than flattening. | 1.0 | Undo revert of Arcade update to ingest Workload Manifest task - Basically, undo https://github.com/dotnet/runtime/pull/71110 (but don't undo the SDK update), and fix up the publish task as described below (copy/pasted from the PR):
```
Runtime will have to take an arcade update, a llvm-project update, and Jacques' change next month, plus a fix to the publish stage invoked here: https://github.com/dotnet/runtime/blob/cd19e8bf35d18cef6f30d896313cbeb1be2a06dd/eng/pipelines/official/jobs/prepare-signed-artifacts.yml#L48-L57. The problem seems to be that you're trying to upload all blobs to `blobArtifacts`, but there are now .msi's with duplicate names. From @joeloff:
> [7:52 PM] Jacques Eloff
it organizes builds by feature band for VS insertions, so MSIs can be duplicated in multiple folders. Copying them the same location won't work.
> because in staging, each sets of files for each feature band is published to a separate VSDROP container and there can be overlap, but not all blobbies have the same files.
Wanted to make sure this was on your radar. The failure happens during the AzDO artifact upload task that happens at the end of that step:
> Async Command Start: Upload Artifact
> ...
> Fail to upload 'D:\a\_work\1\s\artifacts\PackageDownload\IntermediateArtifacts\workloads-vs\Insertion\6.0.300\Microsoft.AOT.win-x64.Cross.android-arm.6.0.7.msi' due to 'Blob is incomplete (missing block). Blob: af95b7c0d1f1ec11b656501ac504b8de, Expected Offset: 0, Actual Offset: 4194304'.
```
The root cause of this was the change to the workload manifest tasks in Arcade: https://github.com/dotnet/arcade/pull/8645
This requires reaction in runtime's usage of those tasks: https://github.com/dotnet/runtime/blob/6212eabc5efa507ee0ca2fc21dfdc327abc73c0a/src/workloads/workloads.csproj#L113-L133
We tried to react with https://github.com/dotnet/runtime/pull/71046, but that caused us to try to publish multiple .msi's with the same name to the same place. Basically we need to publish the output of the new task & preserve its output folder structure, rather than flattening. | infrastructure | undo revert of arcade update to ingest workload manifest task basically undo but don t undo the sdk update and fix up the publish task as described below copy pasted from the pr runtime will have to take an arcade update a llvm project update and jacques change next month plus a fix to the publish stage invoked here the problem seems to be that you re trying to upload all blobs to blobartifacts but there are now msi s with duplicate names from joeloff jacques eloff it organizes builds by feature band for vs insertions so msis can be duplicated in multiple folders copying them the same location won t work because in staging each sets of files for each feature band is published to a separate vsdrop container and there can be overlap but not all blobbies have the same files wanted to make sure this was on your radar the failure happens during the azdo artifact upload task that happens at the end of that step async command start upload artifact fail to upload d a work s artifacts packagedownload intermediateartifacts workloads vs insertion microsoft aot win cross android arm msi due to blob is incomplete missing block blob expected offset actual offset the root cause of this was the change to the workload manifest tasks in arcade this requires reaction in runtime s usage of those tasks we tried to react with but that caused us to try to publish multiple msi s with the same name to the same place basically we need to publish the output of the new task preserve its output folder structure rather than flattening | 1 |
6,267 | 6,279,819,168 | IssuesEvent | 2017-07-18 17:03:10 | gparmer/composite | https://api.github.com/repos/gparmer/composite | opened | Fix cos_kernel_api races | high infrastructure | Any updates to two memory locations in a single cos_kernel_api operation must take great pains to handle races. For example, if we need to increase both the `vasrange_frontier` and the `vasfrontier`, we can likely use a `faa` for the first, and a `cas` to the second, taking into account concurrent changes. The issue is that another thread might call API functions that can return errors that must also understand this race.
TODO: fix all of these instances. | 1.0 | Fix cos_kernel_api races - Any updates to two memory locations in a single cos_kernel_api operation must take great pains to handle races. For example, if we need to increase both the `vasrange_frontier` and the `vasfrontier`, we can likely use a `faa` for the first, and a `cas` to the second, taking into account concurrent changes. The issue is that another thread might call API functions that can return errors that must also understand this race.
TODO: fix all of these instances. | infrastructure | fix cos kernel api races any updates to two memory locations in a single cos kernel api operation must take great pains to handle races for example if we need to increase both the vasrange frontier and the vasfrontier we can likely use a faa for the first and a cas to the second taking into account concurrent changes the issue is that another thread might call api functions that can return errors that must also understand this race todo fix all of these instances | 1 |
76,405 | 26,412,325,840 | IssuesEvent | 2023-01-13 13:17:53 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Element shows "Decryption key withheld" when it wasn't | T-Defect S-Minor O-Occasional | When Element Android (and other clients based on matrix-android-sdk2) gets an `m.room_key_request` for a key that it doesn't have, it sends an `m.room_key.withheld` message with a code of `m.unavailable` (see [implementation](https://github.com/vector-im/element-android/blob/75de805417ffea6cd2b1647e098d1d32f8e3f17b/matrix-sdk-android/src/main/java/org/matrix/android/sdk/internal/crypto/IncomingKeyRequestManager.kt#L403)).
Having received the `m.unavailable` response, Element-web shows "Decryption key withheld", which is misleading at best.
| 1.0 | Element shows "Decryption key withheld" when it wasn't - When Element Android (and other clients based on matrix-android-sdk2) gets an `m.room_key_request` for a key that it doesn't have, it sends an `m.room_key.withheld` message with a code of `m.unavailable` (see [implementation](https://github.com/vector-im/element-android/blob/75de805417ffea6cd2b1647e098d1d32f8e3f17b/matrix-sdk-android/src/main/java/org/matrix/android/sdk/internal/crypto/IncomingKeyRequestManager.kt#L403)).
Having received the `m.unavailable` response, Element-web shows "Decryption key withheld", which is misleading at best.
| non_infrastructure | element shows decryption key withheld when it wasn t when element android and other clients based on matrix android gets an m room key request for a key that it doesn t have it sends an m room key withheld message with a code of m unavailable see having received the m unavailable response element web shows decryption key withheld which is misleading at best | 0 |
400,668 | 27,294,822,532 | IssuesEvent | 2023-02-23 19:24:06 | npm/cli | https://api.github.com/repos/npm/cli | opened | [DOCS] Archived `npm/npm` repo still points to `npm.community`; redirect has bad cert | Needs Triage Documentation | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### This is a CLI Docs Problem, not another kind of Docs Problem.
- [X] This is a CLI Docs Problem.
### Description of Problem
The [archived npm repo](https://github.com/npm/npm) prominently links to http://npm.community/ in several places. This domain now redirects to `npmjs.com` -- reasonable people can disagree over how helpful this is, I guess, but that's not my point here. The problem is that the redirect involves going through a connection to `npm.community` which is signed with a wildcard cert that is only good for `*.github.com`.
### Potential Solution
If you're going to leave a redirect page there, it should be signed with a cert that is valid for the domain that the user is visiting.
(It would be great if the redirect went somewhere more useful for getting actual NPM support, like this repository, or a direct link to the `npm/feedback` repo, rather than the package repository homepage, but that's neither here nor there.)
### Affected URL
http://npm.community/ | 1.0 | [DOCS] Archived `npm/npm` repo still points to `npm.community`; redirect has bad cert - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### This is a CLI Docs Problem, not another kind of Docs Problem.
- [X] This is a CLI Docs Problem.
### Description of Problem
The [archived npm repo](https://github.com/npm/npm) prominently links to http://npm.community/ in several places. This domain now redirects to `npmjs.com` -- reasonable people can disagree over how helpful this is, I guess, but that's not my point here. The problem is that the redirect involves going through a connection to `npm.community` which is signed with a wildcard cert that is only good for `*.github.com`.
### Potential Solution
If you're going to leave a redirect page there, it should be signed with a cert that is valid for the domain that the user is visiting.
(It would be great if the redirect went somewhere more useful for getting actual NPM support, like this repository, or a direct link to the `npm/feedback` repo, rather than the package repository homepage, but that's neither here nor there.)
### Affected URL
http://npm.community/ | non_infrastructure | archived npm npm repo still points to npm community redirect has bad cert is there an existing issue for this i have searched the existing issues this is a cli docs problem not another kind of docs problem this is a cli docs problem description of problem the prominently links to in several places this domain now redirects to npmjs com reasonable people can disagree over how helpful this is i guess but that s not my point here the problem is that the redirect involves going through a connection to npm community which is signed with a wildcard cert that is only good for github com potential solution if you re going to leave a redirect page there it should be signed with a cert that is valid for the domain that the user is visiting it would be great if the redirect went somewhere more useful for getting actual npm support like this repository or a direct link to the npm feedback repo rather than the package repository homepage but that s neither here nor there affected url | 0 |
32,368 | 26,652,575,625 | IssuesEvent | 2023-01-25 14:42:01 | arduino/mdns-discovery | https://api.github.com/repos/arduino/mdns-discovery | closed | Add macos amd64 architecture | type: enhancement topic: infrastructure | ### Describe the request
Add support for macos arm64 architecture. Should be a matter of updating the release workflow with the latest modifications from https://github.com/arduino/tooling-project-assets
### Describe the current behavior
https://github.com/arduino/arduino-cli/issues/1990
### mdns-discovery version
1.0.6
### Operating system
macOS
### Operating system version
Ventura 13.0.1 (22A400)
### Additional context
_No response_
### Issue checklist
- [X] I searched for previous requests in [the issue tracker](https://github.com/arduino/mdns-discovery/issues?q=)
- [X] I verified the feature was still missing when using the latest version
- [X] My request contains all necessary details | 1.0 | Add macos amd64 architecture - ### Describe the request
Add support for macos arm64 architecture. Should be a matter of updating the release workflow with the latest modifications from https://github.com/arduino/tooling-project-assets
### Describe the current behavior
https://github.com/arduino/arduino-cli/issues/1990
### mdns-discovery version
1.0.6
### Operating system
macOS
### Operating system version
Ventura 13.0.1 (22A400)
### Additional context
_No response_
### Issue checklist
- [X] I searched for previous requests in [the issue tracker](https://github.com/arduino/mdns-discovery/issues?q=)
- [X] I verified the feature was still missing when using the latest version
- [X] My request contains all necessary details | infrastructure | add macos architecture describe the request add support for macos architecture should be a matter of updating the release workflow with the latest modifications from describe the current behavior mdns discovery version operating system macos operating system version ventura additional context no response issue checklist i searched for previous requests in i verified the feature was still missing when using the latest version my request contains all necessary details | 1 |
27,905 | 22,589,432,243 | IssuesEvent | 2022-06-28 18:17:06 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | closed | Block `build`, `lint`, and `test` Scripts Without `turbo` | type: enhancement tool: monorepo infrastructure | <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. -->
**Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
Right now it's possible to run `pnpm build` for a plugin, package, or tool. This has a negative impact, however, since it bypasses the `turbo` cache and may leave dependencies unbuilt. We should look into detecting `turbo` execution and do nothing if the commands are not ran using it. The first thing that comes to mind would be looking for an environment variable and seeing if one is set. If not, perhaps we need to upstream a change to Turborepo to add one. | 1.0 | Block `build`, `lint`, and `test` Scripts Without `turbo` - <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. -->
**Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
Right now it's possible to run `pnpm build` for a plugin, package, or tool. This has a negative impact, however, since it bypasses the `turbo` cache and may leave dependencies unbuilt. We should look into detecting `turbo` execution and do nothing if the commands are not ran using it. The first thing that comes to mind would be looking for an environment variable and seeing if one is set. If not, perhaps we need to upstream a change to Turborepo to add one. | infrastructure | block build lint and test scripts without turbo prerequisites mark completed items with an i have checked that my issue type is not listed here my issue is not a security issue support request bug report enhancement or feature request please use the link above if it is issue description right now it s possible to run pnpm build for a plugin package or tool this has a negative impact however since it bypasses the turbo cache and may leave dependencies unbuilt we should look into detecting turbo execution and do nothing if the commands are not ran using it the first thing that comes to mind would be looking for an environment variable and seeing if one is set if not perhaps we need to upstream a change to turborepo to add one | 1 |
31,502 | 7,374,140,068 | IssuesEvent | 2018-03-13 19:22:02 | WikiEducationFoundation/WikiEduDashboard | https://api.github.com/repos/WikiEducationFoundation/WikiEduDashboard | closed | Get rid of `InputMixin` | code quality | We use this mixin for most of the input fields in the app, mainly for handling validations in an extensible way. However, this one in particular seems to be the source of a lot of warnings that will need to be fixed for the React 16 upgrade, and this is the only mixin that isn't just a store subscription. | 1.0 | Get rid of `InputMixin` - We use this mixin for most of the input fields in the app, mainly for handling validations in an extensible way. However, this one in particular seems to be the source of a lot of warnings that will need to be fixed for the React 16 upgrade, and this is the only mixin that isn't just a store subscription. | non_infrastructure | get rid of inputmixin we use this mixin for most of the input fields in the app mainly for handling validations in an extensible way however this one in particular seems to be the source of a lot of warnings that will need to be fixed for the react upgrade and this is the only mixin that isn t just a store subscription | 0 |
96,853 | 28,031,001,955 | IssuesEvent | 2023-03-28 12:23:35 | google/mediapipe | https://api.github.com/repos/google/mediapipe | opened | Issues building _framework_bindings.so for android arm64 | type:build/install | ### OS Platform and Distribution
Linux Mint 5.15.0-56-generic
### Compiler version
_No response_
### Programming Language and version
C++
### Installed using virtualenv? pip? Conda?(if python)
_No response_
### MediaPipe version
0.9.2.1
### Bazel version
5.2.0
### XCode and Tulsi versions (if iOS)
_No response_
### Android SDK and NDK versions (if android)
ndk 21, sdk, unsure
### Android AAR (if android)
None
### OpenCV version (if running on desktop)
_No response_
### Describe the problem
```shell
When trying to build the `_framework_bindings.so` using `bazel build -c opt --config=android_arm64 mediapipe/python:_framework_bindings.so`, however it errors out, specifically in the `mediapipe/python/pybind:packet` part.
It shows
ERROR: /home/brandon/mediapipe/mediapipe/python/pybind/BUILD:96:15: Compiling mediapipe/python/pybind/packet.cc failed: (Exit 1): clang failed: error executing command external/androidndk/ndk/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -gcc-toolchain external/androidndk/ndk/toolchains/aarch64-linux-android-4.9/prebuilt/linux-x86_64 -target ... (remaining 82 arguments skipped)
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
In file included from mediapipe/python/pybind/packet.cc:15:
In file included from ./mediapipe/python/pybind/packet.h:18:
In file included from external/pybind11/include/pybind11/pybind11.h:13:
In file included from external/pybind11/include/pybind11/detail/class.h:12:
In file included from external/pybind11/include/pybind11/detail/../attr.h:13:
In file included from external/pybind11/include/pybind11/detail/common.h:208:
In file included from bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/Python.h:8:
bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/pyconfig.h:9:12: fatal error: 'aarch64-linux-gnu/python3.10/pyconfig.h' file not found
# include <aarch64-linux-gnu/python3.10/pyconfig.h>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
Target //mediapipe/python/pybind:packet failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 1.623s, Critical Path: 1.36s
INFO: 3 processes: 3 internal.
FAILED: Build did NOT complete successfully
```
I already have the pyconfig.h and all other python headers in the `/usr/lib/aarch64-linux-gnu/python3.10/`.
```
### Complete Logs
```shell
INFO: Analyzed target //mediapipe/python/pybind:packet (0 packages loaded, 0 targets configured).
INFO: Found 1 target...
ERROR: /home/brandon/mediapipe/mediapipe/python/pybind/BUILD:96:15: Compiling mediapipe/python/pybind/packet.cc failed: (Exit 1): clang failed: error executing command
(cd /home/brandon/.cache/bazel/_bazel_brandon/97261b983e2c54d2e8d662dd735e7089/sandbox/linux-sandbox/3101/execroot/mediapipe && \
exec env - \
PATH=/home/brandon/.cache/bazelisk/downloads/bazelbuild/bazel-5.2.0-linux-x86_64/bin:/home/brandon/.local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin \
PWD=/proc/self/cwd \
external/androidndk/ndk/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -gcc-toolchain external/androidndk/ndk/toolchains/aarch64-linux-android-4.9/prebuilt/linux-x86_64 -target aarch64-none-linux-android -fpic -isystemexternal/androidndk/ndk/sysroot/usr/include/aarch64-linux-android '-D__ANDROID_API__=21' -no-canonical-prefixes -Wno-invalid-command-line-argument -Wno-unused-command-line-argument -funwind-tables -fstack-protector-strong -fno-addrsig '-Werror=return-type' '-Werror=int-to-pointer-cast' '-Werror=pointer-to-int-cast' '-Werror=implicit-function-declaration' -O2 -g -DNDEBUG -MD -MF bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.d '-frandom-seed=bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.o' -iquote . -iquote bazel-out/arm64-v8a-opt/bin -iquote external/com_google_protobuf -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf -iquote external/zlib -iquote bazel-out/arm64-v8a-opt/bin/external/zlib -iquote external/com_google_absl -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_absl -iquote external/com_github_glog_glog_no_gflags -iquote bazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags -iquote external/pybind11 -iquote bazel-out/arm64-v8a-opt/bin/external/pybind11 -iquote external/local_config_python -iquote bazel-out/arm64-v8a-opt/bin/external/local_config_python -Ibazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags/_virtual_includes/default_glog_headers -isystem external/com_google_protobuf/src -isystem bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf/src -isystem external/zlib -isystem bazel-out/arm64-v8a-opt/bin/external/zlib -isystem external/pybind11/include -isystem bazel-out/arm64-v8a-opt/bin/external/pybind11/include -isystem external/local_config_python/python_include -isystem bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include -w '-std=c++17' -fexceptions '--sysroot=external/androidndk/ndk/platforms/android-21/arch-arm64' -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++/include -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++abi/include -isystem external/androidndk/ndk/sources/android/support/include -isystemexternal/androidndk/ndk/sysroot/usr/include -c mediapipe/python/pybind/packet.cc -o bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.o)
# Configuration: bc113378d8420dfd857078289773f4fbb5cd04e45284558a4ecd0964ef08a12d
# Execution platform: @local_execution_config_platform//:platform
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
In file included from mediapipe/python/pybind/packet.cc:15:
In file included from ./mediapipe/python/pybind/packet.h:18:
In file included from external/pybind11/include/pybind11/pybind11.h:13:
In file included from external/pybind11/include/pybind11/detail/class.h:12:
In file included from external/pybind11/include/pybind11/detail/../attr.h:13:
In file included from external/pybind11/include/pybind11/detail/common.h:208:
In file included from bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/Python.h:8:
bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/pyconfig.h:9:12: fatal error: 'aarch64-linux-gnu/python3.10/pyconfig.h' file not found
# include <aarch64-linux-gnu/python3.10/pyconfig.h>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
ERROR: /home/brandon/mediapipe/mediapipe/python/pybind/BUILD:96:15: Compiling mediapipe/python/pybind/packet.cc failed: (Exit 1): clang failed: error executing command
(cd /home/brandon/.cache/bazel/_bazel_brandon/97261b983e2c54d2e8d662dd735e7089/sandbox/linux-sandbox/3100/execroot/mediapipe && \
exec env - \
PATH=/home/brandon/.cache/bazelisk/downloads/bazelbuild/bazel-5.2.0-linux-x86_64/bin:/home/brandon/.local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin \
PWD=/proc/self/cwd \
external/androidndk/ndk/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -gcc-toolchain external/androidndk/ndk/toolchains/aarch64-linux-android-4.9/prebuilt/linux-x86_64 -target aarch64-none-linux-android -fpic -isystemexternal/androidndk/ndk/sysroot/usr/include/aarch64-linux-android '-D__ANDROID_API__=21' -no-canonical-prefixes -Wno-invalid-command-line-argument -Wno-unused-command-line-argument -funwind-tables -fstack-protector-strong -fno-addrsig '-Werror=return-type' '-Werror=int-to-pointer-cast' '-Werror=pointer-to-int-cast' '-Werror=implicit-function-declaration' -O2 -g -DNDEBUG -MD -MF bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.pic.d '-frandom-seed=bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.pic.o' -fPIC -iquote . -iquote bazel-out/arm64-v8a-opt/bin -iquote external/com_google_protobuf -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf -iquote external/zlib -iquote bazel-out/arm64-v8a-opt/bin/external/zlib -iquote external/com_google_absl -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_absl -iquote external/com_github_glog_glog_no_gflags -iquote bazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags -iquote external/pybind11 -iquote bazel-out/arm64-v8a-opt/bin/external/pybind11 -iquote external/local_config_python -iquote bazel-out/arm64-v8a-opt/bin/external/local_config_python -Ibazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags/_virtual_includes/default_glog_headers -isystem external/com_google_protobuf/src -isystem bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf/src -isystem external/zlib -isystem bazel-out/arm64-v8a-opt/bin/external/zlib -isystem external/pybind11/include -isystem bazel-out/arm64-v8a-opt/bin/external/pybind11/include -isystem external/local_config_python/python_include -isystem bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include -w '-std=c++17' -fexceptions '--sysroot=external/androidndk/ndk/platforms/android-21/arch-arm64' -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++/include -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++abi/include -isystem external/androidndk/ndk/sources/android/support/include -isystemexternal/androidndk/ndk/sysroot/usr/include -c mediapipe/python/pybind/packet.cc -o bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.pic.o)
# Configuration: bc113378d8420dfd857078289773f4fbb5cd04e45284558a4ecd0964ef08a12d
# Execution platform: @local_execution_config_platform//:platform
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
In file included from mediapipe/python/pybind/packet.cc:15:
In file included from ./mediapipe/python/pybind/packet.h:18:
In file included from external/pybind11/include/pybind11/pybind11.h:13:
In file included from external/pybind11/include/pybind11/detail/class.h:12:
In file included from external/pybind11/include/pybind11/detail/../attr.h:13:
In file included from external/pybind11/include/pybind11/detail/common.h:208:
In file included from bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/Python.h:8:
bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/pyconfig.h:9:12: fatal error: 'aarch64-linux-gnu/python3.10/pyconfig.h' file not found
# include <aarch64-linux-gnu/python3.10/pyconfig.h>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
Target //mediapipe/python/pybind:packet failed to build
INFO: Elapsed time: 1.429s, Critical Path: 1.19s
INFO: 3 processes: 3 internal.
FAILED: Build did NOT complete successfully
```
```
| 1.0 | Issues building _framework_bindings.so for android arm64 - ### OS Platform and Distribution
Linux Mint 5.15.0-56-generic
### Compiler version
_No response_
### Programming Language and version
C++
### Installed using virtualenv? pip? Conda?(if python)
_No response_
### MediaPipe version
0.9.2.1
### Bazel version
5.2.0
### XCode and Tulsi versions (if iOS)
_No response_
### Android SDK and NDK versions (if android)
ndk 21, sdk, unsure
### Android AAR (if android)
None
### OpenCV version (if running on desktop)
_No response_
### Describe the problem
```shell
When trying to build the `_framework_bindings.so` using `bazel build -c opt --config=android_arm64 mediapipe/python:_framework_bindings.so`, however it errors out, specifically in the `mediapipe/python/pybind:packet` part.
It shows
ERROR: /home/brandon/mediapipe/mediapipe/python/pybind/BUILD:96:15: Compiling mediapipe/python/pybind/packet.cc failed: (Exit 1): clang failed: error executing command external/androidndk/ndk/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -gcc-toolchain external/androidndk/ndk/toolchains/aarch64-linux-android-4.9/prebuilt/linux-x86_64 -target ... (remaining 82 arguments skipped)
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
In file included from mediapipe/python/pybind/packet.cc:15:
In file included from ./mediapipe/python/pybind/packet.h:18:
In file included from external/pybind11/include/pybind11/pybind11.h:13:
In file included from external/pybind11/include/pybind11/detail/class.h:12:
In file included from external/pybind11/include/pybind11/detail/../attr.h:13:
In file included from external/pybind11/include/pybind11/detail/common.h:208:
In file included from bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/Python.h:8:
bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/pyconfig.h:9:12: fatal error: 'aarch64-linux-gnu/python3.10/pyconfig.h' file not found
# include <aarch64-linux-gnu/python3.10/pyconfig.h>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
Target //mediapipe/python/pybind:packet failed to build
Use --verbose_failures to see the command lines of failed build steps.
INFO: Elapsed time: 1.623s, Critical Path: 1.36s
INFO: 3 processes: 3 internal.
FAILED: Build did NOT complete successfully
```
I already have the pyconfig.h and all other python headers in the `/usr/lib/aarch64-linux-gnu/python3.10/`.
```
### Complete Logs
```shell
INFO: Analyzed target //mediapipe/python/pybind:packet (0 packages loaded, 0 targets configured).
INFO: Found 1 target...
ERROR: /home/brandon/mediapipe/mediapipe/python/pybind/BUILD:96:15: Compiling mediapipe/python/pybind/packet.cc failed: (Exit 1): clang failed: error executing command
(cd /home/brandon/.cache/bazel/_bazel_brandon/97261b983e2c54d2e8d662dd735e7089/sandbox/linux-sandbox/3101/execroot/mediapipe && \
exec env - \
PATH=/home/brandon/.cache/bazelisk/downloads/bazelbuild/bazel-5.2.0-linux-x86_64/bin:/home/brandon/.local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin \
PWD=/proc/self/cwd \
external/androidndk/ndk/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -gcc-toolchain external/androidndk/ndk/toolchains/aarch64-linux-android-4.9/prebuilt/linux-x86_64 -target aarch64-none-linux-android -fpic -isystemexternal/androidndk/ndk/sysroot/usr/include/aarch64-linux-android '-D__ANDROID_API__=21' -no-canonical-prefixes -Wno-invalid-command-line-argument -Wno-unused-command-line-argument -funwind-tables -fstack-protector-strong -fno-addrsig '-Werror=return-type' '-Werror=int-to-pointer-cast' '-Werror=pointer-to-int-cast' '-Werror=implicit-function-declaration' -O2 -g -DNDEBUG -MD -MF bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.d '-frandom-seed=bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.o' -iquote . -iquote bazel-out/arm64-v8a-opt/bin -iquote external/com_google_protobuf -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf -iquote external/zlib -iquote bazel-out/arm64-v8a-opt/bin/external/zlib -iquote external/com_google_absl -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_absl -iquote external/com_github_glog_glog_no_gflags -iquote bazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags -iquote external/pybind11 -iquote bazel-out/arm64-v8a-opt/bin/external/pybind11 -iquote external/local_config_python -iquote bazel-out/arm64-v8a-opt/bin/external/local_config_python -Ibazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags/_virtual_includes/default_glog_headers -isystem external/com_google_protobuf/src -isystem bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf/src -isystem external/zlib -isystem bazel-out/arm64-v8a-opt/bin/external/zlib -isystem external/pybind11/include -isystem bazel-out/arm64-v8a-opt/bin/external/pybind11/include -isystem external/local_config_python/python_include -isystem bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include -w '-std=c++17' -fexceptions '--sysroot=external/androidndk/ndk/platforms/android-21/arch-arm64' -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++/include -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++abi/include -isystem external/androidndk/ndk/sources/android/support/include -isystemexternal/androidndk/ndk/sysroot/usr/include -c mediapipe/python/pybind/packet.cc -o bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.o)
# Configuration: bc113378d8420dfd857078289773f4fbb5cd04e45284558a4ecd0964ef08a12d
# Execution platform: @local_execution_config_platform//:platform
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
In file included from mediapipe/python/pybind/packet.cc:15:
In file included from ./mediapipe/python/pybind/packet.h:18:
In file included from external/pybind11/include/pybind11/pybind11.h:13:
In file included from external/pybind11/include/pybind11/detail/class.h:12:
In file included from external/pybind11/include/pybind11/detail/../attr.h:13:
In file included from external/pybind11/include/pybind11/detail/common.h:208:
In file included from bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/Python.h:8:
bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/pyconfig.h:9:12: fatal error: 'aarch64-linux-gnu/python3.10/pyconfig.h' file not found
# include <aarch64-linux-gnu/python3.10/pyconfig.h>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
ERROR: /home/brandon/mediapipe/mediapipe/python/pybind/BUILD:96:15: Compiling mediapipe/python/pybind/packet.cc failed: (Exit 1): clang failed: error executing command
(cd /home/brandon/.cache/bazel/_bazel_brandon/97261b983e2c54d2e8d662dd735e7089/sandbox/linux-sandbox/3100/execroot/mediapipe && \
exec env - \
PATH=/home/brandon/.cache/bazelisk/downloads/bazelbuild/bazel-5.2.0-linux-x86_64/bin:/home/brandon/.local/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin \
PWD=/proc/self/cwd \
external/androidndk/ndk/toolchains/llvm/prebuilt/linux-x86_64/bin/clang -gcc-toolchain external/androidndk/ndk/toolchains/aarch64-linux-android-4.9/prebuilt/linux-x86_64 -target aarch64-none-linux-android -fpic -isystemexternal/androidndk/ndk/sysroot/usr/include/aarch64-linux-android '-D__ANDROID_API__=21' -no-canonical-prefixes -Wno-invalid-command-line-argument -Wno-unused-command-line-argument -funwind-tables -fstack-protector-strong -fno-addrsig '-Werror=return-type' '-Werror=int-to-pointer-cast' '-Werror=pointer-to-int-cast' '-Werror=implicit-function-declaration' -O2 -g -DNDEBUG -MD -MF bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.pic.d '-frandom-seed=bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.pic.o' -fPIC -iquote . -iquote bazel-out/arm64-v8a-opt/bin -iquote external/com_google_protobuf -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf -iquote external/zlib -iquote bazel-out/arm64-v8a-opt/bin/external/zlib -iquote external/com_google_absl -iquote bazel-out/arm64-v8a-opt/bin/external/com_google_absl -iquote external/com_github_glog_glog_no_gflags -iquote bazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags -iquote external/pybind11 -iquote bazel-out/arm64-v8a-opt/bin/external/pybind11 -iquote external/local_config_python -iquote bazel-out/arm64-v8a-opt/bin/external/local_config_python -Ibazel-out/arm64-v8a-opt/bin/external/com_github_glog_glog_no_gflags/_virtual_includes/default_glog_headers -isystem external/com_google_protobuf/src -isystem bazel-out/arm64-v8a-opt/bin/external/com_google_protobuf/src -isystem external/zlib -isystem bazel-out/arm64-v8a-opt/bin/external/zlib -isystem external/pybind11/include -isystem bazel-out/arm64-v8a-opt/bin/external/pybind11/include -isystem external/local_config_python/python_include -isystem bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include -w '-std=c++17' -fexceptions '--sysroot=external/androidndk/ndk/platforms/android-21/arch-arm64' -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++/include -isystem external/androidndk/ndk/sources/cxx-stl/llvm-libc++abi/include -isystem external/androidndk/ndk/sources/android/support/include -isystemexternal/androidndk/ndk/sysroot/usr/include -c mediapipe/python/pybind/packet.cc -o bazel-out/arm64-v8a-opt/bin/mediapipe/python/pybind/_objs/packet/packet.pic.o)
# Configuration: bc113378d8420dfd857078289773f4fbb5cd04e45284558a4ecd0964ef08a12d
# Execution platform: @local_execution_config_platform//:platform
Use --sandbox_debug to see verbose messages from the sandbox and retain the sandbox build root for debugging
In file included from mediapipe/python/pybind/packet.cc:15:
In file included from ./mediapipe/python/pybind/packet.h:18:
In file included from external/pybind11/include/pybind11/pybind11.h:13:
In file included from external/pybind11/include/pybind11/detail/class.h:12:
In file included from external/pybind11/include/pybind11/detail/../attr.h:13:
In file included from external/pybind11/include/pybind11/detail/common.h:208:
In file included from bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/Python.h:8:
bazel-out/arm64-v8a-opt/bin/external/local_config_python/python_include/pyconfig.h:9:12: fatal error: 'aarch64-linux-gnu/python3.10/pyconfig.h' file not found
# include <aarch64-linux-gnu/python3.10/pyconfig.h>
^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
1 error generated.
Target //mediapipe/python/pybind:packet failed to build
INFO: Elapsed time: 1.429s, Critical Path: 1.19s
INFO: 3 processes: 3 internal.
FAILED: Build did NOT complete successfully
```
```
| non_infrastructure | issues building framework bindings so for android os platform and distribution linux mint generic compiler version no response programming language and version c installed using virtualenv pip conda if python no response mediapipe version bazel version xcode and tulsi versions if ios no response android sdk and ndk versions if android ndk sdk unsure android aar if android none opencv version if running on desktop no response describe the problem shell when trying to build the framework bindings so using bazel build c opt config android mediapipe python framework bindings so however it errors out specifically in the mediapipe python pybind packet part it shows error home brandon mediapipe mediapipe python pybind build compiling mediapipe python pybind packet cc failed exit clang failed error executing command external androidndk ndk toolchains llvm prebuilt linux bin clang gcc toolchain external androidndk ndk toolchains linux android prebuilt linux target remaining arguments skipped use sandbox debug to see verbose messages from the sandbox and retain the sandbox build root for debugging in file included from mediapipe python pybind packet cc in file included from mediapipe python pybind packet h in file included from external include h in file included from external include detail class h in file included from external include detail attr h in file included from external include detail common h in file included from bazel out opt bin external local config python python include python h bazel out opt bin external local config python python include pyconfig h fatal error linux gnu pyconfig h file not found include error generated target mediapipe python pybind packet failed to build use verbose failures to see the command lines of failed build steps info elapsed time critical path info processes internal failed build did not complete successfully i already have the pyconfig h and all other python headers in the usr lib linux gnu complete logs shell info analyzed target mediapipe python pybind packet packages loaded targets configured info found target error home brandon mediapipe mediapipe python pybind build compiling mediapipe python pybind packet cc failed exit clang failed error executing command cd home brandon cache bazel bazel brandon sandbox linux sandbox execroot mediapipe exec env path home brandon cache bazelisk downloads bazelbuild bazel linux bin home brandon local bin usr local sbin usr local bin usr sbin usr bin sbin bin usr games usr local games snap bin pwd proc self cwd external androidndk ndk toolchains llvm prebuilt linux bin clang gcc toolchain external androidndk ndk toolchains linux android prebuilt linux target none linux android fpic isystemexternal androidndk ndk sysroot usr include linux android d android api no canonical prefixes wno invalid command line argument wno unused command line argument funwind tables fstack protector strong fno addrsig werror return type werror int to pointer cast werror pointer to int cast werror implicit function declaration g dndebug md mf bazel out opt bin mediapipe python pybind objs packet packet d frandom seed bazel out opt bin mediapipe python pybind objs packet packet o iquote iquote bazel out opt bin iquote external com google protobuf iquote bazel out opt bin external com google protobuf iquote external zlib iquote bazel out opt bin external zlib iquote external com google absl iquote bazel out opt bin external com google absl iquote external com github glog glog no gflags iquote bazel out opt bin external com github glog glog no gflags iquote external iquote bazel out opt bin external iquote external local config python iquote bazel out opt bin external local config python ibazel out opt bin external com github glog glog no gflags virtual includes default glog headers isystem external com google protobuf src isystem bazel out opt bin external com google protobuf src isystem external zlib isystem bazel out opt bin external zlib isystem external include isystem bazel out opt bin external include isystem external local config python python include isystem bazel out opt bin external local config python python include w std c fexceptions sysroot external androidndk ndk platforms android arch isystem external androidndk ndk sources cxx stl llvm libc include isystem external androidndk ndk sources cxx stl llvm libc abi include isystem external androidndk ndk sources android support include isystemexternal androidndk ndk sysroot usr include c mediapipe python pybind packet cc o bazel out opt bin mediapipe python pybind objs packet packet o configuration execution platform local execution config platform platform use sandbox debug to see verbose messages from the sandbox and retain the sandbox build root for debugging in file included from mediapipe python pybind packet cc in file included from mediapipe python pybind packet h in file included from external include h in file included from external include detail class h in file included from external include detail attr h in file included from external include detail common h in file included from bazel out opt bin external local config python python include python h bazel out opt bin external local config python python include pyconfig h fatal error linux gnu pyconfig h file not found include error generated error home brandon mediapipe mediapipe python pybind build compiling mediapipe python pybind packet cc failed exit clang failed error executing command cd home brandon cache bazel bazel brandon sandbox linux sandbox execroot mediapipe exec env path home brandon cache bazelisk downloads bazelbuild bazel linux bin home brandon local bin usr local sbin usr local bin usr sbin usr bin sbin bin usr games usr local games snap bin pwd proc self cwd external androidndk ndk toolchains llvm prebuilt linux bin clang gcc toolchain external androidndk ndk toolchains linux android prebuilt linux target none linux android fpic isystemexternal androidndk ndk sysroot usr include linux android d android api no canonical prefixes wno invalid command line argument wno unused command line argument funwind tables fstack protector strong fno addrsig werror return type werror int to pointer cast werror pointer to int cast werror implicit function declaration g dndebug md mf bazel out opt bin mediapipe python pybind objs packet packet pic d frandom seed bazel out opt bin mediapipe python pybind objs packet packet pic o fpic iquote iquote bazel out opt bin iquote external com google protobuf iquote bazel out opt bin external com google protobuf iquote external zlib iquote bazel out opt bin external zlib iquote external com google absl iquote bazel out opt bin external com google absl iquote external com github glog glog no gflags iquote bazel out opt bin external com github glog glog no gflags iquote external iquote bazel out opt bin external iquote external local config python iquote bazel out opt bin external local config python ibazel out opt bin external com github glog glog no gflags virtual includes default glog headers isystem external com google protobuf src isystem bazel out opt bin external com google protobuf src isystem external zlib isystem bazel out opt bin external zlib isystem external include isystem bazel out opt bin external include isystem external local config python python include isystem bazel out opt bin external local config python python include w std c fexceptions sysroot external androidndk ndk platforms android arch isystem external androidndk ndk sources cxx stl llvm libc include isystem external androidndk ndk sources cxx stl llvm libc abi include isystem external androidndk ndk sources android support include isystemexternal androidndk ndk sysroot usr include c mediapipe python pybind packet cc o bazel out opt bin mediapipe python pybind objs packet packet pic o configuration execution platform local execution config platform platform use sandbox debug to see verbose messages from the sandbox and retain the sandbox build root for debugging in file included from mediapipe python pybind packet cc in file included from mediapipe python pybind packet h in file included from external include h in file included from external include detail class h in file included from external include detail attr h in file included from external include detail common h in file included from bazel out opt bin external local config python python include python h bazel out opt bin external local config python python include pyconfig h fatal error linux gnu pyconfig h file not found include error generated target mediapipe python pybind packet failed to build info elapsed time critical path info processes internal failed build did not complete successfully | 0 |
272,511 | 8,514,309,670 | IssuesEvent | 2018-10-31 18:12:28 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | closed | Sticky ads hidden in Google AMP viewer on iOS Chrome | Category: Ads Internally Tracked P1: High Priority Type: Bug | ## What's the Issue?
Within the Google viewer on iOS Chrome the bottom toolbar of the new UI doesn’t slide out of the way as it does outside of the viewer. It’s also not pushing fixed position elements up either so the amp-sticky-ad elements are behind the toolbar UI.
## How do we reproduce the issue?
I’m currently seeing this on every page that implements the sticky ad element when viewed through the Google AMP viewer when accessed from the top stories carousel.
## What browsers are affected?
iOS Chrome, latest version
## Which AMP version is affected?
It seems to be an interaction issue between Chrome and the AMP viewer rather than AMP specifically.
| 1.0 | Sticky ads hidden in Google AMP viewer on iOS Chrome - ## What's the Issue?
Within the Google viewer on iOS Chrome the bottom toolbar of the new UI doesn’t slide out of the way as it does outside of the viewer. It’s also not pushing fixed position elements up either so the amp-sticky-ad elements are behind the toolbar UI.
## How do we reproduce the issue?
I’m currently seeing this on every page that implements the sticky ad element when viewed through the Google AMP viewer when accessed from the top stories carousel.
## What browsers are affected?
iOS Chrome, latest version
## Which AMP version is affected?
It seems to be an interaction issue between Chrome and the AMP viewer rather than AMP specifically.
| non_infrastructure | sticky ads hidden in google amp viewer on ios chrome what s the issue within the google viewer on ios chrome the bottom toolbar of the new ui doesn’t slide out of the way as it does outside of the viewer it’s also not pushing fixed position elements up either so the amp sticky ad elements are behind the toolbar ui how do we reproduce the issue i’m currently seeing this on every page that implements the sticky ad element when viewed through the google amp viewer when accessed from the top stories carousel what browsers are affected ios chrome latest version which amp version is affected it seems to be an interaction issue between chrome and the amp viewer rather than amp specifically | 0 |
294,942 | 22,171,359,080 | IssuesEvent | 2022-06-06 01:17:57 | junu0516/airbnb | https://api.github.com/repos/junu0516/airbnb | closed | 금요일 PR 날리기 | 📂 documentation ⚡️ iOS | ### 리뷰어님께
안녕하세요! 이번작업은 실서버를 붙여서 화면에 보여주기 위해 UI 를 우선적으로 작업했습니다.
다음주에 검색 조건 화면 UI(캘린더, 가격 그래프)를 집중적으로 구현할 것 같습니다.
잘부탁드립니다 😊
### 작업
- [x] 실제로 배포된 서버를 붙여서 데이터가 제대로 오는지 확인
- [x] 검색결과 - 지도에 숙소정보를 보여주기
- 숙소 위치 마커찍기
- 숙소를 하단에 컬렉션뷰 형태로 보여주기
- [x] 숙소 상세화면 UI
- [x] 검색 결과 - 카드리스트 화면 UI
### 실행화면

## 고민
### 화면전환시 데이터를 계속 넘겨줘야하는 문제에 대해
``` A-ViewController --(present)--> B-ViewController --(present)--> C-ViewController```
1. A-ViewController 에서 사용하는 데이터를 화면전환시 B-ViewController 에 넘기고
2. B-ViewController 에서 C-ViewController 로 전환시, 같은 데이터를 계속 넘겨줘야 합니다
A,B,C ViewController 가 같은 데이터을 사용합니다.
API 호출을 줄이기 위해서 이렇게 데이터를 계속 넘기는 형태가 되었습니다.
여기서 문제점은
- 화면전환시 하나의 화면을 만들기 위해 생성해야할 인스턴스 꾸러미들(usecase, service)이 있습니다. 이 생성하는 역할을 ViewController 가 가진다는 문제가 있습니다.
- 결국에는 같은 데이터인데, 이 데이터를 넘기는 코드가 A,B,C-ViewController 에 분산되어 있습니다.
결론적으로 ViewController 의 책임이 많아지는데, 이를 해결하기위해 저희가 도출한 방법이 코디네이터 패턴입니다.
이 패턴을 적용한다면, ViewController 는 코디네이터객체에게 어느화면으로 이동할건지 요청만 하면
코디네이터 객체가 해당 화면을 생성(usecase,service) 을 하고 화면 전환을 담당하게 될겁니다.
저희가 생각한 방향이 문제점을 해결하는데 도움이 될지, 레나님은 다른 아이디어가 있으신지 궁금합니다! | 1.0 | 금요일 PR 날리기 - ### 리뷰어님께
안녕하세요! 이번작업은 실서버를 붙여서 화면에 보여주기 위해 UI 를 우선적으로 작업했습니다.
다음주에 검색 조건 화면 UI(캘린더, 가격 그래프)를 집중적으로 구현할 것 같습니다.
잘부탁드립니다 😊
### 작업
- [x] 실제로 배포된 서버를 붙여서 데이터가 제대로 오는지 확인
- [x] 검색결과 - 지도에 숙소정보를 보여주기
- 숙소 위치 마커찍기
- 숙소를 하단에 컬렉션뷰 형태로 보여주기
- [x] 숙소 상세화면 UI
- [x] 검색 결과 - 카드리스트 화면 UI
### 실행화면

## 고민
### 화면전환시 데이터를 계속 넘겨줘야하는 문제에 대해
``` A-ViewController --(present)--> B-ViewController --(present)--> C-ViewController```
1. A-ViewController 에서 사용하는 데이터를 화면전환시 B-ViewController 에 넘기고
2. B-ViewController 에서 C-ViewController 로 전환시, 같은 데이터를 계속 넘겨줘야 합니다
A,B,C ViewController 가 같은 데이터을 사용합니다.
API 호출을 줄이기 위해서 이렇게 데이터를 계속 넘기는 형태가 되었습니다.
여기서 문제점은
- 화면전환시 하나의 화면을 만들기 위해 생성해야할 인스턴스 꾸러미들(usecase, service)이 있습니다. 이 생성하는 역할을 ViewController 가 가진다는 문제가 있습니다.
- 결국에는 같은 데이터인데, 이 데이터를 넘기는 코드가 A,B,C-ViewController 에 분산되어 있습니다.
결론적으로 ViewController 의 책임이 많아지는데, 이를 해결하기위해 저희가 도출한 방법이 코디네이터 패턴입니다.
이 패턴을 적용한다면, ViewController 는 코디네이터객체에게 어느화면으로 이동할건지 요청만 하면
코디네이터 객체가 해당 화면을 생성(usecase,service) 을 하고 화면 전환을 담당하게 될겁니다.
저희가 생각한 방향이 문제점을 해결하는데 도움이 될지, 레나님은 다른 아이디어가 있으신지 궁금합니다! | non_infrastructure | 금요일 pr 날리기 리뷰어님께 안녕하세요 이번작업은 실서버를 붙여서 화면에 보여주기 위해 ui 를 우선적으로 작업했습니다 다음주에 검색 조건 화면 ui 캘린더 가격 그래프 를 집중적으로 구현할 것 같습니다 잘부탁드립니다 😊 작업 실제로 배포된 서버를 붙여서 데이터가 제대로 오는지 확인 검색결과 지도에 숙소정보를 보여주기 숙소 위치 마커찍기 숙소를 하단에 컬렉션뷰 형태로 보여주기 숙소 상세화면 ui 검색 결과 카드리스트 화면 ui 실행화면 고민 화면전환시 데이터를 계속 넘겨줘야하는 문제에 대해 a viewcontroller present b viewcontroller present c viewcontroller a viewcontroller 에서 사용하는 데이터를 화면전환시 b viewcontroller 에 넘기고 b viewcontroller 에서 c viewcontroller 로 전환시 같은 데이터를 계속 넘겨줘야 합니다 a b c viewcontroller 가 같은 데이터을 사용합니다 api 호출을 줄이기 위해서 이렇게 데이터를 계속 넘기는 형태가 되었습니다 여기서 문제점은 화면전환시 하나의 화면을 만들기 위해 생성해야할 인스턴스 꾸러미들 usecase service 이 있습니다 이 생성하는 역할을 viewcontroller 가 가진다는 문제가 있습니다 결국에는 같은 데이터인데 이 데이터를 넘기는 코드가 a b c viewcontroller 에 분산되어 있습니다 결론적으로 viewcontroller 의 책임이 많아지는데 이를 해결하기위해 저희가 도출한 방법이 코디네이터 패턴입니다 이 패턴을 적용한다면 viewcontroller 는 코디네이터객체에게 어느화면으로 이동할건지 요청만 하면 코디네이터 객체가 해당 화면을 생성 usecase service 을 하고 화면 전환을 담당하게 될겁니다 저희가 생각한 방향이 문제점을 해결하는데 도움이 될지 레나님은 다른 아이디어가 있으신지 궁금합니다 | 0 |
3,249 | 4,166,087,432 | IssuesEvent | 2016-06-19 22:50:25 | freenode/web-7.0 | https://api.github.com/repos/freenode/web-7.0 | reopened | Cloudflare throws incorrect 404 | infrastructure | Can we get cloudflare to use the /404 page we have instead of the cloudflare one ( @christeld ) | 1.0 | Cloudflare throws incorrect 404 - Can we get cloudflare to use the /404 page we have instead of the cloudflare one ( @christeld ) | infrastructure | cloudflare throws incorrect can we get cloudflare to use the page we have instead of the cloudflare one christeld | 1 |
326,756 | 24,099,876,354 | IssuesEvent | 2022-09-19 22:50:39 | Roche/rtables | https://api.github.com/repos/Roche/rtables | closed | Visual problem in tibble rendering from all vignettes | bug priority documentation automation | This problem has been there for a bit, and it is possibly a duplicated issue. Here is the rendering on my chrome 105.0.5195.102 (Official Build) (64-bit):

| 1.0 | Visual problem in tibble rendering from all vignettes - This problem has been there for a bit, and it is possibly a duplicated issue. Here is the rendering on my chrome 105.0.5195.102 (Official Build) (64-bit):

| non_infrastructure | visual problem in tibble rendering from all vignettes this problem has been there for a bit and it is possibly a duplicated issue here is the rendering on my chrome official build bit | 0 |
249,321 | 21,158,290,741 | IssuesEvent | 2022-04-07 06:56:15 | zephyrproject-rtos/test_results | https://api.github.com/repos/zephyrproject-rtos/test_results | opened |
IPv4 ARP Reply from other sender hardware has DUT's Protocol Address. error
| area: Tests |
**Describe the bug**
Reply from other sender hardware has DUT's Protocol Address. test is Fail on Zephyr3.0.0 on qemu_x86
**References**
RFC 5227: section 2.4
**Results**
FAIL: icmp.v4 DUT did not respond to ARP conflict.; FAIL: udp.v4 DUT did not respond to ARP conflict.; FAIL: tcp.v4 DUT did not respond to ARP conflict.
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: Zephyr3.0.0
| 1.0 |
IPv4 ARP Reply from other sender hardware has DUT's Protocol Address. error
-
**Describe the bug**
Reply from other sender hardware has DUT's Protocol Address. test is Fail on Zephyr3.0.0 on qemu_x86
**References**
RFC 5227: section 2.4
**Results**
FAIL: icmp.v4 DUT did not respond to ARP conflict.; FAIL: udp.v4 DUT did not respond to ARP conflict.; FAIL: tcp.v4 DUT did not respond to ARP conflict.
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: Zephyr3.0.0
| non_infrastructure | arp reply from other sender hardware has dut s protocol address error describe the bug reply from other sender hardware has dut s protocol address test is fail on on qemu references rfc section results fail icmp dut did not respond to arp conflict fail udp dut did not respond to arp conflict fail tcp dut did not respond to arp conflict environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used | 0 |
11,010 | 8,870,207,865 | IssuesEvent | 2019-01-11 08:49:01 | Snusbolaget/product | https://api.github.com/repos/Snusbolaget/product | closed | Create a role to attach on lambda functions | infrastructure | Create the role that should be attached on lambda functions, with suitable role policies created and attached | 1.0 | Create a role to attach on lambda functions - Create the role that should be attached on lambda functions, with suitable role policies created and attached | infrastructure | create a role to attach on lambda functions create the role that should be attached on lambda functions with suitable role policies created and attached | 1 |
18,960 | 13,178,160,058 | IssuesEvent | 2020-08-12 08:40:35 | pyinstaller/pyinstaller | https://api.github.com/repos/pyinstaller/pyinstaller | closed | Add check for correct names of new-fragments | area:project infrastructure feature | While releasing the last version, I was hindered by some news-fragments having the wrong filename. While this is not much of an issue, it costs yet another few minutes to solve. But for a possible (half-)automated release pipeline, this would make some entries vanish from the change log.
The check could be part of the linting step.
Valid names are in <https://github.com/pyinstaller/pyinstaller/blob/develop/doc/development/changelog-entries.rst>
draft pattern: `([^.]+)\.(feature|bugfix|…)\.rst` -> validate with towncrier | 1.0 | Add check for correct names of new-fragments - While releasing the last version, I was hindered by some news-fragments having the wrong filename. While this is not much of an issue, it costs yet another few minutes to solve. But for a possible (half-)automated release pipeline, this would make some entries vanish from the change log.
The check could be part of the linting step.
Valid names are in <https://github.com/pyinstaller/pyinstaller/blob/develop/doc/development/changelog-entries.rst>
draft pattern: `([^.]+)\.(feature|bugfix|…)\.rst` -> validate with towncrier | infrastructure | add check for correct names of new fragments while releasing the last version i was hindered by some news fragments having the wrong filename while this is not much of an issue it costs yet another few minutes to solve but for a possible half automated release pipeline this would make some entries vanish from the change log the check could be part of the linting step valid names are in draft pattern feature bugfix … rst validate with towncrier | 1 |
34,354 | 29,513,082,210 | IssuesEvent | 2023-06-04 06:41:16 | KHashimoto3/ApartmentBillingPortal | https://api.github.com/repos/KHashimoto3/ApartmentBillingPortal | closed | Firebase hostingを導入し、デプロイできるようにする | frontend infrastructure | # 内容
* frontディレクトリで、firebaseをinitする
* Firebase hostingを設定する
* 最初のサイトをデプロイする | 1.0 | Firebase hostingを導入し、デプロイできるようにする - # 内容
* frontディレクトリで、firebaseをinitする
* Firebase hostingを設定する
* 最初のサイトをデプロイする | infrastructure | firebase hostingを導入し、デプロイできるようにする 内容 frontディレクトリで、firebaseをinitする firebase hostingを設定する 最初のサイトをデプロイする | 1 |
17,570 | 12,454,117,027 | IssuesEvent | 2020-05-27 14:47:48 | reapit/foundations | https://api.github.com/repos/reapit/foundations | closed | Create repository, pipelines and infrastructure for sales progression service | infrastructure platform-team | - Create new repository for codebase
- Create pipelines for CI/CD
- Spin up new infrastructure | 1.0 | Create repository, pipelines and infrastructure for sales progression service - - Create new repository for codebase
- Create pipelines for CI/CD
- Spin up new infrastructure | infrastructure | create repository pipelines and infrastructure for sales progression service create new repository for codebase create pipelines for ci cd spin up new infrastructure | 1 |
24,855 | 7,573,435,781 | IssuesEvent | 2018-04-23 17:46:12 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | opened | Importing existing codebase with new launcher creates 2 build configs | area/app-wizard area/pipelines team/build-cd team/launcher type/bug | When using the new launcher experience to create an application by importing an existing codebase, I ended up with 2 new build configs. This seems to be due to the application name (app-test-1) differing from the Maven artifact ID (mynewvertx) in the repository. I believe the former is created by the launcher and the latter is created by Jenkins.
Steps to reproduce:
1. Using new launcher experience, select "Import an existing codebase".
2. Either choose a GitHub repository with Maven artifact ID different from the application name, or change the application name in the wizard to something else.

3. Wait for pipeline to start, observe 2 build configs (one named after application, one named after Maven artifact ID)

4. Application is built through pipeline twice in build config named after artifact ID:

5. Application is finally built through pipeline in build config named after the application:

Some consequences of this are that only the build config created by the launcher (named after the application) has the appropriate space label attached to it. Thus, this is the only pipeline that is reported to the user in the OSIO pipelines page. | 1.0 | Importing existing codebase with new launcher creates 2 build configs - When using the new launcher experience to create an application by importing an existing codebase, I ended up with 2 new build configs. This seems to be due to the application name (app-test-1) differing from the Maven artifact ID (mynewvertx) in the repository. I believe the former is created by the launcher and the latter is created by Jenkins.
Steps to reproduce:
1. Using new launcher experience, select "Import an existing codebase".
2. Either choose a GitHub repository with Maven artifact ID different from the application name, or change the application name in the wizard to something else.

3. Wait for pipeline to start, observe 2 build configs (one named after application, one named after Maven artifact ID)

4. Application is built through pipeline twice in build config named after artifact ID:

5. Application is finally built through pipeline in build config named after the application:

Some consequences of this are that only the build config created by the launcher (named after the application) has the appropriate space label attached to it. Thus, this is the only pipeline that is reported to the user in the OSIO pipelines page. | non_infrastructure | importing existing codebase with new launcher creates build configs when using the new launcher experience to create an application by importing an existing codebase i ended up with new build configs this seems to be due to the application name app test differing from the maven artifact id mynewvertx in the repository i believe the former is created by the launcher and the latter is created by jenkins steps to reproduce using new launcher experience select import an existing codebase either choose a github repository with maven artifact id different from the application name or change the application name in the wizard to something else wait for pipeline to start observe build configs one named after application one named after maven artifact id application is built through pipeline twice in build config named after artifact id application is finally built through pipeline in build config named after the application some consequences of this are that only the build config created by the launcher named after the application has the appropriate space label attached to it thus this is the only pipeline that is reported to the user in the osio pipelines page | 0 |
19,153 | 13,193,768,457 | IssuesEvent | 2020-08-13 15:43:33 | 18F/tts-tech-portfolio | https://api.github.com/repos/18F/tts-tech-portfolio | closed | Set up continuous deployment for aws-admin | g: accepted i: infrastructure improvement m: feedback needed t: days | ## Background information
Broken out from https://github.com/18F/aws-admin/issues/40
## User stories
As the tech portfolio I want a git workflow to manage deployment of a shared terraform state across all accounts and without state conflict.
## Acceptance criteria
- [ ] AWS baseline terraform is setup with continuous deployment using a persistent/shared .tfstate
| 1.0 | Set up continuous deployment for aws-admin - ## Background information
Broken out from https://github.com/18F/aws-admin/issues/40
## User stories
As the tech portfolio I want a git workflow to manage deployment of a shared terraform state across all accounts and without state conflict.
## Acceptance criteria
- [ ] AWS baseline terraform is setup with continuous deployment using a persistent/shared .tfstate
| infrastructure | set up continuous deployment for aws admin background information broken out from user stories as the tech portfolio i want a git workflow to manage deployment of a shared terraform state across all accounts and without state conflict acceptance criteria aws baseline terraform is setup with continuous deployment using a persistent shared tfstate | 1 |
576,161 | 17,080,429,846 | IssuesEvent | 2021-07-08 03:48:53 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | [LS] `CommonUtil.findNode()` logic is wrong for symbols | Priority/High Team/LanguageServer Type/Bug | **Description:**
`CommonUtil.findNode(Symbol symbol, SyntaxTree syntaxTree)` method's logic to evaluate the end offset is wrong.
```
int start = textDocument.textPositionFrom(symbolRange.startLine());
int len = symbolRange.endLine().offset() - symbolRange.startLine().offset();
```
It doesn't consider the endline.
**Steps to reproduce:**
See description
**Affected Versions:**
SL Beta 1 | 1.0 | [LS] `CommonUtil.findNode()` logic is wrong for symbols - **Description:**
`CommonUtil.findNode(Symbol symbol, SyntaxTree syntaxTree)` method's logic to evaluate the end offset is wrong.
```
int start = textDocument.textPositionFrom(symbolRange.startLine());
int len = symbolRange.endLine().offset() - symbolRange.startLine().offset();
```
It doesn't consider the endline.
**Steps to reproduce:**
See description
**Affected Versions:**
SL Beta 1 | non_infrastructure | commonutil findnode logic is wrong for symbols description commonutil findnode symbol symbol syntaxtree syntaxtree method s logic to evaluate the end offset is wrong int start textdocument textpositionfrom symbolrange startline int len symbolrange endline offset symbolrange startline offset it doesn t consider the endline steps to reproduce see description affected versions sl beta | 0 |
518 | 2,752,459,838 | IssuesEvent | 2015-04-24 15:40:58 | OpenSprites/OpenSprites | https://api.github.com/repos/OpenSprites/OpenSprites | closed | Dedicated Disk for OpenSprites | enhancement feature idea infrastructure input wanted low priority suggestion | When OpenSprites goes into production I may be able to attach a dedicated storage disk to the host node, and then symlink a directory in the OpenSprites webspace to that drive.
It should in theory provide a better R/W speed as well as the benefit of isolating user data from the main node's disk.
Thoughts? Obviously code will need to be adjusted to write data to that specific "disk directory". | 1.0 | Dedicated Disk for OpenSprites - When OpenSprites goes into production I may be able to attach a dedicated storage disk to the host node, and then symlink a directory in the OpenSprites webspace to that drive.
It should in theory provide a better R/W speed as well as the benefit of isolating user data from the main node's disk.
Thoughts? Obviously code will need to be adjusted to write data to that specific "disk directory". | infrastructure | dedicated disk for opensprites when opensprites goes into production i may be able to attach a dedicated storage disk to the host node and then symlink a directory in the opensprites webspace to that drive it should in theory provide a better r w speed as well as the benefit of isolating user data from the main node s disk thoughts obviously code will need to be adjusted to write data to that specific disk directory | 1 |
34,799 | 30,471,433,084 | IssuesEvent | 2023-07-17 13:52:56 | lampepfl/dotty | https://api.github.com/repos/lampepfl/dotty | closed | Benchmark server backlog | area:infrastructure | performance test scheduled: 239 job(s) in queue, 1 running.
_Originally posted by @dotty-bot in https://github.com/lampepfl/dotty/issues/18219#issuecomment-1637535411_
This backlog does not seem to be normal. | 1.0 | Benchmark server backlog - performance test scheduled: 239 job(s) in queue, 1 running.
_Originally posted by @dotty-bot in https://github.com/lampepfl/dotty/issues/18219#issuecomment-1637535411_
This backlog does not seem to be normal. | infrastructure | benchmark server backlog performance test scheduled job s in queue running originally posted by dotty bot in this backlog does not seem to be normal | 1 |
18,558 | 13,055,865,970 | IssuesEvent | 2020-07-30 02:58:07 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | README: tickets have been rearranged (Trac #727) | Incomplete Migration Migrated from Trac infrastructure task | Migrated from https://code.icecube.wisc.edu/ticket/727
```json
{
"status": "closed",
"changetime": "2015-08-17T15:52:36",
"description": "There is now '''one''' location for tickets. Here!\n\nTickets have been slightly re-arranged due to overlap.\n\n\"!IceTray\" tickets with numbers less than 157 have the same ticket number.\n\"!IceCube\" tickets have the same ticket number.\n\"!IceTray\" tickets with ticket numbers >= 157 have been incremented by 348. (i.e. The !IceTray ticket formerly known as 157 is now #505.)\n\nIf you find a ticket that has a ticket link that maybe wrong, file a new ticket referencing this one, and I'll fix it ASAP. ",
"reporter": "nega",
"cc": "",
"resolution": "wontfix",
"_ts": "1439826756684915",
"component": "infrastructure",
"summary": "README: tickets have been rearranged",
"priority": "README",
"keywords": "tickets",
"time": "2014-05-06T22:15:16",
"milestone": "",
"owner": "icecube",
"type": "task"
}
```
| 1.0 | README: tickets have been rearranged (Trac #727) - Migrated from https://code.icecube.wisc.edu/ticket/727
```json
{
"status": "closed",
"changetime": "2015-08-17T15:52:36",
"description": "There is now '''one''' location for tickets. Here!\n\nTickets have been slightly re-arranged due to overlap.\n\n\"!IceTray\" tickets with numbers less than 157 have the same ticket number.\n\"!IceCube\" tickets have the same ticket number.\n\"!IceTray\" tickets with ticket numbers >= 157 have been incremented by 348. (i.e. The !IceTray ticket formerly known as 157 is now #505.)\n\nIf you find a ticket that has a ticket link that maybe wrong, file a new ticket referencing this one, and I'll fix it ASAP. ",
"reporter": "nega",
"cc": "",
"resolution": "wontfix",
"_ts": "1439826756684915",
"component": "infrastructure",
"summary": "README: tickets have been rearranged",
"priority": "README",
"keywords": "tickets",
"time": "2014-05-06T22:15:16",
"milestone": "",
"owner": "icecube",
"type": "task"
}
```
| infrastructure | readme tickets have been rearranged trac migrated from json status closed changetime description there is now one location for tickets here n ntickets have been slightly re arranged due to overlap n n icetray tickets with numbers less than have the same ticket number n icecube tickets have the same ticket number n icetray tickets with ticket numbers have been incremented by i e the icetray ticket formerly known as is now n nif you find a ticket that has a ticket link that maybe wrong file a new ticket referencing this one and i ll fix it asap reporter nega cc resolution wontfix ts component infrastructure summary readme tickets have been rearranged priority readme keywords tickets time milestone owner icecube type task | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.