Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
581 | 2,775,505,941 | IssuesEvent | 2015-05-04 16:26:08 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | ContributingCode wikipage has missing link to CLA, stale link to style guide | Area-Infrastructure | [ContributingCode](https://github.com/dotnet/roslyn/wiki/Contributing-Code) on the wiki doesn't appear to have any link for the CLA, and the link for the style guide seems to have gotten out of date.
For the CLA, you could link to either <https://github.com/dotnet/corefx/wiki/Contribution-License-Agreement-(CLA)> or <https://cla2.dotnetfoundation.org/>.
For the style guide, you probably want to link to <https://github.com/dotnet/corefx/wiki/Coding-style> rather than to <https://github.com/dotnet/corefx/wiki/Contributing#c-coding-style>. | 1.0 | ContributingCode wikipage has missing link to CLA, stale link to style guide - [ContributingCode](https://github.com/dotnet/roslyn/wiki/Contributing-Code) on the wiki doesn't appear to have any link for the CLA, and the link for the style guide seems to have gotten out of date.
For the CLA, you could link to either <https://github.com/dotnet/corefx/wiki/Contribution-License-Agreement-(CLA)> or <https://cla2.dotnetfoundation.org/>.
For the style guide, you probably want to link to <https://github.com/dotnet/corefx/wiki/Coding-style> rather than to <https://github.com/dotnet/corefx/wiki/Contributing#c-coding-style>. | infrastructure | contributingcode wikipage has missing link to cla stale link to style guide on the wiki doesn t appear to have any link for the cla and the link for the style guide seems to have gotten out of date for the cla you could link to either or for the style guide you probably want to link to rather than to | 1 |
7,208 | 7,869,515,607 | IssuesEvent | 2018-06-24 14:52:25 | CodeWarriorsOrganization/CodeWarriorsFinalYearProject | https://api.github.com/repos/CodeWarriorsOrganization/CodeWarriorsFinalYearProject | closed | CW0025/T0025.4 -Implement the Service logic (Priority-2) | implementation service | ### User Story :
> AS a player,
> I NEED to be able to view the total number of turns, SO THAT I can get to know how many turns
> spend.
### Task :
> **Implement the Service logic** | 1.0 | CW0025/T0025.4 -Implement the Service logic (Priority-2) - ### User Story :
> AS a player,
> I NEED to be able to view the total number of turns, SO THAT I can get to know how many turns
> spend.
### Task :
> **Implement the Service logic** | non_infrastructure | implement the service logic priority user story as a player i need to be able to view the total number of turns so that i can get to know how many turns spend task implement the service logic | 0 |
11,865 | 9,476,410,363 | IssuesEvent | 2019-04-19 14:59:49 | opencv/opencv | https://api.github.com/repos/opencv/opencv | closed | Java: test crashes with OpenVINO 2019R1 on MacOSX | bug category: dnn category: infrastructure category: java bindings platform: ios/osx test | [Nightly build](http://pullrequest.opencv.org/buildbot/builders/master_openvino-mac/builds/0)
```
[junit] Running org.opencv.test.calib3d.Calib3dTest
[junit] java.lang.UnsatisfiedLinkError: /Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib: dlopen(/Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib, 1): Library not loaded: @rpath/libtbb.dylib
[junit] Referenced from: /opt/intel/openvino_2019.1.090/deployment_tools/inference_engine/lib/intel64/libinference_engine.dylib
[junit] Reason: image not found
[junit] Running org.opencv.test.calib3d.Calib3dTest
[junit] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0 sec
[junit] Test org.opencv.test.calib3d.Calib3dTest FAILED (crashed)
```
| 1.0 | Java: test crashes with OpenVINO 2019R1 on MacOSX - [Nightly build](http://pullrequest.opencv.org/buildbot/builders/master_openvino-mac/builds/0)
```
[junit] Running org.opencv.test.calib3d.Calib3dTest
[junit] java.lang.UnsatisfiedLinkError: /Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib: dlopen(/Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib, 1): Library not loaded: @rpath/libtbb.dylib
[junit] Referenced from: /opt/intel/openvino_2019.1.090/deployment_tools/inference_engine/lib/intel64/libinference_engine.dylib
[junit] Reason: image not found
[junit] Running org.opencv.test.calib3d.Calib3dTest
[junit] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0 sec
[junit] Test org.opencv.test.calib3d.Calib3dTest FAILED (crashed)
```
| infrastructure | java test crashes with openvino on macosx running org opencv test java lang unsatisfiedlinkerror volumes build storage build master openvino mac build lib libopencv dylib dlopen volumes build storage build master openvino mac build lib libopencv dylib library not loaded rpath libtbb dylib referenced from opt intel openvino deployment tools inference engine lib libinference engine dylib reason image not found running org opencv test tests run failures errors skipped time elapsed sec test org opencv test failed crashed | 1 |
3,597 | 4,427,999,246 | IssuesEvent | 2016-08-16 23:44:50 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | The build.cmd script should accept more common options | Infrastructure | Currently to build the repo as release you have to do this:
```cmd
build /p:ConfigurationGroup=Release
```
This is very error-prone to typos, since you can write something like `/p:ConfigurationGrouo=Release` and go away from your computer for 40 minutes only to realize you built on Debug configuration. We should provide `--debug` and `--release` shortcuts like is done in [the `build.sh` script](https://github.com/dotnet/corefx/blob/master/build.sh#L10), as well as update the documentation/`--help` output. Then, we could call the script like this:
```cmd
build --release
```
Other things we might want to consider: `--skiptests[ bool]` (default false), `--buildtests[ bool]` (default true), `--buildpackages[ bool]` (default true)
cc: @mellinoe
Note: I have knowledge of batch/shell scripting, so I can submit a PR to fix this myself if this is OK.
**edit:** Additionally we should echo the arguments that are being passed to MSBuild to the user, so they know right away if they made a typo. | 1.0 | The build.cmd script should accept more common options - Currently to build the repo as release you have to do this:
```cmd
build /p:ConfigurationGroup=Release
```
This is very error-prone to typos, since you can write something like `/p:ConfigurationGrouo=Release` and go away from your computer for 40 minutes only to realize you built on Debug configuration. We should provide `--debug` and `--release` shortcuts like is done in [the `build.sh` script](https://github.com/dotnet/corefx/blob/master/build.sh#L10), as well as update the documentation/`--help` output. Then, we could call the script like this:
```cmd
build --release
```
Other things we might want to consider: `--skiptests[ bool]` (default false), `--buildtests[ bool]` (default true), `--buildpackages[ bool]` (default true)
cc: @mellinoe
Note: I have knowledge of batch/shell scripting, so I can submit a PR to fix this myself if this is OK.
**edit:** Additionally we should echo the arguments that are being passed to MSBuild to the user, so they know right away if they made a typo. | infrastructure | the build cmd script should accept more common options currently to build the repo as release you have to do this cmd build p configurationgroup release this is very error prone to typos since you can write something like p configurationgrouo release and go away from your computer for minutes only to realize you built on debug configuration we should provide debug and release shortcuts like is done in as well as update the documentation help output then we could call the script like this cmd build release other things we might want to consider skiptests default false buildtests default true buildpackages default true cc mellinoe note i have knowledge of batch shell scripting so i can submit a pr to fix this myself if this is ok edit additionally we should echo the arguments that are being passed to msbuild to the user so they know right away if they made a typo | 1 |
517,991 | 15,021,963,680 | IssuesEvent | 2021-02-01 16:24:14 | BiologicalRecordsCentre/wrappeR | https://api.github.com/repos/BiologicalRecordsCentre/wrappeR | closed | Upload meta data creation script to wrapper | priority - high sprint | Upload meta data creation script to wrapper @mlogie | 1.0 | Upload meta data creation script to wrapper - Upload meta data creation script to wrapper @mlogie | non_infrastructure | upload meta data creation script to wrapper upload meta data creation script to wrapper mlogie | 0 |
102,777 | 11,307,054,780 | IssuesEvent | 2020-01-18 18:19:35 | uhop/stream-json | https://api.github.com/repos/uhop/stream-json | closed | Add type annotations to data-event | documentation enhancement | Hi,
From what I see, the `data` event on the stream accepts "any" as an argument. However in case of `StreamValues` the value is known, so I believe that having an annotation to point to `{key: number, value: any}` would be a good start. What do you think? Not sure how easy to plug that in though given that `StreamValues.withParser()` returns a Chain. | 1.0 | Add type annotations to data-event - Hi,
From what I see, the `data` event on the stream accepts "any" as an argument. However in case of `StreamValues` the value is known, so I believe that having an annotation to point to `{key: number, value: any}` would be a good start. What do you think? Not sure how easy to plug that in though given that `StreamValues.withParser()` returns a Chain. | non_infrastructure | add type annotations to data event hi from what i see the data event on the stream accepts any as an argument however in case of streamvalues the value is known so i believe that having an annotation to point to key number value any would be a good start what do you think not sure how easy to plug that in though given that streamvalues withparser returns a chain | 0 |
900 | 2,997,357,739 | IssuesEvent | 2015-07-23 06:55:19 | gammapy/gammapy | https://api.github.com/repos/gammapy/gammapy | opened | Update install instructions and release notes for conda packages | docs infrastructure | Matt Craig set up binary packages for Gammapy:
https://groups.google.com/forum/#!topic/astropy-dev/u8-r3AW8V4s
https://groups.google.com/forum/#!topic/gammapy/WJjWxHphTGM
- [ ] Update install instructions
- [ ] Update release notes to make a PR against https://github.com/astropy/conda-builder-affiliated to trigger a new build
- [ ] Maybe try to get some of our optional dependencies built there (e.g. iminuit or Gammalib / ctools)
| 1.0 | Update install instructions and release notes for conda packages - Matt Craig set up binary packages for Gammapy:
https://groups.google.com/forum/#!topic/astropy-dev/u8-r3AW8V4s
https://groups.google.com/forum/#!topic/gammapy/WJjWxHphTGM
- [ ] Update install instructions
- [ ] Update release notes to make a PR against https://github.com/astropy/conda-builder-affiliated to trigger a new build
- [ ] Maybe try to get some of our optional dependencies built there (e.g. iminuit or Gammalib / ctools)
| infrastructure | update install instructions and release notes for conda packages matt craig set up binary packages for gammapy update install instructions update release notes to make a pr against to trigger a new build maybe try to get some of our optional dependencies built there e g iminuit or gammalib ctools | 1 |
28,241 | 23,103,585,783 | IssuesEvent | 2022-07-27 06:41:41 | FTBTeam/FTB-Modpack-Issues | https://api.github.com/repos/FTBTeam/FTB-Modpack-Issues | closed | [Bug]: Direwolf20 1.18 Linux fail to download | 3: Infrastructure FTB Presents Direwolf20 1.18 | ### Modpack
Direwolf20 1.18
### Modpack version
2234
### Log Files
491 Successful 266 Failed
https://pste.ch/zoyojitozo.pl
Continue anyway.
https://pste.ch/jovurutaxi.sql
### Describe the bug
With linux, the server does not download fully, having 266 Failed downloads. Additionally, the start.sh is not formatted correctly. https://pste.ch/ijopamevun.bash
### Steps to reproduce
1. wget https://feed-the-beast.com/modpack/95_ftb_presents_direwolf20_1_18/server-files
2. chmod +x linux
3. ./linux 95
4. follow prompts
### Expected behaviour
Direwolf20 1.18 Modpack v.2234 should be downloaded.
### Screenshots
_No response_
### Additional information
I've tried using the Windows server and the download also didn't work correctly. | 1.0 | [Bug]: Direwolf20 1.18 Linux fail to download - ### Modpack
Direwolf20 1.18
### Modpack version
2234
### Log Files
491 Successful 266 Failed
https://pste.ch/zoyojitozo.pl
Continue anyway.
https://pste.ch/jovurutaxi.sql
### Describe the bug
With linux, the server does not download fully, having 266 Failed downloads. Additionally, the start.sh is not formatted correctly. https://pste.ch/ijopamevun.bash
### Steps to reproduce
1. wget https://feed-the-beast.com/modpack/95_ftb_presents_direwolf20_1_18/server-files
2. chmod +x linux
3. ./linux 95
4. follow prompts
### Expected behaviour
Direwolf20 1.18 Modpack v.2234 should be downloaded.
### Screenshots
_No response_
### Additional information
I've tried using the Windows server and the download also didn't work correctly. | infrastructure | linux fail to download modpack modpack version log files successful failed continue anyway describe the bug with linux the server does not download fully having failed downloads additionally the start sh is not formatted correctly steps to reproduce wget chmod x linux linux follow prompts expected behaviour modpack v should be downloaded screenshots no response additional information i ve tried using the windows server and the download also didn t work correctly | 1 |
14,348 | 10,760,573,873 | IssuesEvent | 2019-10-31 18:53:17 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Integration tests sometimes have broken installation | Area-Infrastructure Integration-Test | Tests:
Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorList
Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorsDuringMethodBodyEditing
Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicSquigglesNetCore.VerifySemanticErrorSquiggles
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorsDuringMethodBodyEditing
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorList
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorLevelWarning
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySemanticErrorSquiggles
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySyntaxErrorSquiggles
Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.OpenCSharpThenVBSolution
Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.ProjectReference
| 1.0 | Integration tests sometimes have broken installation - Tests:
Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorList
Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorsDuringMethodBodyEditing
Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicSquigglesNetCore.VerifySemanticErrorSquiggles
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorsDuringMethodBodyEditing
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorList
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorLevelWarning
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySemanticErrorSquiggles
Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySyntaxErrorSquiggles
Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.OpenCSharpThenVBSolution
Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.ProjectReference
| infrastructure | integration tests sometimes have broken installation tests roslyn visualstudio integrationtests visualbasic basicerrorlistnetcore errorlist roslyn visualstudio integrationtests visualbasic basicerrorlistnetcore errorsduringmethodbodyediting roslyn visualstudio integrationtests visualbasic basicsquigglesnetcore verifysemanticerrorsquiggles roslyn visualstudio integrationtests csharp csharperrorlistnetcore errorsduringmethodbodyediting roslyn visualstudio integrationtests csharp csharperrorlistnetcore errorlist roslyn visualstudio integrationtests csharp csharperrorlistnetcore errorlevelwarning roslyn visualstudio integrationtests csharp csharpsquigglesnetcore verifysemanticerrorsquiggles roslyn visualstudio integrationtests csharp csharpsquigglesnetcore verifysyntaxerrorsquiggles roslyn visualstudio integrationtests workspace workspacesnetcore opencsharpthenvbsolution roslyn visualstudio integrationtests workspace workspacesnetcore projectreference | 1 |
202,684 | 7,051,315,157 | IssuesEvent | 2018-01-03 11:09:41 | fusetools/fuselibs-public | https://api.github.com/repos/fusetools/fuselibs-public | closed | Many `...Operator` classes don't deal with invalid input | Priority: High Severity: Bug | There was no mechanism to report lost data before, so these were likely forwarding null. That would explain many of the odd errors I've seen using expression.
The operators can now lose data and they should. Many, like in "Operators.uno" call functions like `Marshal.Add` and assume it works. There is no checking of the argument types.
Some of the conversions were handled with exceptions before. Exceptions are not properly handled in the exception hierarchy, that is, they don't result in lost data. We could instead introduce a special exception and handle it, but that feels far less clean than explicitly handling invalid conversion (these are an expecte condition with user code and bindings). | 1.0 | Many `...Operator` classes don't deal with invalid input - There was no mechanism to report lost data before, so these were likely forwarding null. That would explain many of the odd errors I've seen using expression.
The operators can now lose data and they should. Many, like in "Operators.uno" call functions like `Marshal.Add` and assume it works. There is no checking of the argument types.
Some of the conversions were handled with exceptions before. Exceptions are not properly handled in the exception hierarchy, that is, they don't result in lost data. We could instead introduce a special exception and handle it, but that feels far less clean than explicitly handling invalid conversion (these are an expecte condition with user code and bindings). | non_infrastructure | many operator classes don t deal with invalid input there was no mechanism to report lost data before so these were likely forwarding null that would explain many of the odd errors i ve seen using expression the operators can now lose data and they should many like in operators uno call functions like marshal add and assume it works there is no checking of the argument types some of the conversions were handled with exceptions before exceptions are not properly handled in the exception hierarchy that is they don t result in lost data we could instead introduce a special exception and handle it but that feels far less clean than explicitly handling invalid conversion these are an expecte condition with user code and bindings | 0 |
221,704 | 17,365,858,755 | IssuesEvent | 2021-07-30 07:09:52 | hoechenberger/mne-installers | https://api.github.com/repos/hoechenberger/mne-installers | opened | Run CI tests for Windows installer too | tests windows | Currently, we only build the package, but don't install & test it. | 1.0 | Run CI tests for Windows installer too - Currently, we only build the package, but don't install & test it. | non_infrastructure | run ci tests for windows installer too currently we only build the package but don t install test it | 0 |
27,300 | 21,580,496,156 | IssuesEvent | 2022-05-02 18:10:37 | observingClouds/xbitinfo | https://api.github.com/repos/observingClouds/xbitinfo | closed | Reduce dependencies | infrastructure | pip allows extras when installing, see https://github.com/pangeo-data/climpred/blob/main/setup.py or https://github.com/pydata/xarray/blob/main/setup.cfg for non essential deps
all viz libs could be grouped, also parallel for dask and prefect | 1.0 | Reduce dependencies - pip allows extras when installing, see https://github.com/pangeo-data/climpred/blob/main/setup.py or https://github.com/pydata/xarray/blob/main/setup.cfg for non essential deps
all viz libs could be grouped, also parallel for dask and prefect | infrastructure | reduce dependencies pip allows extras when installing see or for non essential deps all viz libs could be grouped also parallel for dask and prefect | 1 |
97,918 | 20,574,002,159 | IssuesEvent | 2022-03-04 01:09:55 | sourcegraph/sourcegraph | https://api.github.com/repos/sourcegraph/sourcegraph | opened | insights: modify API to take into account frozen insights | team/code-insights backend | 1. Add a new field `frozen` to the `InsightView` type in the API and make sure it accurately reflects whether or not the insight is frozen, (based on the `is_frozen` field in `insight_view`.)
2. When returning insights, do not return any `dataSeries` data for frozen insights.
3. Potentially other restrictions (creating new insights, editing frozen insights, etc.) on the API based on whether or not the user has a Code Insights license. (TBD) | 1.0 | insights: modify API to take into account frozen insights - 1. Add a new field `frozen` to the `InsightView` type in the API and make sure it accurately reflects whether or not the insight is frozen, (based on the `is_frozen` field in `insight_view`.)
2. When returning insights, do not return any `dataSeries` data for frozen insights.
3. Potentially other restrictions (creating new insights, editing frozen insights, etc.) on the API based on whether or not the user has a Code Insights license. (TBD) | non_infrastructure | insights modify api to take into account frozen insights add a new field frozen to the insightview type in the api and make sure it accurately reflects whether or not the insight is frozen based on the is frozen field in insight view when returning insights do not return any dataseries data for frozen insights potentially other restrictions creating new insights editing frozen insights etc on the api based on whether or not the user has a code insights license tbd | 0 |
96,369 | 10,932,060,853 | IssuesEvent | 2019-11-23 15:05:57 | aaFn/Bookmark-search-plus-2 | https://api.github.com/repos/aaFn/Bookmark-search-plus-2 | closed | Can't find ⓘ in address bar | documentation | The instrtuctions for Bookmark-search-plus-2 at Firefox: about additions say, "drag the ⓘ in the address bar or drag web page links into the sidebar, and insert them in the tree at the place you like".
I can't find the ⓘ in the address bar, although I used to be able to. What am I doing wrong? | 1.0 | Can't find ⓘ in address bar - The instrtuctions for Bookmark-search-plus-2 at Firefox: about additions say, "drag the ⓘ in the address bar or drag web page links into the sidebar, and insert them in the tree at the place you like".
I can't find the ⓘ in the address bar, although I used to be able to. What am I doing wrong? | non_infrastructure | can t find ⓘ in address bar the instrtuctions for bookmark search plus at firefox about additions say drag the ⓘ in the address bar or drag web page links into the sidebar and insert them in the tree at the place you like i can t find the ⓘ in the address bar although i used to be able to what am i doing wrong | 0 |
470,124 | 13,531,461,881 | IssuesEvent | 2020-09-15 21:41:51 | grpc/grpc | https://api.github.com/repos/grpc/grpc | opened | Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x | kind/bug priority/P0/RELEASE BLOCKER priority/P2 | This is a blocker for v1.31.x and v1.32.x patches only.
Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x when we patch these branches.
Mark this issue resolved when we no longer need to patch these branches, i.e., when 1.34 is released.
| 2.0 | Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x - This is a blocker for v1.31.x and v1.32.x patches only.
Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x when we patch these branches.
Mark this issue resolved when we no longer need to patch these branches, i.e., when 1.34 is released.
| non_infrastructure | backport to x and x this is a blocker for x and x patches only backport to x and x when we patch these branches mark this issue resolved when we no longer need to patch these branches i e when is released | 0 |
16,046 | 11,809,026,928 | IssuesEvent | 2020-03-19 14:20:34 | google/iree | https://api.github.com/repos/google/iree | closed | LLVMX86CodeGen not linking correctly (cmake/MSVC) | infrastructure | Library is not found during linking of any iree_cc_binary that transitively depends on the llvmjit_driver_module. May be a bad LLVM configuration in our options (not telling the x86 codegen backend to build?) | 1.0 | LLVMX86CodeGen not linking correctly (cmake/MSVC) - Library is not found during linking of any iree_cc_binary that transitively depends on the llvmjit_driver_module. May be a bad LLVM configuration in our options (not telling the x86 codegen backend to build?) | infrastructure | not linking correctly cmake msvc library is not found during linking of any iree cc binary that transitively depends on the llvmjit driver module may be a bad llvm configuration in our options not telling the codegen backend to build | 1 |
97,133 | 20,169,263,072 | IssuesEvent | 2022-02-10 08:54:57 | GeoNode/geonode | https://api.github.com/repos/GeoNode/geonode | closed | Drop legacy dataset upload page and implement client upload url hookset | enhancement code quality master | The legacy upload page is removed. Redirection to the client upload url is implemented with a hookset. This is adopted inside the remaining upload pages (step views). | 1.0 | Drop legacy dataset upload page and implement client upload url hookset - The legacy upload page is removed. Redirection to the client upload url is implemented with a hookset. This is adopted inside the remaining upload pages (step views). | non_infrastructure | drop legacy dataset upload page and implement client upload url hookset the legacy upload page is removed redirection to the client upload url is implemented with a hookset this is adopted inside the remaining upload pages step views | 0 |
159,688 | 20,085,892,657 | IssuesEvent | 2022-02-05 01:07:59 | AkshayMukkavilli/Tensorflow | https://api.github.com/repos/AkshayMukkavilli/Tensorflow | opened | CVE-2021-41200 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2021-41200 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions if `tf.summary.create_file_writer` is called with non-scalar arguments code crashes due to a `CHECK`-fail. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41200>CVE-2021-41200</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-41200 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-41200 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to dependency file: /Tensorflow/src/requirements.txt</p>
<p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. In affected versions if `tf.summary.create_file_writer` is called with non-scalar arguments code crashes due to a `CHECK`-fail. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range.
<p>Publish Date: 2021-11-05
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41200>CVE-2021-41200</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f</a></p>
<p>Release Date: 2021-11-05</p>
<p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file tensorflow src requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning in affected versions if tf summary create file writer is called with non scalar arguments code crashes due to a check fail the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource | 0 |
116,155 | 11,902,477,308 | IssuesEvent | 2020-03-30 14:01:26 | IBM-Blockchain/ansible-collection | https://api.github.com/repos/IBM-Blockchain/ansible-collection | closed | Add missing URLS to Galaxy metadata | documentation | Galaxy is missing the documentation, homepage, and issues URLs - now we have these, we can add them in. | 1.0 | Add missing URLS to Galaxy metadata - Galaxy is missing the documentation, homepage, and issues URLs - now we have these, we can add them in. | non_infrastructure | add missing urls to galaxy metadata galaxy is missing the documentation homepage and issues urls now we have these we can add them in | 0 |
34,104 | 28,242,166,387 | IssuesEvent | 2023-04-06 08:03:46 | localstack/localstack | https://api.github.com/repos/localstack/localstack | closed | bug: Cannot connect to localstack v2.0 in EKS cluster | type: bug status: triage needed area: infrastructure area: networking | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
We have the image localstack/localstack:latest-amd64 deployed into an EKS cluster running SNS and SQS - this picks up the version 2.0 release of localstack.
When the service we have deployed in the same cluster attempts to connect to localstack we receive an error -
`Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused `
It works fine if we revert back to the image -
`localstack/localstack:1.4-amd64`
We have sen this error since the release of 2.0 (which our build picks up as being the latest version.
The full stack is attached.
### Expected Behavior
When the service in the EKS cluster starts up, it shoudl connect successfully to localstack.
### How are you starting LocalStack?
Custom (please describe below)
### Steps To Reproduce
#### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`)
This is done using a helm chart which is passed into a docker run command - docker run -it
### Environment
```markdown
- OS:linux/amd64 JDK 17
- LocalStack:2.0 AMD64
```
### Anything else?
Don't seem able to attach log, but have pasted here -
`Caused by: com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleRetryableException(AmazonHttpClient.java:1219) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1165) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?]
│
Caused by: java.net.ConnectException: Connection refused
at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?]
at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?]
at sun.nio.ch.NioSocketImpl.timedFinishConnect(NioSocketImpl.java:549) ~[?:?]
at sun.nio.ch.NioSocketImpl.connect(NioSocketImpl.java:597) ~[?:?]
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:327) ~[?:?]
at java.net.Socket.connect(Socket.java:633) ~[?:?]
at org.apache.http.conn.socket.PlainConnectionSocketFactory.connectSocket(PlainConnectionSocketFactory.java:75) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.conn.DefaultHttpClientConnectionOperator.connect(DefaultHttpClientConnectionOperator.java:142) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.conn.PoolingHttpClientConnectionManager.connect(PoolingHttpClientConnectionManager.java:376) ~[httpclient-4.5.13.jar:4.5.13]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?]
at com.amazonaws.http.conn.ClientConnectionManagerFactory$Handler.invoke(ClientConnectionManagerFactory.java:76) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.conn.$Proxy133.connect(Unknown Source) ~[?:?]
at org.apache.http.impl.execchain.MainClientExec.establishRoute(MainClientExec.java:393) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:236) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:186) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56) ~[httpclient-4.5.13.jar:4.5.13]
at com.amazonaws.http.apache.client.impl.SdkHttpClient.execute(SdkHttpClient.java:72) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeOneRequest(AmazonHttpClient.java:1346) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1157) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?]
` | 1.0 | bug: Cannot connect to localstack v2.0 in EKS cluster - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
We have the image localstack/localstack:latest-amd64 deployed into an EKS cluster running SNS and SQS - this picks up the version 2.0 release of localstack.
When the service we have deployed in the same cluster attempts to connect to localstack we receive an error -
`Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused `
It works fine if we revert back to the image -
`localstack/localstack:1.4-amd64`
We have sen this error since the release of 2.0 (which our build picks up as being the latest version.
The full stack is attached.
### Expected Behavior
When the service in the EKS cluster starts up, it shoudl connect successfully to localstack.
### How are you starting LocalStack?
Custom (please describe below)
### Steps To Reproduce
#### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`)
This is done using a helm chart which is passed into a docker run command - docker run -it
### Environment
```markdown
- OS:linux/amd64 JDK 17
- LocalStack:2.0 AMD64
```
### Anything else?
Don't seem able to attach log, but have pasted here -
`Caused by: com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleRetryableException(AmazonHttpClient.java:1219) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1165) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?]
│
Caused by: java.net.ConnectException: Connection refused
at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?]
at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?]
at sun.nio.ch.NioSocketImpl.timedFinishConnect(NioSocketImpl.java:549) ~[?:?]
at sun.nio.ch.NioSocketImpl.connect(NioSocketImpl.java:597) ~[?:?]
at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:327) ~[?:?]
at java.net.Socket.connect(Socket.java:633) ~[?:?]
at org.apache.http.conn.socket.PlainConnectionSocketFactory.connectSocket(PlainConnectionSocketFactory.java:75) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.conn.DefaultHttpClientConnectionOperator.connect(DefaultHttpClientConnectionOperator.java:142) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.conn.PoolingHttpClientConnectionManager.connect(PoolingHttpClientConnectionManager.java:376) ~[httpclient-4.5.13.jar:4.5.13]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?]
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?]
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?]
at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?]
at com.amazonaws.http.conn.ClientConnectionManagerFactory$Handler.invoke(ClientConnectionManagerFactory.java:76) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.conn.$Proxy133.connect(Unknown Source) ~[?:?]
at org.apache.http.impl.execchain.MainClientExec.establishRoute(MainClientExec.java:393) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:236) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:186) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83) ~[httpclient-4.5.13.jar:4.5.13]
at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56) ~[httpclient-4.5.13.jar:4.5.13]
at com.amazonaws.http.apache.client.impl.SdkHttpClient.execute(SdkHttpClient.java:72) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeOneRequest(AmazonHttpClient.java:1346) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1157) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?]
at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?]
` | infrastructure | bug cannot connect to localstack in eks cluster is there an existing issue for this i have searched the existing issues current behavior we have the image localstack localstack latest deployed into an eks cluster running sns and sqs this picks up the version release of localstack when the service we have deployed in the same cluster attempts to connect to localstack we receive an error unable to execute http request connect to aws failed connection refused it works fine if we revert back to the image localstack localstack we have sen this error since the release of which our build picks up as being the latest version the full stack is attached expected behavior when the service in the eks cluster starts up it shoudl connect successfully to localstack how are you starting localstack custom please describe below steps to reproduce how are you starting localstack e g bin localstack command arguments or docker compose yml this is done using a helm chart which is passed into a docker run command docker run it environment markdown os linux jdk localstack anything else don t seem able to attach log but have pasted here caused by com amazonaws sdkclientexception unable to execute http request connect to aws failed connection refused at com amazonaws http amazonhttpclient requestexecutor handleretryableexception amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executehelper amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor doexecute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executewithtimer amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor execute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor access amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutionbuilderimpl execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws services sqs amazonsqsclient doinvoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient executecreatequeue amazonsqsclient java at com amazonaws services sqs amazonsqsclient createqueue amazonsqsclient java at com amazonaws services sqs buffered amazonsqsbufferedasyncclient createqueue amazonsqsbufferedasyncclient java │ caused by java net connectexception connection refused at sun nio ch net pollconnect native method at sun nio ch net pollconnectnow net java at sun nio ch niosocketimpl timedfinishconnect niosocketimpl java at sun nio ch niosocketimpl connect niosocketimpl java at java net sockssocketimpl connect sockssocketimpl java at java net socket connect socket java at org apache http conn socket plainconnectionsocketfactory connectsocket plainconnectionsocketfactory java at org apache http impl conn defaulthttpclientconnectionoperator connect defaulthttpclientconnectionoperator java at org apache http impl conn poolinghttpclientconnectionmanager connect poolinghttpclientconnectionmanager java at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com amazonaws http conn clientconnectionmanagerfactory handler invoke clientconnectionmanagerfactory java at com amazonaws http conn connect unknown source at org apache http impl execchain mainclientexec establishroute mainclientexec java at org apache http impl execchain mainclientexec execute mainclientexec java at org apache http impl execchain protocolexec execute protocolexec java at org apache http impl client internalhttpclient doexecute internalhttpclient java at org apache http impl client closeablehttpclient execute closeablehttpclient java at org apache http impl client closeablehttpclient execute closeablehttpclient java at com amazonaws http apache client impl sdkhttpclient execute sdkhttpclient java at com amazonaws http amazonhttpclient requestexecutor executeonerequest amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executehelper amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor doexecute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executewithtimer amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor execute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor access amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutionbuilderimpl execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws services sqs amazonsqsclient doinvoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient executecreatequeue amazonsqsclient java at com amazonaws services sqs amazonsqsclient createqueue amazonsqsclient java at com amazonaws services sqs buffered amazonsqsbufferedasyncclient createqueue amazonsqsbufferedasyncclient java | 1 |
30,439 | 4,615,248,714 | IssuesEvent | 2016-09-26 00:46:58 | Semantic-Org/Semantic-UI | https://api.github.com/repos/Semantic-Org/Semantic-UI | closed | Input type text only displays placeholder text on focus event | Needs Test Case Read the Contributing Guide / Readme Usage Question | I just updated to semantic ui version 2.2.4 in order to use some new features but after upgrading semantic-ui all my inputs only show placeholder text on the user clicks on the input (focus).
I'm trying with a simple input:
`<div className="ui icon input">
<input type="text" placeholder="Search for something"/>
<i className="search icon"></i>
</div>
</div>` | 1.0 | Input type text only displays placeholder text on focus event - I just updated to semantic ui version 2.2.4 in order to use some new features but after upgrading semantic-ui all my inputs only show placeholder text on the user clicks on the input (focus).
I'm trying with a simple input:
`<div className="ui icon input">
<input type="text" placeholder="Search for something"/>
<i className="search icon"></i>
</div>
</div>` | non_infrastructure | input type text only displays placeholder text on focus event i just updated to semantic ui version in order to use some new features but after upgrading semantic ui all my inputs only show placeholder text on the user clicks on the input focus i m trying with a simple input | 0 |
23,211 | 4,894,243,298 | IssuesEvent | 2016-11-19 05:58:26 | F5Networks/f5-icontrol-rest-python | https://api.github.com/repos/F5Networks/f5-icontrol-rest-python | closed | Update README... | documentation | Need to clarify differences (and dependencies) between this repo and f5-common-python.
| 1.0 | Update README... - Need to clarify differences (and dependencies) between this repo and f5-common-python.
| non_infrastructure | update readme need to clarify differences and dependencies between this repo and common python | 0 |
20,662 | 14,099,320,726 | IssuesEvent | 2020-11-06 01:05:06 | noahtalerman/test-issues-kolide | https://api.github.com/repos/noahtalerman/test-issues-kolide | opened | [CLOSED] cleaning up some development files | Component: Development Infrastructure | <a href="https://github.com/marpaia"><img src="https://avatars2.githubusercontent.com/u/927168?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [marpaia](https://github.com/marpaia)**
_Monday Sep 05, 2016 at 21:55 GMT_
_Originally opened as https://github.com/kolide/fleet/pull/115_
----
----
_**[marpaia](https://github.com/marpaia)** included the following code: https://github.com/kolide/fleet/pull/115/commits_
| 1.0 | [CLOSED] cleaning up some development files - <a href="https://github.com/marpaia"><img src="https://avatars2.githubusercontent.com/u/927168?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [marpaia](https://github.com/marpaia)**
_Monday Sep 05, 2016 at 21:55 GMT_
_Originally opened as https://github.com/kolide/fleet/pull/115_
----
----
_**[marpaia](https://github.com/marpaia)** included the following code: https://github.com/kolide/fleet/pull/115/commits_
| infrastructure | cleaning up some development files issue by monday sep at gmt originally opened as included the following code | 1 |
481,212 | 13,882,047,082 | IssuesEvent | 2020-10-18 04:27:52 | python/mypy | https://api.github.com/repos/python/mypy | closed | Can't assign to () | bug false-positive priority-2-low | Mypy incorrectly complains about assignments to `()`. The simplest example of this is
```python
() = []
```
This code executes without a problem, but mypy says `error: can't assign to ()`
A more real world use case might be code like this
```python
(a, b), () = [[1, 2], []]
```
In this case the assignment to `()` is a concise way of asserting that the second list is empty.
* What is the actual behavior/output?
`error: can't assign to ()`
* What is the behavior/output you expect?
No type checking errors
* What are the versions of mypy and Python you are using?
A build from the latest commit as of writing this, 614090b7
* What are the mypy flags you are using?
No flags
If this is a change you decide would be good then I believe it could be implemented by removing the the check on line 1930 of `mypy/semanal.py` and updating the tests.
| 1.0 | Can't assign to () - Mypy incorrectly complains about assignments to `()`. The simplest example of this is
```python
() = []
```
This code executes without a problem, but mypy says `error: can't assign to ()`
A more real world use case might be code like this
```python
(a, b), () = [[1, 2], []]
```
In this case the assignment to `()` is a concise way of asserting that the second list is empty.
* What is the actual behavior/output?
`error: can't assign to ()`
* What is the behavior/output you expect?
No type checking errors
* What are the versions of mypy and Python you are using?
A build from the latest commit as of writing this, 614090b7
* What are the mypy flags you are using?
No flags
If this is a change you decide would be good then I believe it could be implemented by removing the the check on line 1930 of `mypy/semanal.py` and updating the tests.
| non_infrastructure | can t assign to mypy incorrectly complains about assignments to the simplest example of this is python this code executes without a problem but mypy says error can t assign to a more real world use case might be code like this python a b in this case the assignment to is a concise way of asserting that the second list is empty what is the actual behavior output error can t assign to what is the behavior output you expect no type checking errors what are the versions of mypy and python you are using a build from the latest commit as of writing this what are the mypy flags you are using no flags if this is a change you decide would be good then i believe it could be implemented by removing the the check on line of mypy semanal py and updating the tests | 0 |
312,902 | 9,554,724,834 | IssuesEvent | 2019-05-02 23:12:17 | PerezArturo/Medum | https://api.github.com/repos/PerezArturo/Medum | reopened | HU 1- autentificación | priority: high type: US | -Descripción.
Yo como dueño o interesado en rentar una propiedad deseo iniciar sesión mediante mi numero de teléfono.
- Criterios de Aceptación.
1. Dado que el usuario desea iniciar sesión con su numero teléfono deberá ingresar numero para iniciar
2. dado que el usuario ingrese numero y no este registrado deberá registrarse primero para poder iniciar
| 1.0 | HU 1- autentificación - -Descripción.
Yo como dueño o interesado en rentar una propiedad deseo iniciar sesión mediante mi numero de teléfono.
- Criterios de Aceptación.
1. Dado que el usuario desea iniciar sesión con su numero teléfono deberá ingresar numero para iniciar
2. dado que el usuario ingrese numero y no este registrado deberá registrarse primero para poder iniciar
| non_infrastructure | hu autentificación descripción yo como dueño o interesado en rentar una propiedad deseo iniciar sesión mediante mi numero de teléfono criterios de aceptación dado que el usuario desea iniciar sesión con su numero teléfono deberá ingresar numero para iniciar dado que el usuario ingrese numero y no este registrado deberá registrarse primero para poder iniciar | 0 |
110,512 | 16,980,344,145 | IssuesEvent | 2021-06-30 08:02:45 | samuelteixeiras/learnRuby | https://api.github.com/repos/samuelteixeiras/learnRuby | opened | CVE-2012-2661 (Medium) detected in activerecord-3.0.9.gem | security vulnerability | ## CVE-2012-2661 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activerecord-3.0.9.gem</b></p></summary>
<p>Databases on Rails. Build a persistent domain model by mapping database tables to Ruby classes. Strong conventions for associations, validations, aggregations, migrations, and testing come baked-in.</p>
<p>Library home page: <a href="https://rubygems.org/gems/activerecord-3.0.9.gem">https://rubygems.org/gems/activerecord-3.0.9.gem</a></p>
<p>
Dependency Hierarchy:
- rails-3.0.9.gem (Root Library)
- :x: **activerecord-3.0.9.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samuelteixeiras/learnRuby/commit/0c785fd9400921392b8ee5e3e166f30364359ecc">0c785fd9400921392b8ee5e3e166f30364359ecc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Active Record component in Ruby on Rails 3.0.x before 3.0.13, 3.1.x before 3.1.5, and 3.2.x before 3.2.4 does not properly implement the passing of request data to a where method in an ActiveRecord class, which allows remote attackers to conduct certain SQL injection attacks via nested query parameters that leverage unintended recursion, a related issue to CVE-2012-2695.
<p>Publish Date: 2012-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-2661>CVE-2012-2661</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-2661">https://nvd.nist.gov/vuln/detail/CVE-2012-2661</a></p>
<p>Release Date: 2012-06-22</p>
<p>Fix Resolution: 3.0.13,3.1.5,3.2.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2012-2661 (Medium) detected in activerecord-3.0.9.gem - ## CVE-2012-2661 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activerecord-3.0.9.gem</b></p></summary>
<p>Databases on Rails. Build a persistent domain model by mapping database tables to Ruby classes. Strong conventions for associations, validations, aggregations, migrations, and testing come baked-in.</p>
<p>Library home page: <a href="https://rubygems.org/gems/activerecord-3.0.9.gem">https://rubygems.org/gems/activerecord-3.0.9.gem</a></p>
<p>
Dependency Hierarchy:
- rails-3.0.9.gem (Root Library)
- :x: **activerecord-3.0.9.gem** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samuelteixeiras/learnRuby/commit/0c785fd9400921392b8ee5e3e166f30364359ecc">0c785fd9400921392b8ee5e3e166f30364359ecc</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The Active Record component in Ruby on Rails 3.0.x before 3.0.13, 3.1.x before 3.1.5, and 3.2.x before 3.2.4 does not properly implement the passing of request data to a where method in an ActiveRecord class, which allows remote attackers to conduct certain SQL injection attacks via nested query parameters that leverage unintended recursion, a related issue to CVE-2012-2695.
<p>Publish Date: 2012-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-2661>CVE-2012-2661</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-2661">https://nvd.nist.gov/vuln/detail/CVE-2012-2661</a></p>
<p>Release Date: 2012-06-22</p>
<p>Fix Resolution: 3.0.13,3.1.5,3.2.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in activerecord gem cve medium severity vulnerability vulnerable library activerecord gem databases on rails build a persistent domain model by mapping database tables to ruby classes strong conventions for associations validations aggregations migrations and testing come baked in library home page a href dependency hierarchy rails gem root library x activerecord gem vulnerable library found in head commit a href found in base branch master vulnerability details the active record component in ruby on rails x before x before and x before does not properly implement the passing of request data to a where method in an activerecord class which allows remote attackers to conduct certain sql injection attacks via nested query parameters that leverage unintended recursion a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
113,709 | 11,812,538,054 | IssuesEvent | 2020-03-19 20:22:20 | wbarnha/gr-gmuground | https://api.github.com/repos/wbarnha/gr-gmuground | closed | Reorganize apps and examples | documentation | There's a lot of clutter and unexplained components in the software. We should also integrate a method of compiling the hierarchical blocks automatically in `setup.sh` for convenience. | 1.0 | Reorganize apps and examples - There's a lot of clutter and unexplained components in the software. We should also integrate a method of compiling the hierarchical blocks automatically in `setup.sh` for convenience. | non_infrastructure | reorganize apps and examples there s a lot of clutter and unexplained components in the software we should also integrate a method of compiling the hierarchical blocks automatically in setup sh for convenience | 0 |
13,123 | 10,131,762,946 | IssuesEvent | 2019-08-01 20:24:24 | HumanCellAtlas/secondary-analysis | https://api.github.com/repos/HumanCellAtlas/secondary-analysis | closed | Tex reports that some Jenkins jobs write into `~devopsci` on every run. | infrastructure | [~teixeira] reports via [~ferrara]:
can someone figure out why dsde-jenkins is writing `jenkins-dsde-pipelines-run-automated-test` `jenkins-gotc-arrays-dev-nightly-test` and several other directories to `devopsci`'s home directory daily?
That probably comes from this code.
https://github.com/broadinstitute/dsp-jenkins/blob/fd2fd5720925d68a0baa3bc2b86021702dd3edd0/src/main/groovy/org/broadinstitute/dspjenkins/CodeSnippets.groovy#L20
Looks like anything running with `picard` authorization will write there.
Been doing that for a year or so.
It looks like at least these jobs are affected.
`jenkins-dsde-pipelines-run-automated-test`
`jenkins-picard-private-dev-deploy`
`jenkins-picard-private-staging-deploy`
`jenkins-zamboni-nightly-workflows-test`
`jenkins-zamboni-nightly-wgs-job`
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-316)
| 1.0 | Tex reports that some Jenkins jobs write into `~devopsci` on every run. - [~teixeira] reports via [~ferrara]:
can someone figure out why dsde-jenkins is writing `jenkins-dsde-pipelines-run-automated-test` `jenkins-gotc-arrays-dev-nightly-test` and several other directories to `devopsci`'s home directory daily?
That probably comes from this code.
https://github.com/broadinstitute/dsp-jenkins/blob/fd2fd5720925d68a0baa3bc2b86021702dd3edd0/src/main/groovy/org/broadinstitute/dspjenkins/CodeSnippets.groovy#L20
Looks like anything running with `picard` authorization will write there.
Been doing that for a year or so.
It looks like at least these jobs are affected.
`jenkins-dsde-pipelines-run-automated-test`
`jenkins-picard-private-dev-deploy`
`jenkins-picard-private-staging-deploy`
`jenkins-zamboni-nightly-workflows-test`
`jenkins-zamboni-nightly-wgs-job`
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-316)
| infrastructure | tex reports that some jenkins jobs write into devopsci on every run reports via can someone figure out why dsde jenkins is writing jenkins dsde pipelines run automated test jenkins gotc arrays dev nightly test and several other directories to devopsci s home directory daily that probably comes from this code looks like anything running with picard authorization will write there been doing that for a year or so it looks like at least these jobs are affected jenkins dsde pipelines run automated test jenkins picard private dev deploy jenkins picard private staging deploy jenkins zamboni nightly workflows test jenkins zamboni nightly wgs job ┆issue is synchronized with this | 1 |
149,317 | 5,716,327,984 | IssuesEvent | 2017-04-19 14:56:13 | Flexget/Flexget | https://api.github.com/repos/Flexget/Flexget | closed | After timeframe expires, best quality found is downloaded even if a better match to target is available | Enhancement Low priority | From http://flexget.com/ticket/3105
Consider a configuration such as
```
timeframe: 2 hours
target: 720p webdl
quality: 720p-1080p hdtv+
```
and the following scenario:
Flexget finds a suitable episode in two formats: 720p hdtv and 1080p hdtv, whether at the same time or within two hours of each other. Due to them not fulfilling the webdl requirement of the target quality it correctly doesn't download anything yet.
The timeframe then expires and there is still no 720p webrip available.
At this point, Flexget's current behaviour is to download the best quality it has seen, which in this case is the 1080p hdtv.
This may be what is wanted, but it also may not be. It's not what I want, and I would imagine that most users would agree that in this scenario the 720p hdtv would be more appropriate, since the resolution matches the target and the source, though not matching the target, is in the valid quality range. On the other hand the 1080p hdtv, which was downloaded, does not match the target resolution (though it is in the valid range) and the source also does not match the target (though again it is in the valid range).
I think the default behaviour should be to match as many of the target requirements as possible, while of course staying within the quality requirements.
If it seems not all users would want this new default behaviour, it could be configurable. It could either be a "choose closest match: yes or no" option, or it could be more sophisticated with the "target" option expanded. A user who wants "720p webrip" could somehow specify that the 720p part is important, and the webrip part not so important, or some other such combination.
liiight asked in a comment on the old ticket:
> if you don't want 1080p to begin with? why even include it in the original requirement? if that is not a valid option on your part, do not include it.
>
> i'm not saying i'm against adding a 'closest match' feature, that does sound interesting and doable, i'm just wondering.
The answer is that I do want 1080p if no 720p is available, but I prefer 720p. At present I have no way to tell Flexget that I _prefer_ a "worse" quality, or rather I do since that's what I'm saying in the "target", but it isn't taken into account when timeframe expires.
| 1.0 | After timeframe expires, best quality found is downloaded even if a better match to target is available - From http://flexget.com/ticket/3105
Consider a configuration such as
```
timeframe: 2 hours
target: 720p webdl
quality: 720p-1080p hdtv+
```
and the following scenario:
Flexget finds a suitable episode in two formats: 720p hdtv and 1080p hdtv, whether at the same time or within two hours of each other. Due to them not fulfilling the webdl requirement of the target quality it correctly doesn't download anything yet.
The timeframe then expires and there is still no 720p webrip available.
At this point, Flexget's current behaviour is to download the best quality it has seen, which in this case is the 1080p hdtv.
This may be what is wanted, but it also may not be. It's not what I want, and I would imagine that most users would agree that in this scenario the 720p hdtv would be more appropriate, since the resolution matches the target and the source, though not matching the target, is in the valid quality range. On the other hand the 1080p hdtv, which was downloaded, does not match the target resolution (though it is in the valid range) and the source also does not match the target (though again it is in the valid range).
I think the default behaviour should be to match as many of the target requirements as possible, while of course staying within the quality requirements.
If it seems not all users would want this new default behaviour, it could be configurable. It could either be a "choose closest match: yes or no" option, or it could be more sophisticated with the "target" option expanded. A user who wants "720p webrip" could somehow specify that the 720p part is important, and the webrip part not so important, or some other such combination.
liiight asked in a comment on the old ticket:
> if you don't want 1080p to begin with? why even include it in the original requirement? if that is not a valid option on your part, do not include it.
>
> i'm not saying i'm against adding a 'closest match' feature, that does sound interesting and doable, i'm just wondering.
The answer is that I do want 1080p if no 720p is available, but I prefer 720p. At present I have no way to tell Flexget that I _prefer_ a "worse" quality, or rather I do since that's what I'm saying in the "target", but it isn't taken into account when timeframe expires.
| non_infrastructure | after timeframe expires best quality found is downloaded even if a better match to target is available from consider a configuration such as timeframe hours target webdl quality hdtv and the following scenario flexget finds a suitable episode in two formats hdtv and hdtv whether at the same time or within two hours of each other due to them not fulfilling the webdl requirement of the target quality it correctly doesn t download anything yet the timeframe then expires and there is still no webrip available at this point flexget s current behaviour is to download the best quality it has seen which in this case is the hdtv this may be what is wanted but it also may not be it s not what i want and i would imagine that most users would agree that in this scenario the hdtv would be more appropriate since the resolution matches the target and the source though not matching the target is in the valid quality range on the other hand the hdtv which was downloaded does not match the target resolution though it is in the valid range and the source also does not match the target though again it is in the valid range i think the default behaviour should be to match as many of the target requirements as possible while of course staying within the quality requirements if it seems not all users would want this new default behaviour it could be configurable it could either be a choose closest match yes or no option or it could be more sophisticated with the target option expanded a user who wants webrip could somehow specify that the part is important and the webrip part not so important or some other such combination liiight asked in a comment on the old ticket if you don t want to begin with why even include it in the original requirement if that is not a valid option on your part do not include it i m not saying i m against adding a closest match feature that does sound interesting and doable i m just wondering the answer is that i do want if no is available but i prefer at present i have no way to tell flexget that i prefer a worse quality or rather i do since that s what i m saying in the target but it isn t taken into account when timeframe expires | 0 |
2,429 | 3,675,582,480 | IssuesEvent | 2016-02-23 00:15:51 | MozillaFoundation/plan | https://api.github.com/repos/MozillaFoundation/plan | opened | Finalize MLN H1 Production Work documents | Convening Innovation Learning Membership MLN Infrastructure p1 production Size: medium Status: on track | #### RASCI
R: @hannahkane @ldecoursy
A: @chrislarry33 @davidascher
S: @cassiemc @simonwex @adamlofting
C: CMS team
I: eventually everyone
#### Overview
We've pulled out the production work from the MLN H1 Strategy document, and are aiming to organize and present the work in a way that helps all production and program staff understand how our production work enables MLN success. We're also using these documents for planning purposes—specifically to estimate relative effort, to agree upon KPIs and goals, and eventually to prioritize/order the tasks.
#### Links
* [Narrative version](https://docs.google.com/document/d/18tg7DHiI5Ilqa2NuooOKx1fh3D69BEeHb9UHWvD6kJU/edit#heading=h.k63854e1w60v)
* [Granular, task-oriented version with estimates](https://docs.google.com/spreadsheets/d/1Prh4t_mUJLcsjr9Rxprzoy7TDKtxLISaC8jgYzbuscQ/edit#gid=0)
#### Heartbeat deliverables
- [ ] Final versions of documents, including relative effort estimates
- [ ] Plan to socialize with staff
| 1.0 | Finalize MLN H1 Production Work documents - #### RASCI
R: @hannahkane @ldecoursy
A: @chrislarry33 @davidascher
S: @cassiemc @simonwex @adamlofting
C: CMS team
I: eventually everyone
#### Overview
We've pulled out the production work from the MLN H1 Strategy document, and are aiming to organize and present the work in a way that helps all production and program staff understand how our production work enables MLN success. We're also using these documents for planning purposes—specifically to estimate relative effort, to agree upon KPIs and goals, and eventually to prioritize/order the tasks.
#### Links
* [Narrative version](https://docs.google.com/document/d/18tg7DHiI5Ilqa2NuooOKx1fh3D69BEeHb9UHWvD6kJU/edit#heading=h.k63854e1w60v)
* [Granular, task-oriented version with estimates](https://docs.google.com/spreadsheets/d/1Prh4t_mUJLcsjr9Rxprzoy7TDKtxLISaC8jgYzbuscQ/edit#gid=0)
#### Heartbeat deliverables
- [ ] Final versions of documents, including relative effort estimates
- [ ] Plan to socialize with staff
| infrastructure | finalize mln production work documents rasci r hannahkane ldecoursy a davidascher s cassiemc simonwex adamlofting c cms team i eventually everyone overview we ve pulled out the production work from the mln strategy document and are aiming to organize and present the work in a way that helps all production and program staff understand how our production work enables mln success we re also using these documents for planning purposes—specifically to estimate relative effort to agree upon kpis and goals and eventually to prioritize order the tasks links heartbeat deliverables final versions of documents including relative effort estimates plan to socialize with staff | 1 |
254,768 | 21,876,383,920 | IssuesEvent | 2022-05-19 10:31:34 | Tencent/bk-job | https://api.github.com/repos/Tencent/bk-job | closed | bugfix: 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 | kind/bug stage/test stage/uat | **版本/分支/tag**
3.4.4.0
**问题描述**
统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除
**如何复现**
创建脚本1,上线脚本1
创建模板1,引用脚本1
调试模板1,运行
创建脚本2,上线脚本2
编辑模板1,引用脚本1改为引用脚本2
查看脚本1引用次数,发现模板/执行方案引用次数为 0/1
尝试删除脚本1,提示被引用无法删除
**预期结果**
脚本1不再被引用后,查看脚本1引用次数,模板/执行方案引用次数为 0/0
应当能够正常删除脚本1 | 1.0 | bugfix: 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 - **版本/分支/tag**
3.4.4.0
**问题描述**
统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除
**如何复现**
创建脚本1,上线脚本1
创建模板1,引用脚本1
调试模板1,运行
创建脚本2,上线脚本2
编辑模板1,引用脚本1改为引用脚本2
查看脚本1引用次数,发现模板/执行方案引用次数为 0/1
尝试删除脚本1,提示被引用无法删除
**预期结果**
脚本1不再被引用后,查看脚本1引用次数,模板/执行方案引用次数为 0/0
应当能够正常删除脚本1 | non_infrastructure | bugfix 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 版本 分支 tag 问题描述 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 如何复现 , , ,运行 , , ,发现模板 执行方案引用次数为 ,提示被引用无法删除 预期结果 , ,模板 执行方案引用次数为 | 0 |
307,019 | 23,178,176,407 | IssuesEvent | 2022-07-31 18:33:34 | Vector35/binaryninja-api | https://api.github.com/repos/Vector35/binaryninja-api | closed | Documentation for BinaryView.parse_type_string cut-off mid sentence. | API documentation | The documentation for [BinaryView.parse_type_string](https://api.binary.ninja/binaryninja.binaryview-module.html?highlight=parse_type_string#binaryninja.binaryview.BinaryView.parse_type_string) is missing the latter half of the sentence. I assume it would describe that it allows references to types available in the bv.

| 1.0 | Documentation for BinaryView.parse_type_string cut-off mid sentence. - The documentation for [BinaryView.parse_type_string](https://api.binary.ninja/binaryninja.binaryview-module.html?highlight=parse_type_string#binaryninja.binaryview.BinaryView.parse_type_string) is missing the latter half of the sentence. I assume it would describe that it allows references to types available in the bv.

| non_infrastructure | documentation for binaryview parse type string cut off mid sentence the documentation for is missing the latter half of the sentence i assume it would describe that it allows references to types available in the bv | 0 |
29,350 | 23,935,685,902 | IssuesEvent | 2022-09-11 07:26:19 | CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g33 | closed | [Matching Service] Set-up Redis Container | Module/Infrastructure Status/High-Priority Type/Feature | ## Description
Set up the Redis container for queuing.
## Parent Task
- #38 | 1.0 | [Matching Service] Set-up Redis Container - ## Description
Set up the Redis container for queuing.
## Parent Task
- #38 | infrastructure | set up redis container description set up the redis container for queuing parent task | 1 |
13,355 | 8,198,056,341 | IssuesEvent | 2018-08-31 15:12:07 | angular/angular | https://api.github.com/repos/angular/angular | closed | *ngFor Performance Issues on IE 11 and Microsoft Edge | comp: core & compiler comp: performance freq4: critical severity3: broken type: bug/fix | The performance of *ngFor is very slow on IE as compared to Chrome, Firefox and Opera.
The below plunker loops through over 1500 items using *ngFor, it takes less than a second in other browsers whereas it takes several seconds in IE 11 and Microsoft Edge.
http://plnkr.co/edit/sEujClHmuCbrydIiYQYL?p=preview
The code is constantly calling isNan function in es6-shim.js.
The problem seems to be due to the reason that set and get functions of the polyfill are extremely slow (compared to their native counterparts) and take most of the time.
Please have a look at http://stackoverflow.com/questions/36570532/angular-2-performance-ie11-ngfor and the accepted answer.
This performance issue is also noticeable on forms with some dynamically populated drop down lists (8-10 drop downs).
| True | *ngFor Performance Issues on IE 11 and Microsoft Edge - The performance of *ngFor is very slow on IE as compared to Chrome, Firefox and Opera.
The below plunker loops through over 1500 items using *ngFor, it takes less than a second in other browsers whereas it takes several seconds in IE 11 and Microsoft Edge.
http://plnkr.co/edit/sEujClHmuCbrydIiYQYL?p=preview
The code is constantly calling isNan function in es6-shim.js.
The problem seems to be due to the reason that set and get functions of the polyfill are extremely slow (compared to their native counterparts) and take most of the time.
Please have a look at http://stackoverflow.com/questions/36570532/angular-2-performance-ie11-ngfor and the accepted answer.
This performance issue is also noticeable on forms with some dynamically populated drop down lists (8-10 drop downs).
| non_infrastructure | ngfor performance issues on ie and microsoft edge the performance of ngfor is very slow on ie as compared to chrome firefox and opera the below plunker loops through over items using ngfor it takes less than a second in other browsers whereas it takes several seconds in ie and microsoft edge the code is constantly calling isnan function in shim js the problem seems to be due to the reason that set and get functions of the polyfill are extremely slow compared to their native counterparts and take most of the time please have a look at and the accepted answer this performance issue is also noticeable on forms with some dynamically populated drop down lists drop downs | 0 |
202,928 | 23,123,248,783 | IssuesEvent | 2022-07-28 01:02:45 | lukebroganws/secDevLabs | https://api.github.com/repos/lukebroganws/secDevLabs | closed | CVE-2021-35065 (High) detected in multiple libraries - autoclosed | security vulnerability | ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.2.tgz</b>, <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- webpack-4.42.0.tgz
- watchpack-1.7.4.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.2.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.27.7.tgz (Root Library)
- chokidar-3.5.3.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/webpack-dev-server/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/fast-glob/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/webpack-dev-server/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a3/comment-killer/app/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a3/comment-killer/app/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- eslint-6.8.0.tgz
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-base/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.24.7.tgz (Root Library)
- chokidar-1.7.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/secDevLabs/commit/c66f2ad957084d73653f76b825cbe31fa86061ba">c66f2ad957084d73653f76b825cbe31fa86061ba</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | True | CVE-2021-35065 (High) detected in multiple libraries - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.2.tgz</b>, <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary>
<p>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- webpack-4.42.0.tgz
- watchpack-1.7.4.tgz
- watchpack-chokidar2-2.0.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.2.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.27.7.tgz (Root Library)
- chokidar-3.5.3.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-3.1.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent directory path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/webpack-dev-server/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/fast-glob/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/webpack-dev-server/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- chokidar-2.1.8.tgz
- :x: **glob-parent-3.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-5.1.1.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a3/comment-killer/app/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a3/comment-killer/app/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.4.3.tgz (Root Library)
- eslint-6.8.0.tgz
- :x: **glob-parent-5.1.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>glob-parent-2.0.0.tgz</b></p></summary>
<p>Strips glob magic from a string to provide the parent path</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p>
<p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/package.json</p>
<p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-base/node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- browser-sync-2.24.7.tgz (Root Library)
- chokidar-1.7.0.tgz
- :x: **glob-parent-2.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/lukebroganws/secDevLabs/commit/c66f2ad957084d73653f76b825cbe31fa86061ba">c66f2ad957084d73653f76b825cbe31fa86061ba</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p>
<p>Direct dependency fix Resolution (react-scripts): 5.0.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | non_infrastructure | cve high detected in multiple libraries autoclosed cve high severity vulnerability vulnerable libraries glob parent tgz glob parent tgz glob parent tgz glob parent tgz glob parent tgz glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href dependency hierarchy react scripts tgz root library webpack tgz watchpack tgz watchpack tgz chokidar tgz x glob parent tgz vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file owasp apps amarelo designs app static user package json path to vulnerable library owasp apps amarelo designs app static user node modules glob parent package json dependency hierarchy browser sync tgz root library chokidar tgz x glob parent tgz vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file owasp apps stegonography app static admin page package json path to vulnerable library owasp apps stegonography app static admin page node modules glob parent package json owasp apps comment killer app node modules webpack dev server node modules glob parent package json owasp apps comment killer app node modules fast glob node modules glob parent package json owasp apps streaming app frontend node modules webpack dev server node modules glob parent package json dependency hierarchy react scripts tgz root library webpack dev server tgz chokidar tgz x glob parent tgz vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file owasp apps comment killer app package json path to vulnerable library owasp apps comment killer app node modules glob parent package json owasp apps streaming app frontend node modules glob parent package json dependency hierarchy react scripts tgz root library eslint tgz x glob parent tgz vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file owasp apps amarelo designs app static package json path to vulnerable library owasp apps amarelo designs app static node modules glob parent package json owasp apps stegonography app static admin page node modules glob base node modules glob parent package json dependency hierarchy browser sync tgz root library chokidar tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent direct dependency fix resolution react scripts fix resolution glob parent direct dependency fix resolution react scripts fix resolution glob parent direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue | 0 |
123,521 | 26,268,074,173 | IssuesEvent | 2023-01-06 14:27:30 | FerretDB/FerretDB | https://api.github.com/repos/FerretDB/FerretDB | closed | Validate the client's TLS certificate | code/feature | ### What should be done?
It should be possible to configure FerretDB to validate the client's certificates against the given CA certificate and reject connections without valid certificates. See https://www.mongodb.com/docs/manual/tutorial/configure-ssl/#set-up-mongod-and-mongos-with-client-certificate-validation. Standard go library should be enough for that.
That should be a separate flag `--listen-tls-ca-file` that should be empty by default. It should be possible to use TLS as it is now without that flag.
### Where?
https://github.com/FerretDB/FerretDB/blob/main/internal/clientconn/listener.go
### Definition of Done
- unit, integration, or compatibility test should added (at least one of them should be possible to do);
- spot refactorings done;
- user documentation updated (see #1706).
| 1.0 | Validate the client's TLS certificate - ### What should be done?
It should be possible to configure FerretDB to validate the client's certificates against the given CA certificate and reject connections without valid certificates. See https://www.mongodb.com/docs/manual/tutorial/configure-ssl/#set-up-mongod-and-mongos-with-client-certificate-validation. Standard go library should be enough for that.
That should be a separate flag `--listen-tls-ca-file` that should be empty by default. It should be possible to use TLS as it is now without that flag.
### Where?
https://github.com/FerretDB/FerretDB/blob/main/internal/clientconn/listener.go
### Definition of Done
- unit, integration, or compatibility test should added (at least one of them should be possible to do);
- spot refactorings done;
- user documentation updated (see #1706).
| non_infrastructure | validate the client s tls certificate what should be done it should be possible to configure ferretdb to validate the client s certificates against the given ca certificate and reject connections without valid certificates see standard go library should be enough for that that should be a separate flag listen tls ca file that should be empty by default it should be possible to use tls as it is now without that flag where definition of done unit integration or compatibility test should added at least one of them should be possible to do spot refactorings done user documentation updated see | 0 |
311,953 | 23,411,236,467 | IssuesEvent | 2022-08-12 17:46:44 | FearlessSolutions/BCDA-DPC-B2D2-Veteran-Attribution-API | https://api.github.com/repos/FearlessSolutions/BCDA-DPC-B2D2-Veteran-Attribution-API | closed | Tracking multiple sources of truth | documentation | In most project kick-offs there can be multiple sources of truth that team members and stakeholders are each referring to as a guide. Sometimes its a SOO and a PWS, other times its OKRs or the RFQ. Sometimes it's a legacy backlog, other times its a greenfield project. This can cause swirl early as efforts are duplicated and at times conflicting.
There are a few sources of truth circling around in our own project, and I want to ensure we are aligned. There is the original hackmd readme I made as a placeholder (actually for the purpose of tracking multiple sources of truth). That was uploaded to this repo, making the github readme the new source of truth for me. Once the tech challenge was issued, I added the official README to those notes and merged here in github.
There is also a google drive, and a [one note](https://icfonline-my.sharepoint.com/personal/43256_icf_com/_layouts/15/Doc.aspx?sourcedoc={f52a3ebf-cc20-443b-a6dc-d756f416a704}&action=edit&wd=target%28Documentation.one%7Cf10b348a-7bf9-4396-97e2-cd30fc97d193%2FUseful%20Links%7Ce94b6c08-b1e4-465f-bd67-e5df7fcba050%2F%29&wdorigin=NavigationUrl). Due to the short turnaround, I am generally for using whatever tools work best for you, and want to raise awareness to be intentional about tracking github and ensuring anything we have in our own notes is accounted for here in github as well. | 1.0 | Tracking multiple sources of truth - In most project kick-offs there can be multiple sources of truth that team members and stakeholders are each referring to as a guide. Sometimes its a SOO and a PWS, other times its OKRs or the RFQ. Sometimes it's a legacy backlog, other times its a greenfield project. This can cause swirl early as efforts are duplicated and at times conflicting.
There are a few sources of truth circling around in our own project, and I want to ensure we are aligned. There is the original hackmd readme I made as a placeholder (actually for the purpose of tracking multiple sources of truth). That was uploaded to this repo, making the github readme the new source of truth for me. Once the tech challenge was issued, I added the official README to those notes and merged here in github.
There is also a google drive, and a [one note](https://icfonline-my.sharepoint.com/personal/43256_icf_com/_layouts/15/Doc.aspx?sourcedoc={f52a3ebf-cc20-443b-a6dc-d756f416a704}&action=edit&wd=target%28Documentation.one%7Cf10b348a-7bf9-4396-97e2-cd30fc97d193%2FUseful%20Links%7Ce94b6c08-b1e4-465f-bd67-e5df7fcba050%2F%29&wdorigin=NavigationUrl). Due to the short turnaround, I am generally for using whatever tools work best for you, and want to raise awareness to be intentional about tracking github and ensuring anything we have in our own notes is accounted for here in github as well. | non_infrastructure | tracking multiple sources of truth in most project kick offs there can be multiple sources of truth that team members and stakeholders are each referring to as a guide sometimes its a soo and a pws other times its okrs or the rfq sometimes it s a legacy backlog other times its a greenfield project this can cause swirl early as efforts are duplicated and at times conflicting there are a few sources of truth circling around in our own project and i want to ensure we are aligned there is the original hackmd readme i made as a placeholder actually for the purpose of tracking multiple sources of truth that was uploaded to this repo making the github readme the new source of truth for me once the tech challenge was issued i added the official readme to those notes and merged here in github there is also a google drive and a due to the short turnaround i am generally for using whatever tools work best for you and want to raise awareness to be intentional about tracking github and ensuring anything we have in our own notes is accounted for here in github as well | 0 |
19,630 | 13,337,796,931 | IssuesEvent | 2020-08-28 09:52:08 | OCR-D/core | https://api.github.com/repos/OCR-D/core | opened | More idiomatic and consistent unit tests | enhancement help wanted infrastructure | When we began OCR-D/core, we targeted only `unittests` with some enhancements over time in `tests/base.py`. To run the tests, we're actually using pytest, though.
We currently have ~300 tests which use various mechanisms to achieve the same task, often with nested context Managers (`with TemporaryDirectory() as tempdir:` and such).
@M3ssman suggested embracing pytest's features to a larger extent. For example, pytest offers decorators to achieve much of what we currently implement ourselves. For clarity and maintenance, we should refactor the test suite in a more consistent style.
This is obviously nothing that directly improves the user experience, so it is a bit of an unthankful but important task. At least for new tests, we should strive to make use of pytest's features and update the existing tests by and by. | 1.0 | More idiomatic and consistent unit tests - When we began OCR-D/core, we targeted only `unittests` with some enhancements over time in `tests/base.py`. To run the tests, we're actually using pytest, though.
We currently have ~300 tests which use various mechanisms to achieve the same task, often with nested context Managers (`with TemporaryDirectory() as tempdir:` and such).
@M3ssman suggested embracing pytest's features to a larger extent. For example, pytest offers decorators to achieve much of what we currently implement ourselves. For clarity and maintenance, we should refactor the test suite in a more consistent style.
This is obviously nothing that directly improves the user experience, so it is a bit of an unthankful but important task. At least for new tests, we should strive to make use of pytest's features and update the existing tests by and by. | infrastructure | more idiomatic and consistent unit tests when we began ocr d core we targeted only unittests with some enhancements over time in tests base py to run the tests we re actually using pytest though we currently have tests which use various mechanisms to achieve the same task often with nested context managers with temporarydirectory as tempdir and such suggested embracing pytest s features to a larger extent for example pytest offers decorators to achieve much of what we currently implement ourselves for clarity and maintenance we should refactor the test suite in a more consistent style this is obviously nothing that directly improves the user experience so it is a bit of an unthankful but important task at least for new tests we should strive to make use of pytest s features and update the existing tests by and by | 1 |
32,317 | 26,621,786,699 | IssuesEvent | 2023-01-24 11:48:28 | Ladder99/ladder99 | https://api.github.com/repos/Ladder99/ladder99 | closed | Filter to just `l99` services in `l99 status` command | enhancement infrastructure | ob: l99 status is docker ps and it lists all running containers, even those that are not related to l99. IMO it should output only the state of l99 containers. I should output at least the l99 is running or not, ideally each container if it is running or not (even if it is not running), optionally in JSON format.
bb: Yeah, I had included non-ladder99 containers as there were things like fanuc_driver which are currently run separately, but yeah maybe better to filter those out. Adding to list.
cm: L99 status should be limited to containers in the l99 stack. I would not worry about fanuc-driver. We can make it part of the stack eventually.
ob: Why fanuc-driver would be a non-l99 container? IMO if fanuc-driver is used by l99, why wouldn’t it start with --project ladder99 in docker-compose up? Shouldn’t it be set up during l99 setup (l99 start)?
| 1.0 | Filter to just `l99` services in `l99 status` command - ob: l99 status is docker ps and it lists all running containers, even those that are not related to l99. IMO it should output only the state of l99 containers. I should output at least the l99 is running or not, ideally each container if it is running or not (even if it is not running), optionally in JSON format.
bb: Yeah, I had included non-ladder99 containers as there were things like fanuc_driver which are currently run separately, but yeah maybe better to filter those out. Adding to list.
cm: L99 status should be limited to containers in the l99 stack. I would not worry about fanuc-driver. We can make it part of the stack eventually.
ob: Why fanuc-driver would be a non-l99 container? IMO if fanuc-driver is used by l99, why wouldn’t it start with --project ladder99 in docker-compose up? Shouldn’t it be set up during l99 setup (l99 start)?
| infrastructure | filter to just services in status command ob status is docker ps and it lists all running containers even those that are not related to imo it should output only the state of containers i should output at least the is running or not ideally each container if it is running or not even if it is not running optionally in json format bb yeah i had included non containers as there were things like fanuc driver which are currently run separately but yeah maybe better to filter those out adding to list cm status should be limited to containers in the stack i would not worry about fanuc driver we can make it part of the stack eventually ob why fanuc driver would be a non container imo if fanuc driver is used by why wouldn’t it start with project in docker compose up shouldn’t it be set up during setup start | 1 |
293,948 | 25,335,618,100 | IssuesEvent | 2022-11-18 16:34:52 | anoma/namada-trusted-setup | https://api.github.com/repos/anoma/namada-trusted-setup | closed | Repo misc fixes | refactor testing | # Fix
- [x] `cargo check` warnings
- [x] review `masp` import
- [x] all `FIXME`s
- [x] Add a phantom cohort with no tokens at the end of the ceremony to leave enough time to the participants of the last cohort for their contributions
- [x] How to provide the `contributors.json` file to the frontend, two ways:
- Keep the coordinator running forever while shutting down the `join_queue` endpoint (this would not require a phantom cohort)
- Shut the coordinator down and publish the `contributors.json` file to S3
- [x] Remove unused dependencies
- [x] Produce final release tag
- [x] Upload last contribution verified to S3?
- [x] Fix messages and docs
- [x] Rename `phase1` to `phase2`
# Test
- [x] contribution on `AnotherMachine`
- [x] token blacklisting
- [x] Namada keypair generation
- [ ] restart of coordinator?
# Review
- [x] values for the env variables to set
- [x] state of the Coordinator server
- [x] state of the S3 server
- [x] state of the Amazon Parameter Store | 1.0 | Repo misc fixes - # Fix
- [x] `cargo check` warnings
- [x] review `masp` import
- [x] all `FIXME`s
- [x] Add a phantom cohort with no tokens at the end of the ceremony to leave enough time to the participants of the last cohort for their contributions
- [x] How to provide the `contributors.json` file to the frontend, two ways:
- Keep the coordinator running forever while shutting down the `join_queue` endpoint (this would not require a phantom cohort)
- Shut the coordinator down and publish the `contributors.json` file to S3
- [x] Remove unused dependencies
- [x] Produce final release tag
- [x] Upload last contribution verified to S3?
- [x] Fix messages and docs
- [x] Rename `phase1` to `phase2`
# Test
- [x] contribution on `AnotherMachine`
- [x] token blacklisting
- [x] Namada keypair generation
- [ ] restart of coordinator?
# Review
- [x] values for the env variables to set
- [x] state of the Coordinator server
- [x] state of the S3 server
- [x] state of the Amazon Parameter Store | non_infrastructure | repo misc fixes fix cargo check warnings review masp import all fixme s add a phantom cohort with no tokens at the end of the ceremony to leave enough time to the participants of the last cohort for their contributions how to provide the contributors json file to the frontend two ways keep the coordinator running forever while shutting down the join queue endpoint this would not require a phantom cohort shut the coordinator down and publish the contributors json file to remove unused dependencies produce final release tag upload last contribution verified to fix messages and docs rename to test contribution on anothermachine token blacklisting namada keypair generation restart of coordinator review values for the env variables to set state of the coordinator server state of the server state of the amazon parameter store | 0 |
9,185 | 7,857,726,704 | IssuesEvent | 2018-06-21 11:48:28 | hzi-braunschweig/SORMAS-Project | https://api.github.com/repos/hzi-braunschweig/SORMAS-Project | opened | Property fields can depend on other fields | 102XX Infrastructure sormas-app | - [ ] add "app:dependencyParentId" and "app:dependyParentValue" attributes
- [ ] field registers value listener at parent
- [ ] hide field when parent does not have the appropriate value | 1.0 | Property fields can depend on other fields - - [ ] add "app:dependencyParentId" and "app:dependyParentValue" attributes
- [ ] field registers value listener at parent
- [ ] hide field when parent does not have the appropriate value | infrastructure | property fields can depend on other fields add app dependencyparentid and app dependyparentvalue attributes field registers value listener at parent hide field when parent does not have the appropriate value | 1 |
23,473 | 16,239,146,288 | IssuesEvent | 2021-05-07 07:11:10 | microsoft/WindowsTemplateStudio | https://api.github.com/repos/microsoft/WindowsTemplateStudio | closed | Reorganize internal projecttemplate folder structure | Can Close Out Soon Infrastructure | With the addition of ClassLibraries and RuntimeComponents project templates the number of project templates in the [ProjectTemplates](https://github.com/microsoft/WindowsTemplateStudio/tree/dev/code/src/ProjectTemplates) folder is getting to a point where it's difficult to distinguish the different entry points.
We propose to add subfolders for the different platforms and languages to make it easier to find the correct vstemplate.
This is only an internal folder structure change and won't have any impact on the generated code or how templates show in the New Project Dialog.
Proposed folder structure is:
- ProjectTemplates
- WinUI
- CS
- Cpp
- WPF
- CS
- Uwp
- CS
- VB | 1.0 | Reorganize internal projecttemplate folder structure - With the addition of ClassLibraries and RuntimeComponents project templates the number of project templates in the [ProjectTemplates](https://github.com/microsoft/WindowsTemplateStudio/tree/dev/code/src/ProjectTemplates) folder is getting to a point where it's difficult to distinguish the different entry points.
We propose to add subfolders for the different platforms and languages to make it easier to find the correct vstemplate.
This is only an internal folder structure change and won't have any impact on the generated code or how templates show in the New Project Dialog.
Proposed folder structure is:
- ProjectTemplates
- WinUI
- CS
- Cpp
- WPF
- CS
- Uwp
- CS
- VB | infrastructure | reorganize internal projecttemplate folder structure with the addition of classlibraries and runtimecomponents project templates the number of project templates in the folder is getting to a point where it s difficult to distinguish the different entry points we propose to add subfolders for the different platforms and languages to make it easier to find the correct vstemplate this is only an internal folder structure change and won t have any impact on the generated code or how templates show in the new project dialog proposed folder structure is projecttemplates winui cs cpp wpf cs uwp cs vb | 1 |
25,477 | 4,158,871,172 | IssuesEvent | 2016-06-17 06:11:51 | NagiosEnterprises/ncpa | https://api.github.com/repos/NagiosEnterprises/ncpa | closed | Add upgrade option to installer | Feature Request Needs Testing | We should add an upgrade option when someone uses a newer version installer... this doesn't necessarily have to be shown to the user but we should run something to stop the active/passive listeners so that the install can overwrite their current version and then start the new version for them... would be nice if it could somehow know though so you wouldn't have to re-enter the config variables. | 1.0 | Add upgrade option to installer - We should add an upgrade option when someone uses a newer version installer... this doesn't necessarily have to be shown to the user but we should run something to stop the active/passive listeners so that the install can overwrite their current version and then start the new version for them... would be nice if it could somehow know though so you wouldn't have to re-enter the config variables. | non_infrastructure | add upgrade option to installer we should add an upgrade option when someone uses a newer version installer this doesn t necessarily have to be shown to the user but we should run something to stop the active passive listeners so that the install can overwrite their current version and then start the new version for them would be nice if it could somehow know though so you wouldn t have to re enter the config variables | 0 |
356,812 | 25,176,265,762 | IssuesEvent | 2022-11-11 09:31:57 | Tex-Tang/pe | https://api.github.com/repos/Tex-Tang/pe | opened | Out-of-date screenshot in UG | type.DocumentationBug severity.Low | `2 Class-ify's Application Window` section in UG

`4.3.2` section in UG

<!--session: 1668153999061-e63af65f-10dc-4b2f-bfdc-f2a47bc4eed4-->
<!--Version: Web v3.4.4--> | 1.0 | Out-of-date screenshot in UG - `2 Class-ify's Application Window` section in UG

`4.3.2` section in UG

<!--session: 1668153999061-e63af65f-10dc-4b2f-bfdc-f2a47bc4eed4-->
<!--Version: Web v3.4.4--> | non_infrastructure | out of date screenshot in ug class ify s application window section in ug section in ug | 0 |
13,841 | 10,481,995,999 | IssuesEvent | 2019-09-24 10:55:22 | shopsys/shopsys | https://api.github.com/repos/shopsys/shopsys | closed | Build application in kubernetes should be simplified | Infrastructure | <!--- Title should contain short general summary what is the issue about -->
### What is happening
In deploy/build scripts is often used yq to replace placeholders or some values with the proper one right for the environment.
This replacement often looks like
```sh
# Set environment variables to container and initContainer for Google Cloud Storage connection - set it to null
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[0].value 'bucket-name'
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[1].value 'project-id'
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[0].value 'bucket-name'
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[1].value 'project-id'
```
Which transforms (shortened example)
```yaml
spec:
template:
initContainers:
- name: copy-source-codes-to-volume
- name: initialize-database
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: ~
- name: GOOGLE_CLOUD_PROJECT_ID
value: ~
containers:
- name: php-fpm
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: ~
- name: GOOGLE_CLOUD_PROJECT_ID
value: ~
```
into
```yaml
spec:
template:
initContainers:
- name: copy-source-codes-to-volume
- name: initialize-database
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: bucket-name
- name: GOOGLE_CLOUD_PROJECT_ID
value: project-id
containers:
- name: php-fpm
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: bucket-name
- name: GOOGLE_CLOUD_PROJECT_ID
value: project-id
```
This is violation of DRY and most of all very fragile solution.
It's really easy to prepend new env variable as the @PetrHeinz correctly pointed out.
Maybe we could use Kustomize variables (see [vars documentation](https://github.com/kubernetes-sigs/kustomize/blob/master/docs/fields.md#vars) or [an example using vars](https://github.com/kubernetes-sigs/kustomize/blob/master/examples/wordpress/README.md)) to not repeat ourselves like this.
<!--- What are preconditions and your setting e.g. Shopsys version or in case of Docker issues your operating system -->
<!--- Best is to include steps to reproduce this issue if relevant-->
<!--- Tell us what happens -->
### Expected result
Scripts should be less error-prone and simplified. | 1.0 | Build application in kubernetes should be simplified - <!--- Title should contain short general summary what is the issue about -->
### What is happening
In deploy/build scripts is often used yq to replace placeholders or some values with the proper one right for the environment.
This replacement often looks like
```sh
# Set environment variables to container and initContainer for Google Cloud Storage connection - set it to null
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[0].value 'bucket-name'
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[1].value 'project-id'
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[0].value 'bucket-name'
yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[1].value 'project-id'
```
Which transforms (shortened example)
```yaml
spec:
template:
initContainers:
- name: copy-source-codes-to-volume
- name: initialize-database
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: ~
- name: GOOGLE_CLOUD_PROJECT_ID
value: ~
containers:
- name: php-fpm
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: ~
- name: GOOGLE_CLOUD_PROJECT_ID
value: ~
```
into
```yaml
spec:
template:
initContainers:
- name: copy-source-codes-to-volume
- name: initialize-database
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: bucket-name
- name: GOOGLE_CLOUD_PROJECT_ID
value: project-id
containers:
- name: php-fpm
env:
- name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME
value: bucket-name
- name: GOOGLE_CLOUD_PROJECT_ID
value: project-id
```
This is violation of DRY and most of all very fragile solution.
It's really easy to prepend new env variable as the @PetrHeinz correctly pointed out.
Maybe we could use Kustomize variables (see [vars documentation](https://github.com/kubernetes-sigs/kustomize/blob/master/docs/fields.md#vars) or [an example using vars](https://github.com/kubernetes-sigs/kustomize/blob/master/examples/wordpress/README.md)) to not repeat ourselves like this.
<!--- What are preconditions and your setting e.g. Shopsys version or in case of Docker issues your operating system -->
<!--- Best is to include steps to reproduce this issue if relevant-->
<!--- Tell us what happens -->
### Expected result
Scripts should be less error-prone and simplified. | infrastructure | build application in kubernetes should be simplified what is happening in deploy build scripts is often used yq to replace placeholders or some values with the proper one right for the environment this replacement often looks like sh set environment variables to container and initcontainer for google cloud storage connection set it to null yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec containers env value bucket name yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec containers env value project id yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec initcontainers env value bucket name yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec initcontainers env value project id which transforms shortened example yaml spec template initcontainers name copy source codes to volume name initialize database env name google cloud storage bucket name value name google cloud project id value containers name php fpm env name google cloud storage bucket name value name google cloud project id value into yaml spec template initcontainers name copy source codes to volume name initialize database env name google cloud storage bucket name value bucket name name google cloud project id value project id containers name php fpm env name google cloud storage bucket name value bucket name name google cloud project id value project id this is violation of dry and most of all very fragile solution it s really easy to prepend new env variable as the petrheinz correctly pointed out maybe we could use kustomize variables see or to not repeat ourselves like this expected result scripts should be less error prone and simplified | 1 |
24,920 | 17,909,872,241 | IssuesEvent | 2021-09-09 02:43:51 | FullScreenShenanigans/EightBittr | https://api.github.com/repos/FullScreenShenanigans/EightBittr | closed | Switch to fixed package versions in Lerna (instead of independent) | status: accepting prs type: infrastructure | https://github.com/lerna/lerna#how-it-works
It'll be easier to just let Lerna do its magic than manage it manually. | 1.0 | Switch to fixed package versions in Lerna (instead of independent) - https://github.com/lerna/lerna#how-it-works
It'll be easier to just let Lerna do its magic than manage it manually. | infrastructure | switch to fixed package versions in lerna instead of independent it ll be easier to just let lerna do its magic than manage it manually | 1 |
113,497 | 24,429,426,940 | IssuesEvent | 2022-10-06 06:57:01 | WordPress/openverse-frontend | https://api.github.com/repos/WordPress/openverse-frontend | closed | Recent Searches feature (Combobox, localstorage, and more) | 🟨 priority: medium 🏁 status: ready for work ✨ goal: improvement 💻 aspect: code | ## Description
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
We plan to add a new feature to show the user their 'recent searches', in a box below the searchbar. I originally thought this would be a combobox feature, but we shouldn't actually need to implement that until we're doing search _suggestions_, where the user input determines the contents of the box. In this use case we're just showing the nth-most recent searches.
> A preview of the recent searches

Figma Link: https://www.figma.com/file/w60dl1XPUvSaRncv1Utmnb/Openverse-Releases?node-id=133%3A29056
## Dev Tasks
- [ ] Implement the suggestions box
- [ ] Popover for the desktop (above `lg`)
- [ ] Full page modal with a search bar for mobile screens
- [x] Create a composable for storing recent searches
- [x] Store user recent searches in indexedDb (see https://github.com/WordPress/openverse-frontend/issues/479#issuecomment-1252941676 for justification)
- [x] Enable clearing of the recent searches in indexedDb
## Implementation questions
- Do we need a way to let users opt-out of this? (maybe this can wait)
- How many items do we want to show? (4 in the mockup currently) I imagine the search bar will populate and the search will be instantly submitted.
- What will happen when the user clicks on one of the recent searches?
## Design decisions (cc @panchovm)
- [x] Add this to the Figma component library (it's currently only in the page mockups)
- [x] Implement a 'clear' button (will update the mockups for this)
- [x] What will the hover/focus styles look like for the choices? | 1.0 | Recent Searches feature (Combobox, localstorage, and more) - ## Description
<!-- Describe a problem solved by this feature; or delete the section entirely. -->
We plan to add a new feature to show the user their 'recent searches', in a box below the searchbar. I originally thought this would be a combobox feature, but we shouldn't actually need to implement that until we're doing search _suggestions_, where the user input determines the contents of the box. In this use case we're just showing the nth-most recent searches.
> A preview of the recent searches

Figma Link: https://www.figma.com/file/w60dl1XPUvSaRncv1Utmnb/Openverse-Releases?node-id=133%3A29056
## Dev Tasks
- [ ] Implement the suggestions box
- [ ] Popover for the desktop (above `lg`)
- [ ] Full page modal with a search bar for mobile screens
- [x] Create a composable for storing recent searches
- [x] Store user recent searches in indexedDb (see https://github.com/WordPress/openverse-frontend/issues/479#issuecomment-1252941676 for justification)
- [x] Enable clearing of the recent searches in indexedDb
## Implementation questions
- Do we need a way to let users opt-out of this? (maybe this can wait)
- How many items do we want to show? (4 in the mockup currently) I imagine the search bar will populate and the search will be instantly submitted.
- What will happen when the user clicks on one of the recent searches?
## Design decisions (cc @panchovm)
- [x] Add this to the Figma component library (it's currently only in the page mockups)
- [x] Implement a 'clear' button (will update the mockups for this)
- [x] What will the hover/focus styles look like for the choices? | non_infrastructure | recent searches feature combobox localstorage and more description we plan to add a new feature to show the user their recent searches in a box below the searchbar i originally thought this would be a combobox feature but we shouldn t actually need to implement that until we re doing search suggestions where the user input determines the contents of the box in this use case we re just showing the nth most recent searches a preview of the recent searches figma link dev tasks implement the suggestions box popover for the desktop above lg full page modal with a search bar for mobile screens create a composable for storing recent searches store user recent searches in indexeddb see for justification enable clearing of the recent searches in indexeddb implementation questions do we need a way to let users opt out of this maybe this can wait how many items do we want to show in the mockup currently i imagine the search bar will populate and the search will be instantly submitted what will happen when the user clicks on one of the recent searches design decisions cc panchovm add this to the figma component library it s currently only in the page mockups implement a clear button will update the mockups for this what will the hover focus styles look like for the choices | 0 |
6,344 | 6,342,442,438 | IssuesEvent | 2017-07-27 15:29:54 | openshiftio/appdev-documentation | https://api.github.com/repos/openshiftio/appdev-documentation | closed | Release tagging does not work on UNIX machines | Bug Follow-up Has PR Infrastructure Peer Review SME Review | Following up on #391:
As reported by @rhoads-zach in the documentation Slack channel, the `$REPO_HOME/scripts/tagRelease.sh` script does not work on machines with OpenBSD or MacOS because the `date` binary there does not have the `--rfc-3339` flag. I assume this problem also affects the `ci/Jenkinsfile` script as I have used the same invocation there.
I believe the solution is to use `date "+%Y-%m-%d"` instead. I have verified it works on OpenBSD and Fedora. @rhoads-zach or someone with an Apple PC, can you please verify it works on MacOS as well? | 1.0 | Release tagging does not work on UNIX machines - Following up on #391:
As reported by @rhoads-zach in the documentation Slack channel, the `$REPO_HOME/scripts/tagRelease.sh` script does not work on machines with OpenBSD or MacOS because the `date` binary there does not have the `--rfc-3339` flag. I assume this problem also affects the `ci/Jenkinsfile` script as I have used the same invocation there.
I believe the solution is to use `date "+%Y-%m-%d"` instead. I have verified it works on OpenBSD and Fedora. @rhoads-zach or someone with an Apple PC, can you please verify it works on MacOS as well? | infrastructure | release tagging does not work on unix machines following up on as reported by rhoads zach in the documentation slack channel the repo home scripts tagrelease sh script does not work on machines with openbsd or macos because the date binary there does not have the rfc flag i assume this problem also affects the ci jenkinsfile script as i have used the same invocation there i believe the solution is to use date y m d instead i have verified it works on openbsd and fedora rhoads zach or someone with an apple pc can you please verify it works on macos as well | 1 |
21,442 | 14,569,585,692 | IssuesEvent | 2020-12-17 13:17:54 | odrotbohm/moduliths | https://api.github.com/repos/odrotbohm/moduliths | closed | Upgrade to ArchUnit 0.15.0 | in: infrastructure type: task | It now ships with generics analysis, i.e. it detects dependencies stemming from those and generally detects more dependencies previously undiscovered.
https://github.com/TNG/ArchUnit/releases/tag/v0.15.0 | 1.0 | Upgrade to ArchUnit 0.15.0 - It now ships with generics analysis, i.e. it detects dependencies stemming from those and generally detects more dependencies previously undiscovered.
https://github.com/TNG/ArchUnit/releases/tag/v0.15.0 | infrastructure | upgrade to archunit it now ships with generics analysis i e it detects dependencies stemming from those and generally detects more dependencies previously undiscovered | 1 |
2,884 | 3,951,349,600 | IssuesEvent | 2016-04-29 00:31:25 | AtlasOfLivingAustralia/data-management | https://api.github.com/repos/AtlasOfLivingAustralia/data-management | closed | Improved download interface | Downloads Infrastructure | interface to allow field selection from full set of fields - index, QA and additional fields
saving of default set of fields
allow addition of un-sensitised values where appropriate permissions exist
placing export file in an accessible location
functionality to facilitate automation | 1.0 | Improved download interface - interface to allow field selection from full set of fields - index, QA and additional fields
saving of default set of fields
allow addition of un-sensitised values where appropriate permissions exist
placing export file in an accessible location
functionality to facilitate automation | infrastructure | improved download interface interface to allow field selection from full set of fields index qa and additional fields saving of default set of fields allow addition of un sensitised values where appropriate permissions exist placing export file in an accessible location functionality to facilitate automation | 1 |
25,426 | 18,732,923,550 | IssuesEvent | 2021-11-04 01:12:18 | t3kt/raytk | https://api.github.com/repos/t3kt/raytk | closed | Improve the shared shader library system | cleanup infrastructure | Current all the shared libraries need to be available within the shaderLibraries tox. It should be possible to allow them from anywhere, and have some form of deduping. That way an OP could host its own copy of the shared library. | 1.0 | Improve the shared shader library system - Current all the shared libraries need to be available within the shaderLibraries tox. It should be possible to allow them from anywhere, and have some form of deduping. That way an OP could host its own copy of the shared library. | infrastructure | improve the shared shader library system current all the shared libraries need to be available within the shaderlibraries tox it should be possible to allow them from anywhere and have some form of deduping that way an op could host its own copy of the shared library | 1 |
796,519 | 28,116,531,402 | IssuesEvent | 2023-03-31 11:11:25 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | opened | [Priest] Dispersion cannot be used when affected by Maloriak's Flash freeze. | Class: Priest Status: Duplicate Priority: High Raid: Blackwing Descent | Confirmed on PTR and Prod.
Priests specced into shadow cannot use Dispersion when under the effect of Flash Freeze
Flash Freeze is considered a stun and should mean Shadow Priests can use Dispersion.


| 1.0 | [Priest] Dispersion cannot be used when affected by Maloriak's Flash freeze. - Confirmed on PTR and Prod.
Priests specced into shadow cannot use Dispersion when under the effect of Flash Freeze
Flash Freeze is considered a stun and should mean Shadow Priests can use Dispersion.


| non_infrastructure | dispersion cannot be used when affected by maloriak s flash freeze confirmed on ptr and prod priests specced into shadow cannot use dispersion when under the effect of flash freeze flash freeze is considered a stun and should mean shadow priests can use dispersion | 0 |
24,608 | 17,467,186,111 | IssuesEvent | 2021-08-06 18:41:30 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Adhere to the package support policy by building and packaging supported frameworks only | area-Infrastructure-libraries breaking-change needs-breaking-change-doc-created | Continuation of https://github.com/dotnet/runtime/issues/47530
The runtime repository currently produces 102 packages from libraries of which 25 of those contain assets which will be unsupported when we release .NET 6 in November 2021. The list of unsupported assets includes netstandard1.x assets as those apply to unsupported application frameworks only as in any case where a netstandard1.x asset is present, a netstandard2.0 (and in most cases a net461 asset) exists as well which allows to run on .NET Framework 4.6.1 or .NET Core 3.1 and above.
The reasons for trimming out unsupported assets are:
- Simplified build graph and reduced build times
- Less restore size as the NS1.x dependency graph doesn't need to be restored
- Package validation costs
- Fat packages
- Maintenance of code bases which are not tested anymore.
Based on these disadvantages we will remove the remaining unsupported assets from packages in dotnet/runtime.
## List of packages grouped by TargetFrameworkIdentifier
### NETStandard,Version=1.x -> .NETStandard,Version=2.0
- Microsoft.Extensions.HostFactoryResolver.Sources
- System.Collections.Immutable
- System.Diagnostics.DiagnosticSource
- System.IO.Packaging
- System.Numerics.Tensors
- System.Reflection.Context
- System.Reflection.Metadata
- System.Runtime.CompilerServices.Unsafe
- System.Threading.Channels
- System.Threading.Tasks.Dataflow
PR: https://github.com/dotnet/runtime/pull/53283
### NETCoreApp,Version=2.x|3.0 -> .NETCoreApp,Version=3.1
- Microsoft.Win32.SystemEvents
- System.Data.Odbc
- System.Diagnostics.EventLog
- System.ComponentModel.Composition
- System.Diagnostics.PerformanceCounter
- System.DirectoryServices
- System.DirectoryServices.AccountManagement
- System.DirectoryServices.Protocols
- System.Management
- System.Runtime.CompilerServices.Unsafe
- System.Security.Principal.Windows
- System.Speech
- System.Text.Encoding.CodePages
- System.Threading.Channels
PR: https://github.com/dotnet/runtime/pull/53358
### NETFramework,Version=4.5|4.5.1|4.6 -> .NETFramework.Version=4.6.1
- System.Diagnostics.DiagnosticSource
- System.Runtime.CompilerServices.Unsafe
- System.IO.Packaging
- System.ComponentModel.Composition.Registration
- System.ComponentModel.Composition
- System.DirectoryServices
- System.DirectoryServices.AccountManagement
- System.DirectoryServices.Protocols
- System.IO.Packaging
- System.Management
- System.Reflection.Context
- System.Runtime.Caching
- System.Speech
PR: https://github.com/dotnet/runtime/pull/53319
cc @danmoseley @ericstj @terrajobst @Anipik | 1.0 | Adhere to the package support policy by building and packaging supported frameworks only - Continuation of https://github.com/dotnet/runtime/issues/47530
The runtime repository currently produces 102 packages from libraries of which 25 of those contain assets which will be unsupported when we release .NET 6 in November 2021. The list of unsupported assets includes netstandard1.x assets as those apply to unsupported application frameworks only as in any case where a netstandard1.x asset is present, a netstandard2.0 (and in most cases a net461 asset) exists as well which allows to run on .NET Framework 4.6.1 or .NET Core 3.1 and above.
The reasons for trimming out unsupported assets are:
- Simplified build graph and reduced build times
- Less restore size as the NS1.x dependency graph doesn't need to be restored
- Package validation costs
- Fat packages
- Maintenance of code bases which are not tested anymore.
Based on these disadvantages we will remove the remaining unsupported assets from packages in dotnet/runtime.
## List of packages grouped by TargetFrameworkIdentifier
### NETStandard,Version=1.x -> .NETStandard,Version=2.0
- Microsoft.Extensions.HostFactoryResolver.Sources
- System.Collections.Immutable
- System.Diagnostics.DiagnosticSource
- System.IO.Packaging
- System.Numerics.Tensors
- System.Reflection.Context
- System.Reflection.Metadata
- System.Runtime.CompilerServices.Unsafe
- System.Threading.Channels
- System.Threading.Tasks.Dataflow
PR: https://github.com/dotnet/runtime/pull/53283
### NETCoreApp,Version=2.x|3.0 -> .NETCoreApp,Version=3.1
- Microsoft.Win32.SystemEvents
- System.Data.Odbc
- System.Diagnostics.EventLog
- System.ComponentModel.Composition
- System.Diagnostics.PerformanceCounter
- System.DirectoryServices
- System.DirectoryServices.AccountManagement
- System.DirectoryServices.Protocols
- System.Management
- System.Runtime.CompilerServices.Unsafe
- System.Security.Principal.Windows
- System.Speech
- System.Text.Encoding.CodePages
- System.Threading.Channels
PR: https://github.com/dotnet/runtime/pull/53358
### NETFramework,Version=4.5|4.5.1|4.6 -> .NETFramework.Version=4.6.1
- System.Diagnostics.DiagnosticSource
- System.Runtime.CompilerServices.Unsafe
- System.IO.Packaging
- System.ComponentModel.Composition.Registration
- System.ComponentModel.Composition
- System.DirectoryServices
- System.DirectoryServices.AccountManagement
- System.DirectoryServices.Protocols
- System.IO.Packaging
- System.Management
- System.Reflection.Context
- System.Runtime.Caching
- System.Speech
PR: https://github.com/dotnet/runtime/pull/53319
cc @danmoseley @ericstj @terrajobst @Anipik | infrastructure | adhere to the package support policy by building and packaging supported frameworks only continuation of the runtime repository currently produces packages from libraries of which of those contain assets which will be unsupported when we release net in november the list of unsupported assets includes x assets as those apply to unsupported application frameworks only as in any case where a x asset is present a and in most cases a asset exists as well which allows to run on net framework or net core and above the reasons for trimming out unsupported assets are simplified build graph and reduced build times less restore size as the x dependency graph doesn t need to be restored package validation costs fat packages maintenance of code bases which are not tested anymore based on these disadvantages we will remove the remaining unsupported assets from packages in dotnet runtime list of packages grouped by targetframeworkidentifier netstandard version x netstandard version microsoft extensions hostfactoryresolver sources system collections immutable system diagnostics diagnosticsource system io packaging system numerics tensors system reflection context system reflection metadata system runtime compilerservices unsafe system threading channels system threading tasks dataflow pr netcoreapp version x netcoreapp version microsoft systemevents system data odbc system diagnostics eventlog system componentmodel composition system diagnostics performancecounter system directoryservices system directoryservices accountmanagement system directoryservices protocols system management system runtime compilerservices unsafe system security principal windows system speech system text encoding codepages system threading channels pr netframework version netframework version system diagnostics diagnosticsource system runtime compilerservices unsafe system io packaging system componentmodel composition registration system componentmodel composition system directoryservices system directoryservices accountmanagement system directoryservices protocols system io packaging system management system reflection context system runtime caching system speech pr cc danmoseley ericstj terrajobst anipik | 1 |
375 | 2,667,007,321 | IssuesEvent | 2015-03-22 04:39:33 | deis/deis | https://api.github.com/repos/deis/deis | opened | bug(test): during the test if the user app is stopped it continues running | requires-manual-testing test-infrastructure tests | instead of fail.
Steps to reproduce:
- during the test enter in any node.
- stop the app.
- the suite will continue to run instead of throwing an error
Why is relevant?
In case of an error in the deployed app it will restarted until the next output:
```
Mar 22 04:31:15 deis-01 systemd[1]: configsample_v9.cmd.1.service holdoff time over, scheduling restart.
```
then the service will stop and the suite will continue running | 1.0 | bug(test): during the test if the user app is stopped it continues running - instead of fail.
Steps to reproduce:
- during the test enter in any node.
- stop the app.
- the suite will continue to run instead of throwing an error
Why is relevant?
In case of an error in the deployed app it will restarted until the next output:
```
Mar 22 04:31:15 deis-01 systemd[1]: configsample_v9.cmd.1.service holdoff time over, scheduling restart.
```
then the service will stop and the suite will continue running | infrastructure | bug test during the test if the user app is stopped it continues running instead of fail steps to reproduce during the test enter in any node stop the app the suite will continue to run instead of throwing an error why is relevant in case of an error in the deployed app it will restarted until the next output mar deis systemd configsample cmd service holdoff time over scheduling restart then the service will stop and the suite will continue running | 1 |
26,274 | 19,898,992,258 | IssuesEvent | 2022-01-25 04:41:54 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | closed | Merge OppiaTestRule and AccessibilityTestRule | Type: Improvement Where: Infrastructure Priority: Important Status: Not started Project stage: not started | Merge `OppiaTestRule` and `AccessibilityTestRule` in `OppiaTestRule`
Once done, run all test cases on Espresso and post screenshot displaying all tests are passing on Pixel 3 XL. | 1.0 | Merge OppiaTestRule and AccessibilityTestRule - Merge `OppiaTestRule` and `AccessibilityTestRule` in `OppiaTestRule`
Once done, run all test cases on Espresso and post screenshot displaying all tests are passing on Pixel 3 XL. | infrastructure | merge oppiatestrule and accessibilitytestrule merge oppiatestrule and accessibilitytestrule in oppiatestrule once done run all test cases on espresso and post screenshot displaying all tests are passing on pixel xl | 1 |
21,508 | 14,599,829,618 | IssuesEvent | 2020-12-21 05:23:32 | php-coder/mystamps | https://api.github.com/repos/php-coder/mystamps | opened | Install Prometheus and Grafana for monitoring | area/infrastructure | - [ ] install Prometheus
- [ ] install Grafana
- [ ] install node-exporter
- [ ] add ansible roles | 1.0 | Install Prometheus and Grafana for monitoring - - [ ] install Prometheus
- [ ] install Grafana
- [ ] install node-exporter
- [ ] add ansible roles | infrastructure | install prometheus and grafana for monitoring install prometheus install grafana install node exporter add ansible roles | 1 |
183,228 | 6,678,622,058 | IssuesEvent | 2017-10-05 14:47:36 | AdguardTeam/AdguardFilters | https://api.github.com/repos/AdguardTeam/AdguardFilters | closed | http://m.azonline.de/ - sm widget, banners and ad leftovers | Annoyance Incorrect Blocking iOS Browser Priority: Low Resolved Social Widget | [//]: # (***You can delete or ignore strings starting with "[//]:" They will not be visible either way.)
***Description***:
* **Current behaviour**: http://m.azonline.de/NRW/2980745-Prozesse-Steuererklaerung-verspaetet-abgegeben-Berater-verurteilt
[//]: # (Substitute this line with a description of the problem)
[//]: # (Replace %screenshot_url% below with a link to the screenshot of the problem. Also, you can paste image from clipboard instead. It will be automatically loaded.)
<details><summary>Screenshot:</summary>


</details><br/>
***System configuration***
**Filters:**
- ID=0 Name="User filter" Version=1.0.0.0 Enabled=YES
- ID=2 Name="English filter" Version=1.1.30.47 Enabled=YES
- ID=3 Name="Spyware filter" Version=1.0.14.42 Enabled=YES
- ID=4 Name="Social media filter" Version=1.0.23.9 Enabled=YES
- ID=14 Name="Annoyances filter" Version=1.0.18.20 Enabled=YES
- ID=6 Name="German filter" Version=1.0.36.61 Enabled=YES
- ID=11 Name="Mobile Ads filter" Version=1.0.9.64 Enabled=YES
- ID=12 Name="Safari filter" Version=1.0.3.60 Enabled=YES
[//]: # (Substitute this line with the list of your active filters, separated by commas)
[//]: # (Please enter the correct values for your case to the table below)
Information | Value
--- | ---
Operating system: | iOS
Browser: | Safari
AdGuard version: | 1.3.1
Simplified filters (iOS only) | Off
AdGuard DNS: | None
Helpdesk ID (if exists): | 1480792
[//]: # (This template is meant for missed ad/false positive reports, for other type of reports edit it accordingly)
[//]: # (If this is a crash report, include the crashlog with https://gist.github.com/)
| 1.0 | http://m.azonline.de/ - sm widget, banners and ad leftovers - [//]: # (***You can delete or ignore strings starting with "[//]:" They will not be visible either way.)
***Description***:
* **Current behaviour**: http://m.azonline.de/NRW/2980745-Prozesse-Steuererklaerung-verspaetet-abgegeben-Berater-verurteilt
[//]: # (Substitute this line with a description of the problem)
[//]: # (Replace %screenshot_url% below with a link to the screenshot of the problem. Also, you can paste image from clipboard instead. It will be automatically loaded.)
<details><summary>Screenshot:</summary>


</details><br/>
***System configuration***
**Filters:**
- ID=0 Name="User filter" Version=1.0.0.0 Enabled=YES
- ID=2 Name="English filter" Version=1.1.30.47 Enabled=YES
- ID=3 Name="Spyware filter" Version=1.0.14.42 Enabled=YES
- ID=4 Name="Social media filter" Version=1.0.23.9 Enabled=YES
- ID=14 Name="Annoyances filter" Version=1.0.18.20 Enabled=YES
- ID=6 Name="German filter" Version=1.0.36.61 Enabled=YES
- ID=11 Name="Mobile Ads filter" Version=1.0.9.64 Enabled=YES
- ID=12 Name="Safari filter" Version=1.0.3.60 Enabled=YES
[//]: # (Substitute this line with the list of your active filters, separated by commas)
[//]: # (Please enter the correct values for your case to the table below)
Information | Value
--- | ---
Operating system: | iOS
Browser: | Safari
AdGuard version: | 1.3.1
Simplified filters (iOS only) | Off
AdGuard DNS: | None
Helpdesk ID (if exists): | 1480792
[//]: # (This template is meant for missed ad/false positive reports, for other type of reports edit it accordingly)
[//]: # (If this is a crash report, include the crashlog with https://gist.github.com/)
| non_infrastructure | sm widget banners and ad leftovers you can delete or ignore strings starting with they will not be visible either way description current behaviour substitute this line with a description of the problem replace screenshot url below with a link to the screenshot of the problem also you can paste image from clipboard instead it will be automatically loaded screenshot system configuration filters id name user filter version enabled yes id name english filter version enabled yes id name spyware filter version enabled yes id name social media filter version enabled yes id name annoyances filter version enabled yes id name german filter version enabled yes id name mobile ads filter version enabled yes id name safari filter version enabled yes substitute this line with the list of your active filters separated by commas please enter the correct values for your case to the table below information value operating system ios browser safari adguard version simplified filters ios only off adguard dns none helpdesk id if exists this template is meant for missed ad false positive reports for other type of reports edit it accordingly if this is a crash report include the crashlog with | 0 |
3,416 | 4,309,476,831 | IssuesEvent | 2016-07-21 16:07:06 | KhronosGroup/glslang | https://api.github.com/repos/KhronosGroup/glslang | closed | rationalizing warning levels | enhancement Infrastructure | A small change to CMakeLists.txt for discussion: #393.
By default the g++ build was not checking for some classes of warnings that the MSVC build did, for example around unused variables and function parameters. It was easy to accidentally add code like that on a Linux environment. The above WIP PR adds those warnings for g++, and cleans up a few instances in the code where they trigger.
I took a swing at adding the set that clang accepts, but I'm not set up to test this. The auto-build was happy :smile:, but maybe someone knows of a better set. If so, I can drop them in.
I was unable to find a g++ warning to detect the MSVC *"warning C4267: 'initializing' : conversion from 'size_t' to 'unsigned int', possible loss of data"*, which triggers on code such as this:
`int foo = myvector.size();`
Other thoughts?
| 1.0 | rationalizing warning levels - A small change to CMakeLists.txt for discussion: #393.
By default the g++ build was not checking for some classes of warnings that the MSVC build did, for example around unused variables and function parameters. It was easy to accidentally add code like that on a Linux environment. The above WIP PR adds those warnings for g++, and cleans up a few instances in the code where they trigger.
I took a swing at adding the set that clang accepts, but I'm not set up to test this. The auto-build was happy :smile:, but maybe someone knows of a better set. If so, I can drop them in.
I was unable to find a g++ warning to detect the MSVC *"warning C4267: 'initializing' : conversion from 'size_t' to 'unsigned int', possible loss of data"*, which triggers on code such as this:
`int foo = myvector.size();`
Other thoughts?
| infrastructure | rationalizing warning levels a small change to cmakelists txt for discussion by default the g build was not checking for some classes of warnings that the msvc build did for example around unused variables and function parameters it was easy to accidentally add code like that on a linux environment the above wip pr adds those warnings for g and cleans up a few instances in the code where they trigger i took a swing at adding the set that clang accepts but i m not set up to test this the auto build was happy smile but maybe someone knows of a better set if so i can drop them in i was unable to find a g warning to detect the msvc warning initializing conversion from size t to unsigned int possible loss of data which triggers on code such as this int foo myvector size other thoughts | 1 |
23,826 | 16,606,937,459 | IssuesEvent | 2021-06-02 05:54:52 | asciidoctor/asciidoctor | https://api.github.com/repos/asciidoctor/asciidoctor | closed | Import source and build for default stylesheet into this repository | infrastructure | The default stylesheet (for the HTML output) is currently generated by the asciidoctor-stylesheet-factory project. However, that project uses a build which relies on unmaintained software. It's also arduous and obscure to have to make changes to files into another project and import the generated result into this repository to be committed. The whole process will be a lot simpler if that build is located in this project.
To avoid introducing unanticipated changes to the stylesheet, we'll need to work backwards from the minified/distributed end result until we get to something that's maintainable. This issue focuses on the first stage in that migration, which is described below:
1. Unminify data/stylesheets/asciidoctor-default.css (this will become the editable source file)
* The source file will be maintained in plain CSS with some help from PostCSS
2. Set up and configure a npm-based build that uses postcss + cssnano to produce the same or similar minified result as before
3. Use the build to generate data/stylesheets/asciidoctor-default.min.css
4. Update the processor to read the default styles from the data/stylesheets/asciidoctor-default.min.css file instead of data/stylesheets/asciidoctor-default.css file
* If this doesn't work out, the source file could be named data/styelsheets/asciidoctor-default.src.css instead
5. Configure the gemspec to omit the stylesheet source file from the gem
We'll probably want to avoid committing the minified file to the repository and generating it on demand for the test suite and release, but we should do that in a later stage. | 1.0 | Import source and build for default stylesheet into this repository - The default stylesheet (for the HTML output) is currently generated by the asciidoctor-stylesheet-factory project. However, that project uses a build which relies on unmaintained software. It's also arduous and obscure to have to make changes to files into another project and import the generated result into this repository to be committed. The whole process will be a lot simpler if that build is located in this project.
To avoid introducing unanticipated changes to the stylesheet, we'll need to work backwards from the minified/distributed end result until we get to something that's maintainable. This issue focuses on the first stage in that migration, which is described below:
1. Unminify data/stylesheets/asciidoctor-default.css (this will become the editable source file)
* The source file will be maintained in plain CSS with some help from PostCSS
2. Set up and configure a npm-based build that uses postcss + cssnano to produce the same or similar minified result as before
3. Use the build to generate data/stylesheets/asciidoctor-default.min.css
4. Update the processor to read the default styles from the data/stylesheets/asciidoctor-default.min.css file instead of data/stylesheets/asciidoctor-default.css file
* If this doesn't work out, the source file could be named data/styelsheets/asciidoctor-default.src.css instead
5. Configure the gemspec to omit the stylesheet source file from the gem
We'll probably want to avoid committing the minified file to the repository and generating it on demand for the test suite and release, but we should do that in a later stage. | infrastructure | import source and build for default stylesheet into this repository the default stylesheet for the html output is currently generated by the asciidoctor stylesheet factory project however that project uses a build which relies on unmaintained software it s also arduous and obscure to have to make changes to files into another project and import the generated result into this repository to be committed the whole process will be a lot simpler if that build is located in this project to avoid introducing unanticipated changes to the stylesheet we ll need to work backwards from the minified distributed end result until we get to something that s maintainable this issue focuses on the first stage in that migration which is described below unminify data stylesheets asciidoctor default css this will become the editable source file the source file will be maintained in plain css with some help from postcss set up and configure a npm based build that uses postcss cssnano to produce the same or similar minified result as before use the build to generate data stylesheets asciidoctor default min css update the processor to read the default styles from the data stylesheets asciidoctor default min css file instead of data stylesheets asciidoctor default css file if this doesn t work out the source file could be named data styelsheets asciidoctor default src css instead configure the gemspec to omit the stylesheet source file from the gem we ll probably want to avoid committing the minified file to the repository and generating it on demand for the test suite and release but we should do that in a later stage | 1 |
20,838 | 14,194,438,085 | IssuesEvent | 2020-11-15 03:47:13 | opencv/opencv | https://api.github.com/repos/opencv/opencv | opened | OSX framework: enable testing of Objective-C/Swift bindings | category: infrastructure category: swift/objc bindings platform: ios/osx | It would be nice to enable OSX framework testing for OpenCV 4.x through `--run_tests` from `osx/build_framework.py` on MacOSX.
Currently I see this error message:
```
Testing failed:
OpenCVTestTests:
Could not find module 'opencv2' for architecture 'x86_64'; found: x86_64-apple-macos
Testing cancelled because the build failed.
```
@komakai Could you please confirm that this flow should work?
Used this configuration of build environment:
```
$ sw_vers
ProductName: Mac OS X
ProductVersion: 10.14.6
BuildVersion: 18G87
$ clang --version
Apple LLVM version 10.0.1 (clang-1001.0.46.4)
Target: x86_64-apple-darwin18.7.0
Thread model: posix
InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
$ swiftc --version
Apple Swift version 5.0.1 (swiftlang-1001.0.82.4 clang-1001.0.46.5)
Target: x86_64-apple-darwin18.7.0
$ cmake --version
cmake version 3.15.2
MACOSX_DEPLOYMENT_TARGET = 10.12 (default from OpenCV's packaging scripts)
```
---
Extended logs: **TBD** (on GitHub) | 1.0 | OSX framework: enable testing of Objective-C/Swift bindings - It would be nice to enable OSX framework testing for OpenCV 4.x through `--run_tests` from `osx/build_framework.py` on MacOSX.
Currently I see this error message:
```
Testing failed:
OpenCVTestTests:
Could not find module 'opencv2' for architecture 'x86_64'; found: x86_64-apple-macos
Testing cancelled because the build failed.
```
@komakai Could you please confirm that this flow should work?
Used this configuration of build environment:
```
$ sw_vers
ProductName: Mac OS X
ProductVersion: 10.14.6
BuildVersion: 18G87
$ clang --version
Apple LLVM version 10.0.1 (clang-1001.0.46.4)
Target: x86_64-apple-darwin18.7.0
Thread model: posix
InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin
$ swiftc --version
Apple Swift version 5.0.1 (swiftlang-1001.0.82.4 clang-1001.0.46.5)
Target: x86_64-apple-darwin18.7.0
$ cmake --version
cmake version 3.15.2
MACOSX_DEPLOYMENT_TARGET = 10.12 (default from OpenCV's packaging scripts)
```
---
Extended logs: **TBD** (on GitHub) | infrastructure | osx framework enable testing of objective c swift bindings it would be nice to enable osx framework testing for opencv x through run tests from osx build framework py on macosx currently i see this error message testing failed opencvtesttests could not find module for architecture found apple macos testing cancelled because the build failed komakai could you please confirm that this flow should work used this configuration of build environment sw vers productname mac os x productversion buildversion clang version apple llvm version clang target apple thread model posix installeddir applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin swiftc version apple swift version swiftlang clang target apple cmake version cmake version macosx deployment target default from opencv s packaging scripts extended logs tbd on github | 1 |
29,410 | 23,990,666,314 | IssuesEvent | 2022-09-14 00:33:27 | mwaskom/seaborn | https://api.github.com/repos/mwaskom/seaborn | closed | Look into jupytext for managing doc source notebooks | docs infrastructure | Would be nice to be able to edit prose, etc. in plain text, along with not having to track all the cell metadata, etc. | 1.0 | Look into jupytext for managing doc source notebooks - Would be nice to be able to edit prose, etc. in plain text, along with not having to track all the cell metadata, etc. | infrastructure | look into jupytext for managing doc source notebooks would be nice to be able to edit prose etc in plain text along with not having to track all the cell metadata etc | 1 |
142,832 | 5,477,734,210 | IssuesEvent | 2017-03-12 11:37:44 | dhowe/ReadersJS | https://api.github.com/repos/dhowe/ReadersJS | closed | Need perigrams for all texts | priority: High | Currently it seems we only have them for 'the Image': image-perigrams.txt | 1.0 | Need perigrams for all texts - Currently it seems we only have them for 'the Image': image-perigrams.txt | non_infrastructure | need perigrams for all texts currently it seems we only have them for the image image perigrams txt | 0 |
27,014 | 21,006,030,276 | IssuesEvent | 2022-03-29 22:46:13 | coq/coq | https://api.github.com/repos/coq/coq | opened | How to run coq_makefile from the development environment when compiling with dune? | kind: infrastructure | We have `dune exec -- dev/shim/coqide-prelude` to run CoqIDE in the dev environment and a similar command to run `coqtop`. How about the various utilities such as `coq_makefile`? Should I use the old make for this? And if so, how do I switch? Just do a clean?
It will pretty painful to make any improvements to `coq_makefile` if you have to re-install Coq for each compile. (Not that I have a list of changes I want to make in-hand at the moment, but that may change shortly.) | 1.0 | How to run coq_makefile from the development environment when compiling with dune? - We have `dune exec -- dev/shim/coqide-prelude` to run CoqIDE in the dev environment and a similar command to run `coqtop`. How about the various utilities such as `coq_makefile`? Should I use the old make for this? And if so, how do I switch? Just do a clean?
It will pretty painful to make any improvements to `coq_makefile` if you have to re-install Coq for each compile. (Not that I have a list of changes I want to make in-hand at the moment, but that may change shortly.) | infrastructure | how to run coq makefile from the development environment when compiling with dune we have dune exec dev shim coqide prelude to run coqide in the dev environment and a similar command to run coqtop how about the various utilities such as coq makefile should i use the old make for this and if so how do i switch just do a clean it will pretty painful to make any improvements to coq makefile if you have to re install coq for each compile not that i have a list of changes i want to make in hand at the moment but that may change shortly | 1 |
12,578 | 9,855,067,933 | IssuesEvent | 2019-06-19 18:26:20 | patternfly/patternfly-org | https://api.github.com/repos/patternfly/patternfly-org | closed | Improve patternfly-react and patternfly-next src inclusion in the repository | PF4 DEV infrastructure | @dana do you mind filling in Zack with details? | 1.0 | Improve patternfly-react and patternfly-next src inclusion in the repository - @dana do you mind filling in Zack with details? | infrastructure | improve patternfly react and patternfly next src inclusion in the repository dana do you mind filling in zack with details | 1 |
162,849 | 25,603,946,487 | IssuesEvent | 2022-12-01 23:05:33 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | DSN: Update VBA MVP description with items from content audit | Design ⭐️ Facilities Regional office | ## Description
- #10234
- [MVP description](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/facilities/regional-offices/design/README.md)
## Acceptance Criteria
- [x] MVP criteria doc updated
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [x] `⭐️ Facilities`
- [ ] `⭐️ User support`
| 1.0 | DSN: Update VBA MVP description with items from content audit - ## Description
- #10234
- [MVP description](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/facilities/regional-offices/design/README.md)
## Acceptance Criteria
- [x] MVP criteria doc updated
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [x] `⭐️ Facilities`
- [ ] `⭐️ User support`
| non_infrastructure | dsn update vba mvp description with items from content audit description acceptance criteria mvp criteria doc updated cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support | 0 |
27,772 | 22,332,419,715 | IssuesEvent | 2022-06-14 15:30:04 | acts-project/acts | https://api.github.com/repos/acts-project/acts | closed | nodeps build fail | Bug Infrastructure | The linux-nodeps build failed in #1190 with the message:
```
FAILED: thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build thirdparty/boost/boost-prefix/src/boost/stage/lib/libboost_unit_test_framework.so.1.71.0
cd /__w/acts/acts/build/thirdparty/boost/boost-prefix/src/boost && /usr/bin/cmake -P /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake && /usr/bin/cmake -E touch /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build
CMake Error at /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake:49 (message):
Command failed: 1
'./b2' 'variant=release' 'threading=multi' 'runtime-link=shared' 'link=shared' 'cxxflags=-std=c++17 -fPIC -O3' '--with-test'
See also
/__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-*.log
```
I quickly surveyed other recent PR's but didn't see the failure. I cannot reproduce locally either. Still I'll leave this issue up until we're sure this was transient.
| 1.0 | nodeps build fail - The linux-nodeps build failed in #1190 with the message:
```
FAILED: thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build thirdparty/boost/boost-prefix/src/boost/stage/lib/libboost_unit_test_framework.so.1.71.0
cd /__w/acts/acts/build/thirdparty/boost/boost-prefix/src/boost && /usr/bin/cmake -P /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake && /usr/bin/cmake -E touch /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build
CMake Error at /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake:49 (message):
Command failed: 1
'./b2' 'variant=release' 'threading=multi' 'runtime-link=shared' 'link=shared' 'cxxflags=-std=c++17 -fPIC -O3' '--with-test'
See also
/__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-*.log
```
I quickly surveyed other recent PR's but didn't see the failure. I cannot reproduce locally either. Still I'll leave this issue up until we're sure this was transient.
| infrastructure | nodeps build fail the linux nodeps build failed in with the message failed thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build thirdparty boost boost prefix src boost stage lib libboost unit test framework so cd w acts acts build thirdparty boost boost prefix src boost usr bin cmake p w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build release cmake usr bin cmake e touch w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build cmake error at w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build release cmake message command failed variant release threading multi runtime link shared link shared cxxflags std c fpic with test see also w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build log i quickly surveyed other recent pr s but didn t see the failure i cannot reproduce locally either still i ll leave this issue up until we re sure this was transient | 1 |
1,833 | 3,392,861,312 | IssuesEvent | 2015-11-30 21:27:54 | catapult-project/catapult | https://api.github.com/repos/catapult-project/catapult | opened | Non-chromium committers can't give commitable lgtms | Infrastructure | https://codereview.chromium.org/1469203002/#msg19
How do we fix that for @martijncoenen and our other contribs? | 1.0 | Non-chromium committers can't give commitable lgtms - https://codereview.chromium.org/1469203002/#msg19
How do we fix that for @martijncoenen and our other contribs? | infrastructure | non chromium committers can t give commitable lgtms how do we fix that for martijncoenen and our other contribs | 1 |
53,154 | 13,128,685,301 | IssuesEvent | 2020-08-06 12:43:59 | atc0005/send2teams | https://api.github.com/repos/atc0005/send2teams | opened | Use Docker-based GitHub Actions Workflows | CI builds dependencies enhancement linting | Swap out the GitHub Actions used for CI work with custom Docker containers from the atc0005/go-ci project.
refs atc0005/todo#22
| 1.0 | Use Docker-based GitHub Actions Workflows - Swap out the GitHub Actions used for CI work with custom Docker containers from the atc0005/go-ci project.
refs atc0005/todo#22
| non_infrastructure | use docker based github actions workflows swap out the github actions used for ci work with custom docker containers from the go ci project refs todo | 0 |
138,400 | 11,200,552,758 | IssuesEvent | 2020-01-03 22:10:15 | GTNewHorizons/NewHorizons | https://api.github.com/repos/GTNewHorizons/NewHorizons | closed | Singleblock miner bugs | FixedInDev need to be tested | #### Which modpack version are you using?
2.0.8.0pre w/updated GT++
#
#### If in multiplayer; On which server does this happen?
#
#### What did you try to do, and what did you expect to happen?
Crafted (in 2.0.8.0pre) and placed down a HV singleblock miner. Expected it to default to the maximum radius.
#
#### What happened instead? (Attach screenshots if needed)
Defaults to minimum radius (3x3) instead of maximum (49x49), so I had to mash it with a screwdriver a bunch.
#
#### What do you suggest instead/what changes do you propose?
Set the default radius to the maximum, and make the WAILA tooltip for the miner display the current operating radius. | 1.0 | Singleblock miner bugs - #### Which modpack version are you using?
2.0.8.0pre w/updated GT++
#
#### If in multiplayer; On which server does this happen?
#
#### What did you try to do, and what did you expect to happen?
Crafted (in 2.0.8.0pre) and placed down a HV singleblock miner. Expected it to default to the maximum radius.
#
#### What happened instead? (Attach screenshots if needed)
Defaults to minimum radius (3x3) instead of maximum (49x49), so I had to mash it with a screwdriver a bunch.
#
#### What do you suggest instead/what changes do you propose?
Set the default radius to the maximum, and make the WAILA tooltip for the miner display the current operating radius. | non_infrastructure | singleblock miner bugs which modpack version are you using w updated gt if in multiplayer on which server does this happen what did you try to do and what did you expect to happen crafted in and placed down a hv singleblock miner expected it to default to the maximum radius what happened instead attach screenshots if needed defaults to minimum radius instead of maximum so i had to mash it with a screwdriver a bunch what do you suggest instead what changes do you propose set the default radius to the maximum and make the waila tooltip for the miner display the current operating radius | 0 |
67,861 | 9,103,220,012 | IssuesEvent | 2019-02-20 15:27:18 | petermr/tigr2ess | https://api.github.com/repos/petermr/tigr2ess | closed | "Test run of ami-dictionary for the ami-jar - ami20190219." | documentation enhancement | #### Status of test run of ami-dictionary for the ami-jar - ami20190219 - successful.
#### Input - wikipedia link e.g - https://en.wikipedia.org/wiki/Ocimum.
#### Output - xml formatted dictionary file.
#### Run time log.
```
ambarish123@ubuntu:~$ ami-dictionary create --input https://en.wikipedia.org/wiki/Ocimum --informat wikipage --dictionary Ocimum --outformat xml --directory OcimumDir/
Generic values (AMIDictionaryTool)
================================
basename null
cproject
ctree
cTreeList null
dryrun false
excludeBase null
excludeTrees null
file types []
forceMake false
includeBase null
includeTrees null
log4j
logfile null
verbose 0
Specific values (AMIDictionaryTool)
================================
dataCols null
dictionary [Ocimum, --outformat, xml]
dictionaryTop OcimumDir/
href null
hrefCols null
input https://en.wikipedia.org/wiki/Ocimum
informat wikipage
dictInformat null
linkCol null
log4j null
nameCol null
operation create
outformats [xml]
splitCol ,
termCol null
terms null
wikiLinks [wikipedia, wikidata]
0 [main] DEBUG org.contentmine.ami.tools.AMIDictionaryTool - extracting hyperlinks
............!....!!.....!.!......!...!!.!.!!.!!!!!!!!!!!!!!!!!.!!!!.!!!!!.!!!!!!!!!!!!!!!!!.!!!!!!!.....!.!..........!...!. !WP !WP !WP ++++++ !WP ++ !WP ++++++++ !WP !WP +++++++ !WP ++++ !WP !WP + !WP !WP + !WP + !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP + !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP ++++++++++++++++
Missing wikipedia: :
"Genus: Ocimum L."; Isodon inflexus; Kew World Checklist of Selected Plant Families; Mosla scabra; O. campechianum; O. centraliafricanum; O. tenuiflorum; Ocimeae;
```
#### Output file -
[OcimumDir.tar.gz](https://github.com/petermr/tigr2ess/files/2881314/OcimumDir.tar.gz)
| 1.0 | "Test run of ami-dictionary for the ami-jar - ami20190219." - #### Status of test run of ami-dictionary for the ami-jar - ami20190219 - successful.
#### Input - wikipedia link e.g - https://en.wikipedia.org/wiki/Ocimum.
#### Output - xml formatted dictionary file.
#### Run time log.
```
ambarish123@ubuntu:~$ ami-dictionary create --input https://en.wikipedia.org/wiki/Ocimum --informat wikipage --dictionary Ocimum --outformat xml --directory OcimumDir/
Generic values (AMIDictionaryTool)
================================
basename null
cproject
ctree
cTreeList null
dryrun false
excludeBase null
excludeTrees null
file types []
forceMake false
includeBase null
includeTrees null
log4j
logfile null
verbose 0
Specific values (AMIDictionaryTool)
================================
dataCols null
dictionary [Ocimum, --outformat, xml]
dictionaryTop OcimumDir/
href null
hrefCols null
input https://en.wikipedia.org/wiki/Ocimum
informat wikipage
dictInformat null
linkCol null
log4j null
nameCol null
operation create
outformats [xml]
splitCol ,
termCol null
terms null
wikiLinks [wikipedia, wikidata]
0 [main] DEBUG org.contentmine.ami.tools.AMIDictionaryTool - extracting hyperlinks
............!....!!.....!.!......!...!!.!.!!.!!!!!!!!!!!!!!!!!.!!!!.!!!!!.!!!!!!!!!!!!!!!!!.!!!!!!!.....!.!..........!...!. !WP !WP !WP ++++++ !WP ++ !WP ++++++++ !WP !WP +++++++ !WP ++++ !WP !WP + !WP !WP + !WP + !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP + !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP ++++++++++++++++
Missing wikipedia: :
"Genus: Ocimum L."; Isodon inflexus; Kew World Checklist of Selected Plant Families; Mosla scabra; O. campechianum; O. centraliafricanum; O. tenuiflorum; Ocimeae;
```
#### Output file -
[OcimumDir.tar.gz](https://github.com/petermr/tigr2ess/files/2881314/OcimumDir.tar.gz)
| non_infrastructure | test run of ami dictionary for the ami jar status of test run of ami dictionary for the ami jar successful input wikipedia link e g output xml formatted dictionary file run time log ubuntu ami dictionary create input informat wikipage dictionary ocimum outformat xml directory ocimumdir generic values amidictionarytool basename null cproject ctree ctreelist null dryrun false excludebase null excludetrees null file types forcemake false includebase null includetrees null logfile null verbose specific values amidictionarytool datacols null dictionary dictionarytop ocimumdir href null hrefcols null input informat wikipage dictinformat null linkcol null null namecol null operation create outformats splitcol termcol null terms null wikilinks debug org contentmine ami tools amidictionarytool extracting hyperlinks wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp missing wikipedia genus ocimum l isodon inflexus kew world checklist of selected plant families mosla scabra o campechianum o centraliafricanum o tenuiflorum ocimeae output file | 0 |
463,148 | 13,260,508,331 | IssuesEvent | 2020-08-20 18:17:55 | radical-cybertools/radical.saga | https://api.github.com/repos/radical-cybertools/radical.saga | closed | bsub option for NVME in the LSF scheduler | comp:lsf priority:medium topic:resouce:summit type:enhancement | On Summit, Burst Buffer can be activated via the job scheduler option like:
```
-allow_flags NVME
```
There are three options that I can think of now:
- make this option as a default on Summit, then all reserved/allocated nodes have NVMe storage devices mounted under `/mnt/bb/$USER`
- create a special branch to use NVMe
- manual modification directly in the saga code (similarily to SMT level adjustment i.e. #707)
What would be a reasonable choice or is there a better option? | 1.0 | bsub option for NVME in the LSF scheduler - On Summit, Burst Buffer can be activated via the job scheduler option like:
```
-allow_flags NVME
```
There are three options that I can think of now:
- make this option as a default on Summit, then all reserved/allocated nodes have NVMe storage devices mounted under `/mnt/bb/$USER`
- create a special branch to use NVMe
- manual modification directly in the saga code (similarily to SMT level adjustment i.e. #707)
What would be a reasonable choice or is there a better option? | non_infrastructure | bsub option for nvme in the lsf scheduler on summit burst buffer can be activated via the job scheduler option like allow flags nvme there are three options that i can think of now make this option as a default on summit then all reserved allocated nodes have nvme storage devices mounted under mnt bb user create a special branch to use nvme manual modification directly in the saga code similarily to smt level adjustment i e what would be a reasonable choice or is there a better option | 0 |
31,613 | 25,935,472,274 | IssuesEvent | 2022-12-16 13:47:39 | ocaml/ocaml.org | https://api.github.com/repos/ocaml/ocaml.org | closed | Availability of the caml-list archive | bug data infrastructure | On the [Community page](https://ocaml.org/community), the [Mailing list](https://inbox.ocaml.org/caml-list/) link (<https://inbox.ocaml.org/caml-list/>) appears to be dead.
I hit this issue while trying to locate the recommended/current/pleasant archives for the caml-list. I have not found such archives -- the best I had were located on inria.fr, but only contain older posts. Funnily enough, looking for this finds [this reply](https://sympa.inria.fr/sympa/arc/caml-list/2015-12/msg00122.html) in 2015 which contains an issue link that is itself dead (the ocaml.org repo rename made those links invalid) but is actually https://github.com/ocaml/v2.ocaml.org/issues/704.
I'm curious if anyone knows of the current recommend link to caml-list archives. | 1.0 | Availability of the caml-list archive - On the [Community page](https://ocaml.org/community), the [Mailing list](https://inbox.ocaml.org/caml-list/) link (<https://inbox.ocaml.org/caml-list/>) appears to be dead.
I hit this issue while trying to locate the recommended/current/pleasant archives for the caml-list. I have not found such archives -- the best I had were located on inria.fr, but only contain older posts. Funnily enough, looking for this finds [this reply](https://sympa.inria.fr/sympa/arc/caml-list/2015-12/msg00122.html) in 2015 which contains an issue link that is itself dead (the ocaml.org repo rename made those links invalid) but is actually https://github.com/ocaml/v2.ocaml.org/issues/704.
I'm curious if anyone knows of the current recommend link to caml-list archives. | infrastructure | availability of the caml list archive on the the link appears to be dead i hit this issue while trying to locate the recommended current pleasant archives for the caml list i have not found such archives the best i had were located on inria fr but only contain older posts funnily enough looking for this finds in which contains an issue link that is itself dead the ocaml org repo rename made those links invalid but is actually i m curious if anyone knows of the current recommend link to caml list archives | 1 |
2,208 | 3,561,536,532 | IssuesEvent | 2016-01-23 21:25:15 | calblueprint/edge-yl | https://api.github.com/repos/calblueprint/edge-yl | opened | Configure conferences request for rooms page | actions infrastructure stores | If an initial `conference_id` is not passed in as a parameter, first fetch conferences and then fetch the rooms belonging to the first one.
If an initial `conference_id` is passed in as a parameter, fetch rooms belonging to that conference and then fetch conferences (to allow switching conferences). | 1.0 | Configure conferences request for rooms page - If an initial `conference_id` is not passed in as a parameter, first fetch conferences and then fetch the rooms belonging to the first one.
If an initial `conference_id` is passed in as a parameter, fetch rooms belonging to that conference and then fetch conferences (to allow switching conferences). | infrastructure | configure conferences request for rooms page if an initial conference id is not passed in as a parameter first fetch conferences and then fetch the rooms belonging to the first one if an initial conference id is passed in as a parameter fetch rooms belonging to that conference and then fetch conferences to allow switching conferences | 1 |
35,019 | 30,685,483,778 | IssuesEvent | 2023-07-26 12:05:03 | celestiaorg/devops | https://api.github.com/repos/celestiaorg/devops | closed | feat: torch - expose multiaddr as metrics | enhancement kubernetes observability infrastructure devops development tooling | expose the multiaddr as metrics via otel/prometheus
same approach as we have in watchdog to get the values from Grafana | 1.0 | feat: torch - expose multiaddr as metrics - expose the multiaddr as metrics via otel/prometheus
same approach as we have in watchdog to get the values from Grafana | infrastructure | feat torch expose multiaddr as metrics expose the multiaddr as metrics via otel prometheus same approach as we have in watchdog to get the values from grafana | 1 |
109 | 2,524,667,902 | IssuesEvent | 2015-01-20 19:18:40 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | opened | Windows dist builders hit upload exception due to locking bug in buildbot | A-infrastructure | There's a bug in the buildbot DirectoryUpload buildstep on windows that causes exceptions on our dist builders. This happens pretty regularly. It can be avoided by limiting slaves to 1 build at a time, but we don't want to do that on auto builders because it seems to leave them underutilized and lets them shut down to frequently (which then costs time to start up again). | 1.0 | Windows dist builders hit upload exception due to locking bug in buildbot - There's a bug in the buildbot DirectoryUpload buildstep on windows that causes exceptions on our dist builders. This happens pretty regularly. It can be avoided by limiting slaves to 1 build at a time, but we don't want to do that on auto builders because it seems to leave them underutilized and lets them shut down to frequently (which then costs time to start up again). | infrastructure | windows dist builders hit upload exception due to locking bug in buildbot there s a bug in the buildbot directoryupload buildstep on windows that causes exceptions on our dist builders this happens pretty regularly it can be avoided by limiting slaves to build at a time but we don t want to do that on auto builders because it seems to leave them underutilized and lets them shut down to frequently which then costs time to start up again | 1 |
262,642 | 8,272,279,735 | IssuesEvent | 2018-09-16 18:27:56 | javaee/glassfish | https://api.github.com/repos/javaee/glassfish | closed | certificate prompt ignores the --interactive=false option | Component: command_line_interface Priority: Trivial Type: Bug | When secure admin is enabled, a prompt is generated by the AsadminTrustManager class to see if the user wants to trust the certificate. This class uses the following code to set whether the prompt should be generated:
Console cons = System.console();
if (cons != null) {
However, the asadmin command has a --interactive option that is supposed to prevent prompts when set to false. The AsadminTrustManager is ignoring the --interactive option.
#### Affected Versions
[3.1_dev] | 1.0 | certificate prompt ignores the --interactive=false option - When secure admin is enabled, a prompt is generated by the AsadminTrustManager class to see if the user wants to trust the certificate. This class uses the following code to set whether the prompt should be generated:
Console cons = System.console();
if (cons != null) {
However, the asadmin command has a --interactive option that is supposed to prevent prompts when set to false. The AsadminTrustManager is ignoring the --interactive option.
#### Affected Versions
[3.1_dev] | non_infrastructure | certificate prompt ignores the interactive false option when secure admin is enabled a prompt is generated by the asadmintrustmanager class to see if the user wants to trust the certificate this class uses the following code to set whether the prompt should be generated console cons system console if cons null however the asadmin command has a interactive option that is supposed to prevent prompts when set to false the asadmintrustmanager is ignoring the interactive option affected versions | 0 |
30,300 | 24,746,836,242 | IssuesEvent | 2022-10-21 10:24:18 | react-keycloak/react-native-keycloak | https://api.github.com/repos/react-keycloak/react-native-keycloak | closed | Error: fetchTokens failed | infrastructure | **Description**
Library constantly throws "Error: fetchTokens failed" after authorization. The authorization itself is successful. If done from a browser, login and password are valid and it opens the application, but when done from the app it closes in-app-browser and throws the aforementioned error. The keycloak server is setup properly, other web projects use it successfully, problem only occurs in react native implementation.
onEvent triggers with the following message:
`onKeycloakEvent onAuthError {"error": [Error: fetchTokens failed], "error_description": "Failed to refresh token during callback processing"}`
**To Reproduce**
Steps:
1) Start application
2) Press Login
3) Enter email and password
4) Press Sign In
**Expected behavior**
After authorization keycloak updates state to authenticated.
**Smartphone**
- Device: Android Emulator
- OS: Android 11
**Additional context**
Here's the app code (same as the example, but with different app uri scheme and different keycloak url, realm, clientId):
```js
const Login = () => {
const {keycloak} = useKeycloak();
return (
<View style={styles.container}>
<Text>{`Welcome ${keycloak?.authenticated} - ${keycloak?.token}!`}</Text>
<Button onPress={() => keycloak?.login()} title="Login" />
</View>
)
}
const App = () => {
return (
<ReactNativeKeycloakProvider
authClient={keycloak}
initOptions={{ redirectUri: 'posbpay://Homepage' }}
onEvent={(event, error) => {
console.log('onKeycloakEvent', event, error);
}}
>
<Login />
</ReactNativeKeycloakProvider>
);
}
```
Intent filter configuration in AndroidManifest:
```xml
<intent-filter>
<data android:scheme="posbpay" />
<action android:name="android.intent.action.VIEW" />
<category android:name="android.intent.category.DEFAULT" />
<category android:name="android.intent.category.BROWSABLE" />
</intent-filter>
``` | 1.0 | Error: fetchTokens failed - **Description**
Library constantly throws "Error: fetchTokens failed" after authorization. The authorization itself is successful. If done from a browser, login and password are valid and it opens the application, but when done from the app it closes in-app-browser and throws the aforementioned error. The keycloak server is setup properly, other web projects use it successfully, problem only occurs in react native implementation.
onEvent triggers with the following message:
`onKeycloakEvent onAuthError {"error": [Error: fetchTokens failed], "error_description": "Failed to refresh token during callback processing"}`
**To Reproduce**
Steps:
1) Start application
2) Press Login
3) Enter email and password
4) Press Sign In
**Expected behavior**
After authorization keycloak updates state to authenticated.
**Smartphone**
- Device: Android Emulator
- OS: Android 11
**Additional context**
Here's the app code (same as the example, but with different app uri scheme and different keycloak url, realm, clientId):
```js
const Login = () => {
const {keycloak} = useKeycloak();
return (
<View style={styles.container}>
<Text>{`Welcome ${keycloak?.authenticated} - ${keycloak?.token}!`}</Text>
<Button onPress={() => keycloak?.login()} title="Login" />
</View>
)
}
const App = () => {
return (
<ReactNativeKeycloakProvider
authClient={keycloak}
initOptions={{ redirectUri: 'posbpay://Homepage' }}
onEvent={(event, error) => {
console.log('onKeycloakEvent', event, error);
}}
>
<Login />
</ReactNativeKeycloakProvider>
);
}
```
Intent filter configuration in AndroidManifest:
```xml
<intent-filter>
<data android:scheme="posbpay" />
<action android:name="android.intent.action.VIEW" />
<category android:name="android.intent.category.DEFAULT" />
<category android:name="android.intent.category.BROWSABLE" />
</intent-filter>
``` | infrastructure | error fetchtokens failed description library constantly throws error fetchtokens failed after authorization the authorization itself is successful if done from a browser login and password are valid and it opens the application but when done from the app it closes in app browser and throws the aforementioned error the keycloak server is setup properly other web projects use it successfully problem only occurs in react native implementation onevent triggers with the following message onkeycloakevent onautherror error error description failed to refresh token during callback processing to reproduce steps start application press login enter email and password press sign in expected behavior after authorization keycloak updates state to authenticated smartphone device android emulator os android additional context here s the app code same as the example but with different app uri scheme and different keycloak url realm clientid js const login const keycloak usekeycloak return welcome keycloak authenticated keycloak token keycloak login title login const app return reactnativekeycloakprovider authclient keycloak initoptions redirecturi posbpay homepage onevent event error console log onkeycloakevent event error intent filter configuration in androidmanifest xml | 1 |
160 | 2,544,860,635 | IssuesEvent | 2015-01-29 13:35:06 | debops/debops-playbooks | https://api.github.com/repos/debops/debops-playbooks | opened | Add OpenDCIM role | Cluster infrastructure | [OpenDCIM](http://opendcim.org/) is an open source Data Center Infrastructure Management application. This might require support of Basic Authentication in `nginx` to work properly. | 1.0 | Add OpenDCIM role - [OpenDCIM](http://opendcim.org/) is an open source Data Center Infrastructure Management application. This might require support of Basic Authentication in `nginx` to work properly. | infrastructure | add opendcim role is an open source data center infrastructure management application this might require support of basic authentication in nginx to work properly | 1 |
68,250 | 21,569,871,766 | IssuesEvent | 2022-05-02 06:42:02 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | opened | Cannot type the letters 'XP' in the client without it being turned into an emoji. | T-Defect | ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
The letters "XP" to appear (like Windows XP)
#### What happened instead?
Got an emoji. No way to escape it.
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | 1.0 | Cannot type the letters 'XP' in the client without it being turned into an emoji. - ### Steps to reproduce
1. Where are you starting? What can you see?
2. What do you click?
3. More steps…
### Outcome
#### What did you expect?
The letters "XP" to appear (like Windows XP)
#### What happened instead?
Got an emoji. No way to escape it.
### Operating system
_No response_
### Browser information
_No response_
### URL for webapp
_No response_
### Application version
_No response_
### Homeserver
_No response_
### Will you send logs?
No | non_infrastructure | cannot type the letters xp in the client without it being turned into an emoji steps to reproduce where are you starting what can you see what do you click more steps… outcome what did you expect the letters xp to appear like windows xp what happened instead got an emoji no way to escape it operating system no response browser information no response url for webapp no response application version no response homeserver no response will you send logs no | 0 |
28,264 | 23,120,326,103 | IssuesEvent | 2022-07-27 20:44:20 | dotnet/project-system | https://api.github.com/repos/dotnet/project-system | closed | Get APEX nuget packages to upload to a public feed automatically | Feature Request Area-Infrastructure Triage-Approved | currently APEX package creation and publish is a purely manual process
Relates to [AB#1087091](https://devdiv.visualstudio.com/0bdbc590-a062-4c3f-b0f6-9383f67865ee/_workitems/edit/1087091) | 1.0 | Get APEX nuget packages to upload to a public feed automatically - currently APEX package creation and publish is a purely manual process
Relates to [AB#1087091](https://devdiv.visualstudio.com/0bdbc590-a062-4c3f-b0f6-9383f67865ee/_workitems/edit/1087091) | infrastructure | get apex nuget packages to upload to a public feed automatically currently apex package creation and publish is a purely manual process relates to | 1 |
7,953 | 7,160,773,401 | IssuesEvent | 2018-01-28 05:27:18 | srinikoganti/sams | https://api.github.com/repos/srinikoganti/sams | closed | College Infrastructure: Add new infrastructure issue | Infrastructure P1 | Error: Getting server breakdown error when trying to add infrastructure with maximum digits in Hostel capacity text box.
Please observe the below screen shot.


Like shown below, Please check all the digits accepting text box maximum length giving server break down.

| 1.0 | College Infrastructure: Add new infrastructure issue - Error: Getting server breakdown error when trying to add infrastructure with maximum digits in Hostel capacity text box.
Please observe the below screen shot.


Like shown below, Please check all the digits accepting text box maximum length giving server break down.

| infrastructure | college infrastructure add new infrastructure issue error getting server breakdown error when trying to add infrastructure with maximum digits in hostel capacity text box please observe the below screen shot like shown below please check all the digits accepting text box maximum length giving server break down | 1 |
7,900 | 7,129,680,301 | IssuesEvent | 2018-01-22 00:48:26 | Khan/KaTeX | https://api.github.com/repos/Khan/KaTeX | closed | switch from browserify to webpack for building distribution files | infrastructure | Reasons:
- we already have a webpack.config.js with settings for the unminified distribution files
- webpack has cool plugins we can use to better optimize our code, e.g. https://webpack.js.org/plugins/module-concatenation-plugin/ to dedupe helper code that babel inserts, as discussed in #872. | 1.0 | switch from browserify to webpack for building distribution files - Reasons:
- we already have a webpack.config.js with settings for the unminified distribution files
- webpack has cool plugins we can use to better optimize our code, e.g. https://webpack.js.org/plugins/module-concatenation-plugin/ to dedupe helper code that babel inserts, as discussed in #872. | infrastructure | switch from browserify to webpack for building distribution files reasons we already have a webpack config js with settings for the unminified distribution files webpack has cool plugins we can use to better optimize our code e g to dedupe helper code that babel inserts as discussed in | 1 |
1,860 | 3,410,012,252 | IssuesEvent | 2015-12-04 18:11:31 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | reopened | [Test Failure] Test runner infrastructure failure in prtest/win/dbg/unit32 on PR #407 | Area-Infrastructure Contributor Pain Flaky | PR: [#407](https://github.com/dotnet/roslyn-internal/pull/407) ** by @
Failure: http://dotnet-ci.cloudapp.net/job/roslyn-internal_prtest_win_dbg_unit32/324/
**Issue 1: Test runner infrastructure failure**
The Test runner saw an exception that was neither a test failure nor a build failure and must have come from the test runner. | 1.0 | [Test Failure] Test runner infrastructure failure in prtest/win/dbg/unit32 on PR #407 - PR: [#407](https://github.com/dotnet/roslyn-internal/pull/407) ** by @
Failure: http://dotnet-ci.cloudapp.net/job/roslyn-internal_prtest_win_dbg_unit32/324/
**Issue 1: Test runner infrastructure failure**
The Test runner saw an exception that was neither a test failure nor a build failure and must have come from the test runner. | infrastructure | test runner infrastructure failure in prtest win dbg on pr pr by failure issue test runner infrastructure failure the test runner saw an exception that was neither a test failure nor a build failure and must have come from the test runner | 1 |
46,436 | 24,532,634,884 | IssuesEvent | 2022-10-11 17:47:37 | liquibase/liquibase | https://api.github.com/repos/liquibase/liquibase | closed | NetUtils.getLocalHostName() takes 30s | TypeEnhancement Severity3 ImpactLow PRAssociated DBAll IntegrationMaven BBroad ver4.3.5 OSMac performance | <!--- This environment context section helps us quickly review your PR.
Please take a minute to fill-out this information. -->
## Environment
**Liquibase Version**: 4.3.5
**Liquibase Integration & Version**: maven
**Liquibase Extension(s) & Version**:
**Database Vendor & Version**:
**Operating System Type & Version**: MacOS 11.5.2 Big Sur
## Description
During liquibase startup, the call `database.getDefaultSchemaName();` in `liquibase.changelog.ChangeLogParameters.java:61` leads to resolving the current hostname by `liquibase.util.NetUtil.getLocalHostName` in `NetUtil.java:71`. The subsequent line 75 takes 30s to resolve the hostname for the IP given by line 73. Interestingly, line 73 doesn't return 127.0.0.1 (from network interface eth0) but another IP taken from another network interface (eth1 / company internal network IP, here '/192.168.254.80' - yes starting with the slash. This IP can't resolved as fast as normally done by java vm.
`liquibase.util.NetUtil.getLocalHostName()`
```java
InetAddress localhost = getLocalHost(); // => '/192.168.254.80' ... yes with the leading slash
if (localhost != null) {
return localhost.getHostName(); // needs 30s
}
```
## Expected/Desired Behavior
Same as normal JVM resolution, tested by the follwing:
```java
@Test
void testResolveLocalhost() throws UnknownHostException {
InetAddress ia = InetAddress.getLocalHost();
String hostname = ia.getHostName(); // 6ms
}
@Test
void testResolveLiquibase() throws UnknownHostException, SocketException {
String ia = NetUtil.getLocalHostAddress(); //30s
}
```
## Additional Context
There are solution on Stackoverflow for the general host resolving problem in JVM. ( https://stackoverflow.com/questions/39636792/jvm-takes-a-long-time-to-resolve-ip-address-for-localhost/39698914#39698914 ) But the solutiuons given there didn't work.
| True | NetUtils.getLocalHostName() takes 30s - <!--- This environment context section helps us quickly review your PR.
Please take a minute to fill-out this information. -->
## Environment
**Liquibase Version**: 4.3.5
**Liquibase Integration & Version**: maven
**Liquibase Extension(s) & Version**:
**Database Vendor & Version**:
**Operating System Type & Version**: MacOS 11.5.2 Big Sur
## Description
During liquibase startup, the call `database.getDefaultSchemaName();` in `liquibase.changelog.ChangeLogParameters.java:61` leads to resolving the current hostname by `liquibase.util.NetUtil.getLocalHostName` in `NetUtil.java:71`. The subsequent line 75 takes 30s to resolve the hostname for the IP given by line 73. Interestingly, line 73 doesn't return 127.0.0.1 (from network interface eth0) but another IP taken from another network interface (eth1 / company internal network IP, here '/192.168.254.80' - yes starting with the slash. This IP can't resolved as fast as normally done by java vm.
`liquibase.util.NetUtil.getLocalHostName()`
```java
InetAddress localhost = getLocalHost(); // => '/192.168.254.80' ... yes with the leading slash
if (localhost != null) {
return localhost.getHostName(); // needs 30s
}
```
## Expected/Desired Behavior
Same as normal JVM resolution, tested by the follwing:
```java
@Test
void testResolveLocalhost() throws UnknownHostException {
InetAddress ia = InetAddress.getLocalHost();
String hostname = ia.getHostName(); // 6ms
}
@Test
void testResolveLiquibase() throws UnknownHostException, SocketException {
String ia = NetUtil.getLocalHostAddress(); //30s
}
```
## Additional Context
There are solution on Stackoverflow for the general host resolving problem in JVM. ( https://stackoverflow.com/questions/39636792/jvm-takes-a-long-time-to-resolve-ip-address-for-localhost/39698914#39698914 ) But the solutiuons given there didn't work.
| non_infrastructure | netutils getlocalhostname takes this environment context section helps us quickly review your pr please take a minute to fill out this information environment liquibase version liquibase integration version maven liquibase extension s version database vendor version operating system type version macos big sur description during liquibase startup the call database getdefaultschemaname in liquibase changelog changelogparameters java leads to resolving the current hostname by liquibase util netutil getlocalhostname in netutil java the subsequent line takes to resolve the hostname for the ip given by line interestingly line doesn t return from network interface but another ip taken from another network interface company internal network ip here yes starting with the slash this ip can t resolved as fast as normally done by java vm liquibase util netutil getlocalhostname java inetaddress localhost getlocalhost yes with the leading slash if localhost null return localhost gethostname needs expected desired behavior same as normal jvm resolution tested by the follwing java test void testresolvelocalhost throws unknownhostexception inetaddress ia inetaddress getlocalhost string hostname ia gethostname test void testresolveliquibase throws unknownhostexception socketexception string ia netutil getlocalhostaddress additional context there are solution on stackoverflow for the general host resolving problem in jvm but the solutiuons given there didn t work | 0 |
35,435 | 31,280,657,277 | IssuesEvent | 2023-08-22 09:22:28 | Yonatan-Lavie/aws-cdk-mongodb-api | https://api.github.com/repos/Yonatan-Lavie/aws-cdk-mongodb-api | closed | AWS Dev Environment Setup | infrastructure | - create IAM user named aws-cdk-prod/dev
- create IAM policy named `Prod/Dev-Environment-Policy`
- Attach the `Prod-Environment-Policy` to `aws-cdk-prod/dev` IAM User
- Enable Programmatic Access
- Create Access Key
- Store in GitHub Secrets Access key as `AWS_PROD/DEV_ACCESS_KEY` and Secret access key as `AWS_PROD/DEV_SECRET_ACCESS_KEY`
- create a new AWS profile in VSCode for prod/dev
- document | 1.0 | AWS Dev Environment Setup - - create IAM user named aws-cdk-prod/dev
- create IAM policy named `Prod/Dev-Environment-Policy`
- Attach the `Prod-Environment-Policy` to `aws-cdk-prod/dev` IAM User
- Enable Programmatic Access
- Create Access Key
- Store in GitHub Secrets Access key as `AWS_PROD/DEV_ACCESS_KEY` and Secret access key as `AWS_PROD/DEV_SECRET_ACCESS_KEY`
- create a new AWS profile in VSCode for prod/dev
- document | infrastructure | aws dev environment setup create iam user named aws cdk prod dev create iam policy named prod dev environment policy attach the prod environment policy to aws cdk prod dev iam user enable programmatic access create access key store in github secrets access key as aws prod dev access key and secret access key as aws prod dev secret access key create a new aws profile in vscode for prod dev document | 1 |
76,117 | 9,917,466,127 | IssuesEvent | 2019-06-29 00:43:12 | gatsbyjs/gatsby | https://api.github.com/repos/gatsbyjs/gatsby | opened | [docs][guides] improvements to Why Gatsby Uses GraphQL | help wanted topic: feedback type: documentation | ## Summary
In feedback from gatsbyjs.org, some questions surfaced that would be nice improvements to the [Why Gatsby Uses GraphQL Page](https://www.gatsbyjs.org/docs/why-gatsby-uses-graphql/). Specifically:
> "What causes allProductsJson to appear in the graphql results? What/where is it defined? Is this example on github? A link to the working example would be helpful."
For folks following along, the introduction of `allProductsJson` isn't explained clearly enough. It would help to explain where that key comes from, in explicit detail!
@jlengstorf is there also a live example we can link to?
This issue is open for anyone to pick up, once we answer where the live example lives (if there isn't one, we should create one). | 1.0 | [docs][guides] improvements to Why Gatsby Uses GraphQL - ## Summary
In feedback from gatsbyjs.org, some questions surfaced that would be nice improvements to the [Why Gatsby Uses GraphQL Page](https://www.gatsbyjs.org/docs/why-gatsby-uses-graphql/). Specifically:
> "What causes allProductsJson to appear in the graphql results? What/where is it defined? Is this example on github? A link to the working example would be helpful."
For folks following along, the introduction of `allProductsJson` isn't explained clearly enough. It would help to explain where that key comes from, in explicit detail!
@jlengstorf is there also a live example we can link to?
This issue is open for anyone to pick up, once we answer where the live example lives (if there isn't one, we should create one). | non_infrastructure | improvements to why gatsby uses graphql summary in feedback from gatsbyjs org some questions surfaced that would be nice improvements to the specifically what causes allproductsjson to appear in the graphql results what where is it defined is this example on github a link to the working example would be helpful for folks following along the introduction of allproductsjson isn t explained clearly enough it would help to explain where that key comes from in explicit detail jlengstorf is there also a live example we can link to this issue is open for anyone to pick up once we answer where the live example lives if there isn t one we should create one | 0 |
2,533 | 3,739,438,360 | IssuesEvent | 2016-03-09 04:49:04 | polyball/polyball | https://api.github.com/repos/polyball/polyball | closed | Integrate TravisCI (our CI build tool) with slack | Tooling/Infrastructure | It is important that members who break the build are notified so that they can fix it asap. It is currently inconvenient to check Travis-CI.org every time you push. | 1.0 | Integrate TravisCI (our CI build tool) with slack - It is important that members who break the build are notified so that they can fix it asap. It is currently inconvenient to check Travis-CI.org every time you push. | infrastructure | integrate travisci our ci build tool with slack it is important that members who break the build are notified so that they can fix it asap it is currently inconvenient to check travis ci org every time you push | 1 |
133,254 | 18,286,301,395 | IssuesEvent | 2021-10-05 10:39:46 | lowRISC/opentitan | https://api.github.com/repos/lowRISC/opentitan | closed | RFC: Key sideload scheme | RFC:Proposal Component:Security Hotlist:Security | ## Objective
The objective of this RFC is to collect feedback on sideload architecture options, and decide what level of
support to implement in the key manager integrated into Earl Grey.
The RFC is currently in open for comments. Once a consensus has been reached, the RFC will be updated with the
design decision before promoting it to the Technical Committee for final review.
## Background - What is Sideload?
Sideload, or key manager sideload, refers to a direct hardware connection from the OpenTitan key manager module to hardware symmetric crypto components for supplying keys.
When sideload is employed, keys are generated within the key manager and fed directly to the hardware crypto components for usage. At no point during this process are the keys exposed to software. Thus, compromised software can only abuse the keys but cannot permanently expose them [see doc for more details][no-sideload-proposal].
## Design Options
Currently, there are a couple of proposals under consideration. As part of this RFC, we are requesting inputs for
additional proposals.
### Key slot implementation
The implementation may provide a limited number of key slots in the key manager, with support for sideloading keys into the
`HMAC`, `KMAC`, and `AES` engines. The number of key slots is left as an implementation detail which needs
to take into consideration tradeoffs between: use cases versus hardware cost.
Additional considerations:
* Even though software does not have visibility into the key material, it is still responsible for managing the key slot
mapping to consuming applications.
* The number of bits sideloaded to other IP may be constrained to meet a security level up to 256 bits.
* Sideloading keys into `OTBN` is not considered a security feature given that the target does not implement
isolation of critical data from software. The current recommendation is to avoid sideloading into `OTBN`.
### No-Sideload Support
Given the limitations of the key-slot proposal above, and additional considerations, we have put together a justification
to remove support for sideloading from the architecture.
See the [proposal document][no-sideload-proposal] for more details.
[no-sideload-proposal]: https://docs.google.com/document/d/1kTh5E9wgD-X1eQSX3h-67tFvcmHTjjo7ng5bZKYIA-E/edit?ts=5ee97acc#
## Update
See [sideload memo](https://docs.google.com/document/d/1BVxsA6-19eUuempvtiv136LH8Z0GLnc5cpapmk0H1uw/edit?usp=sharing) for the latest thinking and proposals.
The ongoing [key manager update deck](https://docs.google.com/presentation/d/1noUFDWAHlcz5zHByuav6I9z7HrNM3hsiyYwmi2z7N3s/edit?usp=sharing) also details the matching hardware updates for key manager only.
| True | RFC: Key sideload scheme - ## Objective
The objective of this RFC is to collect feedback on sideload architecture options, and decide what level of
support to implement in the key manager integrated into Earl Grey.
The RFC is currently in open for comments. Once a consensus has been reached, the RFC will be updated with the
design decision before promoting it to the Technical Committee for final review.
## Background - What is Sideload?
Sideload, or key manager sideload, refers to a direct hardware connection from the OpenTitan key manager module to hardware symmetric crypto components for supplying keys.
When sideload is employed, keys are generated within the key manager and fed directly to the hardware crypto components for usage. At no point during this process are the keys exposed to software. Thus, compromised software can only abuse the keys but cannot permanently expose them [see doc for more details][no-sideload-proposal].
## Design Options
Currently, there are a couple of proposals under consideration. As part of this RFC, we are requesting inputs for
additional proposals.
### Key slot implementation
The implementation may provide a limited number of key slots in the key manager, with support for sideloading keys into the
`HMAC`, `KMAC`, and `AES` engines. The number of key slots is left as an implementation detail which needs
to take into consideration tradeoffs between: use cases versus hardware cost.
Additional considerations:
* Even though software does not have visibility into the key material, it is still responsible for managing the key slot
mapping to consuming applications.
* The number of bits sideloaded to other IP may be constrained to meet a security level up to 256 bits.
* Sideloading keys into `OTBN` is not considered a security feature given that the target does not implement
isolation of critical data from software. The current recommendation is to avoid sideloading into `OTBN`.
### No-Sideload Support
Given the limitations of the key-slot proposal above, and additional considerations, we have put together a justification
to remove support for sideloading from the architecture.
See the [proposal document][no-sideload-proposal] for more details.
[no-sideload-proposal]: https://docs.google.com/document/d/1kTh5E9wgD-X1eQSX3h-67tFvcmHTjjo7ng5bZKYIA-E/edit?ts=5ee97acc#
## Update
See [sideload memo](https://docs.google.com/document/d/1BVxsA6-19eUuempvtiv136LH8Z0GLnc5cpapmk0H1uw/edit?usp=sharing) for the latest thinking and proposals.
The ongoing [key manager update deck](https://docs.google.com/presentation/d/1noUFDWAHlcz5zHByuav6I9z7HrNM3hsiyYwmi2z7N3s/edit?usp=sharing) also details the matching hardware updates for key manager only.
| non_infrastructure | rfc key sideload scheme objective the objective of this rfc is to collect feedback on sideload architecture options and decide what level of support to implement in the key manager integrated into earl grey the rfc is currently in open for comments once a consensus has been reached the rfc will be updated with the design decision before promoting it to the technical committee for final review background what is sideload sideload or key manager sideload refers to a direct hardware connection from the opentitan key manager module to hardware symmetric crypto components for supplying keys when sideload is employed keys are generated within the key manager and fed directly to the hardware crypto components for usage at no point during this process are the keys exposed to software thus compromised software can only abuse the keys but cannot permanently expose them design options currently there are a couple of proposals under consideration as part of this rfc we are requesting inputs for additional proposals key slot implementation the implementation may provide a limited number of key slots in the key manager with support for sideloading keys into the hmac kmac and aes engines the number of key slots is left as an implementation detail which needs to take into consideration tradeoffs between use cases versus hardware cost additional considerations even though software does not have visibility into the key material it is still responsible for managing the key slot mapping to consuming applications the number of bits sideloaded to other ip may be constrained to meet a security level up to bits sideloading keys into otbn is not considered a security feature given that the target does not implement isolation of critical data from software the current recommendation is to avoid sideloading into otbn no sideload support given the limitations of the key slot proposal above and additional considerations we have put together a justification to remove support for sideloading from the architecture see the for more details update see for the latest thinking and proposals the ongoing also details the matching hardware updates for key manager only | 0 |
7,030 | 3,074,501,641 | IssuesEvent | 2015-08-20 07:45:20 | arangodb/arangodb | https://api.github.com/repos/arangodb/arangodb | closed | Building issues on Windows (V8, Python 3.x and others) | documentation | Quoting a recipe from the cookbook:
> Either version 2.x or 3.x (excluding 3.4) should work - it's used to run V8s GYP. Make sure you add python.exe to your path environment variable; Restarting your running shell may be necessary.
I'm not sure why 3.4 is excluded, but 3.x in general supposed to work. It seems that if 3.x is installed and added to PATH environment variable, V8 compilation fails because of a syntax error:
import sys; print sys.byteorder
`print` is a function in Python 3 and requires parentheses.
Is there a way to make gyp use `ArangoDB\3rdParty\V8-4.3.61\third_party\python_26\python.exe` instead of a local Python installation? | 1.0 | Building issues on Windows (V8, Python 3.x and others) - Quoting a recipe from the cookbook:
> Either version 2.x or 3.x (excluding 3.4) should work - it's used to run V8s GYP. Make sure you add python.exe to your path environment variable; Restarting your running shell may be necessary.
I'm not sure why 3.4 is excluded, but 3.x in general supposed to work. It seems that if 3.x is installed and added to PATH environment variable, V8 compilation fails because of a syntax error:
import sys; print sys.byteorder
`print` is a function in Python 3 and requires parentheses.
Is there a way to make gyp use `ArangoDB\3rdParty\V8-4.3.61\third_party\python_26\python.exe` instead of a local Python installation? | non_infrastructure | building issues on windows python x and others quoting a recipe from the cookbook either version x or x excluding should work it s used to run gyp make sure you add python exe to your path environment variable restarting your running shell may be necessary i m not sure why is excluded but x in general supposed to work it seems that if x is installed and added to path environment variable compilation fails because of a syntax error import sys print sys byteorder print is a function in python and requires parentheses is there a way to make gyp use arangodb third party python python exe instead of a local python installation | 0 |
338,931 | 30,330,312,333 | IssuesEvent | 2023-07-11 05:41:17 | MaterializeInc/materialize | https://api.github.com/repos/MaterializeInc/materialize | closed | QA sign-off for pushdown | C-feature T-testing | ### Feature request
- [x] Enable in CI by default
- [x] RQG runs
- [x] Bring https://github.com/MaterializeInc/materialize/pull/19013 to bear
- [ ] ... | 1.0 | QA sign-off for pushdown - ### Feature request
- [x] Enable in CI by default
- [x] RQG runs
- [x] Bring https://github.com/MaterializeInc/materialize/pull/19013 to bear
- [ ] ... | non_infrastructure | qa sign off for pushdown feature request enable in ci by default rqg runs bring to bear | 0 |
5,462 | 3,930,245,775 | IssuesEvent | 2016-04-25 07:01:29 | kolliSuman/issues | https://api.github.com/repos/kolliSuman/issues | closed | QA_N-Grams Smoothing_Prerequisites_p1 | Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open | Defect Description :
In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required
Actual Result :
In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0Bandwidth : 100Mbps
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/natural-language-processing-iiith/blob/master/test-cases/integration_test-cases/N-Grams%20Smoothing/N-Grams%20Smoothing_13_Prerequisites_p1.org | True | QA_N-Grams Smoothing_Prerequisites_p1 - Defect Description :
In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required
Actual Result :
In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0Bandwidth : 100Mbps
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/natural-language-processing-iiith/blob/master/test-cases/integration_test-cases/N-Grams%20Smoothing/N-Grams%20Smoothing_13_Prerequisites_p1.org | non_infrastructure | qa n grams smoothing prerequisites defect description in the n grams smoothing experiment the minimum requirement to run the experiment is not displayed in the page instead a page or scrolling should appear providing information on minimum requirement to run this experiment information like bandwidth device resolution hardware configuration and software required actual result in the n grams smoothing experiment the minimum requirement to run the experiment is not displayed in the page environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link | 0 |
27,086 | 21,154,695,725 | IssuesEvent | 2022-04-07 00:59:46 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Soil crop parameters not always imported correctly from APSIM 7.10 | bug interface/infrastructure | When importing a .apsim file from APSIM 7.10 the soil crop parameters are not imported if their name starts with a lower case letter e.g. wheat. | 1.0 | Soil crop parameters not always imported correctly from APSIM 7.10 - When importing a .apsim file from APSIM 7.10 the soil crop parameters are not imported if their name starts with a lower case letter e.g. wheat. | infrastructure | soil crop parameters not always imported correctly from apsim when importing a apsim file from apsim the soil crop parameters are not imported if their name starts with a lower case letter e g wheat | 1 |
18,559 | 13,055,867,428 | IssuesEvent | 2020-07-30 02:58:22 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | Use I3_DATA to store photonics spline tables on the buildbots (Trac #742) | Incomplete Migration Migrated from Trac infrastructure task | Migrated from https://code.icecube.wisc.edu/ticket/742
```json
{
"status": "closed",
"changetime": "2015-02-18T21:44:30",
"description": "Some of the Millipede unit tests need the full spline tables to function. At the moment these are read from the canonical path I3_DATA, which for the moment only exists on cvmfs-enabled machines, and so they don't run on the buildbots, and so platform-dependent bugs aren't caught in time.\n\nSimply copying these into the normal test data would be needless bloat, since they're used outside of testing as well. /cvmfs/icecube.wisc.edu/data should be cloned into the buildbots, and I3_DATA set in their environment. If the entire data directory is too large, then /cvmfs/icecube.wisc.edu/data/photon-tables/splines would do in a pinch.",
"reporter": "jvansanten",
"cc": "",
"resolution": "fixed",
"_ts": "1424295870059276",
"component": "infrastructure",
"summary": "Use I3_DATA to store photonics spline tables on the buildbots",
"priority": "normal",
"keywords": "",
"time": "2014-08-30T20:27:54",
"milestone": "",
"owner": "nega",
"type": "task"
}
```
| 1.0 | Use I3_DATA to store photonics spline tables on the buildbots (Trac #742) - Migrated from https://code.icecube.wisc.edu/ticket/742
```json
{
"status": "closed",
"changetime": "2015-02-18T21:44:30",
"description": "Some of the Millipede unit tests need the full spline tables to function. At the moment these are read from the canonical path I3_DATA, which for the moment only exists on cvmfs-enabled machines, and so they don't run on the buildbots, and so platform-dependent bugs aren't caught in time.\n\nSimply copying these into the normal test data would be needless bloat, since they're used outside of testing as well. /cvmfs/icecube.wisc.edu/data should be cloned into the buildbots, and I3_DATA set in their environment. If the entire data directory is too large, then /cvmfs/icecube.wisc.edu/data/photon-tables/splines would do in a pinch.",
"reporter": "jvansanten",
"cc": "",
"resolution": "fixed",
"_ts": "1424295870059276",
"component": "infrastructure",
"summary": "Use I3_DATA to store photonics spline tables on the buildbots",
"priority": "normal",
"keywords": "",
"time": "2014-08-30T20:27:54",
"milestone": "",
"owner": "nega",
"type": "task"
}
```
| infrastructure | use data to store photonics spline tables on the buildbots trac migrated from json status closed changetime description some of the millipede unit tests need the full spline tables to function at the moment these are read from the canonical path data which for the moment only exists on cvmfs enabled machines and so they don t run on the buildbots and so platform dependent bugs aren t caught in time n nsimply copying these into the normal test data would be needless bloat since they re used outside of testing as well cvmfs icecube wisc edu data should be cloned into the buildbots and data set in their environment if the entire data directory is too large then cvmfs icecube wisc edu data photon tables splines would do in a pinch reporter jvansanten cc resolution fixed ts component infrastructure summary use data to store photonics spline tables on the buildbots priority normal keywords time milestone owner nega type task | 1 |
56,999 | 13,962,905,732 | IssuesEvent | 2020-10-25 11:49:16 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | CUDA problem with Linux Kernel 5.9 | module: build module: cuda triaged | ## 🐛 Bug
<!-- A clear and concise description of what the bug is. -->
torch.cuda.is_available() returns True on Ubuntu 20.04 with linux kernel 5.8.13, but returns False when the kernel is upgraded to 5.9-rc8. nvidia-smi and other gpu-related programs work as expected, only pytorch stops detecting the GPU.
## To Reproduce
In a Ubuntu 20.04 OS with linux kernel 5.9-rc8, run:
```python -c 'import torch; print(torch.cuda.is_available())'```
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Pytorch should detect the GPU on Linux kernel 5.9.
## Environment
- PyTorch Version (e.g., 1.0): 1.6.0
- OS (e.g., Linux): Ubuntu 20.04.1 LTS (x86_64)
- How you installed PyTorch (`conda`, `pip`, source): pip
- Python version: 3.8
- CUDA version: 10.2 in pytorch, N/A in the system.
- GPU models and configuration: GeForce RTX 3090
cc @malfet @seemethere @walterddr @ngimel | 1.0 | CUDA problem with Linux Kernel 5.9 - ## 🐛 Bug
<!-- A clear and concise description of what the bug is. -->
torch.cuda.is_available() returns True on Ubuntu 20.04 with linux kernel 5.8.13, but returns False when the kernel is upgraded to 5.9-rc8. nvidia-smi and other gpu-related programs work as expected, only pytorch stops detecting the GPU.
## To Reproduce
In a Ubuntu 20.04 OS with linux kernel 5.9-rc8, run:
```python -c 'import torch; print(torch.cuda.is_available())'```
<!-- If you have a code sample, error messages, stack traces, please provide it here as well -->
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Pytorch should detect the GPU on Linux kernel 5.9.
## Environment
- PyTorch Version (e.g., 1.0): 1.6.0
- OS (e.g., Linux): Ubuntu 20.04.1 LTS (x86_64)
- How you installed PyTorch (`conda`, `pip`, source): pip
- Python version: 3.8
- CUDA version: 10.2 in pytorch, N/A in the system.
- GPU models and configuration: GeForce RTX 3090
cc @malfet @seemethere @walterddr @ngimel | non_infrastructure | cuda problem with linux kernel 🐛 bug torch cuda is available returns true on ubuntu with linux kernel but returns false when the kernel is upgraded to nvidia smi and other gpu related programs work as expected only pytorch stops detecting the gpu to reproduce in a ubuntu os with linux kernel run python c import torch print torch cuda is available expected behavior pytorch should detect the gpu on linux kernel environment pytorch version e g os e g linux ubuntu lts how you installed pytorch conda pip source pip python version cuda version in pytorch n a in the system gpu models and configuration geforce rtx cc malfet seemethere walterddr ngimel | 0 |
690,487 | 23,661,612,225 | IssuesEvent | 2022-08-26 16:06:59 | TheYellowArchitect/doubledamnation | https://api.github.com/repos/TheYellowArchitect/doubledamnation | opened | Level Editor - Pause to disable grabbing/removing tiles | bug good first issue level editor low priority | tl;dr: Try clicking any settings, and not creating tiles in the background.
Spoilers: You can't.
Luckily, its extremely easy to fix, and also takes little time. | 1.0 | Level Editor - Pause to disable grabbing/removing tiles - tl;dr: Try clicking any settings, and not creating tiles in the background.
Spoilers: You can't.
Luckily, its extremely easy to fix, and also takes little time. | non_infrastructure | level editor pause to disable grabbing removing tiles tl dr try clicking any settings and not creating tiles in the background spoilers you can t luckily its extremely easy to fix and also takes little time | 0 |
14,407 | 10,821,991,236 | IssuesEvent | 2019-11-08 20:03:29 | aspnet/AspNetCore | https://api.github.com/repos/aspnet/AspNetCore | closed | restore command in root folder of Asp Net Core project fails | area-infrastructure | Hi,I have cloned aspnetcore project and followed the instruction here in GitHub step-by-step but when ran restore command in the root directory of project I got the following errors.
C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error : Unable to loc
ate the .NET Core SDK. Check that it is installed and that the version specified in global.json (if any) matches the in
stalled version.
C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error MSB4236: The SD
K 'Microsoft.NET.Sdk' specified could not be found.
Build failed.
Please help me in order to solve this issue. | 1.0 | restore command in root folder of Asp Net Core project fails - Hi,I have cloned aspnetcore project and followed the instruction here in GitHub step-by-step but when ran restore command in the root directory of project I got the following errors.
C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error : Unable to loc
ate the .NET Core SDK. Check that it is installed and that the version specified in global.json (if any) matches the in
stalled version.
C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error MSB4236: The SD
K 'Microsoft.NET.Sdk' specified could not be found.
Build failed.
Please help me in order to solve this issue. | infrastructure | restore command in root folder of asp net core project fails hi i have cloned aspnetcore project and followed the instruction here in github step by step but when ran restore command in the root directory of project i got the following errors c users nurlan nuget packages microsoft dotnet arcade sdk beta tools tools proj error unable to loc ate the net core sdk check that it is installed and that the version specified in global json if any matches the in stalled version c users nurlan nuget packages microsoft dotnet arcade sdk beta tools tools proj error the sd k microsoft net sdk specified could not be found build failed please help me in order to solve this issue | 1 |
399,513 | 27,243,670,096 | IssuesEvent | 2023-02-21 23:03:39 | SAP/fundamental | https://api.github.com/repos/SAP/fundamental | closed | Docs: Components should use the base class | Documentation Develop | **Describe the bug**
A clear and concise description of what the bug is.
In some cases, components use a class selector to look for a block class which eliminates the need to always put in a base class.
```
[class*="fd-identifier"] {
}
```
While this works OK, it creates some issues when using Angular with modularized styles.
For example, if I use `<span class="fd-identifier--xl"><span>{{noOfTries}}</span></span>` the styling looks fine, but I cannot easily override the vars in my component CSS.
This will not work ...
```
.fd-identifier {
--fd-identifier-background-color: var(--fd-color-accent-9);
}
```
No changes need to be done in the CSS at this time, but the NJK templates and markup in the docs should be updated to always include the block class along with any modifier classes.
Affects
- [ ] icons
- [ ] image
- [ ] identifier
- [ ] button | 1.0 | Docs: Components should use the base class - **Describe the bug**
A clear and concise description of what the bug is.
In some cases, components use a class selector to look for a block class which eliminates the need to always put in a base class.
```
[class*="fd-identifier"] {
}
```
While this works OK, it creates some issues when using Angular with modularized styles.
For example, if I use `<span class="fd-identifier--xl"><span>{{noOfTries}}</span></span>` the styling looks fine, but I cannot easily override the vars in my component CSS.
This will not work ...
```
.fd-identifier {
--fd-identifier-background-color: var(--fd-color-accent-9);
}
```
No changes need to be done in the CSS at this time, but the NJK templates and markup in the docs should be updated to always include the block class along with any modifier classes.
Affects
- [ ] icons
- [ ] image
- [ ] identifier
- [ ] button | non_infrastructure | docs components should use the base class describe the bug a clear and concise description of what the bug is in some cases components use a class selector to look for a block class which eliminates the need to always put in a base class while this works ok it creates some issues when using angular with modularized styles for example if i use nooftries the styling looks fine but i cannot easily override the vars in my component css this will not work fd identifier fd identifier background color var fd color accent no changes need to be done in the css at this time but the njk templates and markup in the docs should be updated to always include the block class along with any modifier classes affects icons image identifier button | 0 |
28,250 | 23,111,869,342 | IssuesEvent | 2022-07-27 13:40:08 | spring-projects/spring-batch | https://api.github.com/repos/spring-projects/spring-batch | opened | Migrate tests to JUnit Jupiter | in: test in: infrastructure type: feature in: integration in: samples in: core | This issue is to track the migration of the test suite from JUnit 4 to JUnit 5:
- [ ] spring-batch-infrastructure: #4125
- [ ] spring-batch-core
- [ ] spring-batch-test
- [ ] spring-batch-integration: #4124
- [ ] spring-batch-samples | 1.0 | Migrate tests to JUnit Jupiter - This issue is to track the migration of the test suite from JUnit 4 to JUnit 5:
- [ ] spring-batch-infrastructure: #4125
- [ ] spring-batch-core
- [ ] spring-batch-test
- [ ] spring-batch-integration: #4124
- [ ] spring-batch-samples | infrastructure | migrate tests to junit jupiter this issue is to track the migration of the test suite from junit to junit spring batch infrastructure spring batch core spring batch test spring batch integration spring batch samples | 1 |
18,070 | 12,748,923,845 | IssuesEvent | 2020-06-26 21:16:54 | commitdev/zero | https://api.github.com/repos/commitdev/zero | opened | Investigate Ory tools for reusable auth | backend infrastructure | Do another round of investigation into Ory Hydra / Oathkeeper / Kratos for user management and auth. See if they have the features and integrations we want. Last time Bill looked there were some things lacking.
Should be:
- Easy to integrate with nginx ingress controller
- Abstract enough away so that it handles login, then provides auth via header or cookie for backend service to use (probably JWT)
- Should be easy to put a customizeable UI on top | 1.0 | Investigate Ory tools for reusable auth - Do another round of investigation into Ory Hydra / Oathkeeper / Kratos for user management and auth. See if they have the features and integrations we want. Last time Bill looked there were some things lacking.
Should be:
- Easy to integrate with nginx ingress controller
- Abstract enough away so that it handles login, then provides auth via header or cookie for backend service to use (probably JWT)
- Should be easy to put a customizeable UI on top | infrastructure | investigate ory tools for reusable auth do another round of investigation into ory hydra oathkeeper kratos for user management and auth see if they have the features and integrations we want last time bill looked there were some things lacking should be easy to integrate with nginx ingress controller abstract enough away so that it handles login then provides auth via header or cookie for backend service to use probably jwt should be easy to put a customizeable ui on top | 1 |
432,057 | 12,488,647,072 | IssuesEvent | 2020-05-31 15:11:59 | buttercup/buttercup-mobile | https://api.github.com/repos/buttercup/buttercup-mobile | closed | Improve label on button to hide/display info | Effort: Low Priority: Low Status: Available Type: Enhancement | When seeing a entry on a Archive, all sensitive data is blurred. To display, one must click the __view hidden__ button, which is a weird label for it. I suggest to change to something along the lines of:
_Display sensitive information_
_Hide sensitive information_
References:


| 1.0 | Improve label on button to hide/display info - When seeing a entry on a Archive, all sensitive data is blurred. To display, one must click the __view hidden__ button, which is a weird label for it. I suggest to change to something along the lines of:
_Display sensitive information_
_Hide sensitive information_
References:


| non_infrastructure | improve label on button to hide display info when seeing a entry on a archive all sensitive data is blurred to display one must click the view hidden button which is a weird label for it i suggest to change to something along the lines of display sensitive information hide sensitive information references | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.