Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
581
2,775,505,941
IssuesEvent
2015-05-04 16:26:08
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
ContributingCode wikipage has missing link to CLA, stale link to style guide
Area-Infrastructure
[ContributingCode](https://github.com/dotnet/roslyn/wiki/Contributing-Code) on the wiki doesn't appear to have any link for the CLA, and the link for the style guide seems to have gotten out of date. For the CLA, you could link to either <https://github.com/dotnet/corefx/wiki/Contribution-License-Agreement-(CLA)> or <https://cla2.dotnetfoundation.org/>. For the style guide, you probably want to link to <https://github.com/dotnet/corefx/wiki/Coding-style> rather than to <https://github.com/dotnet/corefx/wiki/Contributing#c-coding-style>.
1.0
ContributingCode wikipage has missing link to CLA, stale link to style guide - [ContributingCode](https://github.com/dotnet/roslyn/wiki/Contributing-Code) on the wiki doesn't appear to have any link for the CLA, and the link for the style guide seems to have gotten out of date. For the CLA, you could link to either <https://github.com/dotnet/corefx/wiki/Contribution-License-Agreement-(CLA)> or <https://cla2.dotnetfoundation.org/>. For the style guide, you probably want to link to <https://github.com/dotnet/corefx/wiki/Coding-style> rather than to <https://github.com/dotnet/corefx/wiki/Contributing#c-coding-style>.
infrastructure
contributingcode wikipage has missing link to cla stale link to style guide on the wiki doesn t appear to have any link for the cla and the link for the style guide seems to have gotten out of date for the cla you could link to either or for the style guide you probably want to link to rather than to
1
7,208
7,869,515,607
IssuesEvent
2018-06-24 14:52:25
CodeWarriorsOrganization/CodeWarriorsFinalYearProject
https://api.github.com/repos/CodeWarriorsOrganization/CodeWarriorsFinalYearProject
closed
CW0025/T0025.4 -Implement the Service logic (Priority-2)
implementation service
### User Story : > AS a player, > I NEED to be able to view the total number of turns, SO THAT I can get to know how many turns > spend. ### Task : > **Implement the Service logic**
1.0
CW0025/T0025.4 -Implement the Service logic (Priority-2) - ### User Story : > AS a player, > I NEED to be able to view the total number of turns, SO THAT I can get to know how many turns > spend. ### Task : > **Implement the Service logic**
non_infrastructure
implement the service logic priority user story as a player i need to be able to view the total number of turns so that i can get to know how many turns spend task implement the service logic
0
11,865
9,476,410,363
IssuesEvent
2019-04-19 14:59:49
opencv/opencv
https://api.github.com/repos/opencv/opencv
closed
Java: test crashes with OpenVINO 2019R1 on MacOSX
bug category: dnn category: infrastructure category: java bindings platform: ios/osx test
[Nightly build](http://pullrequest.opencv.org/buildbot/builders/master_openvino-mac/builds/0) ``` [junit] Running org.opencv.test.calib3d.Calib3dTest [junit] java.lang.UnsatisfiedLinkError: /Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib: dlopen(/Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib, 1): Library not loaded: @rpath/libtbb.dylib [junit] Referenced from: /opt/intel/openvino_2019.1.090/deployment_tools/inference_engine/lib/intel64/libinference_engine.dylib [junit] Reason: image not found [junit] Running org.opencv.test.calib3d.Calib3dTest [junit] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0 sec [junit] Test org.opencv.test.calib3d.Calib3dTest FAILED (crashed) ```
1.0
Java: test crashes with OpenVINO 2019R1 on MacOSX - [Nightly build](http://pullrequest.opencv.org/buildbot/builders/master_openvino-mac/builds/0) ``` [junit] Running org.opencv.test.calib3d.Calib3dTest [junit] java.lang.UnsatisfiedLinkError: /Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib: dlopen(/Volumes/build-storage/build/master_openvino-mac/build/lib/libopencv_java410.dylib, 1): Library not loaded: @rpath/libtbb.dylib [junit] Referenced from: /opt/intel/openvino_2019.1.090/deployment_tools/inference_engine/lib/intel64/libinference_engine.dylib [junit] Reason: image not found [junit] Running org.opencv.test.calib3d.Calib3dTest [junit] Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0 sec [junit] Test org.opencv.test.calib3d.Calib3dTest FAILED (crashed) ```
infrastructure
java test crashes with openvino on macosx running org opencv test java lang unsatisfiedlinkerror volumes build storage build master openvino mac build lib libopencv dylib dlopen volumes build storage build master openvino mac build lib libopencv dylib library not loaded rpath libtbb dylib referenced from opt intel openvino deployment tools inference engine lib libinference engine dylib reason image not found running org opencv test tests run failures errors skipped time elapsed sec test org opencv test failed crashed
1
3,597
4,427,999,246
IssuesEvent
2016-08-16 23:44:50
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
The build.cmd script should accept more common options
Infrastructure
Currently to build the repo as release you have to do this: ```cmd build /p:ConfigurationGroup=Release ``` This is very error-prone to typos, since you can write something like `/p:ConfigurationGrouo=Release` and go away from your computer for 40 minutes only to realize you built on Debug configuration. We should provide `--debug` and `--release` shortcuts like is done in [the `build.sh` script](https://github.com/dotnet/corefx/blob/master/build.sh#L10), as well as update the documentation/`--help` output. Then, we could call the script like this: ```cmd build --release ``` Other things we might want to consider: `--skiptests[ bool]` (default false), `--buildtests[ bool]` (default true), `--buildpackages[ bool]` (default true) cc: @mellinoe Note: I have knowledge of batch/shell scripting, so I can submit a PR to fix this myself if this is OK. **edit:** Additionally we should echo the arguments that are being passed to MSBuild to the user, so they know right away if they made a typo.
1.0
The build.cmd script should accept more common options - Currently to build the repo as release you have to do this: ```cmd build /p:ConfigurationGroup=Release ``` This is very error-prone to typos, since you can write something like `/p:ConfigurationGrouo=Release` and go away from your computer for 40 minutes only to realize you built on Debug configuration. We should provide `--debug` and `--release` shortcuts like is done in [the `build.sh` script](https://github.com/dotnet/corefx/blob/master/build.sh#L10), as well as update the documentation/`--help` output. Then, we could call the script like this: ```cmd build --release ``` Other things we might want to consider: `--skiptests[ bool]` (default false), `--buildtests[ bool]` (default true), `--buildpackages[ bool]` (default true) cc: @mellinoe Note: I have knowledge of batch/shell scripting, so I can submit a PR to fix this myself if this is OK. **edit:** Additionally we should echo the arguments that are being passed to MSBuild to the user, so they know right away if they made a typo.
infrastructure
the build cmd script should accept more common options currently to build the repo as release you have to do this cmd build p configurationgroup release this is very error prone to typos since you can write something like p configurationgrouo release and go away from your computer for minutes only to realize you built on debug configuration we should provide debug and release shortcuts like is done in as well as update the documentation help output then we could call the script like this cmd build release other things we might want to consider skiptests default false buildtests default true buildpackages default true cc mellinoe note i have knowledge of batch shell scripting so i can submit a pr to fix this myself if this is ok edit additionally we should echo the arguments that are being passed to msbuild to the user so they know right away if they made a typo
1
517,991
15,021,963,680
IssuesEvent
2021-02-01 16:24:14
BiologicalRecordsCentre/wrappeR
https://api.github.com/repos/BiologicalRecordsCentre/wrappeR
closed
Upload meta data creation script to wrapper
priority - high sprint
Upload meta data creation script to wrapper @mlogie
1.0
Upload meta data creation script to wrapper - Upload meta data creation script to wrapper @mlogie
non_infrastructure
upload meta data creation script to wrapper upload meta data creation script to wrapper mlogie
0
102,777
11,307,054,780
IssuesEvent
2020-01-18 18:19:35
uhop/stream-json
https://api.github.com/repos/uhop/stream-json
closed
Add type annotations to data-event
documentation enhancement
Hi, From what I see, the `data` event on the stream accepts "any" as an argument. However in case of `StreamValues` the value is known, so I believe that having an annotation to point to `{key: number, value: any}` would be a good start. What do you think? Not sure how easy to plug that in though given that `StreamValues.withParser()` returns a Chain.
1.0
Add type annotations to data-event - Hi, From what I see, the `data` event on the stream accepts "any" as an argument. However in case of `StreamValues` the value is known, so I believe that having an annotation to point to `{key: number, value: any}` would be a good start. What do you think? Not sure how easy to plug that in though given that `StreamValues.withParser()` returns a Chain.
non_infrastructure
add type annotations to data event hi from what i see the data event on the stream accepts any as an argument however in case of streamvalues the value is known so i believe that having an annotation to point to key number value any would be a good start what do you think not sure how easy to plug that in though given that streamvalues withparser returns a chain
0
900
2,997,357,739
IssuesEvent
2015-07-23 06:55:19
gammapy/gammapy
https://api.github.com/repos/gammapy/gammapy
opened
Update install instructions and release notes for conda packages
docs infrastructure
Matt Craig set up binary packages for Gammapy: https://groups.google.com/forum/#!topic/astropy-dev/u8-r3AW8V4s https://groups.google.com/forum/#!topic/gammapy/WJjWxHphTGM - [ ] Update install instructions - [ ] Update release notes to make a PR against https://github.com/astropy/conda-builder-affiliated to trigger a new build - [ ] Maybe try to get some of our optional dependencies built there (e.g. iminuit or Gammalib / ctools)
1.0
Update install instructions and release notes for conda packages - Matt Craig set up binary packages for Gammapy: https://groups.google.com/forum/#!topic/astropy-dev/u8-r3AW8V4s https://groups.google.com/forum/#!topic/gammapy/WJjWxHphTGM - [ ] Update install instructions - [ ] Update release notes to make a PR against https://github.com/astropy/conda-builder-affiliated to trigger a new build - [ ] Maybe try to get some of our optional dependencies built there (e.g. iminuit or Gammalib / ctools)
infrastructure
update install instructions and release notes for conda packages matt craig set up binary packages for gammapy update install instructions update release notes to make a pr against to trigger a new build maybe try to get some of our optional dependencies built there e g iminuit or gammalib ctools
1
28,241
23,103,585,783
IssuesEvent
2022-07-27 06:41:41
FTBTeam/FTB-Modpack-Issues
https://api.github.com/repos/FTBTeam/FTB-Modpack-Issues
closed
[Bug]: Direwolf20 1.18 Linux fail to download
3: Infrastructure FTB Presents Direwolf20 1.18
### Modpack Direwolf20 1.18 ### Modpack version 2234 ### Log Files 491 Successful 266 Failed https://pste.ch/zoyojitozo.pl Continue anyway. https://pste.ch/jovurutaxi.sql ### Describe the bug With linux, the server does not download fully, having 266 Failed downloads. Additionally, the start.sh is not formatted correctly. https://pste.ch/ijopamevun.bash ### Steps to reproduce 1. wget https://feed-the-beast.com/modpack/95_ftb_presents_direwolf20_1_18/server-files 2. chmod +x linux 3. ./linux 95 4. follow prompts ### Expected behaviour Direwolf20 1.18 Modpack v.2234 should be downloaded. ### Screenshots _No response_ ### Additional information I've tried using the Windows server and the download also didn't work correctly.
1.0
[Bug]: Direwolf20 1.18 Linux fail to download - ### Modpack Direwolf20 1.18 ### Modpack version 2234 ### Log Files 491 Successful 266 Failed https://pste.ch/zoyojitozo.pl Continue anyway. https://pste.ch/jovurutaxi.sql ### Describe the bug With linux, the server does not download fully, having 266 Failed downloads. Additionally, the start.sh is not formatted correctly. https://pste.ch/ijopamevun.bash ### Steps to reproduce 1. wget https://feed-the-beast.com/modpack/95_ftb_presents_direwolf20_1_18/server-files 2. chmod +x linux 3. ./linux 95 4. follow prompts ### Expected behaviour Direwolf20 1.18 Modpack v.2234 should be downloaded. ### Screenshots _No response_ ### Additional information I've tried using the Windows server and the download also didn't work correctly.
infrastructure
linux fail to download modpack modpack version log files successful failed continue anyway describe the bug with linux the server does not download fully having failed downloads additionally the start sh is not formatted correctly steps to reproduce wget chmod x linux linux follow prompts expected behaviour modpack v should be downloaded screenshots no response additional information i ve tried using the windows server and the download also didn t work correctly
1
14,348
10,760,573,873
IssuesEvent
2019-10-31 18:53:17
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Integration tests sometimes have broken installation
Area-Infrastructure Integration-Test
Tests: Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorList Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorsDuringMethodBodyEditing Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicSquigglesNetCore.VerifySemanticErrorSquiggles Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorsDuringMethodBodyEditing Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorList Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorLevelWarning Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySemanticErrorSquiggles Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySyntaxErrorSquiggles Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.OpenCSharpThenVBSolution Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.ProjectReference
1.0
Integration tests sometimes have broken installation - Tests: Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorList Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicErrorListNetCore.ErrorsDuringMethodBodyEditing Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicSquigglesNetCore.VerifySemanticErrorSquiggles Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorsDuringMethodBodyEditing Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorList Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpErrorListNetCore.ErrorLevelWarning Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySemanticErrorSquiggles Roslyn.VisualStudio.IntegrationTests.CSharp.CSharpSquigglesNetCore.VerifySyntaxErrorSquiggles Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.OpenCSharpThenVBSolution Roslyn.VisualStudio.IntegrationTests.Workspace.WorkspacesNetCore.ProjectReference
infrastructure
integration tests sometimes have broken installation tests roslyn visualstudio integrationtests visualbasic basicerrorlistnetcore errorlist roslyn visualstudio integrationtests visualbasic basicerrorlistnetcore errorsduringmethodbodyediting roslyn visualstudio integrationtests visualbasic basicsquigglesnetcore verifysemanticerrorsquiggles roslyn visualstudio integrationtests csharp csharperrorlistnetcore errorsduringmethodbodyediting roslyn visualstudio integrationtests csharp csharperrorlistnetcore errorlist roslyn visualstudio integrationtests csharp csharperrorlistnetcore errorlevelwarning roslyn visualstudio integrationtests csharp csharpsquigglesnetcore verifysemanticerrorsquiggles roslyn visualstudio integrationtests csharp csharpsquigglesnetcore verifysyntaxerrorsquiggles roslyn visualstudio integrationtests workspace workspacesnetcore opencsharpthenvbsolution roslyn visualstudio integrationtests workspace workspacesnetcore projectreference
1
202,684
7,051,315,157
IssuesEvent
2018-01-03 11:09:41
fusetools/fuselibs-public
https://api.github.com/repos/fusetools/fuselibs-public
closed
Many `...Operator` classes don't deal with invalid input
Priority: High Severity: Bug
There was no mechanism to report lost data before, so these were likely forwarding null. That would explain many of the odd errors I've seen using expression. The operators can now lose data and they should. Many, like in "Operators.uno" call functions like `Marshal.Add` and assume it works. There is no checking of the argument types. Some of the conversions were handled with exceptions before. Exceptions are not properly handled in the exception hierarchy, that is, they don't result in lost data. We could instead introduce a special exception and handle it, but that feels far less clean than explicitly handling invalid conversion (these are an expecte condition with user code and bindings).
1.0
Many `...Operator` classes don't deal with invalid input - There was no mechanism to report lost data before, so these were likely forwarding null. That would explain many of the odd errors I've seen using expression. The operators can now lose data and they should. Many, like in "Operators.uno" call functions like `Marshal.Add` and assume it works. There is no checking of the argument types. Some of the conversions were handled with exceptions before. Exceptions are not properly handled in the exception hierarchy, that is, they don't result in lost data. We could instead introduce a special exception and handle it, but that feels far less clean than explicitly handling invalid conversion (these are an expecte condition with user code and bindings).
non_infrastructure
many operator classes don t deal with invalid input there was no mechanism to report lost data before so these were likely forwarding null that would explain many of the odd errors i ve seen using expression the operators can now lose data and they should many like in operators uno call functions like marshal add and assume it works there is no checking of the argument types some of the conversions were handled with exceptions before exceptions are not properly handled in the exception hierarchy that is they don t result in lost data we could instead introduce a special exception and handle it but that feels far less clean than explicitly handling invalid conversion these are an expecte condition with user code and bindings
0
221,704
17,365,858,755
IssuesEvent
2021-07-30 07:09:52
hoechenberger/mne-installers
https://api.github.com/repos/hoechenberger/mne-installers
opened
Run CI tests for Windows installer too
tests windows
Currently, we only build the package, but don't install & test it.
1.0
Run CI tests for Windows installer too - Currently, we only build the package, but don't install & test it.
non_infrastructure
run ci tests for windows installer too currently we only build the package but don t install test it
0
27,300
21,580,496,156
IssuesEvent
2022-05-02 18:10:37
observingClouds/xbitinfo
https://api.github.com/repos/observingClouds/xbitinfo
closed
Reduce dependencies
infrastructure
pip allows extras when installing, see https://github.com/pangeo-data/climpred/blob/main/setup.py or https://github.com/pydata/xarray/blob/main/setup.cfg for non essential deps all viz libs could be grouped, also parallel for dask and prefect
1.0
Reduce dependencies - pip allows extras when installing, see https://github.com/pangeo-data/climpred/blob/main/setup.py or https://github.com/pydata/xarray/blob/main/setup.cfg for non essential deps all viz libs could be grouped, also parallel for dask and prefect
infrastructure
reduce dependencies pip allows extras when installing see or for non essential deps all viz libs could be grouped also parallel for dask and prefect
1
97,918
20,574,002,159
IssuesEvent
2022-03-04 01:09:55
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
insights: modify API to take into account frozen insights
team/code-insights backend
1. Add a new field `frozen` to the `InsightView` type in the API and make sure it accurately reflects whether or not the insight is frozen, (based on the `is_frozen` field in `insight_view`.) 2. When returning insights, do not return any `dataSeries` data for frozen insights. 3. Potentially other restrictions (creating new insights, editing frozen insights, etc.) on the API based on whether or not the user has a Code Insights license. (TBD)
1.0
insights: modify API to take into account frozen insights - 1. Add a new field `frozen` to the `InsightView` type in the API and make sure it accurately reflects whether or not the insight is frozen, (based on the `is_frozen` field in `insight_view`.) 2. When returning insights, do not return any `dataSeries` data for frozen insights. 3. Potentially other restrictions (creating new insights, editing frozen insights, etc.) on the API based on whether or not the user has a Code Insights license. (TBD)
non_infrastructure
insights modify api to take into account frozen insights add a new field frozen to the insightview type in the api and make sure it accurately reflects whether or not the insight is frozen based on the is frozen field in insight view when returning insights do not return any dataseries data for frozen insights potentially other restrictions creating new insights editing frozen insights etc on the api based on whether or not the user has a code insights license tbd
0
96,369
10,932,060,853
IssuesEvent
2019-11-23 15:05:57
aaFn/Bookmark-search-plus-2
https://api.github.com/repos/aaFn/Bookmark-search-plus-2
closed
Can't find ⓘ in address bar
documentation
The instrtuctions for Bookmark-search-plus-2 at Firefox: about additions say, "drag the ⓘ in the address bar or drag web page links into the sidebar, and insert them in the tree at the place you like". I can't find the ⓘ in the address bar, although I used to be able to. What am I doing wrong?
1.0
Can't find ⓘ in address bar - The instrtuctions for Bookmark-search-plus-2 at Firefox: about additions say, "drag the ⓘ in the address bar or drag web page links into the sidebar, and insert them in the tree at the place you like". I can't find the ⓘ in the address bar, although I used to be able to. What am I doing wrong?
non_infrastructure
can t find ⓘ in address bar the instrtuctions for bookmark search plus at firefox about additions say drag the ⓘ in the address bar or drag web page links into the sidebar and insert them in the tree at the place you like i can t find the ⓘ in the address bar although i used to be able to what am i doing wrong
0
470,124
13,531,461,881
IssuesEvent
2020-09-15 21:41:51
grpc/grpc
https://api.github.com/repos/grpc/grpc
opened
Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x
kind/bug priority/P0/RELEASE BLOCKER priority/P2
This is a blocker for v1.31.x and v1.32.x patches only. Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x when we patch these branches. Mark this issue resolved when we no longer need to patch these branches, i.e., when 1.34 is released.
2.0
Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x - This is a blocker for v1.31.x and v1.32.x patches only. Backport https://github.com/grpc/grpc/pull/24150 to v1.31.x and v1.32.x when we patch these branches. Mark this issue resolved when we no longer need to patch these branches, i.e., when 1.34 is released.
non_infrastructure
backport to x and x this is a blocker for x and x patches only backport to x and x when we patch these branches mark this issue resolved when we no longer need to patch these branches i e when is released
0
16,046
11,809,026,928
IssuesEvent
2020-03-19 14:20:34
google/iree
https://api.github.com/repos/google/iree
closed
LLVMX86CodeGen not linking correctly (cmake/MSVC)
infrastructure
Library is not found during linking of any iree_cc_binary that transitively depends on the llvmjit_driver_module. May be a bad LLVM configuration in our options (not telling the x86 codegen backend to build?)
1.0
LLVMX86CodeGen not linking correctly (cmake/MSVC) - Library is not found during linking of any iree_cc_binary that transitively depends on the llvmjit_driver_module. May be a bad LLVM configuration in our options (not telling the x86 codegen backend to build?)
infrastructure
not linking correctly cmake msvc library is not found during linking of any iree cc binary that transitively depends on the llvmjit driver module may be a bad llvm configuration in our options not telling the codegen backend to build
1
97,133
20,169,263,072
IssuesEvent
2022-02-10 08:54:57
GeoNode/geonode
https://api.github.com/repos/GeoNode/geonode
closed
Drop legacy dataset upload page and implement client upload url hookset
enhancement code quality master
The legacy upload page is removed. Redirection to the client upload url is implemented with a hookset. This is adopted inside the remaining upload pages (step views).
1.0
Drop legacy dataset upload page and implement client upload url hookset - The legacy upload page is removed. Redirection to the client upload url is implemented with a hookset. This is adopted inside the remaining upload pages (step views).
non_infrastructure
drop legacy dataset upload page and implement client upload url hookset the legacy upload page is removed redirection to the client upload url is implemented with a hookset this is adopted inside the remaining upload pages step views
0
159,688
20,085,892,657
IssuesEvent
2022-02-05 01:07:59
AkshayMukkavilli/Tensorflow
https://api.github.com/repos/AkshayMukkavilli/Tensorflow
opened
CVE-2021-41200 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2021-41200 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /Tensorflow/src/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. In affected versions if `tf.summary.create_file_writer` is called with non-scalar arguments code crashes due to a `CHECK`-fail. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range. <p>Publish Date: 2021-11-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41200>CVE-2021-41200</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f</a></p> <p>Release Date: 2021-11-05</p> <p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-41200 (Medium) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-41200 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /Tensorflow/src/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_5ea86033-7612-4210-97f3-8edb65806ddf/20190525011619_2843/20190525011537_depth_0/2/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64/tensorflow-1.13.1.data/purelib/tensorflow</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an open source platform for machine learning. In affected versions if `tf.summary.create_file_writer` is called with non-scalar arguments code crashes due to a `CHECK`-fail. The fix will be included in TensorFlow 2.7.0. We will also cherrypick this commit on TensorFlow 2.6.1, TensorFlow 2.5.2, and TensorFlow 2.4.4, as these are also affected and still in supported range. <p>Publish Date: 2021-11-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-41200>CVE-2021-41200</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-gh8h-7j2j-qv4f</a></p> <p>Release Date: 2021-11-05</p> <p>Fix Resolution: tensorflow - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-cpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0;tensorflow-gpu - 2.4.4, 2.5.2, 2.6.1, 2.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve medium detected in tensorflow whl cve medium severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file tensorflow src requirements txt path to vulnerable library tesource archiveextractor depth tensorflow tensorflow data purelib tensorflow dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning in affected versions if tf summary create file writer is called with non scalar arguments code crashes due to a check fail the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
116,155
11,902,477,308
IssuesEvent
2020-03-30 14:01:26
IBM-Blockchain/ansible-collection
https://api.github.com/repos/IBM-Blockchain/ansible-collection
closed
Add missing URLS to Galaxy metadata
documentation
Galaxy is missing the documentation, homepage, and issues URLs - now we have these, we can add them in.
1.0
Add missing URLS to Galaxy metadata - Galaxy is missing the documentation, homepage, and issues URLs - now we have these, we can add them in.
non_infrastructure
add missing urls to galaxy metadata galaxy is missing the documentation homepage and issues urls now we have these we can add them in
0
34,104
28,242,166,387
IssuesEvent
2023-04-06 08:03:46
localstack/localstack
https://api.github.com/repos/localstack/localstack
closed
bug: Cannot connect to localstack v2.0 in EKS cluster
type: bug status: triage needed area: infrastructure area: networking
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior We have the image localstack/localstack:latest-amd64 deployed into an EKS cluster running SNS and SQS - this picks up the version 2.0 release of localstack. When the service we have deployed in the same cluster attempts to connect to localstack we receive an error - `Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused ` It works fine if we revert back to the image - `localstack/localstack:1.4-amd64` We have sen this error since the release of 2.0 (which our build picks up as being the latest version. The full stack is attached. ### Expected Behavior When the service in the EKS cluster starts up, it shoudl connect successfully to localstack. ### How are you starting LocalStack? Custom (please describe below) ### Steps To Reproduce #### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`) This is done using a helm chart which is passed into a docker run command - docker run -it ### Environment ```markdown - OS:linux/amd64 JDK 17 - LocalStack:2.0 AMD64 ``` ### Anything else? Don't seem able to attach log, but have pasted here - `Caused by: com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused at com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleRetryableException(AmazonHttpClient.java:1219) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1165) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?] │ Caused by: java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.NioSocketImpl.timedFinishConnect(NioSocketImpl.java:549) ~[?:?] at sun.nio.ch.NioSocketImpl.connect(NioSocketImpl.java:597) ~[?:?] at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:327) ~[?:?] at java.net.Socket.connect(Socket.java:633) ~[?:?] at org.apache.http.conn.socket.PlainConnectionSocketFactory.connectSocket(PlainConnectionSocketFactory.java:75) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.conn.DefaultHttpClientConnectionOperator.connect(DefaultHttpClientConnectionOperator.java:142) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.conn.PoolingHttpClientConnectionManager.connect(PoolingHttpClientConnectionManager.java:376) ~[httpclient-4.5.13.jar:4.5.13] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at com.amazonaws.http.conn.ClientConnectionManagerFactory$Handler.invoke(ClientConnectionManagerFactory.java:76) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.conn.$Proxy133.connect(Unknown Source) ~[?:?] at org.apache.http.impl.execchain.MainClientExec.establishRoute(MainClientExec.java:393) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:236) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:186) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56) ~[httpclient-4.5.13.jar:4.5.13] at com.amazonaws.http.apache.client.impl.SdkHttpClient.execute(SdkHttpClient.java:72) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeOneRequest(AmazonHttpClient.java:1346) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1157) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?] `
1.0
bug: Cannot connect to localstack v2.0 in EKS cluster - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior We have the image localstack/localstack:latest-amd64 deployed into an EKS cluster running SNS and SQS - this picks up the version 2.0 release of localstack. When the service we have deployed in the same cluster attempts to connect to localstack we receive an error - `Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused ` It works fine if we revert back to the image - `localstack/localstack:1.4-amd64` We have sen this error since the release of 2.0 (which our build picks up as being the latest version. The full stack is attached. ### Expected Behavior When the service in the EKS cluster starts up, it shoudl connect successfully to localstack. ### How are you starting LocalStack? Custom (please describe below) ### Steps To Reproduce #### How are you starting localstack (e.g., `bin/localstack` command, arguments, or `docker-compose.yml`) This is done using a helm chart which is passed into a docker run command - docker run -it ### Environment ```markdown - OS:linux/amd64 JDK 17 - LocalStack:2.0 AMD64 ``` ### Anything else? Don't seem able to attach log, but have pasted here - `Caused by: com.amazonaws.SdkClientException: Unable to execute HTTP request: Connect to aws:4566 [aws/172.20.92.156] failed: Connection refused at com.amazonaws.http.AmazonHttpClient$RequestExecutor.handleRetryableException(AmazonHttpClient.java:1219) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1165) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?] │ Caused by: java.net.ConnectException: Connection refused at sun.nio.ch.Net.pollConnect(Native Method) ~[?:?] at sun.nio.ch.Net.pollConnectNow(Net.java:672) ~[?:?] at sun.nio.ch.NioSocketImpl.timedFinishConnect(NioSocketImpl.java:549) ~[?:?] at sun.nio.ch.NioSocketImpl.connect(NioSocketImpl.java:597) ~[?:?] at java.net.SocksSocketImpl.connect(SocksSocketImpl.java:327) ~[?:?] at java.net.Socket.connect(Socket.java:633) ~[?:?] at org.apache.http.conn.socket.PlainConnectionSocketFactory.connectSocket(PlainConnectionSocketFactory.java:75) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.conn.DefaultHttpClientConnectionOperator.connect(DefaultHttpClientConnectionOperator.java:142) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.conn.PoolingHttpClientConnectionManager.connect(PoolingHttpClientConnectionManager.java:376) ~[httpclient-4.5.13.jar:4.5.13] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:568) ~[?:?] at com.amazonaws.http.conn.ClientConnectionManagerFactory$Handler.invoke(ClientConnectionManagerFactory.java:76) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.conn.$Proxy133.connect(Unknown Source) ~[?:?] at org.apache.http.impl.execchain.MainClientExec.establishRoute(MainClientExec.java:393) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.execchain.MainClientExec.execute(MainClientExec.java:236) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.execchain.ProtocolExec.execute(ProtocolExec.java:186) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.client.InternalHttpClient.doExecute(InternalHttpClient.java:185) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:83) ~[httpclient-4.5.13.jar:4.5.13] at org.apache.http.impl.client.CloseableHttpClient.execute(CloseableHttpClient.java:56) ~[httpclient-4.5.13.jar:4.5.13] at com.amazonaws.http.apache.client.impl.SdkHttpClient.execute(SdkHttpClient.java:72) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeOneRequest(AmazonHttpClient.java:1346) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeHelper(AmazonHttpClient.java:1157) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.doExecute(AmazonHttpClient.java:814) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.executeWithTimer(AmazonHttpClient.java:781) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.execute(AmazonHttpClient.java:755) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutor.access$500(AmazonHttpClient.java:715) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient$RequestExecutionBuilderImpl.execute(AmazonHttpClient.java:697) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:561) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.http.AmazonHttpClient.execute(AmazonHttpClient.java:541) ~[aws-java-sdk-core-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.doInvoke(AmazonSQSClient.java:2329) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2296) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.invoke(AmazonSQSClient.java:2285) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.executeCreateQueue(AmazonSQSClient.java:890) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.AmazonSQSClient.createQueue(AmazonSQSClient.java:859) ~[aws-java-sdk-sqs-1.12.232.jar:?] at com.amazonaws.services.sqs.buffered.AmazonSQSBufferedAsyncClient.createQueue(AmazonSQSBufferedAsyncClient.java:325) ~[aws-java-sdk-sqs-1.12.232.jar:?] `
infrastructure
bug cannot connect to localstack in eks cluster is there an existing issue for this i have searched the existing issues current behavior we have the image localstack localstack latest deployed into an eks cluster running sns and sqs this picks up the version release of localstack when the service we have deployed in the same cluster attempts to connect to localstack we receive an error unable to execute http request connect to aws failed connection refused it works fine if we revert back to the image localstack localstack we have sen this error since the release of which our build picks up as being the latest version the full stack is attached expected behavior when the service in the eks cluster starts up it shoudl connect successfully to localstack how are you starting localstack custom please describe below steps to reproduce how are you starting localstack e g bin localstack command arguments or docker compose yml this is done using a helm chart which is passed into a docker run command docker run it environment markdown os linux jdk localstack anything else don t seem able to attach log but have pasted here caused by com amazonaws sdkclientexception unable to execute http request connect to aws failed connection refused at com amazonaws http amazonhttpclient requestexecutor handleretryableexception amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executehelper amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor doexecute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executewithtimer amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor execute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor access amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutionbuilderimpl execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws services sqs amazonsqsclient doinvoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient executecreatequeue amazonsqsclient java at com amazonaws services sqs amazonsqsclient createqueue amazonsqsclient java at com amazonaws services sqs buffered amazonsqsbufferedasyncclient createqueue amazonsqsbufferedasyncclient java │ caused by java net connectexception connection refused at sun nio ch net pollconnect native method at sun nio ch net pollconnectnow net java at sun nio ch niosocketimpl timedfinishconnect niosocketimpl java at sun nio ch niosocketimpl connect niosocketimpl java at java net sockssocketimpl connect sockssocketimpl java at java net socket connect socket java at org apache http conn socket plainconnectionsocketfactory connectsocket plainconnectionsocketfactory java at org apache http impl conn defaulthttpclientconnectionoperator connect defaulthttpclientconnectionoperator java at org apache http impl conn poolinghttpclientconnectionmanager connect poolinghttpclientconnectionmanager java at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com amazonaws http conn clientconnectionmanagerfactory handler invoke clientconnectionmanagerfactory java at com amazonaws http conn connect unknown source at org apache http impl execchain mainclientexec establishroute mainclientexec java at org apache http impl execchain mainclientexec execute mainclientexec java at org apache http impl execchain protocolexec execute protocolexec java at org apache http impl client internalhttpclient doexecute internalhttpclient java at org apache http impl client closeablehttpclient execute closeablehttpclient java at org apache http impl client closeablehttpclient execute closeablehttpclient java at com amazonaws http apache client impl sdkhttpclient execute sdkhttpclient java at com amazonaws http amazonhttpclient requestexecutor executeonerequest amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executehelper amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor doexecute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor executewithtimer amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor execute amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutor access amazonhttpclient java at com amazonaws http amazonhttpclient requestexecutionbuilderimpl execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws http amazonhttpclient execute amazonhttpclient java at com amazonaws services sqs amazonsqsclient doinvoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient invoke amazonsqsclient java at com amazonaws services sqs amazonsqsclient executecreatequeue amazonsqsclient java at com amazonaws services sqs amazonsqsclient createqueue amazonsqsclient java at com amazonaws services sqs buffered amazonsqsbufferedasyncclient createqueue amazonsqsbufferedasyncclient java
1
30,439
4,615,248,714
IssuesEvent
2016-09-26 00:46:58
Semantic-Org/Semantic-UI
https://api.github.com/repos/Semantic-Org/Semantic-UI
closed
Input type text only displays placeholder text on focus event
Needs Test Case Read the Contributing Guide / Readme Usage Question
I just updated to semantic ui version 2.2.4 in order to use some new features but after upgrading semantic-ui all my inputs only show placeholder text on the user clicks on the input (focus). I'm trying with a simple input: `<div className="ui icon input"> <input type="text" placeholder="Search for something"/> <i className="search icon"></i> </div> </div>`
1.0
Input type text only displays placeholder text on focus event - I just updated to semantic ui version 2.2.4 in order to use some new features but after upgrading semantic-ui all my inputs only show placeholder text on the user clicks on the input (focus). I'm trying with a simple input: `<div className="ui icon input"> <input type="text" placeholder="Search for something"/> <i className="search icon"></i> </div> </div>`
non_infrastructure
input type text only displays placeholder text on focus event i just updated to semantic ui version in order to use some new features but after upgrading semantic ui all my inputs only show placeholder text on the user clicks on the input focus i m trying with a simple input
0
23,211
4,894,243,298
IssuesEvent
2016-11-19 05:58:26
F5Networks/f5-icontrol-rest-python
https://api.github.com/repos/F5Networks/f5-icontrol-rest-python
closed
Update README...
documentation
Need to clarify differences (and dependencies) between this repo and f5-common-python.
1.0
Update README... - Need to clarify differences (and dependencies) between this repo and f5-common-python.
non_infrastructure
update readme need to clarify differences and dependencies between this repo and common python
0
20,662
14,099,320,726
IssuesEvent
2020-11-06 01:05:06
noahtalerman/test-issues-kolide
https://api.github.com/repos/noahtalerman/test-issues-kolide
opened
[CLOSED] cleaning up some development files
Component: Development Infrastructure
<a href="https://github.com/marpaia"><img src="https://avatars2.githubusercontent.com/u/927168?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [marpaia](https://github.com/marpaia)** _Monday Sep 05, 2016 at 21:55 GMT_ _Originally opened as https://github.com/kolide/fleet/pull/115_ ---- ---- _**[marpaia](https://github.com/marpaia)** included the following code: https://github.com/kolide/fleet/pull/115/commits_
1.0
[CLOSED] cleaning up some development files - <a href="https://github.com/marpaia"><img src="https://avatars2.githubusercontent.com/u/927168?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [marpaia](https://github.com/marpaia)** _Monday Sep 05, 2016 at 21:55 GMT_ _Originally opened as https://github.com/kolide/fleet/pull/115_ ---- ---- _**[marpaia](https://github.com/marpaia)** included the following code: https://github.com/kolide/fleet/pull/115/commits_
infrastructure
cleaning up some development files issue by monday sep at gmt originally opened as included the following code
1
481,212
13,882,047,082
IssuesEvent
2020-10-18 04:27:52
python/mypy
https://api.github.com/repos/python/mypy
closed
Can't assign to ()
bug false-positive priority-2-low
Mypy incorrectly complains about assignments to `()`. The simplest example of this is ```python () = [] ``` This code executes without a problem, but mypy says `error: can't assign to ()` A more real world use case might be code like this ```python (a, b), () = [[1, 2], []] ``` In this case the assignment to `()` is a concise way of asserting that the second list is empty. * What is the actual behavior/output? `error: can't assign to ()` * What is the behavior/output you expect? No type checking errors * What are the versions of mypy and Python you are using? A build from the latest commit as of writing this, 614090b7 * What are the mypy flags you are using? No flags If this is a change you decide would be good then I believe it could be implemented by removing the the check on line 1930 of `mypy/semanal.py` and updating the tests.
1.0
Can't assign to () - Mypy incorrectly complains about assignments to `()`. The simplest example of this is ```python () = [] ``` This code executes without a problem, but mypy says `error: can't assign to ()` A more real world use case might be code like this ```python (a, b), () = [[1, 2], []] ``` In this case the assignment to `()` is a concise way of asserting that the second list is empty. * What is the actual behavior/output? `error: can't assign to ()` * What is the behavior/output you expect? No type checking errors * What are the versions of mypy and Python you are using? A build from the latest commit as of writing this, 614090b7 * What are the mypy flags you are using? No flags If this is a change you decide would be good then I believe it could be implemented by removing the the check on line 1930 of `mypy/semanal.py` and updating the tests.
non_infrastructure
can t assign to mypy incorrectly complains about assignments to the simplest example of this is python this code executes without a problem but mypy says error can t assign to a more real world use case might be code like this python a b in this case the assignment to is a concise way of asserting that the second list is empty what is the actual behavior output error can t assign to what is the behavior output you expect no type checking errors what are the versions of mypy and python you are using a build from the latest commit as of writing this what are the mypy flags you are using no flags if this is a change you decide would be good then i believe it could be implemented by removing the the check on line of mypy semanal py and updating the tests
0
312,902
9,554,724,834
IssuesEvent
2019-05-02 23:12:17
PerezArturo/Medum
https://api.github.com/repos/PerezArturo/Medum
reopened
HU 1- autentificación
priority: high type: US
-Descripción. Yo como dueño o interesado en rentar una propiedad deseo iniciar sesión mediante mi numero de teléfono. - Criterios de Aceptación. 1. Dado que el usuario desea iniciar sesión con su numero teléfono deberá ingresar numero para iniciar 2. dado que el usuario ingrese numero y no este registrado deberá registrarse primero para poder iniciar
1.0
HU 1- autentificación - -Descripción. Yo como dueño o interesado en rentar una propiedad deseo iniciar sesión mediante mi numero de teléfono. - Criterios de Aceptación. 1. Dado que el usuario desea iniciar sesión con su numero teléfono deberá ingresar numero para iniciar 2. dado que el usuario ingrese numero y no este registrado deberá registrarse primero para poder iniciar
non_infrastructure
hu autentificación descripción yo como dueño o interesado en rentar una propiedad deseo iniciar sesión mediante mi numero de teléfono criterios de aceptación dado que el usuario desea iniciar sesión con su numero teléfono deberá ingresar numero para iniciar dado que el usuario ingrese numero y no este registrado deberá registrarse primero para poder iniciar
0
110,512
16,980,344,145
IssuesEvent
2021-06-30 08:02:45
samuelteixeiras/learnRuby
https://api.github.com/repos/samuelteixeiras/learnRuby
opened
CVE-2012-2661 (Medium) detected in activerecord-3.0.9.gem
security vulnerability
## CVE-2012-2661 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activerecord-3.0.9.gem</b></p></summary> <p>Databases on Rails. Build a persistent domain model by mapping database tables to Ruby classes. Strong conventions for associations, validations, aggregations, migrations, and testing come baked-in.</p> <p>Library home page: <a href="https://rubygems.org/gems/activerecord-3.0.9.gem">https://rubygems.org/gems/activerecord-3.0.9.gem</a></p> <p> Dependency Hierarchy: - rails-3.0.9.gem (Root Library) - :x: **activerecord-3.0.9.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samuelteixeiras/learnRuby/commit/0c785fd9400921392b8ee5e3e166f30364359ecc">0c785fd9400921392b8ee5e3e166f30364359ecc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Active Record component in Ruby on Rails 3.0.x before 3.0.13, 3.1.x before 3.1.5, and 3.2.x before 3.2.4 does not properly implement the passing of request data to a where method in an ActiveRecord class, which allows remote attackers to conduct certain SQL injection attacks via nested query parameters that leverage unintended recursion, a related issue to CVE-2012-2695. <p>Publish Date: 2012-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-2661>CVE-2012-2661</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-2661">https://nvd.nist.gov/vuln/detail/CVE-2012-2661</a></p> <p>Release Date: 2012-06-22</p> <p>Fix Resolution: 3.0.13,3.1.5,3.2.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2012-2661 (Medium) detected in activerecord-3.0.9.gem - ## CVE-2012-2661 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>activerecord-3.0.9.gem</b></p></summary> <p>Databases on Rails. Build a persistent domain model by mapping database tables to Ruby classes. Strong conventions for associations, validations, aggregations, migrations, and testing come baked-in.</p> <p>Library home page: <a href="https://rubygems.org/gems/activerecord-3.0.9.gem">https://rubygems.org/gems/activerecord-3.0.9.gem</a></p> <p> Dependency Hierarchy: - rails-3.0.9.gem (Root Library) - :x: **activerecord-3.0.9.gem** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samuelteixeiras/learnRuby/commit/0c785fd9400921392b8ee5e3e166f30364359ecc">0c785fd9400921392b8ee5e3e166f30364359ecc</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The Active Record component in Ruby on Rails 3.0.x before 3.0.13, 3.1.x before 3.1.5, and 3.2.x before 3.2.4 does not properly implement the passing of request data to a where method in an ActiveRecord class, which allows remote attackers to conduct certain SQL injection attacks via nested query parameters that leverage unintended recursion, a related issue to CVE-2012-2695. <p>Publish Date: 2012-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-2661>CVE-2012-2661</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-2661">https://nvd.nist.gov/vuln/detail/CVE-2012-2661</a></p> <p>Release Date: 2012-06-22</p> <p>Fix Resolution: 3.0.13,3.1.5,3.2.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve medium detected in activerecord gem cve medium severity vulnerability vulnerable library activerecord gem databases on rails build a persistent domain model by mapping database tables to ruby classes strong conventions for associations validations aggregations migrations and testing come baked in library home page a href dependency hierarchy rails gem root library x activerecord gem vulnerable library found in head commit a href found in base branch master vulnerability details the active record component in ruby on rails x before x before and x before does not properly implement the passing of request data to a where method in an activerecord class which allows remote attackers to conduct certain sql injection attacks via nested query parameters that leverage unintended recursion a related issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
113,709
11,812,538,054
IssuesEvent
2020-03-19 20:22:20
wbarnha/gr-gmuground
https://api.github.com/repos/wbarnha/gr-gmuground
closed
Reorganize apps and examples
documentation
There's a lot of clutter and unexplained components in the software. We should also integrate a method of compiling the hierarchical blocks automatically in `setup.sh` for convenience.
1.0
Reorganize apps and examples - There's a lot of clutter and unexplained components in the software. We should also integrate a method of compiling the hierarchical blocks automatically in `setup.sh` for convenience.
non_infrastructure
reorganize apps and examples there s a lot of clutter and unexplained components in the software we should also integrate a method of compiling the hierarchical blocks automatically in setup sh for convenience
0
13,123
10,131,762,946
IssuesEvent
2019-08-01 20:24:24
HumanCellAtlas/secondary-analysis
https://api.github.com/repos/HumanCellAtlas/secondary-analysis
closed
Tex reports that some Jenkins jobs write into `~devopsci` on every run.
infrastructure
[~teixeira] reports via [~ferrara]: can someone figure out why dsde-jenkins is writing `jenkins-dsde-pipelines-run-automated-test` `jenkins-gotc-arrays-dev-nightly-test` and several other directories to `devopsci`'s home directory daily? That probably comes from this code. https://github.com/broadinstitute/dsp-jenkins/blob/fd2fd5720925d68a0baa3bc2b86021702dd3edd0/src/main/groovy/org/broadinstitute/dspjenkins/CodeSnippets.groovy#L20 Looks like anything running with `picard` authorization will write there. Been doing that for a year or so. It looks like at least these jobs are affected. `jenkins-dsde-pipelines-run-automated-test` `jenkins-picard-private-dev-deploy` `jenkins-picard-private-staging-deploy` `jenkins-zamboni-nightly-workflows-test` `jenkins-zamboni-nightly-wgs-job` ┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-316)
1.0
Tex reports that some Jenkins jobs write into `~devopsci` on every run. - [~teixeira] reports via [~ferrara]: can someone figure out why dsde-jenkins is writing `jenkins-dsde-pipelines-run-automated-test` `jenkins-gotc-arrays-dev-nightly-test` and several other directories to `devopsci`'s home directory daily? That probably comes from this code. https://github.com/broadinstitute/dsp-jenkins/blob/fd2fd5720925d68a0baa3bc2b86021702dd3edd0/src/main/groovy/org/broadinstitute/dspjenkins/CodeSnippets.groovy#L20 Looks like anything running with `picard` authorization will write there. Been doing that for a year or so. It looks like at least these jobs are affected. `jenkins-dsde-pipelines-run-automated-test` `jenkins-picard-private-dev-deploy` `jenkins-picard-private-staging-deploy` `jenkins-zamboni-nightly-workflows-test` `jenkins-zamboni-nightly-wgs-job` ┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-316)
infrastructure
tex reports that some jenkins jobs write into devopsci on every run reports via can someone figure out why dsde jenkins is writing jenkins dsde pipelines run automated test jenkins gotc arrays dev nightly test and several other directories to devopsci s home directory daily that probably comes from this code looks like anything running with picard authorization will write there been doing that for a year or so it looks like at least these jobs are affected jenkins dsde pipelines run automated test jenkins picard private dev deploy jenkins picard private staging deploy jenkins zamboni nightly workflows test jenkins zamboni nightly wgs job ┆issue is synchronized with this
1
149,317
5,716,327,984
IssuesEvent
2017-04-19 14:56:13
Flexget/Flexget
https://api.github.com/repos/Flexget/Flexget
closed
After timeframe expires, best quality found is downloaded even if a better match to target is available
Enhancement Low priority
From http://flexget.com/ticket/3105 Consider a configuration such as ``` timeframe: 2 hours target: 720p webdl quality: 720p-1080p hdtv+ ``` and the following scenario: Flexget finds a suitable episode in two formats: 720p hdtv and 1080p hdtv, whether at the same time or within two hours of each other. Due to them not fulfilling the webdl requirement of the target quality it correctly doesn't download anything yet. The timeframe then expires and there is still no 720p webrip available. At this point, Flexget's current behaviour is to download the best quality it has seen, which in this case is the 1080p hdtv. This may be what is wanted, but it also may not be. It's not what I want, and I would imagine that most users would agree that in this scenario the 720p hdtv would be more appropriate, since the resolution matches the target and the source, though not matching the target, is in the valid quality range. On the other hand the 1080p hdtv, which was downloaded, does not match the target resolution (though it is in the valid range) and the source also does not match the target (though again it is in the valid range). I think the default behaviour should be to match as many of the target requirements as possible, while of course staying within the quality requirements. If it seems not all users would want this new default behaviour, it could be configurable. It could either be a "choose closest match: yes or no" option, or it could be more sophisticated with the "target" option expanded. A user who wants "720p webrip" could somehow specify that the 720p part is important, and the webrip part not so important, or some other such combination. liiight asked in a comment on the old ticket: > if you don't want 1080p to begin with? why even include it in the original requirement? if that is not a valid option on your part, do not include it. > > i'm not saying i'm against adding a 'closest match' feature, that does sound interesting and doable, i'm just wondering. The answer is that I do want 1080p if no 720p is available, but I prefer 720p. At present I have no way to tell Flexget that I _prefer_ a "worse" quality, or rather I do since that's what I'm saying in the "target", but it isn't taken into account when timeframe expires.
1.0
After timeframe expires, best quality found is downloaded even if a better match to target is available - From http://flexget.com/ticket/3105 Consider a configuration such as ``` timeframe: 2 hours target: 720p webdl quality: 720p-1080p hdtv+ ``` and the following scenario: Flexget finds a suitable episode in two formats: 720p hdtv and 1080p hdtv, whether at the same time or within two hours of each other. Due to them not fulfilling the webdl requirement of the target quality it correctly doesn't download anything yet. The timeframe then expires and there is still no 720p webrip available. At this point, Flexget's current behaviour is to download the best quality it has seen, which in this case is the 1080p hdtv. This may be what is wanted, but it also may not be. It's not what I want, and I would imagine that most users would agree that in this scenario the 720p hdtv would be more appropriate, since the resolution matches the target and the source, though not matching the target, is in the valid quality range. On the other hand the 1080p hdtv, which was downloaded, does not match the target resolution (though it is in the valid range) and the source also does not match the target (though again it is in the valid range). I think the default behaviour should be to match as many of the target requirements as possible, while of course staying within the quality requirements. If it seems not all users would want this new default behaviour, it could be configurable. It could either be a "choose closest match: yes or no" option, or it could be more sophisticated with the "target" option expanded. A user who wants "720p webrip" could somehow specify that the 720p part is important, and the webrip part not so important, or some other such combination. liiight asked in a comment on the old ticket: > if you don't want 1080p to begin with? why even include it in the original requirement? if that is not a valid option on your part, do not include it. > > i'm not saying i'm against adding a 'closest match' feature, that does sound interesting and doable, i'm just wondering. The answer is that I do want 1080p if no 720p is available, but I prefer 720p. At present I have no way to tell Flexget that I _prefer_ a "worse" quality, or rather I do since that's what I'm saying in the "target", but it isn't taken into account when timeframe expires.
non_infrastructure
after timeframe expires best quality found is downloaded even if a better match to target is available from consider a configuration such as timeframe hours target webdl quality hdtv and the following scenario flexget finds a suitable episode in two formats hdtv and hdtv whether at the same time or within two hours of each other due to them not fulfilling the webdl requirement of the target quality it correctly doesn t download anything yet the timeframe then expires and there is still no webrip available at this point flexget s current behaviour is to download the best quality it has seen which in this case is the hdtv this may be what is wanted but it also may not be it s not what i want and i would imagine that most users would agree that in this scenario the hdtv would be more appropriate since the resolution matches the target and the source though not matching the target is in the valid quality range on the other hand the hdtv which was downloaded does not match the target resolution though it is in the valid range and the source also does not match the target though again it is in the valid range i think the default behaviour should be to match as many of the target requirements as possible while of course staying within the quality requirements if it seems not all users would want this new default behaviour it could be configurable it could either be a choose closest match yes or no option or it could be more sophisticated with the target option expanded a user who wants webrip could somehow specify that the part is important and the webrip part not so important or some other such combination liiight asked in a comment on the old ticket if you don t want to begin with why even include it in the original requirement if that is not a valid option on your part do not include it i m not saying i m against adding a closest match feature that does sound interesting and doable i m just wondering the answer is that i do want if no is available but i prefer at present i have no way to tell flexget that i prefer a worse quality or rather i do since that s what i m saying in the target but it isn t taken into account when timeframe expires
0
2,429
3,675,582,480
IssuesEvent
2016-02-23 00:15:51
MozillaFoundation/plan
https://api.github.com/repos/MozillaFoundation/plan
opened
Finalize MLN H1 Production Work documents
Convening Innovation Learning Membership MLN Infrastructure p1 production Size: medium Status: on track
#### RASCI R: @hannahkane @ldecoursy A: @chrislarry33 @davidascher S: @cassiemc @simonwex @adamlofting C: CMS team I: eventually everyone #### Overview We've pulled out the production work from the MLN H1 Strategy document, and are aiming to organize and present the work in a way that helps all production and program staff understand how our production work enables MLN success. We're also using these documents for planning purposes—specifically to estimate relative effort, to agree upon KPIs and goals, and eventually to prioritize/order the tasks. #### Links * [Narrative version](https://docs.google.com/document/d/18tg7DHiI5Ilqa2NuooOKx1fh3D69BEeHb9UHWvD6kJU/edit#heading=h.k63854e1w60v) * [Granular, task-oriented version with estimates](https://docs.google.com/spreadsheets/d/1Prh4t_mUJLcsjr9Rxprzoy7TDKtxLISaC8jgYzbuscQ/edit#gid=0) #### Heartbeat deliverables - [ ] Final versions of documents, including relative effort estimates - [ ] Plan to socialize with staff
1.0
Finalize MLN H1 Production Work documents - #### RASCI R: @hannahkane @ldecoursy A: @chrislarry33 @davidascher S: @cassiemc @simonwex @adamlofting C: CMS team I: eventually everyone #### Overview We've pulled out the production work from the MLN H1 Strategy document, and are aiming to organize and present the work in a way that helps all production and program staff understand how our production work enables MLN success. We're also using these documents for planning purposes—specifically to estimate relative effort, to agree upon KPIs and goals, and eventually to prioritize/order the tasks. #### Links * [Narrative version](https://docs.google.com/document/d/18tg7DHiI5Ilqa2NuooOKx1fh3D69BEeHb9UHWvD6kJU/edit#heading=h.k63854e1w60v) * [Granular, task-oriented version with estimates](https://docs.google.com/spreadsheets/d/1Prh4t_mUJLcsjr9Rxprzoy7TDKtxLISaC8jgYzbuscQ/edit#gid=0) #### Heartbeat deliverables - [ ] Final versions of documents, including relative effort estimates - [ ] Plan to socialize with staff
infrastructure
finalize mln production work documents rasci r hannahkane ldecoursy a davidascher s cassiemc simonwex adamlofting c cms team i eventually everyone overview we ve pulled out the production work from the mln strategy document and are aiming to organize and present the work in a way that helps all production and program staff understand how our production work enables mln success we re also using these documents for planning purposes—specifically to estimate relative effort to agree upon kpis and goals and eventually to prioritize order the tasks links heartbeat deliverables final versions of documents including relative effort estimates plan to socialize with staff
1
254,768
21,876,383,920
IssuesEvent
2022-05-19 10:31:34
Tencent/bk-job
https://api.github.com/repos/Tencent/bk-job
closed
bugfix: 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除
kind/bug stage/test stage/uat
**版本/分支/tag** 3.4.4.0 **问题描述** 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 **如何复现** 创建脚本1,上线脚本1 创建模板1,引用脚本1 调试模板1,运行 创建脚本2,上线脚本2 编辑模板1,引用脚本1改为引用脚本2 查看脚本1引用次数,发现模板/执行方案引用次数为 0/1 尝试删除脚本1,提示被引用无法删除 **预期结果** 脚本1不再被引用后,查看脚本1引用次数,模板/执行方案引用次数为 0/0 应当能够正常删除脚本1
1.0
bugfix: 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 - **版本/分支/tag** 3.4.4.0 **问题描述** 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 **如何复现** 创建脚本1,上线脚本1 创建模板1,引用脚本1 调试模板1,运行 创建脚本2,上线脚本2 编辑模板1,引用脚本1改为引用脚本2 查看脚本1引用次数,发现模板/执行方案引用次数为 0/1 尝试删除脚本1,提示被引用无法删除 **预期结果** 脚本1不再被引用后,查看脚本1引用次数,模板/执行方案引用次数为 0/0 应当能够正常删除脚本1
non_infrastructure
bugfix 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 版本 分支 tag 问题描述 统计脚本被引用次数时未排除调试作业模板产生的调试执行方案导致脚本无法删除 如何复现 , , ,运行 , , ,发现模板 执行方案引用次数为 ,提示被引用无法删除 预期结果 , ,模板 执行方案引用次数为
0
307,019
23,178,176,407
IssuesEvent
2022-07-31 18:33:34
Vector35/binaryninja-api
https://api.github.com/repos/Vector35/binaryninja-api
closed
Documentation for BinaryView.parse_type_string cut-off mid sentence.
API documentation
The documentation for [BinaryView.parse_type_string](https://api.binary.ninja/binaryninja.binaryview-module.html?highlight=parse_type_string#binaryninja.binaryview.BinaryView.parse_type_string) is missing the latter half of the sentence. I assume it would describe that it allows references to types available in the bv. ![image](https://user-images.githubusercontent.com/4206862/182034279-649650d3-f82d-42b5-be3b-69cf7f480cf8.png)
1.0
Documentation for BinaryView.parse_type_string cut-off mid sentence. - The documentation for [BinaryView.parse_type_string](https://api.binary.ninja/binaryninja.binaryview-module.html?highlight=parse_type_string#binaryninja.binaryview.BinaryView.parse_type_string) is missing the latter half of the sentence. I assume it would describe that it allows references to types available in the bv. ![image](https://user-images.githubusercontent.com/4206862/182034279-649650d3-f82d-42b5-be3b-69cf7f480cf8.png)
non_infrastructure
documentation for binaryview parse type string cut off mid sentence the documentation for is missing the latter half of the sentence i assume it would describe that it allows references to types available in the bv
0
29,350
23,935,685,902
IssuesEvent
2022-09-11 07:26:19
CS3219-AY2223S1/cs3219-project-ay2223s1-g33
https://api.github.com/repos/CS3219-AY2223S1/cs3219-project-ay2223s1-g33
closed
[Matching Service] Set-up Redis Container
Module/Infrastructure Status/High-Priority Type/Feature
## Description Set up the Redis container for queuing. ## Parent Task - #38
1.0
[Matching Service] Set-up Redis Container - ## Description Set up the Redis container for queuing. ## Parent Task - #38
infrastructure
set up redis container description set up the redis container for queuing parent task
1
13,355
8,198,056,341
IssuesEvent
2018-08-31 15:12:07
angular/angular
https://api.github.com/repos/angular/angular
closed
*ngFor Performance Issues on IE 11 and Microsoft Edge
comp: core & compiler comp: performance freq4: critical severity3: broken type: bug/fix
The performance of *ngFor is very slow on IE as compared to Chrome, Firefox and Opera. The below plunker loops through over 1500 items using *ngFor, it takes less than a second in other browsers whereas it takes several seconds in IE 11 and Microsoft Edge. http://plnkr.co/edit/sEujClHmuCbrydIiYQYL?p=preview The code is constantly calling isNan function in es6-shim.js. The problem seems to be due to the reason that set and get functions of the polyfill are extremely slow (compared to their native counterparts) and take most of the time. Please have a look at http://stackoverflow.com/questions/36570532/angular-2-performance-ie11-ngfor and the accepted answer. This performance issue is also noticeable on forms with some dynamically populated drop down lists (8-10 drop downs).
True
*ngFor Performance Issues on IE 11 and Microsoft Edge - The performance of *ngFor is very slow on IE as compared to Chrome, Firefox and Opera. The below plunker loops through over 1500 items using *ngFor, it takes less than a second in other browsers whereas it takes several seconds in IE 11 and Microsoft Edge. http://plnkr.co/edit/sEujClHmuCbrydIiYQYL?p=preview The code is constantly calling isNan function in es6-shim.js. The problem seems to be due to the reason that set and get functions of the polyfill are extremely slow (compared to their native counterparts) and take most of the time. Please have a look at http://stackoverflow.com/questions/36570532/angular-2-performance-ie11-ngfor and the accepted answer. This performance issue is also noticeable on forms with some dynamically populated drop down lists (8-10 drop downs).
non_infrastructure
ngfor performance issues on ie and microsoft edge the performance of ngfor is very slow on ie as compared to chrome firefox and opera the below plunker loops through over items using ngfor it takes less than a second in other browsers whereas it takes several seconds in ie and microsoft edge the code is constantly calling isnan function in shim js the problem seems to be due to the reason that set and get functions of the polyfill are extremely slow compared to their native counterparts and take most of the time please have a look at and the accepted answer this performance issue is also noticeable on forms with some dynamically populated drop down lists drop downs
0
202,928
23,123,248,783
IssuesEvent
2022-07-28 01:02:45
lukebroganws/secDevLabs
https://api.github.com/repos/lukebroganws/secDevLabs
closed
CVE-2021-35065 (High) detected in multiple libraries - autoclosed
security vulnerability
## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.2.tgz</b>, <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary> <p> <details><summary><b>glob-parent-3.1.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent directory path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p> <p> Dependency Hierarchy: - react-scripts-3.4.3.tgz (Root Library) - webpack-4.42.0.tgz - watchpack-1.7.4.tgz - watchpack-chokidar2-2.0.0.tgz - chokidar-2.1.8.tgz - :x: **glob-parent-3.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-5.1.2.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.27.7.tgz (Root Library) - chokidar-3.5.3.tgz - :x: **glob-parent-5.1.2.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-3.1.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent directory path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/webpack-dev-server/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/fast-glob/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/webpack-dev-server/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.3.tgz (Root Library) - webpack-dev-server-3.11.0.tgz - chokidar-2.1.8.tgz - :x: **glob-parent-3.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a3/comment-killer/app/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a3/comment-killer/app/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.3.tgz (Root Library) - eslint-6.8.0.tgz - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-2.0.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-base/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.24.7.tgz (Root Library) - chokidar-1.7.0.tgz - :x: **glob-parent-2.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/secDevLabs/commit/c66f2ad957084d73653f76b825cbe31fa86061ba">c66f2ad957084d73653f76b825cbe31fa86061ba</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution (glob-parent): 6.0.1</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2021-35065 (High) detected in multiple libraries - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.2.tgz</b>, <b>glob-parent-3.1.0.tgz</b>, <b>glob-parent-5.1.1.tgz</b>, <b>glob-parent-2.0.0.tgz</b></p></summary> <p> <details><summary><b>glob-parent-3.1.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent directory path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p> <p> Dependency Hierarchy: - react-scripts-3.4.3.tgz (Root Library) - webpack-4.42.0.tgz - watchpack-1.7.4.tgz - watchpack-chokidar2-2.0.0.tgz - chokidar-2.1.8.tgz - :x: **glob-parent-3.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-5.1.2.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/User/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.27.7.tgz (Root Library) - chokidar-3.5.3.tgz - :x: **glob-parent-5.1.2.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-3.1.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent directory path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-3.1.0.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/webpack-dev-server/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/comment-killer/app/node_modules/fast-glob/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/webpack-dev-server/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.3.tgz (Root Library) - webpack-dev-server-3.11.0.tgz - chokidar-2.1.8.tgz - :x: **glob-parent-3.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-5.1.1.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.1.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a3/comment-killer/app/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a3/comment-killer/app/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a3/streaming/app/frontend/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - react-scripts-3.4.3.tgz (Root Library) - eslint-6.8.0.tgz - :x: **glob-parent-5.1.1.tgz** (Vulnerable Library) </details> <details><summary><b>glob-parent-2.0.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p> <p>Path to dependency file: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/package.json</p> <p>Path to vulnerable library: /owasp-top10-2021-apps/a8/amarelo-designs/app/static/node_modules/glob-parent/package.json,/owasp-top10-2021-apps/a5/stegonography/app/static/admin-page/node_modules/glob-base/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.24.7.tgz (Root Library) - chokidar-1.7.0.tgz - :x: **glob-parent-2.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/secDevLabs/commit/c66f2ad957084d73653f76b825cbe31fa86061ba">c66f2ad957084d73653f76b825cbe31fa86061ba</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution (glob-parent): 6.0.1</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.1</p><p>Fix Resolution (glob-parent): 6.0.1</p> <p>Direct dependency fix Resolution (react-scripts): 5.0.1</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_infrastructure
cve high detected in multiple libraries autoclosed cve high severity vulnerability vulnerable libraries glob parent tgz glob parent tgz glob parent tgz glob parent tgz glob parent tgz glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href dependency hierarchy react scripts tgz root library webpack tgz watchpack tgz watchpack tgz chokidar tgz x glob parent tgz vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file owasp apps amarelo designs app static user package json path to vulnerable library owasp apps amarelo designs app static user node modules glob parent package json dependency hierarchy browser sync tgz root library chokidar tgz x glob parent tgz vulnerable library glob parent tgz strips glob magic from a string to provide the parent directory path library home page a href path to dependency file owasp apps stegonography app static admin page package json path to vulnerable library owasp apps stegonography app static admin page node modules glob parent package json owasp apps comment killer app node modules webpack dev server node modules glob parent package json owasp apps comment killer app node modules fast glob node modules glob parent package json owasp apps streaming app frontend node modules webpack dev server node modules glob parent package json dependency hierarchy react scripts tgz root library webpack dev server tgz chokidar tgz x glob parent tgz vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file owasp apps comment killer app package json path to vulnerable library owasp apps comment killer app node modules glob parent package json owasp apps streaming app frontend node modules glob parent package json dependency hierarchy react scripts tgz root library eslint tgz x glob parent tgz vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file owasp apps amarelo designs app static package json path to vulnerable library owasp apps amarelo designs app static node modules glob parent package json owasp apps stegonography app static admin page node modules glob base node modules glob parent package json dependency hierarchy browser sync tgz root library chokidar tgz x glob parent tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package glob parent before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent direct dependency fix resolution react scripts fix resolution glob parent direct dependency fix resolution react scripts fix resolution glob parent direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue
0
123,521
26,268,074,173
IssuesEvent
2023-01-06 14:27:30
FerretDB/FerretDB
https://api.github.com/repos/FerretDB/FerretDB
closed
Validate the client's TLS certificate
code/feature
### What should be done? It should be possible to configure FerretDB to validate the client's certificates against the given CA certificate and reject connections without valid certificates. See https://www.mongodb.com/docs/manual/tutorial/configure-ssl/#set-up-mongod-and-mongos-with-client-certificate-validation. Standard go library should be enough for that. That should be a separate flag `--listen-tls-ca-file` that should be empty by default. It should be possible to use TLS as it is now without that flag. ### Where? https://github.com/FerretDB/FerretDB/blob/main/internal/clientconn/listener.go ### Definition of Done - unit, integration, or compatibility test should added (at least one of them should be possible to do); - spot refactorings done; - user documentation updated (see #1706).
1.0
Validate the client's TLS certificate - ### What should be done? It should be possible to configure FerretDB to validate the client's certificates against the given CA certificate and reject connections without valid certificates. See https://www.mongodb.com/docs/manual/tutorial/configure-ssl/#set-up-mongod-and-mongos-with-client-certificate-validation. Standard go library should be enough for that. That should be a separate flag `--listen-tls-ca-file` that should be empty by default. It should be possible to use TLS as it is now without that flag. ### Where? https://github.com/FerretDB/FerretDB/blob/main/internal/clientconn/listener.go ### Definition of Done - unit, integration, or compatibility test should added (at least one of them should be possible to do); - spot refactorings done; - user documentation updated (see #1706).
non_infrastructure
validate the client s tls certificate what should be done it should be possible to configure ferretdb to validate the client s certificates against the given ca certificate and reject connections without valid certificates see standard go library should be enough for that that should be a separate flag listen tls ca file that should be empty by default it should be possible to use tls as it is now without that flag where definition of done unit integration or compatibility test should added at least one of them should be possible to do spot refactorings done user documentation updated see
0
311,953
23,411,236,467
IssuesEvent
2022-08-12 17:46:44
FearlessSolutions/BCDA-DPC-B2D2-Veteran-Attribution-API
https://api.github.com/repos/FearlessSolutions/BCDA-DPC-B2D2-Veteran-Attribution-API
closed
Tracking multiple sources of truth
documentation
In most project kick-offs there can be multiple sources of truth that team members and stakeholders are each referring to as a guide. Sometimes its a SOO and a PWS, other times its OKRs or the RFQ. Sometimes it's a legacy backlog, other times its a greenfield project. This can cause swirl early as efforts are duplicated and at times conflicting. There are a few sources of truth circling around in our own project, and I want to ensure we are aligned. There is the original hackmd readme I made as a placeholder (actually for the purpose of tracking multiple sources of truth). That was uploaded to this repo, making the github readme the new source of truth for me. Once the tech challenge was issued, I added the official README to those notes and merged here in github. There is also a google drive, and a [one note](https://icfonline-my.sharepoint.com/personal/43256_icf_com/_layouts/15/Doc.aspx?sourcedoc={f52a3ebf-cc20-443b-a6dc-d756f416a704}&action=edit&wd=target%28Documentation.one%7Cf10b348a-7bf9-4396-97e2-cd30fc97d193%2FUseful%20Links%7Ce94b6c08-b1e4-465f-bd67-e5df7fcba050%2F%29&wdorigin=NavigationUrl). Due to the short turnaround, I am generally for using whatever tools work best for you, and want to raise awareness to be intentional about tracking github and ensuring anything we have in our own notes is accounted for here in github as well.
1.0
Tracking multiple sources of truth - In most project kick-offs there can be multiple sources of truth that team members and stakeholders are each referring to as a guide. Sometimes its a SOO and a PWS, other times its OKRs or the RFQ. Sometimes it's a legacy backlog, other times its a greenfield project. This can cause swirl early as efforts are duplicated and at times conflicting. There are a few sources of truth circling around in our own project, and I want to ensure we are aligned. There is the original hackmd readme I made as a placeholder (actually for the purpose of tracking multiple sources of truth). That was uploaded to this repo, making the github readme the new source of truth for me. Once the tech challenge was issued, I added the official README to those notes and merged here in github. There is also a google drive, and a [one note](https://icfonline-my.sharepoint.com/personal/43256_icf_com/_layouts/15/Doc.aspx?sourcedoc={f52a3ebf-cc20-443b-a6dc-d756f416a704}&action=edit&wd=target%28Documentation.one%7Cf10b348a-7bf9-4396-97e2-cd30fc97d193%2FUseful%20Links%7Ce94b6c08-b1e4-465f-bd67-e5df7fcba050%2F%29&wdorigin=NavigationUrl). Due to the short turnaround, I am generally for using whatever tools work best for you, and want to raise awareness to be intentional about tracking github and ensuring anything we have in our own notes is accounted for here in github as well.
non_infrastructure
tracking multiple sources of truth in most project kick offs there can be multiple sources of truth that team members and stakeholders are each referring to as a guide sometimes its a soo and a pws other times its okrs or the rfq sometimes it s a legacy backlog other times its a greenfield project this can cause swirl early as efforts are duplicated and at times conflicting there are a few sources of truth circling around in our own project and i want to ensure we are aligned there is the original hackmd readme i made as a placeholder actually for the purpose of tracking multiple sources of truth that was uploaded to this repo making the github readme the new source of truth for me once the tech challenge was issued i added the official readme to those notes and merged here in github there is also a google drive and a due to the short turnaround i am generally for using whatever tools work best for you and want to raise awareness to be intentional about tracking github and ensuring anything we have in our own notes is accounted for here in github as well
0
19,630
13,337,796,931
IssuesEvent
2020-08-28 09:52:08
OCR-D/core
https://api.github.com/repos/OCR-D/core
opened
More idiomatic and consistent unit tests
enhancement help wanted infrastructure
When we began OCR-D/core, we targeted only `unittests` with some enhancements over time in `tests/base.py`. To run the tests, we're actually using pytest, though. We currently have ~300 tests which use various mechanisms to achieve the same task, often with nested context Managers (`with TemporaryDirectory() as tempdir:` and such). @M3ssman suggested embracing pytest's features to a larger extent. For example, pytest offers decorators to achieve much of what we currently implement ourselves. For clarity and maintenance, we should refactor the test suite in a more consistent style. This is obviously nothing that directly improves the user experience, so it is a bit of an unthankful but important task. At least for new tests, we should strive to make use of pytest's features and update the existing tests by and by.
1.0
More idiomatic and consistent unit tests - When we began OCR-D/core, we targeted only `unittests` with some enhancements over time in `tests/base.py`. To run the tests, we're actually using pytest, though. We currently have ~300 tests which use various mechanisms to achieve the same task, often with nested context Managers (`with TemporaryDirectory() as tempdir:` and such). @M3ssman suggested embracing pytest's features to a larger extent. For example, pytest offers decorators to achieve much of what we currently implement ourselves. For clarity and maintenance, we should refactor the test suite in a more consistent style. This is obviously nothing that directly improves the user experience, so it is a bit of an unthankful but important task. At least for new tests, we should strive to make use of pytest's features and update the existing tests by and by.
infrastructure
more idiomatic and consistent unit tests when we began ocr d core we targeted only unittests with some enhancements over time in tests base py to run the tests we re actually using pytest though we currently have tests which use various mechanisms to achieve the same task often with nested context managers with temporarydirectory as tempdir and such suggested embracing pytest s features to a larger extent for example pytest offers decorators to achieve much of what we currently implement ourselves for clarity and maintenance we should refactor the test suite in a more consistent style this is obviously nothing that directly improves the user experience so it is a bit of an unthankful but important task at least for new tests we should strive to make use of pytest s features and update the existing tests by and by
1
32,317
26,621,786,699
IssuesEvent
2023-01-24 11:48:28
Ladder99/ladder99
https://api.github.com/repos/Ladder99/ladder99
closed
Filter to just `l99` services in `l99 status` command
enhancement infrastructure
ob: l99 status is docker ps and it lists all running containers, even those that are not related to l99. IMO it should output only the state of l99 containers. I should output at least the l99 is running or not, ideally each container if it is running or not (even if it is not running), optionally in JSON format. bb: Yeah, I had included non-ladder99 containers as there were things like fanuc_driver which are currently run separately, but yeah maybe better to filter those out. Adding to list. cm: L99 status should be limited to containers in the l99 stack. I would not worry about fanuc-driver. We can make it part of the stack eventually. ob: Why fanuc-driver would be a non-l99 container? IMO if fanuc-driver is used by l99, why wouldn’t it start with --project ladder99 in docker-compose up? Shouldn’t it be set up during l99 setup (l99 start)?
1.0
Filter to just `l99` services in `l99 status` command - ob: l99 status is docker ps and it lists all running containers, even those that are not related to l99. IMO it should output only the state of l99 containers. I should output at least the l99 is running or not, ideally each container if it is running or not (even if it is not running), optionally in JSON format. bb: Yeah, I had included non-ladder99 containers as there were things like fanuc_driver which are currently run separately, but yeah maybe better to filter those out. Adding to list. cm: L99 status should be limited to containers in the l99 stack. I would not worry about fanuc-driver. We can make it part of the stack eventually. ob: Why fanuc-driver would be a non-l99 container? IMO if fanuc-driver is used by l99, why wouldn’t it start with --project ladder99 in docker-compose up? Shouldn’t it be set up during l99 setup (l99 start)?
infrastructure
filter to just services in status command ob status is docker ps and it lists all running containers even those that are not related to imo it should output only the state of containers i should output at least the is running or not ideally each container if it is running or not even if it is not running optionally in json format bb yeah i had included non containers as there were things like fanuc driver which are currently run separately but yeah maybe better to filter those out adding to list cm status should be limited to containers in the stack i would not worry about fanuc driver we can make it part of the stack eventually ob why fanuc driver would be a non container imo if fanuc driver is used by why wouldn’t it start with project in docker compose up shouldn’t it be set up during setup start
1
293,948
25,335,618,100
IssuesEvent
2022-11-18 16:34:52
anoma/namada-trusted-setup
https://api.github.com/repos/anoma/namada-trusted-setup
closed
Repo misc fixes
refactor testing
# Fix - [x] `cargo check` warnings - [x] review `masp` import - [x] all `FIXME`s - [x] Add a phantom cohort with no tokens at the end of the ceremony to leave enough time to the participants of the last cohort for their contributions - [x] How to provide the `contributors.json` file to the frontend, two ways: - Keep the coordinator running forever while shutting down the `join_queue` endpoint (this would not require a phantom cohort) - Shut the coordinator down and publish the `contributors.json` file to S3 - [x] Remove unused dependencies - [x] Produce final release tag - [x] Upload last contribution verified to S3? - [x] Fix messages and docs - [x] Rename `phase1` to `phase2` # Test - [x] contribution on `AnotherMachine` - [x] token blacklisting - [x] Namada keypair generation - [ ] restart of coordinator? # Review - [x] values for the env variables to set - [x] state of the Coordinator server - [x] state of the S3 server - [x] state of the Amazon Parameter Store
1.0
Repo misc fixes - # Fix - [x] `cargo check` warnings - [x] review `masp` import - [x] all `FIXME`s - [x] Add a phantom cohort with no tokens at the end of the ceremony to leave enough time to the participants of the last cohort for their contributions - [x] How to provide the `contributors.json` file to the frontend, two ways: - Keep the coordinator running forever while shutting down the `join_queue` endpoint (this would not require a phantom cohort) - Shut the coordinator down and publish the `contributors.json` file to S3 - [x] Remove unused dependencies - [x] Produce final release tag - [x] Upload last contribution verified to S3? - [x] Fix messages and docs - [x] Rename `phase1` to `phase2` # Test - [x] contribution on `AnotherMachine` - [x] token blacklisting - [x] Namada keypair generation - [ ] restart of coordinator? # Review - [x] values for the env variables to set - [x] state of the Coordinator server - [x] state of the S3 server - [x] state of the Amazon Parameter Store
non_infrastructure
repo misc fixes fix cargo check warnings review masp import all fixme s add a phantom cohort with no tokens at the end of the ceremony to leave enough time to the participants of the last cohort for their contributions how to provide the contributors json file to the frontend two ways keep the coordinator running forever while shutting down the join queue endpoint this would not require a phantom cohort shut the coordinator down and publish the contributors json file to remove unused dependencies produce final release tag upload last contribution verified to fix messages and docs rename to test contribution on anothermachine token blacklisting namada keypair generation restart of coordinator review values for the env variables to set state of the coordinator server state of the server state of the amazon parameter store
0
9,185
7,857,726,704
IssuesEvent
2018-06-21 11:48:28
hzi-braunschweig/SORMAS-Project
https://api.github.com/repos/hzi-braunschweig/SORMAS-Project
opened
Property fields can depend on other fields
102XX Infrastructure sormas-app
- [ ] add "app:dependencyParentId" and "app:dependyParentValue" attributes - [ ] field registers value listener at parent - [ ] hide field when parent does not have the appropriate value
1.0
Property fields can depend on other fields - - [ ] add "app:dependencyParentId" and "app:dependyParentValue" attributes - [ ] field registers value listener at parent - [ ] hide field when parent does not have the appropriate value
infrastructure
property fields can depend on other fields add app dependencyparentid and app dependyparentvalue attributes field registers value listener at parent hide field when parent does not have the appropriate value
1
23,473
16,239,146,288
IssuesEvent
2021-05-07 07:11:10
microsoft/WindowsTemplateStudio
https://api.github.com/repos/microsoft/WindowsTemplateStudio
closed
Reorganize internal projecttemplate folder structure
Can Close Out Soon Infrastructure
With the addition of ClassLibraries and RuntimeComponents project templates the number of project templates in the [ProjectTemplates](https://github.com/microsoft/WindowsTemplateStudio/tree/dev/code/src/ProjectTemplates) folder is getting to a point where it's difficult to distinguish the different entry points. We propose to add subfolders for the different platforms and languages to make it easier to find the correct vstemplate. This is only an internal folder structure change and won't have any impact on the generated code or how templates show in the New Project Dialog. Proposed folder structure is: - ProjectTemplates - WinUI - CS - Cpp - WPF - CS - Uwp - CS - VB
1.0
Reorganize internal projecttemplate folder structure - With the addition of ClassLibraries and RuntimeComponents project templates the number of project templates in the [ProjectTemplates](https://github.com/microsoft/WindowsTemplateStudio/tree/dev/code/src/ProjectTemplates) folder is getting to a point where it's difficult to distinguish the different entry points. We propose to add subfolders for the different platforms and languages to make it easier to find the correct vstemplate. This is only an internal folder structure change and won't have any impact on the generated code or how templates show in the New Project Dialog. Proposed folder structure is: - ProjectTemplates - WinUI - CS - Cpp - WPF - CS - Uwp - CS - VB
infrastructure
reorganize internal projecttemplate folder structure with the addition of classlibraries and runtimecomponents project templates the number of project templates in the folder is getting to a point where it s difficult to distinguish the different entry points we propose to add subfolders for the different platforms and languages to make it easier to find the correct vstemplate this is only an internal folder structure change and won t have any impact on the generated code or how templates show in the new project dialog proposed folder structure is projecttemplates winui cs cpp wpf cs uwp cs vb
1
25,477
4,158,871,172
IssuesEvent
2016-06-17 06:11:51
NagiosEnterprises/ncpa
https://api.github.com/repos/NagiosEnterprises/ncpa
closed
Add upgrade option to installer
Feature Request Needs Testing
We should add an upgrade option when someone uses a newer version installer... this doesn't necessarily have to be shown to the user but we should run something to stop the active/passive listeners so that the install can overwrite their current version and then start the new version for them... would be nice if it could somehow know though so you wouldn't have to re-enter the config variables.
1.0
Add upgrade option to installer - We should add an upgrade option when someone uses a newer version installer... this doesn't necessarily have to be shown to the user but we should run something to stop the active/passive listeners so that the install can overwrite their current version and then start the new version for them... would be nice if it could somehow know though so you wouldn't have to re-enter the config variables.
non_infrastructure
add upgrade option to installer we should add an upgrade option when someone uses a newer version installer this doesn t necessarily have to be shown to the user but we should run something to stop the active passive listeners so that the install can overwrite their current version and then start the new version for them would be nice if it could somehow know though so you wouldn t have to re enter the config variables
0
356,812
25,176,265,762
IssuesEvent
2022-11-11 09:31:57
Tex-Tang/pe
https://api.github.com/repos/Tex-Tang/pe
opened
Out-of-date screenshot in UG
type.DocumentationBug severity.Low
`2 Class-ify's Application Window` section in UG ![image.png](https://raw.githubusercontent.com/Tex-Tang/pe/main/files/59f4bb35-24ac-4cef-8fe9-12ad4f9fff1f.png) `4.3.2` section in UG ![image.png](https://raw.githubusercontent.com/Tex-Tang/pe/main/files/02820b8d-eaeb-4337-bb86-4e982b0490a6.png) <!--session: 1668153999061-e63af65f-10dc-4b2f-bfdc-f2a47bc4eed4--> <!--Version: Web v3.4.4-->
1.0
Out-of-date screenshot in UG - `2 Class-ify's Application Window` section in UG ![image.png](https://raw.githubusercontent.com/Tex-Tang/pe/main/files/59f4bb35-24ac-4cef-8fe9-12ad4f9fff1f.png) `4.3.2` section in UG ![image.png](https://raw.githubusercontent.com/Tex-Tang/pe/main/files/02820b8d-eaeb-4337-bb86-4e982b0490a6.png) <!--session: 1668153999061-e63af65f-10dc-4b2f-bfdc-f2a47bc4eed4--> <!--Version: Web v3.4.4-->
non_infrastructure
out of date screenshot in ug class ify s application window section in ug section in ug
0
13,841
10,481,995,999
IssuesEvent
2019-09-24 10:55:22
shopsys/shopsys
https://api.github.com/repos/shopsys/shopsys
closed
Build application in kubernetes should be simplified
Infrastructure
<!--- Title should contain short general summary what is the issue about --> ### What is happening In deploy/build scripts is often used yq to replace placeholders or some values with the proper one right for the environment. This replacement often looks like ```sh # Set environment variables to container and initContainer for Google Cloud Storage connection - set it to null yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[0].value 'bucket-name' yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[1].value 'project-id' yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[0].value 'bucket-name' yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[1].value 'project-id' ``` Which transforms (shortened example) ```yaml spec: template: initContainers: - name: copy-source-codes-to-volume - name: initialize-database env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: ~ - name: GOOGLE_CLOUD_PROJECT_ID value: ~ containers: - name: php-fpm env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: ~ - name: GOOGLE_CLOUD_PROJECT_ID value: ~ ``` into ```yaml spec: template: initContainers: - name: copy-source-codes-to-volume - name: initialize-database env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: bucket-name - name: GOOGLE_CLOUD_PROJECT_ID value: project-id containers: - name: php-fpm env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: bucket-name - name: GOOGLE_CLOUD_PROJECT_ID value: project-id ``` This is violation of DRY and most of all very fragile solution. It's really easy to prepend new env variable as the @PetrHeinz correctly pointed out. Maybe we could use Kustomize variables (see [vars documentation](https://github.com/kubernetes-sigs/kustomize/blob/master/docs/fields.md#vars) or [an example using vars](https://github.com/kubernetes-sigs/kustomize/blob/master/examples/wordpress/README.md)) to not repeat ourselves like this. <!--- What are preconditions and your setting e.g. Shopsys version or in case of Docker issues your operating system --> <!--- Best is to include steps to reproduce this issue if relevant--> <!--- Tell us what happens --> ### Expected result Scripts should be less error-prone and simplified.
1.0
Build application in kubernetes should be simplified - <!--- Title should contain short general summary what is the issue about --> ### What is happening In deploy/build scripts is often used yq to replace placeholders or some values with the proper one right for the environment. This replacement often looks like ```sh # Set environment variables to container and initContainer for Google Cloud Storage connection - set it to null yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[0].value 'bucket-name' yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.containers[0].env[1].value 'project-id' yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[0].value 'bucket-name' yq write --inplace project-base/kubernetes/deployments/webserver-php-fpm.yml spec.template.spec.initContainers[1].env[1].value 'project-id' ``` Which transforms (shortened example) ```yaml spec: template: initContainers: - name: copy-source-codes-to-volume - name: initialize-database env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: ~ - name: GOOGLE_CLOUD_PROJECT_ID value: ~ containers: - name: php-fpm env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: ~ - name: GOOGLE_CLOUD_PROJECT_ID value: ~ ``` into ```yaml spec: template: initContainers: - name: copy-source-codes-to-volume - name: initialize-database env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: bucket-name - name: GOOGLE_CLOUD_PROJECT_ID value: project-id containers: - name: php-fpm env: - name: GOOGLE_CLOUD_STORAGE_BUCKET_NAME value: bucket-name - name: GOOGLE_CLOUD_PROJECT_ID value: project-id ``` This is violation of DRY and most of all very fragile solution. It's really easy to prepend new env variable as the @PetrHeinz correctly pointed out. Maybe we could use Kustomize variables (see [vars documentation](https://github.com/kubernetes-sigs/kustomize/blob/master/docs/fields.md#vars) or [an example using vars](https://github.com/kubernetes-sigs/kustomize/blob/master/examples/wordpress/README.md)) to not repeat ourselves like this. <!--- What are preconditions and your setting e.g. Shopsys version or in case of Docker issues your operating system --> <!--- Best is to include steps to reproduce this issue if relevant--> <!--- Tell us what happens --> ### Expected result Scripts should be less error-prone and simplified.
infrastructure
build application in kubernetes should be simplified what is happening in deploy build scripts is often used yq to replace placeholders or some values with the proper one right for the environment this replacement often looks like sh set environment variables to container and initcontainer for google cloud storage connection set it to null yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec containers env value bucket name yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec containers env value project id yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec initcontainers env value bucket name yq write inplace project base kubernetes deployments webserver php fpm yml spec template spec initcontainers env value project id which transforms shortened example yaml spec template initcontainers name copy source codes to volume name initialize database env name google cloud storage bucket name value name google cloud project id value containers name php fpm env name google cloud storage bucket name value name google cloud project id value into yaml spec template initcontainers name copy source codes to volume name initialize database env name google cloud storage bucket name value bucket name name google cloud project id value project id containers name php fpm env name google cloud storage bucket name value bucket name name google cloud project id value project id this is violation of dry and most of all very fragile solution it s really easy to prepend new env variable as the petrheinz correctly pointed out maybe we could use kustomize variables see or to not repeat ourselves like this expected result scripts should be less error prone and simplified
1
24,920
17,909,872,241
IssuesEvent
2021-09-09 02:43:51
FullScreenShenanigans/EightBittr
https://api.github.com/repos/FullScreenShenanigans/EightBittr
closed
Switch to fixed package versions in Lerna (instead of independent)
status: accepting prs type: infrastructure
https://github.com/lerna/lerna#how-it-works It'll be easier to just let Lerna do its magic than manage it manually.
1.0
Switch to fixed package versions in Lerna (instead of independent) - https://github.com/lerna/lerna#how-it-works It'll be easier to just let Lerna do its magic than manage it manually.
infrastructure
switch to fixed package versions in lerna instead of independent it ll be easier to just let lerna do its magic than manage it manually
1
113,497
24,429,426,940
IssuesEvent
2022-10-06 06:57:01
WordPress/openverse-frontend
https://api.github.com/repos/WordPress/openverse-frontend
closed
Recent Searches feature (Combobox, localstorage, and more)
🟨 priority: medium 🏁 status: ready for work ✨ goal: improvement 💻 aspect: code
## Description <!-- Describe a problem solved by this feature; or delete the section entirely. --> We plan to add a new feature to show the user their 'recent searches', in a box below the searchbar. I originally thought this would be a combobox feature, but we shouldn't actually need to implement that until we're doing search _suggestions_, where the user input determines the contents of the box. In this use case we're just showing the nth-most recent searches. > A preview of the recent searches ![image](https://user-images.githubusercontent.com/6351754/144633420-a3aeadac-187a-455d-9b97-1358d22b7046.png) Figma Link: https://www.figma.com/file/w60dl1XPUvSaRncv1Utmnb/Openverse-Releases?node-id=133%3A29056 ## Dev Tasks - [ ] Implement the suggestions box - [ ] Popover for the desktop (above `lg`) - [ ] Full page modal with a search bar for mobile screens - [x] Create a composable for storing recent searches - [x] Store user recent searches in indexedDb (see https://github.com/WordPress/openverse-frontend/issues/479#issuecomment-1252941676 for justification) - [x] Enable clearing of the recent searches in indexedDb ## Implementation questions - Do we need a way to let users opt-out of this? (maybe this can wait) - How many items do we want to show? (4 in the mockup currently) I imagine the search bar will populate and the search will be instantly submitted. - What will happen when the user clicks on one of the recent searches? ## Design decisions (cc @panchovm) - [x] Add this to the Figma component library (it's currently only in the page mockups) - [x] Implement a 'clear' button (will update the mockups for this) - [x] What will the hover/focus styles look like for the choices?
1.0
Recent Searches feature (Combobox, localstorage, and more) - ## Description <!-- Describe a problem solved by this feature; or delete the section entirely. --> We plan to add a new feature to show the user their 'recent searches', in a box below the searchbar. I originally thought this would be a combobox feature, but we shouldn't actually need to implement that until we're doing search _suggestions_, where the user input determines the contents of the box. In this use case we're just showing the nth-most recent searches. > A preview of the recent searches ![image](https://user-images.githubusercontent.com/6351754/144633420-a3aeadac-187a-455d-9b97-1358d22b7046.png) Figma Link: https://www.figma.com/file/w60dl1XPUvSaRncv1Utmnb/Openverse-Releases?node-id=133%3A29056 ## Dev Tasks - [ ] Implement the suggestions box - [ ] Popover for the desktop (above `lg`) - [ ] Full page modal with a search bar for mobile screens - [x] Create a composable for storing recent searches - [x] Store user recent searches in indexedDb (see https://github.com/WordPress/openverse-frontend/issues/479#issuecomment-1252941676 for justification) - [x] Enable clearing of the recent searches in indexedDb ## Implementation questions - Do we need a way to let users opt-out of this? (maybe this can wait) - How many items do we want to show? (4 in the mockup currently) I imagine the search bar will populate and the search will be instantly submitted. - What will happen when the user clicks on one of the recent searches? ## Design decisions (cc @panchovm) - [x] Add this to the Figma component library (it's currently only in the page mockups) - [x] Implement a 'clear' button (will update the mockups for this) - [x] What will the hover/focus styles look like for the choices?
non_infrastructure
recent searches feature combobox localstorage and more description we plan to add a new feature to show the user their recent searches in a box below the searchbar i originally thought this would be a combobox feature but we shouldn t actually need to implement that until we re doing search suggestions where the user input determines the contents of the box in this use case we re just showing the nth most recent searches a preview of the recent searches figma link dev tasks implement the suggestions box popover for the desktop above lg full page modal with a search bar for mobile screens create a composable for storing recent searches store user recent searches in indexeddb see for justification enable clearing of the recent searches in indexeddb implementation questions do we need a way to let users opt out of this maybe this can wait how many items do we want to show in the mockup currently i imagine the search bar will populate and the search will be instantly submitted what will happen when the user clicks on one of the recent searches design decisions cc panchovm add this to the figma component library it s currently only in the page mockups implement a clear button will update the mockups for this what will the hover focus styles look like for the choices
0
6,344
6,342,442,438
IssuesEvent
2017-07-27 15:29:54
openshiftio/appdev-documentation
https://api.github.com/repos/openshiftio/appdev-documentation
closed
Release tagging does not work on UNIX machines
Bug Follow-up Has PR Infrastructure Peer Review SME Review
Following up on #391: As reported by @rhoads-zach in the documentation Slack channel, the `$REPO_HOME/scripts/tagRelease.sh` script does not work on machines with OpenBSD or MacOS because the `date` binary there does not have the `--rfc-3339` flag. I assume this problem also affects the `ci/Jenkinsfile` script as I have used the same invocation there. I believe the solution is to use `date "+%Y-%m-%d"` instead. I have verified it works on OpenBSD and Fedora. @rhoads-zach or someone with an Apple PC, can you please verify it works on MacOS as well?
1.0
Release tagging does not work on UNIX machines - Following up on #391: As reported by @rhoads-zach in the documentation Slack channel, the `$REPO_HOME/scripts/tagRelease.sh` script does not work on machines with OpenBSD or MacOS because the `date` binary there does not have the `--rfc-3339` flag. I assume this problem also affects the `ci/Jenkinsfile` script as I have used the same invocation there. I believe the solution is to use `date "+%Y-%m-%d"` instead. I have verified it works on OpenBSD and Fedora. @rhoads-zach or someone with an Apple PC, can you please verify it works on MacOS as well?
infrastructure
release tagging does not work on unix machines following up on as reported by rhoads zach in the documentation slack channel the repo home scripts tagrelease sh script does not work on machines with openbsd or macos because the date binary there does not have the rfc flag i assume this problem also affects the ci jenkinsfile script as i have used the same invocation there i believe the solution is to use date y m d instead i have verified it works on openbsd and fedora rhoads zach or someone with an apple pc can you please verify it works on macos as well
1
21,442
14,569,585,692
IssuesEvent
2020-12-17 13:17:54
odrotbohm/moduliths
https://api.github.com/repos/odrotbohm/moduliths
closed
Upgrade to ArchUnit 0.15.0
in: infrastructure type: task
It now ships with generics analysis, i.e. it detects dependencies stemming from those and generally detects more dependencies previously undiscovered. https://github.com/TNG/ArchUnit/releases/tag/v0.15.0
1.0
Upgrade to ArchUnit 0.15.0 - It now ships with generics analysis, i.e. it detects dependencies stemming from those and generally detects more dependencies previously undiscovered. https://github.com/TNG/ArchUnit/releases/tag/v0.15.0
infrastructure
upgrade to archunit it now ships with generics analysis i e it detects dependencies stemming from those and generally detects more dependencies previously undiscovered
1
2,884
3,951,349,600
IssuesEvent
2016-04-29 00:31:25
AtlasOfLivingAustralia/data-management
https://api.github.com/repos/AtlasOfLivingAustralia/data-management
closed
Improved download interface
Downloads Infrastructure
interface to allow field selection from full set of fields - index, QA and additional fields saving of default set of fields allow addition of un-sensitised values where appropriate permissions exist placing export file in an accessible location functionality to facilitate automation
1.0
Improved download interface - interface to allow field selection from full set of fields - index, QA and additional fields saving of default set of fields allow addition of un-sensitised values where appropriate permissions exist placing export file in an accessible location functionality to facilitate automation
infrastructure
improved download interface interface to allow field selection from full set of fields index qa and additional fields saving of default set of fields allow addition of un sensitised values where appropriate permissions exist placing export file in an accessible location functionality to facilitate automation
1
25,426
18,732,923,550
IssuesEvent
2021-11-04 01:12:18
t3kt/raytk
https://api.github.com/repos/t3kt/raytk
closed
Improve the shared shader library system
cleanup infrastructure
Current all the shared libraries need to be available within the shaderLibraries tox. It should be possible to allow them from anywhere, and have some form of deduping. That way an OP could host its own copy of the shared library.
1.0
Improve the shared shader library system - Current all the shared libraries need to be available within the shaderLibraries tox. It should be possible to allow them from anywhere, and have some form of deduping. That way an OP could host its own copy of the shared library.
infrastructure
improve the shared shader library system current all the shared libraries need to be available within the shaderlibraries tox it should be possible to allow them from anywhere and have some form of deduping that way an op could host its own copy of the shared library
1
796,519
28,116,531,402
IssuesEvent
2023-03-31 11:11:25
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
opened
[Priest] Dispersion cannot be used when affected by Maloriak's Flash freeze.
Class: Priest Status: Duplicate Priority: High Raid: Blackwing Descent
Confirmed on PTR and Prod. Priests specced into shadow cannot use Dispersion when under the effect of Flash Freeze Flash Freeze is considered a stun and should mean Shadow Priests can use Dispersion. ![image](https://user-images.githubusercontent.com/121137397/229104916-d06a7952-1425-49e4-93ac-b41a4598fd22.png) ![image](https://user-images.githubusercontent.com/121137397/229104997-97aa27f0-53ab-44e7-8a03-f3113917af06.png)
1.0
[Priest] Dispersion cannot be used when affected by Maloriak's Flash freeze. - Confirmed on PTR and Prod. Priests specced into shadow cannot use Dispersion when under the effect of Flash Freeze Flash Freeze is considered a stun and should mean Shadow Priests can use Dispersion. ![image](https://user-images.githubusercontent.com/121137397/229104916-d06a7952-1425-49e4-93ac-b41a4598fd22.png) ![image](https://user-images.githubusercontent.com/121137397/229104997-97aa27f0-53ab-44e7-8a03-f3113917af06.png)
non_infrastructure
dispersion cannot be used when affected by maloriak s flash freeze confirmed on ptr and prod priests specced into shadow cannot use dispersion when under the effect of flash freeze flash freeze is considered a stun and should mean shadow priests can use dispersion
0
24,608
17,467,186,111
IssuesEvent
2021-08-06 18:41:30
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Adhere to the package support policy by building and packaging supported frameworks only
area-Infrastructure-libraries breaking-change needs-breaking-change-doc-created
Continuation of https://github.com/dotnet/runtime/issues/47530 The runtime repository currently produces 102 packages from libraries of which 25 of those contain assets which will be unsupported when we release .NET 6 in November 2021. The list of unsupported assets includes netstandard1.x assets as those apply to unsupported application frameworks only as in any case where a netstandard1.x asset is present, a netstandard2.0 (and in most cases a net461 asset) exists as well which allows to run on .NET Framework 4.6.1 or .NET Core 3.1 and above. The reasons for trimming out unsupported assets are: - Simplified build graph and reduced build times - Less restore size as the NS1.x dependency graph doesn't need to be restored - Package validation costs - Fat packages - Maintenance of code bases which are not tested anymore. Based on these disadvantages we will remove the remaining unsupported assets from packages in dotnet/runtime. ## List of packages grouped by TargetFrameworkIdentifier ### NETStandard,Version=1.x -> .NETStandard,Version=2.0 - Microsoft.Extensions.HostFactoryResolver.Sources - System.Collections.Immutable - System.Diagnostics.DiagnosticSource - System.IO.Packaging - System.Numerics.Tensors - System.Reflection.Context - System.Reflection.Metadata - System.Runtime.CompilerServices.Unsafe - System.Threading.Channels - System.Threading.Tasks.Dataflow PR: https://github.com/dotnet/runtime/pull/53283 ### NETCoreApp,Version=2.x|3.0 -> .NETCoreApp,Version=3.1 - Microsoft.Win32.SystemEvents - System.Data.Odbc - System.Diagnostics.EventLog - System.ComponentModel.Composition - System.Diagnostics.PerformanceCounter - System.DirectoryServices - System.DirectoryServices.AccountManagement - System.DirectoryServices.Protocols - System.Management - System.Runtime.CompilerServices.Unsafe - System.Security.Principal.Windows - System.Speech - System.Text.Encoding.CodePages - System.Threading.Channels PR: https://github.com/dotnet/runtime/pull/53358 ### NETFramework,Version=4.5|4.5.1|4.6 -> .NETFramework.Version=4.6.1 - System.Diagnostics.DiagnosticSource - System.Runtime.CompilerServices.Unsafe - System.IO.Packaging - System.ComponentModel.Composition.Registration - System.ComponentModel.Composition - System.DirectoryServices - System.DirectoryServices.AccountManagement - System.DirectoryServices.Protocols - System.IO.Packaging - System.Management - System.Reflection.Context - System.Runtime.Caching - System.Speech PR: https://github.com/dotnet/runtime/pull/53319 cc @danmoseley @ericstj @terrajobst @Anipik
1.0
Adhere to the package support policy by building and packaging supported frameworks only - Continuation of https://github.com/dotnet/runtime/issues/47530 The runtime repository currently produces 102 packages from libraries of which 25 of those contain assets which will be unsupported when we release .NET 6 in November 2021. The list of unsupported assets includes netstandard1.x assets as those apply to unsupported application frameworks only as in any case where a netstandard1.x asset is present, a netstandard2.0 (and in most cases a net461 asset) exists as well which allows to run on .NET Framework 4.6.1 or .NET Core 3.1 and above. The reasons for trimming out unsupported assets are: - Simplified build graph and reduced build times - Less restore size as the NS1.x dependency graph doesn't need to be restored - Package validation costs - Fat packages - Maintenance of code bases which are not tested anymore. Based on these disadvantages we will remove the remaining unsupported assets from packages in dotnet/runtime. ## List of packages grouped by TargetFrameworkIdentifier ### NETStandard,Version=1.x -> .NETStandard,Version=2.0 - Microsoft.Extensions.HostFactoryResolver.Sources - System.Collections.Immutable - System.Diagnostics.DiagnosticSource - System.IO.Packaging - System.Numerics.Tensors - System.Reflection.Context - System.Reflection.Metadata - System.Runtime.CompilerServices.Unsafe - System.Threading.Channels - System.Threading.Tasks.Dataflow PR: https://github.com/dotnet/runtime/pull/53283 ### NETCoreApp,Version=2.x|3.0 -> .NETCoreApp,Version=3.1 - Microsoft.Win32.SystemEvents - System.Data.Odbc - System.Diagnostics.EventLog - System.ComponentModel.Composition - System.Diagnostics.PerformanceCounter - System.DirectoryServices - System.DirectoryServices.AccountManagement - System.DirectoryServices.Protocols - System.Management - System.Runtime.CompilerServices.Unsafe - System.Security.Principal.Windows - System.Speech - System.Text.Encoding.CodePages - System.Threading.Channels PR: https://github.com/dotnet/runtime/pull/53358 ### NETFramework,Version=4.5|4.5.1|4.6 -> .NETFramework.Version=4.6.1 - System.Diagnostics.DiagnosticSource - System.Runtime.CompilerServices.Unsafe - System.IO.Packaging - System.ComponentModel.Composition.Registration - System.ComponentModel.Composition - System.DirectoryServices - System.DirectoryServices.AccountManagement - System.DirectoryServices.Protocols - System.IO.Packaging - System.Management - System.Reflection.Context - System.Runtime.Caching - System.Speech PR: https://github.com/dotnet/runtime/pull/53319 cc @danmoseley @ericstj @terrajobst @Anipik
infrastructure
adhere to the package support policy by building and packaging supported frameworks only continuation of the runtime repository currently produces packages from libraries of which of those contain assets which will be unsupported when we release net in november the list of unsupported assets includes x assets as those apply to unsupported application frameworks only as in any case where a x asset is present a and in most cases a asset exists as well which allows to run on net framework or net core and above the reasons for trimming out unsupported assets are simplified build graph and reduced build times less restore size as the x dependency graph doesn t need to be restored package validation costs fat packages maintenance of code bases which are not tested anymore based on these disadvantages we will remove the remaining unsupported assets from packages in dotnet runtime list of packages grouped by targetframeworkidentifier netstandard version x netstandard version microsoft extensions hostfactoryresolver sources system collections immutable system diagnostics diagnosticsource system io packaging system numerics tensors system reflection context system reflection metadata system runtime compilerservices unsafe system threading channels system threading tasks dataflow pr netcoreapp version x netcoreapp version microsoft systemevents system data odbc system diagnostics eventlog system componentmodel composition system diagnostics performancecounter system directoryservices system directoryservices accountmanagement system directoryservices protocols system management system runtime compilerservices unsafe system security principal windows system speech system text encoding codepages system threading channels pr netframework version netframework version system diagnostics diagnosticsource system runtime compilerservices unsafe system io packaging system componentmodel composition registration system componentmodel composition system directoryservices system directoryservices accountmanagement system directoryservices protocols system io packaging system management system reflection context system runtime caching system speech pr cc danmoseley ericstj terrajobst anipik
1
375
2,667,007,321
IssuesEvent
2015-03-22 04:39:33
deis/deis
https://api.github.com/repos/deis/deis
opened
bug(test): during the test if the user app is stopped it continues running
requires-manual-testing test-infrastructure tests
instead of fail. Steps to reproduce: - during the test enter in any node. - stop the app. - the suite will continue to run instead of throwing an error Why is relevant? In case of an error in the deployed app it will restarted until the next output: ``` Mar 22 04:31:15 deis-01 systemd[1]: configsample_v9.cmd.1.service holdoff time over, scheduling restart. ``` then the service will stop and the suite will continue running
1.0
bug(test): during the test if the user app is stopped it continues running - instead of fail. Steps to reproduce: - during the test enter in any node. - stop the app. - the suite will continue to run instead of throwing an error Why is relevant? In case of an error in the deployed app it will restarted until the next output: ``` Mar 22 04:31:15 deis-01 systemd[1]: configsample_v9.cmd.1.service holdoff time over, scheduling restart. ``` then the service will stop and the suite will continue running
infrastructure
bug test during the test if the user app is stopped it continues running instead of fail steps to reproduce during the test enter in any node stop the app the suite will continue to run instead of throwing an error why is relevant in case of an error in the deployed app it will restarted until the next output mar deis systemd configsample cmd service holdoff time over scheduling restart then the service will stop and the suite will continue running
1
26,274
19,898,992,258
IssuesEvent
2022-01-25 04:41:54
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
closed
Merge OppiaTestRule and AccessibilityTestRule
Type: Improvement Where: Infrastructure Priority: Important Status: Not started Project stage: not started
Merge `OppiaTestRule` and `AccessibilityTestRule` in `OppiaTestRule` Once done, run all test cases on Espresso and post screenshot displaying all tests are passing on Pixel 3 XL.
1.0
Merge OppiaTestRule and AccessibilityTestRule - Merge `OppiaTestRule` and `AccessibilityTestRule` in `OppiaTestRule` Once done, run all test cases on Espresso and post screenshot displaying all tests are passing on Pixel 3 XL.
infrastructure
merge oppiatestrule and accessibilitytestrule merge oppiatestrule and accessibilitytestrule in oppiatestrule once done run all test cases on espresso and post screenshot displaying all tests are passing on pixel xl
1
21,508
14,599,829,618
IssuesEvent
2020-12-21 05:23:32
php-coder/mystamps
https://api.github.com/repos/php-coder/mystamps
opened
Install Prometheus and Grafana for monitoring
area/infrastructure
- [ ] install Prometheus - [ ] install Grafana - [ ] install node-exporter - [ ] add ansible roles
1.0
Install Prometheus and Grafana for monitoring - - [ ] install Prometheus - [ ] install Grafana - [ ] install node-exporter - [ ] add ansible roles
infrastructure
install prometheus and grafana for monitoring install prometheus install grafana install node exporter add ansible roles
1
183,228
6,678,622,058
IssuesEvent
2017-10-05 14:47:36
AdguardTeam/AdguardFilters
https://api.github.com/repos/AdguardTeam/AdguardFilters
closed
http://m.azonline.de/ - sm widget, banners and ad leftovers
Annoyance Incorrect Blocking iOS Browser Priority: Low Resolved Social Widget
[//]: # (***You can delete or ignore strings starting with "[//]:" They will not be visible either way.) ***Description***: * **Current behaviour**: http://m.azonline.de/NRW/2980745-Prozesse-Steuererklaerung-verspaetet-abgegeben-Berater-verurteilt [//]: # (Substitute this line with a description of the problem) [//]: # (Replace %screenshot_url% below with a link to the screenshot of the problem. Also, you can paste image from clipboard instead. It will be automatically loaded.) <details><summary>Screenshot:</summary> ![image](https://user-images.githubusercontent.com/15856304/30413914-9df13904-9929-11e7-8890-f2207be2e551.png) ![image](https://user-images.githubusercontent.com/15856304/30413921-a4c76c62-9929-11e7-8fc5-09d11311f08e.png) </details><br/> ***System configuration*** **Filters:** - ID=0 Name="User filter" Version=1.0.0.0 Enabled=YES - ID=2 Name="English filter" Version=1.1.30.47 Enabled=YES - ID=3 Name="Spyware filter" Version=1.0.14.42 Enabled=YES - ID=4 Name="Social media filter" Version=1.0.23.9 Enabled=YES - ID=14 Name="Annoyances filter" Version=1.0.18.20 Enabled=YES - ID=6 Name="German filter" Version=1.0.36.61 Enabled=YES - ID=11 Name="Mobile Ads filter" Version=1.0.9.64 Enabled=YES - ID=12 Name="Safari filter" Version=1.0.3.60 Enabled=YES [//]: # (Substitute this line with the list of your active filters, separated by commas) [//]: # (Please enter the correct values for your case to the table below) Information | Value --- | --- Operating system: | iOS Browser: | Safari AdGuard version: | 1.3.1 Simplified filters (iOS only) | Off AdGuard DNS: | None Helpdesk ID (if exists): | 1480792 [//]: # (This template is meant for missed ad/false positive reports, for other type of reports edit it accordingly) [//]: # (If this is a crash report, include the crashlog with https://gist.github.com/)
1.0
http://m.azonline.de/ - sm widget, banners and ad leftovers - [//]: # (***You can delete or ignore strings starting with "[//]:" They will not be visible either way.) ***Description***: * **Current behaviour**: http://m.azonline.de/NRW/2980745-Prozesse-Steuererklaerung-verspaetet-abgegeben-Berater-verurteilt [//]: # (Substitute this line with a description of the problem) [//]: # (Replace %screenshot_url% below with a link to the screenshot of the problem. Also, you can paste image from clipboard instead. It will be automatically loaded.) <details><summary>Screenshot:</summary> ![image](https://user-images.githubusercontent.com/15856304/30413914-9df13904-9929-11e7-8890-f2207be2e551.png) ![image](https://user-images.githubusercontent.com/15856304/30413921-a4c76c62-9929-11e7-8fc5-09d11311f08e.png) </details><br/> ***System configuration*** **Filters:** - ID=0 Name="User filter" Version=1.0.0.0 Enabled=YES - ID=2 Name="English filter" Version=1.1.30.47 Enabled=YES - ID=3 Name="Spyware filter" Version=1.0.14.42 Enabled=YES - ID=4 Name="Social media filter" Version=1.0.23.9 Enabled=YES - ID=14 Name="Annoyances filter" Version=1.0.18.20 Enabled=YES - ID=6 Name="German filter" Version=1.0.36.61 Enabled=YES - ID=11 Name="Mobile Ads filter" Version=1.0.9.64 Enabled=YES - ID=12 Name="Safari filter" Version=1.0.3.60 Enabled=YES [//]: # (Substitute this line with the list of your active filters, separated by commas) [//]: # (Please enter the correct values for your case to the table below) Information | Value --- | --- Operating system: | iOS Browser: | Safari AdGuard version: | 1.3.1 Simplified filters (iOS only) | Off AdGuard DNS: | None Helpdesk ID (if exists): | 1480792 [//]: # (This template is meant for missed ad/false positive reports, for other type of reports edit it accordingly) [//]: # (If this is a crash report, include the crashlog with https://gist.github.com/)
non_infrastructure
sm widget banners and ad leftovers you can delete or ignore strings starting with they will not be visible either way description current behaviour substitute this line with a description of the problem replace screenshot url below with a link to the screenshot of the problem also you can paste image from clipboard instead it will be automatically loaded screenshot system configuration filters id name user filter version enabled yes id name english filter version enabled yes id name spyware filter version enabled yes id name social media filter version enabled yes id name annoyances filter version enabled yes id name german filter version enabled yes id name mobile ads filter version enabled yes id name safari filter version enabled yes substitute this line with the list of your active filters separated by commas please enter the correct values for your case to the table below information value operating system ios browser safari adguard version simplified filters ios only off adguard dns none helpdesk id if exists this template is meant for missed ad false positive reports for other type of reports edit it accordingly if this is a crash report include the crashlog with
0
3,416
4,309,476,831
IssuesEvent
2016-07-21 16:07:06
KhronosGroup/glslang
https://api.github.com/repos/KhronosGroup/glslang
closed
rationalizing warning levels
enhancement Infrastructure
A small change to CMakeLists.txt for discussion: #393. By default the g++ build was not checking for some classes of warnings that the MSVC build did, for example around unused variables and function parameters. It was easy to accidentally add code like that on a Linux environment. The above WIP PR adds those warnings for g++, and cleans up a few instances in the code where they trigger. I took a swing at adding the set that clang accepts, but I'm not set up to test this. The auto-build was happy :smile:, but maybe someone knows of a better set. If so, I can drop them in. I was unable to find a g++ warning to detect the MSVC *"warning C4267: 'initializing' : conversion from 'size_t' to 'unsigned int', possible loss of data"*, which triggers on code such as this: `int foo = myvector.size();` Other thoughts?
1.0
rationalizing warning levels - A small change to CMakeLists.txt for discussion: #393. By default the g++ build was not checking for some classes of warnings that the MSVC build did, for example around unused variables and function parameters. It was easy to accidentally add code like that on a Linux environment. The above WIP PR adds those warnings for g++, and cleans up a few instances in the code where they trigger. I took a swing at adding the set that clang accepts, but I'm not set up to test this. The auto-build was happy :smile:, but maybe someone knows of a better set. If so, I can drop them in. I was unable to find a g++ warning to detect the MSVC *"warning C4267: 'initializing' : conversion from 'size_t' to 'unsigned int', possible loss of data"*, which triggers on code such as this: `int foo = myvector.size();` Other thoughts?
infrastructure
rationalizing warning levels a small change to cmakelists txt for discussion by default the g build was not checking for some classes of warnings that the msvc build did for example around unused variables and function parameters it was easy to accidentally add code like that on a linux environment the above wip pr adds those warnings for g and cleans up a few instances in the code where they trigger i took a swing at adding the set that clang accepts but i m not set up to test this the auto build was happy smile but maybe someone knows of a better set if so i can drop them in i was unable to find a g warning to detect the msvc warning initializing conversion from size t to unsigned int possible loss of data which triggers on code such as this int foo myvector size other thoughts
1
23,826
16,606,937,459
IssuesEvent
2021-06-02 05:54:52
asciidoctor/asciidoctor
https://api.github.com/repos/asciidoctor/asciidoctor
closed
Import source and build for default stylesheet into this repository
infrastructure
The default stylesheet (for the HTML output) is currently generated by the asciidoctor-stylesheet-factory project. However, that project uses a build which relies on unmaintained software. It's also arduous and obscure to have to make changes to files into another project and import the generated result into this repository to be committed. The whole process will be a lot simpler if that build is located in this project. To avoid introducing unanticipated changes to the stylesheet, we'll need to work backwards from the minified/distributed end result until we get to something that's maintainable. This issue focuses on the first stage in that migration, which is described below: 1. Unminify data/stylesheets/asciidoctor-default.css (this will become the editable source file) * The source file will be maintained in plain CSS with some help from PostCSS 2. Set up and configure a npm-based build that uses postcss + cssnano to produce the same or similar minified result as before 3. Use the build to generate data/stylesheets/asciidoctor-default.min.css 4. Update the processor to read the default styles from the data/stylesheets/asciidoctor-default.min.css file instead of data/stylesheets/asciidoctor-default.css file * If this doesn't work out, the source file could be named data/styelsheets/asciidoctor-default.src.css instead 5. Configure the gemspec to omit the stylesheet source file from the gem We'll probably want to avoid committing the minified file to the repository and generating it on demand for the test suite and release, but we should do that in a later stage.
1.0
Import source and build for default stylesheet into this repository - The default stylesheet (for the HTML output) is currently generated by the asciidoctor-stylesheet-factory project. However, that project uses a build which relies on unmaintained software. It's also arduous and obscure to have to make changes to files into another project and import the generated result into this repository to be committed. The whole process will be a lot simpler if that build is located in this project. To avoid introducing unanticipated changes to the stylesheet, we'll need to work backwards from the minified/distributed end result until we get to something that's maintainable. This issue focuses on the first stage in that migration, which is described below: 1. Unminify data/stylesheets/asciidoctor-default.css (this will become the editable source file) * The source file will be maintained in plain CSS with some help from PostCSS 2. Set up and configure a npm-based build that uses postcss + cssnano to produce the same or similar minified result as before 3. Use the build to generate data/stylesheets/asciidoctor-default.min.css 4. Update the processor to read the default styles from the data/stylesheets/asciidoctor-default.min.css file instead of data/stylesheets/asciidoctor-default.css file * If this doesn't work out, the source file could be named data/styelsheets/asciidoctor-default.src.css instead 5. Configure the gemspec to omit the stylesheet source file from the gem We'll probably want to avoid committing the minified file to the repository and generating it on demand for the test suite and release, but we should do that in a later stage.
infrastructure
import source and build for default stylesheet into this repository the default stylesheet for the html output is currently generated by the asciidoctor stylesheet factory project however that project uses a build which relies on unmaintained software it s also arduous and obscure to have to make changes to files into another project and import the generated result into this repository to be committed the whole process will be a lot simpler if that build is located in this project to avoid introducing unanticipated changes to the stylesheet we ll need to work backwards from the minified distributed end result until we get to something that s maintainable this issue focuses on the first stage in that migration which is described below unminify data stylesheets asciidoctor default css this will become the editable source file the source file will be maintained in plain css with some help from postcss set up and configure a npm based build that uses postcss cssnano to produce the same or similar minified result as before use the build to generate data stylesheets asciidoctor default min css update the processor to read the default styles from the data stylesheets asciidoctor default min css file instead of data stylesheets asciidoctor default css file if this doesn t work out the source file could be named data styelsheets asciidoctor default src css instead configure the gemspec to omit the stylesheet source file from the gem we ll probably want to avoid committing the minified file to the repository and generating it on demand for the test suite and release but we should do that in a later stage
1
20,838
14,194,438,085
IssuesEvent
2020-11-15 03:47:13
opencv/opencv
https://api.github.com/repos/opencv/opencv
opened
OSX framework: enable testing of Objective-C/Swift bindings
category: infrastructure category: swift/objc bindings platform: ios/osx
It would be nice to enable OSX framework testing for OpenCV 4.x through `--run_tests` from `osx/build_framework.py` on MacOSX. Currently I see this error message: ``` Testing failed: OpenCVTestTests: Could not find module 'opencv2' for architecture 'x86_64'; found: x86_64-apple-macos Testing cancelled because the build failed. ``` @komakai Could you please confirm that this flow should work? Used this configuration of build environment: ``` $ sw_vers ProductName: Mac OS X ProductVersion: 10.14.6 BuildVersion: 18G87 $ clang --version Apple LLVM version 10.0.1 (clang-1001.0.46.4) Target: x86_64-apple-darwin18.7.0 Thread model: posix InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin $ swiftc --version Apple Swift version 5.0.1 (swiftlang-1001.0.82.4 clang-1001.0.46.5) Target: x86_64-apple-darwin18.7.0 $ cmake --version cmake version 3.15.2 MACOSX_DEPLOYMENT_TARGET = 10.12 (default from OpenCV's packaging scripts) ``` --- Extended logs: **TBD** (on GitHub)
1.0
OSX framework: enable testing of Objective-C/Swift bindings - It would be nice to enable OSX framework testing for OpenCV 4.x through `--run_tests` from `osx/build_framework.py` on MacOSX. Currently I see this error message: ``` Testing failed: OpenCVTestTests: Could not find module 'opencv2' for architecture 'x86_64'; found: x86_64-apple-macos Testing cancelled because the build failed. ``` @komakai Could you please confirm that this flow should work? Used this configuration of build environment: ``` $ sw_vers ProductName: Mac OS X ProductVersion: 10.14.6 BuildVersion: 18G87 $ clang --version Apple LLVM version 10.0.1 (clang-1001.0.46.4) Target: x86_64-apple-darwin18.7.0 Thread model: posix InstalledDir: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin $ swiftc --version Apple Swift version 5.0.1 (swiftlang-1001.0.82.4 clang-1001.0.46.5) Target: x86_64-apple-darwin18.7.0 $ cmake --version cmake version 3.15.2 MACOSX_DEPLOYMENT_TARGET = 10.12 (default from OpenCV's packaging scripts) ``` --- Extended logs: **TBD** (on GitHub)
infrastructure
osx framework enable testing of objective c swift bindings it would be nice to enable osx framework testing for opencv x through run tests from osx build framework py on macosx currently i see this error message testing failed opencvtesttests could not find module for architecture found apple macos testing cancelled because the build failed komakai could you please confirm that this flow should work used this configuration of build environment sw vers productname mac os x productversion buildversion clang version apple llvm version clang target apple thread model posix installeddir applications xcode app contents developer toolchains xcodedefault xctoolchain usr bin swiftc version apple swift version swiftlang clang target apple cmake version cmake version macosx deployment target default from opencv s packaging scripts extended logs tbd on github
1
29,410
23,990,666,314
IssuesEvent
2022-09-14 00:33:27
mwaskom/seaborn
https://api.github.com/repos/mwaskom/seaborn
closed
Look into jupytext for managing doc source notebooks
docs infrastructure
Would be nice to be able to edit prose, etc. in plain text, along with not having to track all the cell metadata, etc.
1.0
Look into jupytext for managing doc source notebooks - Would be nice to be able to edit prose, etc. in plain text, along with not having to track all the cell metadata, etc.
infrastructure
look into jupytext for managing doc source notebooks would be nice to be able to edit prose etc in plain text along with not having to track all the cell metadata etc
1
142,832
5,477,734,210
IssuesEvent
2017-03-12 11:37:44
dhowe/ReadersJS
https://api.github.com/repos/dhowe/ReadersJS
closed
Need perigrams for all texts
priority: High
Currently it seems we only have them for 'the Image': image-perigrams.txt
1.0
Need perigrams for all texts - Currently it seems we only have them for 'the Image': image-perigrams.txt
non_infrastructure
need perigrams for all texts currently it seems we only have them for the image image perigrams txt
0
27,014
21,006,030,276
IssuesEvent
2022-03-29 22:46:13
coq/coq
https://api.github.com/repos/coq/coq
opened
How to run coq_makefile from the development environment when compiling with dune?
kind: infrastructure
We have `dune exec -- dev/shim/coqide-prelude` to run CoqIDE in the dev environment and a similar command to run `coqtop`. How about the various utilities such as `coq_makefile`? Should I use the old make for this? And if so, how do I switch? Just do a clean? It will pretty painful to make any improvements to `coq_makefile` if you have to re-install Coq for each compile. (Not that I have a list of changes I want to make in-hand at the moment, but that may change shortly.)
1.0
How to run coq_makefile from the development environment when compiling with dune? - We have `dune exec -- dev/shim/coqide-prelude` to run CoqIDE in the dev environment and a similar command to run `coqtop`. How about the various utilities such as `coq_makefile`? Should I use the old make for this? And if so, how do I switch? Just do a clean? It will pretty painful to make any improvements to `coq_makefile` if you have to re-install Coq for each compile. (Not that I have a list of changes I want to make in-hand at the moment, but that may change shortly.)
infrastructure
how to run coq makefile from the development environment when compiling with dune we have dune exec dev shim coqide prelude to run coqide in the dev environment and a similar command to run coqtop how about the various utilities such as coq makefile should i use the old make for this and if so how do i switch just do a clean it will pretty painful to make any improvements to coq makefile if you have to re install coq for each compile not that i have a list of changes i want to make in hand at the moment but that may change shortly
1
12,578
9,855,067,933
IssuesEvent
2019-06-19 18:26:20
patternfly/patternfly-org
https://api.github.com/repos/patternfly/patternfly-org
closed
Improve patternfly-react and patternfly-next src inclusion in the repository
PF4 DEV infrastructure
@dana do you mind filling in Zack with details?
1.0
Improve patternfly-react and patternfly-next src inclusion in the repository - @dana do you mind filling in Zack with details?
infrastructure
improve patternfly react and patternfly next src inclusion in the repository dana do you mind filling in zack with details
1
162,849
25,603,946,487
IssuesEvent
2022-12-01 23:05:33
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
DSN: Update VBA MVP description with items from content audit
Design ⭐️ Facilities Regional office
## Description - #10234 - [MVP description](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/facilities/regional-offices/design/README.md) ## Acceptance Criteria - [x] MVP criteria doc updated ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [x] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
DSN: Update VBA MVP description with items from content audit - ## Description - #10234 - [MVP description](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/products/facilities/regional-offices/design/README.md) ## Acceptance Criteria - [x] MVP criteria doc updated ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [x] `⭐️ Facilities` - [ ] `⭐️ User support`
non_infrastructure
dsn update vba mvp description with items from content audit description acceptance criteria mvp criteria doc updated cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
0
27,772
22,332,419,715
IssuesEvent
2022-06-14 15:30:04
acts-project/acts
https://api.github.com/repos/acts-project/acts
closed
nodeps build fail
Bug Infrastructure
The linux-nodeps build failed in #1190 with the message: ``` FAILED: thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build thirdparty/boost/boost-prefix/src/boost/stage/lib/libboost_unit_test_framework.so.1.71.0 cd /__w/acts/acts/build/thirdparty/boost/boost-prefix/src/boost && /usr/bin/cmake -P /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake && /usr/bin/cmake -E touch /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build CMake Error at /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake:49 (message): Command failed: 1 './b2' 'variant=release' 'threading=multi' 'runtime-link=shared' 'link=shared' 'cxxflags=-std=c++17 -fPIC -O3' '--with-test' See also /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-*.log ``` I quickly surveyed other recent PR's but didn't see the failure. I cannot reproduce locally either. Still I'll leave this issue up until we're sure this was transient.
1.0
nodeps build fail - The linux-nodeps build failed in #1190 with the message: ``` FAILED: thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build thirdparty/boost/boost-prefix/src/boost/stage/lib/libboost_unit_test_framework.so.1.71.0 cd /__w/acts/acts/build/thirdparty/boost/boost-prefix/src/boost && /usr/bin/cmake -P /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake && /usr/bin/cmake -E touch /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build CMake Error at /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-Release.cmake:49 (message): Command failed: 1 './b2' 'variant=release' 'threading=multi' 'runtime-link=shared' 'link=shared' 'cxxflags=-std=c++17 -fPIC -O3' '--with-test' See also /__w/acts/acts/build/thirdparty/boost/boost_build_unit_test_framework-prefix/src/boost_build_unit_test_framework-stamp/boost_build_unit_test_framework-build-*.log ``` I quickly surveyed other recent PR's but didn't see the failure. I cannot reproduce locally either. Still I'll leave this issue up until we're sure this was transient.
infrastructure
nodeps build fail the linux nodeps build failed in with the message failed thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build thirdparty boost boost prefix src boost stage lib libboost unit test framework so cd w acts acts build thirdparty boost boost prefix src boost usr bin cmake p w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build release cmake usr bin cmake e touch w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build cmake error at w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build release cmake message command failed variant release threading multi runtime link shared link shared cxxflags std c fpic with test see also w acts acts build thirdparty boost boost build unit test framework prefix src boost build unit test framework stamp boost build unit test framework build log i quickly surveyed other recent pr s but didn t see the failure i cannot reproduce locally either still i ll leave this issue up until we re sure this was transient
1
1,833
3,392,861,312
IssuesEvent
2015-11-30 21:27:54
catapult-project/catapult
https://api.github.com/repos/catapult-project/catapult
opened
Non-chromium committers can't give commitable lgtms
Infrastructure
https://codereview.chromium.org/1469203002/#msg19 How do we fix that for @martijncoenen and our other contribs?
1.0
Non-chromium committers can't give commitable lgtms - https://codereview.chromium.org/1469203002/#msg19 How do we fix that for @martijncoenen and our other contribs?
infrastructure
non chromium committers can t give commitable lgtms how do we fix that for martijncoenen and our other contribs
1
53,154
13,128,685,301
IssuesEvent
2020-08-06 12:43:59
atc0005/send2teams
https://api.github.com/repos/atc0005/send2teams
opened
Use Docker-based GitHub Actions Workflows
CI builds dependencies enhancement linting
Swap out the GitHub Actions used for CI work with custom Docker containers from the atc0005/go-ci project. refs atc0005/todo#22
1.0
Use Docker-based GitHub Actions Workflows - Swap out the GitHub Actions used for CI work with custom Docker containers from the atc0005/go-ci project. refs atc0005/todo#22
non_infrastructure
use docker based github actions workflows swap out the github actions used for ci work with custom docker containers from the go ci project refs todo
0
138,400
11,200,552,758
IssuesEvent
2020-01-03 22:10:15
GTNewHorizons/NewHorizons
https://api.github.com/repos/GTNewHorizons/NewHorizons
closed
Singleblock miner bugs
FixedInDev need to be tested
#### Which modpack version are you using? 2.0.8.0pre w/updated GT++ # #### If in multiplayer; On which server does this happen? # #### What did you try to do, and what did you expect to happen? Crafted (in 2.0.8.0pre) and placed down a HV singleblock miner. Expected it to default to the maximum radius. # #### What happened instead? (Attach screenshots if needed) Defaults to minimum radius (3x3) instead of maximum (49x49), so I had to mash it with a screwdriver a bunch. # #### What do you suggest instead/what changes do you propose? Set the default radius to the maximum, and make the WAILA tooltip for the miner display the current operating radius.
1.0
Singleblock miner bugs - #### Which modpack version are you using? 2.0.8.0pre w/updated GT++ # #### If in multiplayer; On which server does this happen? # #### What did you try to do, and what did you expect to happen? Crafted (in 2.0.8.0pre) and placed down a HV singleblock miner. Expected it to default to the maximum radius. # #### What happened instead? (Attach screenshots if needed) Defaults to minimum radius (3x3) instead of maximum (49x49), so I had to mash it with a screwdriver a bunch. # #### What do you suggest instead/what changes do you propose? Set the default radius to the maximum, and make the WAILA tooltip for the miner display the current operating radius.
non_infrastructure
singleblock miner bugs which modpack version are you using w updated gt if in multiplayer on which server does this happen what did you try to do and what did you expect to happen crafted in and placed down a hv singleblock miner expected it to default to the maximum radius what happened instead attach screenshots if needed defaults to minimum radius instead of maximum so i had to mash it with a screwdriver a bunch what do you suggest instead what changes do you propose set the default radius to the maximum and make the waila tooltip for the miner display the current operating radius
0
67,861
9,103,220,012
IssuesEvent
2019-02-20 15:27:18
petermr/tigr2ess
https://api.github.com/repos/petermr/tigr2ess
closed
"Test run of ami-dictionary for the ami-jar - ami20190219."
documentation enhancement
#### Status of test run of ami-dictionary for the ami-jar - ami20190219 - successful. #### Input - wikipedia link e.g - https://en.wikipedia.org/wiki/Ocimum. #### Output - xml formatted dictionary file. #### Run time log. ``` ambarish123@ubuntu:~$ ami-dictionary create --input https://en.wikipedia.org/wiki/Ocimum --informat wikipage --dictionary Ocimum --outformat xml --directory OcimumDir/ Generic values (AMIDictionaryTool) ================================ basename null cproject ctree cTreeList null dryrun false excludeBase null excludeTrees null file types [] forceMake false includeBase null includeTrees null log4j logfile null verbose 0 Specific values (AMIDictionaryTool) ================================ dataCols null dictionary [Ocimum, --outformat, xml] dictionaryTop OcimumDir/ href null hrefCols null input https://en.wikipedia.org/wiki/Ocimum informat wikipage dictInformat null linkCol null log4j null nameCol null operation create outformats [xml] splitCol , termCol null terms null wikiLinks [wikipedia, wikidata] 0 [main] DEBUG org.contentmine.ami.tools.AMIDictionaryTool - extracting hyperlinks ............!....!!.....!.!......!...!!.!.!!.!!!!!!!!!!!!!!!!!.!!!!.!!!!!.!!!!!!!!!!!!!!!!!.!!!!!!!.....!.!..........!...!. !WP !WP !WP ++++++ !WP ++ !WP ++++++++ !WP !WP +++++++ !WP ++++ !WP !WP + !WP !WP + !WP + !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP + !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP ++++++++++++++++ Missing wikipedia: : "Genus: Ocimum L."; Isodon inflexus; Kew World Checklist of Selected Plant Families; Mosla scabra; O. campechianum; O. centraliafricanum; O. tenuiflorum; Ocimeae; ``` #### Output file - [OcimumDir.tar.gz](https://github.com/petermr/tigr2ess/files/2881314/OcimumDir.tar.gz)
1.0
"Test run of ami-dictionary for the ami-jar - ami20190219." - #### Status of test run of ami-dictionary for the ami-jar - ami20190219 - successful. #### Input - wikipedia link e.g - https://en.wikipedia.org/wiki/Ocimum. #### Output - xml formatted dictionary file. #### Run time log. ``` ambarish123@ubuntu:~$ ami-dictionary create --input https://en.wikipedia.org/wiki/Ocimum --informat wikipage --dictionary Ocimum --outformat xml --directory OcimumDir/ Generic values (AMIDictionaryTool) ================================ basename null cproject ctree cTreeList null dryrun false excludeBase null excludeTrees null file types [] forceMake false includeBase null includeTrees null log4j logfile null verbose 0 Specific values (AMIDictionaryTool) ================================ dataCols null dictionary [Ocimum, --outformat, xml] dictionaryTop OcimumDir/ href null hrefCols null input https://en.wikipedia.org/wiki/Ocimum informat wikipage dictInformat null linkCol null log4j null nameCol null operation create outformats [xml] splitCol , termCol null terms null wikiLinks [wikipedia, wikidata] 0 [main] DEBUG org.contentmine.ami.tools.AMIDictionaryTool - extracting hyperlinks ............!....!!.....!.!......!...!!.!.!!.!!!!!!!!!!!!!!!!!.!!!!.!!!!!.!!!!!!!!!!!!!!!!!.!!!!!!!.....!.!..........!...!. !WP !WP !WP ++++++ !WP ++ !WP ++++++++ !WP !WP +++++++ !WP ++++ !WP !WP + !WP !WP + !WP + !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP + !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP !WP + !WP !WP !WP !WP !WP !WP !WP ++++++++++++++++ Missing wikipedia: : "Genus: Ocimum L."; Isodon inflexus; Kew World Checklist of Selected Plant Families; Mosla scabra; O. campechianum; O. centraliafricanum; O. tenuiflorum; Ocimeae; ``` #### Output file - [OcimumDir.tar.gz](https://github.com/petermr/tigr2ess/files/2881314/OcimumDir.tar.gz)
non_infrastructure
test run of ami dictionary for the ami jar status of test run of ami dictionary for the ami jar successful input wikipedia link e g output xml formatted dictionary file run time log ubuntu ami dictionary create input informat wikipage dictionary ocimum outformat xml directory ocimumdir generic values amidictionarytool basename null cproject ctree ctreelist null dryrun false excludebase null excludetrees null file types forcemake false includebase null includetrees null logfile null verbose specific values amidictionarytool datacols null dictionary dictionarytop ocimumdir href null hrefcols null input informat wikipage dictinformat null linkcol null null namecol null operation create outformats splitcol termcol null terms null wikilinks debug org contentmine ami tools amidictionarytool extracting hyperlinks wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp wp missing wikipedia genus ocimum l isodon inflexus kew world checklist of selected plant families mosla scabra o campechianum o centraliafricanum o tenuiflorum ocimeae output file
0
463,148
13,260,508,331
IssuesEvent
2020-08-20 18:17:55
radical-cybertools/radical.saga
https://api.github.com/repos/radical-cybertools/radical.saga
closed
bsub option for NVME in the LSF scheduler
comp:lsf priority:medium topic:resouce:summit type:enhancement
On Summit, Burst Buffer can be activated via the job scheduler option like: ``` -allow_flags NVME ``` There are three options that I can think of now: - make this option as a default on Summit, then all reserved/allocated nodes have NVMe storage devices mounted under `/mnt/bb/$USER` - create a special branch to use NVMe - manual modification directly in the saga code (similarily to SMT level adjustment i.e. #707) What would be a reasonable choice or is there a better option?
1.0
bsub option for NVME in the LSF scheduler - On Summit, Burst Buffer can be activated via the job scheduler option like: ``` -allow_flags NVME ``` There are three options that I can think of now: - make this option as a default on Summit, then all reserved/allocated nodes have NVMe storage devices mounted under `/mnt/bb/$USER` - create a special branch to use NVMe - manual modification directly in the saga code (similarily to SMT level adjustment i.e. #707) What would be a reasonable choice or is there a better option?
non_infrastructure
bsub option for nvme in the lsf scheduler on summit burst buffer can be activated via the job scheduler option like allow flags nvme there are three options that i can think of now make this option as a default on summit then all reserved allocated nodes have nvme storage devices mounted under mnt bb user create a special branch to use nvme manual modification directly in the saga code similarily to smt level adjustment i e what would be a reasonable choice or is there a better option
0
31,613
25,935,472,274
IssuesEvent
2022-12-16 13:47:39
ocaml/ocaml.org
https://api.github.com/repos/ocaml/ocaml.org
closed
Availability of the caml-list archive
bug data infrastructure
On the [Community page](https://ocaml.org/community), the [Mailing list](https://inbox.ocaml.org/caml-list/) link (<https://inbox.ocaml.org/caml-list/>) appears to be dead. I hit this issue while trying to locate the recommended/current/pleasant archives for the caml-list. I have not found such archives -- the best I had were located on inria.fr, but only contain older posts. Funnily enough, looking for this finds [this reply](https://sympa.inria.fr/sympa/arc/caml-list/2015-12/msg00122.html) in 2015 which contains an issue link that is itself dead (the ocaml.org repo rename made those links invalid) but is actually https://github.com/ocaml/v2.ocaml.org/issues/704. I'm curious if anyone knows of the current recommend link to caml-list archives.
1.0
Availability of the caml-list archive - On the [Community page](https://ocaml.org/community), the [Mailing list](https://inbox.ocaml.org/caml-list/) link (<https://inbox.ocaml.org/caml-list/>) appears to be dead. I hit this issue while trying to locate the recommended/current/pleasant archives for the caml-list. I have not found such archives -- the best I had were located on inria.fr, but only contain older posts. Funnily enough, looking for this finds [this reply](https://sympa.inria.fr/sympa/arc/caml-list/2015-12/msg00122.html) in 2015 which contains an issue link that is itself dead (the ocaml.org repo rename made those links invalid) but is actually https://github.com/ocaml/v2.ocaml.org/issues/704. I'm curious if anyone knows of the current recommend link to caml-list archives.
infrastructure
availability of the caml list archive on the the link appears to be dead i hit this issue while trying to locate the recommended current pleasant archives for the caml list i have not found such archives the best i had were located on inria fr but only contain older posts funnily enough looking for this finds in which contains an issue link that is itself dead the ocaml org repo rename made those links invalid but is actually i m curious if anyone knows of the current recommend link to caml list archives
1
2,208
3,561,536,532
IssuesEvent
2016-01-23 21:25:15
calblueprint/edge-yl
https://api.github.com/repos/calblueprint/edge-yl
opened
Configure conferences request for rooms page
actions infrastructure stores
If an initial `conference_id` is not passed in as a parameter, first fetch conferences and then fetch the rooms belonging to the first one. If an initial `conference_id` is passed in as a parameter, fetch rooms belonging to that conference and then fetch conferences (to allow switching conferences).
1.0
Configure conferences request for rooms page - If an initial `conference_id` is not passed in as a parameter, first fetch conferences and then fetch the rooms belonging to the first one. If an initial `conference_id` is passed in as a parameter, fetch rooms belonging to that conference and then fetch conferences (to allow switching conferences).
infrastructure
configure conferences request for rooms page if an initial conference id is not passed in as a parameter first fetch conferences and then fetch the rooms belonging to the first one if an initial conference id is passed in as a parameter fetch rooms belonging to that conference and then fetch conferences to allow switching conferences
1
35,019
30,685,483,778
IssuesEvent
2023-07-26 12:05:03
celestiaorg/devops
https://api.github.com/repos/celestiaorg/devops
closed
feat: torch - expose multiaddr as metrics
enhancement kubernetes observability infrastructure devops development tooling
expose the multiaddr as metrics via otel/prometheus same approach as we have in watchdog to get the values from Grafana
1.0
feat: torch - expose multiaddr as metrics - expose the multiaddr as metrics via otel/prometheus same approach as we have in watchdog to get the values from Grafana
infrastructure
feat torch expose multiaddr as metrics expose the multiaddr as metrics via otel prometheus same approach as we have in watchdog to get the values from grafana
1
109
2,524,667,902
IssuesEvent
2015-01-20 19:18:40
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
opened
Windows dist builders hit upload exception due to locking bug in buildbot
A-infrastructure
There's a bug in the buildbot DirectoryUpload buildstep on windows that causes exceptions on our dist builders. This happens pretty regularly. It can be avoided by limiting slaves to 1 build at a time, but we don't want to do that on auto builders because it seems to leave them underutilized and lets them shut down to frequently (which then costs time to start up again).
1.0
Windows dist builders hit upload exception due to locking bug in buildbot - There's a bug in the buildbot DirectoryUpload buildstep on windows that causes exceptions on our dist builders. This happens pretty regularly. It can be avoided by limiting slaves to 1 build at a time, but we don't want to do that on auto builders because it seems to leave them underutilized and lets them shut down to frequently (which then costs time to start up again).
infrastructure
windows dist builders hit upload exception due to locking bug in buildbot there s a bug in the buildbot directoryupload buildstep on windows that causes exceptions on our dist builders this happens pretty regularly it can be avoided by limiting slaves to build at a time but we don t want to do that on auto builders because it seems to leave them underutilized and lets them shut down to frequently which then costs time to start up again
1
262,642
8,272,279,735
IssuesEvent
2018-09-16 18:27:56
javaee/glassfish
https://api.github.com/repos/javaee/glassfish
closed
certificate prompt ignores the --interactive=false option
Component: command_line_interface Priority: Trivial Type: Bug
When secure admin is enabled, a prompt is generated by the AsadminTrustManager class to see if the user wants to trust the certificate. This class uses the following code to set whether the prompt should be generated: Console cons = System.console(); if (cons != null) { However, the asadmin command has a --interactive option that is supposed to prevent prompts when set to false. The AsadminTrustManager is ignoring the --interactive option. #### Affected Versions [3.1_dev]
1.0
certificate prompt ignores the --interactive=false option - When secure admin is enabled, a prompt is generated by the AsadminTrustManager class to see if the user wants to trust the certificate. This class uses the following code to set whether the prompt should be generated: Console cons = System.console(); if (cons != null) { However, the asadmin command has a --interactive option that is supposed to prevent prompts when set to false. The AsadminTrustManager is ignoring the --interactive option. #### Affected Versions [3.1_dev]
non_infrastructure
certificate prompt ignores the interactive false option when secure admin is enabled a prompt is generated by the asadmintrustmanager class to see if the user wants to trust the certificate this class uses the following code to set whether the prompt should be generated console cons system console if cons null however the asadmin command has a interactive option that is supposed to prevent prompts when set to false the asadmintrustmanager is ignoring the interactive option affected versions
0
30,300
24,746,836,242
IssuesEvent
2022-10-21 10:24:18
react-keycloak/react-native-keycloak
https://api.github.com/repos/react-keycloak/react-native-keycloak
closed
Error: fetchTokens failed
infrastructure
**Description** Library constantly throws "Error: fetchTokens failed" after authorization. The authorization itself is successful. If done from a browser, login and password are valid and it opens the application, but when done from the app it closes in-app-browser and throws the aforementioned error. The keycloak server is setup properly, other web projects use it successfully, problem only occurs in react native implementation. onEvent triggers with the following message: `onKeycloakEvent onAuthError {"error": [Error: fetchTokens failed], "error_description": "Failed to refresh token during callback processing"}` **To Reproduce** Steps: 1) Start application 2) Press Login 3) Enter email and password 4) Press Sign In **Expected behavior** After authorization keycloak updates state to authenticated. **Smartphone** - Device: Android Emulator - OS: Android 11 **Additional context** Here's the app code (same as the example, but with different app uri scheme and different keycloak url, realm, clientId): ```js const Login = () => { const {keycloak} = useKeycloak(); return ( <View style={styles.container}> <Text>{`Welcome ${keycloak?.authenticated} - ${keycloak?.token}!`}</Text> <Button onPress={() => keycloak?.login()} title="Login" /> </View> ) } const App = () => { return ( <ReactNativeKeycloakProvider authClient={keycloak} initOptions={{ redirectUri: 'posbpay://Homepage' }} onEvent={(event, error) => { console.log('onKeycloakEvent', event, error); }} > <Login /> </ReactNativeKeycloakProvider> ); } ``` Intent filter configuration in AndroidManifest: ```xml <intent-filter> <data android:scheme="posbpay" /> <action android:name="android.intent.action.VIEW" /> <category android:name="android.intent.category.DEFAULT" /> <category android:name="android.intent.category.BROWSABLE" /> </intent-filter> ```
1.0
Error: fetchTokens failed - **Description** Library constantly throws "Error: fetchTokens failed" after authorization. The authorization itself is successful. If done from a browser, login and password are valid and it opens the application, but when done from the app it closes in-app-browser and throws the aforementioned error. The keycloak server is setup properly, other web projects use it successfully, problem only occurs in react native implementation. onEvent triggers with the following message: `onKeycloakEvent onAuthError {"error": [Error: fetchTokens failed], "error_description": "Failed to refresh token during callback processing"}` **To Reproduce** Steps: 1) Start application 2) Press Login 3) Enter email and password 4) Press Sign In **Expected behavior** After authorization keycloak updates state to authenticated. **Smartphone** - Device: Android Emulator - OS: Android 11 **Additional context** Here's the app code (same as the example, but with different app uri scheme and different keycloak url, realm, clientId): ```js const Login = () => { const {keycloak} = useKeycloak(); return ( <View style={styles.container}> <Text>{`Welcome ${keycloak?.authenticated} - ${keycloak?.token}!`}</Text> <Button onPress={() => keycloak?.login()} title="Login" /> </View> ) } const App = () => { return ( <ReactNativeKeycloakProvider authClient={keycloak} initOptions={{ redirectUri: 'posbpay://Homepage' }} onEvent={(event, error) => { console.log('onKeycloakEvent', event, error); }} > <Login /> </ReactNativeKeycloakProvider> ); } ``` Intent filter configuration in AndroidManifest: ```xml <intent-filter> <data android:scheme="posbpay" /> <action android:name="android.intent.action.VIEW" /> <category android:name="android.intent.category.DEFAULT" /> <category android:name="android.intent.category.BROWSABLE" /> </intent-filter> ```
infrastructure
error fetchtokens failed description library constantly throws error fetchtokens failed after authorization the authorization itself is successful if done from a browser login and password are valid and it opens the application but when done from the app it closes in app browser and throws the aforementioned error the keycloak server is setup properly other web projects use it successfully problem only occurs in react native implementation onevent triggers with the following message onkeycloakevent onautherror error error description failed to refresh token during callback processing to reproduce steps start application press login enter email and password press sign in expected behavior after authorization keycloak updates state to authenticated smartphone device android emulator os android additional context here s the app code same as the example but with different app uri scheme and different keycloak url realm clientid js const login const keycloak usekeycloak return welcome keycloak authenticated keycloak token keycloak login title login const app return reactnativekeycloakprovider authclient keycloak initoptions redirecturi posbpay homepage onevent event error console log onkeycloakevent event error intent filter configuration in androidmanifest xml
1
160
2,544,860,635
IssuesEvent
2015-01-29 13:35:06
debops/debops-playbooks
https://api.github.com/repos/debops/debops-playbooks
opened
Add OpenDCIM role
Cluster infrastructure
[OpenDCIM](http://opendcim.org/) is an open source Data Center Infrastructure Management application. This might require support of Basic Authentication in `nginx` to work properly.
1.0
Add OpenDCIM role - [OpenDCIM](http://opendcim.org/) is an open source Data Center Infrastructure Management application. This might require support of Basic Authentication in `nginx` to work properly.
infrastructure
add opendcim role is an open source data center infrastructure management application this might require support of basic authentication in nginx to work properly
1
68,250
21,569,871,766
IssuesEvent
2022-05-02 06:42:02
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Cannot type the letters 'XP' in the client without it being turned into an emoji.
T-Defect
### Steps to reproduce 1. Where are you starting? What can you see? 2. What do you click? 3. More steps… ### Outcome #### What did you expect? The letters "XP" to appear (like Windows XP) #### What happened instead? Got an emoji. No way to escape it. ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
1.0
Cannot type the letters 'XP' in the client without it being turned into an emoji. - ### Steps to reproduce 1. Where are you starting? What can you see? 2. What do you click? 3. More steps… ### Outcome #### What did you expect? The letters "XP" to appear (like Windows XP) #### What happened instead? Got an emoji. No way to escape it. ### Operating system _No response_ ### Browser information _No response_ ### URL for webapp _No response_ ### Application version _No response_ ### Homeserver _No response_ ### Will you send logs? No
non_infrastructure
cannot type the letters xp in the client without it being turned into an emoji steps to reproduce where are you starting what can you see what do you click more steps… outcome what did you expect the letters xp to appear like windows xp what happened instead got an emoji no way to escape it operating system no response browser information no response url for webapp no response application version no response homeserver no response will you send logs no
0
28,264
23,120,326,103
IssuesEvent
2022-07-27 20:44:20
dotnet/project-system
https://api.github.com/repos/dotnet/project-system
closed
Get APEX nuget packages to upload to a public feed automatically
Feature Request Area-Infrastructure Triage-Approved
currently APEX package creation and publish is a purely manual process Relates to [AB#1087091](https://devdiv.visualstudio.com/0bdbc590-a062-4c3f-b0f6-9383f67865ee/_workitems/edit/1087091)
1.0
Get APEX nuget packages to upload to a public feed automatically - currently APEX package creation and publish is a purely manual process Relates to [AB#1087091](https://devdiv.visualstudio.com/0bdbc590-a062-4c3f-b0f6-9383f67865ee/_workitems/edit/1087091)
infrastructure
get apex nuget packages to upload to a public feed automatically currently apex package creation and publish is a purely manual process relates to
1
7,953
7,160,773,401
IssuesEvent
2018-01-28 05:27:18
srinikoganti/sams
https://api.github.com/repos/srinikoganti/sams
closed
College Infrastructure: Add new infrastructure issue
Infrastructure P1
Error: Getting server breakdown error when trying to add infrastructure with maximum digits in Hostel capacity text box. Please observe the below screen shot. ![hosteldigits error](https://user-images.githubusercontent.com/33898419/34152168-add8cbf2-e4d3-11e7-81b6-a86a0c9d34f2.png) ![hosteldigiterror](https://user-images.githubusercontent.com/33898419/34152289-19c6046a-e4d4-11e7-9c6e-7014aa28b848.png) Like shown below, Please check all the digits accepting text box maximum length giving server break down. ![allcapacity](https://user-images.githubusercontent.com/33898419/34152358-56139de2-e4d4-11e7-8357-d0a60cefd6b3.png)
1.0
College Infrastructure: Add new infrastructure issue - Error: Getting server breakdown error when trying to add infrastructure with maximum digits in Hostel capacity text box. Please observe the below screen shot. ![hosteldigits error](https://user-images.githubusercontent.com/33898419/34152168-add8cbf2-e4d3-11e7-81b6-a86a0c9d34f2.png) ![hosteldigiterror](https://user-images.githubusercontent.com/33898419/34152289-19c6046a-e4d4-11e7-9c6e-7014aa28b848.png) Like shown below, Please check all the digits accepting text box maximum length giving server break down. ![allcapacity](https://user-images.githubusercontent.com/33898419/34152358-56139de2-e4d4-11e7-8357-d0a60cefd6b3.png)
infrastructure
college infrastructure add new infrastructure issue error getting server breakdown error when trying to add infrastructure with maximum digits in hostel capacity text box please observe the below screen shot like shown below please check all the digits accepting text box maximum length giving server break down
1
7,900
7,129,680,301
IssuesEvent
2018-01-22 00:48:26
Khan/KaTeX
https://api.github.com/repos/Khan/KaTeX
closed
switch from browserify to webpack for building distribution files
infrastructure
Reasons: - we already have a webpack.config.js with settings for the unminified distribution files - webpack has cool plugins we can use to better optimize our code, e.g. https://webpack.js.org/plugins/module-concatenation-plugin/ to dedupe helper code that babel inserts, as discussed in #872.
1.0
switch from browserify to webpack for building distribution files - Reasons: - we already have a webpack.config.js with settings for the unminified distribution files - webpack has cool plugins we can use to better optimize our code, e.g. https://webpack.js.org/plugins/module-concatenation-plugin/ to dedupe helper code that babel inserts, as discussed in #872.
infrastructure
switch from browserify to webpack for building distribution files reasons we already have a webpack config js with settings for the unminified distribution files webpack has cool plugins we can use to better optimize our code e g to dedupe helper code that babel inserts as discussed in
1
1,860
3,410,012,252
IssuesEvent
2015-12-04 18:11:31
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
reopened
[Test Failure] Test runner infrastructure failure in prtest/win/dbg/unit32 on PR #407
Area-Infrastructure Contributor Pain Flaky
PR: [#407](https://github.com/dotnet/roslyn-internal/pull/407) ** by @ Failure: http://dotnet-ci.cloudapp.net/job/roslyn-internal_prtest_win_dbg_unit32/324/ **Issue 1: Test runner infrastructure failure** The Test runner saw an exception that was neither a test failure nor a build failure and must have come from the test runner.
1.0
[Test Failure] Test runner infrastructure failure in prtest/win/dbg/unit32 on PR #407 - PR: [#407](https://github.com/dotnet/roslyn-internal/pull/407) ** by @ Failure: http://dotnet-ci.cloudapp.net/job/roslyn-internal_prtest_win_dbg_unit32/324/ **Issue 1: Test runner infrastructure failure** The Test runner saw an exception that was neither a test failure nor a build failure and must have come from the test runner.
infrastructure
test runner infrastructure failure in prtest win dbg on pr pr by failure issue test runner infrastructure failure the test runner saw an exception that was neither a test failure nor a build failure and must have come from the test runner
1
46,436
24,532,634,884
IssuesEvent
2022-10-11 17:47:37
liquibase/liquibase
https://api.github.com/repos/liquibase/liquibase
closed
NetUtils.getLocalHostName() takes 30s
TypeEnhancement Severity3 ImpactLow PRAssociated DBAll IntegrationMaven BBroad ver4.3.5 OSMac performance
<!--- This environment context section helps us quickly review your PR. Please take a minute to fill-out this information. --> ## Environment **Liquibase Version**: 4.3.5 **Liquibase Integration & Version**: maven **Liquibase Extension(s) & Version**: **Database Vendor & Version**: **Operating System Type & Version**: MacOS 11.5.2 Big Sur ## Description During liquibase startup, the call `database.getDefaultSchemaName();` in `liquibase.changelog.ChangeLogParameters.java:61` leads to resolving the current hostname by `liquibase.util.NetUtil.getLocalHostName` in `NetUtil.java:71`. The subsequent line 75 takes 30s to resolve the hostname for the IP given by line 73. Interestingly, line 73 doesn't return 127.0.0.1 (from network interface eth0) but another IP taken from another network interface (eth1 / company internal network IP, here '/192.168.254.80' - yes starting with the slash. This IP can't resolved as fast as normally done by java vm. `liquibase.util.NetUtil.getLocalHostName()` ```java InetAddress localhost = getLocalHost(); // => '/192.168.254.80' ... yes with the leading slash if (localhost != null) { return localhost.getHostName(); // needs 30s } ``` ## Expected/Desired Behavior Same as normal JVM resolution, tested by the follwing: ```java @Test void testResolveLocalhost() throws UnknownHostException { InetAddress ia = InetAddress.getLocalHost(); String hostname = ia.getHostName(); // 6ms } @Test void testResolveLiquibase() throws UnknownHostException, SocketException { String ia = NetUtil.getLocalHostAddress(); //30s } ``` ## Additional Context There are solution on Stackoverflow for the general host resolving problem in JVM. ( https://stackoverflow.com/questions/39636792/jvm-takes-a-long-time-to-resolve-ip-address-for-localhost/39698914#39698914 ) But the solutiuons given there didn't work.
True
NetUtils.getLocalHostName() takes 30s - <!--- This environment context section helps us quickly review your PR. Please take a minute to fill-out this information. --> ## Environment **Liquibase Version**: 4.3.5 **Liquibase Integration & Version**: maven **Liquibase Extension(s) & Version**: **Database Vendor & Version**: **Operating System Type & Version**: MacOS 11.5.2 Big Sur ## Description During liquibase startup, the call `database.getDefaultSchemaName();` in `liquibase.changelog.ChangeLogParameters.java:61` leads to resolving the current hostname by `liquibase.util.NetUtil.getLocalHostName` in `NetUtil.java:71`. The subsequent line 75 takes 30s to resolve the hostname for the IP given by line 73. Interestingly, line 73 doesn't return 127.0.0.1 (from network interface eth0) but another IP taken from another network interface (eth1 / company internal network IP, here '/192.168.254.80' - yes starting with the slash. This IP can't resolved as fast as normally done by java vm. `liquibase.util.NetUtil.getLocalHostName()` ```java InetAddress localhost = getLocalHost(); // => '/192.168.254.80' ... yes with the leading slash if (localhost != null) { return localhost.getHostName(); // needs 30s } ``` ## Expected/Desired Behavior Same as normal JVM resolution, tested by the follwing: ```java @Test void testResolveLocalhost() throws UnknownHostException { InetAddress ia = InetAddress.getLocalHost(); String hostname = ia.getHostName(); // 6ms } @Test void testResolveLiquibase() throws UnknownHostException, SocketException { String ia = NetUtil.getLocalHostAddress(); //30s } ``` ## Additional Context There are solution on Stackoverflow for the general host resolving problem in JVM. ( https://stackoverflow.com/questions/39636792/jvm-takes-a-long-time-to-resolve-ip-address-for-localhost/39698914#39698914 ) But the solutiuons given there didn't work.
non_infrastructure
netutils getlocalhostname takes this environment context section helps us quickly review your pr please take a minute to fill out this information environment liquibase version liquibase integration version maven liquibase extension s version database vendor version operating system type version macos big sur description during liquibase startup the call database getdefaultschemaname in liquibase changelog changelogparameters java leads to resolving the current hostname by liquibase util netutil getlocalhostname in netutil java the subsequent line takes to resolve the hostname for the ip given by line interestingly line doesn t return from network interface but another ip taken from another network interface company internal network ip here yes starting with the slash this ip can t resolved as fast as normally done by java vm liquibase util netutil getlocalhostname java inetaddress localhost getlocalhost yes with the leading slash if localhost null return localhost gethostname needs expected desired behavior same as normal jvm resolution tested by the follwing java test void testresolvelocalhost throws unknownhostexception inetaddress ia inetaddress getlocalhost string hostname ia gethostname test void testresolveliquibase throws unknownhostexception socketexception string ia netutil getlocalhostaddress additional context there are solution on stackoverflow for the general host resolving problem in jvm but the solutiuons given there didn t work
0
35,435
31,280,657,277
IssuesEvent
2023-08-22 09:22:28
Yonatan-Lavie/aws-cdk-mongodb-api
https://api.github.com/repos/Yonatan-Lavie/aws-cdk-mongodb-api
closed
AWS Dev Environment Setup
infrastructure
- create IAM user named aws-cdk-prod/dev - create IAM policy named `Prod/Dev-Environment-Policy` - Attach the `Prod-Environment-Policy` to `aws-cdk-prod/dev` IAM User - Enable Programmatic Access - Create Access Key - Store in GitHub Secrets Access key as `AWS_PROD/DEV_ACCESS_KEY` and Secret access key as `AWS_PROD/DEV_SECRET_ACCESS_KEY` - create a new AWS profile in VSCode for prod/dev - document
1.0
AWS Dev Environment Setup - - create IAM user named aws-cdk-prod/dev - create IAM policy named `Prod/Dev-Environment-Policy` - Attach the `Prod-Environment-Policy` to `aws-cdk-prod/dev` IAM User - Enable Programmatic Access - Create Access Key - Store in GitHub Secrets Access key as `AWS_PROD/DEV_ACCESS_KEY` and Secret access key as `AWS_PROD/DEV_SECRET_ACCESS_KEY` - create a new AWS profile in VSCode for prod/dev - document
infrastructure
aws dev environment setup create iam user named aws cdk prod dev create iam policy named prod dev environment policy attach the prod environment policy to aws cdk prod dev iam user enable programmatic access create access key store in github secrets access key as aws prod dev access key and secret access key as aws prod dev secret access key create a new aws profile in vscode for prod dev document
1
76,117
9,917,466,127
IssuesEvent
2019-06-29 00:43:12
gatsbyjs/gatsby
https://api.github.com/repos/gatsbyjs/gatsby
opened
[docs][guides] improvements to Why Gatsby Uses GraphQL
help wanted topic: feedback type: documentation
## Summary In feedback from gatsbyjs.org, some questions surfaced that would be nice improvements to the [Why Gatsby Uses GraphQL Page](https://www.gatsbyjs.org/docs/why-gatsby-uses-graphql/). Specifically: > "What causes allProductsJson to appear in the graphql results? What/where is it defined? Is this example on github? A link to the working example would be helpful." For folks following along, the introduction of `allProductsJson` isn't explained clearly enough. It would help to explain where that key comes from, in explicit detail! @jlengstorf is there also a live example we can link to? This issue is open for anyone to pick up, once we answer where the live example lives (if there isn't one, we should create one).
1.0
[docs][guides] improvements to Why Gatsby Uses GraphQL - ## Summary In feedback from gatsbyjs.org, some questions surfaced that would be nice improvements to the [Why Gatsby Uses GraphQL Page](https://www.gatsbyjs.org/docs/why-gatsby-uses-graphql/). Specifically: > "What causes allProductsJson to appear in the graphql results? What/where is it defined? Is this example on github? A link to the working example would be helpful." For folks following along, the introduction of `allProductsJson` isn't explained clearly enough. It would help to explain where that key comes from, in explicit detail! @jlengstorf is there also a live example we can link to? This issue is open for anyone to pick up, once we answer where the live example lives (if there isn't one, we should create one).
non_infrastructure
improvements to why gatsby uses graphql summary in feedback from gatsbyjs org some questions surfaced that would be nice improvements to the specifically what causes allproductsjson to appear in the graphql results what where is it defined is this example on github a link to the working example would be helpful for folks following along the introduction of allproductsjson isn t explained clearly enough it would help to explain where that key comes from in explicit detail jlengstorf is there also a live example we can link to this issue is open for anyone to pick up once we answer where the live example lives if there isn t one we should create one
0
2,533
3,739,438,360
IssuesEvent
2016-03-09 04:49:04
polyball/polyball
https://api.github.com/repos/polyball/polyball
closed
Integrate TravisCI (our CI build tool) with slack
Tooling/Infrastructure
It is important that members who break the build are notified so that they can fix it asap. It is currently inconvenient to check Travis-CI.org every time you push.
1.0
Integrate TravisCI (our CI build tool) with slack - It is important that members who break the build are notified so that they can fix it asap. It is currently inconvenient to check Travis-CI.org every time you push.
infrastructure
integrate travisci our ci build tool with slack it is important that members who break the build are notified so that they can fix it asap it is currently inconvenient to check travis ci org every time you push
1
133,254
18,286,301,395
IssuesEvent
2021-10-05 10:39:46
lowRISC/opentitan
https://api.github.com/repos/lowRISC/opentitan
closed
RFC: Key sideload scheme
RFC:Proposal Component:Security Hotlist:Security
## Objective The objective of this RFC is to collect feedback on sideload architecture options, and decide what level of support to implement in the key manager integrated into Earl Grey. The RFC is currently in open for comments. Once a consensus has been reached, the RFC will be updated with the design decision before promoting it to the Technical Committee for final review. ## Background - What is Sideload? Sideload, or key manager sideload, refers to a direct hardware connection from the OpenTitan key manager module to hardware symmetric crypto components for supplying keys. When sideload is employed, keys are generated within the key manager and fed directly to the hardware crypto components for usage. At no point during this process are the keys exposed to software. Thus, compromised software can only abuse the keys but cannot permanently expose them [see doc for more details][no-sideload-proposal]. ## Design Options Currently, there are a couple of proposals under consideration. As part of this RFC, we are requesting inputs for additional proposals. ### Key slot implementation The implementation may provide a limited number of key slots in the key manager, with support for sideloading keys into the `HMAC`, `KMAC`, and `AES` engines. The number of key slots is left as an implementation detail which needs to take into consideration tradeoffs between: use cases versus hardware cost. Additional considerations: * Even though software does not have visibility into the key material, it is still responsible for managing the key slot mapping to consuming applications. * The number of bits sideloaded to other IP may be constrained to meet a security level up to 256 bits. * Sideloading keys into `OTBN` is not considered a security feature given that the target does not implement isolation of critical data from software. The current recommendation is to avoid sideloading into `OTBN`. ### No-Sideload Support Given the limitations of the key-slot proposal above, and additional considerations, we have put together a justification to remove support for sideloading from the architecture. See the [proposal document][no-sideload-proposal] for more details. [no-sideload-proposal]: https://docs.google.com/document/d/1kTh5E9wgD-X1eQSX3h-67tFvcmHTjjo7ng5bZKYIA-E/edit?ts=5ee97acc# ## Update See [sideload memo](https://docs.google.com/document/d/1BVxsA6-19eUuempvtiv136LH8Z0GLnc5cpapmk0H1uw/edit?usp=sharing) for the latest thinking and proposals. The ongoing [key manager update deck](https://docs.google.com/presentation/d/1noUFDWAHlcz5zHByuav6I9z7HrNM3hsiyYwmi2z7N3s/edit?usp=sharing) also details the matching hardware updates for key manager only.
True
RFC: Key sideload scheme - ## Objective The objective of this RFC is to collect feedback on sideload architecture options, and decide what level of support to implement in the key manager integrated into Earl Grey. The RFC is currently in open for comments. Once a consensus has been reached, the RFC will be updated with the design decision before promoting it to the Technical Committee for final review. ## Background - What is Sideload? Sideload, or key manager sideload, refers to a direct hardware connection from the OpenTitan key manager module to hardware symmetric crypto components for supplying keys. When sideload is employed, keys are generated within the key manager and fed directly to the hardware crypto components for usage. At no point during this process are the keys exposed to software. Thus, compromised software can only abuse the keys but cannot permanently expose them [see doc for more details][no-sideload-proposal]. ## Design Options Currently, there are a couple of proposals under consideration. As part of this RFC, we are requesting inputs for additional proposals. ### Key slot implementation The implementation may provide a limited number of key slots in the key manager, with support for sideloading keys into the `HMAC`, `KMAC`, and `AES` engines. The number of key slots is left as an implementation detail which needs to take into consideration tradeoffs between: use cases versus hardware cost. Additional considerations: * Even though software does not have visibility into the key material, it is still responsible for managing the key slot mapping to consuming applications. * The number of bits sideloaded to other IP may be constrained to meet a security level up to 256 bits. * Sideloading keys into `OTBN` is not considered a security feature given that the target does not implement isolation of critical data from software. The current recommendation is to avoid sideloading into `OTBN`. ### No-Sideload Support Given the limitations of the key-slot proposal above, and additional considerations, we have put together a justification to remove support for sideloading from the architecture. See the [proposal document][no-sideload-proposal] for more details. [no-sideload-proposal]: https://docs.google.com/document/d/1kTh5E9wgD-X1eQSX3h-67tFvcmHTjjo7ng5bZKYIA-E/edit?ts=5ee97acc# ## Update See [sideload memo](https://docs.google.com/document/d/1BVxsA6-19eUuempvtiv136LH8Z0GLnc5cpapmk0H1uw/edit?usp=sharing) for the latest thinking and proposals. The ongoing [key manager update deck](https://docs.google.com/presentation/d/1noUFDWAHlcz5zHByuav6I9z7HrNM3hsiyYwmi2z7N3s/edit?usp=sharing) also details the matching hardware updates for key manager only.
non_infrastructure
rfc key sideload scheme objective the objective of this rfc is to collect feedback on sideload architecture options and decide what level of support to implement in the key manager integrated into earl grey the rfc is currently in open for comments once a consensus has been reached the rfc will be updated with the design decision before promoting it to the technical committee for final review background what is sideload sideload or key manager sideload refers to a direct hardware connection from the opentitan key manager module to hardware symmetric crypto components for supplying keys when sideload is employed keys are generated within the key manager and fed directly to the hardware crypto components for usage at no point during this process are the keys exposed to software thus compromised software can only abuse the keys but cannot permanently expose them design options currently there are a couple of proposals under consideration as part of this rfc we are requesting inputs for additional proposals key slot implementation the implementation may provide a limited number of key slots in the key manager with support for sideloading keys into the hmac kmac and aes engines the number of key slots is left as an implementation detail which needs to take into consideration tradeoffs between use cases versus hardware cost additional considerations even though software does not have visibility into the key material it is still responsible for managing the key slot mapping to consuming applications the number of bits sideloaded to other ip may be constrained to meet a security level up to bits sideloading keys into otbn is not considered a security feature given that the target does not implement isolation of critical data from software the current recommendation is to avoid sideloading into otbn no sideload support given the limitations of the key slot proposal above and additional considerations we have put together a justification to remove support for sideloading from the architecture see the for more details update see for the latest thinking and proposals the ongoing also details the matching hardware updates for key manager only
0
7,030
3,074,501,641
IssuesEvent
2015-08-20 07:45:20
arangodb/arangodb
https://api.github.com/repos/arangodb/arangodb
closed
Building issues on Windows (V8, Python 3.x and others)
documentation
Quoting a recipe from the cookbook: > Either version 2.x or 3.x (excluding 3.4) should work - it's used to run V8s GYP. Make sure you add python.exe to your path environment variable; Restarting your running shell may be necessary. I'm not sure why 3.4 is excluded, but 3.x in general supposed to work. It seems that if 3.x is installed and added to PATH environment variable, V8 compilation fails because of a syntax error: import sys; print sys.byteorder `print` is a function in Python 3 and requires parentheses. Is there a way to make gyp use `ArangoDB\3rdParty\V8-4.3.61\third_party\python_26\python.exe` instead of a local Python installation?
1.0
Building issues on Windows (V8, Python 3.x and others) - Quoting a recipe from the cookbook: > Either version 2.x or 3.x (excluding 3.4) should work - it's used to run V8s GYP. Make sure you add python.exe to your path environment variable; Restarting your running shell may be necessary. I'm not sure why 3.4 is excluded, but 3.x in general supposed to work. It seems that if 3.x is installed and added to PATH environment variable, V8 compilation fails because of a syntax error: import sys; print sys.byteorder `print` is a function in Python 3 and requires parentheses. Is there a way to make gyp use `ArangoDB\3rdParty\V8-4.3.61\third_party\python_26\python.exe` instead of a local Python installation?
non_infrastructure
building issues on windows python x and others quoting a recipe from the cookbook either version x or x excluding should work it s used to run gyp make sure you add python exe to your path environment variable restarting your running shell may be necessary i m not sure why is excluded but x in general supposed to work it seems that if x is installed and added to path environment variable compilation fails because of a syntax error import sys print sys byteorder print is a function in python and requires parentheses is there a way to make gyp use arangodb third party python python exe instead of a local python installation
0
338,931
30,330,312,333
IssuesEvent
2023-07-11 05:41:17
MaterializeInc/materialize
https://api.github.com/repos/MaterializeInc/materialize
closed
QA sign-off for pushdown
C-feature T-testing
### Feature request - [x] Enable in CI by default - [x] RQG runs - [x] Bring https://github.com/MaterializeInc/materialize/pull/19013 to bear - [ ] ...
1.0
QA sign-off for pushdown - ### Feature request - [x] Enable in CI by default - [x] RQG runs - [x] Bring https://github.com/MaterializeInc/materialize/pull/19013 to bear - [ ] ...
non_infrastructure
qa sign off for pushdown feature request enable in ci by default rqg runs bring to bear
0
5,462
3,930,245,775
IssuesEvent
2016-04-25 07:01:29
kolliSuman/issues
https://api.github.com/repos/kolliSuman/issues
closed
QA_N-Grams Smoothing_Prerequisites_p1
Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open
Defect Description : In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required Actual Result : In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0Bandwidth : 100Mbps Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/natural-language-processing-iiith/blob/master/test-cases/integration_test-cases/N-Grams%20Smoothing/N-Grams%20Smoothing_13_Prerequisites_p1.org
True
QA_N-Grams Smoothing_Prerequisites_p1 - Defect Description : In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required Actual Result : In the "N-Grams Smoothing" experiment, the minimum requirement to run the experiment is not displayed in the page Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0Bandwidth : 100Mbps Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Test Step Link: https://github.com/Virtual-Labs/natural-language-processing-iiith/blob/master/test-cases/integration_test-cases/N-Grams%20Smoothing/N-Grams%20Smoothing_13_Prerequisites_p1.org
non_infrastructure
qa n grams smoothing prerequisites defect description in the n grams smoothing experiment the minimum requirement to run the experiment is not displayed in the page instead a page or scrolling should appear providing information on minimum requirement to run this experiment information like bandwidth device resolution hardware configuration and software required actual result in the n grams smoothing experiment the minimum requirement to run the experiment is not displayed in the page environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link
0
27,086
21,154,695,725
IssuesEvent
2022-04-07 00:59:46
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Soil crop parameters not always imported correctly from APSIM 7.10
bug interface/infrastructure
When importing a .apsim file from APSIM 7.10 the soil crop parameters are not imported if their name starts with a lower case letter e.g. wheat.
1.0
Soil crop parameters not always imported correctly from APSIM 7.10 - When importing a .apsim file from APSIM 7.10 the soil crop parameters are not imported if their name starts with a lower case letter e.g. wheat.
infrastructure
soil crop parameters not always imported correctly from apsim when importing a apsim file from apsim the soil crop parameters are not imported if their name starts with a lower case letter e g wheat
1
18,559
13,055,867,428
IssuesEvent
2020-07-30 02:58:22
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
Use I3_DATA to store photonics spline tables on the buildbots (Trac #742)
Incomplete Migration Migrated from Trac infrastructure task
Migrated from https://code.icecube.wisc.edu/ticket/742 ```json { "status": "closed", "changetime": "2015-02-18T21:44:30", "description": "Some of the Millipede unit tests need the full spline tables to function. At the moment these are read from the canonical path I3_DATA, which for the moment only exists on cvmfs-enabled machines, and so they don't run on the buildbots, and so platform-dependent bugs aren't caught in time.\n\nSimply copying these into the normal test data would be needless bloat, since they're used outside of testing as well. /cvmfs/icecube.wisc.edu/data should be cloned into the buildbots, and I3_DATA set in their environment. If the entire data directory is too large, then /cvmfs/icecube.wisc.edu/data/photon-tables/splines would do in a pinch.", "reporter": "jvansanten", "cc": "", "resolution": "fixed", "_ts": "1424295870059276", "component": "infrastructure", "summary": "Use I3_DATA to store photonics spline tables on the buildbots", "priority": "normal", "keywords": "", "time": "2014-08-30T20:27:54", "milestone": "", "owner": "nega", "type": "task" } ```
1.0
Use I3_DATA to store photonics spline tables on the buildbots (Trac #742) - Migrated from https://code.icecube.wisc.edu/ticket/742 ```json { "status": "closed", "changetime": "2015-02-18T21:44:30", "description": "Some of the Millipede unit tests need the full spline tables to function. At the moment these are read from the canonical path I3_DATA, which for the moment only exists on cvmfs-enabled machines, and so they don't run on the buildbots, and so platform-dependent bugs aren't caught in time.\n\nSimply copying these into the normal test data would be needless bloat, since they're used outside of testing as well. /cvmfs/icecube.wisc.edu/data should be cloned into the buildbots, and I3_DATA set in their environment. If the entire data directory is too large, then /cvmfs/icecube.wisc.edu/data/photon-tables/splines would do in a pinch.", "reporter": "jvansanten", "cc": "", "resolution": "fixed", "_ts": "1424295870059276", "component": "infrastructure", "summary": "Use I3_DATA to store photonics spline tables on the buildbots", "priority": "normal", "keywords": "", "time": "2014-08-30T20:27:54", "milestone": "", "owner": "nega", "type": "task" } ```
infrastructure
use data to store photonics spline tables on the buildbots trac migrated from json status closed changetime description some of the millipede unit tests need the full spline tables to function at the moment these are read from the canonical path data which for the moment only exists on cvmfs enabled machines and so they don t run on the buildbots and so platform dependent bugs aren t caught in time n nsimply copying these into the normal test data would be needless bloat since they re used outside of testing as well cvmfs icecube wisc edu data should be cloned into the buildbots and data set in their environment if the entire data directory is too large then cvmfs icecube wisc edu data photon tables splines would do in a pinch reporter jvansanten cc resolution fixed ts component infrastructure summary use data to store photonics spline tables on the buildbots priority normal keywords time milestone owner nega type task
1
56,999
13,962,905,732
IssuesEvent
2020-10-25 11:49:16
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
CUDA problem with Linux Kernel 5.9
module: build module: cuda triaged
## 🐛 Bug <!-- A clear and concise description of what the bug is. --> torch.cuda.is_available() returns True on Ubuntu 20.04 with linux kernel 5.8.13, but returns False when the kernel is upgraded to 5.9-rc8. nvidia-smi and other gpu-related programs work as expected, only pytorch stops detecting the GPU. ## To Reproduce In a Ubuntu 20.04 OS with linux kernel 5.9-rc8, run: ```python -c 'import torch; print(torch.cuda.is_available())'``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> Pytorch should detect the GPU on Linux kernel 5.9. ## Environment - PyTorch Version (e.g., 1.0): 1.6.0 - OS (e.g., Linux): Ubuntu 20.04.1 LTS (x86_64) - How you installed PyTorch (`conda`, `pip`, source): pip - Python version: 3.8 - CUDA version: 10.2 in pytorch, N/A in the system. - GPU models and configuration: GeForce RTX 3090 cc @malfet @seemethere @walterddr @ngimel
1.0
CUDA problem with Linux Kernel 5.9 - ## 🐛 Bug <!-- A clear and concise description of what the bug is. --> torch.cuda.is_available() returns True on Ubuntu 20.04 with linux kernel 5.8.13, but returns False when the kernel is upgraded to 5.9-rc8. nvidia-smi and other gpu-related programs work as expected, only pytorch stops detecting the GPU. ## To Reproduce In a Ubuntu 20.04 OS with linux kernel 5.9-rc8, run: ```python -c 'import torch; print(torch.cuda.is_available())'``` <!-- If you have a code sample, error messages, stack traces, please provide it here as well --> ## Expected behavior <!-- A clear and concise description of what you expected to happen. --> Pytorch should detect the GPU on Linux kernel 5.9. ## Environment - PyTorch Version (e.g., 1.0): 1.6.0 - OS (e.g., Linux): Ubuntu 20.04.1 LTS (x86_64) - How you installed PyTorch (`conda`, `pip`, source): pip - Python version: 3.8 - CUDA version: 10.2 in pytorch, N/A in the system. - GPU models and configuration: GeForce RTX 3090 cc @malfet @seemethere @walterddr @ngimel
non_infrastructure
cuda problem with linux kernel 🐛 bug torch cuda is available returns true on ubuntu with linux kernel but returns false when the kernel is upgraded to nvidia smi and other gpu related programs work as expected only pytorch stops detecting the gpu to reproduce in a ubuntu os with linux kernel run python c import torch print torch cuda is available expected behavior pytorch should detect the gpu on linux kernel environment pytorch version e g os e g linux ubuntu lts how you installed pytorch conda pip source pip python version cuda version in pytorch n a in the system gpu models and configuration geforce rtx cc malfet seemethere walterddr ngimel
0
690,487
23,661,612,225
IssuesEvent
2022-08-26 16:06:59
TheYellowArchitect/doubledamnation
https://api.github.com/repos/TheYellowArchitect/doubledamnation
opened
Level Editor - Pause to disable grabbing/removing tiles
bug good first issue level editor low priority
tl;dr: Try clicking any settings, and not creating tiles in the background. Spoilers: You can't. Luckily, its extremely easy to fix, and also takes little time.
1.0
Level Editor - Pause to disable grabbing/removing tiles - tl;dr: Try clicking any settings, and not creating tiles in the background. Spoilers: You can't. Luckily, its extremely easy to fix, and also takes little time.
non_infrastructure
level editor pause to disable grabbing removing tiles tl dr try clicking any settings and not creating tiles in the background spoilers you can t luckily its extremely easy to fix and also takes little time
0
14,407
10,821,991,236
IssuesEvent
2019-11-08 20:03:29
aspnet/AspNetCore
https://api.github.com/repos/aspnet/AspNetCore
closed
restore command in root folder of Asp Net Core project fails
area-infrastructure
Hi,I have cloned aspnetcore project and followed the instruction here in GitHub step-by-step but when ran restore command in the root directory of project I got the following errors. C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error : Unable to loc ate the .NET Core SDK. Check that it is installed and that the version specified in global.json (if any) matches the in stalled version. C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error MSB4236: The SD K 'Microsoft.NET.Sdk' specified could not be found. Build failed. Please help me in order to solve this issue.
1.0
restore command in root folder of Asp Net Core project fails - Hi,I have cloned aspnetcore project and followed the instruction here in GitHub step-by-step but when ran restore command in the root directory of project I got the following errors. C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error : Unable to loc ate the .NET Core SDK. Check that it is installed and that the version specified in global.json (if any) matches the in stalled version. C:\Users\Nurlan\.nuget\packages\microsoft.dotnet.arcade.sdk\1.0.0-beta.19517.3\tools\Tools.proj : error MSB4236: The SD K 'Microsoft.NET.Sdk' specified could not be found. Build failed. Please help me in order to solve this issue.
infrastructure
restore command in root folder of asp net core project fails hi i have cloned aspnetcore project and followed the instruction here in github step by step but when ran restore command in the root directory of project i got the following errors c users nurlan nuget packages microsoft dotnet arcade sdk beta tools tools proj error unable to loc ate the net core sdk check that it is installed and that the version specified in global json if any matches the in stalled version c users nurlan nuget packages microsoft dotnet arcade sdk beta tools tools proj error the sd k microsoft net sdk specified could not be found build failed please help me in order to solve this issue
1
399,513
27,243,670,096
IssuesEvent
2023-02-21 23:03:39
SAP/fundamental
https://api.github.com/repos/SAP/fundamental
closed
Docs: Components should use the base class
Documentation Develop
**Describe the bug** A clear and concise description of what the bug is. In some cases, components use a class selector to look for a block class which eliminates the need to always put in a base class. ``` [class*="fd-identifier"] { } ``` While this works OK, it creates some issues when using Angular with modularized styles. For example, if I use `<span class="fd-identifier--xl"><span>{{noOfTries}}</span></span>` the styling looks fine, but I cannot easily override the vars in my component CSS. This will not work ... ``` .fd-identifier { --fd-identifier-background-color: var(--fd-color-accent-9); } ``` No changes need to be done in the CSS at this time, but the NJK templates and markup in the docs should be updated to always include the block class along with any modifier classes. Affects - [ ] icons - [ ] image - [ ] identifier - [ ] button
1.0
Docs: Components should use the base class - **Describe the bug** A clear and concise description of what the bug is. In some cases, components use a class selector to look for a block class which eliminates the need to always put in a base class. ``` [class*="fd-identifier"] { } ``` While this works OK, it creates some issues when using Angular with modularized styles. For example, if I use `<span class="fd-identifier--xl"><span>{{noOfTries}}</span></span>` the styling looks fine, but I cannot easily override the vars in my component CSS. This will not work ... ``` .fd-identifier { --fd-identifier-background-color: var(--fd-color-accent-9); } ``` No changes need to be done in the CSS at this time, but the NJK templates and markup in the docs should be updated to always include the block class along with any modifier classes. Affects - [ ] icons - [ ] image - [ ] identifier - [ ] button
non_infrastructure
docs components should use the base class describe the bug a clear and concise description of what the bug is in some cases components use a class selector to look for a block class which eliminates the need to always put in a base class while this works ok it creates some issues when using angular with modularized styles for example if i use nooftries the styling looks fine but i cannot easily override the vars in my component css this will not work fd identifier fd identifier background color var fd color accent no changes need to be done in the css at this time but the njk templates and markup in the docs should be updated to always include the block class along with any modifier classes affects icons image identifier button
0
28,250
23,111,869,342
IssuesEvent
2022-07-27 13:40:08
spring-projects/spring-batch
https://api.github.com/repos/spring-projects/spring-batch
opened
Migrate tests to JUnit Jupiter
in: test in: infrastructure type: feature in: integration in: samples in: core
This issue is to track the migration of the test suite from JUnit 4 to JUnit 5: - [ ] spring-batch-infrastructure: #4125 - [ ] spring-batch-core - [ ] spring-batch-test - [ ] spring-batch-integration: #4124 - [ ] spring-batch-samples
1.0
Migrate tests to JUnit Jupiter - This issue is to track the migration of the test suite from JUnit 4 to JUnit 5: - [ ] spring-batch-infrastructure: #4125 - [ ] spring-batch-core - [ ] spring-batch-test - [ ] spring-batch-integration: #4124 - [ ] spring-batch-samples
infrastructure
migrate tests to junit jupiter this issue is to track the migration of the test suite from junit to junit spring batch infrastructure spring batch core spring batch test spring batch integration spring batch samples
1
18,070
12,748,923,845
IssuesEvent
2020-06-26 21:16:54
commitdev/zero
https://api.github.com/repos/commitdev/zero
opened
Investigate Ory tools for reusable auth
backend infrastructure
Do another round of investigation into Ory Hydra / Oathkeeper / Kratos for user management and auth. See if they have the features and integrations we want. Last time Bill looked there were some things lacking. Should be: - Easy to integrate with nginx ingress controller - Abstract enough away so that it handles login, then provides auth via header or cookie for backend service to use (probably JWT) - Should be easy to put a customizeable UI on top
1.0
Investigate Ory tools for reusable auth - Do another round of investigation into Ory Hydra / Oathkeeper / Kratos for user management and auth. See if they have the features and integrations we want. Last time Bill looked there were some things lacking. Should be: - Easy to integrate with nginx ingress controller - Abstract enough away so that it handles login, then provides auth via header or cookie for backend service to use (probably JWT) - Should be easy to put a customizeable UI on top
infrastructure
investigate ory tools for reusable auth do another round of investigation into ory hydra oathkeeper kratos for user management and auth see if they have the features and integrations we want last time bill looked there were some things lacking should be easy to integrate with nginx ingress controller abstract enough away so that it handles login then provides auth via header or cookie for backend service to use probably jwt should be easy to put a customizeable ui on top
1
432,057
12,488,647,072
IssuesEvent
2020-05-31 15:11:59
buttercup/buttercup-mobile
https://api.github.com/repos/buttercup/buttercup-mobile
closed
Improve label on button to hide/display info
Effort: Low Priority: Low Status: Available Type: Enhancement
When seeing a entry on a Archive, all sensitive data is blurred. To display, one must click the __view hidden__ button, which is a weird label for it. I suggest to change to something along the lines of: _Display sensitive information_ _Hide sensitive information_ References: ![no-info](https://user-images.githubusercontent.com/861658/56237534-98e95200-6062-11e9-8500-b88449f00e7a.jpeg) ![with-info](https://user-images.githubusercontent.com/861658/56237536-9a1a7f00-6062-11e9-80c0-02e7867eb31e.jpeg)
1.0
Improve label on button to hide/display info - When seeing a entry on a Archive, all sensitive data is blurred. To display, one must click the __view hidden__ button, which is a weird label for it. I suggest to change to something along the lines of: _Display sensitive information_ _Hide sensitive information_ References: ![no-info](https://user-images.githubusercontent.com/861658/56237534-98e95200-6062-11e9-8500-b88449f00e7a.jpeg) ![with-info](https://user-images.githubusercontent.com/861658/56237536-9a1a7f00-6062-11e9-80c0-02e7867eb31e.jpeg)
non_infrastructure
improve label on button to hide display info when seeing a entry on a archive all sensitive data is blurred to display one must click the view hidden button which is a weird label for it i suggest to change to something along the lines of display sensitive information hide sensitive information references
0