Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
665
labels
stringlengths
4
554
body
stringlengths
3
235k
index
stringclasses
6 values
text_combine
stringlengths
96
235k
label
stringclasses
2 values
text
stringlengths
96
196k
binary_label
int64
0
1
325,182
9,920,482,432
IssuesEvent
2019-06-30 09:44:45
gicentre/litvis
https://api.github.com/repos/gicentre/litvis
reopened
Some function types generate parsing warning in markdown mode
bug priority/p2
For the record, we still need to implement markdown output for a wider range of types. Currently some types can generate parsing warnings. For example ```` ```elm {m} myFunction : List ( Int, Int ) myFunction = [ ( 1, 2 ) ] ``` ```` generates the warning : `Could not parse "[(1,2)]" (litvis:expression-value)` highlighting the entire code block. We should also test against those examples listed in #13 and #14. If this isn't fixable in the short term, we should probably generate a more beginner friendly and correct warning (the problem isn't with the contents of the function, but the output type of the code block) such as `Markdown output not supported for this type. Could you use 'raw' (or 'r') output instead?`
1.0
Some function types generate parsing warning in markdown mode - For the record, we still need to implement markdown output for a wider range of types. Currently some types can generate parsing warnings. For example ```` ```elm {m} myFunction : List ( Int, Int ) myFunction = [ ( 1, 2 ) ] ``` ```` generates the warning : `Could not parse "[(1,2)]" (litvis:expression-value)` highlighting the entire code block. We should also test against those examples listed in #13 and #14. If this isn't fixable in the short term, we should probably generate a more beginner friendly and correct warning (the problem isn't with the contents of the function, but the output type of the code block) such as `Markdown output not supported for this type. Could you use 'raw' (or 'r') output instead?`
non_infrastructure
some function types generate parsing warning in markdown mode for the record we still need to implement markdown output for a wider range of types currently some types can generate parsing warnings for example elm m myfunction list int int myfunction generates the warning could not parse litvis expression value highlighting the entire code block we should also test against those examples listed in and if this isn t fixable in the short term we should probably generate a more beginner friendly and correct warning the problem isn t with the contents of the function but the output type of the code block such as markdown output not supported for this type could you use raw or r output instead
0
21,984
14,948,627,925
IssuesEvent
2021-01-26 10:20:41
RasaHQ/rasa
https://api.github.com/repos/RasaHQ/rasa
opened
Add integration test for Rasa Chat Channel (Rasa X)
area:rasa-oss :ferris_wheel: area:rasa-oss/channels 📞 area:rasa-oss/infrastructure :bullettrain_front: type:maintenance :wrench:
**Description of Problem**: We should add integration tests for the `Rasa Chat` channel which test the entire roundtrip of sending a message, processing it via Rasa Open Source and receiving the bot's response. **Overview of the Solution**: TBD **Blockers** * This task is dependent on https://github.com/RasaHQ/rasa/issues/7804. * we need some Rasa X deployment to test against which could make this test a little bit different than the other integration tests **Definition of Done**: - [ ] Tests are added
1.0
Add integration test for Rasa Chat Channel (Rasa X) - **Description of Problem**: We should add integration tests for the `Rasa Chat` channel which test the entire roundtrip of sending a message, processing it via Rasa Open Source and receiving the bot's response. **Overview of the Solution**: TBD **Blockers** * This task is dependent on https://github.com/RasaHQ/rasa/issues/7804. * we need some Rasa X deployment to test against which could make this test a little bit different than the other integration tests **Definition of Done**: - [ ] Tests are added
infrastructure
add integration test for rasa chat channel rasa x description of problem we should add integration tests for the rasa chat channel which test the entire roundtrip of sending a message processing it via rasa open source and receiving the bot s response overview of the solution tbd blockers this task is dependent on we need some rasa x deployment to test against which could make this test a little bit different than the other integration tests definition of done tests are added
1
5,653
5,849,344,770
IssuesEvent
2017-05-10 23:33:16
dotnet/corert
https://api.github.com/repos/dotnet/corert
opened
Remove un-needed target when upgrading to v2 BuildTools
Infrastructure
We have done some change in the PR #3580 to change the license Url for packaging. because corert repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct
1.0
Remove un-needed target when upgrading to v2 BuildTools - We have done some change in the PR #3580 to change the license Url for packaging. because corert repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct
infrastructure
remove un needed target when upgrading to buildtools we have done some change in the pr to change the license url for packaging because corert repo not using of the buildtools we had to hack this by adding the temporary target overridelicenseurl in the dir targets files this target should be removed after upgrading to of the buildtools and need to ensure when removing it the license url in the packages nuspec files is correct
1
27,679
22,148,616,896
IssuesEvent
2022-06-03 14:31:23
abacus-network/abacus-monorepo
https://api.github.com/repos/abacus-network/abacus-monorepo
closed
Update infra to specify the InterchainGasPaymaster in Relayer settings
mainnet infrastructure
* Blocked by: * https://github.com/abacus-network/abacus-monorepo/issues/478 * Following https://github.com/abacus-network/abacus-monorepo/issues/478, we should update `infra` too ensure that relayers are properly configured with the InterchainGasPaymaster address
1.0
Update infra to specify the InterchainGasPaymaster in Relayer settings - * Blocked by: * https://github.com/abacus-network/abacus-monorepo/issues/478 * Following https://github.com/abacus-network/abacus-monorepo/issues/478, we should update `infra` too ensure that relayers are properly configured with the InterchainGasPaymaster address
infrastructure
update infra to specify the interchaingaspaymaster in relayer settings blocked by following we should update infra too ensure that relayers are properly configured with the interchaingaspaymaster address
1
159,270
20,048,346,004
IssuesEvent
2022-02-03 01:07:34
kapseliboi/RedGL2
https://api.github.com/repos/kapseliboi/RedGL2
opened
CVE-2021-32804 (High) detected in tar-4.4.8.tgz
security vulnerability
## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.8.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p> <p> Dependency Hierarchy: - gulp-4.0.2.tgz (Root Library) - glob-watcher-5.0.3.tgz - chokidar-2.1.6.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - :x: **tar-4.4.8.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-32804 (High) detected in tar-4.4.8.tgz - ## CVE-2021-32804 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.8.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p> <p> Dependency Hierarchy: - gulp-4.0.2.tgz (Root Library) - glob-watcher-5.0.3.tgz - chokidar-2.1.6.tgz - fsevents-1.2.9.tgz - node-pre-gyp-0.12.0.tgz - :x: **tar-4.4.8.tgz** (Vulnerable Library) <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar. <p>Publish Date: 2021-08-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p> <p>Release Date: 2021-08-03</p> <p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href dependency hierarchy gulp tgz root library glob watcher tgz chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in base branch dev vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
0
717,938
24,697,365,653
IssuesEvent
2022-10-19 13:05:46
ant-media/Ant-Media-Server
https://api.github.com/repos/ant-media/Ant-Media-Server
closed
iOS SDK Screen capture is pixelating when screen changes
Priority: Urgent
Being [discussed here](https://github.com/ant-media/Ant-Media-Server/issues/4278) ### Steps to reproduce 1. Run iOS Sample Project and Choose to Publish the screen 2. Play the stream on Ant Media Server with WebRTC 3. Swipe left-right in the screen ### Expected behavior There should be smooth video play ### Actual behavior Pixelating the video. Check the recorded video here https://user-images.githubusercontent.com/3456251/187892806-19fbfea0-d7fa-4898-9bd2-9f8ab8c19f05.mp4 Join [community.antmedia.io](https://community.antmedia.io) and be a part of Ant Media Server Community.
1.0
iOS SDK Screen capture is pixelating when screen changes - Being [discussed here](https://github.com/ant-media/Ant-Media-Server/issues/4278) ### Steps to reproduce 1. Run iOS Sample Project and Choose to Publish the screen 2. Play the stream on Ant Media Server with WebRTC 3. Swipe left-right in the screen ### Expected behavior There should be smooth video play ### Actual behavior Pixelating the video. Check the recorded video here https://user-images.githubusercontent.com/3456251/187892806-19fbfea0-d7fa-4898-9bd2-9f8ab8c19f05.mp4 Join [community.antmedia.io](https://community.antmedia.io) and be a part of Ant Media Server Community.
non_infrastructure
ios sdk screen capture is pixelating when screen changes being steps to reproduce run ios sample project and choose to publish the screen play the stream on ant media server with webrtc swipe left right in the screen expected behavior there should be smooth video play actual behavior pixelating the video check the recorded video here join and be a part of ant media server community
0
666,686
22,363,652,204
IssuesEvent
2022-06-15 23:59:07
AlbertaSat/ex2_ground_station_website
https://api.github.com/repos/AlbertaSat/ex2_ground_station_website
closed
Verify accuracy of Pyorbital against GPredict
enhancement high priority
### Brief Feature Description Using the same TLE data (eg. ISS), compare calculated passover times between what Pyorbital calculates and what [GPredict](http://gpredict.oz9aec.net/) calculates. ### Additional Details If these discrepencies are consistent by an offset, compensate for it in the calculation in automation.py. Consult with Mission Design about acceptable tolerances. If the discrepencies are too high, find a way to use GPredict's data instead over Pyorbital as that is our way of determining satellite location (among other things). ### Priority High
1.0
Verify accuracy of Pyorbital against GPredict - ### Brief Feature Description Using the same TLE data (eg. ISS), compare calculated passover times between what Pyorbital calculates and what [GPredict](http://gpredict.oz9aec.net/) calculates. ### Additional Details If these discrepencies are consistent by an offset, compensate for it in the calculation in automation.py. Consult with Mission Design about acceptable tolerances. If the discrepencies are too high, find a way to use GPredict's data instead over Pyorbital as that is our way of determining satellite location (among other things). ### Priority High
non_infrastructure
verify accuracy of pyorbital against gpredict brief feature description using the same tle data eg iss compare calculated passover times between what pyorbital calculates and what calculates additional details if these discrepencies are consistent by an offset compensate for it in the calculation in automation py consult with mission design about acceptable tolerances if the discrepencies are too high find a way to use gpredict s data instead over pyorbital as that is our way of determining satellite location among other things priority high
0
4,109
4,807,080,688
IssuesEvent
2016-11-02 20:24:18
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
closed
Move bootstrap build back to MSBuild 15.0
Area-Infrastructure
Our bootstrap build temporarily needs to go back to MSBuild 14.0. There is a bug in MSBuild 15.0 which prevents us from fully loading our bootstrap components and hence invalidates our build. https://github.com/Microsoft/msbuild/issues/1183 This bug tracks the removal of work arounds on our end.
1.0
Move bootstrap build back to MSBuild 15.0 - Our bootstrap build temporarily needs to go back to MSBuild 14.0. There is a bug in MSBuild 15.0 which prevents us from fully loading our bootstrap components and hence invalidates our build. https://github.com/Microsoft/msbuild/issues/1183 This bug tracks the removal of work arounds on our end.
infrastructure
move bootstrap build back to msbuild our bootstrap build temporarily needs to go back to msbuild there is a bug in msbuild which prevents us from fully loading our bootstrap components and hence invalidates our build this bug tracks the removal of work arounds on our end
1
2,008
3,461,657,951
IssuesEvent
2015-12-20 08:30:28
t3kt/vjzual2
https://api.github.com/repos/t3kt/vjzual2
closed
get rid of unnecessary module extension classes
cleanup infrastructure
unless it does something other than initialize parameters, remove it. simplify things and whatnot.
1.0
get rid of unnecessary module extension classes - unless it does something other than initialize parameters, remove it. simplify things and whatnot.
infrastructure
get rid of unnecessary module extension classes unless it does something other than initialize parameters remove it simplify things and whatnot
1
5,030
2,760,108,114
IssuesEvent
2015-04-28 09:56:12
MatchmakerExchange/mme-apis
https://api.github.com/repos/MatchmakerExchange/mme-apis
closed
Test data does not conform to v1.0 specification
bug PR submitted Testing
Sigh. Apologies. 1) variant 'referenceName' includes 'chr', which it shouldn't 2) the 'institution' field of contact is misnamed 'organization'
1.0
Test data does not conform to v1.0 specification - Sigh. Apologies. 1) variant 'referenceName' includes 'chr', which it shouldn't 2) the 'institution' field of contact is misnamed 'organization'
non_infrastructure
test data does not conform to specification sigh apologies variant referencename includes chr which it shouldn t the institution field of contact is misnamed organization
0
18,565
25,828,633,405
IssuesEvent
2022-12-12 14:39:41
ldtteam/Structurize
https://api.github.com/repos/ldtteam/Structurize
closed
Rubidium visual glitch
Compatibility: Mod
### Is there an existing compatibility issue for this mod? - [X] I have searched the existing issues. ### Are you using the latest Structurize and your mod Version? - [X] I am running the latest **alpha** version of Structurize for my Minecraft version. I am also running the latest versions of other mods that are part of my problem. ### Did you ask on Discord? - [X] I checked the MineColonies/Structurize [Wiki](https://wiki.minecolonies.ldtteam.com/) or I asked on [discord](https://discord.minecolonies.com/). ### Minecraft Version 1.19 ### Structurize Version 1.19.2-1.0.468-ALPHA ### Your Mod Name and Version rubidium-0.6.2 ### Related Mods and their Versions forge: 43.1.23 minecolonies: 1.19.2-1.0.1137-ALPHA domum_ornamentum: 1.19-1.0.64-ALPHA-universal ### Current Behavior ![2022-12-11_19 04 24](https://user-images.githubusercontent.com/120334856/206924043-19621124-4caa-4299-ae8f-64f5c1d692ae.png) ![2022-12-11_19 26 24](https://user-images.githubusercontent.com/120334856/206924046-5c747c4e-7ecb-4f91-8ed5-eef63913c784.png) when previewing a building or a shape a visual glitch happens ### Expected Behavior the visual glitch doesnt happen ### Reproduction Steps 1.get build or shape tool 2.right click and try to preview a building or a shape 3.a visual glitch happens ### Logs https://pastebin.com/d8hG3jkx ### Anything else? _No response_ ### Footer --- ## Viewers * Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments. * Add a comment if you have any insights or background information that isn't already part of the conversation.
True
Rubidium visual glitch - ### Is there an existing compatibility issue for this mod? - [X] I have searched the existing issues. ### Are you using the latest Structurize and your mod Version? - [X] I am running the latest **alpha** version of Structurize for my Minecraft version. I am also running the latest versions of other mods that are part of my problem. ### Did you ask on Discord? - [X] I checked the MineColonies/Structurize [Wiki](https://wiki.minecolonies.ldtteam.com/) or I asked on [discord](https://discord.minecolonies.com/). ### Minecraft Version 1.19 ### Structurize Version 1.19.2-1.0.468-ALPHA ### Your Mod Name and Version rubidium-0.6.2 ### Related Mods and their Versions forge: 43.1.23 minecolonies: 1.19.2-1.0.1137-ALPHA domum_ornamentum: 1.19-1.0.64-ALPHA-universal ### Current Behavior ![2022-12-11_19 04 24](https://user-images.githubusercontent.com/120334856/206924043-19621124-4caa-4299-ae8f-64f5c1d692ae.png) ![2022-12-11_19 26 24](https://user-images.githubusercontent.com/120334856/206924046-5c747c4e-7ecb-4f91-8ed5-eef63913c784.png) when previewing a building or a shape a visual glitch happens ### Expected Behavior the visual glitch doesnt happen ### Reproduction Steps 1.get build or shape tool 2.right click and try to preview a building or a shape 3.a visual glitch happens ### Logs https://pastebin.com/d8hG3jkx ### Anything else? _No response_ ### Footer --- ## Viewers * Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments. * Add a comment if you have any insights or background information that isn't already part of the conversation.
non_infrastructure
rubidium visual glitch is there an existing compatibility issue for this mod i have searched the existing issues are you using the latest structurize and your mod version i am running the latest alpha version of structurize for my minecraft version i am also running the latest versions of other mods that are part of my problem did you ask on discord i checked the minecolonies structurize or i asked on minecraft version structurize version alpha your mod name and version rubidium related mods and their versions forge minecolonies alpha domum ornamentum alpha universal current behavior when previewing a building or a shape a visual glitch happens expected behavior the visual glitch doesnt happen reproduction steps get build or shape tool right click and try to preview a building or a shape a visual glitch happens logs anything else no response footer viewers add a thumbs up to the bug report if you are also affected this helps the bug report become more visible to the team and doesn t clutter the comments add a comment if you have any insights or background information that isn t already part of the conversation
0
658
2,824,189,557
IssuesEvent
2015-05-21 13:31:47
dotnet/wcf
https://api.github.com/repos/dotnet/wcf
opened
Provide a mechanism to pass command line parameters into tests
Infrastructure
Our scenario tests will require the ability to redirect the URL of the test service to alternate machines. One can also imagine other test inputs that might feed some [Theory] based test. One technique to investigate is a pre-build step that accepts MSBuild properties of a known format, generates a temporary .cs file with the information (a Dictionary?), and adds that file to the @(Compile) collection. The tests could be written to consult the dictionary for expected names. Also we could create a [Theory] style test that pulls multiple values. The implementation should not be tied only to scenario tests but available to all tests. This would be more generally useful than just for WCF, so we should consider working with https://github.com/dotnet/buildtools
1.0
Provide a mechanism to pass command line parameters into tests - Our scenario tests will require the ability to redirect the URL of the test service to alternate machines. One can also imagine other test inputs that might feed some [Theory] based test. One technique to investigate is a pre-build step that accepts MSBuild properties of a known format, generates a temporary .cs file with the information (a Dictionary?), and adds that file to the @(Compile) collection. The tests could be written to consult the dictionary for expected names. Also we could create a [Theory] style test that pulls multiple values. The implementation should not be tied only to scenario tests but available to all tests. This would be more generally useful than just for WCF, so we should consider working with https://github.com/dotnet/buildtools
infrastructure
provide a mechanism to pass command line parameters into tests our scenario tests will require the ability to redirect the url of the test service to alternate machines one can also imagine other test inputs that might feed some based test one technique to investigate is a pre build step that accepts msbuild properties of a known format generates a temporary cs file with the information a dictionary and adds that file to the compile collection the tests could be written to consult the dictionary for expected names also we could create a style test that pulls multiple values the implementation should not be tied only to scenario tests but available to all tests this would be more generally useful than just for wcf so we should consider working with
1
330,124
28,349,894,256
IssuesEvent
2023-04-12 01:18:25
aodn/nrmn-application
https://api.github.com/repos/aodn/nrmn-application
closed
Ordering of sites slows the loading of the List Site page
ready to test planned systest
Noticed a time lag to load the List Site page since the ordering of the site list has been implemented.
2.0
Ordering of sites slows the loading of the List Site page - Noticed a time lag to load the List Site page since the ordering of the site list has been implemented.
non_infrastructure
ordering of sites slows the loading of the list site page noticed a time lag to load the list site page since the ordering of the site list has been implemented
0
99,821
4,065,750,427
IssuesEvent
2016-05-26 12:37:35
vikilabs-code/hombot-issues
https://api.github.com/repos/vikilabs-code/hombot-issues
closed
No themes available theme
high-priority IOS
When you delete all the themes and then you go back and select one touch themes again you get an error no themes available. @artakshar will help you recreate the issue.
1.0
No themes available theme - When you delete all the themes and then you go back and select one touch themes again you get an error no themes available. @artakshar will help you recreate the issue.
non_infrastructure
no themes available theme when you delete all the themes and then you go back and select one touch themes again you get an error no themes available artakshar will help you recreate the issue
0
15,004
11,297,258,273
IssuesEvent
2020-01-17 05:16:28
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
closed
Docs: Changelog fixup for 3.15.0
chore infrastructure
## overview This ticket is a placeholder reminder to fix the comparison diff for the 3.14.1 hotfix; It should be compared against 3.14.0 _not_ 3.13.2. We should also add calibrate to bottom as a current limitation in api v2 beta.
1.0
Docs: Changelog fixup for 3.15.0 - ## overview This ticket is a placeholder reminder to fix the comparison diff for the 3.14.1 hotfix; It should be compared against 3.14.0 _not_ 3.13.2. We should also add calibrate to bottom as a current limitation in api v2 beta.
infrastructure
docs changelog fixup for overview this ticket is a placeholder reminder to fix the comparison diff for the hotfix it should be compared against not we should also add calibrate to bottom as a current limitation in api beta
1
34,494
6,339,686,789
IssuesEvent
2017-07-27 08:58:54
lumapps/lumX
https://api.github.com/repos/lumapps/lumX
opened
Fix AJAX examples
documentation
omdbapi requires a key now so we could try to use something else in our examples?
1.0
Fix AJAX examples - omdbapi requires a key now so we could try to use something else in our examples?
non_infrastructure
fix ajax examples omdbapi requires a key now so we could try to use something else in our examples
0
16,941
12,152,151,701
IssuesEvent
2020-04-24 21:31:38
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
closed
Enable unsafe sysctl (to allow per pod disable of IPv6)
Infrastructure closed
https://trello.com/c/o4U3LtW9/93-enable-unsafe-sysctl-to-allow-per-pod-disable-of-ipv6 Looking to allow pods to disable IPv6. see the following: https://access.redhat.com/solutions/3340721 We have requests to be able to disable IPv6 for certain projects/namespaces.
1.0
Enable unsafe sysctl (to allow per pod disable of IPv6) - https://trello.com/c/o4U3LtW9/93-enable-unsafe-sysctl-to-allow-per-pod-disable-of-ipv6 Looking to allow pods to disable IPv6. see the following: https://access.redhat.com/solutions/3340721 We have requests to be able to disable IPv6 for certain projects/namespaces.
infrastructure
enable unsafe sysctl to allow per pod disable of looking to allow pods to disable see the following we have requests to be able to disable for certain projects namespaces
1
11,240
9,014,241,991
IssuesEvent
2019-02-05 21:48:17
webhintio/hint
https://api.github.com/repos/webhintio/hint
closed
Update @types/sinon from 5.0.7 to 7.0.5 [0.5]
area:infrastructure type:new-feature type:testing
# 🚀 Feature request ## Description PR #1770 is failing. Need to find out why and fix it.
1.0
Update @types/sinon from 5.0.7 to 7.0.5 [0.5] - # 🚀 Feature request ## Description PR #1770 is failing. Need to find out why and fix it.
infrastructure
update types sinon from to 🚀 feature request description pr is failing need to find out why and fix it
1
16,879
12,152,145,520
IssuesEvent
2020-04-24 21:30:47
BCDevOps/developer-experience
https://api.github.com/repos/BCDevOps/developer-experience
closed
WARN: Low Metrics Cassandra Space - 15% free
Infrastructure closed medium priority
https://trello.com/c/T5f2yg80/164-warn-low-metrics-cassandra-space-15-free Cassandra is slowly filling up as the platform grows and needs to have its storage expanded.
1.0
WARN: Low Metrics Cassandra Space - 15% free - https://trello.com/c/T5f2yg80/164-warn-low-metrics-cassandra-space-15-free Cassandra is slowly filling up as the platform grows and needs to have its storage expanded.
infrastructure
warn low metrics cassandra space free cassandra is slowly filling up as the platform grows and needs to have its storage expanded
1
334,118
10,136,151,191
IssuesEvent
2019-08-02 12:14:07
CheckiO/EoC-battle
https://api.github.com/repos/CheckiO/EoC-battle
closed
Throw async errors from REF to EXEC
complex:middle priority:normal section:referee
Not every error we can find on Client Side so we should make a say to send an error from referee to Client so the user can see the error in console during the battle review.
1.0
Throw async errors from REF to EXEC - Not every error we can find on Client Side so we should make a say to send an error from referee to Client so the user can see the error in console during the battle review.
non_infrastructure
throw async errors from ref to exec not every error we can find on client side so we should make a say to send an error from referee to client so the user can see the error in console during the battle review
0
144,768
13,125,346,421
IssuesEvent
2020-08-06 06:28:12
RedHatInsights/insights-results-aggregator-mock
https://api.github.com/repos/RedHatInsights/insights-results-aggregator-mock
closed
Update pull request template to be compatible with other repositories in a project
documentation
Update pull request template to be compatible with other repositories in a project
1.0
Update pull request template to be compatible with other repositories in a project - Update pull request template to be compatible with other repositories in a project
non_infrastructure
update pull request template to be compatible with other repositories in a project update pull request template to be compatible with other repositories in a project
0
15,642
11,631,944,535
IssuesEvent
2020-02-28 03:12:02
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Need some extra math methods for a weather data project
interface/infrastructure newfeature
In particular I need AverageMonthlyAverages and YearlyAverages
1.0
Need some extra math methods for a weather data project - In particular I need AverageMonthlyAverages and YearlyAverages
infrastructure
need some extra math methods for a weather data project in particular i need averagemonthlyaverages and yearlyaverages
1
22,828
15,531,502,127
IssuesEvent
2021-03-14 00:01:07
algorand/go-algorand
https://api.github.com/repos/algorand/go-algorand
closed
TEAL - arguments counts and traps
Infrastructure TEAL 3
Application contracts can check the number of arguments they have been provided via `NumAppArgs`, but normal arguments do not have a similar field or opcode to look at. For application contracts, there's no discussion in the docs about what happens if you attempt to read an argument after this limit. I expect it should trap. For non-application contracts, I'd expect the same behavior, but again, this is not discussed. Finally, application contracts in the same group can read each other's arguments via `gtxna` but non-application contracts cannot read each other's arguments (nor can application contracts read the arguments of non-application contracts in the same group.)
1.0
TEAL - arguments counts and traps - Application contracts can check the number of arguments they have been provided via `NumAppArgs`, but normal arguments do not have a similar field or opcode to look at. For application contracts, there's no discussion in the docs about what happens if you attempt to read an argument after this limit. I expect it should trap. For non-application contracts, I'd expect the same behavior, but again, this is not discussed. Finally, application contracts in the same group can read each other's arguments via `gtxna` but non-application contracts cannot read each other's arguments (nor can application contracts read the arguments of non-application contracts in the same group.)
infrastructure
teal arguments counts and traps application contracts can check the number of arguments they have been provided via numappargs but normal arguments do not have a similar field or opcode to look at for application contracts there s no discussion in the docs about what happens if you attempt to read an argument after this limit i expect it should trap for non application contracts i d expect the same behavior but again this is not discussed finally application contracts in the same group can read each other s arguments via gtxna but non application contracts cannot read each other s arguments nor can application contracts read the arguments of non application contracts in the same group
1
9,821
8,183,520,506
IssuesEvent
2018-08-29 09:18:07
coq/coq
https://api.github.com/repos/coq/coq
closed
Coq makefile: print-pretty-timed-diff mixes up "before" and "after"
kind: infrastructure
#### Version 8.8 #### Operating system NixOS #### Description of the problem The makefile template passes arguments into `make-both-time-files.py` in the opposite order they should be in. [From `CoqMakefile.in`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/CoqMakefile.in#L346): ```make COQMAKE_BOTH_TIME_FILES ?= "$(COQLIB)/tools/make-both-time-files.py" print-pretty-timed-diff:: $(HIDE)$(COQMAKE_BOTH_TIME_FILES) --sort-by=$(TIMING_SORT_BY) $(TIME_OF_BUILD_BEFORE_FILE) $(TIME_OF_BUILD_AFTER_FILE) $(TIME_OF_PRETTY_BOTH_BUILD_FILE) $(TIME_OF_PRETTY_BUILD_EXTRA_FILES) ``` but [from `make-both-time-files.py`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/make-both-time-files.py#L6): ```python USAGE = 'Usage: %s [--sort-by=auto|absolute|diff] AFTER_FILE_NAME BEFORE_FILE_NAME [OUTPUT_FILE_NAME ..]' % sys.argv[0] ``` cc: @JasonGross
1.0
Coq makefile: print-pretty-timed-diff mixes up "before" and "after" - #### Version 8.8 #### Operating system NixOS #### Description of the problem The makefile template passes arguments into `make-both-time-files.py` in the opposite order they should be in. [From `CoqMakefile.in`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/CoqMakefile.in#L346): ```make COQMAKE_BOTH_TIME_FILES ?= "$(COQLIB)/tools/make-both-time-files.py" print-pretty-timed-diff:: $(HIDE)$(COQMAKE_BOTH_TIME_FILES) --sort-by=$(TIMING_SORT_BY) $(TIME_OF_BUILD_BEFORE_FILE) $(TIME_OF_BUILD_AFTER_FILE) $(TIME_OF_PRETTY_BOTH_BUILD_FILE) $(TIME_OF_PRETTY_BUILD_EXTRA_FILES) ``` but [from `make-both-time-files.py`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/make-both-time-files.py#L6): ```python USAGE = 'Usage: %s [--sort-by=auto|absolute|diff] AFTER_FILE_NAME BEFORE_FILE_NAME [OUTPUT_FILE_NAME ..]' % sys.argv[0] ``` cc: @JasonGross
infrastructure
coq makefile print pretty timed diff mixes up before and after version operating system nixos description of the problem the makefile template passes arguments into make both time files py in the opposite order they should be in make coqmake both time files coqlib tools make both time files py print pretty timed diff hide coqmake both time files sort by timing sort by time of build before file time of build after file time of pretty both build file time of pretty build extra files but python usage usage s after file name before file name sys argv cc jasongross
1
5
2,490,813,283
IssuesEvent
2015-01-02 20:13:41
Jasig/cas
https://api.github.com/repos/Jasig/cas
closed
CAS/Environment info in the logs
3.5.0 4.0 Configuration Infrastructure Task Trivial
When troubleshooting CAS and reviewing logs, it useful to know the following to better understand the cause of issue (Figuring out this data sometimes is difficult for the novice adopter!) - CAS server version - Java version, vendor, vm info - OS version, vendor, etc - Possibly the container info via the ServletContext These would show up under DEBUG/TRACE levels on startup. It would also be cool if CAS could output its own banner like spring boot :) https://github.com/spring-projects/spring-boot/blob/master/spring-boot/src/main/java/org/springframework/boot/SpringBootBanner.java
1.0
CAS/Environment info in the logs - When troubleshooting CAS and reviewing logs, it useful to know the following to better understand the cause of issue (Figuring out this data sometimes is difficult for the novice adopter!) - CAS server version - Java version, vendor, vm info - OS version, vendor, etc - Possibly the container info via the ServletContext These would show up under DEBUG/TRACE levels on startup. It would also be cool if CAS could output its own banner like spring boot :) https://github.com/spring-projects/spring-boot/blob/master/spring-boot/src/main/java/org/springframework/boot/SpringBootBanner.java
infrastructure
cas environment info in the logs when troubleshooting cas and reviewing logs it useful to know the following to better understand the cause of issue figuring out this data sometimes is difficult for the novice adopter cas server version java version vendor vm info os version vendor etc possibly the container info via the servletcontext these would show up under debug trace levels on startup it would also be cool if cas could output its own banner like spring boot
1
348,973
10,455,289,359
IssuesEvent
2019-09-19 20:56:28
kubeapps/kubeapps
https://api.github.com/repos/kubeapps/kubeapps
closed
Kubeapps shows "default" namespace if user cannot list namespaces
component/dashboard good first issue kind/bug priority/low size/S
If a service account is created for a user in a specific namespace (and without permission to list namespaces), and the associated credential is used to login, the namespace switcher shows "default" rather than the specific namespace. Kubeapps should default to the namespace of the service account rather than the "default" namespace which to which the user may not have access. Ideally, kubeapps would remember the selected namespace so when returning (without re-authing) it remains selected.
1.0
Kubeapps shows "default" namespace if user cannot list namespaces - If a service account is created for a user in a specific namespace (and without permission to list namespaces), and the associated credential is used to login, the namespace switcher shows "default" rather than the specific namespace. Kubeapps should default to the namespace of the service account rather than the "default" namespace which to which the user may not have access. Ideally, kubeapps would remember the selected namespace so when returning (without re-authing) it remains selected.
non_infrastructure
kubeapps shows default namespace if user cannot list namespaces if a service account is created for a user in a specific namespace and without permission to list namespaces and the associated credential is used to login the namespace switcher shows default rather than the specific namespace kubeapps should default to the namespace of the service account rather than the default namespace which to which the user may not have access ideally kubeapps would remember the selected namespace so when returning without re authing it remains selected
0
34,524
30,053,867,673
IssuesEvent
2023-06-28 04:29:12
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
closed
Automate Changelog Generation
tool: monorepo infrastructure
**Prerequisites (mark completed items with an [x]):** - [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose - [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is). **Issue Description:** In order to make working with changelogs easier, we should implement a feature for automatically generating them from the PR description. We should add a GitHub Action that: - Validation - The PR description should have a section for entering the changelog. - If the changelog does not follow the correct format we can fail a PR check. - PRs should NOT require a changelog for things that won't ever be published. Markdown files, `.github`, configuration like ESLint, etc. Only code changes that are shipped to users should trigger it. - If there are already changelog files present we should validate them using the JP changelogger and error if they're invalid. - Parsing - Any `push` or editing of the PR description should trigger this. - **Instead, only run prior to the merge taking place. This avoids conflicts with existing changelog files.** - Parse the format so we can construct a JP Changelogger command from it. - Scan for any projects that have been changed (we might even be able to use `pnpm --filter=trunk...HEAD`?) and check if they've got files that have changed that are valid. - If they do and there isn't already a changelog file from this PR, run the command we generated and add the one from the PR description. - Remove local changelog check and make the generation optional outside of the PR checks. As is standard now, we should put whatever makes sense into the `monorepo-utils` CLI tool. I might suggest it going as far as using GitHub's API to review the PR's description and avoid needing to do any parsing in the GitHub Action. Maybe instead of using the JP changelogger to make the changelog we can just use [this script](https://gist.github.com/rrennick/56285a3d3580aa1c5d38d50d9ba5a6cd). Given the scope of this issue, it may even make sense to just remove the JP changelogger entirely? There's some meat to discuss here.
1.0
Automate Changelog Generation - **Prerequisites (mark completed items with an [x]):** - [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose - [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is). **Issue Description:** In order to make working with changelogs easier, we should implement a feature for automatically generating them from the PR description. We should add a GitHub Action that: - Validation - The PR description should have a section for entering the changelog. - If the changelog does not follow the correct format we can fail a PR check. - PRs should NOT require a changelog for things that won't ever be published. Markdown files, `.github`, configuration like ESLint, etc. Only code changes that are shipped to users should trigger it. - If there are already changelog files present we should validate them using the JP changelogger and error if they're invalid. - Parsing - Any `push` or editing of the PR description should trigger this. - **Instead, only run prior to the merge taking place. This avoids conflicts with existing changelog files.** - Parse the format so we can construct a JP Changelogger command from it. - Scan for any projects that have been changed (we might even be able to use `pnpm --filter=trunk...HEAD`?) and check if they've got files that have changed that are valid. - If they do and there isn't already a changelog file from this PR, run the command we generated and add the one from the PR description. - Remove local changelog check and make the generation optional outside of the PR checks. As is standard now, we should put whatever makes sense into the `monorepo-utils` CLI tool. I might suggest it going as far as using GitHub's API to review the PR's description and avoid needing to do any parsing in the GitHub Action. Maybe instead of using the JP changelogger to make the changelog we can just use [this script](https://gist.github.com/rrennick/56285a3d3580aa1c5d38d50d9ba5a6cd). Given the scope of this issue, it may even make sense to just remove the JP changelogger entirely? There's some meat to discuss here.
infrastructure
automate changelog generation prerequisites mark completed items with an i have checked that my issue type is not listed here my issue is not a security issue support request bug report enhancement or feature request please use the link above if it is issue description in order to make working with changelogs easier we should implement a feature for automatically generating them from the pr description we should add a github action that validation the pr description should have a section for entering the changelog if the changelog does not follow the correct format we can fail a pr check prs should not require a changelog for things that won t ever be published markdown files github configuration like eslint etc only code changes that are shipped to users should trigger it if there are already changelog files present we should validate them using the jp changelogger and error if they re invalid parsing any push or editing of the pr description should trigger this instead only run prior to the merge taking place this avoids conflicts with existing changelog files parse the format so we can construct a jp changelogger command from it scan for any projects that have been changed we might even be able to use pnpm filter trunk head and check if they ve got files that have changed that are valid if they do and there isn t already a changelog file from this pr run the command we generated and add the one from the pr description remove local changelog check and make the generation optional outside of the pr checks as is standard now we should put whatever makes sense into the monorepo utils cli tool i might suggest it going as far as using github s api to review the pr s description and avoid needing to do any parsing in the github action maybe instead of using the jp changelogger to make the changelog we can just use given the scope of this issue it may even make sense to just remove the jp changelogger entirely there s some meat to discuss here
1
462,880
13,255,601,134
IssuesEvent
2020-08-20 11:14:10
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.cbs.com - Video is not supported on Opera browser
browser-opera priority-normal severity-critical
<!-- @browser: Opera 63.0.3368 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.80 Safari/537.36 OPR/63.0.3368.22 (Edition beta) --> <!-- @reported_with: --> **URL**: https://www.cbs.com/shows/cbs_this_morning/video/cK6tv4NFGkyNVVjonYuUnjEg5oRIxE2_/how-jeffrey-epstein-s-accusers-are-reacting-to-his-apparent-suicide/ **Browser / Version**: Opera 63.0.3368 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Video or audio doesn't play **Description**: Website puts up notice to use either Chrome or Firefox when trying to watch video **Steps to Reproduce**: I subscribe to CBS All Access. It seems the site is browser sniffing, not letting you watch video in either Opera Stable or Beta, even when switching the user agent to Chrome or Firefox. [![Screenshot Description](https://webcompat.com/uploads/2019/8/7c728241-3c25-46ee-8db3-65e452238874-thumb.jpeg)](https://webcompat.com/uploads/2019/8/7c728241-3c25-46ee-8db3-65e452238874.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.cbs.com - Video is not supported on Opera browser - <!-- @browser: Opera 63.0.3368 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.80 Safari/537.36 OPR/63.0.3368.22 (Edition beta) --> <!-- @reported_with: --> **URL**: https://www.cbs.com/shows/cbs_this_morning/video/cK6tv4NFGkyNVVjonYuUnjEg5oRIxE2_/how-jeffrey-epstein-s-accusers-are-reacting-to-his-apparent-suicide/ **Browser / Version**: Opera 63.0.3368 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Video or audio doesn't play **Description**: Website puts up notice to use either Chrome or Firefox when trying to watch video **Steps to Reproduce**: I subscribe to CBS All Access. It seems the site is browser sniffing, not letting you watch video in either Opera Stable or Beta, even when switching the user agent to Chrome or Firefox. [![Screenshot Description](https://webcompat.com/uploads/2019/8/7c728241-3c25-46ee-8db3-65e452238874-thumb.jpeg)](https://webcompat.com/uploads/2019/8/7c728241-3c25-46ee-8db3-65e452238874.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_infrastructure
video is not supported on opera browser url browser version opera operating system windows tested another browser yes problem type video or audio doesn t play description website puts up notice to use either chrome or firefox when trying to watch video steps to reproduce i subscribe to cbs all access it seems the site is browser sniffing not letting you watch video in either opera stable or beta even when switching the user agent to chrome or firefox browser configuration none from with ❤️
0
3,060
4,031,893,140
IssuesEvent
2016-05-18 18:43:10
trackpete/exiletools-indexer
https://api.github.com/repos/trackpete/exiletools-indexer
closed
v4 New ExileTools Infrastructure
infrastructure / exiletools.com setup
This issue will serve as a placeholder for the plans to implement the new Infrastructure required for production deployment of both the ExileTools v4 Indexer as well as the associated production infrastructure. This will give interested parties additional insight into the current production environment as well as illustrating how the organic growth of ExileTools and the adoption of various tools over time has led to a inefficient infrastructure. # Current Infrastructure Overview ## Hardware / Software ### pwx * Primary Server * 64GB RAM, 16 cpu cores * ~800GB of compressed + deduplicated data on NFS Back End Services include: 1. MariaDB (only used for Ladder API and league information at this time) 2. Jenkins (manages execution of Ladder API updates) 3. ElasticSearch Cluster Master (stats, items, and other indexes) 4. All Indexer related pipeline tools, such as river-watch, etc. Front End Services include: 1. Apache, serves ALL primary web content for exiletools.com as well as the older Ladder API front end 2. Varnish (all requests to Apache go through Varnish for front end caching) ### pwx2 * Secondary Server * 64GB RAM, 16 cpu cores Back End Services include: 1. ElasticSearch Cluster Secondary. Kibana is hosted here as well. 2. Tyk.io API Gateway (handles incoming API requests by firing them off at localhost:9200 for ElasticSearch). Redis and MongoDB are used by Tyk. Front End Services include: 1. nginx for any incoming API requests. nginx modifies basic http auth into something Tyk can read as well as provides a short front-end cache for all API requests (including POSTS) based on checksums of the request payload. ### pwx3 * Secondary Server, ES only * 48GB RAM, 12 cpu cores Back End Services include: 1. ElasticSearch Cluster member. Only purpose for this server, by adding another 30GB JVM it increases the amount of items stored in memory on the shards. ### haproxy * Small VM running front end software load balancer instance which provides SSL offload as well as routing to either Varnish on pwx for primary requests or Nginx on pwx2 for ES Index requests. All external inbound traffic first passes through this system. ## ElasticSearch 1. Runs on NFS - slow! 2. Three 30GB JVM's spread across three machines, with each index having two shards per JVM, means the indexes can hold ~40GB of data in memory between them while still supporting full redundancy in the event of node failure. This is not enough memory to manage a long running item index. 3. Internal programs access the ES cluster directly, while external programs must go through haproxy->nginx->tyk->ES. # The Future / Planned Changes These changes will be deployed into production at some point after or while deploying the v4 Indexer Pipeline. I may wait to make full infrastructure stack changes until the Perandus leagues are over. I will announce my plans on twitter when they are nailed down. 1. New Backend Machine: 96GB RAM, 24 cpu cores, with 2x240GB SSD's. This will be the *only* ElasticSearch machine moving forward, with two JVM's running on it. It will be a single point of failure, but the performance gains are worth it. This machine will also run Kafka and the entire v4 Indexer Pipeline. 2. New Hybrid Services Machine: Probably a small 16GB VM with 8 cores or so. This machine will run the Tyk API Gateway (2.0), Apache, and a very small MariaDB instance for the ladder (which really needs to be transitioned to ElasticSearch). 3. New Front-end Machine: Probably a small 16GB VM with 8 cores or so. This machine will run haproxy as well as a Varnish cache for Apache and an nginx cache for Tyk (why nginx? it's just sooo much easier to handle caching for POST requests in nginx). # Other Big Changes I think that when I move to Tyk 2.0 I will drop the requirement for API keys and authorization headers and just fully open the index - instead, I will just apply rate limiting by end user IP address. The main reason I originally asked for people to sign up for API keys was so that I had a list of the userbase, but at this point there are something like 500+ API keys out there and less than 20 are in active daily use, so it's not very useful or accurate. I'm also considering picking up some "new" hardware off ebay instead of just deploying on one primary box. Let's see where this goes first though.
1.0
v4 New ExileTools Infrastructure - This issue will serve as a placeholder for the plans to implement the new Infrastructure required for production deployment of both the ExileTools v4 Indexer as well as the associated production infrastructure. This will give interested parties additional insight into the current production environment as well as illustrating how the organic growth of ExileTools and the adoption of various tools over time has led to a inefficient infrastructure. # Current Infrastructure Overview ## Hardware / Software ### pwx * Primary Server * 64GB RAM, 16 cpu cores * ~800GB of compressed + deduplicated data on NFS Back End Services include: 1. MariaDB (only used for Ladder API and league information at this time) 2. Jenkins (manages execution of Ladder API updates) 3. ElasticSearch Cluster Master (stats, items, and other indexes) 4. All Indexer related pipeline tools, such as river-watch, etc. Front End Services include: 1. Apache, serves ALL primary web content for exiletools.com as well as the older Ladder API front end 2. Varnish (all requests to Apache go through Varnish for front end caching) ### pwx2 * Secondary Server * 64GB RAM, 16 cpu cores Back End Services include: 1. ElasticSearch Cluster Secondary. Kibana is hosted here as well. 2. Tyk.io API Gateway (handles incoming API requests by firing them off at localhost:9200 for ElasticSearch). Redis and MongoDB are used by Tyk. Front End Services include: 1. nginx for any incoming API requests. nginx modifies basic http auth into something Tyk can read as well as provides a short front-end cache for all API requests (including POSTS) based on checksums of the request payload. ### pwx3 * Secondary Server, ES only * 48GB RAM, 12 cpu cores Back End Services include: 1. ElasticSearch Cluster member. Only purpose for this server, by adding another 30GB JVM it increases the amount of items stored in memory on the shards. ### haproxy * Small VM running front end software load balancer instance which provides SSL offload as well as routing to either Varnish on pwx for primary requests or Nginx on pwx2 for ES Index requests. All external inbound traffic first passes through this system. ## ElasticSearch 1. Runs on NFS - slow! 2. Three 30GB JVM's spread across three machines, with each index having two shards per JVM, means the indexes can hold ~40GB of data in memory between them while still supporting full redundancy in the event of node failure. This is not enough memory to manage a long running item index. 3. Internal programs access the ES cluster directly, while external programs must go through haproxy->nginx->tyk->ES. # The Future / Planned Changes These changes will be deployed into production at some point after or while deploying the v4 Indexer Pipeline. I may wait to make full infrastructure stack changes until the Perandus leagues are over. I will announce my plans on twitter when they are nailed down. 1. New Backend Machine: 96GB RAM, 24 cpu cores, with 2x240GB SSD's. This will be the *only* ElasticSearch machine moving forward, with two JVM's running on it. It will be a single point of failure, but the performance gains are worth it. This machine will also run Kafka and the entire v4 Indexer Pipeline. 2. New Hybrid Services Machine: Probably a small 16GB VM with 8 cores or so. This machine will run the Tyk API Gateway (2.0), Apache, and a very small MariaDB instance for the ladder (which really needs to be transitioned to ElasticSearch). 3. New Front-end Machine: Probably a small 16GB VM with 8 cores or so. This machine will run haproxy as well as a Varnish cache for Apache and an nginx cache for Tyk (why nginx? it's just sooo much easier to handle caching for POST requests in nginx). # Other Big Changes I think that when I move to Tyk 2.0 I will drop the requirement for API keys and authorization headers and just fully open the index - instead, I will just apply rate limiting by end user IP address. The main reason I originally asked for people to sign up for API keys was so that I had a list of the userbase, but at this point there are something like 500+ API keys out there and less than 20 are in active daily use, so it's not very useful or accurate. I'm also considering picking up some "new" hardware off ebay instead of just deploying on one primary box. Let's see where this goes first though.
infrastructure
new exiletools infrastructure this issue will serve as a placeholder for the plans to implement the new infrastructure required for production deployment of both the exiletools indexer as well as the associated production infrastructure this will give interested parties additional insight into the current production environment as well as illustrating how the organic growth of exiletools and the adoption of various tools over time has led to a inefficient infrastructure current infrastructure overview hardware software pwx primary server ram cpu cores of compressed deduplicated data on nfs back end services include mariadb only used for ladder api and league information at this time jenkins manages execution of ladder api updates elasticsearch cluster master stats items and other indexes all indexer related pipeline tools such as river watch etc front end services include apache serves all primary web content for exiletools com as well as the older ladder api front end varnish all requests to apache go through varnish for front end caching secondary server ram cpu cores back end services include elasticsearch cluster secondary kibana is hosted here as well tyk io api gateway handles incoming api requests by firing them off at localhost for elasticsearch redis and mongodb are used by tyk front end services include nginx for any incoming api requests nginx modifies basic http auth into something tyk can read as well as provides a short front end cache for all api requests including posts based on checksums of the request payload secondary server es only ram cpu cores back end services include elasticsearch cluster member only purpose for this server by adding another jvm it increases the amount of items stored in memory on the shards haproxy small vm running front end software load balancer instance which provides ssl offload as well as routing to either varnish on pwx for primary requests or nginx on for es index requests all external inbound traffic first passes through this system elasticsearch runs on nfs slow three jvm s spread across three machines with each index having two shards per jvm means the indexes can hold of data in memory between them while still supporting full redundancy in the event of node failure this is not enough memory to manage a long running item index internal programs access the es cluster directly while external programs must go through haproxy nginx tyk es the future planned changes these changes will be deployed into production at some point after or while deploying the indexer pipeline i may wait to make full infrastructure stack changes until the perandus leagues are over i will announce my plans on twitter when they are nailed down new backend machine ram cpu cores with ssd s this will be the only elasticsearch machine moving forward with two jvm s running on it it will be a single point of failure but the performance gains are worth it this machine will also run kafka and the entire indexer pipeline new hybrid services machine probably a small vm with cores or so this machine will run the tyk api gateway apache and a very small mariadb instance for the ladder which really needs to be transitioned to elasticsearch new front end machine probably a small vm with cores or so this machine will run haproxy as well as a varnish cache for apache and an nginx cache for tyk why nginx it s just sooo much easier to handle caching for post requests in nginx other big changes i think that when i move to tyk i will drop the requirement for api keys and authorization headers and just fully open the index instead i will just apply rate limiting by end user ip address the main reason i originally asked for people to sign up for api keys was so that i had a list of the userbase but at this point there are something like api keys out there and less than are in active daily use so it s not very useful or accurate i m also considering picking up some new hardware off ebay instead of just deploying on one primary box let s see where this goes first though
1
671,484
22,763,143,062
IssuesEvent
2022-07-07 23:41:56
touretzkyds/KnowledgeGraphDemo
https://api.github.com/repos/touretzkyds/KnowledgeGraphDemo
closed
taxon name input cleanup
bug high priority
We need to clean up the input we receive in the taxon name text box. Delete any leading or trailing whitespace. Replace any internal whitespace (spaces or tabs) with a single space. Otherwise the name lookup will fail, as just happened to me when I left a trailing space on a genus name.
1.0
taxon name input cleanup - We need to clean up the input we receive in the taxon name text box. Delete any leading or trailing whitespace. Replace any internal whitespace (spaces or tabs) with a single space. Otherwise the name lookup will fail, as just happened to me when I left a trailing space on a genus name.
non_infrastructure
taxon name input cleanup we need to clean up the input we receive in the taxon name text box delete any leading or trailing whitespace replace any internal whitespace spaces or tabs with a single space otherwise the name lookup will fail as just happened to me when i left a trailing space on a genus name
0
10,788
8,722,281,795
IssuesEvent
2018-12-09 10:48:36
coq/coq
https://api.github.com/repos/coq/coq
closed
Change extensions of log files in WIndows build to _log.txt and _err.txt so that they can be viewed immediately in gitlab
kind: infrastructure platform: Windows priority: low
In the windows build log files are now browsable with artifacts browse, but it is not possible to look at log files in the browser cause of the extension. So change the file extension to .txt for all .log and .err files.
1.0
Change extensions of log files in WIndows build to _log.txt and _err.txt so that they can be viewed immediately in gitlab - In the windows build log files are now browsable with artifacts browse, but it is not possible to look at log files in the browser cause of the extension. So change the file extension to .txt for all .log and .err files.
infrastructure
change extensions of log files in windows build to log txt and err txt so that they can be viewed immediately in gitlab in the windows build log files are now browsable with artifacts browse but it is not possible to look at log files in the browser cause of the extension so change the file extension to txt for all log and err files
1
34,985
30,656,790,404
IssuesEvent
2023-07-25 12:41:24
Sinyks/AMT2020-StackOverflowSimpleVersion
https://api.github.com/repos/Sinyks/AMT2020-StackOverflowSimpleVersion
closed
add business logic for tags
Domain infrastructure
- [x] domain - [x] database (repository) - [ ] management facade - [ ] tags DTO
1.0
add business logic for tags - - [x] domain - [x] database (repository) - [ ] management facade - [ ] tags DTO
infrastructure
add business logic for tags domain database repository management facade tags dto
1
358,535
10,618,538,470
IssuesEvent
2019-10-13 05:31:51
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
opened
/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md
lang/zh priority/P0 sync/update version/1.16 welcome
Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md) Diff 查看原始文档更新差异命令: ```bash git diff release-1.14 release-1.16 -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md ```
1.0
/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md) Diff 查看原始文档更新差异命令: ```bash git diff release-1.14 release-1.16 -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md ```
non_infrastructure
docs reference setup tools kubeadm generated kubeadm alpha md source file diff 查看原始文档更新差异命令 bash git diff release release content en docs reference setup tools kubeadm generated kubeadm alpha md
0
252
2,595,806,656
IssuesEvent
2015-02-20 16:35:54
Starcounter/Starcounter
https://api.github.com/repos/Starcounter/Starcounter
closed
Out of memory error for three running databases on build server
Follow-up Infrastructure P/High stability issue
I try to run my regression test with three simple apps on the build server. Each app creates own database. When third database creates a database I get the error (see, e.g., [this build](https://scbuildserver/viewLog.html?buildId=24172&buildTypeId=Starcounter_DevelopDaily&tab=buildLog): >[14:41:05][Step 14/42] [14:41:04, Critical, Starcounter (selectonnull2362db)] [14:41:05][Step 14/42] System.OutOfMemoryException: ScErrOutOfMemory (SCERR1003): Operation failed because needed memory couldn't be allocated. [14:41:05][Step 14/42] Version: 2.0.8424.2. [14:41:05][Step 14/42] Help page: https://github.com/Starcounter/Starcounter/wiki/SCERR1003. [14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.ConnectDatabase(UInt32 schedulerCount, UInt64 hmenv, UInt64 hlogs) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 557 [14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Setup(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 208 [14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Main(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 55 [14:41:05][Step 14/42] HResult=-2147024882 [14:41:05][Step 14/42] HelpLink=https://github.com/Starcounter/Starcounter/wiki/SCERR1003 Currently I disabled the third test app, since it will fail nightly build due to this problem. After I fix #2362 I will return to this one and describe how to reproduce. If anyone can hint about the problem, post here.
1.0
Out of memory error for three running databases on build server - I try to run my regression test with three simple apps on the build server. Each app creates own database. When third database creates a database I get the error (see, e.g., [this build](https://scbuildserver/viewLog.html?buildId=24172&buildTypeId=Starcounter_DevelopDaily&tab=buildLog): >[14:41:05][Step 14/42] [14:41:04, Critical, Starcounter (selectonnull2362db)] [14:41:05][Step 14/42] System.OutOfMemoryException: ScErrOutOfMemory (SCERR1003): Operation failed because needed memory couldn't be allocated. [14:41:05][Step 14/42] Version: 2.0.8424.2. [14:41:05][Step 14/42] Help page: https://github.com/Starcounter/Starcounter/wiki/SCERR1003. [14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.ConnectDatabase(UInt32 schedulerCount, UInt64 hmenv, UInt64 hlogs) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 557 [14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Setup(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 208 [14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Main(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 55 [14:41:05][Step 14/42] HResult=-2147024882 [14:41:05][Step 14/42] HelpLink=https://github.com/Starcounter/Starcounter/wiki/SCERR1003 Currently I disabled the third test app, since it will fail nightly build due to this problem. After I fix #2362 I will return to this one and describe how to reproduce. If anyone can hint about the problem, post here.
infrastructure
out of memory error for three running databases on build server i try to run my regression test with three simple apps on the build server each app creates own database when third database creates a database i get the error see e g system outofmemoryexception scerroutofmemory operation failed because needed memory couldn t be allocated version help page at starcounterinternal bootstrap control connectdatabase schedulercount hmenv hlogs in c teamcity buildagent work build daily develop src starcounter bootstrap control cs line at starcounterinternal bootstrap control setup string args in c teamcity buildagent work build daily develop src starcounter bootstrap control cs line at starcounterinternal bootstrap control main string args in c teamcity buildagent work build daily develop src starcounter bootstrap control cs line hresult helplink currently i disabled the third test app since it will fail nightly build due to this problem after i fix i will return to this one and describe how to reproduce if anyone can hint about the problem post here
1
281
2,610,651,253
IssuesEvent
2015-02-26 21:38:41
uProxy/uproxy
https://api.github.com/repos/uProxy/uproxy
opened
set setBufferedLogFilter to help reduce memory usage
C:Infrastructure C:Networking P1
Although we call `setConsoleFilter`, we are not setting `setBufferedLogFilter`. This may be leading to *lots* of debug messages building up in RAM. Let's set it and see what difference it makes. Here's where it should be set in Simple SOCKS: https://github.com/uProxy/uproxy-networking/blob/dev/src/simple-socks/freedom-module.ts#L13
1.0
set setBufferedLogFilter to help reduce memory usage - Although we call `setConsoleFilter`, we are not setting `setBufferedLogFilter`. This may be leading to *lots* of debug messages building up in RAM. Let's set it and see what difference it makes. Here's where it should be set in Simple SOCKS: https://github.com/uProxy/uproxy-networking/blob/dev/src/simple-socks/freedom-module.ts#L13
infrastructure
set setbufferedlogfilter to help reduce memory usage although we call setconsolefilter we are not setting setbufferedlogfilter this may be leading to lots of debug messages building up in ram let s set it and see what difference it makes here s where it should be set in simple socks
1
26,372
11,300,324,782
IssuesEvent
2020-01-17 13:22:13
aleksaa01/WorkTodo
https://api.github.com/repos/aleksaa01/WorkTodo
closed
Storage is not cleaned up upon logout
bug security
When you logout, token in storage is set to null, but cards, tasks and preferences are not deleted. This is obviously a security flaw.
True
Storage is not cleaned up upon logout - When you logout, token in storage is set to null, but cards, tasks and preferences are not deleted. This is obviously a security flaw.
non_infrastructure
storage is not cleaned up upon logout when you logout token in storage is set to null but cards tasks and preferences are not deleted this is obviously a security flaw
0
13,311
10,198,237,403
IssuesEvent
2019-08-13 04:16:45
AtlasOfLivingAustralia/data-management
https://api.github.com/repos/AtlasOfLivingAustralia/data-management
closed
Cassandra server changes
Infrastructure Performance
- [x] Document proposed cassandra server changes for CAB - [x] Testing (performance and load) plan - [x] Deployment plan - [x] Rollback plan
1.0
Cassandra server changes - - [x] Document proposed cassandra server changes for CAB - [x] Testing (performance and load) plan - [x] Deployment plan - [x] Rollback plan
infrastructure
cassandra server changes document proposed cassandra server changes for cab testing performance and load plan deployment plan rollback plan
1
302,143
26,128,248,570
IssuesEvent
2022-12-28 22:22:11
omansak/libvideo
https://api.github.com/repos/omansak/libvideo
closed
Unable to download a specific video
todo need test
Hi there! I've notice that it looks like the library is not able to download a specific video. When i try to download https://www.youtube.com/watch?v=b2YUw1rqbTY the library stops working without any error... I can defently live with that :) I just would like to inform you about this issuse I've found. Thanks for your great job!
1.0
Unable to download a specific video - Hi there! I've notice that it looks like the library is not able to download a specific video. When i try to download https://www.youtube.com/watch?v=b2YUw1rqbTY the library stops working without any error... I can defently live with that :) I just would like to inform you about this issuse I've found. Thanks for your great job!
non_infrastructure
unable to download a specific video hi there i ve notice that it looks like the library is not able to download a specific video when i try to download the library stops working without any error i can defently live with that i just would like to inform you about this issuse i ve found thanks for your great job
0
507,982
14,686,218,138
IssuesEvent
2021-01-01 13:50:42
googleapis/python-dialogflow-cx
https://api.github.com/repos/googleapis/python-dialogflow-cx
closed
Synthesis failed for python-dialogflow-cx
api: dialogflow autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate python-dialogflow-cx. :broken_heart: Here's the output from running `synth.py`: ``` /.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_pages.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_security_settings_service.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_sessions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_session_entity_types.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_transition_route_groups.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_versions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_webhooks.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_agents.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_entity_types.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_environments.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_experiments.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_intents.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_flows.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_pages.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_sessions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_security_settings_service.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_session_entity_types.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_transition_route_groups.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_versions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_webhooks.py All done! ✨ 🍰 ✨ 271 files reformatted, 5 files left unchanged. nox > Session blacken was successful. 2020-12-31 05:48:08,762 synthtool [DEBUG] > Wrote metadata to synth.metadata. 2020-12-31 05:48:08,829 autosynth [INFO] > Changed files: 2020-12-31 05:48:08,829 autosynth [INFO] > M google/cloud/dialogflowcx_v3/__init__.py M google/cloud/dialogflowcx_v3beta1/__init__.py M synth.metadata ?? dialogflow-cx-v3-py.tar.gz 2020-12-31 05:48:08,829 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%s 2020-12-31 05:48:08,832 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%b%n%nSource-Author: %an <%ae>%nSource-Date: %ad 2020-12-31 05:48:08,836 autosynth [DEBUG] > Running: git add -A 2020-12-31 05:48:08,841 autosynth [DEBUG] > Running: git status --porcelain 2020-12-31 05:48:08,847 autosynth [DEBUG] > Running: git commit -m feat: add experiments API Source-Author: Yoshi Automation Bot <yoshi-automation@google.com> Source-Date: Wed Dec 30 08:14:21 2020 -0800 Source-Repo: googleapis/python-dialogflow-cx Source-Sha: 5381512872ca2492ddabcbdd7ccde5f054aed011 Source-Link: https://github.com/googleapis/python-dialogflow-cx/commit/5381512872ca2492ddabcbdd7ccde5f054aed011 [autosynth-self-0 76444fc] feat: add experiments API 4 files changed, 8 insertions(+), 7 deletions(-) create mode 100644 dialogflow-cx-v3-py.tar.gz 2020-12-31 05:48:08,854 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 76444fc feat: add experiments API 2020-12-31 05:48:08,859 autosynth [DEBUG] > Running: git checkout autosynth-self Switched to branch 'autosynth-self' 2020-12-31 05:48:08,864 autosynth [DEBUG] > Running: git diff HEAD..autosynth-self-0 -- . :(exclude)synth.metadata 2020-12-31 05:48:08,869 autosynth [DEBUG] > Running: git diff HEAD autosynth-self-0 2020-12-31 05:48:08,872 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line error: dialogflow-cx-v3-py.tar.gz: patch does not apply 2020-12-31 05:48:08,876 autosynth [ERROR] > Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1. 2020-12-31 05:48:09,066 autosynth [DEBUG] > Running: git checkout 5381512872ca2492ddabcbdd7ccde5f054aed011 Note: checking out '5381512872ca2492ddabcbdd7ccde5f054aed011'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 5381512 feat: add experiments API (#36) 2020-12-31 05:48:09,071 autosynth [DEBUG] > Running: git checkout 373861061648b5fe5e0ac4f8a38b32d639ee93e4 HEAD is now at 3738610 chore: exclude `.nox` directories from linting (#885) 2020-12-31 05:48:09,076 autosynth [DEBUG] > Running: git checkout 69697504d9eba1d064820c3085b4750767be6d08 HEAD is now at 69697504 fix: fixed config for Python and Java grpc lib 2020-12-31 05:48:09,089 autosynth [DEBUG] > Running: git branch -f autosynth-0 2020-12-31 05:48:09,092 autosynth [DEBUG] > Running: git checkout autosynth-0 Switched to branch 'autosynth-0' 2020-12-31 05:48:09,096 autosynth [DEBUG] > Running: git merge --ff-only autosynth-self-0 Updating 5381512..76444fc Fast-forward dialogflow-cx-v3-py.tar.gz | Bin 0 -> 49152 bytes google/cloud/dialogflowcx_v3/__init__.py | 6 +++--- google/cloud/dialogflowcx_v3beta1/__init__.py | 6 +++--- synth.metadata | 3 ++- 4 files changed, 8 insertions(+), 7 deletions(-) create mode 100644 dialogflow-cx-v3-py.tar.gz 2020-12-31 05:48:09,105 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 76444fc feat: add experiments API 2020-12-31 05:48:09,110 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-12-31 05:48:09,116 autosynth [DEBUG] > Running: git diff HEAD..autosynth-0 -- . :(exclude)synth.metadata 2020-12-31 05:48:09,120 autosynth [DEBUG] > Running: git diff HEAD autosynth-0 2020-12-31 05:48:09,124 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line error: dialogflow-cx-v3-py.tar.gz: patch does not apply 2020-12-31 05:48:09,127 autosynth [DEBUG] > Running: git clean -fdx Removing .nox/ Removing __pycache__/ Traceback (most recent call last): File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 92, in synthesize_loop synthesize_inner_loop(fork, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop toolbox.patch_merge_version(0, comment) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version git.patch_merge(sub_branch, patch_file_path) File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir) File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call subprocess.check_call(command, **args) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 334, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 100, in synthesize_loop synthesize_loop_single_pr(toolbox, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 126, in synthesize_loop_single_pr synthesize_inner_loop(toolbox, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop toolbox.patch_merge_version(0, comment) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version git.patch_merge(sub_branch, patch_file_path) File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir) File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call subprocess.check_call(command, **args) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/a0f4ccaf-319e-46b9-9b15-5ac1e95ea79c/targets/github%2Fsynthtool;config=default/tests;query=python-dialogflow-cx;failed=false).
1.0
Synthesis failed for python-dialogflow-cx - Hello! Autosynth couldn't regenerate python-dialogflow-cx. :broken_heart: Here's the output from running `synth.py`: ``` /.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_pages.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_security_settings_service.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_sessions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_session_entity_types.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_transition_route_groups.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_versions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_webhooks.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_agents.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_entity_types.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_environments.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_experiments.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_intents.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_flows.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_pages.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_sessions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_security_settings_service.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_session_entity_types.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_transition_route_groups.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_versions.py reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_webhooks.py All done! ✨ 🍰 ✨ 271 files reformatted, 5 files left unchanged. nox > Session blacken was successful. 2020-12-31 05:48:08,762 synthtool [DEBUG] > Wrote metadata to synth.metadata. 2020-12-31 05:48:08,829 autosynth [INFO] > Changed files: 2020-12-31 05:48:08,829 autosynth [INFO] > M google/cloud/dialogflowcx_v3/__init__.py M google/cloud/dialogflowcx_v3beta1/__init__.py M synth.metadata ?? dialogflow-cx-v3-py.tar.gz 2020-12-31 05:48:08,829 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%s 2020-12-31 05:48:08,832 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%b%n%nSource-Author: %an <%ae>%nSource-Date: %ad 2020-12-31 05:48:08,836 autosynth [DEBUG] > Running: git add -A 2020-12-31 05:48:08,841 autosynth [DEBUG] > Running: git status --porcelain 2020-12-31 05:48:08,847 autosynth [DEBUG] > Running: git commit -m feat: add experiments API Source-Author: Yoshi Automation Bot <yoshi-automation@google.com> Source-Date: Wed Dec 30 08:14:21 2020 -0800 Source-Repo: googleapis/python-dialogflow-cx Source-Sha: 5381512872ca2492ddabcbdd7ccde5f054aed011 Source-Link: https://github.com/googleapis/python-dialogflow-cx/commit/5381512872ca2492ddabcbdd7ccde5f054aed011 [autosynth-self-0 76444fc] feat: add experiments API 4 files changed, 8 insertions(+), 7 deletions(-) create mode 100644 dialogflow-cx-v3-py.tar.gz 2020-12-31 05:48:08,854 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 76444fc feat: add experiments API 2020-12-31 05:48:08,859 autosynth [DEBUG] > Running: git checkout autosynth-self Switched to branch 'autosynth-self' 2020-12-31 05:48:08,864 autosynth [DEBUG] > Running: git diff HEAD..autosynth-self-0 -- . :(exclude)synth.metadata 2020-12-31 05:48:08,869 autosynth [DEBUG] > Running: git diff HEAD autosynth-self-0 2020-12-31 05:48:08,872 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line error: dialogflow-cx-v3-py.tar.gz: patch does not apply 2020-12-31 05:48:08,876 autosynth [ERROR] > Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1. 2020-12-31 05:48:09,066 autosynth [DEBUG] > Running: git checkout 5381512872ca2492ddabcbdd7ccde5f054aed011 Note: checking out '5381512872ca2492ddabcbdd7ccde5f054aed011'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by performing another checkout. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -b with the checkout command again. Example: git checkout -b <new-branch-name> HEAD is now at 5381512 feat: add experiments API (#36) 2020-12-31 05:48:09,071 autosynth [DEBUG] > Running: git checkout 373861061648b5fe5e0ac4f8a38b32d639ee93e4 HEAD is now at 3738610 chore: exclude `.nox` directories from linting (#885) 2020-12-31 05:48:09,076 autosynth [DEBUG] > Running: git checkout 69697504d9eba1d064820c3085b4750767be6d08 HEAD is now at 69697504 fix: fixed config for Python and Java grpc lib 2020-12-31 05:48:09,089 autosynth [DEBUG] > Running: git branch -f autosynth-0 2020-12-31 05:48:09,092 autosynth [DEBUG] > Running: git checkout autosynth-0 Switched to branch 'autosynth-0' 2020-12-31 05:48:09,096 autosynth [DEBUG] > Running: git merge --ff-only autosynth-self-0 Updating 5381512..76444fc Fast-forward dialogflow-cx-v3-py.tar.gz | Bin 0 -> 49152 bytes google/cloud/dialogflowcx_v3/__init__.py | 6 +++--- google/cloud/dialogflowcx_v3beta1/__init__.py | 6 +++--- synth.metadata | 3 ++- 4 files changed, 8 insertions(+), 7 deletions(-) create mode 100644 dialogflow-cx-v3-py.tar.gz 2020-12-31 05:48:09,105 autosynth [DEBUG] > Running: git reset --hard HEAD HEAD is now at 76444fc feat: add experiments API 2020-12-31 05:48:09,110 autosynth [DEBUG] > Running: git checkout autosynth Switched to branch 'autosynth' 2020-12-31 05:48:09,116 autosynth [DEBUG] > Running: git diff HEAD..autosynth-0 -- . :(exclude)synth.metadata 2020-12-31 05:48:09,120 autosynth [DEBUG] > Running: git diff HEAD autosynth-0 2020-12-31 05:48:09,124 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line error: dialogflow-cx-v3-py.tar.gz: patch does not apply 2020-12-31 05:48:09,127 autosynth [DEBUG] > Running: git clean -fdx Removing .nox/ Removing __pycache__/ Traceback (most recent call last): File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 92, in synthesize_loop synthesize_inner_loop(fork, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop toolbox.patch_merge_version(0, comment) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version git.patch_merge(sub_branch, patch_file_path) File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir) File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call subprocess.check_call(command, **args) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1. During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 334, in _inner_main commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 100, in synthesize_loop synthesize_loop_single_pr(toolbox, change_pusher, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 126, in synthesize_loop_single_pr synthesize_inner_loop(toolbox, synthesizer) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop toolbox.patch_merge_version(0, comment) File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version git.patch_merge(sub_branch, patch_file_path) File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir) File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call subprocess.check_call(command, **args) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch']' returned non-zero exit status 1. ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/a0f4ccaf-319e-46b9-9b15-5ac1e95ea79c/targets/github%2Fsynthtool;config=default/tests;query=python-dialogflow-cx;failed=false).
non_infrastructure
synthesis failed for python dialogflow cx hello autosynth couldn t regenerate python dialogflow cx broken heart here s the output from running synth py cache synthtool python dialogflow cx tests unit gapic dialogflowcx test pages py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test security settings service py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test sessions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test session entity types py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test transition route groups py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test versions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test webhooks py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test agents py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test entity types py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test environments py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test experiments py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test intents py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test flows py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test pages py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test sessions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test security settings service py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test session entity types py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test transition route groups py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test versions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test webhooks py all done ✨ 🍰 ✨ files reformatted files left unchanged nox session blacken was successful synthtool wrote metadata to synth metadata autosynth changed files autosynth m google cloud dialogflowcx init py m google cloud dialogflowcx init py m synth metadata dialogflow cx py tar gz autosynth running git log no decorate pretty s autosynth running git log no decorate pretty b n nsource author an nsource date ad autosynth running git add a autosynth running git status porcelain autosynth running git commit m feat add experiments api source author yoshi automation bot source date wed dec source repo googleapis python dialogflow cx source sha source link feat add experiments api files changed insertions deletions create mode dialogflow cx py tar gz autosynth running git reset hard head head is now at feat add experiments api autosynth running git checkout autosynth self switched to branch autosynth self autosynth running git diff head autosynth self exclude synth metadata autosynth running git diff head autosynth self autosynth running git apply tmpfs tmp autosynth self patch error cannot apply binary patch to dialogflow cx py tar gz without full index line error dialogflow cx py tar gz patch does not apply autosynth command returned non zero exit status autosynth running git checkout note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at feat add experiments api autosynth running git checkout head is now at chore exclude nox directories from linting autosynth running git checkout head is now at fix fixed config for python and java grpc lib autosynth running git branch f autosynth autosynth running git checkout autosynth switched to branch autosynth autosynth running git merge ff only autosynth self updating fast forward dialogflow cx py tar gz bin bytes google cloud dialogflowcx init py google cloud dialogflowcx init py synth metadata files changed insertions deletions create mode dialogflow cx py tar gz autosynth running git reset hard head head is now at feat add experiments api autosynth running git checkout autosynth switched to branch autosynth autosynth running git diff head autosynth exclude synth metadata autosynth running git diff head autosynth autosynth running git apply tmpfs tmp autosynth patch error cannot apply binary patch to dialogflow cx py tar gz without full index line error dialogflow cx py tar gz patch does not apply autosynth running git clean fdx removing nox removing pycache traceback most recent call last file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize inner loop fork synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop toolbox patch merge version comment file tmpfs src github synthtool autosynth synth toolbox py line in patch merge version git patch merge sub branch patch file path file tmpfs src github synthtool autosynth git py line in patch merge executor check call cwd git repo dir file tmpfs src github synthtool autosynth executor py line in check call subprocess check call command args file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status during handling of the above exception another exception occurred traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize loop single pr toolbox change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop single pr synthesize inner loop toolbox synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop toolbox patch merge version comment file tmpfs src github synthtool autosynth synth toolbox py line in patch merge version git patch merge sub branch patch file path file tmpfs src github synthtool autosynth git py line in patch merge executor check call cwd git repo dir file tmpfs src github synthtool autosynth executor py line in check call subprocess check call command args file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log
0
115,425
24,762,865,739
IssuesEvent
2022-10-22 05:51:13
log2timeline/plaso
https://api.github.com/repos/log2timeline/plaso
closed
refactor usbstor winreg plugin
enhancement parsers analysis code health
**Description of problem:** https://github.com/log2timeline/plaso/blob/master/plaso/parsers/winreg_plugins/usbstor.py#L45 The usbstor winreg plugin current assumes connection times based on the key last written time: - [x] change parser to produce event objects with last written time - [x] ~~add setup api information~~ - setupapi parser has been added - [x] ~~move the determination of the connection times to an analysis plugin~~ - out of scope
1.0
refactor usbstor winreg plugin - **Description of problem:** https://github.com/log2timeline/plaso/blob/master/plaso/parsers/winreg_plugins/usbstor.py#L45 The usbstor winreg plugin current assumes connection times based on the key last written time: - [x] change parser to produce event objects with last written time - [x] ~~add setup api information~~ - setupapi parser has been added - [x] ~~move the determination of the connection times to an analysis plugin~~ - out of scope
non_infrastructure
refactor usbstor winreg plugin description of problem the usbstor winreg plugin current assumes connection times based on the key last written time change parser to produce event objects with last written time add setup api information setupapi parser has been added move the determination of the connection times to an analysis plugin out of scope
0
8,388
7,375,375,097
IssuesEvent
2018-03-14 00:06:23
Microsoft/TypeScript
https://api.github.com/repos/Microsoft/TypeScript
closed
Nightly builds not published to npm (starting 3/8/2017)
Infrastructure
For any others who may file this, we're aware of the issue and will have it back up by next week. Thanks!
1.0
Nightly builds not published to npm (starting 3/8/2017) - For any others who may file this, we're aware of the issue and will have it back up by next week. Thanks!
infrastructure
nightly builds not published to npm starting for any others who may file this we re aware of the issue and will have it back up by next week thanks
1
413,195
12,061,292,904
IssuesEvent
2020-04-15 23:20:28
grpc/grpc
https://api.github.com/repos/grpc/grpc
closed
Timeout: unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate
disposition/stale kind/bug lang/Python priority/P2
``` ++ dirname /var/local/git/grpc/tools/run_tests/helper_scripts/run_python.sh + cd /var/local/git/grpc/tools/run_tests/helper_scripts/../../.. ++ realpath py36_native/bin/python + PYTHON=/var/local/git/grpc/py36_native/bin/python3.6 ++ pwd + ROOT=/var/local/git/grpc + /var/local/git/grpc/py36_native/bin/python3.6 /var/local/git/grpc/src/python/grpcio_tests/setup.py test_lite running test_lite Running unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call [95mTesting gRPC Python... [0m[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call [0mRunning unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_stream_stream_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_stream_unary_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_unary_call [0mRunning unit._exit_test.ExitTest.test_in_flight_unary_stream_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_unary_unary_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_unary_call [0mRunning unit._exit_test.ExitTest.test_poll_connectivity [92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity [0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server [92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity_no_server [0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate ``` https://source.cloud.google.com/results/invocations/848afa18-6b20-45ce-959b-09b094219029/targets/github%2Fgrpc%2Frun_tests%2Fpython_linux_opt_native%2Fpy36_native.test.unit._exit_test.ExitTest/tests
1.0
Timeout: unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate - ``` ++ dirname /var/local/git/grpc/tools/run_tests/helper_scripts/run_python.sh + cd /var/local/git/grpc/tools/run_tests/helper_scripts/../../.. ++ realpath py36_native/bin/python + PYTHON=/var/local/git/grpc/py36_native/bin/python3.6 ++ pwd + ROOT=/var/local/git/grpc + /var/local/git/grpc/py36_native/bin/python3.6 /var/local/git/grpc/src/python/grpcio_tests/setup.py test_lite running test_lite Running unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call [95mTesting gRPC Python... [0m[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call [0mRunning unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_stream_stream_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_stream_unary_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_unary_call [0mRunning unit._exit_test.ExitTest.test_in_flight_unary_stream_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_stream_call [0mRunning unit._exit_test.ExitTest.test_in_flight_unary_unary_call [92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_unary_call [0mRunning unit._exit_test.ExitTest.test_poll_connectivity [92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity [0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server [92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity_no_server [0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate ``` https://source.cloud.google.com/results/invocations/848afa18-6b20-45ce-959b-09b094219029/targets/github%2Fgrpc%2Frun_tests%2Fpython_linux_opt_native%2Fpy36_native.test.unit._exit_test.ExitTest/tests
non_infrastructure
timeout unit exit test exittest test poll connectivity no server terminate dirname var local git grpc tools run tests helper scripts run python sh cd var local git grpc tools run tests helper scripts realpath native bin python python var local git grpc native bin pwd root var local git grpc var local git grpc native bin var local git grpc src python grpcio tests setup py test lite running test lite running unit exit test exittest test in flight partial stream stream call grpc python unit exit test exittest test in flight partial stream stream call unit exit test exittest test in flight partial stream unary call unit exit test exittest test in flight partial stream unary call unit exit test exittest test in flight partial unary stream call unit exit test exittest test in flight partial unary stream call unit exit test exittest test in flight stream stream call unit exit test exittest test in flight stream stream call unit exit test exittest test in flight stream unary call unit exit test exittest test in flight stream unary call unit exit test exittest test in flight unary stream call unit exit test exittest test in flight unary stream call unit exit test exittest test in flight unary unary call unit exit test exittest test in flight unary unary call unit exit test exittest test poll connectivity unit exit test exittest test poll connectivity unit exit test exittest test poll connectivity no server unit exit test exittest test poll connectivity no server unit exit test exittest test poll connectivity no server terminate
0
24,563
17,399,634,610
IssuesEvent
2021-08-02 17:41:42
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
Several issues with MessagePack when building asp net core 5
area-infrastructure
I'm stuck with building asp dot net core 5.0 on Windows 10 I followed the instruction but have no luck with it. The output of build.cmd reporting me several errors that some files are missed in MessagePack > CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\T4\MessagePackReader.Integers.cs". > CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\MessagePackSerializationException.cs". fix_symlinks.bat in MessagaPack.UnityClient doesn't fix this. Could anybody help me with it what I missed? My steps are (commit 3f873035c9ee96c4e4e93a9d3966baf3ef02581a) 1. restore.cmd 2. build.cmd Thank you!
1.0
Several issues with MessagePack when building asp net core 5 - I'm stuck with building asp dot net core 5.0 on Windows 10 I followed the instruction but have no luck with it. The output of build.cmd reporting me several errors that some files are missed in MessagePack > CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\T4\MessagePackReader.Integers.cs". > CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\MessagePackSerializationException.cs". fix_symlinks.bat in MessagaPack.UnityClient doesn't fix this. Could anybody help me with it what I missed? My steps are (commit 3f873035c9ee96c4e4e93a9d3966baf3ef02581a) 1. restore.cmd 2. build.cmd Thank you!
infrastructure
several issues with messagepack when building asp net core i m stuck with building asp dot net core on windows i followed the instruction but have no luck with it the output of build cmd reporting me several errors that some files are missed in messagepack file not found mvc source aspnetcore src submodules messagepack csharp src messagepack unityclient assets scripts messagepack messagepackreader integers cs file not found mvc source aspnetcore src submodules messagepack csharp src messagepack unityclient assets scripts messagepack messagepackserializationexception cs fix symlinks bat in messagapack unityclient doesn t fix this could anybody help me with it what i missed my steps are commit restore cmd build cmd thank you
1
7,378
6,928,076,459
IssuesEvent
2017-12-01 02:24:03
codeforboston/cliff-effects
https://api.github.com/repos/codeforboston/cliff-effects
closed
Stabilize Conditional Logic for Header
infrastructure
Currently the `Header` file has a [hacky line](https://github.com/codeforboston/cliff-effects/blob/dev/src/components/Header.js#L14) that makes it only display if the current path is '/'. This is pretty fragile because changing the path can cause the header to display on the home page. Action item here is to implement a more robust toggle for the header visibility.
1.0
Stabilize Conditional Logic for Header - Currently the `Header` file has a [hacky line](https://github.com/codeforboston/cliff-effects/blob/dev/src/components/Header.js#L14) that makes it only display if the current path is '/'. This is pretty fragile because changing the path can cause the header to display on the home page. Action item here is to implement a more robust toggle for the header visibility.
infrastructure
stabilize conditional logic for header currently the header file has a that makes it only display if the current path is this is pretty fragile because changing the path can cause the header to display on the home page action item here is to implement a more robust toggle for the header visibility
1
11,768
9,418,522,409
IssuesEvent
2019-04-10 19:26:48
OpenLiberty/openliberty.io
https://api.github.com/repos/OpenLiberty/openliberty.io
closed
Automate tests for interactive guides
infrastructure
We need to explore what we can do arounds having automated tests for the interactive guides. My first gut feeling is to write some simple Selenium tests that we can run before we publish to openliberty.io.
1.0
Automate tests for interactive guides - We need to explore what we can do arounds having automated tests for the interactive guides. My first gut feeling is to write some simple Selenium tests that we can run before we publish to openliberty.io.
infrastructure
automate tests for interactive guides we need to explore what we can do arounds having automated tests for the interactive guides my first gut feeling is to write some simple selenium tests that we can run before we publish to openliberty io
1
16,542
12,032,868,915
IssuesEvent
2020-04-13 13:08:33
spockframework/spock
https://api.github.com/repos/spockframework/spock
opened
Test Spock with Java 13 and 14
infrastructure
Spock 2 should work with Java 14 (having Groovy 2.5.10 and 3.0.x) on a classpath, but It would be good to run internal Spock tests with with Java 13 and 14 on CI server. It is not impossible to run tests with Java 14 with older Gradle versions, but Gradle 6.3 is required for running Java 14 in general, so the upgrade is the most straightforward way to go. Unfortunately it is currently blocked by the Gradle OSGI plugin removal which has to be replaced with something else (if Spock 2 would keep OSGI manifest support at all) - #1139
1.0
Test Spock with Java 13 and 14 - Spock 2 should work with Java 14 (having Groovy 2.5.10 and 3.0.x) on a classpath, but It would be good to run internal Spock tests with with Java 13 and 14 on CI server. It is not impossible to run tests with Java 14 with older Gradle versions, but Gradle 6.3 is required for running Java 14 in general, so the upgrade is the most straightforward way to go. Unfortunately it is currently blocked by the Gradle OSGI plugin removal which has to be replaced with something else (if Spock 2 would keep OSGI manifest support at all) - #1139
infrastructure
test spock with java and spock should work with java having groovy and x on a classpath but it would be good to run internal spock tests with with java and on ci server it is not impossible to run tests with java with older gradle versions but gradle is required for running java in general so the upgrade is the most straightforward way to go unfortunately it is currently blocked by the gradle osgi plugin removal which has to be replaced with something else if spock would keep osgi manifest support at all
1
20,829
31,150,009,829
IssuesEvent
2023-08-16 09:18:12
sebastianbergmann/phpunit
https://api.github.com/repos/sebastianbergmann/phpunit
opened
Remove `TestCase::assertStringNotMatchesFormat()` and `TestCase::assertStringNotMatchesFormatFile()`
type/backward-compatibility feature/assertion
Follows #5472
True
Remove `TestCase::assertStringNotMatchesFormat()` and `TestCase::assertStringNotMatchesFormatFile()` - Follows #5472
non_infrastructure
remove testcase assertstringnotmatchesformat and testcase assertstringnotmatchesformatfile follows
0
12,040
9,556,162,145
IssuesEvent
2019-05-03 07:24:49
ImpressCMS/impresscms
https://api.github.com/repos/ImpressCMS/impresscms
closed
Add Slack badge to README
enhancement infrastructure
I think Slack badge on https://github.com/GitbookIO/gitbook looks good. And I think it would be good to add something like this to our README. Maybe more users would join our chat system and maybe more users would be involved. Yes, I know that @fiammybe created another chat system on Gitter but we don't have so many users that I don't see any any point to have multiple chat systems for a while. I think Slack could be good enough for us for a while. This badge is driven by [Slackin](https://github.com/rauchg/slackin). I think we could host this on Heroku or similar service as free hosting. That's not hard. But before that I need to get custom slack API app token from @fiammybe (yeah that's why I added you to here for this task!)
1.0
Add Slack badge to README - I think Slack badge on https://github.com/GitbookIO/gitbook looks good. And I think it would be good to add something like this to our README. Maybe more users would join our chat system and maybe more users would be involved. Yes, I know that @fiammybe created another chat system on Gitter but we don't have so many users that I don't see any any point to have multiple chat systems for a while. I think Slack could be good enough for us for a while. This badge is driven by [Slackin](https://github.com/rauchg/slackin). I think we could host this on Heroku or similar service as free hosting. That's not hard. But before that I need to get custom slack API app token from @fiammybe (yeah that's why I added you to here for this task!)
infrastructure
add slack badge to readme i think slack badge on looks good and i think it would be good to add something like this to our readme maybe more users would join our chat system and maybe more users would be involved yes i know that fiammybe created another chat system on gitter but we don t have so many users that i don t see any any point to have multiple chat systems for a while i think slack could be good enough for us for a while this badge is driven by i think we could host this on heroku or similar service as free hosting that s not hard but before that i need to get custom slack api app token from fiammybe yeah that s why i added you to here for this task
1
31,943
26,263,484,244
IssuesEvent
2023-01-06 10:13:08
GitGuardian/ggshield
https://api.github.com/repos/GitGuardian/ggshield
closed
ggshield should not install its tests
bug status:: confirmed Infrastructure
## Environment - ggshield version: 1.13.5 - Operating system (Linux, macOS, Windows): any - Operating system version: any - Python version: 3.8.10 ## Describe the bug Installing ggshield with `pip install` installs the tests directory. Steps to reproduce: 1. Create and activate a virtualenv ``` python -m venv bug ``` 2. Check it contains no tests directory: ``` ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests ls: cannot access '/home/agateau/tmp/bug/lib/python3.8/site-packages/tests': No such file or directory ``` 3. Install ggshield: ``` pip install ~/src/ggshield ``` Actual result: ``` ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests cmd core iac output scan unit ``` Expected result: There should be no `tests` directory in the `site-packages` directory. ## Useful resources https://setuptools.pypa.io/en/latest/userguide/package_discovery.html ## TODO - [ ] Fix the issue - [ ] Setup a tool like [check-wheel-contents](https://pypi.org/project/check-wheel-contents/) in the CI so that this cannot happen again
1.0
ggshield should not install its tests - ## Environment - ggshield version: 1.13.5 - Operating system (Linux, macOS, Windows): any - Operating system version: any - Python version: 3.8.10 ## Describe the bug Installing ggshield with `pip install` installs the tests directory. Steps to reproduce: 1. Create and activate a virtualenv ``` python -m venv bug ``` 2. Check it contains no tests directory: ``` ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests ls: cannot access '/home/agateau/tmp/bug/lib/python3.8/site-packages/tests': No such file or directory ``` 3. Install ggshield: ``` pip install ~/src/ggshield ``` Actual result: ``` ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests cmd core iac output scan unit ``` Expected result: There should be no `tests` directory in the `site-packages` directory. ## Useful resources https://setuptools.pypa.io/en/latest/userguide/package_discovery.html ## TODO - [ ] Fix the issue - [ ] Setup a tool like [check-wheel-contents](https://pypi.org/project/check-wheel-contents/) in the CI so that this cannot happen again
infrastructure
ggshield should not install its tests environment ggshield version operating system linux macos windows any operating system version any python version describe the bug installing ggshield with pip install installs the tests directory steps to reproduce create and activate a virtualenv python m venv bug check it contains no tests directory ls virtual env lib site packages tests ls cannot access home agateau tmp bug lib site packages tests no such file or directory install ggshield pip install src ggshield actual result ls virtual env lib site packages tests cmd core iac output scan unit expected result there should be no tests directory in the site packages directory useful resources todo fix the issue setup a tool like in the ci so that this cannot happen again
1
290,745
8,903,452,750
IssuesEvent
2019-01-17 10:31:11
open-learning-exchange/planet
https://api.github.com/repos/open-learning-exchange/planet
closed
Adding a new course overwrites last edited course
priority
Specifically when you add an exam to the new course.
1.0
Adding a new course overwrites last edited course - Specifically when you add an exam to the new course.
non_infrastructure
adding a new course overwrites last edited course specifically when you add an exam to the new course
0
182,135
6,667,446,369
IssuesEvent
2017-10-03 12:37:55
spring-projects/spring-boot
https://api.github.com/repos/spring-projects/spring-boot
reopened
Make TomcatEmbeddedWebappClassLoader parallel capable
for: team-attention priority: normal type: enhancement
With reference to [DATAMONGO-1792](https://jira.spring.io/browse/DATAMONGO-1792), the synchronization of `TomcatEmbeddedWebappClassLoader.loadClass(String, boolean)` is inefficient. Rather than making the method itself `synchronized` it should synchronize on the response from `(getClassLoadingLock(className)`. We should also verify that the JVM considers the class loader to be parallel capable (its superclass is registered as such).
1.0
Make TomcatEmbeddedWebappClassLoader parallel capable - With reference to [DATAMONGO-1792](https://jira.spring.io/browse/DATAMONGO-1792), the synchronization of `TomcatEmbeddedWebappClassLoader.loadClass(String, boolean)` is inefficient. Rather than making the method itself `synchronized` it should synchronize on the response from `(getClassLoadingLock(className)`. We should also verify that the JVM considers the class loader to be parallel capable (its superclass is registered as such).
non_infrastructure
make tomcatembeddedwebappclassloader parallel capable with reference to the synchronization of tomcatembeddedwebappclassloader loadclass string boolean is inefficient rather than making the method itself synchronized it should synchronize on the response from getclassloadinglock classname we should also verify that the jvm considers the class loader to be parallel capable its superclass is registered as such
0
56,068
11,501,100,802
IssuesEvent
2020-02-12 16:36:13
pace-neutrons/Pace-Project-Plan
https://api.github.com/repos/pace-neutrons/Pace-Project-Plan
closed
Euphonic Calculations: API/Internal Data Structure Design: Review current/proposed designs
Euphonic Calculations Interfacing to 3rd Party Code
Hold a session to give an overview and review of Euphonic's current API/data structures Sub issue of #46 Will feed into #14, #15, #16, #17
1.0
Euphonic Calculations: API/Internal Data Structure Design: Review current/proposed designs - Hold a session to give an overview and review of Euphonic's current API/data structures Sub issue of #46 Will feed into #14, #15, #16, #17
non_infrastructure
euphonic calculations api internal data structure design review current proposed designs hold a session to give an overview and review of euphonic s current api data structures sub issue of will feed into
0
15,560
11,589,268,730
IssuesEvent
2020-02-24 01:16:32
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Detailed descriptions can be too long for the user interface
interface/infrastructure newfeature
The "Description" attribute of properties is sometimes overly long when used in the context of the user interface. For example, SWCON currently has the description: "Fractional amount of water above DUL that can drain under gravity per day (SWCON) (/d)". That's all useful information, but it's rather long when used as a column header in a table. Further, in the SoilWater input views, the units attribute is also added automatically, so the header ends up with "(/d) (/d)" at the end. I'd welcome discussion about the best way to address these problems. As a starter, I'd suggest that units of measurement not be included in the Description attribute, but be provided only in the Units attribute; having it in both is just redundant. I think it might also be useful to have a "ShortDescription" (or "Label") attribute, and a "LongDescription" attribute, The short version could be used in the column headers of the GUI, and perhaps the long version could be displayed when the user requests to see it. For example, the header might just read "SWCON", but the user would be able to see the longer, detailed description by some mechanism (perhaps by mouse over, or perhaps from a popup-menu item) Any comments?
1.0
Detailed descriptions can be too long for the user interface - The "Description" attribute of properties is sometimes overly long when used in the context of the user interface. For example, SWCON currently has the description: "Fractional amount of water above DUL that can drain under gravity per day (SWCON) (/d)". That's all useful information, but it's rather long when used as a column header in a table. Further, in the SoilWater input views, the units attribute is also added automatically, so the header ends up with "(/d) (/d)" at the end. I'd welcome discussion about the best way to address these problems. As a starter, I'd suggest that units of measurement not be included in the Description attribute, but be provided only in the Units attribute; having it in both is just redundant. I think it might also be useful to have a "ShortDescription" (or "Label") attribute, and a "LongDescription" attribute, The short version could be used in the column headers of the GUI, and perhaps the long version could be displayed when the user requests to see it. For example, the header might just read "SWCON", but the user would be able to see the longer, detailed description by some mechanism (perhaps by mouse over, or perhaps from a popup-menu item) Any comments?
infrastructure
detailed descriptions can be too long for the user interface the description attribute of properties is sometimes overly long when used in the context of the user interface for example swcon currently has the description fractional amount of water above dul that can drain under gravity per day swcon d that s all useful information but it s rather long when used as a column header in a table further in the soilwater input views the units attribute is also added automatically so the header ends up with d d at the end i d welcome discussion about the best way to address these problems as a starter i d suggest that units of measurement not be included in the description attribute but be provided only in the units attribute having it in both is just redundant i think it might also be useful to have a shortdescription or label attribute and a longdescription attribute the short version could be used in the column headers of the gui and perhaps the long version could be displayed when the user requests to see it for example the header might just read swcon but the user would be able to see the longer detailed description by some mechanism perhaps by mouse over or perhaps from a popup menu item any comments
1
27,678
22,148,105,849
IssuesEvent
2022-06-03 14:04:38
enarx/enarx
https://api.github.com/repos/enarx/enarx
closed
[Infrastructure]: Release improvements
infrastructure triage release
### Is there an existing issue for this? - [X] I have searched the existing issues ### Code of Conduct - [X] I agree to follow this project's Code of Conduct ### Description ## `helper/release_enarx.sh` - Bugs - lower case 'release 0.5.1' to make sure that commit checker passes - lower case 'release' for PR to make sure commit checker passes - Sleep 60 seconds between published crates to avoid race condition - Features - fast mode to run single image - dry-run false always holds on to artifacts/repo - maybe jump to release and skip checks. scream like a banshee and check if you want to do this
1.0
[Infrastructure]: Release improvements - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Code of Conduct - [X] I agree to follow this project's Code of Conduct ### Description ## `helper/release_enarx.sh` - Bugs - lower case 'release 0.5.1' to make sure that commit checker passes - lower case 'release' for PR to make sure commit checker passes - Sleep 60 seconds between published crates to avoid race condition - Features - fast mode to run single image - dry-run false always holds on to artifacts/repo - maybe jump to release and skip checks. scream like a banshee and check if you want to do this
infrastructure
release improvements is there an existing issue for this i have searched the existing issues code of conduct i agree to follow this project s code of conduct description helper release enarx sh bugs lower case release to make sure that commit checker passes lower case release for pr to make sure commit checker passes sleep seconds between published crates to avoid race condition features fast mode to run single image dry run false always holds on to artifacts repo maybe jump to release and skip checks scream like a banshee and check if you want to do this
1
283,343
30,913,274,459
IssuesEvent
2023-08-05 01:31:24
hshivhare67/kernel_v4.19.72_CVE-2022-42896_new
https://api.github.com/repos/hshivhare67/kernel_v4.19.72_CVE-2022-42896_new
reopened
CVE-2021-29154 (High) detected in linuxlinux-4.19.279
Mend: dependency security vulnerability
## CVE-2021-29154 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> BPF JIT compilers in the Linux kernel through 5.11.12 have incorrect computation of branch displacements, allowing them to execute arbitrary code within the kernel context. This affects arch/x86/net/bpf_jit_comp.c and arch/x86/net/bpf_jit_comp32.c. <p>Publish Date: 2021-04-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29154>CVE-2021-29154</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-29154">https://www.linuxkernelcves.com/cves/CVE-2021-29154</a></p> <p>Release Date: 2021-04-08</p> <p>Fix Resolution: v4.4.266, v4.9.266, v4.14.230, v4.19.186, v5.4.111, v5.10.29, v5.11.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29154 (High) detected in linuxlinux-4.19.279 - ## CVE-2021-29154 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> BPF JIT compilers in the Linux kernel through 5.11.12 have incorrect computation of branch displacements, allowing them to execute arbitrary code within the kernel context. This affects arch/x86/net/bpf_jit_comp.c and arch/x86/net/bpf_jit_comp32.c. <p>Publish Date: 2021-04-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29154>CVE-2021-29154</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-29154">https://www.linuxkernelcves.com/cves/CVE-2021-29154</a></p> <p>Release Date: 2021-04-08</p> <p>Fix Resolution: v4.4.266, v4.9.266, v4.14.230, v4.19.186, v5.4.111, v5.10.29, v5.11.13</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_infrastructure
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files vulnerability details bpf jit compilers in the linux kernel through have incorrect computation of branch displacements allowing them to execute arbitrary code within the kernel context this affects arch net bpf jit comp c and arch net bpf jit c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
9,326
7,921,732,065
IssuesEvent
2018-07-05 08:33:02
ressec/hemajoo-foundation
https://api.github.com/repos/ressec/hemajoo-foundation
closed
Travis build for hemajoo-foundation fails
Bug Critical Development Infrastructure
## Cause The error reported in the build log is: `Caused by: java.lang.ClassNotFoundException: org.apache.maven.doxia.siterenderer.DocumentContent` ## Resolution Resolved by changing in the parent pom the version number of the Maven site plugin: ``` <plugin> <groupId>org.apache.maven.plugins</groupId> <artifactId>maven-site-plugin</artifactId> <version>3.7.1</version> . . . </plugin> ```
1.0
Travis build for hemajoo-foundation fails - ## Cause The error reported in the build log is: `Caused by: java.lang.ClassNotFoundException: org.apache.maven.doxia.siterenderer.DocumentContent` ## Resolution Resolved by changing in the parent pom the version number of the Maven site plugin: ``` <plugin> <groupId>org.apache.maven.plugins</groupId> <artifactId>maven-site-plugin</artifactId> <version>3.7.1</version> . . . </plugin> ```
infrastructure
travis build for hemajoo foundation fails cause the error reported in the build log is caused by java lang classnotfoundexception org apache maven doxia siterenderer documentcontent resolution resolved by changing in the parent pom the version number of the maven site plugin org apache maven plugins maven site plugin
1
64,324
26,689,284,338
IssuesEvent
2023-01-27 02:08:10
cityofaustin/atd-data-tech
https://api.github.com/repos/cityofaustin/atd-data-tech
closed
Project: Placemaking Program
Service: Apps Epic Workgroup: SMO Product: Smart Mobility Office Project Index Project: Placemaking Program
Intake and tracking platform that aids evaluation, milestone management, and a directory for Artists and Art Partners. ## Background The importance and need of developing a management methodology through Placemaking is essential to have a user-friendly interface to manage processes, collaborate seamlessly with team members, automate tasks, and track key milestones. This becomes important as placemaking sub-programs are expanding and becoming more complex with limited resources. Additionally, there have been some complications with the current application form and intake process currently in place. The intake form currently places all submissions together, rather than separating each intake application. ## Scope & Deliverables EOI intake form, evaluation platform with key milestones for projects, and directory to track ## Desired Outcomes A user-friendly platform that can be used by internal and external agencies to track key milestones. The ability to track beyond when applications come in, but the need for lengthy community outreach before installation is scheduled. ## Timebox November-December 2022 ## Concepts and references [Google form](https://api.knack.com/v1/applications/595d00ebd315cc4cb98daff4/download/asset/630fa509bb94060021f36aa7/austintransportation_smartmobilityofficegoogleforms.pdf) [Artbox website](https://www.austintexas.gov/page/artbox-program) ![Placemaking.png](https://images.zenhubusercontent.com/5d68810b05d7792664be403f/3740b13c-f18b-474d-aa27-7a0d84b1bb55) --- This Github issue represents a project of Austin Transportation's [Data & Technology Services](https://austinmobility.io/) team. Project status is documented regularly in the comments below. <!-- Don't forget to: - Add a new "Project:" label here: https://github.com/cityofaustin/atd-data-tech/labels. Use the hex code #3D3D3D. - Add a project evaluation here: https://atd.knack.com/dts#project-evaluation/ -->
1.0
Project: Placemaking Program - Intake and tracking platform that aids evaluation, milestone management, and a directory for Artists and Art Partners. ## Background The importance and need of developing a management methodology through Placemaking is essential to have a user-friendly interface to manage processes, collaborate seamlessly with team members, automate tasks, and track key milestones. This becomes important as placemaking sub-programs are expanding and becoming more complex with limited resources. Additionally, there have been some complications with the current application form and intake process currently in place. The intake form currently places all submissions together, rather than separating each intake application. ## Scope & Deliverables EOI intake form, evaluation platform with key milestones for projects, and directory to track ## Desired Outcomes A user-friendly platform that can be used by internal and external agencies to track key milestones. The ability to track beyond when applications come in, but the need for lengthy community outreach before installation is scheduled. ## Timebox November-December 2022 ## Concepts and references [Google form](https://api.knack.com/v1/applications/595d00ebd315cc4cb98daff4/download/asset/630fa509bb94060021f36aa7/austintransportation_smartmobilityofficegoogleforms.pdf) [Artbox website](https://www.austintexas.gov/page/artbox-program) ![Placemaking.png](https://images.zenhubusercontent.com/5d68810b05d7792664be403f/3740b13c-f18b-474d-aa27-7a0d84b1bb55) --- This Github issue represents a project of Austin Transportation's [Data & Technology Services](https://austinmobility.io/) team. Project status is documented regularly in the comments below. <!-- Don't forget to: - Add a new "Project:" label here: https://github.com/cityofaustin/atd-data-tech/labels. Use the hex code #3D3D3D. - Add a project evaluation here: https://atd.knack.com/dts#project-evaluation/ -->
non_infrastructure
project placemaking program intake and tracking platform that aids evaluation milestone management and a directory for artists and art partners background the importance and need of developing a management methodology through placemaking is essential to have a user friendly interface to manage processes collaborate seamlessly with team members automate tasks and track key milestones this becomes important as placemaking sub programs are expanding and becoming more complex with limited resources additionally there have been some complications with the current application form and intake process currently in place the intake form currently places all submissions together rather than separating each intake application scope deliverables eoi intake form evaluation platform with key milestones for projects and directory to track desired outcomes a user friendly platform that can be used by internal and external agencies to track key milestones the ability to track beyond when applications come in but the need for lengthy community outreach before installation is scheduled timebox november december concepts and references this github issue represents a project of austin transportation s team project status is documented regularly in the comments below don t forget to add a new project label here use the hex code add a project evaluation here
0
209,219
16,187,586,786
IssuesEvent
2021-05-04 00:40:27
google/iree
https://api.github.com/repos/google/iree
reopened
[New documentation] Write docs for "Building Python bindings"
bindings/python documentation ✏️
* Live preview of this page: https://scotttodd.github.io/iree/building-from-source/python/, which will eventually live on https://google.github.io/iree/ after we switch over the docs. * Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/building-from-source/python.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though. * See the README here for instructions on previewing changes locally: https://github.com/google/iree/tree/staging-docs-rework/mkdocs * mkdocs + mkdocs material offer a bunch of nice features, like https://squidfunk.github.io/mkdocs-material/reference/code-blocks/ (line numbers, highlighting specific lines, etc.) I'd like for this page to aggregate the "building from source" aspects of our current Python pages: * https://google.github.io/iree/get-started/getting-started-python * https://google.github.io/iree/get-started/getting-started-tensorflow * https://google.github.io/iree/developing-iree/tensorflow-integrations (maybe) This will be linked from the "Python bindings" page (preview: https://scotttodd.github.io/iree/bindings/python/) and possibly from the Python frontend pages, in cases where users will benefit from building from source instead of installing from our published packages.
1.0
[New documentation] Write docs for "Building Python bindings" - * Live preview of this page: https://scotttodd.github.io/iree/building-from-source/python/, which will eventually live on https://google.github.io/iree/ after we switch over the docs. * Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/building-from-source/python.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though. * See the README here for instructions on previewing changes locally: https://github.com/google/iree/tree/staging-docs-rework/mkdocs * mkdocs + mkdocs material offer a bunch of nice features, like https://squidfunk.github.io/mkdocs-material/reference/code-blocks/ (line numbers, highlighting specific lines, etc.) I'd like for this page to aggregate the "building from source" aspects of our current Python pages: * https://google.github.io/iree/get-started/getting-started-python * https://google.github.io/iree/get-started/getting-started-tensorflow * https://google.github.io/iree/developing-iree/tensorflow-integrations (maybe) This will be linked from the "Python bindings" page (preview: https://scotttodd.github.io/iree/bindings/python/) and possibly from the Python frontend pages, in cases where users will benefit from building from source instead of installing from our published packages.
non_infrastructure
write docs for building python bindings live preview of this page which will eventually live on after we switch over the docs source currently lives on the staging docs rework branch here prs or direct commits to the branch are okay we ll want to be careful about force pushes rebases though see the readme here for instructions on previewing changes locally mkdocs mkdocs material offer a bunch of nice features like line numbers highlighting specific lines etc i d like for this page to aggregate the building from source aspects of our current python pages maybe this will be linked from the python bindings page preview and possibly from the python frontend pages in cases where users will benefit from building from source instead of installing from our published packages
0
8,902
7,724,934,260
IssuesEvent
2018-05-24 16:22:58
GoogleCloudPlatform/forseti-security
https://api.github.com/repos/GoogleCloudPlatform/forseti-security
closed
Move cron lock file out of /tmp (will be cleaned up by system)
module: infrastructure priority: p1 release-testing: 2.0 RC3 triaged: yes
The cron lock file is currrently placed in `/tmp`, which will be cleaned up by the system everyday. So, this should be moved somewhere else so that we avoid the chance that a cron job is running, and then the lock file gets cleaned up. ``` ubuntu@forseti-server-vm-4330:~/forseti-security$ ls /tmp SERVERENGINE_SOCKETMANAGER_2018-05-22T02:23:05Z_2743 deployment.log forseti_cron_runner.lock ```
1.0
Move cron lock file out of /tmp (will be cleaned up by system) - The cron lock file is currrently placed in `/tmp`, which will be cleaned up by the system everyday. So, this should be moved somewhere else so that we avoid the chance that a cron job is running, and then the lock file gets cleaned up. ``` ubuntu@forseti-server-vm-4330:~/forseti-security$ ls /tmp SERVERENGINE_SOCKETMANAGER_2018-05-22T02:23:05Z_2743 deployment.log forseti_cron_runner.lock ```
infrastructure
move cron lock file out of tmp will be cleaned up by system the cron lock file is currrently placed in tmp which will be cleaned up by the system everyday so this should be moved somewhere else so that we avoid the chance that a cron job is running and then the lock file gets cleaned up ubuntu forseti server vm forseti security ls tmp serverengine socketmanager deployment log forseti cron runner lock
1
532,386
15,555,255,049
IssuesEvent
2021-03-16 05:45:42
AY2021S2-CS2113-F10-2/tp
https://api.github.com/repos/AY2021S2-CS2113-F10-2/tp
closed
H-13: Add classes for basic framework
priority.High type.Task
Add classes including User, FitCenter, RecordList, Record, and enum RecordType
1.0
H-13: Add classes for basic framework - Add classes including User, FitCenter, RecordList, Record, and enum RecordType
non_infrastructure
h add classes for basic framework add classes including user fitcenter recordlist record and enum recordtype
0
25,196
18,251,282,760
IssuesEvent
2021-10-02 08:45:32
kgorbenko/SeasonStats
https://api.github.com/repos/kgorbenko/SeasonStats
closed
Configure ESLint
enhancement infrastructure
We need to audit default `create-react-app` ESLint config and decide if we need to somehow change it.
1.0
Configure ESLint - We need to audit default `create-react-app` ESLint config and decide if we need to somehow change it.
infrastructure
configure eslint we need to audit default create react app eslint config and decide if we need to somehow change it
1
131,435
10,696,062,294
IssuesEvent
2019-10-23 14:07:06
fedora-infra/bodhi
https://api.github.com/repos/fedora-infra/bodhi
closed
Convert bodhi/tests/client/test___init__.py to PyTest
EasyFix Low Priority Tests
[bodhi/tests/client/test___init__.py](https://github.com/fedora-infra/bodhi/blob/develop/bodhi/tests/client/test___init__.py) still contains classes that derive from `unittest.TestCase`. We use PyTest, so they could be migrated to not derive from `unittest.TestCase`, or derive from `bodhi.tests.server.base.BasePyTestCase` in case they need to use the test database or WSGI app. This also allows the use of the simpler assert constructs instead of the self.assert*() methods. Instead of unittest `setUp()` or `tearDown()` methods, PyTest uses `setup_method(...)` or `teardown_method(...)`, or analog methods to setup/teardown stuff for different scopes. The following PR is an example of the changes that might need to be done to this file: #3612
1.0
Convert bodhi/tests/client/test___init__.py to PyTest - [bodhi/tests/client/test___init__.py](https://github.com/fedora-infra/bodhi/blob/develop/bodhi/tests/client/test___init__.py) still contains classes that derive from `unittest.TestCase`. We use PyTest, so they could be migrated to not derive from `unittest.TestCase`, or derive from `bodhi.tests.server.base.BasePyTestCase` in case they need to use the test database or WSGI app. This also allows the use of the simpler assert constructs instead of the self.assert*() methods. Instead of unittest `setUp()` or `tearDown()` methods, PyTest uses `setup_method(...)` or `teardown_method(...)`, or analog methods to setup/teardown stuff for different scopes. The following PR is an example of the changes that might need to be done to this file: #3612
non_infrastructure
convert bodhi tests client test init py to pytest still contains classes that derive from unittest testcase we use pytest so they could be migrated to not derive from unittest testcase or derive from bodhi tests server base basepytestcase in case they need to use the test database or wsgi app this also allows the use of the simpler assert constructs instead of the self assert methods instead of unittest setup or teardown methods pytest uses setup method or teardown method or analog methods to setup teardown stuff for different scopes the following pr is an example of the changes that might need to be done to this file
0
33,873
27,963,294,548
IssuesEvent
2023-03-24 17:15:59
mitodl/ol-infrastructure
https://api.github.com/repos/mitodl/ol-infrastructure
opened
Send Fastly logs into Grafana for debugging and alerting
Monitoring and Alerting product:ocw product:infrastructure
# User Story - As a platform engineer I would like to be alerted when a Fastly property is experiencing a spike of error conditions - As an application engineer I would like to be able to diagnose errors all the way from Fastly through to the backend system # Description/Context We need a path for sending Fastly logs into a data source connected to Grafana. The most likely destination is to use Vector as a router for proxying log data into the Grana cloud managed Loki storage system. # Acceptance Criteria - [ ] Fastly properties all log as JSON to S3 with a standard message schema - [ ] Fastly logs are sent to Loki via a Vector proxy using HTTPS or Syslog inputs - [ ] Vector rejects log data generated by bot traffic as identified by user agent information - [ ] Fastly logs are available as a data source in Grafana
1.0
Send Fastly logs into Grafana for debugging and alerting - # User Story - As a platform engineer I would like to be alerted when a Fastly property is experiencing a spike of error conditions - As an application engineer I would like to be able to diagnose errors all the way from Fastly through to the backend system # Description/Context We need a path for sending Fastly logs into a data source connected to Grafana. The most likely destination is to use Vector as a router for proxying log data into the Grana cloud managed Loki storage system. # Acceptance Criteria - [ ] Fastly properties all log as JSON to S3 with a standard message schema - [ ] Fastly logs are sent to Loki via a Vector proxy using HTTPS or Syslog inputs - [ ] Vector rejects log data generated by bot traffic as identified by user agent information - [ ] Fastly logs are available as a data source in Grafana
infrastructure
send fastly logs into grafana for debugging and alerting user story as a platform engineer i would like to be alerted when a fastly property is experiencing a spike of error conditions as an application engineer i would like to be able to diagnose errors all the way from fastly through to the backend system description context we need a path for sending fastly logs into a data source connected to grafana the most likely destination is to use vector as a router for proxying log data into the grana cloud managed loki storage system acceptance criteria fastly properties all log as json to with a standard message schema fastly logs are sent to loki via a vector proxy using https or syslog inputs vector rejects log data generated by bot traffic as identified by user agent information fastly logs are available as a data source in grafana
1
47,257
10,059,308,245
IssuesEvent
2019-07-22 15:45:39
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
Release Namex-July 15-Phase1
Namex Code With Us
Prior to moving into production, the following steps should be completed or confirmed: - [x] Checked devop channels for upgrades, patching - [x] Update version number in the code being released (Katie did this already) - [x] Create a draft release in GitHub and confirm the correct commits are present (DONE) - [x] Add version # and release # to this ticket - DONE. - [x] schedule the release with staff/clients (daytime's best or when staff are available for rollback)-DONE - [x] All dev/test pipeline test suites green (none for nae-examination and no e2e ready yet.) - [x] Tag prod to previous-prod if pipeline cant be used. - [x] Tag prod to test if pipeline can be used to deploy. - [x] Release the code to production and complete smoke test - [x] Finalise/publish the release in GitHub, tagging it **NameX Smoke Test Script** - [x] Log in - [x] Go to search - [x] Filter for last 30 days - [x] Go to ZZZ record (usually on Hold) - [x] edit name 1, click save - [x] edit address line 1 + comment, click save - [x] check NRO/oracle tables for changes - [x] reject name 1 with a conflict - [x] approve name 2 with quick approve - [x] wait 5 minutes, then check in NRO - [x] reload NR in namex - [x] reset NR - [x] undo all decisions - [x] change name 1 - [x] do a conditional approval use mouse to verify behaviour - [x] RESET - [x] reject name choice 1 and 2 and quick approve name choice 3 - [x] undo all decisions - [x] set applicant info back to original values - [x] put it on hold - [x] watch [Kibana](https://kibana.pathfinder.gov.bc.ca/) logs for 5 minutes
1.0
Release Namex-July 15-Phase1 - Prior to moving into production, the following steps should be completed or confirmed: - [x] Checked devop channels for upgrades, patching - [x] Update version number in the code being released (Katie did this already) - [x] Create a draft release in GitHub and confirm the correct commits are present (DONE) - [x] Add version # and release # to this ticket - DONE. - [x] schedule the release with staff/clients (daytime's best or when staff are available for rollback)-DONE - [x] All dev/test pipeline test suites green (none for nae-examination and no e2e ready yet.) - [x] Tag prod to previous-prod if pipeline cant be used. - [x] Tag prod to test if pipeline can be used to deploy. - [x] Release the code to production and complete smoke test - [x] Finalise/publish the release in GitHub, tagging it **NameX Smoke Test Script** - [x] Log in - [x] Go to search - [x] Filter for last 30 days - [x] Go to ZZZ record (usually on Hold) - [x] edit name 1, click save - [x] edit address line 1 + comment, click save - [x] check NRO/oracle tables for changes - [x] reject name 1 with a conflict - [x] approve name 2 with quick approve - [x] wait 5 minutes, then check in NRO - [x] reload NR in namex - [x] reset NR - [x] undo all decisions - [x] change name 1 - [x] do a conditional approval use mouse to verify behaviour - [x] RESET - [x] reject name choice 1 and 2 and quick approve name choice 3 - [x] undo all decisions - [x] set applicant info back to original values - [x] put it on hold - [x] watch [Kibana](https://kibana.pathfinder.gov.bc.ca/) logs for 5 minutes
non_infrastructure
release namex july prior to moving into production the following steps should be completed or confirmed checked devop channels for upgrades patching update version number in the code being released katie did this already create a draft release in github and confirm the correct commits are present done add version and release to this ticket done schedule the release with staff clients daytime s best or when staff are available for rollback done all dev test pipeline test suites green none for nae examination and no ready yet tag prod to previous prod if pipeline cant be used tag prod to test if pipeline can be used to deploy release the code to production and complete smoke test finalise publish the release in github tagging it namex smoke test script log in go to search filter for last days go to zzz record usually on hold edit name click save edit address line comment click save check nro oracle tables for changes reject name with a conflict approve name with quick approve wait minutes then check in nro reload nr in namex reset nr undo all decisions change name do a conditional approval use mouse to verify behaviour reset reject name choice and and quick approve name choice undo all decisions set applicant info back to original values put it on hold watch logs for minutes
0
32,132
26,464,245,057
IssuesEvent
2023-01-16 21:11:30
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Report cannot handle overloaded methods
bug interface/infrastructure
Report throws "Ambiguous" error when trying to report an overloaded method.
1.0
Report cannot handle overloaded methods - Report throws "Ambiguous" error when trying to report an overloaded method.
infrastructure
report cannot handle overloaded methods report throws ambiguous error when trying to report an overloaded method
1
19,675
13,387,522,061
IssuesEvent
2020-09-02 16:06:23
jmcgeheeiv/pyfakefs
https://api.github.com/repos/jmcgeheeiv/pyfakefs
closed
Tests failing on Ubuntu 20.04 (focal) with python 3.8.2
infrastructure
I'm trying to package pyfakefs for Ubuntu focal and the build on launchpad fails. I have reproduced the same error on a Virtual Box Ubuntu: a bunch of python3 related packages were already installed for another package (debhelper, dh-python, python3-all, python3-setuptools, tox, python3-pytest, python3-pytest-cov), so I just git cloned the repo (but I also tested with the latest release) and in master branch, ran `pytest-3`. I've pasted the results [here](https://pastebin.com/6i7dGi0B). You can check the build log, that contains the same tests errors [here](https://launchpadlibrarian.net/494885084/buildlog_ubuntu-focal-amd64.python3-pyfakefs_4.1.0-0ubuntu1_BUILDING.txt.gz) (quite at bottom of the page). **Environment** ``` nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import platform; print(platform.platform())" Linux-5.4.0-42-generic-x86_64-with-glibc2.29 nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import sys; print('Python', sys.version)" Python 3.8.2 (default, Jul 16 2020, 14:00:26) [GCC 9.3.0] nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "from pyfakefs.fake_filesystem import __version__; print('pyfakefs', __version__)" pyfakefs 4.1.0 nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ ```
1.0
Tests failing on Ubuntu 20.04 (focal) with python 3.8.2 - I'm trying to package pyfakefs for Ubuntu focal and the build on launchpad fails. I have reproduced the same error on a Virtual Box Ubuntu: a bunch of python3 related packages were already installed for another package (debhelper, dh-python, python3-all, python3-setuptools, tox, python3-pytest, python3-pytest-cov), so I just git cloned the repo (but I also tested with the latest release) and in master branch, ran `pytest-3`. I've pasted the results [here](https://pastebin.com/6i7dGi0B). You can check the build log, that contains the same tests errors [here](https://launchpadlibrarian.net/494885084/buildlog_ubuntu-focal-amd64.python3-pyfakefs_4.1.0-0ubuntu1_BUILDING.txt.gz) (quite at bottom of the page). **Environment** ``` nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import platform; print(platform.platform())" Linux-5.4.0-42-generic-x86_64-with-glibc2.29 nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import sys; print('Python', sys.version)" Python 3.8.2 (default, Jul 16 2020, 14:00:26) [GCC 9.3.0] nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "from pyfakefs.fake_filesystem import __version__; print('pyfakefs', __version__)" pyfakefs 4.1.0 nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ ```
infrastructure
tests failing on ubuntu focal with python i m trying to package pyfakefs for ubuntu focal and the build on launchpad fails i have reproduced the same error on a virtual box ubuntu a bunch of related packages were already installed for another package debhelper dh python all setuptools tox pytest pytest cov so i just git cloned the repo but i also tested with the latest release and in master branch ran pytest i ve pasted the results you can check the build log that contains the same tests errors quite at bottom of the page environment nico dev pyfakefs python c import platform print platform platform linux generic with nico dev pyfakefs python c import sys print python sys version python default jul nico dev pyfakefs python c from pyfakefs fake filesystem import version print pyfakefs version pyfakefs nico dev pyfakefs
1
29,046
23,690,430,307
IssuesEvent
2022-08-29 10:16:12
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
opened
Optimize agents using
infrastructure
**Description** Now build of UTBot Java requires many agents: up to 29. 27 jobs in this build can be ran as concurrent. All of these jobs are doing repeatable task: building UTBot Java project. The task takes up to 7 minutes. [GitHub allocates up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) for the organization at the same time. As the result, we don't use allocated resources effectively. For the efficient use of resources we can reduce the number of concurrent jobs. Some of the jobs are doing nothing time-consuming, thus, they can be combined with another jobs. **Expected behavior** Some of the jobs are combined together with another to provide efficient use of allocated runners. **Environment** GitHub Actions **Potential alternatives** No alternatives. **Context** Not applicable.
1.0
Optimize agents using - **Description** Now build of UTBot Java requires many agents: up to 29. 27 jobs in this build can be ran as concurrent. All of these jobs are doing repeatable task: building UTBot Java project. The task takes up to 7 minutes. [GitHub allocates up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) for the organization at the same time. As the result, we don't use allocated resources effectively. For the efficient use of resources we can reduce the number of concurrent jobs. Some of the jobs are doing nothing time-consuming, thus, they can be combined with another jobs. **Expected behavior** Some of the jobs are combined together with another to provide efficient use of allocated runners. **Environment** GitHub Actions **Potential alternatives** No alternatives. **Context** Not applicable.
infrastructure
optimize agents using description now build of utbot java requires many agents up to jobs in this build can be ran as concurrent all of these jobs are doing repeatable task building utbot java project the task takes up to minutes for the organization at the same time as the result we don t use allocated resources effectively for the efficient use of resources we can reduce the number of concurrent jobs some of the jobs are doing nothing time consuming thus they can be combined with another jobs expected behavior some of the jobs are combined together with another to provide efficient use of allocated runners environment github actions potential alternatives no alternatives context not applicable
1
89,787
15,837,539,989
IssuesEvent
2021-04-06 20:56:25
TIBCOSoftware/tci-flogo
https://api.github.com/repos/TIBCOSoftware/tci-flogo
closed
WS-2019-0063 (High) detected in js-yaml-3.7.0.tgz - autoclosed
security vulnerability
## WS-2019-0063 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p> <p> Dependency Hierarchy: - css-loader-0.28.11.tgz (Root Library) - cssnano-3.10.0.tgz - postcss-svgo-2.1.6.tgz - svgo-0.7.2.tgz - :x: **js-yaml-3.7.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: js-yaml - 3.13.1</p> </p> </details> <p></p>
True
WS-2019-0063 (High) detected in js-yaml-3.7.0.tgz - autoclosed - ## WS-2019-0063 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p> <p> Dependency Hierarchy: - css-loader-0.28.11.tgz (Root Library) - cssnano-3.10.0.tgz - postcss-svgo-2.1.6.tgz - svgo-0.7.2.tgz - :x: **js-yaml-3.7.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file. <p>Publish Date: 2019-04-05 <p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p> <p>Release Date: 2019-04-05</p> <p>Fix Resolution: js-yaml - 3.13.1</p> </p> </details> <p></p>
non_infrastructure
ws high detected in js yaml tgz autoclosed ws high severity vulnerability vulnerable library js yaml tgz yaml parser and serializer library home page a href dependency hierarchy css loader tgz root library cssnano tgz postcss svgo tgz svgo tgz x js yaml tgz vulnerable library vulnerability details js yaml prior to are vulnerable to code injection the load function may execute arbitrary code injected through a malicious yaml file publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution js yaml
0
33,674
27,706,920,005
IssuesEvent
2023-03-14 11:51:56
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
PR/issue labeling bot isn't labeling
area-Infrastructure
Nothing is getting automatically labeled by area, for at least the last day or two.
1.0
PR/issue labeling bot isn't labeling - Nothing is getting automatically labeled by area, for at least the last day or two.
infrastructure
pr issue labeling bot isn t labeling nothing is getting automatically labeled by area for at least the last day or two
1
26,420
20,103,419,792
IssuesEvent
2022-02-07 08:04:24
deckhouse/deckhouse
https://api.github.com/repos/deckhouse/deckhouse
closed
[node-manager] NodeUser multiple ssh keys
type/feature-request area/cluster-and-infrastructure
NodeUser CR adds system users on nodes, and only one ssh key per user is allowed.
1.0
[node-manager] NodeUser multiple ssh keys - NodeUser CR adds system users on nodes, and only one ssh key per user is allowed.
infrastructure
nodeuser multiple ssh keys nodeuser cr adds system users on nodes and only one ssh key per user is allowed
1
19,365
13,224,537,111
IssuesEvent
2020-08-17 19:20:18
algorand/indexer
https://api.github.com/repos/algorand/indexer
closed
Postgres Tx that can be ReadOnly should be
Infrastructure
in idb/postgres.go set TxOption.ReadOnly on all transactions that are read-only. Current 'read only' mode works on Postgres database users that only have read permissions on tables, but when the whole database is in read-only mode (a hot spare or follower replicated database) it fails on transactions that _might_ write.
1.0
Postgres Tx that can be ReadOnly should be - in idb/postgres.go set TxOption.ReadOnly on all transactions that are read-only. Current 'read only' mode works on Postgres database users that only have read permissions on tables, but when the whole database is in read-only mode (a hot spare or follower replicated database) it fails on transactions that _might_ write.
infrastructure
postgres tx that can be readonly should be in idb postgres go set txoption readonly on all transactions that are read only current read only mode works on postgres database users that only have read permissions on tables but when the whole database is in read only mode a hot spare or follower replicated database it fails on transactions that might write
1
582
2,775,996,000
IssuesEvent
2015-05-04 19:14:16
fsr-itse/1327
https://api.github.com/repos/fsr-itse/1327
closed
Provide libraries locally
infrastructure minor
The project should not import any libraries from Google APIs or any other remote server.
1.0
Provide libraries locally - The project should not import any libraries from Google APIs or any other remote server.
infrastructure
provide libraries locally the project should not import any libraries from google apis or any other remote server
1
29,613
24,107,205,390
IssuesEvent
2022-09-20 08:27:51
UnitTestBot/UTBotJava
https://api.github.com/repos/UnitTestBot/UTBotJava
closed
Reduce concurrent jobs number for projects
infrastructure
**Description** Now tests for every separate project run in separate job. These jobs are concurrent. [GitHub allocates for organization up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) that can run at the same time. **Expected behavior** Reduce concurrent jobs number for projects by combining some projects builds together. **Environment** GitHub Actions. **Potential alternatives** No alternatives. **Context** No context.
1.0
Reduce concurrent jobs number for projects - **Description** Now tests for every separate project run in separate job. These jobs are concurrent. [GitHub allocates for organization up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) that can run at the same time. **Expected behavior** Reduce concurrent jobs number for projects by combining some projects builds together. **Environment** GitHub Actions. **Potential alternatives** No alternatives. **Context** No context.
infrastructure
reduce concurrent jobs number for projects description now tests for every separate project run in separate job these jobs are concurrent that can run at the same time expected behavior reduce concurrent jobs number for projects by combining some projects builds together environment github actions potential alternatives no alternatives context no context
1
235,616
25,955,245,786
IssuesEvent
2022-12-18 05:41:41
Dima2022/easybuggy
https://api.github.com/repos/Dima2022/easybuggy
closed
CVE-2021-4104 (High) detected in log4j-1.2.13.jar - autoclosed
security vulnerability
## CVE-2021-4104 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary> <p>Log4j</p> <p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.13.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/easybuggy/commit/aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f">aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2021-12-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4104>CVE-2021-4104</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-4104">https://nvd.nist.gov/vuln/detail/CVE-2021-4104</a></p> <p>Release Date: 2021-12-14</p> <p>Fix Resolution: uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
True
CVE-2021-4104 (High) detected in log4j-1.2.13.jar - autoclosed - ## CVE-2021-4104 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary> <p>Log4j</p> <p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p> <p> Dependency Hierarchy: - :x: **log4j-1.2.13.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/easybuggy/commit/aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f">aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions. <p>Publish Date: 2021-12-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4104>CVE-2021-4104</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-4104">https://nvd.nist.gov/vuln/detail/CVE-2021-4104</a></p> <p>Release Date: 2021-12-14</p> <p>Fix Resolution: uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue
non_infrastructure
cve high detected in jar autoclosed cve high severity vulnerability vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details jmsappender in is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration the attacker can provide topicbindingname and topicconnectionfactorybindingname configurations causing jmsappender to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects when specifically configured to use jmsappender which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution uom parent module module uom se javadoc module parfait examples module manual si units javadoc module unit api module module unit api javadoc module parfait module module javadoc uom systems javadoc module uom lib javadoc module uom systems module module uom se module module uom lib module module parfait javadoc module pcp parfait agent module si units module module rescue worker helmet automatic remediation is available for this issue
0
130,679
27,744,418,329
IssuesEvent
2023-03-15 16:07:40
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
duplicate submarine in editor (again)
Bug Need more info Code
### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? once again, there's duplicate entries in the submarine loader ![image](https://user-images.githubusercontent.com/5103517/205137033-d54368e8-c3f8-46ad-b3d9-c5735e704eda.png) ![image](https://user-images.githubusercontent.com/5103517/205137058-22c037f4-2f55-4324-b4db-f10473081651.png) but only for this one. my other submarines are not duplicate ### Reproduction steps _No response_ ### Bug prevalence Happens every now and then ### Version 0.19.14.0 ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
1.0
duplicate submarine in editor (again) - ### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? once again, there's duplicate entries in the submarine loader ![image](https://user-images.githubusercontent.com/5103517/205137033-d54368e8-c3f8-46ad-b3d9-c5735e704eda.png) ![image](https://user-images.githubusercontent.com/5103517/205137058-22c037f4-2f55-4324-b4db-f10473081651.png) but only for this one. my other submarines are not duplicate ### Reproduction steps _No response_ ### Bug prevalence Happens every now and then ### Version 0.19.14.0 ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
non_infrastructure
duplicate submarine in editor again disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened once again there s duplicate entries in the submarine loader but only for this one my other submarines are not duplicate reproduction steps no response bug prevalence happens every now and then version no response which operating system did you encounter this bug on windows relevant error messages and crash reports no response
0
45,615
2,937,573,977
IssuesEvent
2015-07-01 03:18:21
isuPatches/RockThePatch
https://api.github.com/repos/isuPatches/RockThePatch
closed
GitHub API has a rate limit and messages that should be shown
Affected- Develop Bug Bug Severity - Priority 3 (medium) Bug Source - External Source Bug Type - Other Fixed In - Develop Found By Me Found In - Develop Resolution - Fixed Status - Fixed And Has A Commit
Sometimes there's a hiccup and I need to show the GitHub API message or an error dialog instead of attempting to still show the github issues for bugs and projects.
1.0
GitHub API has a rate limit and messages that should be shown - Sometimes there's a hiccup and I need to show the GitHub API message or an error dialog instead of attempting to still show the github issues for bugs and projects.
non_infrastructure
github api has a rate limit and messages that should be shown sometimes there s a hiccup and i need to show the github api message or an error dialog instead of attempting to still show the github issues for bugs and projects
0
13,320
10,209,188,033
IssuesEvent
2019-08-14 12:06:42
eclipse/antenna
https://api.github.com/repos/eclipse/antenna
closed
Java9 travis tests currently always fail with 10min timeout
bug infrastructure wontfix
### Summary of the Bug see e.g. - https://travis-ci.org/eclipse/antenna/jobs/550295410 ### Acceptance Criteria - [ ] travis tests no longer fail ### Definition of Done - [ ] Acceptance criteria fulfilled - [ ] A test case is created to reproduce the bug - [ ] A PR is created, the CI infrastructure reports green, the bug test case proves that bug is fixed - [ ] The PR is reviewed and approved - [ ] No TODOs left in the code unless explained in the ticket, if something else is still open, this is summarized in a comment in the issue - [ ] Documentation is updated
1.0
Java9 travis tests currently always fail with 10min timeout - ### Summary of the Bug see e.g. - https://travis-ci.org/eclipse/antenna/jobs/550295410 ### Acceptance Criteria - [ ] travis tests no longer fail ### Definition of Done - [ ] Acceptance criteria fulfilled - [ ] A test case is created to reproduce the bug - [ ] A PR is created, the CI infrastructure reports green, the bug test case proves that bug is fixed - [ ] The PR is reviewed and approved - [ ] No TODOs left in the code unless explained in the ticket, if something else is still open, this is summarized in a comment in the issue - [ ] Documentation is updated
infrastructure
travis tests currently always fail with timeout summary of the bug see e g acceptance criteria travis tests no longer fail definition of done acceptance criteria fulfilled a test case is created to reproduce the bug a pr is created the ci infrastructure reports green the bug test case proves that bug is fixed the pr is reviewed and approved no todos left in the code unless explained in the ticket if something else is still open this is summarized in a comment in the issue documentation is updated
1
374,841
11,096,189,789
IssuesEvent
2019-12-16 10:37:46
JuPedSim/jpscore
https://api.github.com/repos/JuPedSim/jpscore
closed
Python integration test script
Priority: Medium Status: Ready Type: Enhancement
For the integration tests we need a "central" python script for the execution of the test cases. This script should sort the integration tests by approximate runtime in three different groups: slow, medium and fast. It should accept command-line parameters for setting the JPScore binary and setting the desired group (slow, medium, fast, all). This script should also manage where log files from the test execution are written. Additionally we need a command line parameter accepting a list of test to execute. This should be used to execute a single test in ctest/cmake. ACs: - Test script in python for executing integration tests - Command-line parameter for JPScore binary path - Grouped test depending on approximate runtime - Command line parameter for selecting the group to execute - Command line parameter to define a list of tests to execute (see also: #491)
1.0
Python integration test script - For the integration tests we need a "central" python script for the execution of the test cases. This script should sort the integration tests by approximate runtime in three different groups: slow, medium and fast. It should accept command-line parameters for setting the JPScore binary and setting the desired group (slow, medium, fast, all). This script should also manage where log files from the test execution are written. Additionally we need a command line parameter accepting a list of test to execute. This should be used to execute a single test in ctest/cmake. ACs: - Test script in python for executing integration tests - Command-line parameter for JPScore binary path - Grouped test depending on approximate runtime - Command line parameter for selecting the group to execute - Command line parameter to define a list of tests to execute (see also: #491)
non_infrastructure
python integration test script for the integration tests we need a central python script for the execution of the test cases this script should sort the integration tests by approximate runtime in three different groups slow medium and fast it should accept command line parameters for setting the jpscore binary and setting the desired group slow medium fast all this script should also manage where log files from the test execution are written additionally we need a command line parameter accepting a list of test to execute this should be used to execute a single test in ctest cmake acs test script in python for executing integration tests command line parameter for jpscore binary path grouped test depending on approximate runtime command line parameter for selecting the group to execute command line parameter to define a list of tests to execute see also
0
30,855
25,119,715,523
IssuesEvent
2022-11-09 06:55:43
astropy/astroquery
https://api.github.com/repos/astropy/astroquery
closed
Revise keyring version limitation in tests
testing infrastructure
Once this fix end up in a conda packaged version we may lift the current limitation of keyring to <12. https://github.com/jaraco/keyring/pull/320
1.0
Revise keyring version limitation in tests - Once this fix end up in a conda packaged version we may lift the current limitation of keyring to <12. https://github.com/jaraco/keyring/pull/320
infrastructure
revise keyring version limitation in tests once this fix end up in a conda packaged version we may lift the current limitation of keyring to
1
7,454
6,963,845,311
IssuesEvent
2017-12-08 19:05:13
KSP-CKAN/CKAN
https://api.github.com/repos/KSP-CKAN/CKAN
opened
CKAN's GitHub downloads are breaking the rules
ckan.dll Cmdline GUI Infrastructure Metadata Netkan Schema Spec
# GitHub downloading needs a rewrite (I debated whether to add this as a comment to #1817, but it seems like too much text and detail for that.) ## Problems Currently if CKAN downloads many files from GitHub at the same time, they often fail with HTTP status code 403-Forbidden. #1817 contains an example, but these reports are common and I've definitely seen it happen myself several times. ## Background The GitHub API uses 403 codes for throttling; you get 60 unauthenticated requests per hour, and any beyond that return a 403. I encountered this while working on [an unrelated project](https://hebarusan.github.io/releases-timeline/), and I had to use a GitHub token to allow 5000/hour, passed in the HTTP request headers: ```http Authentication: token <OAuth token here> ``` Currently CKAN's downloads do **not** go through the GitHub API, so this does not necessarily indicate exactly what's going on with them. However, it establishes that 403-Forbidden is sometimes used for throttling, and it becomes more relevant later in discussion of the API. [Sample API data for releases](https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924), minus the author and uploader fields since they're long and not relevant to this issue: ```json { "url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924", "assets_url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets", "upload_url": "https://uploads.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets{?name,label}", "html_url": "https://github.com/HebaruSan/Astrogator/releases/tag/v0.7.8", "id": 7538924, "tag_name": "v0.7.8", "target_commitish": "master", "name": "Frictionless toilet", "draft": false, "prerelease": false, "created_at": "2017-08-28T06:00:32Z", "published_at": "2017-08-28T06:09:07Z", "assets": [ { "url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/assets/4682631", "id": 4682631, "name": "Astrogator.zip", "label": null, "content_type": "application/zip", "state": "uploaded", "size": 114372, "download_count": 5243, "created_at": "2017-08-28T06:08:01Z", "updated_at": "2017-08-28T06:08:02Z", "browser_download_url": "https://github.com/HebaruSan/Astrogator/releases/download/v0.7.8/Astrogator.zip" } ], "tarball_url": "https://api.github.com/repos/HebaruSan/Astrogator/tarball/v0.7.8", "zipball_url": "https://api.github.com/repos/HebaruSan/Astrogator/zipball/v0.7.8", "body": "Fix glitches when the settings file is invalid" } ``` The zip file that we want to download is associated with `assets[0]`, and there are two fields for it, `url` and `browser_download_url`. This becomes important later. ## Investigation summary I used the "Contact GitHub" link to reach out to GitHub about how their download throttling works. Surprisingly, the person who replied understood exactly what I was talking about and how to fix it :+1:. It turns out that these problems happen because CKAN is not using GitHub as intended. From my conversation with the very helpful support person: > If I understood your message correctly, it seems like you're programmatically downloading resources from github.com, is that right? If that's so, then you shouldn't be doing that. GitHub.com wasn't build for programmatic use like that, it was built for humans. For programmatic use, you should be using the API. The API has well defined rate limits and caching behavior you can rely on, while GitHub.com doesn't. That doesn't mean that github.com doesn't have any rate limits, it only means that you can be rate limited at any time and without warning. > > So, we'd like to ask you to switch and use the API for downloading the data you need, and respect the defined rate limits (that's what a good citizen app should be doing, instead of hitting github.com): ("good citizen" was my phrasing in my original message, so don't take that as an unprovoked criticism of our civic virtues.) > If I'm interpreting that code snippet correctly, you're using the browser_download_url link, which, as the name suggests, is intended to be used by human users via a browser. > > For downloading release assets via the API, you should be using this endpoint: > > https://developer.github.com/v3/repos/releases/#get-a-single-release-asset > > Notice this note: "To download the asset's binary content, set the Accept header of the request to application/octet-stream. The API will either redirect the client to the location, or stream it directly if possible. API clients should handle both a 200 or 302 response." > > That would be the "url" field of a particular asset (which are listed when you fetch a release e.g. via https://developer.github.com/v3/repos/releases/#get-a-single-release), but with the addition of the special Accept header. Key points: - The URL from the field we're using currently (`browser_download_url`) is for users and browsers only, not applications. It can be throttled, but there is no explicit policy or workaround. - We should be using the GitHub API for downloads. Currently we use it in the Netkan code that finds new releases, but for downloads we effectively impersonate a browser. - This can be done by requesting the `url` field instead of `browser_download_url` and setting a custom HTTP header: ```http Accept: application/octet-stream ``` I tested this with wget, and setting the Accept header did indeed give me the download. Without this header, it returns a JSON object describing the asset. ## Changes needed to stop abusing `browser_download_url` ### GitHub-specific downloading metadata & logic When downloading from GitHub, we need to send the custom HTTP header. This cannot be accomplished simply by swapping out the bad URL for the good URL in the `download` metadata field. Proposed new metadata field: - `github_download` - The `assets[0].url` value from the API Specific changes: - The spec/schema would need to be updated to allow this field. - Netkan would need to be updated to generate this field. - CKAN would need to be updated to check for the presence of this field, which would then trigger an alternate download method that sets the custom header. ### UI to handle 403 statuses If a GitHub download returns a 403 status, we should handle the exception and notify the user that their downloads are being throttled. We could direct them to the setting (see below) and [web page dealing with GitHub auth tokens](https://github.com/settings/tokens), and/or advise them to wait 60 minutes for their limit to reset. https://api.github.com/rate_limit can be used to get the exact limit and timing numbers. ### GitHub token handling Users will be limited to 60 GitHub downloads per hour, because this is the limit of the GitHub API. 140+ mod installs are pretty commonly mentioned on the forums, and reinstalling everything from scratch is a common method for dealing with compatible upgrades, so some users would probably encounter this limit and not appreciate the 60-minute wait to be able to download more. The only way around this is to use a GitHub auth token, which boosts the limit to 5000/hour per token. It would be nice to ship a single internal auth token for all of CKAN, since then users would have the 5000/hour limit by default without having to worry about any of the details. More responses from the GitHub contact person: > Including a single pre-defined token with the app so that this token is used by all users of the app is possible. You could create a scopeless token here https://github.com/settings/tokens/new and include that. A scopeless token doesn't have any special permissions -- it can be used for read-only access of public data. So, it would be safe in that way. However, someone could easily take your token from the app, and then drain the API quota for the user who owns the token by making lots of unnecessary API requests. At that point, the app would stop working for everyone who uses the app. Deliberate abuse like that is unlikely, but assuming 200 downloads per active user per hour, a 5000/hour limit across all CKAN users would support 25 active users in a given hour. The number of active users at a given time isn't known, but the latest CKAN release has over 60000 downloads, so it's probably more than 25. If we were able to determine the limit we needed per hour, we could divide it by 5000 and then generate that many tokens and pick one randomly per request, but that might not be in the spirit of the API's rules. #### A setting We could create a new settings field called GitHub Auth Token, where the user could fill in their own tokens to allow more downloads. This could be instead of or in addition to any built-in tokens we may or may not use, and it should support all the UIs. #### Multi-pass approach 1. Try with no authentication at all. This would succeed for the first 60 requests per user per hour, probably the majority but not all. 2. For the remaining requests that fail, retry with a single hard coded auth token. As long as we only use this as a fallback, the 5000/hour limit would only apply to downloads in excess the 60/hour. ### Migration concerns If Netkan was updated to use this new scheme tomorrow, current CKAN clients would break unless the old `download` field was still populated. So we should not remove support for the old metadata immediately; GitHub downloads should use both `download` and `github_download` until all clients are updated. ## Or just download serially The [API docs](https://developer.github.com/v3/guides/best-practices-for-integrators/) say: > Make requests for a single user or client ID serially. Do not make requests for a single user or client ID concurrently. So even with a token, CKAN's parallel download method would still be in violation of the letter of the law. As a halfway measure, we could try scaling back the parallelization of downloads. 1. Check whether a download URL contains "github.com" 2. If so, add it to a pool of downloads to be handled serially 3. Handle all other downloads normally 4. When a download finishes, if it contains "github.com", then start a new download from the pool This might solve the problem without messing with all the API/token stuff. We would still technically be misusing GitHub, but users should no longer encounter failed downloads as frequently.
1.0
CKAN's GitHub downloads are breaking the rules - # GitHub downloading needs a rewrite (I debated whether to add this as a comment to #1817, but it seems like too much text and detail for that.) ## Problems Currently if CKAN downloads many files from GitHub at the same time, they often fail with HTTP status code 403-Forbidden. #1817 contains an example, but these reports are common and I've definitely seen it happen myself several times. ## Background The GitHub API uses 403 codes for throttling; you get 60 unauthenticated requests per hour, and any beyond that return a 403. I encountered this while working on [an unrelated project](https://hebarusan.github.io/releases-timeline/), and I had to use a GitHub token to allow 5000/hour, passed in the HTTP request headers: ```http Authentication: token <OAuth token here> ``` Currently CKAN's downloads do **not** go through the GitHub API, so this does not necessarily indicate exactly what's going on with them. However, it establishes that 403-Forbidden is sometimes used for throttling, and it becomes more relevant later in discussion of the API. [Sample API data for releases](https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924), minus the author and uploader fields since they're long and not relevant to this issue: ```json { "url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924", "assets_url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets", "upload_url": "https://uploads.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets{?name,label}", "html_url": "https://github.com/HebaruSan/Astrogator/releases/tag/v0.7.8", "id": 7538924, "tag_name": "v0.7.8", "target_commitish": "master", "name": "Frictionless toilet", "draft": false, "prerelease": false, "created_at": "2017-08-28T06:00:32Z", "published_at": "2017-08-28T06:09:07Z", "assets": [ { "url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/assets/4682631", "id": 4682631, "name": "Astrogator.zip", "label": null, "content_type": "application/zip", "state": "uploaded", "size": 114372, "download_count": 5243, "created_at": "2017-08-28T06:08:01Z", "updated_at": "2017-08-28T06:08:02Z", "browser_download_url": "https://github.com/HebaruSan/Astrogator/releases/download/v0.7.8/Astrogator.zip" } ], "tarball_url": "https://api.github.com/repos/HebaruSan/Astrogator/tarball/v0.7.8", "zipball_url": "https://api.github.com/repos/HebaruSan/Astrogator/zipball/v0.7.8", "body": "Fix glitches when the settings file is invalid" } ``` The zip file that we want to download is associated with `assets[0]`, and there are two fields for it, `url` and `browser_download_url`. This becomes important later. ## Investigation summary I used the "Contact GitHub" link to reach out to GitHub about how their download throttling works. Surprisingly, the person who replied understood exactly what I was talking about and how to fix it :+1:. It turns out that these problems happen because CKAN is not using GitHub as intended. From my conversation with the very helpful support person: > If I understood your message correctly, it seems like you're programmatically downloading resources from github.com, is that right? If that's so, then you shouldn't be doing that. GitHub.com wasn't build for programmatic use like that, it was built for humans. For programmatic use, you should be using the API. The API has well defined rate limits and caching behavior you can rely on, while GitHub.com doesn't. That doesn't mean that github.com doesn't have any rate limits, it only means that you can be rate limited at any time and without warning. > > So, we'd like to ask you to switch and use the API for downloading the data you need, and respect the defined rate limits (that's what a good citizen app should be doing, instead of hitting github.com): ("good citizen" was my phrasing in my original message, so don't take that as an unprovoked criticism of our civic virtues.) > If I'm interpreting that code snippet correctly, you're using the browser_download_url link, which, as the name suggests, is intended to be used by human users via a browser. > > For downloading release assets via the API, you should be using this endpoint: > > https://developer.github.com/v3/repos/releases/#get-a-single-release-asset > > Notice this note: "To download the asset's binary content, set the Accept header of the request to application/octet-stream. The API will either redirect the client to the location, or stream it directly if possible. API clients should handle both a 200 or 302 response." > > That would be the "url" field of a particular asset (which are listed when you fetch a release e.g. via https://developer.github.com/v3/repos/releases/#get-a-single-release), but with the addition of the special Accept header. Key points: - The URL from the field we're using currently (`browser_download_url`) is for users and browsers only, not applications. It can be throttled, but there is no explicit policy or workaround. - We should be using the GitHub API for downloads. Currently we use it in the Netkan code that finds new releases, but for downloads we effectively impersonate a browser. - This can be done by requesting the `url` field instead of `browser_download_url` and setting a custom HTTP header: ```http Accept: application/octet-stream ``` I tested this with wget, and setting the Accept header did indeed give me the download. Without this header, it returns a JSON object describing the asset. ## Changes needed to stop abusing `browser_download_url` ### GitHub-specific downloading metadata & logic When downloading from GitHub, we need to send the custom HTTP header. This cannot be accomplished simply by swapping out the bad URL for the good URL in the `download` metadata field. Proposed new metadata field: - `github_download` - The `assets[0].url` value from the API Specific changes: - The spec/schema would need to be updated to allow this field. - Netkan would need to be updated to generate this field. - CKAN would need to be updated to check for the presence of this field, which would then trigger an alternate download method that sets the custom header. ### UI to handle 403 statuses If a GitHub download returns a 403 status, we should handle the exception and notify the user that their downloads are being throttled. We could direct them to the setting (see below) and [web page dealing with GitHub auth tokens](https://github.com/settings/tokens), and/or advise them to wait 60 minutes for their limit to reset. https://api.github.com/rate_limit can be used to get the exact limit and timing numbers. ### GitHub token handling Users will be limited to 60 GitHub downloads per hour, because this is the limit of the GitHub API. 140+ mod installs are pretty commonly mentioned on the forums, and reinstalling everything from scratch is a common method for dealing with compatible upgrades, so some users would probably encounter this limit and not appreciate the 60-minute wait to be able to download more. The only way around this is to use a GitHub auth token, which boosts the limit to 5000/hour per token. It would be nice to ship a single internal auth token for all of CKAN, since then users would have the 5000/hour limit by default without having to worry about any of the details. More responses from the GitHub contact person: > Including a single pre-defined token with the app so that this token is used by all users of the app is possible. You could create a scopeless token here https://github.com/settings/tokens/new and include that. A scopeless token doesn't have any special permissions -- it can be used for read-only access of public data. So, it would be safe in that way. However, someone could easily take your token from the app, and then drain the API quota for the user who owns the token by making lots of unnecessary API requests. At that point, the app would stop working for everyone who uses the app. Deliberate abuse like that is unlikely, but assuming 200 downloads per active user per hour, a 5000/hour limit across all CKAN users would support 25 active users in a given hour. The number of active users at a given time isn't known, but the latest CKAN release has over 60000 downloads, so it's probably more than 25. If we were able to determine the limit we needed per hour, we could divide it by 5000 and then generate that many tokens and pick one randomly per request, but that might not be in the spirit of the API's rules. #### A setting We could create a new settings field called GitHub Auth Token, where the user could fill in their own tokens to allow more downloads. This could be instead of or in addition to any built-in tokens we may or may not use, and it should support all the UIs. #### Multi-pass approach 1. Try with no authentication at all. This would succeed for the first 60 requests per user per hour, probably the majority but not all. 2. For the remaining requests that fail, retry with a single hard coded auth token. As long as we only use this as a fallback, the 5000/hour limit would only apply to downloads in excess the 60/hour. ### Migration concerns If Netkan was updated to use this new scheme tomorrow, current CKAN clients would break unless the old `download` field was still populated. So we should not remove support for the old metadata immediately; GitHub downloads should use both `download` and `github_download` until all clients are updated. ## Or just download serially The [API docs](https://developer.github.com/v3/guides/best-practices-for-integrators/) say: > Make requests for a single user or client ID serially. Do not make requests for a single user or client ID concurrently. So even with a token, CKAN's parallel download method would still be in violation of the letter of the law. As a halfway measure, we could try scaling back the parallelization of downloads. 1. Check whether a download URL contains "github.com" 2. If so, add it to a pool of downloads to be handled serially 3. Handle all other downloads normally 4. When a download finishes, if it contains "github.com", then start a new download from the pool This might solve the problem without messing with all the API/token stuff. We would still technically be misusing GitHub, but users should no longer encounter failed downloads as frequently.
infrastructure
ckan s github downloads are breaking the rules github downloading needs a rewrite i debated whether to add this as a comment to but it seems like too much text and detail for that problems currently if ckan downloads many files from github at the same time they often fail with http status code forbidden contains an example but these reports are common and i ve definitely seen it happen myself several times background the github api uses codes for throttling you get unauthenticated requests per hour and any beyond that return a i encountered this while working on and i had to use a github token to allow hour passed in the http request headers http authentication token currently ckan s downloads do not go through the github api so this does not necessarily indicate exactly what s going on with them however it establishes that forbidden is sometimes used for throttling and it becomes more relevant later in discussion of the api minus the author and uploader fields since they re long and not relevant to this issue json url assets url upload url html url id tag name target commitish master name frictionless toilet draft false prerelease false created at published at assets url id name astrogator zip label null content type application zip state uploaded size download count created at updated at browser download url tarball url zipball url body fix glitches when the settings file is invalid the zip file that we want to download is associated with assets and there are two fields for it url and browser download url this becomes important later investigation summary i used the contact github link to reach out to github about how their download throttling works surprisingly the person who replied understood exactly what i was talking about and how to fix it it turns out that these problems happen because ckan is not using github as intended from my conversation with the very helpful support person if i understood your message correctly it seems like you re programmatically downloading resources from github com is that right if that s so then you shouldn t be doing that github com wasn t build for programmatic use like that it was built for humans for programmatic use you should be using the api the api has well defined rate limits and caching behavior you can rely on while github com doesn t that doesn t mean that github com doesn t have any rate limits it only means that you can be rate limited at any time and without warning so we d like to ask you to switch and use the api for downloading the data you need and respect the defined rate limits that s what a good citizen app should be doing instead of hitting github com good citizen was my phrasing in my original message so don t take that as an unprovoked criticism of our civic virtues if i m interpreting that code snippet correctly you re using the browser download url link which as the name suggests is intended to be used by human users via a browser for downloading release assets via the api you should be using this endpoint notice this note to download the asset s binary content set the accept header of the request to application octet stream the api will either redirect the client to the location or stream it directly if possible api clients should handle both a or response that would be the url field of a particular asset which are listed when you fetch a release e g via but with the addition of the special accept header key points the url from the field we re using currently browser download url is for users and browsers only not applications it can be throttled but there is no explicit policy or workaround we should be using the github api for downloads currently we use it in the netkan code that finds new releases but for downloads we effectively impersonate a browser this can be done by requesting the url field instead of browser download url and setting a custom http header http accept application octet stream i tested this with wget and setting the accept header did indeed give me the download without this header it returns a json object describing the asset changes needed to stop abusing browser download url github specific downloading metadata logic when downloading from github we need to send the custom http header this cannot be accomplished simply by swapping out the bad url for the good url in the download metadata field proposed new metadata field github download the assets url value from the api specific changes the spec schema would need to be updated to allow this field netkan would need to be updated to generate this field ckan would need to be updated to check for the presence of this field which would then trigger an alternate download method that sets the custom header ui to handle statuses if a github download returns a status we should handle the exception and notify the user that their downloads are being throttled we could direct them to the setting see below and and or advise them to wait minutes for their limit to reset can be used to get the exact limit and timing numbers github token handling users will be limited to github downloads per hour because this is the limit of the github api mod installs are pretty commonly mentioned on the forums and reinstalling everything from scratch is a common method for dealing with compatible upgrades so some users would probably encounter this limit and not appreciate the minute wait to be able to download more the only way around this is to use a github auth token which boosts the limit to hour per token it would be nice to ship a single internal auth token for all of ckan since then users would have the hour limit by default without having to worry about any of the details more responses from the github contact person including a single pre defined token with the app so that this token is used by all users of the app is possible you could create a scopeless token here and include that a scopeless token doesn t have any special permissions it can be used for read only access of public data so it would be safe in that way however someone could easily take your token from the app and then drain the api quota for the user who owns the token by making lots of unnecessary api requests at that point the app would stop working for everyone who uses the app deliberate abuse like that is unlikely but assuming downloads per active user per hour a hour limit across all ckan users would support active users in a given hour the number of active users at a given time isn t known but the latest ckan release has over downloads so it s probably more than if we were able to determine the limit we needed per hour we could divide it by and then generate that many tokens and pick one randomly per request but that might not be in the spirit of the api s rules a setting we could create a new settings field called github auth token where the user could fill in their own tokens to allow more downloads this could be instead of or in addition to any built in tokens we may or may not use and it should support all the uis multi pass approach try with no authentication at all this would succeed for the first requests per user per hour probably the majority but not all for the remaining requests that fail retry with a single hard coded auth token as long as we only use this as a fallback the hour limit would only apply to downloads in excess the hour migration concerns if netkan was updated to use this new scheme tomorrow current ckan clients would break unless the old download field was still populated so we should not remove support for the old metadata immediately github downloads should use both download and github download until all clients are updated or just download serially the say make requests for a single user or client id serially do not make requests for a single user or client id concurrently so even with a token ckan s parallel download method would still be in violation of the letter of the law as a halfway measure we could try scaling back the parallelization of downloads check whether a download url contains github com if so add it to a pool of downloads to be handled serially handle all other downloads normally when a download finishes if it contains github com then start a new download from the pool this might solve the problem without messing with all the api token stuff we would still technically be misusing github but users should no longer encounter failed downloads as frequently
1
12,544
9,819,543,776
IssuesEvent
2019-06-13 22:23:55
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
Implement Preview strategies for commands/command groups/parameters
Infrastructure
We currently have a mechanism to show that extensions are in preview, but we have no such thing for specific CLI commands, command groups, or parameters. Even within a GA module, certain parameters may be added in preview, for example `--ddos-protection` for `vnet create`. It would be good to standardize the experience for these: perhaps issue a warning if a preview command or parameter is used so that users are warned that these things may break and thus to avoid using them in scripts. Displaying this warning or not could be a configurable setting. --- ### Environment summary Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows) ``` azure-cli (2.0.30) ```
1.0
Implement Preview strategies for commands/command groups/parameters - We currently have a mechanism to show that extensions are in preview, but we have no such thing for specific CLI commands, command groups, or parameters. Even within a GA module, certain parameters may be added in preview, for example `--ddos-protection` for `vnet create`. It would be good to standardize the experience for these: perhaps issue a warning if a preview command or parameter is used so that users are warned that these things may break and thus to avoid using them in scripts. Displaying this warning or not could be a configurable setting. --- ### Environment summary Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows) ``` azure-cli (2.0.30) ```
infrastructure
implement preview strategies for commands command groups parameters we currently have a mechanism to show that extensions are in preview but we have no such thing for specific cli commands command groups or parameters even within a ga module certain parameters may be added in preview for example ddos protection for vnet create it would be good to standardize the experience for these perhaps issue a warning if a preview command or parameter is used so that users are warned that these things may break and thus to avoid using them in scripts displaying this warning or not could be a configurable setting environment summary install method e g pip interactive script apt get docker msi edge build cli version az version os version shell type e g bash cmd exe bash on windows azure cli
1
35,138
30,781,405,970
IssuesEvent
2023-07-31 10:15:39
Altinn/altinn-platform
https://api.github.com/repos/Altinn/altinn-platform
closed
Rollout Grafana/Prometheus apps cluster
solution/apps area/infrastructure feature-complete
```[tasklist] ### Rollout - [x] at21 - [x] at2[2-4] - [x] ttd-tt02 - [x] tt02 - [x] ttd-prod - [x] prod ```
1.0
Rollout Grafana/Prometheus apps cluster - ```[tasklist] ### Rollout - [x] at21 - [x] at2[2-4] - [x] ttd-tt02 - [x] tt02 - [x] ttd-prod - [x] prod ```
infrastructure
rollout grafana prometheus apps cluster rollout ttd ttd prod prod
1
71,115
23,453,165,146
IssuesEvent
2022-08-16 06:15:13
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
closed
[🐛 Bug]: DotNet WebDriver msedge(Ver:103.0.1264.77)Use EdgeOptions DebuggerAddress is 127.0.0.1:50521 Error
C-dotnet I-defect
### What happened? OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid from invalid argument: cannot parse capability: ms:edgeOptions from invalid argument: unrecognized msedge option: excludeSwitches 在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse) 在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters) 在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities) 在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities) 在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout) 在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout) 在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning)) ### How can we reproduce the issue? ```shell var options = new EdgeOptions(); var debugHost = "127.0.0.1"; var debugPort = 50521; options.DebuggerAddress = $"{debugHost}:{debugPort}"; driver = new EdgeDriver(driverService, options, TimeSpan.FromSeconds(40)); ``` ### Relevant log output ```shell OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid from invalid argument: cannot parse capability: ms:edgeOptions from invalid argument: unrecognized msedge option: excludeSwitches 在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse) 在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters) 在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities) 在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities) 在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout) 在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout) 在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning)) ``` ### Operating System Windows 11 ### Selenium version Dotnet 4.3 ### What are the browser(s) and version(s) where you see this issue? msedge 103.0.1264.77 ### What are the browser driver(s) and version(s) where you see this issue? msedgedriver 103.0.1264.77 ### Are you using Selenium Grid? No
1.0
[🐛 Bug]: DotNet WebDriver msedge(Ver:103.0.1264.77)Use EdgeOptions DebuggerAddress is 127.0.0.1:50521 Error - ### What happened? OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid from invalid argument: cannot parse capability: ms:edgeOptions from invalid argument: unrecognized msedge option: excludeSwitches 在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse) 在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters) 在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities) 在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities) 在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout) 在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout) 在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning)) ### How can we reproduce the issue? ```shell var options = new EdgeOptions(); var debugHost = "127.0.0.1"; var debugPort = 50521; options.DebuggerAddress = $"{debugHost}:{debugPort}"; driver = new EdgeDriver(driverService, options, TimeSpan.FromSeconds(40)); ``` ### Relevant log output ```shell OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid from invalid argument: cannot parse capability: ms:edgeOptions from invalid argument: unrecognized msedge option: excludeSwitches 在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse) 在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters) 在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities) 在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities) 在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout) 在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout) 在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning)) ``` ### Operating System Windows 11 ### Selenium version Dotnet 4.3 ### What are the browser(s) and version(s) where you see this issue? msedge 103.0.1264.77 ### What are the browser driver(s) and version(s) where you see this issue? msedgedriver 103.0.1264.77 ### Are you using Selenium Grid? No
non_infrastructure
dotnet webdriver msedge ver )use edgeoptions debuggeraddress is error what happened openqa selenium webdriverargumentexception invalid argument entry of firstmatch is invalid from invalid argument cannot parse capability ms edgeoptions from invalid argument unrecognized msedge option excludeswitches 在 openqa selenium webdriver unpackandthrowonerror response errorresponse 在 openqa selenium webdriver execute string drivercommandtoexecute dictionary parameters 在 openqa selenium webdriver startsession icapabilities desiredcapabilities 在 openqa selenium webdriver ctor icommandexecutor executor icapabilities capabilities 在 openqa selenium chromium chromiumdriver ctor chromiumdriverservice service chromiumoptions options timespan commandtimeout 在 openqa selenium edge edgedriver ctor edgedriverservice service edgeoptions options timespan commandtimeout 在 xrbrowser mainform createedgedriver boolean isrunning ) how can we reproduce the issue shell var options new edgeoptions var debughost var debugport options debuggeraddress debughost debugport driver new edgedriver driverservice options timespan fromseconds relevant log output shell openqa selenium webdriverargumentexception invalid argument entry of firstmatch is invalid from invalid argument cannot parse capability ms edgeoptions from invalid argument unrecognized msedge option excludeswitches 在 openqa selenium webdriver unpackandthrowonerror response errorresponse 在 openqa selenium webdriver execute string drivercommandtoexecute dictionary parameters 在 openqa selenium webdriver startsession icapabilities desiredcapabilities 在 openqa selenium webdriver ctor icommandexecutor executor icapabilities capabilities 在 openqa selenium chromium chromiumdriver ctor chromiumdriverservice service chromiumoptions options timespan commandtimeout 在 openqa selenium edge edgedriver ctor edgedriverservice service edgeoptions options timespan commandtimeout 在 xrbrowser mainform createedgedriver boolean isrunning ) operating system windows selenium version dotnet what are the browser s and version s where you see this issue msedge what are the browser driver s and version s where you see this issue msedgedriver are you using selenium grid no
0
410
2,857,052,418
IssuesEvent
2015-06-02 17:50:05
astropy/astroquery
https://api.github.com/repos/astropy/astroquery
opened
Add module for Pan-STARRS dust mapping service
New Service
See http://argonaut.skymaps.info/ for the service. http://argonaut.skymaps.info/usage#API gives a lot of the necessary query code, but it doesn't show use with astropy or anything (and requires doing things like constructing the API calls yourself). One possible concern - it's not clear how "stable" this site will be because I think it's just a few authors hosting it rather than an "organization".
1.0
Add module for Pan-STARRS dust mapping service - See http://argonaut.skymaps.info/ for the service. http://argonaut.skymaps.info/usage#API gives a lot of the necessary query code, but it doesn't show use with astropy or anything (and requires doing things like constructing the API calls yourself). One possible concern - it's not clear how "stable" this site will be because I think it's just a few authors hosting it rather than an "organization".
non_infrastructure
add module for pan starrs dust mapping service see for the service gives a lot of the necessary query code but it doesn t show use with astropy or anything and requires doing things like constructing the api calls yourself one possible concern it s not clear how stable this site will be because i think it s just a few authors hosting it rather than an organization
0
27,432
21,707,919,881
IssuesEvent
2022-05-10 11:22:37
coatrack/coatrack
https://api.github.com/repos/coatrack/coatrack
opened
Integration of CoatRack into FriDa Project
infrastructure
This is an epic issue and therefore probably requires to create sub-issues for each sub-task. Preparation on CoatRack side: - [ ] CoatRack should be deployable using an nginx proxy for certificate management. The nginx component is to be added to the `docker-compose.yml`. - [ ] The CI pipeline shall be adapted to provide an option to deploy CoatRack (with nginx) on a Hetzner VM manually. Integration of CoatRack and FriDa: - [ ] Set up a Hetzner VM and deploy CoatRack with nginx. - [ ] Rent a VM - [ ] Register domain, e.g. `frida-integration.coatrack.atb-bremen.de` - [ ] Deploy CoatRack - [ ] Gateway creation in CoatRack Web Application whose target URL is an instance of the Tempman Project. - [ ] Context Broker subscription which directs data from the Broker to the CoatRack Gateway. - [ ] To be discussed: How do we ensure that the connection is secure? Shall we assume the user has a public IP and a certificate to run the Gateway as a usual server? Should we, for reasons of simplicity, deploy the proxy to the same VM as the CoatRack setup?
1.0
Integration of CoatRack into FriDa Project - This is an epic issue and therefore probably requires to create sub-issues for each sub-task. Preparation on CoatRack side: - [ ] CoatRack should be deployable using an nginx proxy for certificate management. The nginx component is to be added to the `docker-compose.yml`. - [ ] The CI pipeline shall be adapted to provide an option to deploy CoatRack (with nginx) on a Hetzner VM manually. Integration of CoatRack and FriDa: - [ ] Set up a Hetzner VM and deploy CoatRack with nginx. - [ ] Rent a VM - [ ] Register domain, e.g. `frida-integration.coatrack.atb-bremen.de` - [ ] Deploy CoatRack - [ ] Gateway creation in CoatRack Web Application whose target URL is an instance of the Tempman Project. - [ ] Context Broker subscription which directs data from the Broker to the CoatRack Gateway. - [ ] To be discussed: How do we ensure that the connection is secure? Shall we assume the user has a public IP and a certificate to run the Gateway as a usual server? Should we, for reasons of simplicity, deploy the proxy to the same VM as the CoatRack setup?
infrastructure
integration of coatrack into frida project this is an epic issue and therefore probably requires to create sub issues for each sub task preparation on coatrack side coatrack should be deployable using an nginx proxy for certificate management the nginx component is to be added to the docker compose yml the ci pipeline shall be adapted to provide an option to deploy coatrack with nginx on a hetzner vm manually integration of coatrack and frida set up a hetzner vm and deploy coatrack with nginx rent a vm register domain e g frida integration coatrack atb bremen de deploy coatrack gateway creation in coatrack web application whose target url is an instance of the tempman project context broker subscription which directs data from the broker to the coatrack gateway to be discussed how do we ensure that the connection is secure shall we assume the user has a public ip and a certificate to run the gateway as a usual server should we for reasons of simplicity deploy the proxy to the same vm as the coatrack setup
1
24,812
17,792,611,312
IssuesEvent
2021-08-31 18:00:55
hackforla/website
https://api.github.com/repos/hackforla/website
opened
Standardize how 'Overview' links open pdf files
role: front end Size: Large Feature: Infrastructure
### Overview We need to standardize how the 'Overview' link for each project page acts when a user clicks it. Currently, the 'Overview' links act differently for each project. Clicking the link should open a new tab to the pdf file directly. Once this issue is approved, we will convert each action item into an issue and create a pull request to close the corresponding issue. ### Details When clicking the 'Overview' link, each project page acts differently. Some open to a google docs page, others open to a github page that hosts the pdf file, and another downloads the pdf file locally to the machine. There needs to be a standard in how the pdf files open. Developers decided during the backend meeting that pdfs should be opened as a separate tab to the file directly, since modern browsers have the ability to view pdfs. The process for a developer would involve: - Using the newly created directory for pdf files: [/assets/pdfs](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs) - Place a copy of a project's pdf overview file into the /assets/pdfs directory - Changing the 'Overview' link for a project to point to the pdf file living under /assets/pdfs - Ensure that the links in both the home page and the project's dedicated page open the pdf file properly after the change ### Action Items - [ ] Change the 'Overview' link for 100 Automations project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Access The Data project to open as a pdf file in the browser instead of opening a google doc - [ ] Change the 'Overview' link for BallotNav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Civic Opportunity Project project to open as a pdf file in the browser instead of downloading a pdf file - [ ] Change the 'Overview' link for Civic Tech Index project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Civic Tech Structure project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Engage project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Expunge Assist project to open as a pdf file in the browser instead of opening a google doc - [ ] Change the 'Overview' link for GreenEarthOS project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for LA TDM Calculator project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Youth Justice Nav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file ### Resources/Instructions [Project Page](https://www.hackforla.org/projects/) Each individual landing page for affected projects: [100 automations as an example](https://www.hackforla.org/projects/100-automations) [Location for pdf files](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs) [Jekyll Static Files](https://jekyllrb.com/docs/static-files/)
1.0
Standardize how 'Overview' links open pdf files - ### Overview We need to standardize how the 'Overview' link for each project page acts when a user clicks it. Currently, the 'Overview' links act differently for each project. Clicking the link should open a new tab to the pdf file directly. Once this issue is approved, we will convert each action item into an issue and create a pull request to close the corresponding issue. ### Details When clicking the 'Overview' link, each project page acts differently. Some open to a google docs page, others open to a github page that hosts the pdf file, and another downloads the pdf file locally to the machine. There needs to be a standard in how the pdf files open. Developers decided during the backend meeting that pdfs should be opened as a separate tab to the file directly, since modern browsers have the ability to view pdfs. The process for a developer would involve: - Using the newly created directory for pdf files: [/assets/pdfs](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs) - Place a copy of a project's pdf overview file into the /assets/pdfs directory - Changing the 'Overview' link for a project to point to the pdf file living under /assets/pdfs - Ensure that the links in both the home page and the project's dedicated page open the pdf file properly after the change ### Action Items - [ ] Change the 'Overview' link for 100 Automations project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Access The Data project to open as a pdf file in the browser instead of opening a google doc - [ ] Change the 'Overview' link for BallotNav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Civic Opportunity Project project to open as a pdf file in the browser instead of downloading a pdf file - [ ] Change the 'Overview' link for Civic Tech Index project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Civic Tech Structure project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Engage project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Expunge Assist project to open as a pdf file in the browser instead of opening a google doc - [ ] Change the 'Overview' link for GreenEarthOS project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for LA TDM Calculator project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file - [ ] Change the 'Overview' link for Youth Justice Nav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file ### Resources/Instructions [Project Page](https://www.hackforla.org/projects/) Each individual landing page for affected projects: [100 automations as an example](https://www.hackforla.org/projects/100-automations) [Location for pdf files](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs) [Jekyll Static Files](https://jekyllrb.com/docs/static-files/)
infrastructure
standardize how overview links open pdf files overview we need to standardize how the overview link for each project page acts when a user clicks it currently the overview links act differently for each project clicking the link should open a new tab to the pdf file directly once this issue is approved we will convert each action item into an issue and create a pull request to close the corresponding issue details when clicking the overview link each project page acts differently some open to a google docs page others open to a github page that hosts the pdf file and another downloads the pdf file locally to the machine there needs to be a standard in how the pdf files open developers decided during the backend meeting that pdfs should be opened as a separate tab to the file directly since modern browsers have the ability to view pdfs the process for a developer would involve using the newly created directory for pdf files place a copy of a project s pdf overview file into the assets pdfs directory changing the overview link for a project to point to the pdf file living under assets pdfs ensure that the links in both the home page and the project s dedicated page open the pdf file properly after the change action items change the overview link for automations project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for access the data project to open as a pdf file in the browser instead of opening a google doc change the overview link for ballotnav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for civic opportunity project project to open as a pdf file in the browser instead of downloading a pdf file change the overview link for civic tech index project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for civic tech structure project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for engage project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for expunge assist project to open as a pdf file in the browser instead of opening a google doc change the overview link for greenearthos project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for la tdm calculator project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for youth justice nav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file resources instructions each individual landing page for affected projects
1
7,724
7,059,161,483
IssuesEvent
2018-01-04 23:46:42
dotnet/core-setup
https://api.github.com/repos/dotnet/core-setup
closed
Debian stretch packages?
area-Infrastructure
Are there packages for debian stretch? The [official debian instructions](https://www.microsoft.com/net/core#linuxdebian) don't work [because of issues with dependencies](https://github.com/dotnet/cli/issues/2018#issuecomment-239301891) and the packages in this repo can't be installed for similar reasons.
1.0
Debian stretch packages? - Are there packages for debian stretch? The [official debian instructions](https://www.microsoft.com/net/core#linuxdebian) don't work [because of issues with dependencies](https://github.com/dotnet/cli/issues/2018#issuecomment-239301891) and the packages in this repo can't be installed for similar reasons.
infrastructure
debian stretch packages are there packages for debian stretch the don t work and the packages in this repo can t be installed for similar reasons
1
53,965
13,890,088,965
IssuesEvent
2020-10-19 08:48:55
edwardhu/yarn
https://api.github.com/repos/edwardhu/yarn
opened
CVE-2018-3728 (High) detected in hoek-2.16.3.tgz
security vulnerability
## CVE-2018-3728 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary> <p>General purpose node utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p> <p>Path to dependency file: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock</p> <p>Path to vulnerable library: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock,yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/node_modules/hoek/package.json,yarn/__tests__/fixtures/install/install-bin-links-eslint/node_modules/hoek/package.json</p> <p> Dependency Hierarchy: - karma-2.0.5.tgz (Root Library) - log4js-2.11.0.tgz - loggly-1.1.1.tgz - request-2.75.0.tgz - hawk-3.1.3.tgz - :x: **hoek-2.16.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/edwardhu/yarn/commit/2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9">2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects. <p>Publish Date: 2018-03-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3728">https://nvd.nist.gov/vuln/detail/CVE-2018-3728</a></p> <p>Release Date: 2018-03-30</p> <p>Fix Resolution: 4.2.1,5.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hoek","packageVersion":"2.16.3","isTransitiveDependency":true,"dependencyTree":"karma:2.0.5;log4js:2.11.0;loggly:1.1.1;request:2.75.0;hawk:3.1.3;hoek:2.16.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.1,5.0.3"}],"vulnerabilityIdentifier":"CVE-2018-3728","vulnerabilityDetails":"hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via \u0027merge\u0027 and \u0027applyToDefaults\u0027 functions, which allows a malicious user to modify the prototype of \"Object\" via __proto__, causing the addition or modification of an existing property that will exist on all objects.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-3728 (High) detected in hoek-2.16.3.tgz - ## CVE-2018-3728 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary> <p>General purpose node utilities</p> <p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p> <p>Path to dependency file: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock</p> <p>Path to vulnerable library: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock,yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/node_modules/hoek/package.json,yarn/__tests__/fixtures/install/install-bin-links-eslint/node_modules/hoek/package.json</p> <p> Dependency Hierarchy: - karma-2.0.5.tgz (Root Library) - log4js-2.11.0.tgz - loggly-1.1.1.tgz - request-2.75.0.tgz - hawk-3.1.3.tgz - :x: **hoek-2.16.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/edwardhu/yarn/commit/2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9">2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects. <p>Publish Date: 2018-03-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3728">https://nvd.nist.gov/vuln/detail/CVE-2018-3728</a></p> <p>Release Date: 2018-03-30</p> <p>Fix Resolution: 4.2.1,5.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hoek","packageVersion":"2.16.3","isTransitiveDependency":true,"dependencyTree":"karma:2.0.5;log4js:2.11.0;loggly:1.1.1;request:2.75.0;hawk:3.1.3;hoek:2.16.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.1,5.0.3"}],"vulnerabilityIdentifier":"CVE-2018-3728","vulnerabilityDetails":"hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via \u0027merge\u0027 and \u0027applyToDefaults\u0027 functions, which allows a malicious user to modify the prototype of \"Object\" via __proto__, causing the addition or modification of an existing property that will exist on all objects.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_infrastructure
cve high detected in hoek tgz cve high severity vulnerability vulnerable library hoek tgz general purpose node utilities library home page a href path to dependency file yarn tests fixtures outdated wanted newer than current yarn lock path to vulnerable library yarn tests fixtures outdated wanted newer than current yarn lock yarn tests fixtures install install update auth no offline integrity offline mirror abab package node modules hoek package json yarn tests fixtures install install bin links eslint node modules hoek package json dependency hierarchy karma tgz root library tgz loggly tgz request tgz hawk tgz x hoek tgz vulnerable library found in head commit a href found in base branch master vulnerability details hoek node module before and x before suffers from a modification of assumed immutable data maid vulnerability via merge and applytodefaults functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails hoek node module before and x before suffers from a modification of assumed immutable data maid vulnerability via and functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects vulnerabilityurl
0
34,649
30,250,297,682
IssuesEvent
2023-07-06 19:58:07
tskit-dev/tstrait
https://api.github.com/repos/tskit-dev/tstrait
closed
Add testing
Infrastructure
As this is pure-python I think you could get away with linux-only testing for now. That will remove the need for conda in the setup.
1.0
Add testing - As this is pure-python I think you could get away with linux-only testing for now. That will remove the need for conda in the setup.
infrastructure
add testing as this is pure python i think you could get away with linux only testing for now that will remove the need for conda in the setup
1
13,740
10,439,871,687
IssuesEvent
2019-09-18 07:27:17
oppia/oppia-android
https://api.github.com/repos/oppia/oppia-android
opened
Fix Robolectric tests on Android Studio 3.5+
Priority: Essential Status: In implementation Type: Bug Where: Infrastructure
For some reason starting with Android Studio 3.5, all Robolectric tests appear to fail pulling in key generated classpath elements such as protobuf classes. Robolectric seems to be the same, so this appears to be a change in how Android Studio/bundled Gradle handle the classpath when running these tests. #59 may circumvent this problem by avoiding Gradle, and given the time I've put into investigating this issue that may be a faster solution.
1.0
Fix Robolectric tests on Android Studio 3.5+ - For some reason starting with Android Studio 3.5, all Robolectric tests appear to fail pulling in key generated classpath elements such as protobuf classes. Robolectric seems to be the same, so this appears to be a change in how Android Studio/bundled Gradle handle the classpath when running these tests. #59 may circumvent this problem by avoiding Gradle, and given the time I've put into investigating this issue that may be a faster solution.
infrastructure
fix robolectric tests on android studio for some reason starting with android studio all robolectric tests appear to fail pulling in key generated classpath elements such as protobuf classes robolectric seems to be the same so this appears to be a change in how android studio bundled gradle handle the classpath when running these tests may circumvent this problem by avoiding gradle and given the time i ve put into investigating this issue that may be a faster solution
1
34,491
30,024,373,445
IssuesEvent
2023-06-27 04:09:18
dotnet/project-system
https://api.github.com/repos/dotnet/project-system
closed
View Code button missing on Solution Explorer toolbar for WinForms and WPF apps
Area-Infrastructure
## Minimal Reproduction Scenario 1. Create a Winforms .Net Core application. 2. In solution Explorer, focus on Form1.cs node, then observe the toolbar items at the top of solution explorer window. ## Current Behavior Missing "View Code" button in Solution Explorer ![Actual](https://github.com/dotnet/project-system/assets/45864985/3cd83107-eff7-40d8-be95-19f848000b4e) ## Expected Behavior View Code button is there, like the result of focusing on .Net framework form ![Expected](https://github.com/dotnet/project-system/assets/45864985/c73cca90-6453-487d-afcd-f8ff4961f997) ## Test Environment 17.7.0 Preview 3.0 [33824.57.main] ## Have you experienced this same issue with .NET Framework designer? No ## Have you experienced this same issue with previous previews of VisualStudio? Yes, 17.3.6 have this issue ## More Info 1. .NET Framework project does not repro this issue. 2. This issue occurs in the .NET Core WPF project too.
1.0
View Code button missing on Solution Explorer toolbar for WinForms and WPF apps - ## Minimal Reproduction Scenario 1. Create a Winforms .Net Core application. 2. In solution Explorer, focus on Form1.cs node, then observe the toolbar items at the top of solution explorer window. ## Current Behavior Missing "View Code" button in Solution Explorer ![Actual](https://github.com/dotnet/project-system/assets/45864985/3cd83107-eff7-40d8-be95-19f848000b4e) ## Expected Behavior View Code button is there, like the result of focusing on .Net framework form ![Expected](https://github.com/dotnet/project-system/assets/45864985/c73cca90-6453-487d-afcd-f8ff4961f997) ## Test Environment 17.7.0 Preview 3.0 [33824.57.main] ## Have you experienced this same issue with .NET Framework designer? No ## Have you experienced this same issue with previous previews of VisualStudio? Yes, 17.3.6 have this issue ## More Info 1. .NET Framework project does not repro this issue. 2. This issue occurs in the .NET Core WPF project too.
infrastructure
view code button missing on solution explorer toolbar for winforms and wpf apps minimal reproduction scenario create a winforms net core application in solution explorer focus on cs node then observe the toolbar items at the top of solution explorer window current behavior missing view code button in solution explorer expected behavior view code button is there like the result of focusing on net framework form test environment preview have you experienced this same issue with net framework designer no have you experienced this same issue with previous previews of visualstudio yes have this issue more info net framework project does not repro this issue this issue occurs in the net core wpf project too
1
118,271
4,733,665,164
IssuesEvent
2016-10-19 11:59:38
cdnjs/cdnjs
https://api.github.com/repos/cdnjs/cdnjs
closed
[Request] Update flux to 3.0.0
High Priority Library - Request to Add/Update
**Library name:** flux **Git repository url:** https://github.com/facebook/flux **npm package url(optional):** https://www.npmjs.com/package/flux **Files:** https://github.com/facebook/flux/tree/master/dist
1.0
[Request] Update flux to 3.0.0 - **Library name:** flux **Git repository url:** https://github.com/facebook/flux **npm package url(optional):** https://www.npmjs.com/package/flux **Files:** https://github.com/facebook/flux/tree/master/dist
non_infrastructure
update flux to library name flux git repository url npm package url optional files
0
18,999
13,184,826,328
IssuesEvent
2020-08-12 20:10:18
Kemmey/Kemmey-TeslaWatch-Public
https://api.github.com/repos/Kemmey/Kemmey-TeslaWatch-Public
closed
Reset iWatch
AppStore infrastructure issue
I reset my watch. I can’t get the Tesla for iWatch back. In the App Store it doesn’t let me click on anything (only says “purchased”) and the app is no where to be found. This is likely a issue I need to address with Apple but I realized that as I was typing so I will continue with this avenue. App is great btw. I had the Tesla Remote app forever but it would never succeed in waking my M3. Yours worked like a charm, and quick!
1.0
Reset iWatch - I reset my watch. I can’t get the Tesla for iWatch back. In the App Store it doesn’t let me click on anything (only says “purchased”) and the app is no where to be found. This is likely a issue I need to address with Apple but I realized that as I was typing so I will continue with this avenue. App is great btw. I had the Tesla Remote app forever but it would never succeed in waking my M3. Yours worked like a charm, and quick!
infrastructure
reset iwatch i reset my watch i can’t get the tesla for iwatch back in the app store it doesn’t let me click on anything only says “purchased” and the app is no where to be found this is likely a issue i need to address with apple but i realized that as i was typing so i will continue with this avenue app is great btw i had the tesla remote app forever but it would never succeed in waking my yours worked like a charm and quick
1
6,413
3,801,383,606
IssuesEvent
2016-03-23 22:40:01
Homebrew/homebrew-science
https://api.github.com/repos/Homebrew/homebrew-science
closed
astral: error: 404 Not Found
Bioinformatics build-error
``` ==> Installing astral from homebrew/science ==> Downloading https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip curl: (22) The requested URL returned error: 404 Not Found Error: Failed to download resource "astral" Download failed: https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip ```
1.0
astral: error: 404 Not Found - ``` ==> Installing astral from homebrew/science ==> Downloading https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip curl: (22) The requested URL returned error: 404 Not Found Error: Failed to download resource "astral" Download failed: https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip ```
non_infrastructure
astral error not found installing astral from homebrew science downloading curl the requested url returned error not found error failed to download resource astral download failed
0
226,925
17,367,220,323
IssuesEvent
2021-07-30 09:00:04
vlang/v
https://api.github.com/repos/vlang/v
closed
Wrong escaping in HTML docs
Bug Unit: Documentation Unit: vdoc
`<` and `>` is getting replaced/escaped with `&lt;` and `&gt;` in markdown code in the generated HTML docs Example in the description here: https://modules.vlang.io/index.html#compare_strings The the time of writing it looks like this: ![image](https://user-images.githubusercontent.com/768942/118396766-50416000-b651-11eb-89a6-ba70c6b943e1.png)
1.0
Wrong escaping in HTML docs - `<` and `>` is getting replaced/escaped with `&lt;` and `&gt;` in markdown code in the generated HTML docs Example in the description here: https://modules.vlang.io/index.html#compare_strings The the time of writing it looks like this: ![image](https://user-images.githubusercontent.com/768942/118396766-50416000-b651-11eb-89a6-ba70c6b943e1.png)
non_infrastructure
wrong escaping in html docs   is getting replaced escaped with  lt  and  gt  in markdown code in the generated html docs example in the description here the the time of writing it looks like this
0
23,219
15,947,748,919
IssuesEvent
2021-04-15 04:17:01
OpenHistoricalMap/issues
https://api.github.com/repos/OpenHistoricalMap/issues
closed
Add Nominatim to staging and production deploy process
infrastructure
Right now, Nominatim is not part of the deployment process, and we reference the Production database on both Staging and Production, which can lead to some confusion since the staging mMaptiles and Inspector and iD all reference staging data. So we'll need to bring Nominatim into the deployment fold. This issue might need more notes about tech approach, but I'll leave that to you!
1.0
Add Nominatim to staging and production deploy process - Right now, Nominatim is not part of the deployment process, and we reference the Production database on both Staging and Production, which can lead to some confusion since the staging mMaptiles and Inspector and iD all reference staging data. So we'll need to bring Nominatim into the deployment fold. This issue might need more notes about tech approach, but I'll leave that to you!
infrastructure
add nominatim to staging and production deploy process right now nominatim is not part of the deployment process and we reference the production database on both staging and production which can lead to some confusion since the staging mmaptiles and inspector and id all reference staging data so we ll need to bring nominatim into the deployment fold this issue might need more notes about tech approach but i ll leave that to you
1
99,438
11,140,838,655
IssuesEvent
2019-12-21 17:29:35
prezly/prop-types-php
https://api.github.com/repos/prezly/prop-types-php
closed
There is no PHP code in the repo
bug documentation good first issue
The README states that's a complete PHP port of React PropTypes, but there is no PHP code in the repository. Probably I misunderstood and it's a complete port of _production_ React PropTypes, which don't exist? 🤔
1.0
There is no PHP code in the repo - The README states that's a complete PHP port of React PropTypes, but there is no PHP code in the repository. Probably I misunderstood and it's a complete port of _production_ React PropTypes, which don't exist? 🤔
non_infrastructure
there is no php code in the repo the readme states that s a complete php port of react proptypes but there is no php code in the repository probably i misunderstood and it s a complete port of production react proptypes which don t exist 🤔
0
286,928
21,629,871,887
IssuesEvent
2022-05-05 08:36:14
process-analytics/bpmn-visualization-R
https://api.github.com/repos/process-analytics/bpmn-visualization-R
opened
[DOC] Some overlays are missing in the hero image
documentation
![image](https://user-images.githubusercontent.com/27200110/166888462-66bd079a-a7bc-4880-9688-1fd1ecbf3848.png) ### Notes **TODO explain how the image was created and how to change it**
1.0
[DOC] Some overlays are missing in the hero image - ![image](https://user-images.githubusercontent.com/27200110/166888462-66bd079a-a7bc-4880-9688-1fd1ecbf3848.png) ### Notes **TODO explain how the image was created and how to change it**
non_infrastructure
some overlays are missing in the hero image notes todo explain how the image was created and how to change it
0
15,541
11,579,298,779
IssuesEvent
2020-02-21 17:37:02
eventespresso/event-espresso-core
https://api.github.com/repos/eventespresso/event-espresso-core
closed
Complete server-side mutations
EDTR v2 category:models-and-data-infrastructure
The server-side mutations for Datetime, Ticket and Price need to be completed in files inside `core/domain/services/graphql/data/mutations`. Primary fields are already taken care of. Rest of the fields need to be added. **Note**: Mutation input should properly be sanitized and converted to proper data type.
1.0
Complete server-side mutations - The server-side mutations for Datetime, Ticket and Price need to be completed in files inside `core/domain/services/graphql/data/mutations`. Primary fields are already taken care of. Rest of the fields need to be added. **Note**: Mutation input should properly be sanitized and converted to proper data type.
infrastructure
complete server side mutations the server side mutations for datetime ticket and price need to be completed in files inside core domain services graphql data mutations primary fields are already taken care of rest of the fields need to be added note mutation input should properly be sanitized and converted to proper data type
1
33,060
27,183,115,957
IssuesEvent
2023-02-18 22:05:11
surge-synthesizer/surge
https://api.github.com/repos/surge-synthesizer/surge
closed
Display version reporting error
Infrastructure Bug Report
"Surge XT 1.1.2.d3568a9 is installed but the plugin is reporting Version 1.1.0 to the host. In the About dialog, the version is ok"
1.0
Display version reporting error - "Surge XT 1.1.2.d3568a9 is installed but the plugin is reporting Version 1.1.0 to the host. In the About dialog, the version is ok"
infrastructure
display version reporting error surge xt is installed but the plugin is reporting version to the host in the about dialog the version is ok
1