Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
325,182 | 9,920,482,432 | IssuesEvent | 2019-06-30 09:44:45 | gicentre/litvis | https://api.github.com/repos/gicentre/litvis | reopened | Some function types generate parsing warning in markdown mode | bug priority/p2 | For the record, we still need to implement markdown output for a wider range of types.
Currently some types can generate parsing warnings. For example
````
```elm {m}
myFunction : List ( Int, Int )
myFunction =
[ ( 1, 2 ) ]
```
````
generates the warning : `Could not parse "[(1,2)]" (litvis:expression-value)` highlighting the entire code block.
We should also test against those examples listed in #13 and #14.
If this isn't fixable in the short term, we should probably generate a more beginner friendly and correct warning (the problem isn't with the contents of the function, but the output type of the code block) such as `Markdown output not supported for this type. Could you use 'raw' (or 'r') output instead?` | 1.0 | Some function types generate parsing warning in markdown mode - For the record, we still need to implement markdown output for a wider range of types.
Currently some types can generate parsing warnings. For example
````
```elm {m}
myFunction : List ( Int, Int )
myFunction =
[ ( 1, 2 ) ]
```
````
generates the warning : `Could not parse "[(1,2)]" (litvis:expression-value)` highlighting the entire code block.
We should also test against those examples listed in #13 and #14.
If this isn't fixable in the short term, we should probably generate a more beginner friendly and correct warning (the problem isn't with the contents of the function, but the output type of the code block) such as `Markdown output not supported for this type. Could you use 'raw' (or 'r') output instead?` | non_infrastructure | some function types generate parsing warning in markdown mode for the record we still need to implement markdown output for a wider range of types currently some types can generate parsing warnings for example elm m myfunction list int int myfunction generates the warning could not parse litvis expression value highlighting the entire code block we should also test against those examples listed in and if this isn t fixable in the short term we should probably generate a more beginner friendly and correct warning the problem isn t with the contents of the function but the output type of the code block such as markdown output not supported for this type could you use raw or r output instead | 0 |
21,984 | 14,948,627,925 | IssuesEvent | 2021-01-26 10:20:41 | RasaHQ/rasa | https://api.github.com/repos/RasaHQ/rasa | opened | Add integration test for Rasa Chat Channel (Rasa X) | area:rasa-oss :ferris_wheel: area:rasa-oss/channels 📞 area:rasa-oss/infrastructure :bullettrain_front: type:maintenance :wrench: | **Description of Problem**:
We should add integration tests for the `Rasa Chat` channel which test the entire roundtrip of sending a message, processing it via Rasa Open Source and receiving the bot's response.
**Overview of the Solution**:
TBD
**Blockers**
* This task is dependent on https://github.com/RasaHQ/rasa/issues/7804.
* we need some Rasa X deployment to test against which could make this test a little bit different than the other integration tests
**Definition of Done**:
- [ ] Tests are added
| 1.0 | Add integration test for Rasa Chat Channel (Rasa X) - **Description of Problem**:
We should add integration tests for the `Rasa Chat` channel which test the entire roundtrip of sending a message, processing it via Rasa Open Source and receiving the bot's response.
**Overview of the Solution**:
TBD
**Blockers**
* This task is dependent on https://github.com/RasaHQ/rasa/issues/7804.
* we need some Rasa X deployment to test against which could make this test a little bit different than the other integration tests
**Definition of Done**:
- [ ] Tests are added
| infrastructure | add integration test for rasa chat channel rasa x description of problem we should add integration tests for the rasa chat channel which test the entire roundtrip of sending a message processing it via rasa open source and receiving the bot s response overview of the solution tbd blockers this task is dependent on we need some rasa x deployment to test against which could make this test a little bit different than the other integration tests definition of done tests are added | 1 |
5,653 | 5,849,344,770 | IssuesEvent | 2017-05-10 23:33:16 | dotnet/corert | https://api.github.com/repos/dotnet/corert | opened | Remove un-needed target when upgrading to v2 BuildTools | Infrastructure | We have done some change in the PR #3580 to change the license Url for packaging. because corert repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct | 1.0 | Remove un-needed target when upgrading to v2 BuildTools - We have done some change in the PR #3580 to change the license Url for packaging. because corert repo not using v2 of the buildtools we had to hack this by adding the temporary target OverrideLicenseUrl in the dir.targets files. This target should be removed after upgrading to v2 of the buildtools and need to ensure when removing it the license Url in the packages nuspec files is correct | infrastructure | remove un needed target when upgrading to buildtools we have done some change in the pr to change the license url for packaging because corert repo not using of the buildtools we had to hack this by adding the temporary target overridelicenseurl in the dir targets files this target should be removed after upgrading to of the buildtools and need to ensure when removing it the license url in the packages nuspec files is correct | 1 |
27,679 | 22,148,616,896 | IssuesEvent | 2022-06-03 14:31:23 | abacus-network/abacus-monorepo | https://api.github.com/repos/abacus-network/abacus-monorepo | closed | Update infra to specify the InterchainGasPaymaster in Relayer settings | mainnet infrastructure | * Blocked by:
* https://github.com/abacus-network/abacus-monorepo/issues/478
* Following https://github.com/abacus-network/abacus-monorepo/issues/478, we should update `infra` too ensure that relayers are properly configured with the InterchainGasPaymaster address | 1.0 | Update infra to specify the InterchainGasPaymaster in Relayer settings - * Blocked by:
* https://github.com/abacus-network/abacus-monorepo/issues/478
* Following https://github.com/abacus-network/abacus-monorepo/issues/478, we should update `infra` too ensure that relayers are properly configured with the InterchainGasPaymaster address | infrastructure | update infra to specify the interchaingaspaymaster in relayer settings blocked by following we should update infra too ensure that relayers are properly configured with the interchaingaspaymaster address | 1 |
159,270 | 20,048,346,004 | IssuesEvent | 2022-02-03 01:07:34 | kapseliboi/RedGL2 | https://api.github.com/repos/kapseliboi/RedGL2 | opened | CVE-2021-32804 (High) detected in tar-4.4.8.tgz | security vulnerability | ## CVE-2021-32804 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.8.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- glob-watcher-5.0.3.tgz
- chokidar-2.1.6.tgz
- fsevents-1.2.9.tgz
- node-pre-gyp-0.12.0.tgz
- :x: **tar-4.4.8.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-32804 (High) detected in tar-4.4.8.tgz - ## CVE-2021-32804 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-4.4.8.tgz</b></p></summary>
<p>tar for node</p>
<p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-4.4.8.tgz">https://registry.npmjs.org/tar/-/tar-4.4.8.tgz</a></p>
<p>
Dependency Hierarchy:
- gulp-4.0.2.tgz (Root Library)
- glob-watcher-5.0.3.tgz
- chokidar-2.1.6.tgz
- fsevents-1.2.9.tgz
- node-pre-gyp-0.12.0.tgz
- :x: **tar-4.4.8.tgz** (Vulnerable Library)
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The npm package "tar" (aka node-tar) before versions 6.1.1, 5.0.6, 4.4.14, and 3.3.2 has a arbitrary File Creation/Overwrite vulnerability due to insufficient absolute path sanitization. node-tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the `preservePaths` flag is not set to `true`. This is achieved by stripping the absolute path root from any absolute file paths contained in a tar file. For example `/home/user/.bashrc` would turn into `home/user/.bashrc`. This logic was insufficient when file paths contained repeated path roots such as `////home/user/.bashrc`. `node-tar` would only strip a single path root from such paths. When given an absolute file path with repeating path roots, the resulting path (e.g. `///home/user/.bashrc`) would still resolve to an absolute path, thus allowing arbitrary file creation and overwrite. This issue was addressed in releases 3.2.2, 4.4.14, 5.0.6 and 6.1.1. Users may work around this vulnerability without upgrading by creating a custom `onentry` method which sanitizes the `entry.path` or a `filter` method which removes entries with absolute paths. See referenced GitHub Advisory for details. Be aware of CVE-2021-32803 which fixes a similar bug in later versions of tar.
<p>Publish Date: 2021-08-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32804>CVE-2021-32804</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9">https://github.com/npm/node-tar/security/advisories/GHSA-3jfq-g458-7qm9</a></p>
<p>Release Date: 2021-08-03</p>
<p>Fix Resolution: tar - 3.2.2, 4.4.14, 5.0.6, 6.1.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href dependency hierarchy gulp tgz root library glob watcher tgz chokidar tgz fsevents tgz node pre gyp tgz x tar tgz vulnerable library found in base branch dev vulnerability details the npm package tar aka node tar before versions and has a arbitrary file creation overwrite vulnerability due to insufficient absolute path sanitization node tar aims to prevent extraction of absolute file paths by turning absolute paths into relative paths when the preservepaths flag is not set to true this is achieved by stripping the absolute path root from any absolute file paths contained in a tar file for example home user bashrc would turn into home user bashrc this logic was insufficient when file paths contained repeated path roots such as home user bashrc node tar would only strip a single path root from such paths when given an absolute file path with repeating path roots the resulting path e g home user bashrc would still resolve to an absolute path thus allowing arbitrary file creation and overwrite this issue was addressed in releases and users may work around this vulnerability without upgrading by creating a custom onentry method which sanitizes the entry path or a filter method which removes entries with absolute paths see referenced github advisory for details be aware of cve which fixes a similar bug in later versions of tar publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource | 0 |
717,938 | 24,697,365,653 | IssuesEvent | 2022-10-19 13:05:46 | ant-media/Ant-Media-Server | https://api.github.com/repos/ant-media/Ant-Media-Server | closed | iOS SDK Screen capture is pixelating when screen changes | Priority: Urgent | Being [discussed here](https://github.com/ant-media/Ant-Media-Server/issues/4278)
### Steps to reproduce
1. Run iOS Sample Project and Choose to Publish the screen
2. Play the stream on Ant Media Server with WebRTC
3. Swipe left-right in the screen
### Expected behavior
There should be smooth video play
### Actual behavior
Pixelating the video. Check the recorded video here
https://user-images.githubusercontent.com/3456251/187892806-19fbfea0-d7fa-4898-9bd2-9f8ab8c19f05.mp4
Join [community.antmedia.io](https://community.antmedia.io) and be a part of Ant Media Server Community.
| 1.0 | iOS SDK Screen capture is pixelating when screen changes - Being [discussed here](https://github.com/ant-media/Ant-Media-Server/issues/4278)
### Steps to reproduce
1. Run iOS Sample Project and Choose to Publish the screen
2. Play the stream on Ant Media Server with WebRTC
3. Swipe left-right in the screen
### Expected behavior
There should be smooth video play
### Actual behavior
Pixelating the video. Check the recorded video here
https://user-images.githubusercontent.com/3456251/187892806-19fbfea0-d7fa-4898-9bd2-9f8ab8c19f05.mp4
Join [community.antmedia.io](https://community.antmedia.io) and be a part of Ant Media Server Community.
| non_infrastructure | ios sdk screen capture is pixelating when screen changes being steps to reproduce run ios sample project and choose to publish the screen play the stream on ant media server with webrtc swipe left right in the screen expected behavior there should be smooth video play actual behavior pixelating the video check the recorded video here join and be a part of ant media server community | 0 |
666,686 | 22,363,652,204 | IssuesEvent | 2022-06-15 23:59:07 | AlbertaSat/ex2_ground_station_website | https://api.github.com/repos/AlbertaSat/ex2_ground_station_website | closed | Verify accuracy of Pyorbital against GPredict | enhancement high priority | ### Brief Feature Description
Using the same TLE data (eg. ISS), compare calculated passover times between what Pyorbital calculates and what [GPredict](http://gpredict.oz9aec.net/) calculates.
### Additional Details
If these discrepencies are consistent by an offset, compensate for it in the calculation in automation.py.
Consult with Mission Design about acceptable tolerances.
If the discrepencies are too high, find a way to use GPredict's data instead over Pyorbital as that is our way of determining satellite location (among other things).
### Priority
High | 1.0 | Verify accuracy of Pyorbital against GPredict - ### Brief Feature Description
Using the same TLE data (eg. ISS), compare calculated passover times between what Pyorbital calculates and what [GPredict](http://gpredict.oz9aec.net/) calculates.
### Additional Details
If these discrepencies are consistent by an offset, compensate for it in the calculation in automation.py.
Consult with Mission Design about acceptable tolerances.
If the discrepencies are too high, find a way to use GPredict's data instead over Pyorbital as that is our way of determining satellite location (among other things).
### Priority
High | non_infrastructure | verify accuracy of pyorbital against gpredict brief feature description using the same tle data eg iss compare calculated passover times between what pyorbital calculates and what calculates additional details if these discrepencies are consistent by an offset compensate for it in the calculation in automation py consult with mission design about acceptable tolerances if the discrepencies are too high find a way to use gpredict s data instead over pyorbital as that is our way of determining satellite location among other things priority high | 0 |
4,109 | 4,807,080,688 | IssuesEvent | 2016-11-02 20:24:18 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Move bootstrap build back to MSBuild 15.0 | Area-Infrastructure | Our bootstrap build temporarily needs to go back to MSBuild 14.0. There is a bug in MSBuild 15.0 which prevents us from fully loading our bootstrap components and hence invalidates our build.
https://github.com/Microsoft/msbuild/issues/1183
This bug tracks the removal of work arounds on our end.
| 1.0 | Move bootstrap build back to MSBuild 15.0 - Our bootstrap build temporarily needs to go back to MSBuild 14.0. There is a bug in MSBuild 15.0 which prevents us from fully loading our bootstrap components and hence invalidates our build.
https://github.com/Microsoft/msbuild/issues/1183
This bug tracks the removal of work arounds on our end.
| infrastructure | move bootstrap build back to msbuild our bootstrap build temporarily needs to go back to msbuild there is a bug in msbuild which prevents us from fully loading our bootstrap components and hence invalidates our build this bug tracks the removal of work arounds on our end | 1 |
2,008 | 3,461,657,951 | IssuesEvent | 2015-12-20 08:30:28 | t3kt/vjzual2 | https://api.github.com/repos/t3kt/vjzual2 | closed | get rid of unnecessary module extension classes | cleanup infrastructure | unless it does something other than initialize parameters, remove it.
simplify things and whatnot. | 1.0 | get rid of unnecessary module extension classes - unless it does something other than initialize parameters, remove it.
simplify things and whatnot. | infrastructure | get rid of unnecessary module extension classes unless it does something other than initialize parameters remove it simplify things and whatnot | 1 |
5,030 | 2,760,108,114 | IssuesEvent | 2015-04-28 09:56:12 | MatchmakerExchange/mme-apis | https://api.github.com/repos/MatchmakerExchange/mme-apis | closed | Test data does not conform to v1.0 specification | bug PR submitted Testing | Sigh. Apologies.
1) variant 'referenceName' includes 'chr', which it shouldn't
2) the 'institution' field of contact is misnamed 'organization' | 1.0 | Test data does not conform to v1.0 specification - Sigh. Apologies.
1) variant 'referenceName' includes 'chr', which it shouldn't
2) the 'institution' field of contact is misnamed 'organization' | non_infrastructure | test data does not conform to specification sigh apologies variant referencename includes chr which it shouldn t the institution field of contact is misnamed organization | 0 |
18,565 | 25,828,633,405 | IssuesEvent | 2022-12-12 14:39:41 | ldtteam/Structurize | https://api.github.com/repos/ldtteam/Structurize | closed | Rubidium visual glitch | Compatibility: Mod | ### Is there an existing compatibility issue for this mod?
- [X] I have searched the existing issues.
### Are you using the latest Structurize and your mod Version?
- [X] I am running the latest **alpha** version of Structurize for my Minecraft version.
I am also running the latest versions of other mods that are part of my problem.
### Did you ask on Discord?
- [X] I checked the MineColonies/Structurize [Wiki](https://wiki.minecolonies.ldtteam.com/) or I asked on [discord](https://discord.minecolonies.com/).
### Minecraft Version
1.19
### Structurize Version
1.19.2-1.0.468-ALPHA
### Your Mod Name and Version
rubidium-0.6.2
### Related Mods and their Versions
forge: 43.1.23
minecolonies: 1.19.2-1.0.1137-ALPHA
domum_ornamentum: 1.19-1.0.64-ALPHA-universal
### Current Behavior


when previewing a building or a shape a visual glitch happens
### Expected Behavior
the visual glitch doesnt happen
### Reproduction Steps
1.get build or shape tool
2.right click and try to preview a building or a shape
3.a visual glitch happens
### Logs
https://pastebin.com/d8hG3jkx
### Anything else?
_No response_
### Footer
---
## Viewers
* Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments.
* Add a comment if you have any insights or background information that isn't already part of the conversation.
| True | Rubidium visual glitch - ### Is there an existing compatibility issue for this mod?
- [X] I have searched the existing issues.
### Are you using the latest Structurize and your mod Version?
- [X] I am running the latest **alpha** version of Structurize for my Minecraft version.
I am also running the latest versions of other mods that are part of my problem.
### Did you ask on Discord?
- [X] I checked the MineColonies/Structurize [Wiki](https://wiki.minecolonies.ldtteam.com/) or I asked on [discord](https://discord.minecolonies.com/).
### Minecraft Version
1.19
### Structurize Version
1.19.2-1.0.468-ALPHA
### Your Mod Name and Version
rubidium-0.6.2
### Related Mods and their Versions
forge: 43.1.23
minecolonies: 1.19.2-1.0.1137-ALPHA
domum_ornamentum: 1.19-1.0.64-ALPHA-universal
### Current Behavior


when previewing a building or a shape a visual glitch happens
### Expected Behavior
the visual glitch doesnt happen
### Reproduction Steps
1.get build or shape tool
2.right click and try to preview a building or a shape
3.a visual glitch happens
### Logs
https://pastebin.com/d8hG3jkx
### Anything else?
_No response_
### Footer
---
## Viewers
* Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments.
* Add a comment if you have any insights or background information that isn't already part of the conversation.
| non_infrastructure | rubidium visual glitch is there an existing compatibility issue for this mod i have searched the existing issues are you using the latest structurize and your mod version i am running the latest alpha version of structurize for my minecraft version i am also running the latest versions of other mods that are part of my problem did you ask on discord i checked the minecolonies structurize or i asked on minecraft version structurize version alpha your mod name and version rubidium related mods and their versions forge minecolonies alpha domum ornamentum alpha universal current behavior when previewing a building or a shape a visual glitch happens expected behavior the visual glitch doesnt happen reproduction steps get build or shape tool right click and try to preview a building or a shape a visual glitch happens logs anything else no response footer viewers add a thumbs up to the bug report if you are also affected this helps the bug report become more visible to the team and doesn t clutter the comments add a comment if you have any insights or background information that isn t already part of the conversation | 0 |
658 | 2,824,189,557 | IssuesEvent | 2015-05-21 13:31:47 | dotnet/wcf | https://api.github.com/repos/dotnet/wcf | opened | Provide a mechanism to pass command line parameters into tests | Infrastructure | Our scenario tests will require the ability to redirect the URL of the test service to alternate machines. One can also imagine other test inputs that might feed some [Theory] based test.
One technique to investigate is a pre-build step that accepts MSBuild properties of a known format, generates a temporary .cs file with the information (a Dictionary?), and adds that file to the @(Compile) collection. The tests could be written to consult the dictionary for expected names. Also we could create a [Theory] style test that pulls multiple values. The implementation should not be tied only to scenario tests but available to all tests.
This would be more generally useful than just for WCF, so we should consider working with https://github.com/dotnet/buildtools | 1.0 | Provide a mechanism to pass command line parameters into tests - Our scenario tests will require the ability to redirect the URL of the test service to alternate machines. One can also imagine other test inputs that might feed some [Theory] based test.
One technique to investigate is a pre-build step that accepts MSBuild properties of a known format, generates a temporary .cs file with the information (a Dictionary?), and adds that file to the @(Compile) collection. The tests could be written to consult the dictionary for expected names. Also we could create a [Theory] style test that pulls multiple values. The implementation should not be tied only to scenario tests but available to all tests.
This would be more generally useful than just for WCF, so we should consider working with https://github.com/dotnet/buildtools | infrastructure | provide a mechanism to pass command line parameters into tests our scenario tests will require the ability to redirect the url of the test service to alternate machines one can also imagine other test inputs that might feed some based test one technique to investigate is a pre build step that accepts msbuild properties of a known format generates a temporary cs file with the information a dictionary and adds that file to the compile collection the tests could be written to consult the dictionary for expected names also we could create a style test that pulls multiple values the implementation should not be tied only to scenario tests but available to all tests this would be more generally useful than just for wcf so we should consider working with | 1 |
330,124 | 28,349,894,256 | IssuesEvent | 2023-04-12 01:18:25 | aodn/nrmn-application | https://api.github.com/repos/aodn/nrmn-application | closed | Ordering of sites slows the loading of the List Site page | ready to test planned systest | Noticed a time lag to load the List Site page since the ordering of the site list has been implemented.
| 2.0 | Ordering of sites slows the loading of the List Site page - Noticed a time lag to load the List Site page since the ordering of the site list has been implemented.
| non_infrastructure | ordering of sites slows the loading of the list site page noticed a time lag to load the list site page since the ordering of the site list has been implemented | 0 |
99,821 | 4,065,750,427 | IssuesEvent | 2016-05-26 12:37:35 | vikilabs-code/hombot-issues | https://api.github.com/repos/vikilabs-code/hombot-issues | closed | No themes available theme | high-priority IOS | When you delete all the themes and then you go back and select one touch themes again you get an error no themes available. @artakshar will help you recreate the issue. | 1.0 | No themes available theme - When you delete all the themes and then you go back and select one touch themes again you get an error no themes available. @artakshar will help you recreate the issue. | non_infrastructure | no themes available theme when you delete all the themes and then you go back and select one touch themes again you get an error no themes available artakshar will help you recreate the issue | 0 |
15,004 | 11,297,258,273 | IssuesEvent | 2020-01-17 05:16:28 | Opentrons/opentrons | https://api.github.com/repos/Opentrons/opentrons | closed | Docs: Changelog fixup for 3.15.0 | chore infrastructure | ## overview
This ticket is a placeholder reminder to fix the comparison diff for the 3.14.1 hotfix; It should be compared against 3.14.0 _not_ 3.13.2.
We should also add calibrate to bottom as a current limitation in api v2 beta. | 1.0 | Docs: Changelog fixup for 3.15.0 - ## overview
This ticket is a placeholder reminder to fix the comparison diff for the 3.14.1 hotfix; It should be compared against 3.14.0 _not_ 3.13.2.
We should also add calibrate to bottom as a current limitation in api v2 beta. | infrastructure | docs changelog fixup for overview this ticket is a placeholder reminder to fix the comparison diff for the hotfix it should be compared against not we should also add calibrate to bottom as a current limitation in api beta | 1 |
34,494 | 6,339,686,789 | IssuesEvent | 2017-07-27 08:58:54 | lumapps/lumX | https://api.github.com/repos/lumapps/lumX | opened | Fix AJAX examples | documentation | omdbapi requires a key now so we could try to use something else in our examples? | 1.0 | Fix AJAX examples - omdbapi requires a key now so we could try to use something else in our examples? | non_infrastructure | fix ajax examples omdbapi requires a key now so we could try to use something else in our examples | 0 |
16,941 | 12,152,151,701 | IssuesEvent | 2020-04-24 21:31:38 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | Enable unsafe sysctl (to allow per pod disable of IPv6) | Infrastructure closed | https://trello.com/c/o4U3LtW9/93-enable-unsafe-sysctl-to-allow-per-pod-disable-of-ipv6
Looking to allow pods to disable IPv6.
see the following:
https://access.redhat.com/solutions/3340721
We have requests to be able to disable IPv6 for certain projects/namespaces. | 1.0 | Enable unsafe sysctl (to allow per pod disable of IPv6) - https://trello.com/c/o4U3LtW9/93-enable-unsafe-sysctl-to-allow-per-pod-disable-of-ipv6
Looking to allow pods to disable IPv6.
see the following:
https://access.redhat.com/solutions/3340721
We have requests to be able to disable IPv6 for certain projects/namespaces. | infrastructure | enable unsafe sysctl to allow per pod disable of looking to allow pods to disable see the following we have requests to be able to disable for certain projects namespaces | 1 |
11,240 | 9,014,241,991 | IssuesEvent | 2019-02-05 21:48:17 | webhintio/hint | https://api.github.com/repos/webhintio/hint | closed | Update @types/sinon from 5.0.7 to 7.0.5 [0.5] | area:infrastructure type:new-feature type:testing | # 🚀 Feature request
## Description
PR #1770 is failing. Need to find out why and fix it. | 1.0 | Update @types/sinon from 5.0.7 to 7.0.5 [0.5] - # 🚀 Feature request
## Description
PR #1770 is failing. Need to find out why and fix it. | infrastructure | update types sinon from to 🚀 feature request description pr is failing need to find out why and fix it | 1 |
16,879 | 12,152,145,520 | IssuesEvent | 2020-04-24 21:30:47 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | WARN: Low Metrics Cassandra Space - 15% free | Infrastructure closed medium priority | https://trello.com/c/T5f2yg80/164-warn-low-metrics-cassandra-space-15-free
Cassandra is slowly filling up as the platform grows and needs to have its storage expanded. | 1.0 | WARN: Low Metrics Cassandra Space - 15% free - https://trello.com/c/T5f2yg80/164-warn-low-metrics-cassandra-space-15-free
Cassandra is slowly filling up as the platform grows and needs to have its storage expanded. | infrastructure | warn low metrics cassandra space free cassandra is slowly filling up as the platform grows and needs to have its storage expanded | 1 |
334,118 | 10,136,151,191 | IssuesEvent | 2019-08-02 12:14:07 | CheckiO/EoC-battle | https://api.github.com/repos/CheckiO/EoC-battle | closed | Throw async errors from REF to EXEC | complex:middle priority:normal section:referee | Not every error we can find on Client Side so we should make a say to send an error from referee to Client so the user can see the error in console during the battle review.
| 1.0 | Throw async errors from REF to EXEC - Not every error we can find on Client Side so we should make a say to send an error from referee to Client so the user can see the error in console during the battle review.
| non_infrastructure | throw async errors from ref to exec not every error we can find on client side so we should make a say to send an error from referee to client so the user can see the error in console during the battle review | 0 |
144,768 | 13,125,346,421 | IssuesEvent | 2020-08-06 06:28:12 | RedHatInsights/insights-results-aggregator-mock | https://api.github.com/repos/RedHatInsights/insights-results-aggregator-mock | closed | Update pull request template to be compatible with other repositories in a project | documentation | Update pull request template to be compatible with other repositories in a project | 1.0 | Update pull request template to be compatible with other repositories in a project - Update pull request template to be compatible with other repositories in a project | non_infrastructure | update pull request template to be compatible with other repositories in a project update pull request template to be compatible with other repositories in a project | 0 |
15,642 | 11,631,944,535 | IssuesEvent | 2020-02-28 03:12:02 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Need some extra math methods for a weather data project | interface/infrastructure newfeature | In particular I need AverageMonthlyAverages and YearlyAverages | 1.0 | Need some extra math methods for a weather data project - In particular I need AverageMonthlyAverages and YearlyAverages | infrastructure | need some extra math methods for a weather data project in particular i need averagemonthlyaverages and yearlyaverages | 1 |
22,828 | 15,531,502,127 | IssuesEvent | 2021-03-14 00:01:07 | algorand/go-algorand | https://api.github.com/repos/algorand/go-algorand | closed | TEAL - arguments counts and traps | Infrastructure TEAL 3 | Application contracts can check the number of arguments they have been provided via `NumAppArgs`, but normal arguments do not have a similar field or opcode to look at.
For application contracts, there's no discussion in the docs about what happens if you attempt to read an argument after this limit. I expect it should trap. For non-application contracts, I'd expect the same behavior, but again, this is not discussed.
Finally, application contracts in the same group can read each other's arguments via `gtxna` but non-application contracts cannot read each other's arguments (nor can application contracts read the arguments of non-application contracts in the same group.) | 1.0 | TEAL - arguments counts and traps - Application contracts can check the number of arguments they have been provided via `NumAppArgs`, but normal arguments do not have a similar field or opcode to look at.
For application contracts, there's no discussion in the docs about what happens if you attempt to read an argument after this limit. I expect it should trap. For non-application contracts, I'd expect the same behavior, but again, this is not discussed.
Finally, application contracts in the same group can read each other's arguments via `gtxna` but non-application contracts cannot read each other's arguments (nor can application contracts read the arguments of non-application contracts in the same group.) | infrastructure | teal arguments counts and traps application contracts can check the number of arguments they have been provided via numappargs but normal arguments do not have a similar field or opcode to look at for application contracts there s no discussion in the docs about what happens if you attempt to read an argument after this limit i expect it should trap for non application contracts i d expect the same behavior but again this is not discussed finally application contracts in the same group can read each other s arguments via gtxna but non application contracts cannot read each other s arguments nor can application contracts read the arguments of non application contracts in the same group | 1 |
9,821 | 8,183,520,506 | IssuesEvent | 2018-08-29 09:18:07 | coq/coq | https://api.github.com/repos/coq/coq | closed | Coq makefile: print-pretty-timed-diff mixes up "before" and "after" | kind: infrastructure | #### Version
8.8
#### Operating system
NixOS
#### Description of the problem
The makefile template passes arguments into `make-both-time-files.py` in the opposite order they should be in. [From `CoqMakefile.in`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/CoqMakefile.in#L346):
```make
COQMAKE_BOTH_TIME_FILES ?= "$(COQLIB)/tools/make-both-time-files.py"
print-pretty-timed-diff::
$(HIDE)$(COQMAKE_BOTH_TIME_FILES) --sort-by=$(TIMING_SORT_BY) $(TIME_OF_BUILD_BEFORE_FILE) $(TIME_OF_BUILD_AFTER_FILE) $(TIME_OF_PRETTY_BOTH_BUILD_FILE) $(TIME_OF_PRETTY_BUILD_EXTRA_FILES)
```
but [from `make-both-time-files.py`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/make-both-time-files.py#L6):
```python
USAGE = 'Usage: %s [--sort-by=auto|absolute|diff] AFTER_FILE_NAME BEFORE_FILE_NAME [OUTPUT_FILE_NAME ..]' % sys.argv[0]
```
cc: @JasonGross | 1.0 | Coq makefile: print-pretty-timed-diff mixes up "before" and "after" - #### Version
8.8
#### Operating system
NixOS
#### Description of the problem
The makefile template passes arguments into `make-both-time-files.py` in the opposite order they should be in. [From `CoqMakefile.in`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/CoqMakefile.in#L346):
```make
COQMAKE_BOTH_TIME_FILES ?= "$(COQLIB)/tools/make-both-time-files.py"
print-pretty-timed-diff::
$(HIDE)$(COQMAKE_BOTH_TIME_FILES) --sort-by=$(TIMING_SORT_BY) $(TIME_OF_BUILD_BEFORE_FILE) $(TIME_OF_BUILD_AFTER_FILE) $(TIME_OF_PRETTY_BOTH_BUILD_FILE) $(TIME_OF_PRETTY_BUILD_EXTRA_FILES)
```
but [from `make-both-time-files.py`](https://github.com/coq/coq/blob/a44b4be1bf8b96e941216cd10cfb5981a825c3fa/tools/make-both-time-files.py#L6):
```python
USAGE = 'Usage: %s [--sort-by=auto|absolute|diff] AFTER_FILE_NAME BEFORE_FILE_NAME [OUTPUT_FILE_NAME ..]' % sys.argv[0]
```
cc: @JasonGross | infrastructure | coq makefile print pretty timed diff mixes up before and after version operating system nixos description of the problem the makefile template passes arguments into make both time files py in the opposite order they should be in make coqmake both time files coqlib tools make both time files py print pretty timed diff hide coqmake both time files sort by timing sort by time of build before file time of build after file time of pretty both build file time of pretty build extra files but python usage usage s after file name before file name sys argv cc jasongross | 1 |
5 | 2,490,813,283 | IssuesEvent | 2015-01-02 20:13:41 | Jasig/cas | https://api.github.com/repos/Jasig/cas | closed | CAS/Environment info in the logs | 3.5.0 4.0 Configuration Infrastructure Task Trivial | When troubleshooting CAS and reviewing logs, it useful to know the following to better understand the cause of issue (Figuring out this data sometimes is difficult for the novice adopter!)
- CAS server version
- Java version, vendor, vm info
- OS version, vendor, etc
- Possibly the container info via the ServletContext
These would show up under DEBUG/TRACE levels on startup. It would also be cool if CAS could output its own banner like spring boot :)
https://github.com/spring-projects/spring-boot/blob/master/spring-boot/src/main/java/org/springframework/boot/SpringBootBanner.java
| 1.0 | CAS/Environment info in the logs - When troubleshooting CAS and reviewing logs, it useful to know the following to better understand the cause of issue (Figuring out this data sometimes is difficult for the novice adopter!)
- CAS server version
- Java version, vendor, vm info
- OS version, vendor, etc
- Possibly the container info via the ServletContext
These would show up under DEBUG/TRACE levels on startup. It would also be cool if CAS could output its own banner like spring boot :)
https://github.com/spring-projects/spring-boot/blob/master/spring-boot/src/main/java/org/springframework/boot/SpringBootBanner.java
| infrastructure | cas environment info in the logs when troubleshooting cas and reviewing logs it useful to know the following to better understand the cause of issue figuring out this data sometimes is difficult for the novice adopter cas server version java version vendor vm info os version vendor etc possibly the container info via the servletcontext these would show up under debug trace levels on startup it would also be cool if cas could output its own banner like spring boot | 1 |
348,973 | 10,455,289,359 | IssuesEvent | 2019-09-19 20:56:28 | kubeapps/kubeapps | https://api.github.com/repos/kubeapps/kubeapps | closed | Kubeapps shows "default" namespace if user cannot list namespaces | component/dashboard good first issue kind/bug priority/low size/S | If a service account is created for a user in a specific namespace (and without permission to list namespaces), and the associated credential is used to login, the namespace switcher shows "default" rather than the specific namespace.
Kubeapps should default to the namespace of the service account rather than the "default" namespace which to which the user may not have access.
Ideally, kubeapps would remember the selected namespace so when returning (without re-authing) it remains selected. | 1.0 | Kubeapps shows "default" namespace if user cannot list namespaces - If a service account is created for a user in a specific namespace (and without permission to list namespaces), and the associated credential is used to login, the namespace switcher shows "default" rather than the specific namespace.
Kubeapps should default to the namespace of the service account rather than the "default" namespace which to which the user may not have access.
Ideally, kubeapps would remember the selected namespace so when returning (without re-authing) it remains selected. | non_infrastructure | kubeapps shows default namespace if user cannot list namespaces if a service account is created for a user in a specific namespace and without permission to list namespaces and the associated credential is used to login the namespace switcher shows default rather than the specific namespace kubeapps should default to the namespace of the service account rather than the default namespace which to which the user may not have access ideally kubeapps would remember the selected namespace so when returning without re authing it remains selected | 0 |
34,524 | 30,053,867,673 | IssuesEvent | 2023-06-28 04:29:12 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | closed | Automate Changelog Generation | tool: monorepo infrastructure | **Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
In order to make working with changelogs easier, we should implement a feature for automatically generating them from the PR description. We should add a GitHub Action that:
- Validation
- The PR description should have a section for entering the changelog.
- If the changelog does not follow the correct format we can fail a PR check.
- PRs should NOT require a changelog for things that won't ever be published. Markdown files, `.github`, configuration like ESLint, etc. Only code changes that are shipped to users should trigger it.
- If there are already changelog files present we should validate them using the JP changelogger and error if they're invalid.
- Parsing
- Any `push` or editing of the PR description should trigger this.
- **Instead, only run prior to the merge taking place. This avoids conflicts with existing changelog files.**
- Parse the format so we can construct a JP Changelogger command from it.
- Scan for any projects that have been changed (we might even be able to use `pnpm --filter=trunk...HEAD`?) and check if they've got files that have changed that are valid.
- If they do and there isn't already a changelog file from this PR, run the command we generated and add the one from the PR description.
- Remove local changelog check and make the generation optional outside of the PR checks.
As is standard now, we should put whatever makes sense into the `monorepo-utils` CLI tool. I might suggest it going as far as using GitHub's API to review the PR's description and avoid needing to do any parsing in the GitHub Action. Maybe instead of using the JP changelogger to make the changelog we can just use [this script](https://gist.github.com/rrennick/56285a3d3580aa1c5d38d50d9ba5a6cd).
Given the scope of this issue, it may even make sense to just remove the JP changelogger entirely? There's some meat to discuss here. | 1.0 | Automate Changelog Generation - **Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
In order to make working with changelogs easier, we should implement a feature for automatically generating them from the PR description. We should add a GitHub Action that:
- Validation
- The PR description should have a section for entering the changelog.
- If the changelog does not follow the correct format we can fail a PR check.
- PRs should NOT require a changelog for things that won't ever be published. Markdown files, `.github`, configuration like ESLint, etc. Only code changes that are shipped to users should trigger it.
- If there are already changelog files present we should validate them using the JP changelogger and error if they're invalid.
- Parsing
- Any `push` or editing of the PR description should trigger this.
- **Instead, only run prior to the merge taking place. This avoids conflicts with existing changelog files.**
- Parse the format so we can construct a JP Changelogger command from it.
- Scan for any projects that have been changed (we might even be able to use `pnpm --filter=trunk...HEAD`?) and check if they've got files that have changed that are valid.
- If they do and there isn't already a changelog file from this PR, run the command we generated and add the one from the PR description.
- Remove local changelog check and make the generation optional outside of the PR checks.
As is standard now, we should put whatever makes sense into the `monorepo-utils` CLI tool. I might suggest it going as far as using GitHub's API to review the PR's description and avoid needing to do any parsing in the GitHub Action. Maybe instead of using the JP changelogger to make the changelog we can just use [this script](https://gist.github.com/rrennick/56285a3d3580aa1c5d38d50d9ba5a6cd).
Given the scope of this issue, it may even make sense to just remove the JP changelogger entirely? There's some meat to discuss here. | infrastructure | automate changelog generation prerequisites mark completed items with an i have checked that my issue type is not listed here my issue is not a security issue support request bug report enhancement or feature request please use the link above if it is issue description in order to make working with changelogs easier we should implement a feature for automatically generating them from the pr description we should add a github action that validation the pr description should have a section for entering the changelog if the changelog does not follow the correct format we can fail a pr check prs should not require a changelog for things that won t ever be published markdown files github configuration like eslint etc only code changes that are shipped to users should trigger it if there are already changelog files present we should validate them using the jp changelogger and error if they re invalid parsing any push or editing of the pr description should trigger this instead only run prior to the merge taking place this avoids conflicts with existing changelog files parse the format so we can construct a jp changelogger command from it scan for any projects that have been changed we might even be able to use pnpm filter trunk head and check if they ve got files that have changed that are valid if they do and there isn t already a changelog file from this pr run the command we generated and add the one from the pr description remove local changelog check and make the generation optional outside of the pr checks as is standard now we should put whatever makes sense into the monorepo utils cli tool i might suggest it going as far as using github s api to review the pr s description and avoid needing to do any parsing in the github action maybe instead of using the jp changelogger to make the changelog we can just use given the scope of this issue it may even make sense to just remove the jp changelogger entirely there s some meat to discuss here | 1 |
462,880 | 13,255,601,134 | IssuesEvent | 2020-08-20 11:14:10 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.cbs.com - Video is not supported on Opera browser | browser-opera priority-normal severity-critical | <!-- @browser: Opera 63.0.3368 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.80 Safari/537.36 OPR/63.0.3368.22 (Edition beta) -->
<!-- @reported_with: -->
**URL**: https://www.cbs.com/shows/cbs_this_morning/video/cK6tv4NFGkyNVVjonYuUnjEg5oRIxE2_/how-jeffrey-epstein-s-accusers-are-reacting-to-his-apparent-suicide/
**Browser / Version**: Opera 63.0.3368
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Video or audio doesn't play
**Description**: Website puts up notice to use either Chrome or Firefox when trying to watch video
**Steps to Reproduce**:
I subscribe to CBS All Access. It seems the site is browser sniffing, not letting you watch video in either Opera Stable or Beta, even when switching the user agent to Chrome or Firefox.
[](https://webcompat.com/uploads/2019/8/7c728241-3c25-46ee-8db3-65e452238874.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.cbs.com - Video is not supported on Opera browser - <!-- @browser: Opera 63.0.3368 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/76.0.3809.80 Safari/537.36 OPR/63.0.3368.22 (Edition beta) -->
<!-- @reported_with: -->
**URL**: https://www.cbs.com/shows/cbs_this_morning/video/cK6tv4NFGkyNVVjonYuUnjEg5oRIxE2_/how-jeffrey-epstein-s-accusers-are-reacting-to-his-apparent-suicide/
**Browser / Version**: Opera 63.0.3368
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Video or audio doesn't play
**Description**: Website puts up notice to use either Chrome or Firefox when trying to watch video
**Steps to Reproduce**:
I subscribe to CBS All Access. It seems the site is browser sniffing, not letting you watch video in either Opera Stable or Beta, even when switching the user agent to Chrome or Firefox.
[](https://webcompat.com/uploads/2019/8/7c728241-3c25-46ee-8db3-65e452238874.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_infrastructure | video is not supported on opera browser url browser version opera operating system windows tested another browser yes problem type video or audio doesn t play description website puts up notice to use either chrome or firefox when trying to watch video steps to reproduce i subscribe to cbs all access it seems the site is browser sniffing not letting you watch video in either opera stable or beta even when switching the user agent to chrome or firefox browser configuration none from with ❤️ | 0 |
3,060 | 4,031,893,140 | IssuesEvent | 2016-05-18 18:43:10 | trackpete/exiletools-indexer | https://api.github.com/repos/trackpete/exiletools-indexer | closed | v4 New ExileTools Infrastructure | infrastructure / exiletools.com setup | This issue will serve as a placeholder for the plans to implement the new Infrastructure required for production deployment of both the ExileTools v4 Indexer as well as the associated production infrastructure.
This will give interested parties additional insight into the current production environment as well as illustrating how the organic growth of ExileTools and the adoption of various tools over time has led to a inefficient infrastructure.
# Current Infrastructure Overview
## Hardware / Software
### pwx
* Primary Server
* 64GB RAM, 16 cpu cores
* ~800GB of compressed + deduplicated data on NFS
Back End Services include:
1. MariaDB (only used for Ladder API and league information at this time)
2. Jenkins (manages execution of Ladder API updates)
3. ElasticSearch Cluster Master (stats, items, and other indexes)
4. All Indexer related pipeline tools, such as river-watch, etc.
Front End Services include:
1. Apache, serves ALL primary web content for exiletools.com as well as the older Ladder API front end
2. Varnish (all requests to Apache go through Varnish for front end caching)
### pwx2
* Secondary Server
* 64GB RAM, 16 cpu cores
Back End Services include:
1. ElasticSearch Cluster Secondary. Kibana is hosted here as well.
2. Tyk.io API Gateway (handles incoming API requests by firing them off at localhost:9200 for ElasticSearch). Redis and MongoDB are used by Tyk.
Front End Services include:
1. nginx for any incoming API requests. nginx modifies basic http auth into something Tyk can read as well as provides a short front-end cache for all API requests (including POSTS) based on checksums of the request payload.
### pwx3
* Secondary Server, ES only
* 48GB RAM, 12 cpu cores
Back End Services include:
1. ElasticSearch Cluster member. Only purpose for this server, by adding another 30GB JVM it increases the amount of items stored in memory on the shards.
### haproxy
* Small VM running front end software load balancer instance which provides SSL offload as well as routing to either Varnish on pwx for primary requests or Nginx on pwx2 for ES Index requests. All external inbound traffic first passes through this system.
## ElasticSearch
1. Runs on NFS - slow!
2. Three 30GB JVM's spread across three machines, with each index having two shards per JVM, means the indexes can hold ~40GB of data in memory between them while still supporting full redundancy in the event of node failure. This is not enough memory to manage a long running item index.
3. Internal programs access the ES cluster directly, while external programs must go through haproxy->nginx->tyk->ES.
# The Future / Planned Changes
These changes will be deployed into production at some point after or while deploying the v4 Indexer Pipeline. I may wait to make full infrastructure stack changes until the Perandus leagues are over. I will announce my plans on twitter when they are nailed down.
1. New Backend Machine: 96GB RAM, 24 cpu cores, with 2x240GB SSD's. This will be the *only* ElasticSearch machine moving forward, with two JVM's running on it. It will be a single point of failure, but the performance gains are worth it. This machine will also run Kafka and the entire v4 Indexer Pipeline.
2. New Hybrid Services Machine: Probably a small 16GB VM with 8 cores or so. This machine will run the Tyk API Gateway (2.0), Apache, and a very small MariaDB instance for the ladder (which really needs to be transitioned to ElasticSearch).
3. New Front-end Machine: Probably a small 16GB VM with 8 cores or so. This machine will run haproxy as well as a Varnish cache for Apache and an nginx cache for Tyk (why nginx? it's just sooo much easier to handle caching for POST requests in nginx).
# Other Big Changes
I think that when I move to Tyk 2.0 I will drop the requirement for API keys and authorization headers and just fully open the index - instead, I will just apply rate limiting by end user IP address. The main reason I originally asked for people to sign up for API keys was so that I had a list of the userbase, but at this point there are something like 500+ API keys out there and less than 20 are in active daily use, so it's not very useful or accurate.
I'm also considering picking up some "new" hardware off ebay instead of just deploying on one primary box. Let's see where this goes first though.
| 1.0 | v4 New ExileTools Infrastructure - This issue will serve as a placeholder for the plans to implement the new Infrastructure required for production deployment of both the ExileTools v4 Indexer as well as the associated production infrastructure.
This will give interested parties additional insight into the current production environment as well as illustrating how the organic growth of ExileTools and the adoption of various tools over time has led to a inefficient infrastructure.
# Current Infrastructure Overview
## Hardware / Software
### pwx
* Primary Server
* 64GB RAM, 16 cpu cores
* ~800GB of compressed + deduplicated data on NFS
Back End Services include:
1. MariaDB (only used for Ladder API and league information at this time)
2. Jenkins (manages execution of Ladder API updates)
3. ElasticSearch Cluster Master (stats, items, and other indexes)
4. All Indexer related pipeline tools, such as river-watch, etc.
Front End Services include:
1. Apache, serves ALL primary web content for exiletools.com as well as the older Ladder API front end
2. Varnish (all requests to Apache go through Varnish for front end caching)
### pwx2
* Secondary Server
* 64GB RAM, 16 cpu cores
Back End Services include:
1. ElasticSearch Cluster Secondary. Kibana is hosted here as well.
2. Tyk.io API Gateway (handles incoming API requests by firing them off at localhost:9200 for ElasticSearch). Redis and MongoDB are used by Tyk.
Front End Services include:
1. nginx for any incoming API requests. nginx modifies basic http auth into something Tyk can read as well as provides a short front-end cache for all API requests (including POSTS) based on checksums of the request payload.
### pwx3
* Secondary Server, ES only
* 48GB RAM, 12 cpu cores
Back End Services include:
1. ElasticSearch Cluster member. Only purpose for this server, by adding another 30GB JVM it increases the amount of items stored in memory on the shards.
### haproxy
* Small VM running front end software load balancer instance which provides SSL offload as well as routing to either Varnish on pwx for primary requests or Nginx on pwx2 for ES Index requests. All external inbound traffic first passes through this system.
## ElasticSearch
1. Runs on NFS - slow!
2. Three 30GB JVM's spread across three machines, with each index having two shards per JVM, means the indexes can hold ~40GB of data in memory between them while still supporting full redundancy in the event of node failure. This is not enough memory to manage a long running item index.
3. Internal programs access the ES cluster directly, while external programs must go through haproxy->nginx->tyk->ES.
# The Future / Planned Changes
These changes will be deployed into production at some point after or while deploying the v4 Indexer Pipeline. I may wait to make full infrastructure stack changes until the Perandus leagues are over. I will announce my plans on twitter when they are nailed down.
1. New Backend Machine: 96GB RAM, 24 cpu cores, with 2x240GB SSD's. This will be the *only* ElasticSearch machine moving forward, with two JVM's running on it. It will be a single point of failure, but the performance gains are worth it. This machine will also run Kafka and the entire v4 Indexer Pipeline.
2. New Hybrid Services Machine: Probably a small 16GB VM with 8 cores or so. This machine will run the Tyk API Gateway (2.0), Apache, and a very small MariaDB instance for the ladder (which really needs to be transitioned to ElasticSearch).
3. New Front-end Machine: Probably a small 16GB VM with 8 cores or so. This machine will run haproxy as well as a Varnish cache for Apache and an nginx cache for Tyk (why nginx? it's just sooo much easier to handle caching for POST requests in nginx).
# Other Big Changes
I think that when I move to Tyk 2.0 I will drop the requirement for API keys and authorization headers and just fully open the index - instead, I will just apply rate limiting by end user IP address. The main reason I originally asked for people to sign up for API keys was so that I had a list of the userbase, but at this point there are something like 500+ API keys out there and less than 20 are in active daily use, so it's not very useful or accurate.
I'm also considering picking up some "new" hardware off ebay instead of just deploying on one primary box. Let's see where this goes first though.
| infrastructure | new exiletools infrastructure this issue will serve as a placeholder for the plans to implement the new infrastructure required for production deployment of both the exiletools indexer as well as the associated production infrastructure this will give interested parties additional insight into the current production environment as well as illustrating how the organic growth of exiletools and the adoption of various tools over time has led to a inefficient infrastructure current infrastructure overview hardware software pwx primary server ram cpu cores of compressed deduplicated data on nfs back end services include mariadb only used for ladder api and league information at this time jenkins manages execution of ladder api updates elasticsearch cluster master stats items and other indexes all indexer related pipeline tools such as river watch etc front end services include apache serves all primary web content for exiletools com as well as the older ladder api front end varnish all requests to apache go through varnish for front end caching secondary server ram cpu cores back end services include elasticsearch cluster secondary kibana is hosted here as well tyk io api gateway handles incoming api requests by firing them off at localhost for elasticsearch redis and mongodb are used by tyk front end services include nginx for any incoming api requests nginx modifies basic http auth into something tyk can read as well as provides a short front end cache for all api requests including posts based on checksums of the request payload secondary server es only ram cpu cores back end services include elasticsearch cluster member only purpose for this server by adding another jvm it increases the amount of items stored in memory on the shards haproxy small vm running front end software load balancer instance which provides ssl offload as well as routing to either varnish on pwx for primary requests or nginx on for es index requests all external inbound traffic first passes through this system elasticsearch runs on nfs slow three jvm s spread across three machines with each index having two shards per jvm means the indexes can hold of data in memory between them while still supporting full redundancy in the event of node failure this is not enough memory to manage a long running item index internal programs access the es cluster directly while external programs must go through haproxy nginx tyk es the future planned changes these changes will be deployed into production at some point after or while deploying the indexer pipeline i may wait to make full infrastructure stack changes until the perandus leagues are over i will announce my plans on twitter when they are nailed down new backend machine ram cpu cores with ssd s this will be the only elasticsearch machine moving forward with two jvm s running on it it will be a single point of failure but the performance gains are worth it this machine will also run kafka and the entire indexer pipeline new hybrid services machine probably a small vm with cores or so this machine will run the tyk api gateway apache and a very small mariadb instance for the ladder which really needs to be transitioned to elasticsearch new front end machine probably a small vm with cores or so this machine will run haproxy as well as a varnish cache for apache and an nginx cache for tyk why nginx it s just sooo much easier to handle caching for post requests in nginx other big changes i think that when i move to tyk i will drop the requirement for api keys and authorization headers and just fully open the index instead i will just apply rate limiting by end user ip address the main reason i originally asked for people to sign up for api keys was so that i had a list of the userbase but at this point there are something like api keys out there and less than are in active daily use so it s not very useful or accurate i m also considering picking up some new hardware off ebay instead of just deploying on one primary box let s see where this goes first though | 1 |
671,484 | 22,763,143,062 | IssuesEvent | 2022-07-07 23:41:56 | touretzkyds/KnowledgeGraphDemo | https://api.github.com/repos/touretzkyds/KnowledgeGraphDemo | closed | taxon name input cleanup | bug high priority | We need to clean up the input we receive in the taxon name text box. Delete any leading or trailing whitespace. Replace any internal whitespace (spaces or tabs) with a single space. Otherwise the name lookup will fail, as just happened to me when I left a trailing space on a genus name. | 1.0 | taxon name input cleanup - We need to clean up the input we receive in the taxon name text box. Delete any leading or trailing whitespace. Replace any internal whitespace (spaces or tabs) with a single space. Otherwise the name lookup will fail, as just happened to me when I left a trailing space on a genus name. | non_infrastructure | taxon name input cleanup we need to clean up the input we receive in the taxon name text box delete any leading or trailing whitespace replace any internal whitespace spaces or tabs with a single space otherwise the name lookup will fail as just happened to me when i left a trailing space on a genus name | 0 |
10,788 | 8,722,281,795 | IssuesEvent | 2018-12-09 10:48:36 | coq/coq | https://api.github.com/repos/coq/coq | closed | Change extensions of log files in WIndows build to _log.txt and _err.txt so that they can be viewed immediately in gitlab | kind: infrastructure platform: Windows priority: low | In the windows build log files are now browsable with artifacts browse, but it is not possible to look at log files in the browser cause of the extension. So change the file extension to .txt for all .log and .err files. | 1.0 | Change extensions of log files in WIndows build to _log.txt and _err.txt so that they can be viewed immediately in gitlab - In the windows build log files are now browsable with artifacts browse, but it is not possible to look at log files in the browser cause of the extension. So change the file extension to .txt for all .log and .err files. | infrastructure | change extensions of log files in windows build to log txt and err txt so that they can be viewed immediately in gitlab in the windows build log files are now browsable with artifacts browse but it is not possible to look at log files in the browser cause of the extension so change the file extension to txt for all log and err files | 1 |
34,985 | 30,656,790,404 | IssuesEvent | 2023-07-25 12:41:24 | Sinyks/AMT2020-StackOverflowSimpleVersion | https://api.github.com/repos/Sinyks/AMT2020-StackOverflowSimpleVersion | closed | add business logic for tags | Domain infrastructure | - [x] domain
- [x] database (repository)
- [ ] management facade
- [ ] tags DTO | 1.0 | add business logic for tags - - [x] domain
- [x] database (repository)
- [ ] management facade
- [ ] tags DTO | infrastructure | add business logic for tags domain database repository management facade tags dto | 1 |
358,535 | 10,618,538,470 | IssuesEvent | 2019-10-13 05:31:51 | k8smeetup/website-tasks | https://api.github.com/repos/k8smeetup/website-tasks | opened | /docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md | lang/zh priority/P0 sync/update version/1.16 welcome | Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md)
Diff 查看原始文档更新差异命令:
```bash
git diff release-1.14 release-1.16 -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md
``` | 1.0 | /docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md - Source File: [/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md](https://github.com/kubernetes/website/blob/release-1.16/content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md)
Diff 查看原始文档更新差异命令:
```bash
git diff release-1.14 release-1.16 -- content/en/docs/reference/setup-tools/kubeadm/generated/kubeadm_alpha.md
``` | non_infrastructure | docs reference setup tools kubeadm generated kubeadm alpha md source file diff 查看原始文档更新差异命令 bash git diff release release content en docs reference setup tools kubeadm generated kubeadm alpha md | 0 |
252 | 2,595,806,656 | IssuesEvent | 2015-02-20 16:35:54 | Starcounter/Starcounter | https://api.github.com/repos/Starcounter/Starcounter | closed | Out of memory error for three running databases on build server | Follow-up Infrastructure P/High stability issue | I try to run my regression test with three simple apps on the build server. Each app creates own database. When third database creates a database I get the error (see, e.g., [this build](https://scbuildserver/viewLog.html?buildId=24172&buildTypeId=Starcounter_DevelopDaily&tab=buildLog):
>[14:41:05][Step 14/42] [14:41:04, Critical, Starcounter (selectonnull2362db)]
[14:41:05][Step 14/42] System.OutOfMemoryException: ScErrOutOfMemory (SCERR1003): Operation failed because needed memory couldn't be allocated.
[14:41:05][Step 14/42] Version: 2.0.8424.2.
[14:41:05][Step 14/42] Help page: https://github.com/Starcounter/Starcounter/wiki/SCERR1003.
[14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.ConnectDatabase(UInt32 schedulerCount, UInt64 hmenv, UInt64 hlogs) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 557
[14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Setup(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 208
[14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Main(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 55
[14:41:05][Step 14/42] HResult=-2147024882
[14:41:05][Step 14/42] HelpLink=https://github.com/Starcounter/Starcounter/wiki/SCERR1003
Currently I disabled the third test app, since it will fail nightly build due to this problem. After I fix #2362 I will return to this one and describe how to reproduce.
If anyone can hint about the problem, post here. | 1.0 | Out of memory error for three running databases on build server - I try to run my regression test with three simple apps on the build server. Each app creates own database. When third database creates a database I get the error (see, e.g., [this build](https://scbuildserver/viewLog.html?buildId=24172&buildTypeId=Starcounter_DevelopDaily&tab=buildLog):
>[14:41:05][Step 14/42] [14:41:04, Critical, Starcounter (selectonnull2362db)]
[14:41:05][Step 14/42] System.OutOfMemoryException: ScErrOutOfMemory (SCERR1003): Operation failed because needed memory couldn't be allocated.
[14:41:05][Step 14/42] Version: 2.0.8424.2.
[14:41:05][Step 14/42] Help page: https://github.com/Starcounter/Starcounter/wiki/SCERR1003.
[14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.ConnectDatabase(UInt32 schedulerCount, UInt64 hmenv, UInt64 hlogs) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 557
[14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Setup(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 208
[14:41:05][Step 14/42] at StarcounterInternal.Bootstrap.Control.Main(String[] args) in c:\TeamCity\TeamCity8\buildAgent\work\build-daily-develop-2.0.8424.2\Level1\src\Starcounter.Bootstrap\Control.cs:line 55
[14:41:05][Step 14/42] HResult=-2147024882
[14:41:05][Step 14/42] HelpLink=https://github.com/Starcounter/Starcounter/wiki/SCERR1003
Currently I disabled the third test app, since it will fail nightly build due to this problem. After I fix #2362 I will return to this one and describe how to reproduce.
If anyone can hint about the problem, post here. | infrastructure | out of memory error for three running databases on build server i try to run my regression test with three simple apps on the build server each app creates own database when third database creates a database i get the error see e g system outofmemoryexception scerroutofmemory operation failed because needed memory couldn t be allocated version help page at starcounterinternal bootstrap control connectdatabase schedulercount hmenv hlogs in c teamcity buildagent work build daily develop src starcounter bootstrap control cs line at starcounterinternal bootstrap control setup string args in c teamcity buildagent work build daily develop src starcounter bootstrap control cs line at starcounterinternal bootstrap control main string args in c teamcity buildagent work build daily develop src starcounter bootstrap control cs line hresult helplink currently i disabled the third test app since it will fail nightly build due to this problem after i fix i will return to this one and describe how to reproduce if anyone can hint about the problem post here | 1 |
281 | 2,610,651,253 | IssuesEvent | 2015-02-26 21:38:41 | uProxy/uproxy | https://api.github.com/repos/uProxy/uproxy | opened | set setBufferedLogFilter to help reduce memory usage | C:Infrastructure C:Networking P1 | Although we call `setConsoleFilter`, we are not setting `setBufferedLogFilter`.
This may be leading to *lots* of debug messages building up in RAM. Let's set it and see what difference it makes.
Here's where it should be set in Simple SOCKS:
https://github.com/uProxy/uproxy-networking/blob/dev/src/simple-socks/freedom-module.ts#L13 | 1.0 | set setBufferedLogFilter to help reduce memory usage - Although we call `setConsoleFilter`, we are not setting `setBufferedLogFilter`.
This may be leading to *lots* of debug messages building up in RAM. Let's set it and see what difference it makes.
Here's where it should be set in Simple SOCKS:
https://github.com/uProxy/uproxy-networking/blob/dev/src/simple-socks/freedom-module.ts#L13 | infrastructure | set setbufferedlogfilter to help reduce memory usage although we call setconsolefilter we are not setting setbufferedlogfilter this may be leading to lots of debug messages building up in ram let s set it and see what difference it makes here s where it should be set in simple socks | 1 |
26,372 | 11,300,324,782 | IssuesEvent | 2020-01-17 13:22:13 | aleksaa01/WorkTodo | https://api.github.com/repos/aleksaa01/WorkTodo | closed | Storage is not cleaned up upon logout | bug security | When you logout, token in storage is set to null, but cards, tasks and preferences are not deleted.
This is obviously a security flaw. | True | Storage is not cleaned up upon logout - When you logout, token in storage is set to null, but cards, tasks and preferences are not deleted.
This is obviously a security flaw. | non_infrastructure | storage is not cleaned up upon logout when you logout token in storage is set to null but cards tasks and preferences are not deleted this is obviously a security flaw | 0 |
13,311 | 10,198,237,403 | IssuesEvent | 2019-08-13 04:16:45 | AtlasOfLivingAustralia/data-management | https://api.github.com/repos/AtlasOfLivingAustralia/data-management | closed | Cassandra server changes | Infrastructure Performance | - [x] Document proposed cassandra server changes for CAB
- [x] Testing (performance and load) plan
- [x] Deployment plan
- [x] Rollback plan | 1.0 | Cassandra server changes - - [x] Document proposed cassandra server changes for CAB
- [x] Testing (performance and load) plan
- [x] Deployment plan
- [x] Rollback plan | infrastructure | cassandra server changes document proposed cassandra server changes for cab testing performance and load plan deployment plan rollback plan | 1 |
302,143 | 26,128,248,570 | IssuesEvent | 2022-12-28 22:22:11 | omansak/libvideo | https://api.github.com/repos/omansak/libvideo | closed | Unable to download a specific video | todo need test | Hi there! I've notice that it looks like the library is not able to download a specific video.
When i try to download https://www.youtube.com/watch?v=b2YUw1rqbTY the library stops working without any error...
I can defently live with that :) I just would like to inform you about this issuse I've found.
Thanks for your great job! | 1.0 | Unable to download a specific video - Hi there! I've notice that it looks like the library is not able to download a specific video.
When i try to download https://www.youtube.com/watch?v=b2YUw1rqbTY the library stops working without any error...
I can defently live with that :) I just would like to inform you about this issuse I've found.
Thanks for your great job! | non_infrastructure | unable to download a specific video hi there i ve notice that it looks like the library is not able to download a specific video when i try to download the library stops working without any error i can defently live with that i just would like to inform you about this issuse i ve found thanks for your great job | 0 |
507,982 | 14,686,218,138 | IssuesEvent | 2021-01-01 13:50:42 | googleapis/python-dialogflow-cx | https://api.github.com/repos/googleapis/python-dialogflow-cx | closed | Synthesis failed for python-dialogflow-cx | api: dialogflow autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate python-dialogflow-cx. :broken_heart:
Here's the output from running `synth.py`:
```
/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_pages.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_security_settings_service.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_sessions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_session_entity_types.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_transition_route_groups.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_versions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_webhooks.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_agents.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_entity_types.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_environments.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_experiments.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_intents.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_flows.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_pages.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_sessions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_security_settings_service.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_session_entity_types.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_transition_route_groups.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_versions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_webhooks.py
All done! ✨ 🍰 ✨
271 files reformatted, 5 files left unchanged.
nox > Session blacken was successful.
2020-12-31 05:48:08,762 synthtool [DEBUG] > Wrote metadata to synth.metadata.
2020-12-31 05:48:08,829 autosynth [INFO] > Changed files:
2020-12-31 05:48:08,829 autosynth [INFO] > M google/cloud/dialogflowcx_v3/__init__.py
M google/cloud/dialogflowcx_v3beta1/__init__.py
M synth.metadata
?? dialogflow-cx-v3-py.tar.gz
2020-12-31 05:48:08,829 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%s
2020-12-31 05:48:08,832 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%b%n%nSource-Author: %an <%ae>%nSource-Date: %ad
2020-12-31 05:48:08,836 autosynth [DEBUG] > Running: git add -A
2020-12-31 05:48:08,841 autosynth [DEBUG] > Running: git status --porcelain
2020-12-31 05:48:08,847 autosynth [DEBUG] > Running: git commit -m feat: add experiments API
Source-Author: Yoshi Automation Bot <yoshi-automation@google.com>
Source-Date: Wed Dec 30 08:14:21 2020 -0800
Source-Repo: googleapis/python-dialogflow-cx
Source-Sha: 5381512872ca2492ddabcbdd7ccde5f054aed011
Source-Link: https://github.com/googleapis/python-dialogflow-cx/commit/5381512872ca2492ddabcbdd7ccde5f054aed011
[autosynth-self-0 76444fc] feat: add experiments API
4 files changed, 8 insertions(+), 7 deletions(-)
create mode 100644 dialogflow-cx-v3-py.tar.gz
2020-12-31 05:48:08,854 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 76444fc feat: add experiments API
2020-12-31 05:48:08,859 autosynth [DEBUG] > Running: git checkout autosynth-self
Switched to branch 'autosynth-self'
2020-12-31 05:48:08,864 autosynth [DEBUG] > Running: git diff HEAD..autosynth-self-0 -- . :(exclude)synth.metadata
2020-12-31 05:48:08,869 autosynth [DEBUG] > Running: git diff HEAD autosynth-self-0
2020-12-31 05:48:08,872 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch
error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line
error: dialogflow-cx-v3-py.tar.gz: patch does not apply
2020-12-31 05:48:08,876 autosynth [ERROR] > Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1.
2020-12-31 05:48:09,066 autosynth [DEBUG] > Running: git checkout 5381512872ca2492ddabcbdd7ccde5f054aed011
Note: checking out '5381512872ca2492ddabcbdd7ccde5f054aed011'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 5381512 feat: add experiments API (#36)
2020-12-31 05:48:09,071 autosynth [DEBUG] > Running: git checkout 373861061648b5fe5e0ac4f8a38b32d639ee93e4
HEAD is now at 3738610 chore: exclude `.nox` directories from linting (#885)
2020-12-31 05:48:09,076 autosynth [DEBUG] > Running: git checkout 69697504d9eba1d064820c3085b4750767be6d08
HEAD is now at 69697504 fix: fixed config for Python and Java grpc lib
2020-12-31 05:48:09,089 autosynth [DEBUG] > Running: git branch -f autosynth-0
2020-12-31 05:48:09,092 autosynth [DEBUG] > Running: git checkout autosynth-0
Switched to branch 'autosynth-0'
2020-12-31 05:48:09,096 autosynth [DEBUG] > Running: git merge --ff-only autosynth-self-0
Updating 5381512..76444fc
Fast-forward
dialogflow-cx-v3-py.tar.gz | Bin 0 -> 49152 bytes
google/cloud/dialogflowcx_v3/__init__.py | 6 +++---
google/cloud/dialogflowcx_v3beta1/__init__.py | 6 +++---
synth.metadata | 3 ++-
4 files changed, 8 insertions(+), 7 deletions(-)
create mode 100644 dialogflow-cx-v3-py.tar.gz
2020-12-31 05:48:09,105 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 76444fc feat: add experiments API
2020-12-31 05:48:09,110 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-12-31 05:48:09,116 autosynth [DEBUG] > Running: git diff HEAD..autosynth-0 -- . :(exclude)synth.metadata
2020-12-31 05:48:09,120 autosynth [DEBUG] > Running: git diff HEAD autosynth-0
2020-12-31 05:48:09,124 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch
error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line
error: dialogflow-cx-v3-py.tar.gz: patch does not apply
2020-12-31 05:48:09,127 autosynth [DEBUG] > Running: git clean -fdx
Removing .nox/
Removing __pycache__/
Traceback (most recent call last):
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 92, in synthesize_loop
synthesize_inner_loop(fork, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop
toolbox.patch_merge_version(0, comment)
File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version
git.patch_merge(sub_branch, patch_file_path)
File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge
executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir)
File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call
subprocess.check_call(command, **args)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 334, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 100, in synthesize_loop
synthesize_loop_single_pr(toolbox, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 126, in synthesize_loop_single_pr
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop
toolbox.patch_merge_version(0, comment)
File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version
git.patch_merge(sub_branch, patch_file_path)
File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge
executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir)
File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call
subprocess.check_call(command, **args)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/a0f4ccaf-319e-46b9-9b15-5ac1e95ea79c/targets/github%2Fsynthtool;config=default/tests;query=python-dialogflow-cx;failed=false).
| 1.0 | Synthesis failed for python-dialogflow-cx - Hello! Autosynth couldn't regenerate python-dialogflow-cx. :broken_heart:
Here's the output from running `synth.py`:
```
/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_pages.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_security_settings_service.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_sessions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_session_entity_types.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_transition_route_groups.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_versions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3/test_webhooks.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_agents.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_entity_types.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_environments.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_experiments.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_intents.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_flows.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_pages.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_sessions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_security_settings_service.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_session_entity_types.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_transition_route_groups.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_versions.py
reformatted /home/kbuilder/.cache/synthtool/python-dialogflow-cx/tests/unit/gapic/dialogflowcx_v3beta1/test_webhooks.py
All done! ✨ 🍰 ✨
271 files reformatted, 5 files left unchanged.
nox > Session blacken was successful.
2020-12-31 05:48:08,762 synthtool [DEBUG] > Wrote metadata to synth.metadata.
2020-12-31 05:48:08,829 autosynth [INFO] > Changed files:
2020-12-31 05:48:08,829 autosynth [INFO] > M google/cloud/dialogflowcx_v3/__init__.py
M google/cloud/dialogflowcx_v3beta1/__init__.py
M synth.metadata
?? dialogflow-cx-v3-py.tar.gz
2020-12-31 05:48:08,829 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%s
2020-12-31 05:48:08,832 autosynth [DEBUG] > Running: git log 5381512872ca2492ddabcbdd7ccde5f054aed011 -1 --no-decorate --pretty=%b%n%nSource-Author: %an <%ae>%nSource-Date: %ad
2020-12-31 05:48:08,836 autosynth [DEBUG] > Running: git add -A
2020-12-31 05:48:08,841 autosynth [DEBUG] > Running: git status --porcelain
2020-12-31 05:48:08,847 autosynth [DEBUG] > Running: git commit -m feat: add experiments API
Source-Author: Yoshi Automation Bot <yoshi-automation@google.com>
Source-Date: Wed Dec 30 08:14:21 2020 -0800
Source-Repo: googleapis/python-dialogflow-cx
Source-Sha: 5381512872ca2492ddabcbdd7ccde5f054aed011
Source-Link: https://github.com/googleapis/python-dialogflow-cx/commit/5381512872ca2492ddabcbdd7ccde5f054aed011
[autosynth-self-0 76444fc] feat: add experiments API
4 files changed, 8 insertions(+), 7 deletions(-)
create mode 100644 dialogflow-cx-v3-py.tar.gz
2020-12-31 05:48:08,854 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 76444fc feat: add experiments API
2020-12-31 05:48:08,859 autosynth [DEBUG] > Running: git checkout autosynth-self
Switched to branch 'autosynth-self'
2020-12-31 05:48:08,864 autosynth [DEBUG] > Running: git diff HEAD..autosynth-self-0 -- . :(exclude)synth.metadata
2020-12-31 05:48:08,869 autosynth [DEBUG] > Running: git diff HEAD autosynth-self-0
2020-12-31 05:48:08,872 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch
error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line
error: dialogflow-cx-v3-py.tar.gz: patch does not apply
2020-12-31 05:48:08,876 autosynth [ERROR] > Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1.
2020-12-31 05:48:09,066 autosynth [DEBUG] > Running: git checkout 5381512872ca2492ddabcbdd7ccde5f054aed011
Note: checking out '5381512872ca2492ddabcbdd7ccde5f054aed011'.
You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by performing another checkout.
If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -b with the checkout command again. Example:
git checkout -b <new-branch-name>
HEAD is now at 5381512 feat: add experiments API (#36)
2020-12-31 05:48:09,071 autosynth [DEBUG] > Running: git checkout 373861061648b5fe5e0ac4f8a38b32d639ee93e4
HEAD is now at 3738610 chore: exclude `.nox` directories from linting (#885)
2020-12-31 05:48:09,076 autosynth [DEBUG] > Running: git checkout 69697504d9eba1d064820c3085b4750767be6d08
HEAD is now at 69697504 fix: fixed config for Python and Java grpc lib
2020-12-31 05:48:09,089 autosynth [DEBUG] > Running: git branch -f autosynth-0
2020-12-31 05:48:09,092 autosynth [DEBUG] > Running: git checkout autosynth-0
Switched to branch 'autosynth-0'
2020-12-31 05:48:09,096 autosynth [DEBUG] > Running: git merge --ff-only autosynth-self-0
Updating 5381512..76444fc
Fast-forward
dialogflow-cx-v3-py.tar.gz | Bin 0 -> 49152 bytes
google/cloud/dialogflowcx_v3/__init__.py | 6 +++---
google/cloud/dialogflowcx_v3beta1/__init__.py | 6 +++---
synth.metadata | 3 ++-
4 files changed, 8 insertions(+), 7 deletions(-)
create mode 100644 dialogflow-cx-v3-py.tar.gz
2020-12-31 05:48:09,105 autosynth [DEBUG] > Running: git reset --hard HEAD
HEAD is now at 76444fc feat: add experiments API
2020-12-31 05:48:09,110 autosynth [DEBUG] > Running: git checkout autosynth
Switched to branch 'autosynth'
2020-12-31 05:48:09,116 autosynth [DEBUG] > Running: git diff HEAD..autosynth-0 -- . :(exclude)synth.metadata
2020-12-31 05:48:09,120 autosynth [DEBUG] > Running: git diff HEAD autosynth-0
2020-12-31 05:48:09,124 autosynth [DEBUG] > Running: git apply /tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch
error: cannot apply binary patch to 'dialogflow-cx-v3-py.tar.gz' without full index line
error: dialogflow-cx-v3-py.tar.gz: patch does not apply
2020-12-31 05:48:09,127 autosynth [DEBUG] > Running: git clean -fdx
Removing .nox/
Removing __pycache__/
Traceback (most recent call last):
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 92, in synthesize_loop
synthesize_inner_loop(fork, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop
toolbox.patch_merge_version(0, comment)
File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version
git.patch_merge(sub_branch, patch_file_path)
File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge
executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir)
File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call
subprocess.check_call(command, **args)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-self-0.patch']' returned non-zero exit status 1.
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 354, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 189, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 334, in _inner_main
commit_count = synthesize_loop(x, multiple_prs, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 100, in synthesize_loop
synthesize_loop_single_pr(toolbox, change_pusher, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 126, in synthesize_loop_single_pr
synthesize_inner_loop(toolbox, synthesizer)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 149, in synthesize_inner_loop
toolbox.patch_merge_version(0, comment)
File "/tmpfs/src/github/synthtool/autosynth/synth_toolbox.py", line 182, in patch_merge_version
git.patch_merge(sub_branch, patch_file_path)
File "/tmpfs/src/github/synthtool/autosynth/git.py", line 151, in patch_merge
executor.check_call(["git", "apply", patch_file_path], cwd=git_repo_dir)
File "/tmpfs/src/github/synthtool/autosynth/executor.py", line 29, in check_call
subprocess.check_call(command, **args)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 311, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command '['git', 'apply', '/tmpfs/tmp/tmphk3d8qsl/autosynth-0.patch']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](http://sponge2/results/invocations/a0f4ccaf-319e-46b9-9b15-5ac1e95ea79c/targets/github%2Fsynthtool;config=default/tests;query=python-dialogflow-cx;failed=false).
| non_infrastructure | synthesis failed for python dialogflow cx hello autosynth couldn t regenerate python dialogflow cx broken heart here s the output from running synth py cache synthtool python dialogflow cx tests unit gapic dialogflowcx test pages py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test security settings service py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test sessions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test session entity types py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test transition route groups py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test versions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test webhooks py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test agents py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test entity types py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test environments py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test experiments py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test intents py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test flows py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test pages py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test sessions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test security settings service py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test session entity types py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test transition route groups py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test versions py reformatted home kbuilder cache synthtool python dialogflow cx tests unit gapic dialogflowcx test webhooks py all done ✨ 🍰 ✨ files reformatted files left unchanged nox session blacken was successful synthtool wrote metadata to synth metadata autosynth changed files autosynth m google cloud dialogflowcx init py m google cloud dialogflowcx init py m synth metadata dialogflow cx py tar gz autosynth running git log no decorate pretty s autosynth running git log no decorate pretty b n nsource author an nsource date ad autosynth running git add a autosynth running git status porcelain autosynth running git commit m feat add experiments api source author yoshi automation bot source date wed dec source repo googleapis python dialogflow cx source sha source link feat add experiments api files changed insertions deletions create mode dialogflow cx py tar gz autosynth running git reset hard head head is now at feat add experiments api autosynth running git checkout autosynth self switched to branch autosynth self autosynth running git diff head autosynth self exclude synth metadata autosynth running git diff head autosynth self autosynth running git apply tmpfs tmp autosynth self patch error cannot apply binary patch to dialogflow cx py tar gz without full index line error dialogflow cx py tar gz patch does not apply autosynth command returned non zero exit status autosynth running git checkout note checking out you are in detached head state you can look around make experimental changes and commit them and you can discard any commits you make in this state without impacting any branches by performing another checkout if you want to create a new branch to retain commits you create you may do so now or later by using b with the checkout command again example git checkout b head is now at feat add experiments api autosynth running git checkout head is now at chore exclude nox directories from linting autosynth running git checkout head is now at fix fixed config for python and java grpc lib autosynth running git branch f autosynth autosynth running git checkout autosynth switched to branch autosynth autosynth running git merge ff only autosynth self updating fast forward dialogflow cx py tar gz bin bytes google cloud dialogflowcx init py google cloud dialogflowcx init py synth metadata files changed insertions deletions create mode dialogflow cx py tar gz autosynth running git reset hard head head is now at feat add experiments api autosynth running git checkout autosynth switched to branch autosynth autosynth running git diff head autosynth exclude synth metadata autosynth running git diff head autosynth autosynth running git apply tmpfs tmp autosynth patch error cannot apply binary patch to dialogflow cx py tar gz without full index line error dialogflow cx py tar gz patch does not apply autosynth running git clean fdx removing nox removing pycache traceback most recent call last file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize inner loop fork synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop toolbox patch merge version comment file tmpfs src github synthtool autosynth synth toolbox py line in patch merge version git patch merge sub branch patch file path file tmpfs src github synthtool autosynth git py line in patch merge executor check call cwd git repo dir file tmpfs src github synthtool autosynth executor py line in check call subprocess check call command args file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status during handling of the above exception another exception occurred traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main commit count synthesize loop x multiple prs change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop synthesize loop single pr toolbox change pusher synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize loop single pr synthesize inner loop toolbox synthesizer file tmpfs src github synthtool autosynth synth py line in synthesize inner loop toolbox patch merge version comment file tmpfs src github synthtool autosynth synth toolbox py line in patch merge version git patch merge sub branch patch file path file tmpfs src github synthtool autosynth git py line in patch merge executor check call cwd git repo dir file tmpfs src github synthtool autosynth executor py line in check call subprocess check call command args file home kbuilder pyenv versions lib subprocess py line in check call raise calledprocesserror retcode cmd subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 0 |
115,425 | 24,762,865,739 | IssuesEvent | 2022-10-22 05:51:13 | log2timeline/plaso | https://api.github.com/repos/log2timeline/plaso | closed | refactor usbstor winreg plugin | enhancement parsers analysis code health | **Description of problem:**
https://github.com/log2timeline/plaso/blob/master/plaso/parsers/winreg_plugins/usbstor.py#L45
The usbstor winreg plugin current assumes connection times based on the key last written time:
- [x] change parser to produce event objects with last written time
- [x] ~~add setup api information~~ - setupapi parser has been added
- [x] ~~move the determination of the connection times to an analysis plugin~~ - out of scope
| 1.0 | refactor usbstor winreg plugin - **Description of problem:**
https://github.com/log2timeline/plaso/blob/master/plaso/parsers/winreg_plugins/usbstor.py#L45
The usbstor winreg plugin current assumes connection times based on the key last written time:
- [x] change parser to produce event objects with last written time
- [x] ~~add setup api information~~ - setupapi parser has been added
- [x] ~~move the determination of the connection times to an analysis plugin~~ - out of scope
| non_infrastructure | refactor usbstor winreg plugin description of problem the usbstor winreg plugin current assumes connection times based on the key last written time change parser to produce event objects with last written time add setup api information setupapi parser has been added move the determination of the connection times to an analysis plugin out of scope | 0 |
8,388 | 7,375,375,097 | IssuesEvent | 2018-03-14 00:06:23 | Microsoft/TypeScript | https://api.github.com/repos/Microsoft/TypeScript | closed | Nightly builds not published to npm (starting 3/8/2017) | Infrastructure | For any others who may file this, we're aware of the issue and will have it back up by next week. Thanks! | 1.0 | Nightly builds not published to npm (starting 3/8/2017) - For any others who may file this, we're aware of the issue and will have it back up by next week. Thanks! | infrastructure | nightly builds not published to npm starting for any others who may file this we re aware of the issue and will have it back up by next week thanks | 1 |
413,195 | 12,061,292,904 | IssuesEvent | 2020-04-15 23:20:28 | grpc/grpc | https://api.github.com/repos/grpc/grpc | closed | Timeout: unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate | disposition/stale kind/bug lang/Python priority/P2 | ```
++ dirname /var/local/git/grpc/tools/run_tests/helper_scripts/run_python.sh
+ cd /var/local/git/grpc/tools/run_tests/helper_scripts/../../..
++ realpath py36_native/bin/python
+ PYTHON=/var/local/git/grpc/py36_native/bin/python3.6
++ pwd
+ ROOT=/var/local/git/grpc
+ /var/local/git/grpc/py36_native/bin/python3.6 /var/local/git/grpc/src/python/grpcio_tests/setup.py test_lite
running test_lite
Running unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call
[95mTesting gRPC Python...
[0m[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_stream_stream_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_stream_unary_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_unary_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_unary_stream_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_unary_unary_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_unary_call
[0mRunning unit._exit_test.ExitTest.test_poll_connectivity
[92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity
[0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server
[92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity_no_server
[0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate
```
https://source.cloud.google.com/results/invocations/848afa18-6b20-45ce-959b-09b094219029/targets/github%2Fgrpc%2Frun_tests%2Fpython_linux_opt_native%2Fpy36_native.test.unit._exit_test.ExitTest/tests | 1.0 | Timeout: unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate - ```
++ dirname /var/local/git/grpc/tools/run_tests/helper_scripts/run_python.sh
+ cd /var/local/git/grpc/tools/run_tests/helper_scripts/../../..
++ realpath py36_native/bin/python
+ PYTHON=/var/local/git/grpc/py36_native/bin/python3.6
++ pwd
+ ROOT=/var/local/git/grpc
+ /var/local/git/grpc/py36_native/bin/python3.6 /var/local/git/grpc/src/python/grpcio_tests/setup.py test_lite
running test_lite
Running unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call
[95mTesting gRPC Python...
[0m[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_stream_unary_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_partial_unary_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_stream_stream_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_stream_unary_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_stream_unary_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_unary_stream_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_stream_call
[0mRunning unit._exit_test.ExitTest.test_in_flight_unary_unary_call
[92mSUCCESS unit._exit_test.ExitTest.test_in_flight_unary_unary_call
[0mRunning unit._exit_test.ExitTest.test_poll_connectivity
[92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity
[0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server
[92mSUCCESS unit._exit_test.ExitTest.test_poll_connectivity_no_server
[0mRunning unit._exit_test.ExitTest.test_poll_connectivity_no_server_terminate
```
https://source.cloud.google.com/results/invocations/848afa18-6b20-45ce-959b-09b094219029/targets/github%2Fgrpc%2Frun_tests%2Fpython_linux_opt_native%2Fpy36_native.test.unit._exit_test.ExitTest/tests | non_infrastructure | timeout unit exit test exittest test poll connectivity no server terminate dirname var local git grpc tools run tests helper scripts run python sh cd var local git grpc tools run tests helper scripts realpath native bin python python var local git grpc native bin pwd root var local git grpc var local git grpc native bin var local git grpc src python grpcio tests setup py test lite running test lite running unit exit test exittest test in flight partial stream stream call grpc python unit exit test exittest test in flight partial stream stream call unit exit test exittest test in flight partial stream unary call unit exit test exittest test in flight partial stream unary call unit exit test exittest test in flight partial unary stream call unit exit test exittest test in flight partial unary stream call unit exit test exittest test in flight stream stream call unit exit test exittest test in flight stream stream call unit exit test exittest test in flight stream unary call unit exit test exittest test in flight stream unary call unit exit test exittest test in flight unary stream call unit exit test exittest test in flight unary stream call unit exit test exittest test in flight unary unary call unit exit test exittest test in flight unary unary call unit exit test exittest test poll connectivity unit exit test exittest test poll connectivity unit exit test exittest test poll connectivity no server unit exit test exittest test poll connectivity no server unit exit test exittest test poll connectivity no server terminate | 0 |
24,563 | 17,399,634,610 | IssuesEvent | 2021-08-02 17:41:42 | dotnet/aspnetcore | https://api.github.com/repos/dotnet/aspnetcore | closed | Several issues with MessagePack when building asp net core 5 | area-infrastructure | I'm stuck with building asp dot net core 5.0 on Windows 10
I followed the instruction but have no luck with it. The output of build.cmd reporting me several errors that some files are missed in MessagePack
> CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\T4\MessagePackReader.Integers.cs".
> CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\MessagePackSerializationException.cs".
fix_symlinks.bat in MessagaPack.UnityClient doesn't fix this.
Could anybody help me with it what I missed?
My steps are (commit 3f873035c9ee96c4e4e93a9d3966baf3ef02581a)
1. restore.cmd
2. build.cmd
Thank you! | 1.0 | Several issues with MessagePack when building asp net core 5 - I'm stuck with building asp dot net core 5.0 on Windows 10
I followed the instruction but have no luck with it. The output of build.cmd reporting me several errors that some files are missed in MessagePack
> CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\T4\MessagePackReader.Integers.cs".
> CCS2001: File not found "..mvc\source\aspnetcore\src\submodules\MessagePack-CSharp\src\MessagePack.UnityClient\Assets\Scripts\MessagePack\MessagePackSerializationException.cs".
fix_symlinks.bat in MessagaPack.UnityClient doesn't fix this.
Could anybody help me with it what I missed?
My steps are (commit 3f873035c9ee96c4e4e93a9d3966baf3ef02581a)
1. restore.cmd
2. build.cmd
Thank you! | infrastructure | several issues with messagepack when building asp net core i m stuck with building asp dot net core on windows i followed the instruction but have no luck with it the output of build cmd reporting me several errors that some files are missed in messagepack file not found mvc source aspnetcore src submodules messagepack csharp src messagepack unityclient assets scripts messagepack messagepackreader integers cs file not found mvc source aspnetcore src submodules messagepack csharp src messagepack unityclient assets scripts messagepack messagepackserializationexception cs fix symlinks bat in messagapack unityclient doesn t fix this could anybody help me with it what i missed my steps are commit restore cmd build cmd thank you | 1 |
7,378 | 6,928,076,459 | IssuesEvent | 2017-12-01 02:24:03 | codeforboston/cliff-effects | https://api.github.com/repos/codeforboston/cliff-effects | closed | Stabilize Conditional Logic for Header | infrastructure | Currently the `Header` file has a [hacky line](https://github.com/codeforboston/cliff-effects/blob/dev/src/components/Header.js#L14) that makes it only display if the current path is '/'. This is pretty fragile because changing the path can cause the header to display on the home page.
Action item here is to implement a more robust toggle for the header visibility. | 1.0 | Stabilize Conditional Logic for Header - Currently the `Header` file has a [hacky line](https://github.com/codeforboston/cliff-effects/blob/dev/src/components/Header.js#L14) that makes it only display if the current path is '/'. This is pretty fragile because changing the path can cause the header to display on the home page.
Action item here is to implement a more robust toggle for the header visibility. | infrastructure | stabilize conditional logic for header currently the header file has a that makes it only display if the current path is this is pretty fragile because changing the path can cause the header to display on the home page action item here is to implement a more robust toggle for the header visibility | 1 |
11,768 | 9,418,522,409 | IssuesEvent | 2019-04-10 19:26:48 | OpenLiberty/openliberty.io | https://api.github.com/repos/OpenLiberty/openliberty.io | closed | Automate tests for interactive guides | infrastructure | We need to explore what we can do arounds having automated tests for the interactive guides. My first gut feeling is to write some simple Selenium tests that we can run before we publish to openliberty.io. | 1.0 | Automate tests for interactive guides - We need to explore what we can do arounds having automated tests for the interactive guides. My first gut feeling is to write some simple Selenium tests that we can run before we publish to openliberty.io. | infrastructure | automate tests for interactive guides we need to explore what we can do arounds having automated tests for the interactive guides my first gut feeling is to write some simple selenium tests that we can run before we publish to openliberty io | 1 |
16,542 | 12,032,868,915 | IssuesEvent | 2020-04-13 13:08:33 | spockframework/spock | https://api.github.com/repos/spockframework/spock | opened | Test Spock with Java 13 and 14 | infrastructure | Spock 2 should work with Java 14 (having Groovy 2.5.10 and 3.0.x) on a classpath, but It would be good to run internal Spock tests with with Java 13 and 14 on CI server.
It is not impossible to run tests with Java 14 with older Gradle versions, but Gradle 6.3 is required for running Java 14 in general, so the upgrade is the most straightforward way to go. Unfortunately it is currently blocked by the Gradle OSGI plugin removal which has to be replaced with something else (if Spock 2 would keep OSGI manifest support at all) - #1139
| 1.0 | Test Spock with Java 13 and 14 - Spock 2 should work with Java 14 (having Groovy 2.5.10 and 3.0.x) on a classpath, but It would be good to run internal Spock tests with with Java 13 and 14 on CI server.
It is not impossible to run tests with Java 14 with older Gradle versions, but Gradle 6.3 is required for running Java 14 in general, so the upgrade is the most straightforward way to go. Unfortunately it is currently blocked by the Gradle OSGI plugin removal which has to be replaced with something else (if Spock 2 would keep OSGI manifest support at all) - #1139
| infrastructure | test spock with java and spock should work with java having groovy and x on a classpath but it would be good to run internal spock tests with with java and on ci server it is not impossible to run tests with java with older gradle versions but gradle is required for running java in general so the upgrade is the most straightforward way to go unfortunately it is currently blocked by the gradle osgi plugin removal which has to be replaced with something else if spock would keep osgi manifest support at all | 1 |
20,829 | 31,150,009,829 | IssuesEvent | 2023-08-16 09:18:12 | sebastianbergmann/phpunit | https://api.github.com/repos/sebastianbergmann/phpunit | opened | Remove `TestCase::assertStringNotMatchesFormat()` and `TestCase::assertStringNotMatchesFormatFile()` | type/backward-compatibility feature/assertion | Follows #5472 | True | Remove `TestCase::assertStringNotMatchesFormat()` and `TestCase::assertStringNotMatchesFormatFile()` - Follows #5472 | non_infrastructure | remove testcase assertstringnotmatchesformat and testcase assertstringnotmatchesformatfile follows | 0 |
12,040 | 9,556,162,145 | IssuesEvent | 2019-05-03 07:24:49 | ImpressCMS/impresscms | https://api.github.com/repos/ImpressCMS/impresscms | closed | Add Slack badge to README | enhancement infrastructure | I think Slack badge on https://github.com/GitbookIO/gitbook looks good. And I think it would be good to add something like this to our README. Maybe more users would join our chat system and maybe more users would be involved. Yes, I know that @fiammybe created another chat system on Gitter but we don't have so many users that I don't see any any point to have multiple chat systems for a while. I think Slack could be good enough for us for a while.
This badge is driven by [Slackin](https://github.com/rauchg/slackin). I think we could host this on Heroku or similar service as free hosting. That's not hard. But before that I need to get custom slack API app token from @fiammybe (yeah that's why I added you to here for this task!) | 1.0 | Add Slack badge to README - I think Slack badge on https://github.com/GitbookIO/gitbook looks good. And I think it would be good to add something like this to our README. Maybe more users would join our chat system and maybe more users would be involved. Yes, I know that @fiammybe created another chat system on Gitter but we don't have so many users that I don't see any any point to have multiple chat systems for a while. I think Slack could be good enough for us for a while.
This badge is driven by [Slackin](https://github.com/rauchg/slackin). I think we could host this on Heroku or similar service as free hosting. That's not hard. But before that I need to get custom slack API app token from @fiammybe (yeah that's why I added you to here for this task!) | infrastructure | add slack badge to readme i think slack badge on looks good and i think it would be good to add something like this to our readme maybe more users would join our chat system and maybe more users would be involved yes i know that fiammybe created another chat system on gitter but we don t have so many users that i don t see any any point to have multiple chat systems for a while i think slack could be good enough for us for a while this badge is driven by i think we could host this on heroku or similar service as free hosting that s not hard but before that i need to get custom slack api app token from fiammybe yeah that s why i added you to here for this task | 1 |
31,943 | 26,263,484,244 | IssuesEvent | 2023-01-06 10:13:08 | GitGuardian/ggshield | https://api.github.com/repos/GitGuardian/ggshield | closed | ggshield should not install its tests | bug status:: confirmed Infrastructure | ## Environment
- ggshield version: 1.13.5
- Operating system (Linux, macOS, Windows): any
- Operating system version: any
- Python version: 3.8.10
## Describe the bug
Installing ggshield with `pip install` installs the tests directory.
Steps to reproduce:
1. Create and activate a virtualenv
```
python -m venv bug
```
2. Check it contains no tests directory:
```
ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests
ls: cannot access '/home/agateau/tmp/bug/lib/python3.8/site-packages/tests': No such file or directory
```
3. Install ggshield:
```
pip install ~/src/ggshield
```
Actual result:
```
ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests
cmd core iac output scan unit
```
Expected result:
There should be no `tests` directory in the `site-packages` directory.
## Useful resources
https://setuptools.pypa.io/en/latest/userguide/package_discovery.html
## TODO
- [ ] Fix the issue
- [ ] Setup a tool like [check-wheel-contents](https://pypi.org/project/check-wheel-contents/) in the CI so that this cannot happen again | 1.0 | ggshield should not install its tests - ## Environment
- ggshield version: 1.13.5
- Operating system (Linux, macOS, Windows): any
- Operating system version: any
- Python version: 3.8.10
## Describe the bug
Installing ggshield with `pip install` installs the tests directory.
Steps to reproduce:
1. Create and activate a virtualenv
```
python -m venv bug
```
2. Check it contains no tests directory:
```
ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests
ls: cannot access '/home/agateau/tmp/bug/lib/python3.8/site-packages/tests': No such file or directory
```
3. Install ggshield:
```
pip install ~/src/ggshield
```
Actual result:
```
ls $VIRTUAL_ENV/lib/python3.8/site-packages/tests
cmd core iac output scan unit
```
Expected result:
There should be no `tests` directory in the `site-packages` directory.
## Useful resources
https://setuptools.pypa.io/en/latest/userguide/package_discovery.html
## TODO
- [ ] Fix the issue
- [ ] Setup a tool like [check-wheel-contents](https://pypi.org/project/check-wheel-contents/) in the CI so that this cannot happen again | infrastructure | ggshield should not install its tests environment ggshield version operating system linux macos windows any operating system version any python version describe the bug installing ggshield with pip install installs the tests directory steps to reproduce create and activate a virtualenv python m venv bug check it contains no tests directory ls virtual env lib site packages tests ls cannot access home agateau tmp bug lib site packages tests no such file or directory install ggshield pip install src ggshield actual result ls virtual env lib site packages tests cmd core iac output scan unit expected result there should be no tests directory in the site packages directory useful resources todo fix the issue setup a tool like in the ci so that this cannot happen again | 1 |
290,745 | 8,903,452,750 | IssuesEvent | 2019-01-17 10:31:11 | open-learning-exchange/planet | https://api.github.com/repos/open-learning-exchange/planet | closed | Adding a new course overwrites last edited course | priority | Specifically when you add an exam to the new course. | 1.0 | Adding a new course overwrites last edited course - Specifically when you add an exam to the new course. | non_infrastructure | adding a new course overwrites last edited course specifically when you add an exam to the new course | 0 |
182,135 | 6,667,446,369 | IssuesEvent | 2017-10-03 12:37:55 | spring-projects/spring-boot | https://api.github.com/repos/spring-projects/spring-boot | reopened | Make TomcatEmbeddedWebappClassLoader parallel capable | for: team-attention priority: normal type: enhancement | With reference to [DATAMONGO-1792](https://jira.spring.io/browse/DATAMONGO-1792), the synchronization of `TomcatEmbeddedWebappClassLoader.loadClass(String, boolean)` is inefficient. Rather than making the method itself `synchronized` it should synchronize on the response from `(getClassLoadingLock(className)`. We should also verify that the JVM considers the class loader to be parallel capable (its superclass is registered as such). | 1.0 | Make TomcatEmbeddedWebappClassLoader parallel capable - With reference to [DATAMONGO-1792](https://jira.spring.io/browse/DATAMONGO-1792), the synchronization of `TomcatEmbeddedWebappClassLoader.loadClass(String, boolean)` is inefficient. Rather than making the method itself `synchronized` it should synchronize on the response from `(getClassLoadingLock(className)`. We should also verify that the JVM considers the class loader to be parallel capable (its superclass is registered as such). | non_infrastructure | make tomcatembeddedwebappclassloader parallel capable with reference to the synchronization of tomcatembeddedwebappclassloader loadclass string boolean is inefficient rather than making the method itself synchronized it should synchronize on the response from getclassloadinglock classname we should also verify that the jvm considers the class loader to be parallel capable its superclass is registered as such | 0 |
56,068 | 11,501,100,802 | IssuesEvent | 2020-02-12 16:36:13 | pace-neutrons/Pace-Project-Plan | https://api.github.com/repos/pace-neutrons/Pace-Project-Plan | closed | Euphonic Calculations: API/Internal Data Structure Design: Review current/proposed designs | Euphonic Calculations Interfacing to 3rd Party Code | Hold a session to give an overview and review of Euphonic's current API/data structures
Sub issue of #46
Will feed into #14, #15, #16, #17 | 1.0 | Euphonic Calculations: API/Internal Data Structure Design: Review current/proposed designs - Hold a session to give an overview and review of Euphonic's current API/data structures
Sub issue of #46
Will feed into #14, #15, #16, #17 | non_infrastructure | euphonic calculations api internal data structure design review current proposed designs hold a session to give an overview and review of euphonic s current api data structures sub issue of will feed into | 0 |
15,560 | 11,589,268,730 | IssuesEvent | 2020-02-24 01:16:32 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Detailed descriptions can be too long for the user interface | interface/infrastructure newfeature | The "Description" attribute of properties is sometimes overly long when used in the context of the user interface. For example, SWCON currently has the description: "Fractional amount of water above DUL that can drain under gravity per day (SWCON) (/d)". That's all useful information, but it's rather long when used as a column header in a table. Further, in the SoilWater input views, the units attribute is also added automatically, so the header ends up with "(/d) (/d)" at the end.
I'd welcome discussion about the best way to address these problems. As a starter, I'd suggest that units of measurement not be included in the Description attribute, but be provided only in the Units attribute; having it in both is just redundant. I think it might also be useful to have a "ShortDescription" (or "Label") attribute, and a "LongDescription" attribute, The short version could be used in the column headers of the GUI, and perhaps the long version could be displayed when the user requests to see it. For example, the header might just read "SWCON", but the user would be able to see the longer, detailed description by some mechanism (perhaps by mouse over, or perhaps from a popup-menu item)
Any comments? | 1.0 | Detailed descriptions can be too long for the user interface - The "Description" attribute of properties is sometimes overly long when used in the context of the user interface. For example, SWCON currently has the description: "Fractional amount of water above DUL that can drain under gravity per day (SWCON) (/d)". That's all useful information, but it's rather long when used as a column header in a table. Further, in the SoilWater input views, the units attribute is also added automatically, so the header ends up with "(/d) (/d)" at the end.
I'd welcome discussion about the best way to address these problems. As a starter, I'd suggest that units of measurement not be included in the Description attribute, but be provided only in the Units attribute; having it in both is just redundant. I think it might also be useful to have a "ShortDescription" (or "Label") attribute, and a "LongDescription" attribute, The short version could be used in the column headers of the GUI, and perhaps the long version could be displayed when the user requests to see it. For example, the header might just read "SWCON", but the user would be able to see the longer, detailed description by some mechanism (perhaps by mouse over, or perhaps from a popup-menu item)
Any comments? | infrastructure | detailed descriptions can be too long for the user interface the description attribute of properties is sometimes overly long when used in the context of the user interface for example swcon currently has the description fractional amount of water above dul that can drain under gravity per day swcon d that s all useful information but it s rather long when used as a column header in a table further in the soilwater input views the units attribute is also added automatically so the header ends up with d d at the end i d welcome discussion about the best way to address these problems as a starter i d suggest that units of measurement not be included in the description attribute but be provided only in the units attribute having it in both is just redundant i think it might also be useful to have a shortdescription or label attribute and a longdescription attribute the short version could be used in the column headers of the gui and perhaps the long version could be displayed when the user requests to see it for example the header might just read swcon but the user would be able to see the longer detailed description by some mechanism perhaps by mouse over or perhaps from a popup menu item any comments | 1 |
27,678 | 22,148,105,849 | IssuesEvent | 2022-06-03 14:04:38 | enarx/enarx | https://api.github.com/repos/enarx/enarx | closed | [Infrastructure]: Release improvements | infrastructure triage release | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct
### Description
## `helper/release_enarx.sh`
- Bugs
- lower case 'release 0.5.1' to make sure that commit checker passes
- lower case 'release' for PR to make sure commit checker passes
- Sleep 60 seconds between published crates to avoid race condition
- Features
- fast mode to run single image
- dry-run false always holds on to artifacts/repo
- maybe jump to release and skip checks. scream like a banshee and check if you want to do this | 1.0 | [Infrastructure]: Release improvements - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Code of Conduct
- [X] I agree to follow this project's Code of Conduct
### Description
## `helper/release_enarx.sh`
- Bugs
- lower case 'release 0.5.1' to make sure that commit checker passes
- lower case 'release' for PR to make sure commit checker passes
- Sleep 60 seconds between published crates to avoid race condition
- Features
- fast mode to run single image
- dry-run false always holds on to artifacts/repo
- maybe jump to release and skip checks. scream like a banshee and check if you want to do this | infrastructure | release improvements is there an existing issue for this i have searched the existing issues code of conduct i agree to follow this project s code of conduct description helper release enarx sh bugs lower case release to make sure that commit checker passes lower case release for pr to make sure commit checker passes sleep seconds between published crates to avoid race condition features fast mode to run single image dry run false always holds on to artifacts repo maybe jump to release and skip checks scream like a banshee and check if you want to do this | 1 |
283,343 | 30,913,274,459 | IssuesEvent | 2023-08-05 01:31:24 | hshivhare67/kernel_v4.19.72_CVE-2022-42896_new | https://api.github.com/repos/hshivhare67/kernel_v4.19.72_CVE-2022-42896_new | reopened | CVE-2021-29154 (High) detected in linuxlinux-4.19.279 | Mend: dependency security vulnerability | ## CVE-2021-29154 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
BPF JIT compilers in the Linux kernel through 5.11.12 have incorrect computation of branch displacements, allowing them to execute arbitrary code within the kernel context. This affects arch/x86/net/bpf_jit_comp.c and arch/x86/net/bpf_jit_comp32.c.
<p>Publish Date: 2021-04-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29154>CVE-2021-29154</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-29154">https://www.linuxkernelcves.com/cves/CVE-2021-29154</a></p>
<p>Release Date: 2021-04-08</p>
<p>Fix Resolution: v4.4.266, v4.9.266, v4.14.230, v4.19.186, v5.4.111, v5.10.29, v5.11.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-29154 (High) detected in linuxlinux-4.19.279 - ## CVE-2021-29154 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.279</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
BPF JIT compilers in the Linux kernel through 5.11.12 have incorrect computation of branch displacements, allowing them to execute arbitrary code within the kernel context. This affects arch/x86/net/bpf_jit_comp.c and arch/x86/net/bpf_jit_comp32.c.
<p>Publish Date: 2021-04-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29154>CVE-2021-29154</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-29154">https://www.linuxkernelcves.com/cves/CVE-2021-29154</a></p>
<p>Release Date: 2021-04-08</p>
<p>Fix Resolution: v4.4.266, v4.9.266, v4.14.230, v4.19.186, v5.4.111, v5.10.29, v5.11.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in base branch master vulnerable source files vulnerability details bpf jit compilers in the linux kernel through have incorrect computation of branch displacements allowing them to execute arbitrary code within the kernel context this affects arch net bpf jit comp c and arch net bpf jit c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
9,326 | 7,921,732,065 | IssuesEvent | 2018-07-05 08:33:02 | ressec/hemajoo-foundation | https://api.github.com/repos/ressec/hemajoo-foundation | closed | Travis build for hemajoo-foundation fails | Bug Critical Development Infrastructure | ## Cause
The error reported in the build log is:
`Caused by: java.lang.ClassNotFoundException: org.apache.maven.doxia.siterenderer.DocumentContent`
## Resolution
Resolved by changing in the parent pom the version number of the Maven site plugin:
```
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-site-plugin</artifactId>
<version>3.7.1</version>
. . .
</plugin>
```
| 1.0 | Travis build for hemajoo-foundation fails - ## Cause
The error reported in the build log is:
`Caused by: java.lang.ClassNotFoundException: org.apache.maven.doxia.siterenderer.DocumentContent`
## Resolution
Resolved by changing in the parent pom the version number of the Maven site plugin:
```
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-site-plugin</artifactId>
<version>3.7.1</version>
. . .
</plugin>
```
| infrastructure | travis build for hemajoo foundation fails cause the error reported in the build log is caused by java lang classnotfoundexception org apache maven doxia siterenderer documentcontent resolution resolved by changing in the parent pom the version number of the maven site plugin org apache maven plugins maven site plugin | 1 |
64,324 | 26,689,284,338 | IssuesEvent | 2023-01-27 02:08:10 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Project: Placemaking Program | Service: Apps Epic Workgroup: SMO Product: Smart Mobility Office Project Index Project: Placemaking Program | Intake and tracking platform that aids evaluation, milestone management, and a directory for Artists and Art Partners.
## Background
The importance and need of developing a management methodology through Placemaking is essential to have a user-friendly interface to manage processes, collaborate seamlessly with team members, automate tasks, and track key milestones. This becomes important as placemaking sub-programs are expanding and becoming more complex with limited resources. Additionally, there have been some complications with the current application form and intake process currently in place. The intake form currently places all submissions together, rather than separating each intake application.
## Scope & Deliverables
EOI intake form, evaluation platform with key milestones for projects, and directory to track
## Desired Outcomes
A user-friendly platform that can be used by internal and external agencies to track key milestones. The ability to track beyond when applications come in, but the need for lengthy community outreach before installation is scheduled.
## Timebox
November-December 2022
## Concepts and references
[Google form](https://api.knack.com/v1/applications/595d00ebd315cc4cb98daff4/download/asset/630fa509bb94060021f36aa7/austintransportation_smartmobilityofficegoogleforms.pdf)
[Artbox website](https://www.austintexas.gov/page/artbox-program)

---
This Github issue represents a project of Austin Transportation's [Data & Technology Services](https://austinmobility.io/) team. Project status is documented regularly in the comments below.
<!-- Don't forget to:
- Add a new "Project:" label here: https://github.com/cityofaustin/atd-data-tech/labels. Use the hex code #3D3D3D.
- Add a project evaluation here: https://atd.knack.com/dts#project-evaluation/
-->
| 1.0 | Project: Placemaking Program - Intake and tracking platform that aids evaluation, milestone management, and a directory for Artists and Art Partners.
## Background
The importance and need of developing a management methodology through Placemaking is essential to have a user-friendly interface to manage processes, collaborate seamlessly with team members, automate tasks, and track key milestones. This becomes important as placemaking sub-programs are expanding and becoming more complex with limited resources. Additionally, there have been some complications with the current application form and intake process currently in place. The intake form currently places all submissions together, rather than separating each intake application.
## Scope & Deliverables
EOI intake form, evaluation platform with key milestones for projects, and directory to track
## Desired Outcomes
A user-friendly platform that can be used by internal and external agencies to track key milestones. The ability to track beyond when applications come in, but the need for lengthy community outreach before installation is scheduled.
## Timebox
November-December 2022
## Concepts and references
[Google form](https://api.knack.com/v1/applications/595d00ebd315cc4cb98daff4/download/asset/630fa509bb94060021f36aa7/austintransportation_smartmobilityofficegoogleforms.pdf)
[Artbox website](https://www.austintexas.gov/page/artbox-program)

---
This Github issue represents a project of Austin Transportation's [Data & Technology Services](https://austinmobility.io/) team. Project status is documented regularly in the comments below.
<!-- Don't forget to:
- Add a new "Project:" label here: https://github.com/cityofaustin/atd-data-tech/labels. Use the hex code #3D3D3D.
- Add a project evaluation here: https://atd.knack.com/dts#project-evaluation/
-->
| non_infrastructure | project placemaking program intake and tracking platform that aids evaluation milestone management and a directory for artists and art partners background the importance and need of developing a management methodology through placemaking is essential to have a user friendly interface to manage processes collaborate seamlessly with team members automate tasks and track key milestones this becomes important as placemaking sub programs are expanding and becoming more complex with limited resources additionally there have been some complications with the current application form and intake process currently in place the intake form currently places all submissions together rather than separating each intake application scope deliverables eoi intake form evaluation platform with key milestones for projects and directory to track desired outcomes a user friendly platform that can be used by internal and external agencies to track key milestones the ability to track beyond when applications come in but the need for lengthy community outreach before installation is scheduled timebox november december concepts and references this github issue represents a project of austin transportation s team project status is documented regularly in the comments below don t forget to add a new project label here use the hex code add a project evaluation here | 0 |
209,219 | 16,187,586,786 | IssuesEvent | 2021-05-04 00:40:27 | google/iree | https://api.github.com/repos/google/iree | reopened | [New documentation] Write docs for "Building Python bindings" | bindings/python documentation ✏️ | * Live preview of this page: https://scotttodd.github.io/iree/building-from-source/python/, which will eventually live on https://google.github.io/iree/ after we switch over the docs.
* Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/building-from-source/python.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though.
* See the README here for instructions on previewing changes locally: https://github.com/google/iree/tree/staging-docs-rework/mkdocs
* mkdocs + mkdocs material offer a bunch of nice features, like https://squidfunk.github.io/mkdocs-material/reference/code-blocks/ (line numbers, highlighting specific lines, etc.)
I'd like for this page to aggregate the "building from source" aspects of our current Python pages:
* https://google.github.io/iree/get-started/getting-started-python
* https://google.github.io/iree/get-started/getting-started-tensorflow
* https://google.github.io/iree/developing-iree/tensorflow-integrations (maybe)
This will be linked from the "Python bindings" page (preview: https://scotttodd.github.io/iree/bindings/python/) and possibly from the Python frontend pages, in cases where users will benefit from building from source instead of installing from our published packages. | 1.0 | [New documentation] Write docs for "Building Python bindings" - * Live preview of this page: https://scotttodd.github.io/iree/building-from-source/python/, which will eventually live on https://google.github.io/iree/ after we switch over the docs.
* Source currently lives on the `staging-docs-rework` branch here: https://github.com/google/iree/blob/staging-docs-rework/mkdocs/docs/building-from-source/python.md . PRs or direct commits to the branch are okay. We'll want to be careful about force pushes / rebases though.
* See the README here for instructions on previewing changes locally: https://github.com/google/iree/tree/staging-docs-rework/mkdocs
* mkdocs + mkdocs material offer a bunch of nice features, like https://squidfunk.github.io/mkdocs-material/reference/code-blocks/ (line numbers, highlighting specific lines, etc.)
I'd like for this page to aggregate the "building from source" aspects of our current Python pages:
* https://google.github.io/iree/get-started/getting-started-python
* https://google.github.io/iree/get-started/getting-started-tensorflow
* https://google.github.io/iree/developing-iree/tensorflow-integrations (maybe)
This will be linked from the "Python bindings" page (preview: https://scotttodd.github.io/iree/bindings/python/) and possibly from the Python frontend pages, in cases where users will benefit from building from source instead of installing from our published packages. | non_infrastructure | write docs for building python bindings live preview of this page which will eventually live on after we switch over the docs source currently lives on the staging docs rework branch here prs or direct commits to the branch are okay we ll want to be careful about force pushes rebases though see the readme here for instructions on previewing changes locally mkdocs mkdocs material offer a bunch of nice features like line numbers highlighting specific lines etc i d like for this page to aggregate the building from source aspects of our current python pages maybe this will be linked from the python bindings page preview and possibly from the python frontend pages in cases where users will benefit from building from source instead of installing from our published packages | 0 |
8,902 | 7,724,934,260 | IssuesEvent | 2018-05-24 16:22:58 | GoogleCloudPlatform/forseti-security | https://api.github.com/repos/GoogleCloudPlatform/forseti-security | closed | Move cron lock file out of /tmp (will be cleaned up by system) | module: infrastructure priority: p1 release-testing: 2.0 RC3 triaged: yes | The cron lock file is currrently placed in `/tmp`, which will be cleaned up by the system everyday. So, this should be moved somewhere else so that we avoid the chance that a cron job is running, and then the lock file gets cleaned up.
```
ubuntu@forseti-server-vm-4330:~/forseti-security$ ls /tmp
SERVERENGINE_SOCKETMANAGER_2018-05-22T02:23:05Z_2743
deployment.log
forseti_cron_runner.lock
``` | 1.0 | Move cron lock file out of /tmp (will be cleaned up by system) - The cron lock file is currrently placed in `/tmp`, which will be cleaned up by the system everyday. So, this should be moved somewhere else so that we avoid the chance that a cron job is running, and then the lock file gets cleaned up.
```
ubuntu@forseti-server-vm-4330:~/forseti-security$ ls /tmp
SERVERENGINE_SOCKETMANAGER_2018-05-22T02:23:05Z_2743
deployment.log
forseti_cron_runner.lock
``` | infrastructure | move cron lock file out of tmp will be cleaned up by system the cron lock file is currrently placed in tmp which will be cleaned up by the system everyday so this should be moved somewhere else so that we avoid the chance that a cron job is running and then the lock file gets cleaned up ubuntu forseti server vm forseti security ls tmp serverengine socketmanager deployment log forseti cron runner lock | 1 |
532,386 | 15,555,255,049 | IssuesEvent | 2021-03-16 05:45:42 | AY2021S2-CS2113-F10-2/tp | https://api.github.com/repos/AY2021S2-CS2113-F10-2/tp | closed | H-13: Add classes for basic framework | priority.High type.Task | Add classes including User, FitCenter, RecordList, Record, and enum RecordType | 1.0 | H-13: Add classes for basic framework - Add classes including User, FitCenter, RecordList, Record, and enum RecordType | non_infrastructure | h add classes for basic framework add classes including user fitcenter recordlist record and enum recordtype | 0 |
25,196 | 18,251,282,760 | IssuesEvent | 2021-10-02 08:45:32 | kgorbenko/SeasonStats | https://api.github.com/repos/kgorbenko/SeasonStats | closed | Configure ESLint | enhancement infrastructure | We need to audit default `create-react-app` ESLint config and decide if we need to somehow change it. | 1.0 | Configure ESLint - We need to audit default `create-react-app` ESLint config and decide if we need to somehow change it. | infrastructure | configure eslint we need to audit default create react app eslint config and decide if we need to somehow change it | 1 |
131,435 | 10,696,062,294 | IssuesEvent | 2019-10-23 14:07:06 | fedora-infra/bodhi | https://api.github.com/repos/fedora-infra/bodhi | closed | Convert bodhi/tests/client/test___init__.py to PyTest | EasyFix Low Priority Tests | [bodhi/tests/client/test___init__.py](https://github.com/fedora-infra/bodhi/blob/develop/bodhi/tests/client/test___init__.py) still contains classes that derive from `unittest.TestCase`. We use PyTest, so they could be migrated to not derive from `unittest.TestCase`, or derive from `bodhi.tests.server.base.BasePyTestCase` in case they need to use the test database or WSGI app.
This also allows the use of the simpler assert constructs instead of the self.assert*() methods. Instead of unittest `setUp()` or `tearDown()` methods, PyTest uses `setup_method(...)` or `teardown_method(...)`, or analog methods to setup/teardown stuff for different scopes.
The following PR is an example of the changes that might need to be done to this file:
#3612 | 1.0 | Convert bodhi/tests/client/test___init__.py to PyTest - [bodhi/tests/client/test___init__.py](https://github.com/fedora-infra/bodhi/blob/develop/bodhi/tests/client/test___init__.py) still contains classes that derive from `unittest.TestCase`. We use PyTest, so they could be migrated to not derive from `unittest.TestCase`, or derive from `bodhi.tests.server.base.BasePyTestCase` in case they need to use the test database or WSGI app.
This also allows the use of the simpler assert constructs instead of the self.assert*() methods. Instead of unittest `setUp()` or `tearDown()` methods, PyTest uses `setup_method(...)` or `teardown_method(...)`, or analog methods to setup/teardown stuff for different scopes.
The following PR is an example of the changes that might need to be done to this file:
#3612 | non_infrastructure | convert bodhi tests client test init py to pytest still contains classes that derive from unittest testcase we use pytest so they could be migrated to not derive from unittest testcase or derive from bodhi tests server base basepytestcase in case they need to use the test database or wsgi app this also allows the use of the simpler assert constructs instead of the self assert methods instead of unittest setup or teardown methods pytest uses setup method or teardown method or analog methods to setup teardown stuff for different scopes the following pr is an example of the changes that might need to be done to this file | 0 |
33,873 | 27,963,294,548 | IssuesEvent | 2023-03-24 17:15:59 | mitodl/ol-infrastructure | https://api.github.com/repos/mitodl/ol-infrastructure | opened | Send Fastly logs into Grafana for debugging and alerting | Monitoring and Alerting product:ocw product:infrastructure | # User Story
- As a platform engineer I would like to be alerted when a Fastly property is experiencing a spike of error conditions
- As an application engineer I would like to be able to diagnose errors all the way from Fastly through to the backend system
# Description/Context
We need a path for sending Fastly logs into a data source connected to Grafana. The most likely destination is to use Vector as a router for proxying log data into the Grana cloud managed Loki storage system.
# Acceptance Criteria
- [ ] Fastly properties all log as JSON to S3 with a standard message schema
- [ ] Fastly logs are sent to Loki via a Vector proxy using HTTPS or Syslog inputs
- [ ] Vector rejects log data generated by bot traffic as identified by user agent information
- [ ] Fastly logs are available as a data source in Grafana | 1.0 | Send Fastly logs into Grafana for debugging and alerting - # User Story
- As a platform engineer I would like to be alerted when a Fastly property is experiencing a spike of error conditions
- As an application engineer I would like to be able to diagnose errors all the way from Fastly through to the backend system
# Description/Context
We need a path for sending Fastly logs into a data source connected to Grafana. The most likely destination is to use Vector as a router for proxying log data into the Grana cloud managed Loki storage system.
# Acceptance Criteria
- [ ] Fastly properties all log as JSON to S3 with a standard message schema
- [ ] Fastly logs are sent to Loki via a Vector proxy using HTTPS or Syslog inputs
- [ ] Vector rejects log data generated by bot traffic as identified by user agent information
- [ ] Fastly logs are available as a data source in Grafana | infrastructure | send fastly logs into grafana for debugging and alerting user story as a platform engineer i would like to be alerted when a fastly property is experiencing a spike of error conditions as an application engineer i would like to be able to diagnose errors all the way from fastly through to the backend system description context we need a path for sending fastly logs into a data source connected to grafana the most likely destination is to use vector as a router for proxying log data into the grana cloud managed loki storage system acceptance criteria fastly properties all log as json to with a standard message schema fastly logs are sent to loki via a vector proxy using https or syslog inputs vector rejects log data generated by bot traffic as identified by user agent information fastly logs are available as a data source in grafana | 1 |
47,257 | 10,059,308,245 | IssuesEvent | 2019-07-22 15:45:39 | bcgov/entity | https://api.github.com/repos/bcgov/entity | closed | Release Namex-July 15-Phase1 | Namex Code With Us | Prior to moving into production, the following steps should be completed or confirmed:
- [x] Checked devop channels for upgrades, patching
- [x] Update version number in the code being released (Katie did this already)
- [x] Create a draft release in GitHub and confirm the correct commits are present (DONE)
- [x] Add version # and release # to this ticket - DONE.
- [x] schedule the release with staff/clients (daytime's best or when staff are available for rollback)-DONE
- [x] All dev/test pipeline test suites green (none for nae-examination and no e2e ready yet.)
- [x] Tag prod to previous-prod if pipeline cant be used.
- [x] Tag prod to test if pipeline can be used to deploy.
- [x] Release the code to production and complete smoke test
- [x] Finalise/publish the release in GitHub, tagging it
**NameX Smoke Test Script**
- [x] Log in
- [x] Go to search
- [x] Filter for last 30 days
- [x] Go to ZZZ record (usually on Hold)
- [x] edit name 1, click save
- [x] edit address line 1 + comment, click save
- [x] check NRO/oracle tables for changes
- [x] reject name 1 with a conflict
- [x] approve name 2 with quick approve
- [x] wait 5 minutes, then check in NRO
- [x] reload NR in namex
- [x] reset NR
- [x] undo all decisions
- [x] change name 1
- [x] do a conditional approval use mouse to verify behaviour
- [x] RESET
- [x] reject name choice 1 and 2 and quick approve name choice 3
- [x] undo all decisions
- [x] set applicant info back to original values
- [x] put it on hold
- [x] watch [Kibana](https://kibana.pathfinder.gov.bc.ca/) logs for 5 minutes
| 1.0 | Release Namex-July 15-Phase1 - Prior to moving into production, the following steps should be completed or confirmed:
- [x] Checked devop channels for upgrades, patching
- [x] Update version number in the code being released (Katie did this already)
- [x] Create a draft release in GitHub and confirm the correct commits are present (DONE)
- [x] Add version # and release # to this ticket - DONE.
- [x] schedule the release with staff/clients (daytime's best or when staff are available for rollback)-DONE
- [x] All dev/test pipeline test suites green (none for nae-examination and no e2e ready yet.)
- [x] Tag prod to previous-prod if pipeline cant be used.
- [x] Tag prod to test if pipeline can be used to deploy.
- [x] Release the code to production and complete smoke test
- [x] Finalise/publish the release in GitHub, tagging it
**NameX Smoke Test Script**
- [x] Log in
- [x] Go to search
- [x] Filter for last 30 days
- [x] Go to ZZZ record (usually on Hold)
- [x] edit name 1, click save
- [x] edit address line 1 + comment, click save
- [x] check NRO/oracle tables for changes
- [x] reject name 1 with a conflict
- [x] approve name 2 with quick approve
- [x] wait 5 minutes, then check in NRO
- [x] reload NR in namex
- [x] reset NR
- [x] undo all decisions
- [x] change name 1
- [x] do a conditional approval use mouse to verify behaviour
- [x] RESET
- [x] reject name choice 1 and 2 and quick approve name choice 3
- [x] undo all decisions
- [x] set applicant info back to original values
- [x] put it on hold
- [x] watch [Kibana](https://kibana.pathfinder.gov.bc.ca/) logs for 5 minutes
| non_infrastructure | release namex july prior to moving into production the following steps should be completed or confirmed checked devop channels for upgrades patching update version number in the code being released katie did this already create a draft release in github and confirm the correct commits are present done add version and release to this ticket done schedule the release with staff clients daytime s best or when staff are available for rollback done all dev test pipeline test suites green none for nae examination and no ready yet tag prod to previous prod if pipeline cant be used tag prod to test if pipeline can be used to deploy release the code to production and complete smoke test finalise publish the release in github tagging it namex smoke test script log in go to search filter for last days go to zzz record usually on hold edit name click save edit address line comment click save check nro oracle tables for changes reject name with a conflict approve name with quick approve wait minutes then check in nro reload nr in namex reset nr undo all decisions change name do a conditional approval use mouse to verify behaviour reset reject name choice and and quick approve name choice undo all decisions set applicant info back to original values put it on hold watch logs for minutes | 0 |
32,132 | 26,464,245,057 | IssuesEvent | 2023-01-16 21:11:30 | APSIMInitiative/ApsimX | https://api.github.com/repos/APSIMInitiative/ApsimX | closed | Report cannot handle overloaded methods | bug interface/infrastructure | Report throws "Ambiguous" error when trying to report an overloaded method. | 1.0 | Report cannot handle overloaded methods - Report throws "Ambiguous" error when trying to report an overloaded method. | infrastructure | report cannot handle overloaded methods report throws ambiguous error when trying to report an overloaded method | 1 |
19,675 | 13,387,522,061 | IssuesEvent | 2020-09-02 16:06:23 | jmcgeheeiv/pyfakefs | https://api.github.com/repos/jmcgeheeiv/pyfakefs | closed | Tests failing on Ubuntu 20.04 (focal) with python 3.8.2 | infrastructure | I'm trying to package pyfakefs for Ubuntu focal and the build on launchpad fails. I have reproduced the same error on a Virtual Box Ubuntu: a bunch of python3 related packages were already installed for another package (debhelper, dh-python, python3-all, python3-setuptools, tox, python3-pytest, python3-pytest-cov), so I just git cloned the repo (but I also tested with the latest release) and in master branch, ran `pytest-3`.
I've pasted the results [here](https://pastebin.com/6i7dGi0B).
You can check the build log, that contains the same tests errors [here](https://launchpadlibrarian.net/494885084/buildlog_ubuntu-focal-amd64.python3-pyfakefs_4.1.0-0ubuntu1_BUILDING.txt.gz) (quite at bottom of the page).
**Environment**
```
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import platform; print(platform.platform())"
Linux-5.4.0-42-generic-x86_64-with-glibc2.29
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import sys; print('Python', sys.version)"
Python 3.8.2 (default, Jul 16 2020, 14:00:26)
[GCC 9.3.0]
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "from pyfakefs.fake_filesystem import __version__; print('pyfakefs', __version__)"
pyfakefs 4.1.0
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$
```
| 1.0 | Tests failing on Ubuntu 20.04 (focal) with python 3.8.2 - I'm trying to package pyfakefs for Ubuntu focal and the build on launchpad fails. I have reproduced the same error on a Virtual Box Ubuntu: a bunch of python3 related packages were already installed for another package (debhelper, dh-python, python3-all, python3-setuptools, tox, python3-pytest, python3-pytest-cov), so I just git cloned the repo (but I also tested with the latest release) and in master branch, ran `pytest-3`.
I've pasted the results [here](https://pastebin.com/6i7dGi0B).
You can check the build log, that contains the same tests errors [here](https://launchpadlibrarian.net/494885084/buildlog_ubuntu-focal-amd64.python3-pyfakefs_4.1.0-0ubuntu1_BUILDING.txt.gz) (quite at bottom of the page).
**Environment**
```
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import platform; print(platform.platform())"
Linux-5.4.0-42-generic-x86_64-with-glibc2.29
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "import sys; print('Python', sys.version)"
Python 3.8.2 (default, Jul 16 2020, 14:00:26)
[GCC 9.3.0]
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$ python -c "from pyfakefs.fake_filesystem import __version__; print('pyfakefs', __version__)"
pyfakefs 4.1.0
nico@xubuntu2004:~/dev/pyfakefs-4.1.0$
```
| infrastructure | tests failing on ubuntu focal with python i m trying to package pyfakefs for ubuntu focal and the build on launchpad fails i have reproduced the same error on a virtual box ubuntu a bunch of related packages were already installed for another package debhelper dh python all setuptools tox pytest pytest cov so i just git cloned the repo but i also tested with the latest release and in master branch ran pytest i ve pasted the results you can check the build log that contains the same tests errors quite at bottom of the page environment nico dev pyfakefs python c import platform print platform platform linux generic with nico dev pyfakefs python c import sys print python sys version python default jul nico dev pyfakefs python c from pyfakefs fake filesystem import version print pyfakefs version pyfakefs nico dev pyfakefs | 1 |
29,046 | 23,690,430,307 | IssuesEvent | 2022-08-29 10:16:12 | UnitTestBot/UTBotJava | https://api.github.com/repos/UnitTestBot/UTBotJava | opened | Optimize agents using | infrastructure | **Description**
Now build of UTBot Java requires many agents: up to 29. 27 jobs in this build can be ran as concurrent. All of these jobs are doing repeatable task: building UTBot Java project. The task takes up to 7 minutes.
[GitHub allocates up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) for the organization at the same time. As the result, we don't use allocated resources effectively. For the efficient use of resources we can reduce the number of concurrent jobs. Some of the jobs are doing nothing time-consuming, thus, they can be combined with another jobs.
**Expected behavior**
Some of the jobs are combined together with another to provide efficient use of allocated runners.
**Environment**
GitHub Actions
**Potential alternatives**
No alternatives.
**Context**
Not applicable.
| 1.0 | Optimize agents using - **Description**
Now build of UTBot Java requires many agents: up to 29. 27 jobs in this build can be ran as concurrent. All of these jobs are doing repeatable task: building UTBot Java project. The task takes up to 7 minutes.
[GitHub allocates up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) for the organization at the same time. As the result, we don't use allocated resources effectively. For the efficient use of resources we can reduce the number of concurrent jobs. Some of the jobs are doing nothing time-consuming, thus, they can be combined with another jobs.
**Expected behavior**
Some of the jobs are combined together with another to provide efficient use of allocated runners.
**Environment**
GitHub Actions
**Potential alternatives**
No alternatives.
**Context**
Not applicable.
| infrastructure | optimize agents using description now build of utbot java requires many agents up to jobs in this build can be ran as concurrent all of these jobs are doing repeatable task building utbot java project the task takes up to minutes for the organization at the same time as the result we don t use allocated resources effectively for the efficient use of resources we can reduce the number of concurrent jobs some of the jobs are doing nothing time consuming thus they can be combined with another jobs expected behavior some of the jobs are combined together with another to provide efficient use of allocated runners environment github actions potential alternatives no alternatives context not applicable | 1 |
89,787 | 15,837,539,989 | IssuesEvent | 2021-04-06 20:56:25 | TIBCOSoftware/tci-flogo | https://api.github.com/repos/TIBCOSoftware/tci-flogo | closed | WS-2019-0063 (High) detected in js-yaml-3.7.0.tgz - autoclosed | security vulnerability | ## WS-2019-0063 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>
Dependency Hierarchy:
- css-loader-0.28.11.tgz (Root Library)
- cssnano-3.10.0.tgz
- postcss-svgo-2.1.6.tgz
- svgo-0.7.2.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file.
<p>Publish Date: 2019-04-05
<p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p>
<p>Release Date: 2019-04-05</p>
<p>Fix Resolution: js-yaml - 3.13.1</p>
</p>
</details>
<p></p>
| True | WS-2019-0063 (High) detected in js-yaml-3.7.0.tgz - autoclosed - ## WS-2019-0063 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>
Dependency Hierarchy:
- css-loader-0.28.11.tgz (Root Library)
- cssnano-3.10.0.tgz
- postcss-svgo-2.1.6.tgz
- svgo-0.7.2.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file.
<p>Publish Date: 2019-04-05
<p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p>
<p>Release Date: 2019-04-05</p>
<p>Fix Resolution: js-yaml - 3.13.1</p>
</p>
</details>
<p></p>
| non_infrastructure | ws high detected in js yaml tgz autoclosed ws high severity vulnerability vulnerable library js yaml tgz yaml parser and serializer library home page a href dependency hierarchy css loader tgz root library cssnano tgz postcss svgo tgz svgo tgz x js yaml tgz vulnerable library vulnerability details js yaml prior to are vulnerable to code injection the load function may execute arbitrary code injected through a malicious yaml file publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution js yaml | 0 |
33,674 | 27,706,920,005 | IssuesEvent | 2023-03-14 11:51:56 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | PR/issue labeling bot isn't labeling | area-Infrastructure | Nothing is getting automatically labeled by area, for at least the last day or two. | 1.0 | PR/issue labeling bot isn't labeling - Nothing is getting automatically labeled by area, for at least the last day or two. | infrastructure | pr issue labeling bot isn t labeling nothing is getting automatically labeled by area for at least the last day or two | 1 |
26,420 | 20,103,419,792 | IssuesEvent | 2022-02-07 08:04:24 | deckhouse/deckhouse | https://api.github.com/repos/deckhouse/deckhouse | closed | [node-manager] NodeUser multiple ssh keys | type/feature-request area/cluster-and-infrastructure | NodeUser CR adds system users on nodes, and only one ssh key per user is allowed.
| 1.0 | [node-manager] NodeUser multiple ssh keys - NodeUser CR adds system users on nodes, and only one ssh key per user is allowed.
| infrastructure | nodeuser multiple ssh keys nodeuser cr adds system users on nodes and only one ssh key per user is allowed | 1 |
19,365 | 13,224,537,111 | IssuesEvent | 2020-08-17 19:20:18 | algorand/indexer | https://api.github.com/repos/algorand/indexer | closed | Postgres Tx that can be ReadOnly should be | Infrastructure | in idb/postgres.go set TxOption.ReadOnly on all transactions that are read-only.
Current 'read only' mode works on Postgres database users that only have read permissions on tables, but when the whole database is in read-only mode (a hot spare or follower replicated database) it fails on transactions that _might_ write. | 1.0 | Postgres Tx that can be ReadOnly should be - in idb/postgres.go set TxOption.ReadOnly on all transactions that are read-only.
Current 'read only' mode works on Postgres database users that only have read permissions on tables, but when the whole database is in read-only mode (a hot spare or follower replicated database) it fails on transactions that _might_ write. | infrastructure | postgres tx that can be readonly should be in idb postgres go set txoption readonly on all transactions that are read only current read only mode works on postgres database users that only have read permissions on tables but when the whole database is in read only mode a hot spare or follower replicated database it fails on transactions that might write | 1 |
582 | 2,775,996,000 | IssuesEvent | 2015-05-04 19:14:16 | fsr-itse/1327 | https://api.github.com/repos/fsr-itse/1327 | closed | Provide libraries locally | infrastructure minor | The project should not import any libraries from Google APIs or any other remote server. | 1.0 | Provide libraries locally - The project should not import any libraries from Google APIs or any other remote server. | infrastructure | provide libraries locally the project should not import any libraries from google apis or any other remote server | 1 |
29,613 | 24,107,205,390 | IssuesEvent | 2022-09-20 08:27:51 | UnitTestBot/UTBotJava | https://api.github.com/repos/UnitTestBot/UTBotJava | closed | Reduce concurrent jobs number for projects | infrastructure | **Description**
Now tests for every separate project run in separate job. These jobs are concurrent. [GitHub allocates for organization up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) that can run at the same time.
**Expected behavior**
Reduce concurrent jobs number for projects by combining some projects builds together.
**Environment**
GitHub Actions.
**Potential alternatives**
No alternatives.
**Context**
No context.
| 1.0 | Reduce concurrent jobs number for projects - **Description**
Now tests for every separate project run in separate job. These jobs are concurrent. [GitHub allocates for organization up to 20 runners](https://docs.github.com/en/actions/learn-github-actions/usage-limits-billing-and-administration) that can run at the same time.
**Expected behavior**
Reduce concurrent jobs number for projects by combining some projects builds together.
**Environment**
GitHub Actions.
**Potential alternatives**
No alternatives.
**Context**
No context.
| infrastructure | reduce concurrent jobs number for projects description now tests for every separate project run in separate job these jobs are concurrent that can run at the same time expected behavior reduce concurrent jobs number for projects by combining some projects builds together environment github actions potential alternatives no alternatives context no context | 1 |
235,616 | 25,955,245,786 | IssuesEvent | 2022-12-18 05:41:41 | Dima2022/easybuggy | https://api.github.com/repos/Dima2022/easybuggy | closed | CVE-2021-4104 (High) detected in log4j-1.2.13.jar - autoclosed | security vulnerability | ## CVE-2021-4104 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/easybuggy/commit/aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f">aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2021-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-4104">https://nvd.nist.gov/vuln/detail/CVE-2021-4104</a></p>
<p>Release Date: 2021-12-14</p>
<p>Fix Resolution: uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | True | CVE-2021-4104 (High) detected in log4j-1.2.13.jar - autoclosed - ## CVE-2021-4104 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.13.jar</b></p></summary>
<p>Log4j</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/">http://logging.apache.org/log4j/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /target/easybuggy-1-SNAPSHOT/WEB-INF/lib/log4j-1.2.13.jar,/home/wss-scanner/.m2/repository/log4j/log4j/1.2.13/log4j-1.2.13.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.13.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Dima2022/easybuggy/commit/aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f">aac8d904642c672c7ec784cc1cbbcf6e9c4c9c3f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
JMSAppender in Log4j 1.2 is vulnerable to deserialization of untrusted data when the attacker has write access to the Log4j configuration. The attacker can provide TopicBindingName and TopicConnectionFactoryBindingName configurations causing JMSAppender to perform JNDI requests that result in remote code execution in a similar fashion to CVE-2021-44228. Note this issue only affects Log4j 1.2 when specifically configured to use JMSAppender, which is not the default. Apache Log4j 1.2 reached end of life in August 2015. Users should upgrade to Log4j 2 as it addresses numerous other issues from the previous versions.
<p>Publish Date: 2021-12-14
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-4104>CVE-2021-4104</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-4104">https://nvd.nist.gov/vuln/detail/CVE-2021-4104</a></p>
<p>Release Date: 2021-12-14</p>
<p>Fix Resolution: uom-parent - 1.0.3-3.module,1.0.3-3.module;uom-se-javadoc - 1.0.4-3.module;parfait-examples - 0.5.4-4.module;log4j-manual - 1.2.17-16;si-units-javadoc - 0.6.5-2.module;unit-api - 1.0-5.module,1.0-5.module;unit-api-javadoc - 1.0-5.module;parfait - 0.5.4-4.module,0.5.4-4.module;log4j-javadoc - 1.2.17-16;uom-systems-javadoc - 0.7-1.module;uom-lib-javadoc - 1.0.1-6.module;uom-systems - 0.7-1.module,0.7-1.module;log4j - 1.2.17-16,1.2.17-16;uom-se - 1.0.4-3.module,1.0.4-3.module;uom-lib - 1.0.1-6.module,1.0.1-6.module;parfait-javadoc - 0.5.4-4.module;pcp-parfait-agent - 0.5.4-4.module;si-units - 0.6.5-2.module,0.6.5-2.module</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue | non_infrastructure | cve high detected in jar autoclosed cve high severity vulnerability vulnerable library jar library home page a href path to dependency file pom xml path to vulnerable library target easybuggy snapshot web inf lib jar home wss scanner repository jar dependency hierarchy x jar vulnerable library found in head commit a href found in base branch master vulnerability details jmsappender in is vulnerable to deserialization of untrusted data when the attacker has write access to the configuration the attacker can provide topicbindingname and topicconnectionfactorybindingname configurations causing jmsappender to perform jndi requests that result in remote code execution in a similar fashion to cve note this issue only affects when specifically configured to use jmsappender which is not the default apache reached end of life in august users should upgrade to as it addresses numerous other issues from the previous versions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution uom parent module module uom se javadoc module parfait examples module manual si units javadoc module unit api module module unit api javadoc module parfait module module javadoc uom systems javadoc module uom lib javadoc module uom systems module module uom se module module uom lib module module parfait javadoc module pcp parfait agent module si units module module rescue worker helmet automatic remediation is available for this issue | 0 |
130,679 | 27,744,418,329 | IssuesEvent | 2023-03-15 16:07:40 | Regalis11/Barotrauma | https://api.github.com/repos/Regalis11/Barotrauma | closed | duplicate submarine in editor (again) | Bug Need more info Code | ### Disclaimers
- [X] I have searched the issue tracker to check if the issue has already been reported.
- [ ] My issue happened while using mods.
### What happened?
once again, there's duplicate entries in the submarine loader


but only for this one. my other submarines are not duplicate
### Reproduction steps
_No response_
### Bug prevalence
Happens every now and then
### Version
0.19.14.0
### -
_No response_
### Which operating system did you encounter this bug on?
Windows
### Relevant error messages and crash reports
_No response_ | 1.0 | duplicate submarine in editor (again) - ### Disclaimers
- [X] I have searched the issue tracker to check if the issue has already been reported.
- [ ] My issue happened while using mods.
### What happened?
once again, there's duplicate entries in the submarine loader


but only for this one. my other submarines are not duplicate
### Reproduction steps
_No response_
### Bug prevalence
Happens every now and then
### Version
0.19.14.0
### -
_No response_
### Which operating system did you encounter this bug on?
Windows
### Relevant error messages and crash reports
_No response_ | non_infrastructure | duplicate submarine in editor again disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened once again there s duplicate entries in the submarine loader but only for this one my other submarines are not duplicate reproduction steps no response bug prevalence happens every now and then version no response which operating system did you encounter this bug on windows relevant error messages and crash reports no response | 0 |
45,615 | 2,937,573,977 | IssuesEvent | 2015-07-01 03:18:21 | isuPatches/RockThePatch | https://api.github.com/repos/isuPatches/RockThePatch | closed | GitHub API has a rate limit and messages that should be shown | Affected- Develop Bug Bug Severity - Priority 3 (medium) Bug Source - External Source Bug Type - Other Fixed In - Develop Found By Me Found In - Develop Resolution - Fixed Status - Fixed And Has A Commit | Sometimes there's a hiccup and I need to show the GitHub API message or an error dialog instead of attempting to still show the github issues for bugs and projects. | 1.0 | GitHub API has a rate limit and messages that should be shown - Sometimes there's a hiccup and I need to show the GitHub API message or an error dialog instead of attempting to still show the github issues for bugs and projects. | non_infrastructure | github api has a rate limit and messages that should be shown sometimes there s a hiccup and i need to show the github api message or an error dialog instead of attempting to still show the github issues for bugs and projects | 0 |
13,320 | 10,209,188,033 | IssuesEvent | 2019-08-14 12:06:42 | eclipse/antenna | https://api.github.com/repos/eclipse/antenna | closed | Java9 travis tests currently always fail with 10min timeout | bug infrastructure wontfix | ### Summary of the Bug
see e.g.
- https://travis-ci.org/eclipse/antenna/jobs/550295410
### Acceptance Criteria
- [ ] travis tests no longer fail
### Definition of Done
- [ ] Acceptance criteria fulfilled
- [ ] A test case is created to reproduce the bug
- [ ] A PR is created, the CI infrastructure reports green, the bug test case proves that bug is fixed
- [ ] The PR is reviewed and approved
- [ ] No TODOs left in the code unless explained in the ticket, if something else is still open, this is summarized in a comment in the issue
- [ ] Documentation is updated
| 1.0 | Java9 travis tests currently always fail with 10min timeout - ### Summary of the Bug
see e.g.
- https://travis-ci.org/eclipse/antenna/jobs/550295410
### Acceptance Criteria
- [ ] travis tests no longer fail
### Definition of Done
- [ ] Acceptance criteria fulfilled
- [ ] A test case is created to reproduce the bug
- [ ] A PR is created, the CI infrastructure reports green, the bug test case proves that bug is fixed
- [ ] The PR is reviewed and approved
- [ ] No TODOs left in the code unless explained in the ticket, if something else is still open, this is summarized in a comment in the issue
- [ ] Documentation is updated
| infrastructure | travis tests currently always fail with timeout summary of the bug see e g acceptance criteria travis tests no longer fail definition of done acceptance criteria fulfilled a test case is created to reproduce the bug a pr is created the ci infrastructure reports green the bug test case proves that bug is fixed the pr is reviewed and approved no todos left in the code unless explained in the ticket if something else is still open this is summarized in a comment in the issue documentation is updated | 1 |
374,841 | 11,096,189,789 | IssuesEvent | 2019-12-16 10:37:46 | JuPedSim/jpscore | https://api.github.com/repos/JuPedSim/jpscore | closed | Python integration test script | Priority: Medium Status: Ready Type: Enhancement | For the integration tests we need a "central" python script for the execution of the test cases. This script should sort the integration tests by approximate runtime in three different groups: slow, medium and fast.
It should accept command-line parameters for setting the JPScore binary and setting the desired group (slow, medium, fast, all). This script should also manage where log files from the test execution are written. Additionally we need a command line parameter accepting a list of test to execute. This should be used to execute a single test in ctest/cmake.
ACs:
- Test script in python for executing integration tests
- Command-line parameter for JPScore binary path
- Grouped test depending on approximate runtime
- Command line parameter for selecting the group to execute
- Command line parameter to define a list of tests to execute
(see also: #491) | 1.0 | Python integration test script - For the integration tests we need a "central" python script for the execution of the test cases. This script should sort the integration tests by approximate runtime in three different groups: slow, medium and fast.
It should accept command-line parameters for setting the JPScore binary and setting the desired group (slow, medium, fast, all). This script should also manage where log files from the test execution are written. Additionally we need a command line parameter accepting a list of test to execute. This should be used to execute a single test in ctest/cmake.
ACs:
- Test script in python for executing integration tests
- Command-line parameter for JPScore binary path
- Grouped test depending on approximate runtime
- Command line parameter for selecting the group to execute
- Command line parameter to define a list of tests to execute
(see also: #491) | non_infrastructure | python integration test script for the integration tests we need a central python script for the execution of the test cases this script should sort the integration tests by approximate runtime in three different groups slow medium and fast it should accept command line parameters for setting the jpscore binary and setting the desired group slow medium fast all this script should also manage where log files from the test execution are written additionally we need a command line parameter accepting a list of test to execute this should be used to execute a single test in ctest cmake acs test script in python for executing integration tests command line parameter for jpscore binary path grouped test depending on approximate runtime command line parameter for selecting the group to execute command line parameter to define a list of tests to execute see also | 0 |
30,855 | 25,119,715,523 | IssuesEvent | 2022-11-09 06:55:43 | astropy/astroquery | https://api.github.com/repos/astropy/astroquery | closed | Revise keyring version limitation in tests | testing infrastructure | Once this fix end up in a conda packaged version we may lift the current limitation of keyring to <12.
https://github.com/jaraco/keyring/pull/320 | 1.0 | Revise keyring version limitation in tests - Once this fix end up in a conda packaged version we may lift the current limitation of keyring to <12.
https://github.com/jaraco/keyring/pull/320 | infrastructure | revise keyring version limitation in tests once this fix end up in a conda packaged version we may lift the current limitation of keyring to | 1 |
7,454 | 6,963,845,311 | IssuesEvent | 2017-12-08 19:05:13 | KSP-CKAN/CKAN | https://api.github.com/repos/KSP-CKAN/CKAN | opened | CKAN's GitHub downloads are breaking the rules | ckan.dll Cmdline GUI Infrastructure Metadata Netkan Schema Spec | # GitHub downloading needs a rewrite
(I debated whether to add this as a comment to #1817, but it seems like too much text and detail for that.)
## Problems
Currently if CKAN downloads many files from GitHub at the same time, they often fail with HTTP status code 403-Forbidden. #1817 contains an example, but these reports are common and I've definitely seen it happen myself several times.
## Background
The GitHub API uses 403 codes for throttling; you get 60 unauthenticated requests per hour, and any beyond that return a 403. I encountered this while working on [an unrelated project](https://hebarusan.github.io/releases-timeline/), and I had to use a GitHub token to allow 5000/hour, passed in the HTTP request headers:
```http
Authentication: token <OAuth token here>
```
Currently CKAN's downloads do **not** go through the GitHub API, so this does not necessarily indicate exactly what's going on with them. However, it establishes that 403-Forbidden is sometimes used for throttling, and it becomes more relevant later in discussion of the API.
[Sample API data for releases](https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924), minus the author and uploader fields since they're long and not relevant to this issue:
```json
{
"url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924",
"assets_url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets",
"upload_url": "https://uploads.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets{?name,label}",
"html_url": "https://github.com/HebaruSan/Astrogator/releases/tag/v0.7.8",
"id": 7538924,
"tag_name": "v0.7.8",
"target_commitish": "master",
"name": "Frictionless toilet",
"draft": false,
"prerelease": false,
"created_at": "2017-08-28T06:00:32Z",
"published_at": "2017-08-28T06:09:07Z",
"assets": [
{
"url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/assets/4682631",
"id": 4682631,
"name": "Astrogator.zip",
"label": null,
"content_type": "application/zip",
"state": "uploaded",
"size": 114372,
"download_count": 5243,
"created_at": "2017-08-28T06:08:01Z",
"updated_at": "2017-08-28T06:08:02Z",
"browser_download_url": "https://github.com/HebaruSan/Astrogator/releases/download/v0.7.8/Astrogator.zip"
}
],
"tarball_url": "https://api.github.com/repos/HebaruSan/Astrogator/tarball/v0.7.8",
"zipball_url": "https://api.github.com/repos/HebaruSan/Astrogator/zipball/v0.7.8",
"body": "Fix glitches when the settings file is invalid"
}
```
The zip file that we want to download is associated with `assets[0]`, and there are two fields for it, `url` and `browser_download_url`. This becomes important later.
## Investigation summary
I used the "Contact GitHub" link to reach out to GitHub about how their download throttling works. Surprisingly, the person who replied understood exactly what I was talking about and how to fix it :+1:. It turns out that these problems happen because CKAN is not using GitHub as intended. From my conversation with the very helpful support person:
> If I understood your message correctly, it seems like you're programmatically downloading resources from github.com, is that right? If that's so, then you shouldn't be doing that. GitHub.com wasn't build for programmatic use like that, it was built for humans. For programmatic use, you should be using the API. The API has well defined rate limits and caching behavior you can rely on, while GitHub.com doesn't. That doesn't mean that github.com doesn't have any rate limits, it only means that you can be rate limited at any time and without warning.
>
> So, we'd like to ask you to switch and use the API for downloading the data you need, and respect the defined rate limits (that's what a good citizen app should be doing, instead of hitting github.com):
("good citizen" was my phrasing in my original message, so don't take that as an unprovoked criticism of our civic virtues.)
> If I'm interpreting that code snippet correctly, you're using the browser_download_url link, which, as the name suggests, is intended to be used by human users via a browser.
>
> For downloading release assets via the API, you should be using this endpoint:
>
> https://developer.github.com/v3/repos/releases/#get-a-single-release-asset
>
> Notice this note: "To download the asset's binary content, set the Accept header of the request to application/octet-stream. The API will either redirect the client to the location, or stream it directly if possible. API clients should handle both a 200 or 302 response."
>
> That would be the "url" field of a particular asset (which are listed when you fetch a release e.g. via https://developer.github.com/v3/repos/releases/#get-a-single-release), but with the addition of the special Accept header.
Key points:
- The URL from the field we're using currently (`browser_download_url`) is for users and browsers only, not applications. It can be throttled, but there is no explicit policy or workaround.
- We should be using the GitHub API for downloads. Currently we use it in the Netkan code that finds new releases, but for downloads we effectively impersonate a browser.
- This can be done by requesting the `url` field instead of `browser_download_url` and setting a custom HTTP header:
```http
Accept: application/octet-stream
```
I tested this with wget, and setting the Accept header did indeed give me the download. Without this header, it returns a JSON object describing the asset.
## Changes needed to stop abusing `browser_download_url`
### GitHub-specific downloading metadata & logic
When downloading from GitHub, we need to send the custom HTTP header. This cannot be accomplished simply by swapping out the bad URL for the good URL in the `download` metadata field.
Proposed new metadata field:
- `github_download` - The `assets[0].url` value from the API
Specific changes:
- The spec/schema would need to be updated to allow this field.
- Netkan would need to be updated to generate this field.
- CKAN would need to be updated to check for the presence of this field, which would then trigger an alternate download method that sets the custom header.
### UI to handle 403 statuses
If a GitHub download returns a 403 status, we should handle the exception and notify the user that their downloads are being throttled. We could direct them to the setting (see below) and [web page dealing with GitHub auth tokens](https://github.com/settings/tokens), and/or advise them to wait 60 minutes for their limit to reset. https://api.github.com/rate_limit can be used to get the exact limit and timing numbers.
### GitHub token handling
Users will be limited to 60 GitHub downloads per hour, because this is the limit of the GitHub API. 140+ mod installs are pretty commonly mentioned on the forums, and reinstalling everything from scratch is a common method for dealing with compatible upgrades, so some users would probably encounter this limit and not appreciate the 60-minute wait to be able to download more. The only way around this is to use a GitHub auth token, which boosts the limit to 5000/hour per token.
It would be nice to ship a single internal auth token for all of CKAN, since then users would have the 5000/hour limit by default without having to worry about any of the details. More responses from the GitHub contact person:
> Including a single pre-defined token with the app so that this token is used by all users of the app is possible. You could create a scopeless token here https://github.com/settings/tokens/new and include that. A scopeless token doesn't have any special permissions -- it can be used for read-only access of public data. So, it would be safe in that way. However, someone could easily take your token from the app, and then drain the API quota for the user who owns the token by making lots of unnecessary API requests. At that point, the app would stop working for everyone who uses the app.
Deliberate abuse like that is unlikely, but assuming 200 downloads per active user per hour, a 5000/hour limit across all CKAN users would support 25 active users in a given hour. The number of active users at a given time isn't known, but the latest CKAN release has over 60000 downloads, so it's probably more than 25. If we were able to determine the limit we needed per hour, we could divide it by 5000 and then generate that many tokens and pick one randomly per request, but that might not be in the spirit of the API's rules.
#### A setting
We could create a new settings field called GitHub Auth Token, where the user could fill in their own tokens to allow more downloads. This could be instead of or in addition to any built-in tokens we may or may not use, and it should support all the UIs.
#### Multi-pass approach
1. Try with no authentication at all. This would succeed for the first 60 requests per user per hour, probably the majority but not all.
2. For the remaining requests that fail, retry with a single hard coded auth token. As long as we only use this as a fallback, the 5000/hour limit would only apply to downloads in excess the 60/hour.
### Migration concerns
If Netkan was updated to use this new scheme tomorrow, current CKAN clients would break unless the old `download` field was still populated. So we should not remove support for the old metadata immediately; GitHub downloads should use both `download` and `github_download` until all clients are updated.
## Or just download serially
The [API docs](https://developer.github.com/v3/guides/best-practices-for-integrators/) say:
> Make requests for a single user or client ID serially. Do not make requests for a single user or client ID concurrently.
So even with a token, CKAN's parallel download method would still be in violation of the letter of the law.
As a halfway measure, we could try scaling back the parallelization of downloads.
1. Check whether a download URL contains "github.com"
2. If so, add it to a pool of downloads to be handled serially
3. Handle all other downloads normally
4. When a download finishes, if it contains "github.com", then start a new download from the pool
This might solve the problem without messing with all the API/token stuff. We would still technically be misusing GitHub, but users should no longer encounter failed downloads as frequently. | 1.0 | CKAN's GitHub downloads are breaking the rules - # GitHub downloading needs a rewrite
(I debated whether to add this as a comment to #1817, but it seems like too much text and detail for that.)
## Problems
Currently if CKAN downloads many files from GitHub at the same time, they often fail with HTTP status code 403-Forbidden. #1817 contains an example, but these reports are common and I've definitely seen it happen myself several times.
## Background
The GitHub API uses 403 codes for throttling; you get 60 unauthenticated requests per hour, and any beyond that return a 403. I encountered this while working on [an unrelated project](https://hebarusan.github.io/releases-timeline/), and I had to use a GitHub token to allow 5000/hour, passed in the HTTP request headers:
```http
Authentication: token <OAuth token here>
```
Currently CKAN's downloads do **not** go through the GitHub API, so this does not necessarily indicate exactly what's going on with them. However, it establishes that 403-Forbidden is sometimes used for throttling, and it becomes more relevant later in discussion of the API.
[Sample API data for releases](https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924), minus the author and uploader fields since they're long and not relevant to this issue:
```json
{
"url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924",
"assets_url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets",
"upload_url": "https://uploads.github.com/repos/HebaruSan/Astrogator/releases/7538924/assets{?name,label}",
"html_url": "https://github.com/HebaruSan/Astrogator/releases/tag/v0.7.8",
"id": 7538924,
"tag_name": "v0.7.8",
"target_commitish": "master",
"name": "Frictionless toilet",
"draft": false,
"prerelease": false,
"created_at": "2017-08-28T06:00:32Z",
"published_at": "2017-08-28T06:09:07Z",
"assets": [
{
"url": "https://api.github.com/repos/HebaruSan/Astrogator/releases/assets/4682631",
"id": 4682631,
"name": "Astrogator.zip",
"label": null,
"content_type": "application/zip",
"state": "uploaded",
"size": 114372,
"download_count": 5243,
"created_at": "2017-08-28T06:08:01Z",
"updated_at": "2017-08-28T06:08:02Z",
"browser_download_url": "https://github.com/HebaruSan/Astrogator/releases/download/v0.7.8/Astrogator.zip"
}
],
"tarball_url": "https://api.github.com/repos/HebaruSan/Astrogator/tarball/v0.7.8",
"zipball_url": "https://api.github.com/repos/HebaruSan/Astrogator/zipball/v0.7.8",
"body": "Fix glitches when the settings file is invalid"
}
```
The zip file that we want to download is associated with `assets[0]`, and there are two fields for it, `url` and `browser_download_url`. This becomes important later.
## Investigation summary
I used the "Contact GitHub" link to reach out to GitHub about how their download throttling works. Surprisingly, the person who replied understood exactly what I was talking about and how to fix it :+1:. It turns out that these problems happen because CKAN is not using GitHub as intended. From my conversation with the very helpful support person:
> If I understood your message correctly, it seems like you're programmatically downloading resources from github.com, is that right? If that's so, then you shouldn't be doing that. GitHub.com wasn't build for programmatic use like that, it was built for humans. For programmatic use, you should be using the API. The API has well defined rate limits and caching behavior you can rely on, while GitHub.com doesn't. That doesn't mean that github.com doesn't have any rate limits, it only means that you can be rate limited at any time and without warning.
>
> So, we'd like to ask you to switch and use the API for downloading the data you need, and respect the defined rate limits (that's what a good citizen app should be doing, instead of hitting github.com):
("good citizen" was my phrasing in my original message, so don't take that as an unprovoked criticism of our civic virtues.)
> If I'm interpreting that code snippet correctly, you're using the browser_download_url link, which, as the name suggests, is intended to be used by human users via a browser.
>
> For downloading release assets via the API, you should be using this endpoint:
>
> https://developer.github.com/v3/repos/releases/#get-a-single-release-asset
>
> Notice this note: "To download the asset's binary content, set the Accept header of the request to application/octet-stream. The API will either redirect the client to the location, or stream it directly if possible. API clients should handle both a 200 or 302 response."
>
> That would be the "url" field of a particular asset (which are listed when you fetch a release e.g. via https://developer.github.com/v3/repos/releases/#get-a-single-release), but with the addition of the special Accept header.
Key points:
- The URL from the field we're using currently (`browser_download_url`) is for users and browsers only, not applications. It can be throttled, but there is no explicit policy or workaround.
- We should be using the GitHub API for downloads. Currently we use it in the Netkan code that finds new releases, but for downloads we effectively impersonate a browser.
- This can be done by requesting the `url` field instead of `browser_download_url` and setting a custom HTTP header:
```http
Accept: application/octet-stream
```
I tested this with wget, and setting the Accept header did indeed give me the download. Without this header, it returns a JSON object describing the asset.
## Changes needed to stop abusing `browser_download_url`
### GitHub-specific downloading metadata & logic
When downloading from GitHub, we need to send the custom HTTP header. This cannot be accomplished simply by swapping out the bad URL for the good URL in the `download` metadata field.
Proposed new metadata field:
- `github_download` - The `assets[0].url` value from the API
Specific changes:
- The spec/schema would need to be updated to allow this field.
- Netkan would need to be updated to generate this field.
- CKAN would need to be updated to check for the presence of this field, which would then trigger an alternate download method that sets the custom header.
### UI to handle 403 statuses
If a GitHub download returns a 403 status, we should handle the exception and notify the user that their downloads are being throttled. We could direct them to the setting (see below) and [web page dealing with GitHub auth tokens](https://github.com/settings/tokens), and/or advise them to wait 60 minutes for their limit to reset. https://api.github.com/rate_limit can be used to get the exact limit and timing numbers.
### GitHub token handling
Users will be limited to 60 GitHub downloads per hour, because this is the limit of the GitHub API. 140+ mod installs are pretty commonly mentioned on the forums, and reinstalling everything from scratch is a common method for dealing with compatible upgrades, so some users would probably encounter this limit and not appreciate the 60-minute wait to be able to download more. The only way around this is to use a GitHub auth token, which boosts the limit to 5000/hour per token.
It would be nice to ship a single internal auth token for all of CKAN, since then users would have the 5000/hour limit by default without having to worry about any of the details. More responses from the GitHub contact person:
> Including a single pre-defined token with the app so that this token is used by all users of the app is possible. You could create a scopeless token here https://github.com/settings/tokens/new and include that. A scopeless token doesn't have any special permissions -- it can be used for read-only access of public data. So, it would be safe in that way. However, someone could easily take your token from the app, and then drain the API quota for the user who owns the token by making lots of unnecessary API requests. At that point, the app would stop working for everyone who uses the app.
Deliberate abuse like that is unlikely, but assuming 200 downloads per active user per hour, a 5000/hour limit across all CKAN users would support 25 active users in a given hour. The number of active users at a given time isn't known, but the latest CKAN release has over 60000 downloads, so it's probably more than 25. If we were able to determine the limit we needed per hour, we could divide it by 5000 and then generate that many tokens and pick one randomly per request, but that might not be in the spirit of the API's rules.
#### A setting
We could create a new settings field called GitHub Auth Token, where the user could fill in their own tokens to allow more downloads. This could be instead of or in addition to any built-in tokens we may or may not use, and it should support all the UIs.
#### Multi-pass approach
1. Try with no authentication at all. This would succeed for the first 60 requests per user per hour, probably the majority but not all.
2. For the remaining requests that fail, retry with a single hard coded auth token. As long as we only use this as a fallback, the 5000/hour limit would only apply to downloads in excess the 60/hour.
### Migration concerns
If Netkan was updated to use this new scheme tomorrow, current CKAN clients would break unless the old `download` field was still populated. So we should not remove support for the old metadata immediately; GitHub downloads should use both `download` and `github_download` until all clients are updated.
## Or just download serially
The [API docs](https://developer.github.com/v3/guides/best-practices-for-integrators/) say:
> Make requests for a single user or client ID serially. Do not make requests for a single user or client ID concurrently.
So even with a token, CKAN's parallel download method would still be in violation of the letter of the law.
As a halfway measure, we could try scaling back the parallelization of downloads.
1. Check whether a download URL contains "github.com"
2. If so, add it to a pool of downloads to be handled serially
3. Handle all other downloads normally
4. When a download finishes, if it contains "github.com", then start a new download from the pool
This might solve the problem without messing with all the API/token stuff. We would still technically be misusing GitHub, but users should no longer encounter failed downloads as frequently. | infrastructure | ckan s github downloads are breaking the rules github downloading needs a rewrite i debated whether to add this as a comment to but it seems like too much text and detail for that problems currently if ckan downloads many files from github at the same time they often fail with http status code forbidden contains an example but these reports are common and i ve definitely seen it happen myself several times background the github api uses codes for throttling you get unauthenticated requests per hour and any beyond that return a i encountered this while working on and i had to use a github token to allow hour passed in the http request headers http authentication token currently ckan s downloads do not go through the github api so this does not necessarily indicate exactly what s going on with them however it establishes that forbidden is sometimes used for throttling and it becomes more relevant later in discussion of the api minus the author and uploader fields since they re long and not relevant to this issue json url assets url upload url html url id tag name target commitish master name frictionless toilet draft false prerelease false created at published at assets url id name astrogator zip label null content type application zip state uploaded size download count created at updated at browser download url tarball url zipball url body fix glitches when the settings file is invalid the zip file that we want to download is associated with assets and there are two fields for it url and browser download url this becomes important later investigation summary i used the contact github link to reach out to github about how their download throttling works surprisingly the person who replied understood exactly what i was talking about and how to fix it it turns out that these problems happen because ckan is not using github as intended from my conversation with the very helpful support person if i understood your message correctly it seems like you re programmatically downloading resources from github com is that right if that s so then you shouldn t be doing that github com wasn t build for programmatic use like that it was built for humans for programmatic use you should be using the api the api has well defined rate limits and caching behavior you can rely on while github com doesn t that doesn t mean that github com doesn t have any rate limits it only means that you can be rate limited at any time and without warning so we d like to ask you to switch and use the api for downloading the data you need and respect the defined rate limits that s what a good citizen app should be doing instead of hitting github com good citizen was my phrasing in my original message so don t take that as an unprovoked criticism of our civic virtues if i m interpreting that code snippet correctly you re using the browser download url link which as the name suggests is intended to be used by human users via a browser for downloading release assets via the api you should be using this endpoint notice this note to download the asset s binary content set the accept header of the request to application octet stream the api will either redirect the client to the location or stream it directly if possible api clients should handle both a or response that would be the url field of a particular asset which are listed when you fetch a release e g via but with the addition of the special accept header key points the url from the field we re using currently browser download url is for users and browsers only not applications it can be throttled but there is no explicit policy or workaround we should be using the github api for downloads currently we use it in the netkan code that finds new releases but for downloads we effectively impersonate a browser this can be done by requesting the url field instead of browser download url and setting a custom http header http accept application octet stream i tested this with wget and setting the accept header did indeed give me the download without this header it returns a json object describing the asset changes needed to stop abusing browser download url github specific downloading metadata logic when downloading from github we need to send the custom http header this cannot be accomplished simply by swapping out the bad url for the good url in the download metadata field proposed new metadata field github download the assets url value from the api specific changes the spec schema would need to be updated to allow this field netkan would need to be updated to generate this field ckan would need to be updated to check for the presence of this field which would then trigger an alternate download method that sets the custom header ui to handle statuses if a github download returns a status we should handle the exception and notify the user that their downloads are being throttled we could direct them to the setting see below and and or advise them to wait minutes for their limit to reset can be used to get the exact limit and timing numbers github token handling users will be limited to github downloads per hour because this is the limit of the github api mod installs are pretty commonly mentioned on the forums and reinstalling everything from scratch is a common method for dealing with compatible upgrades so some users would probably encounter this limit and not appreciate the minute wait to be able to download more the only way around this is to use a github auth token which boosts the limit to hour per token it would be nice to ship a single internal auth token for all of ckan since then users would have the hour limit by default without having to worry about any of the details more responses from the github contact person including a single pre defined token with the app so that this token is used by all users of the app is possible you could create a scopeless token here and include that a scopeless token doesn t have any special permissions it can be used for read only access of public data so it would be safe in that way however someone could easily take your token from the app and then drain the api quota for the user who owns the token by making lots of unnecessary api requests at that point the app would stop working for everyone who uses the app deliberate abuse like that is unlikely but assuming downloads per active user per hour a hour limit across all ckan users would support active users in a given hour the number of active users at a given time isn t known but the latest ckan release has over downloads so it s probably more than if we were able to determine the limit we needed per hour we could divide it by and then generate that many tokens and pick one randomly per request but that might not be in the spirit of the api s rules a setting we could create a new settings field called github auth token where the user could fill in their own tokens to allow more downloads this could be instead of or in addition to any built in tokens we may or may not use and it should support all the uis multi pass approach try with no authentication at all this would succeed for the first requests per user per hour probably the majority but not all for the remaining requests that fail retry with a single hard coded auth token as long as we only use this as a fallback the hour limit would only apply to downloads in excess the hour migration concerns if netkan was updated to use this new scheme tomorrow current ckan clients would break unless the old download field was still populated so we should not remove support for the old metadata immediately github downloads should use both download and github download until all clients are updated or just download serially the say make requests for a single user or client id serially do not make requests for a single user or client id concurrently so even with a token ckan s parallel download method would still be in violation of the letter of the law as a halfway measure we could try scaling back the parallelization of downloads check whether a download url contains github com if so add it to a pool of downloads to be handled serially handle all other downloads normally when a download finishes if it contains github com then start a new download from the pool this might solve the problem without messing with all the api token stuff we would still technically be misusing github but users should no longer encounter failed downloads as frequently | 1 |
12,544 | 9,819,543,776 | IssuesEvent | 2019-06-13 22:23:55 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | Implement Preview strategies for commands/command groups/parameters | Infrastructure | We currently have a mechanism to show that extensions are in preview, but we have no such thing for specific CLI commands, command groups, or parameters. Even within a GA module, certain parameters may be added in preview, for example `--ddos-protection` for `vnet create`. It would be good to standardize the experience for these: perhaps issue a warning if a preview command or parameter is used so that users are warned that these things may break and thus to avoid using them in scripts. Displaying this warning or not could be a configurable setting.
---
### Environment summary
Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows)
```
azure-cli (2.0.30)
```
| 1.0 | Implement Preview strategies for commands/command groups/parameters - We currently have a mechanism to show that extensions are in preview, but we have no such thing for specific CLI commands, command groups, or parameters. Even within a GA module, certain parameters may be added in preview, for example `--ddos-protection` for `vnet create`. It would be good to standardize the experience for these: perhaps issue a warning if a preview command or parameter is used so that users are warned that these things may break and thus to avoid using them in scripts. Displaying this warning or not could be a configurable setting.
---
### Environment summary
Install Method (e.g. pip, interactive script, apt-get, Docker, MSI, edge build) / CLI version (`az --version`) / OS version / Shell Type (e.g. bash, cmd.exe, Bash on Windows)
```
azure-cli (2.0.30)
```
| infrastructure | implement preview strategies for commands command groups parameters we currently have a mechanism to show that extensions are in preview but we have no such thing for specific cli commands command groups or parameters even within a ga module certain parameters may be added in preview for example ddos protection for vnet create it would be good to standardize the experience for these perhaps issue a warning if a preview command or parameter is used so that users are warned that these things may break and thus to avoid using them in scripts displaying this warning or not could be a configurable setting environment summary install method e g pip interactive script apt get docker msi edge build cli version az version os version shell type e g bash cmd exe bash on windows azure cli | 1 |
35,138 | 30,781,405,970 | IssuesEvent | 2023-07-31 10:15:39 | Altinn/altinn-platform | https://api.github.com/repos/Altinn/altinn-platform | closed | Rollout Grafana/Prometheus apps cluster | solution/apps area/infrastructure feature-complete |
```[tasklist]
### Rollout
- [x] at21
- [x] at2[2-4]
- [x] ttd-tt02
- [x] tt02
- [x] ttd-prod
- [x] prod
``` | 1.0 | Rollout Grafana/Prometheus apps cluster -
```[tasklist]
### Rollout
- [x] at21
- [x] at2[2-4]
- [x] ttd-tt02
- [x] tt02
- [x] ttd-prod
- [x] prod
``` | infrastructure | rollout grafana prometheus apps cluster rollout ttd ttd prod prod | 1 |
71,115 | 23,453,165,146 | IssuesEvent | 2022-08-16 06:15:13 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | closed | [🐛 Bug]: DotNet WebDriver msedge(Ver:103.0.1264.77)Use EdgeOptions DebuggerAddress is 127.0.0.1:50521 Error | C-dotnet I-defect | ### What happened?
OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid
from invalid argument: cannot parse capability: ms:edgeOptions
from invalid argument: unrecognized msedge option: excludeSwitches
在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse)
在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities)
在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities)
在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout)
在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning))
### How can we reproduce the issue?
```shell
var options = new EdgeOptions();
var debugHost = "127.0.0.1";
var debugPort = 50521;
options.DebuggerAddress = $"{debugHost}:{debugPort}";
driver = new EdgeDriver(driverService, options, TimeSpan.FromSeconds(40));
```
### Relevant log output
```shell
OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid
from invalid argument: cannot parse capability: ms:edgeOptions
from invalid argument: unrecognized msedge option: excludeSwitches
在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse)
在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities)
在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities)
在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout)
在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning))
```
### Operating System
Windows 11
### Selenium version
Dotnet 4.3
### What are the browser(s) and version(s) where you see this issue?
msedge 103.0.1264.77
### What are the browser driver(s) and version(s) where you see this issue?
msedgedriver 103.0.1264.77
### Are you using Selenium Grid?
No | 1.0 | [🐛 Bug]: DotNet WebDriver msedge(Ver:103.0.1264.77)Use EdgeOptions DebuggerAddress is 127.0.0.1:50521 Error - ### What happened?
OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid
from invalid argument: cannot parse capability: ms:edgeOptions
from invalid argument: unrecognized msedge option: excludeSwitches
在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse)
在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities)
在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities)
在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout)
在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning))
### How can we reproduce the issue?
```shell
var options = new EdgeOptions();
var debugHost = "127.0.0.1";
var debugPort = 50521;
options.DebuggerAddress = $"{debugHost}:{debugPort}";
driver = new EdgeDriver(driverService, options, TimeSpan.FromSeconds(40));
```
### Relevant log output
```shell
OpenQA.Selenium.WebDriverArgumentException: invalid argument: entry 0 of 'firstMatch' is invalid
from invalid argument: cannot parse capability: ms:edgeOptions
from invalid argument: unrecognized msedge option: excludeSwitches
在 OpenQA.Selenium.WebDriver.UnpackAndThrowOnError(Response errorResponse)
在 OpenQA.Selenium.WebDriver.Execute(String driverCommandToExecute, Dictionary`2 parameters)
在 OpenQA.Selenium.WebDriver.StartSession(ICapabilities desiredCapabilities)
在 OpenQA.Selenium.WebDriver..ctor(ICommandExecutor executor, ICapabilities capabilities)
在 OpenQA.Selenium.Chromium.ChromiumDriver..ctor(ChromiumDriverService service, ChromiumOptions options, TimeSpan commandTimeout)
在 OpenQA.Selenium.Edge.EdgeDriver..ctor(EdgeDriverService service, EdgeOptions options, TimeSpan commandTimeout)
在 XRBrowser.MainForm.CreateEdgeDriver(Boolean& isRunning))
```
### Operating System
Windows 11
### Selenium version
Dotnet 4.3
### What are the browser(s) and version(s) where you see this issue?
msedge 103.0.1264.77
### What are the browser driver(s) and version(s) where you see this issue?
msedgedriver 103.0.1264.77
### Are you using Selenium Grid?
No | non_infrastructure | dotnet webdriver msedge ver )use edgeoptions debuggeraddress is error what happened openqa selenium webdriverargumentexception invalid argument entry of firstmatch is invalid from invalid argument cannot parse capability ms edgeoptions from invalid argument unrecognized msedge option excludeswitches 在 openqa selenium webdriver unpackandthrowonerror response errorresponse 在 openqa selenium webdriver execute string drivercommandtoexecute dictionary parameters 在 openqa selenium webdriver startsession icapabilities desiredcapabilities 在 openqa selenium webdriver ctor icommandexecutor executor icapabilities capabilities 在 openqa selenium chromium chromiumdriver ctor chromiumdriverservice service chromiumoptions options timespan commandtimeout 在 openqa selenium edge edgedriver ctor edgedriverservice service edgeoptions options timespan commandtimeout 在 xrbrowser mainform createedgedriver boolean isrunning ) how can we reproduce the issue shell var options new edgeoptions var debughost var debugport options debuggeraddress debughost debugport driver new edgedriver driverservice options timespan fromseconds relevant log output shell openqa selenium webdriverargumentexception invalid argument entry of firstmatch is invalid from invalid argument cannot parse capability ms edgeoptions from invalid argument unrecognized msedge option excludeswitches 在 openqa selenium webdriver unpackandthrowonerror response errorresponse 在 openqa selenium webdriver execute string drivercommandtoexecute dictionary parameters 在 openqa selenium webdriver startsession icapabilities desiredcapabilities 在 openqa selenium webdriver ctor icommandexecutor executor icapabilities capabilities 在 openqa selenium chromium chromiumdriver ctor chromiumdriverservice service chromiumoptions options timespan commandtimeout 在 openqa selenium edge edgedriver ctor edgedriverservice service edgeoptions options timespan commandtimeout 在 xrbrowser mainform createedgedriver boolean isrunning ) operating system windows selenium version dotnet what are the browser s and version s where you see this issue msedge what are the browser driver s and version s where you see this issue msedgedriver are you using selenium grid no | 0 |
410 | 2,857,052,418 | IssuesEvent | 2015-06-02 17:50:05 | astropy/astroquery | https://api.github.com/repos/astropy/astroquery | opened | Add module for Pan-STARRS dust mapping service | New Service | See http://argonaut.skymaps.info/ for the service. http://argonaut.skymaps.info/usage#API gives a lot of the necessary query code, but it doesn't show use with astropy or anything (and requires doing things like constructing the API calls yourself).
One possible concern - it's not clear how "stable" this site will be because I think it's just a few authors hosting it rather than an "organization". | 1.0 | Add module for Pan-STARRS dust mapping service - See http://argonaut.skymaps.info/ for the service. http://argonaut.skymaps.info/usage#API gives a lot of the necessary query code, but it doesn't show use with astropy or anything (and requires doing things like constructing the API calls yourself).
One possible concern - it's not clear how "stable" this site will be because I think it's just a few authors hosting it rather than an "organization". | non_infrastructure | add module for pan starrs dust mapping service see for the service gives a lot of the necessary query code but it doesn t show use with astropy or anything and requires doing things like constructing the api calls yourself one possible concern it s not clear how stable this site will be because i think it s just a few authors hosting it rather than an organization | 0 |
27,432 | 21,707,919,881 | IssuesEvent | 2022-05-10 11:22:37 | coatrack/coatrack | https://api.github.com/repos/coatrack/coatrack | opened | Integration of CoatRack into FriDa Project | infrastructure | This is an epic issue and therefore probably requires to create sub-issues for each sub-task.
Preparation on CoatRack side:
- [ ] CoatRack should be deployable using an nginx proxy for certificate management. The nginx component is to be added to the `docker-compose.yml`.
- [ ] The CI pipeline shall be adapted to provide an option to deploy CoatRack (with nginx) on a Hetzner VM manually.
Integration of CoatRack and FriDa:
- [ ] Set up a Hetzner VM and deploy CoatRack with nginx.
- [ ] Rent a VM
- [ ] Register domain, e.g. `frida-integration.coatrack.atb-bremen.de`
- [ ] Deploy CoatRack
- [ ] Gateway creation in CoatRack Web Application whose target URL is an instance of the Tempman Project.
- [ ] Context Broker subscription which directs data from the Broker to the CoatRack Gateway.
- [ ] To be discussed: How do we ensure that the connection is secure? Shall we assume the user has a public IP and a certificate to run the Gateway as a usual server? Should we, for reasons of simplicity, deploy the proxy to the same VM as the CoatRack setup? | 1.0 | Integration of CoatRack into FriDa Project - This is an epic issue and therefore probably requires to create sub-issues for each sub-task.
Preparation on CoatRack side:
- [ ] CoatRack should be deployable using an nginx proxy for certificate management. The nginx component is to be added to the `docker-compose.yml`.
- [ ] The CI pipeline shall be adapted to provide an option to deploy CoatRack (with nginx) on a Hetzner VM manually.
Integration of CoatRack and FriDa:
- [ ] Set up a Hetzner VM and deploy CoatRack with nginx.
- [ ] Rent a VM
- [ ] Register domain, e.g. `frida-integration.coatrack.atb-bremen.de`
- [ ] Deploy CoatRack
- [ ] Gateway creation in CoatRack Web Application whose target URL is an instance of the Tempman Project.
- [ ] Context Broker subscription which directs data from the Broker to the CoatRack Gateway.
- [ ] To be discussed: How do we ensure that the connection is secure? Shall we assume the user has a public IP and a certificate to run the Gateway as a usual server? Should we, for reasons of simplicity, deploy the proxy to the same VM as the CoatRack setup? | infrastructure | integration of coatrack into frida project this is an epic issue and therefore probably requires to create sub issues for each sub task preparation on coatrack side coatrack should be deployable using an nginx proxy for certificate management the nginx component is to be added to the docker compose yml the ci pipeline shall be adapted to provide an option to deploy coatrack with nginx on a hetzner vm manually integration of coatrack and frida set up a hetzner vm and deploy coatrack with nginx rent a vm register domain e g frida integration coatrack atb bremen de deploy coatrack gateway creation in coatrack web application whose target url is an instance of the tempman project context broker subscription which directs data from the broker to the coatrack gateway to be discussed how do we ensure that the connection is secure shall we assume the user has a public ip and a certificate to run the gateway as a usual server should we for reasons of simplicity deploy the proxy to the same vm as the coatrack setup | 1 |
24,812 | 17,792,611,312 | IssuesEvent | 2021-08-31 18:00:55 | hackforla/website | https://api.github.com/repos/hackforla/website | opened | Standardize how 'Overview' links open pdf files | role: front end Size: Large Feature: Infrastructure | ### Overview
We need to standardize how the 'Overview' link for each project page acts when a user clicks it. Currently, the 'Overview' links act differently for each project. Clicking the link should open a new tab to the pdf file directly. Once this issue is approved, we will convert each action item into an issue and create a pull request to close the corresponding issue.
### Details
When clicking the 'Overview' link, each project page acts differently. Some open to a google docs page, others open to a github page that hosts the pdf file, and another downloads the pdf file locally to the machine. There needs to be a standard in how the pdf files open. Developers decided during the backend meeting that pdfs should be opened as a separate tab to the file directly, since modern browsers have the ability to view pdfs.
The process for a developer would involve:
- Using the newly created directory for pdf files: [/assets/pdfs](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs)
- Place a copy of a project's pdf overview file into the /assets/pdfs directory
- Changing the 'Overview' link for a project to point to the pdf file living under /assets/pdfs
- Ensure that the links in both the home page and the project's dedicated page open the pdf file properly after the change
### Action Items
- [ ] Change the 'Overview' link for 100 Automations project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Access The Data project to open as a pdf file in the browser instead of opening a google doc
- [ ] Change the 'Overview' link for BallotNav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Civic Opportunity Project project to open as a pdf file in the browser instead of downloading a pdf file
- [ ] Change the 'Overview' link for Civic Tech Index project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Civic Tech Structure project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Engage project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Expunge Assist project to open as a pdf file in the browser instead of opening a google doc
- [ ] Change the 'Overview' link for GreenEarthOS project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for LA TDM Calculator project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Youth Justice Nav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
### Resources/Instructions
[Project Page](https://www.hackforla.org/projects/)
Each individual landing page for affected projects: [100 automations as an example](https://www.hackforla.org/projects/100-automations)
[Location for pdf files](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs)
[Jekyll Static Files](https://jekyllrb.com/docs/static-files/) | 1.0 | Standardize how 'Overview' links open pdf files - ### Overview
We need to standardize how the 'Overview' link for each project page acts when a user clicks it. Currently, the 'Overview' links act differently for each project. Clicking the link should open a new tab to the pdf file directly. Once this issue is approved, we will convert each action item into an issue and create a pull request to close the corresponding issue.
### Details
When clicking the 'Overview' link, each project page acts differently. Some open to a google docs page, others open to a github page that hosts the pdf file, and another downloads the pdf file locally to the machine. There needs to be a standard in how the pdf files open. Developers decided during the backend meeting that pdfs should be opened as a separate tab to the file directly, since modern browsers have the ability to view pdfs.
The process for a developer would involve:
- Using the newly created directory for pdf files: [/assets/pdfs](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs)
- Place a copy of a project's pdf overview file into the /assets/pdfs directory
- Changing the 'Overview' link for a project to point to the pdf file living under /assets/pdfs
- Ensure that the links in both the home page and the project's dedicated page open the pdf file properly after the change
### Action Items
- [ ] Change the 'Overview' link for 100 Automations project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Access The Data project to open as a pdf file in the browser instead of opening a google doc
- [ ] Change the 'Overview' link for BallotNav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Civic Opportunity Project project to open as a pdf file in the browser instead of downloading a pdf file
- [ ] Change the 'Overview' link for Civic Tech Index project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Civic Tech Structure project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Engage project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Expunge Assist project to open as a pdf file in the browser instead of opening a google doc
- [ ] Change the 'Overview' link for GreenEarthOS project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for LA TDM Calculator project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
- [ ] Change the 'Overview' link for Youth Justice Nav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file
### Resources/Instructions
[Project Page](https://www.hackforla.org/projects/)
Each individual landing page for affected projects: [100 automations as an example](https://www.hackforla.org/projects/100-automations)
[Location for pdf files](https://github.com/hackforla/website/tree/gh-pages/assets/pdfs)
[Jekyll Static Files](https://jekyllrb.com/docs/static-files/) | infrastructure | standardize how overview links open pdf files overview we need to standardize how the overview link for each project page acts when a user clicks it currently the overview links act differently for each project clicking the link should open a new tab to the pdf file directly once this issue is approved we will convert each action item into an issue and create a pull request to close the corresponding issue details when clicking the overview link each project page acts differently some open to a google docs page others open to a github page that hosts the pdf file and another downloads the pdf file locally to the machine there needs to be a standard in how the pdf files open developers decided during the backend meeting that pdfs should be opened as a separate tab to the file directly since modern browsers have the ability to view pdfs the process for a developer would involve using the newly created directory for pdf files place a copy of a project s pdf overview file into the assets pdfs directory changing the overview link for a project to point to the pdf file living under assets pdfs ensure that the links in both the home page and the project s dedicated page open the pdf file properly after the change action items change the overview link for automations project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for access the data project to open as a pdf file in the browser instead of opening a google doc change the overview link for ballotnav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for civic opportunity project project to open as a pdf file in the browser instead of downloading a pdf file change the overview link for civic tech index project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for civic tech structure project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for engage project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for expunge assist project to open as a pdf file in the browser instead of opening a google doc change the overview link for greenearthos project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for la tdm calculator project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file change the overview link for youth justice nav project to open as a pdf file in the browser instead of opening a github page that hosts the pdf file resources instructions each individual landing page for affected projects | 1 |
7,724 | 7,059,161,483 | IssuesEvent | 2018-01-04 23:46:42 | dotnet/core-setup | https://api.github.com/repos/dotnet/core-setup | closed | Debian stretch packages? | area-Infrastructure | Are there packages for debian stretch?
The [official debian instructions](https://www.microsoft.com/net/core#linuxdebian) don't work [because of issues with dependencies](https://github.com/dotnet/cli/issues/2018#issuecomment-239301891) and the packages in this repo can't be installed for similar reasons. | 1.0 | Debian stretch packages? - Are there packages for debian stretch?
The [official debian instructions](https://www.microsoft.com/net/core#linuxdebian) don't work [because of issues with dependencies](https://github.com/dotnet/cli/issues/2018#issuecomment-239301891) and the packages in this repo can't be installed for similar reasons. | infrastructure | debian stretch packages are there packages for debian stretch the don t work and the packages in this repo can t be installed for similar reasons | 1 |
53,965 | 13,890,088,965 | IssuesEvent | 2020-10-19 08:48:55 | edwardhu/yarn | https://api.github.com/repos/edwardhu/yarn | opened | CVE-2018-3728 (High) detected in hoek-2.16.3.tgz | security vulnerability | ## CVE-2018-3728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>Path to dependency file: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock</p>
<p>Path to vulnerable library: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock,yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/node_modules/hoek/package.json,yarn/__tests__/fixtures/install/install-bin-links-eslint/node_modules/hoek/package.json</p>
<p>
Dependency Hierarchy:
- karma-2.0.5.tgz (Root Library)
- log4js-2.11.0.tgz
- loggly-1.1.1.tgz
- request-2.75.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/edwardhu/yarn/commit/2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9">2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3728">https://nvd.nist.gov/vuln/detail/CVE-2018-3728</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution: 4.2.1,5.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hoek","packageVersion":"2.16.3","isTransitiveDependency":true,"dependencyTree":"karma:2.0.5;log4js:2.11.0;loggly:1.1.1;request:2.75.0;hawk:3.1.3;hoek:2.16.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.1,5.0.3"}],"vulnerabilityIdentifier":"CVE-2018-3728","vulnerabilityDetails":"hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via \u0027merge\u0027 and \u0027applyToDefaults\u0027 functions, which allows a malicious user to modify the prototype of \"Object\" via __proto__, causing the addition or modification of an existing property that will exist on all objects.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-3728 (High) detected in hoek-2.16.3.tgz - ## CVE-2018-3728 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hoek-2.16.3.tgz</b></p></summary>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>Path to dependency file: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock</p>
<p>Path to vulnerable library: yarn/__tests__/fixtures/outdated/wanted-newer-than-current/yarn.lock,yarn/__tests__/fixtures/install/install-update-auth-no-offline-integrity/offline-mirror/abab-2.0.0/package/node_modules/hoek/package.json,yarn/__tests__/fixtures/install/install-bin-links-eslint/node_modules/hoek/package.json</p>
<p>
Dependency Hierarchy:
- karma-2.0.5.tgz (Root Library)
- log4js-2.11.0.tgz
- loggly-1.1.1.tgz
- request-2.75.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/edwardhu/yarn/commit/2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9">2ee054bb235ddfee52d2edbc5ea95cd8becdb9c9</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3728">https://nvd.nist.gov/vuln/detail/CVE-2018-3728</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution: 4.2.1,5.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hoek","packageVersion":"2.16.3","isTransitiveDependency":true,"dependencyTree":"karma:2.0.5;log4js:2.11.0;loggly:1.1.1;request:2.75.0;hawk:3.1.3;hoek:2.16.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.2.1,5.0.3"}],"vulnerabilityIdentifier":"CVE-2018-3728","vulnerabilityDetails":"hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via \u0027merge\u0027 and \u0027applyToDefaults\u0027 functions, which allows a malicious user to modify the prototype of \"Object\" via __proto__, causing the addition or modification of an existing property that will exist on all objects.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve high detected in hoek tgz cve high severity vulnerability vulnerable library hoek tgz general purpose node utilities library home page a href path to dependency file yarn tests fixtures outdated wanted newer than current yarn lock path to vulnerable library yarn tests fixtures outdated wanted newer than current yarn lock yarn tests fixtures install install update auth no offline integrity offline mirror abab package node modules hoek package json yarn tests fixtures install install bin links eslint node modules hoek package json dependency hierarchy karma tgz root library tgz loggly tgz request tgz hawk tgz x hoek tgz vulnerable library found in head commit a href found in base branch master vulnerability details hoek node module before and x before suffers from a modification of assumed immutable data maid vulnerability via merge and applytodefaults functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails hoek node module before and x before suffers from a modification of assumed immutable data maid vulnerability via and functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects vulnerabilityurl | 0 |
34,649 | 30,250,297,682 | IssuesEvent | 2023-07-06 19:58:07 | tskit-dev/tstrait | https://api.github.com/repos/tskit-dev/tstrait | closed | Add testing | Infrastructure | As this is pure-python I think you could get away with linux-only testing for now. That will remove the need for conda in the setup. | 1.0 | Add testing - As this is pure-python I think you could get away with linux-only testing for now. That will remove the need for conda in the setup. | infrastructure | add testing as this is pure python i think you could get away with linux only testing for now that will remove the need for conda in the setup | 1 |
13,740 | 10,439,871,687 | IssuesEvent | 2019-09-18 07:27:17 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | opened | Fix Robolectric tests on Android Studio 3.5+ | Priority: Essential Status: In implementation Type: Bug Where: Infrastructure | For some reason starting with Android Studio 3.5, all Robolectric tests appear to fail pulling in key generated classpath elements such as protobuf classes. Robolectric seems to be the same, so this appears to be a change in how Android Studio/bundled Gradle handle the classpath when running these tests.
#59 may circumvent this problem by avoiding Gradle, and given the time I've put into investigating this issue that may be a faster solution. | 1.0 | Fix Robolectric tests on Android Studio 3.5+ - For some reason starting with Android Studio 3.5, all Robolectric tests appear to fail pulling in key generated classpath elements such as protobuf classes. Robolectric seems to be the same, so this appears to be a change in how Android Studio/bundled Gradle handle the classpath when running these tests.
#59 may circumvent this problem by avoiding Gradle, and given the time I've put into investigating this issue that may be a faster solution. | infrastructure | fix robolectric tests on android studio for some reason starting with android studio all robolectric tests appear to fail pulling in key generated classpath elements such as protobuf classes robolectric seems to be the same so this appears to be a change in how android studio bundled gradle handle the classpath when running these tests may circumvent this problem by avoiding gradle and given the time i ve put into investigating this issue that may be a faster solution | 1 |
34,491 | 30,024,373,445 | IssuesEvent | 2023-06-27 04:09:18 | dotnet/project-system | https://api.github.com/repos/dotnet/project-system | closed | View Code button missing on Solution Explorer toolbar for WinForms and WPF apps | Area-Infrastructure | ## Minimal Reproduction Scenario
1. Create a Winforms .Net Core application.
2. In solution Explorer, focus on Form1.cs node, then observe the toolbar items at the top of solution explorer window.
## Current Behavior
Missing "View Code" button in Solution Explorer

## Expected Behavior
View Code button is there, like the result of focusing on .Net framework form

## Test Environment
17.7.0 Preview 3.0 [33824.57.main]
## Have you experienced this same issue with .NET Framework designer?
No
## Have you experienced this same issue with previous previews of VisualStudio?
Yes, 17.3.6 have this issue
## More Info
1. .NET Framework project does not repro this issue.
2. This issue occurs in the .NET Core WPF project too. | 1.0 | View Code button missing on Solution Explorer toolbar for WinForms and WPF apps - ## Minimal Reproduction Scenario
1. Create a Winforms .Net Core application.
2. In solution Explorer, focus on Form1.cs node, then observe the toolbar items at the top of solution explorer window.
## Current Behavior
Missing "View Code" button in Solution Explorer

## Expected Behavior
View Code button is there, like the result of focusing on .Net framework form

## Test Environment
17.7.0 Preview 3.0 [33824.57.main]
## Have you experienced this same issue with .NET Framework designer?
No
## Have you experienced this same issue with previous previews of VisualStudio?
Yes, 17.3.6 have this issue
## More Info
1. .NET Framework project does not repro this issue.
2. This issue occurs in the .NET Core WPF project too. | infrastructure | view code button missing on solution explorer toolbar for winforms and wpf apps minimal reproduction scenario create a winforms net core application in solution explorer focus on cs node then observe the toolbar items at the top of solution explorer window current behavior missing view code button in solution explorer expected behavior view code button is there like the result of focusing on net framework form test environment preview have you experienced this same issue with net framework designer no have you experienced this same issue with previous previews of visualstudio yes have this issue more info net framework project does not repro this issue this issue occurs in the net core wpf project too | 1 |
118,271 | 4,733,665,164 | IssuesEvent | 2016-10-19 11:59:38 | cdnjs/cdnjs | https://api.github.com/repos/cdnjs/cdnjs | closed | [Request] Update flux to 3.0.0 | High Priority Library - Request to Add/Update | **Library name:** flux
**Git repository url:** https://github.com/facebook/flux
**npm package url(optional):** https://www.npmjs.com/package/flux
**Files:** https://github.com/facebook/flux/tree/master/dist | 1.0 | [Request] Update flux to 3.0.0 - **Library name:** flux
**Git repository url:** https://github.com/facebook/flux
**npm package url(optional):** https://www.npmjs.com/package/flux
**Files:** https://github.com/facebook/flux/tree/master/dist | non_infrastructure | update flux to library name flux git repository url npm package url optional files | 0 |
18,999 | 13,184,826,328 | IssuesEvent | 2020-08-12 20:10:18 | Kemmey/Kemmey-TeslaWatch-Public | https://api.github.com/repos/Kemmey/Kemmey-TeslaWatch-Public | closed | Reset iWatch | AppStore infrastructure issue | I reset my watch. I can’t get the Tesla for iWatch back. In the App Store it doesn’t let me click on anything (only says “purchased”) and the app is no where to be found. This is likely a issue I need to address with Apple but I realized that as I was typing so I will continue with this avenue. App is great btw. I had the Tesla Remote app forever but it would never succeed in waking my M3. Yours worked like a charm, and quick! | 1.0 | Reset iWatch - I reset my watch. I can’t get the Tesla for iWatch back. In the App Store it doesn’t let me click on anything (only says “purchased”) and the app is no where to be found. This is likely a issue I need to address with Apple but I realized that as I was typing so I will continue with this avenue. App is great btw. I had the Tesla Remote app forever but it would never succeed in waking my M3. Yours worked like a charm, and quick! | infrastructure | reset iwatch i reset my watch i can’t get the tesla for iwatch back in the app store it doesn’t let me click on anything only says “purchased” and the app is no where to be found this is likely a issue i need to address with apple but i realized that as i was typing so i will continue with this avenue app is great btw i had the tesla remote app forever but it would never succeed in waking my yours worked like a charm and quick | 1 |
6,413 | 3,801,383,606 | IssuesEvent | 2016-03-23 22:40:01 | Homebrew/homebrew-science | https://api.github.com/repos/Homebrew/homebrew-science | closed | astral: error: 404 Not Found | Bioinformatics build-error | ```
==> Installing astral from homebrew/science
==> Downloading https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip
curl: (22) The requested URL returned error: 404 Not Found
Error: Failed to download resource "astral"
Download failed: https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip
``` | 1.0 | astral: error: 404 Not Found - ```
==> Installing astral from homebrew/science
==> Downloading https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip
curl: (22) The requested URL returned error: 404 Not Found
Error: Failed to download resource "astral"
Download failed: https://github.com/smirarab/ASTRAL/raw/master/Astral.4.7.8.zip
``` | non_infrastructure | astral error not found installing astral from homebrew science downloading curl the requested url returned error not found error failed to download resource astral download failed | 0 |
226,925 | 17,367,220,323 | IssuesEvent | 2021-07-30 09:00:04 | vlang/v | https://api.github.com/repos/vlang/v | closed | Wrong escaping in HTML docs | Bug Unit: Documentation Unit: vdoc | `<` and `>` is getting replaced/escaped with `<` and `>` in markdown code in the generated HTML docs
Example in the description here:
https://modules.vlang.io/index.html#compare_strings
The the time of writing it looks like this:

| 1.0 | Wrong escaping in HTML docs - `<` and `>` is getting replaced/escaped with `<` and `>` in markdown code in the generated HTML docs
Example in the description here:
https://modules.vlang.io/index.html#compare_strings
The the time of writing it looks like this:

| non_infrastructure | wrong escaping in html docs is getting replaced escaped with lt and gt in markdown code in the generated html docs example in the description here the the time of writing it looks like this | 0 |
23,219 | 15,947,748,919 | IssuesEvent | 2021-04-15 04:17:01 | OpenHistoricalMap/issues | https://api.github.com/repos/OpenHistoricalMap/issues | closed | Add Nominatim to staging and production deploy process | infrastructure | Right now, Nominatim is not part of the deployment process, and we reference the Production database on both Staging and Production, which can lead to some confusion since the staging mMaptiles and Inspector and iD all reference staging data.
So we'll need to bring Nominatim into the deployment fold. This issue might need more notes about tech approach, but I'll leave that to you! | 1.0 | Add Nominatim to staging and production deploy process - Right now, Nominatim is not part of the deployment process, and we reference the Production database on both Staging and Production, which can lead to some confusion since the staging mMaptiles and Inspector and iD all reference staging data.
So we'll need to bring Nominatim into the deployment fold. This issue might need more notes about tech approach, but I'll leave that to you! | infrastructure | add nominatim to staging and production deploy process right now nominatim is not part of the deployment process and we reference the production database on both staging and production which can lead to some confusion since the staging mmaptiles and inspector and id all reference staging data so we ll need to bring nominatim into the deployment fold this issue might need more notes about tech approach but i ll leave that to you | 1 |
99,438 | 11,140,838,655 | IssuesEvent | 2019-12-21 17:29:35 | prezly/prop-types-php | https://api.github.com/repos/prezly/prop-types-php | closed | There is no PHP code in the repo | bug documentation good first issue | The README states that's a complete PHP port of React PropTypes, but there is no PHP code in the repository.
Probably I misunderstood and it's a complete port of _production_ React PropTypes, which don't exist? 🤔 | 1.0 | There is no PHP code in the repo - The README states that's a complete PHP port of React PropTypes, but there is no PHP code in the repository.
Probably I misunderstood and it's a complete port of _production_ React PropTypes, which don't exist? 🤔 | non_infrastructure | there is no php code in the repo the readme states that s a complete php port of react proptypes but there is no php code in the repository probably i misunderstood and it s a complete port of production react proptypes which don t exist 🤔 | 0 |
286,928 | 21,629,871,887 | IssuesEvent | 2022-05-05 08:36:14 | process-analytics/bpmn-visualization-R | https://api.github.com/repos/process-analytics/bpmn-visualization-R | opened | [DOC] Some overlays are missing in the hero image | documentation |

### Notes
**TODO explain how the image was created and how to change it**
| 1.0 | [DOC] Some overlays are missing in the hero image -

### Notes
**TODO explain how the image was created and how to change it**
| non_infrastructure | some overlays are missing in the hero image notes todo explain how the image was created and how to change it | 0 |
15,541 | 11,579,298,779 | IssuesEvent | 2020-02-21 17:37:02 | eventespresso/event-espresso-core | https://api.github.com/repos/eventespresso/event-espresso-core | closed | Complete server-side mutations | EDTR v2 category:models-and-data-infrastructure | The server-side mutations for Datetime, Ticket and Price need to be completed in files inside `core/domain/services/graphql/data/mutations`. Primary fields are already taken care of. Rest of the fields need to be added.
**Note**: Mutation input should properly be sanitized and converted to proper data type. | 1.0 | Complete server-side mutations - The server-side mutations for Datetime, Ticket and Price need to be completed in files inside `core/domain/services/graphql/data/mutations`. Primary fields are already taken care of. Rest of the fields need to be added.
**Note**: Mutation input should properly be sanitized and converted to proper data type. | infrastructure | complete server side mutations the server side mutations for datetime ticket and price need to be completed in files inside core domain services graphql data mutations primary fields are already taken care of rest of the fields need to be added note mutation input should properly be sanitized and converted to proper data type | 1 |
33,060 | 27,183,115,957 | IssuesEvent | 2023-02-18 22:05:11 | surge-synthesizer/surge | https://api.github.com/repos/surge-synthesizer/surge | closed | Display version reporting error | Infrastructure Bug Report | "Surge XT 1.1.2.d3568a9 is installed but the plugin is reporting Version 1.1.0 to the host. In the About dialog, the version is ok" | 1.0 | Display version reporting error - "Surge XT 1.1.2.d3568a9 is installed but the plugin is reporting Version 1.1.0 to the host. In the About dialog, the version is ok" | infrastructure | display version reporting error surge xt is installed but the plugin is reporting version to the host in the about dialog the version is ok | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.