Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
5,789 | 30,660,611,093 | IssuesEvent | 2023-07-25 14:47:12 | jupyter-naas/awesome-notebooks | https://api.github.com/repos/jupyter-naas/awesome-notebooks | closed | Pandas - Apply custom styles on column | templates maintainer | This notebook will show how to apply custom styles on a column of a Pandas DataFrame. It is usefull for data analysis and data visualization.
| True | Pandas - Apply custom styles on column - This notebook will show how to apply custom styles on a column of a Pandas DataFrame. It is usefull for data analysis and data visualization.
| main | pandas apply custom styles on column this notebook will show how to apply custom styles on a column of a pandas dataframe it is usefull for data analysis and data visualization | 1 |
2,263 | 7,961,970,660 | IssuesEvent | 2018-07-13 12:50:27 | cucumber/aruba | https://api.github.com/repos/cucumber/aruba | closed | Add more logging for childprocess | needs feedback by maintainer stale type: new feature | <!-- These sections are meant as guidance for you, to help you give the kind of information we'll need to help with your issue. If a section doesn't seem to fit, just skip it.
In general: Please provide as much information as you can to help us solving your problem -->
## Summary
`ChildProcess` has an API for a logger to troubleshoot issues with this gem. This might be a valuable addition for aruba to write those logs on request as well
<!--- Provide a general summary description of the issue -->
## Expected Behavior
* Aruba has a configuration flag to activate this logger
* Those logs are written to a separate file
* The path to the file is written to console
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
<!--- Feel free to use Given / Then / Then if that helps, but please add some plain-language context too -->
## Context & Motivation
This might help to troubleshoot issues with "ChildProcess" itself.
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
| True | Add more logging for childprocess - <!-- These sections are meant as guidance for you, to help you give the kind of information we'll need to help with your issue. If a section doesn't seem to fit, just skip it.
In general: Please provide as much information as you can to help us solving your problem -->
## Summary
`ChildProcess` has an API for a logger to troubleshoot issues with this gem. This might be a valuable addition for aruba to write those logs on request as well
<!--- Provide a general summary description of the issue -->
## Expected Behavior
* Aruba has a configuration flag to activate this logger
* Those logs are written to a separate file
* The path to the file is written to console
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
<!--- Feel free to use Given / Then / Then if that helps, but please add some plain-language context too -->
## Context & Motivation
This might help to troubleshoot issues with "ChildProcess" itself.
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
| main | add more logging for childprocess these sections are meant as guidance for you to help you give the kind of information we ll need to help with your issue if a section doesn t seem to fit just skip it in general please provide as much information as you can to help us solving your problem summary childprocess has an api for a logger to troubleshoot issues with this gem this might be a valuable addition for aruba to write those logs on request as well expected behavior aruba has a configuration flag to activate this logger those logs are written to a separate file the path to the file is written to console context motivation this might help to troubleshoot issues with childprocess itself | 1 |
5,608 | 28,069,213,189 | IssuesEvent | 2023-03-29 17:44:14 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | opened | Make it possible to skip either frontend or backend CI jobs | enhancement maintainability to be reviewed CI/CD | A lot of pull requests either affect the backend or frontend but currently one can just skip all of CI or none, which is especially annoying when multiple jobs gets queued.
### Proposed solution
We could add two commit message keywords and conditions to frontend respective backend related tests/quality checks:
```
if: ${{ !contains(github.event.head_commit.message, '#frontend-only') }}
```
```
if: ${{ !contains(github.event.head_commit.message, '#backend-only') }}
```
So that contributors can skip jobs that aren't relevant to their contribution. All jobs would still run on merges. | True | Make it possible to skip either frontend or backend CI jobs - A lot of pull requests either affect the backend or frontend but currently one can just skip all of CI or none, which is especially annoying when multiple jobs gets queued.
### Proposed solution
We could add two commit message keywords and conditions to frontend respective backend related tests/quality checks:
```
if: ${{ !contains(github.event.head_commit.message, '#frontend-only') }}
```
```
if: ${{ !contains(github.event.head_commit.message, '#backend-only') }}
```
So that contributors can skip jobs that aren't relevant to their contribution. All jobs would still run on merges. | main | make it possible to skip either frontend or backend ci jobs a lot of pull requests either affect the backend or frontend but currently one can just skip all of ci or none which is especially annoying when multiple jobs gets queued proposed solution we could add two commit message keywords and conditions to frontend respective backend related tests quality checks if contains github event head commit message frontend only if contains github event head commit message backend only so that contributors can skip jobs that aren t relevant to their contribution all jobs would still run on merges | 1 |
554,081 | 16,388,596,207 | IssuesEvent | 2021-05-17 13:37:54 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.fedex.com - site is not usable | browser-firefox-ios os-ios priority-important | <!-- @browser: Firefox iOS 33.1 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/74045 -->
**URL**: https://www.fedex.com/fedextrack/?trknbr=787095647716
**Browser / Version**: Firefox iOS 33.1
**Operating System**: iOS 14.4.2
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
After signing into account, infinite page reload cycle m m tracking date
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.fedex.com - site is not usable - <!-- @browser: Firefox iOS 33.1 -->
<!-- @ua_header: Mozilla/5.0 (iPhone; CPU OS 14_4_2 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/33.1 Mobile/15E148 Safari/605.1.15 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/74045 -->
**URL**: https://www.fedex.com/fedextrack/?trknbr=787095647716
**Browser / Version**: Firefox iOS 33.1
**Operating System**: iOS 14.4.2
**Tested Another Browser**: Yes Safari
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
After signing into account, infinite page reload cycle m m tracking date
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | site is not usable url browser version firefox ios operating system ios tested another browser yes safari problem type site is not usable description page not loading correctly steps to reproduce after signing into account infinite page reload cycle m m tracking date browser configuration none from with ❤️ | 0 |
96,866 | 8,635,029,298 | IssuesEvent | 2018-11-22 19:51:54 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Manual test run on OS X for 0.56.x - Release Hotfix 2 | OS/macOS release-notes/exclude tests | ## Per release specialty tests
- [x] Upgrade to Chromium 70.0.3538.110.([#2200](https://github.com/brave/brave-browser/issues/2200))
### Installer
- [x] Check that installer is close to the size of last release
- [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### Data(Upgrade from previous release)
- [x] Make sure that data from the last version appears in the new version OK
- [x] With data from the last version, verify that
- [x] bookmarks on the bookmark toolbar and bookmark folders can be opened
- [x] cookies are preserved
- [x] installed extensions are retained and work correctly
- [x] opened tabs can be reloaded
- [x] stored passwords are preserved
- [x] unpinned tabs can be pinned
### Widevine
- [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [x] Test that you can stream on Netflix on a fresh profile after installing Widevine
### Geolocation
- [ ] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation shows correct location
- [x] Check that https://browserleaks.com/geo works and shows correct location
- [x] Check that https://html5demos.com/geo/ works but doesn't require an accurate location
### Crash Reporting
- [x] Check that loading `brave://crash` causes the new tab to crash
- [x] Check that `brave://crashes` lists all the crashes and includes both Crash Report ID & Local Crash ID
- [x] Verify the crash ID matches the report on brave stats
### Bravery settings
- [x] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/
- [x] Verify that toggling `Ads and trackers blocked` works as expected
- [x] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [x] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly
- [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
### Fingerprint Tests
- [ ] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00`
- [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on
- [x] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/
- [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on
### Content tests
- [x] Open a page with an input control and type some misspellings on a textbox, make sure they are underlined
- [x] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text
- [x] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run)
### Session storage
- [x] Temporarily move away your browser profile and test that a new profile is created when browser is launched
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [x] Test that windows and tabs restore when closed, including active tab
- [x] Ensure that the tabs in the above session are being lazy loaded when the session is restored
## Chromium upgrade tests
#### Adblock
- [x] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/`
#### Components
- [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
| 1.0 | Manual test run on OS X for 0.56.x - Release Hotfix 2 - ## Per release specialty tests
- [x] Upgrade to Chromium 70.0.3538.110.([#2200](https://github.com/brave/brave-browser/issues/2200))
### Installer
- [x] Check that installer is close to the size of last release
- [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### Data(Upgrade from previous release)
- [x] Make sure that data from the last version appears in the new version OK
- [x] With data from the last version, verify that
- [x] bookmarks on the bookmark toolbar and bookmark folders can be opened
- [x] cookies are preserved
- [x] installed extensions are retained and work correctly
- [x] opened tabs can be reloaded
- [x] stored passwords are preserved
- [x] unpinned tabs can be pinned
### Widevine
- [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [x] Test that you can stream on Netflix on a fresh profile after installing Widevine
### Geolocation
- [ ] Check that https://developer.mozilla.org/en-US/docs/Web/API/Geolocation/Using_geolocation shows correct location
- [x] Check that https://browserleaks.com/geo works and shows correct location
- [x] Check that https://html5demos.com/geo/ works but doesn't require an accurate location
### Crash Reporting
- [x] Check that loading `brave://crash` causes the new tab to crash
- [x] Check that `brave://crashes` lists all the crashes and includes both Crash Report ID & Local Crash ID
- [x] Verify the crash ID matches the report on brave stats
### Bravery settings
- [x] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/
- [x] Verify that toggling `Ads and trackers blocked` works as expected
- [x] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [x] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly
- [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
### Fingerprint Tests
- [ ] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00`
- [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on
- [x] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/
- [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on
### Content tests
- [x] Open a page with an input control and type some misspellings on a textbox, make sure they are underlined
- [x] Make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text
- [x] Test that https://mixed-script.badssl.com/ shows up as grey not red (no mixed content scripts are run)
### Session storage
- [x] Temporarily move away your browser profile and test that a new profile is created when browser is launched
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [x] Test that windows and tabs restore when closed, including active tab
- [x] Ensure that the tabs in the above session are being lazy loaded when the session is restored
## Chromium upgrade tests
#### Adblock
- [x] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/`
#### Components
- [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
| non_main | manual test run on os x for x release hotfix per release specialty tests upgrade to chromium installer check that installer is close to the size of last release check signature if os run spctl assess verbose applications brave browser beta app and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window data upgrade from previous release make sure that data from the last version appears in the new version ok with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved unpinned tabs can be pinned widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine geolocation check that shows correct location check that works and shows correct location check that works but doesn t require an accurate location crash reporting check that loading brave crash causes the new tab to crash check that brave crashes lists all the crashes and includes both crash report id local crash id verify the crash id matches the report on brave stats bravery settings verify that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to verify that toggling ads and trackers blocked works as expected visit verify that safe browsing via our proxy works for all the listed items visit and then turn on script blocking page should not load allow it from the script blocking ui in the url bar and it should load the page correctly test that party storage results are blank at when party cookies are blocked and not blank when party cookies are unblocked fingerprint tests visit ensure blocked items are listed in shields result window should show got canvas fingerprint and got webgl fingerprint test that audio fingerprint is blocked at only when block all fingerprinting protection is on test that brave browser isn t detected on test that doesn t leak ip address when block all fingerprinting protection is on content tests open a page with an input control and type some misspellings on a textbox make sure they are underlined make sure that right clicking on a word with suggestions gives a suggestion and that clicking on the suggestion replaces the text test that shows up as grey not red no mixed content scripts are run session storage temporarily move away your browser profile and test that a new profile is created when browser is launched macos library application support bravesoftware windows userprofile appdata local bravesoftware linux ubuntu config bravesoftware test that windows and tabs restore when closed including active tab ensure that the tabs in the above session are being lazy loaded when the session is restored chromium upgrade tests adblock verify referrer blocking works properly for tld visit and verify adblock works properly similar to components delete adblock folder from browser profile and restart browser visit brave components and verify brave ad block updater downloads and update the component repeat for all brave components | 0 |
874 | 4,540,095,870 | IssuesEvent | 2016-09-09 13:37:41 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | eos eapi failed commands | affects_2.1 bug_report networking waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
networking/eos_command
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
config file = /home/admin-0/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
-->
##### OS / ENVIRONMENT
<!---
N/A
-->
##### SUMMARY
<!--- Explain the problem briefly -->
Running the eos_command 'show version' using the eapi transport works fine. When running the command 'show running-configuration section Et1' it returns a failure. Both work fine when using cli as the transport.
Switch is an Arista 7150S running 4.16.7M
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
PLAYBOOK: test_arista_command.yml **********************************************
1 plays in test_arista_command.yml
PLAY [arista_test] *************************************************************
TASK [setup] *******************************************************************
<10.24.1.14> ESTABLISH LOCAL CONNECTION FOR USER: reynolds <10.24.1.14> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python && sleep 0'
<10.24.1.13> ESTABLISH LOCAL CONNECTION FOR USER: reynolds <10.24.1.13> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python && sleep 0'
ok: [myswitch]
TASK [test_arista_command : test command] **************************************
task path: /home/reynolds/wc/cfg/ansible/roles/test_arista_command/tasks/main.yml:1
<10.24.1.13> ESTABLISH LOCAL CONNECTION FOR USER: reynolds <10.24.1.13> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python && sleep 0'
fatal: [myswitch]: FAILED! => {"changed": false, "code": 1003, "commands": ["show running-config section Et1"], "data": [{}, {"errors": ["Command cannot be used over the API at this time. To see ASCII output, set format='text' in your request"]}], "failed": true, "invocation": {"module_args": {"auth_pass": null, "authorize": true, "commands": ["show running-config section Et1"], "host": "10.24.1.13", "interval": 1, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": null, "retries": 10, "ssh_keyfile": null, "transport": "eapi", "url_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "url_username": "ansible", "use_ssl": false, "username": "ansible", "waitfor": null}, "module_name": "eos_command"}, "message": "CLI command 2 of 2 'show running-config section Et1' failed: unconverted command", "msg": "json-rpc error"}
```
| True | eos eapi failed commands - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
networking/eos_command
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
config file = /home/admin-0/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
-->
##### OS / ENVIRONMENT
<!---
N/A
-->
##### SUMMARY
<!--- Explain the problem briefly -->
Running the eos_command 'show version' using the eapi transport works fine. When running the command 'show running-configuration section Et1' it returns a failure. Both work fine when using cli as the transport.
Switch is an Arista 7150S running 4.16.7M
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
PLAYBOOK: test_arista_command.yml **********************************************
1 plays in test_arista_command.yml
PLAY [arista_test] *************************************************************
TASK [setup] *******************************************************************
<10.24.1.14> ESTABLISH LOCAL CONNECTION FOR USER: reynolds <10.24.1.14> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python && sleep 0'
<10.24.1.13> ESTABLISH LOCAL CONNECTION FOR USER: reynolds <10.24.1.13> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python && sleep 0'
ok: [myswitch]
TASK [test_arista_command : test command] **************************************
task path: /home/reynolds/wc/cfg/ansible/roles/test_arista_command/tasks/main.yml:1
<10.24.1.13> ESTABLISH LOCAL CONNECTION FOR USER: reynolds <10.24.1.13> EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python && sleep 0'
fatal: [myswitch]: FAILED! => {"changed": false, "code": 1003, "commands": ["show running-config section Et1"], "data": [{}, {"errors": ["Command cannot be used over the API at this time. To see ASCII output, set format='text' in your request"]}], "failed": true, "invocation": {"module_args": {"auth_pass": null, "authorize": true, "commands": ["show running-config section Et1"], "host": "10.24.1.13", "interval": 1, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": null, "retries": 10, "ssh_keyfile": null, "transport": "eapi", "url_password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "url_username": "ansible", "use_ssl": false, "username": "ansible", "waitfor": null}, "module_name": "eos_command"}, "message": "CLI command 2 of 2 'show running-config section Et1' failed: unconverted command", "msg": "json-rpc error"}
```
| main | eos eapi failed commands issue type bug report component name networking eos command ansible version ansible config file home admin ansible ansible cfg configured module search path default w o overrides configuration os environment n a summary running the eos command show version using the eapi transport works fine when running the command show running configuration section it returns a failure both work fine when using cli as the transport switch is an arista running steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used expected results actual results playbook test arista command yml plays in test arista command yml play task establish local connection for user reynolds exec bin sh c lang c lc all c lc messages c usr bin python sleep establish local connection for user reynolds exec bin sh c lang c lc all c lc messages c usr bin python sleep ok task task path home reynolds wc cfg ansible roles test arista command tasks main yml establish local connection for user reynolds exec bin sh c lang c lc all c lc messages c usr bin python sleep fatal failed changed false code commands data failed true invocation module args auth pass null authorize true commands host interval password value specified in no log parameter port null provider null retries ssh keyfile null transport eapi url password value specified in no log parameter url username ansible use ssl false username ansible waitfor null module name eos command message cli command of show running config section failed unconverted command msg json rpc error | 1 |
379,549 | 26,375,083,876 | IssuesEvent | 2023-01-12 01:16:56 | criblio/appscope | https://api.github.com/repos/criblio/appscope | opened | 1.2.2 docs tweaks | documentation | ### Steps To Reproduce
_No response_
### Environment
```markdown
- AppScope:
- OS:
- Architecture:
- Kernel:
```
### Requested priority
None
### Relevant log output
_No response_ | 1.0 | 1.2.2 docs tweaks - ### Steps To Reproduce
_No response_
### Environment
```markdown
- AppScope:
- OS:
- Architecture:
- Kernel:
```
### Requested priority
None
### Relevant log output
_No response_ | non_main | docs tweaks steps to reproduce no response environment markdown appscope os architecture kernel requested priority none relevant log output no response | 0 |
52,043 | 12,842,776,587 | IssuesEvent | 2020-07-08 02:55:28 | inspireui/support | https://api.github.com/repos/inspireui/support | closed | pod install faild | FluxStore ios-build-fail |
** Product = Fluxstore Pro - Flutter
** version = 1.7.5
** Testing Device/Simulator = problem while building app
I can't install the pod file. I already attach screenshot app command.
error::===
[!] Error installing FBAudienceNetwork
[!] /usr/bin/curl -f -L -o /var/folders/13/0hdhg6mx3w35k1893tjgb2mr0000gn/T/d20200705-1271-xeiny3/file.zip https://developers.facebook.com/resources/FBAudienceNetwork-5.8.0.zip --create-dirs --netrc-optional --retry 2
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- 0:00:02 --:--:-- 0
Warning: Transient problem: HTTP error Will retry in 1 seconds. 2 retries
Warning: left.
0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0
Warning: Transient problem: HTTP error Will retry in 2 seconds. 1 retries
Warning: left.
0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0
curl: (22) The requested URL returned error: 500
[!] Automatically assigning platform `iOS` with version `10.3` on target `Runner` because no platform was specified. Please specify a platform for this target in your Podfile. See `https://guides.cocoapods.org/syntax/podfile.html#platform`.
[!] Automatically assigning platform `iOS` with version `10.3` on target `OneSignalNotificationServiceExtension` because no platform was specified. Please specify a platform for this target in your Podfile. See `https://guides.cocoapods.org/syntax/podfile.html#platform`.
<img width="1440" alt="Screen Shot 2020-07-05 at 7 27 11 PM" src="https://user-images.githubusercontent.com/17203863/86536076-d6e7c580-bef5-11ea-8b81-2b98323212d2.png">
_**
| 1.0 | pod install faild -
** Product = Fluxstore Pro - Flutter
** version = 1.7.5
** Testing Device/Simulator = problem while building app
I can't install the pod file. I already attach screenshot app command.
error::===
[!] Error installing FBAudienceNetwork
[!] /usr/bin/curl -f -L -o /var/folders/13/0hdhg6mx3w35k1893tjgb2mr0000gn/T/d20200705-1271-xeiny3/file.zip https://developers.facebook.com/resources/FBAudienceNetwork-5.8.0.zip --create-dirs --netrc-optional --retry 2
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
0 0 0 0 0 0 0 0 --:--:-- 0:00:02 --:--:-- 0
Warning: Transient problem: HTTP error Will retry in 1 seconds. 2 retries
Warning: left.
0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0
Warning: Transient problem: HTTP error Will retry in 2 seconds. 1 retries
Warning: left.
0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0
curl: (22) The requested URL returned error: 500
[!] Automatically assigning platform `iOS` with version `10.3` on target `Runner` because no platform was specified. Please specify a platform for this target in your Podfile. See `https://guides.cocoapods.org/syntax/podfile.html#platform`.
[!] Automatically assigning platform `iOS` with version `10.3` on target `OneSignalNotificationServiceExtension` because no platform was specified. Please specify a platform for this target in your Podfile. See `https://guides.cocoapods.org/syntax/podfile.html#platform`.
<img width="1440" alt="Screen Shot 2020-07-05 at 7 27 11 PM" src="https://user-images.githubusercontent.com/17203863/86536076-d6e7c580-bef5-11ea-8b81-2b98323212d2.png">
_**
| non_main | pod install faild product fluxstore pro flutter version testing device simulator problem while building app i can t install the pod file i already attach screenshot app command error error installing fbaudiencenetwork usr bin curl f l o var folders t file zip create dirs netrc optional retry total received xferd average speed time time time current dload upload total spent left speed warning transient problem http error will retry in seconds retries warning left warning transient problem http error will retry in seconds retries warning left curl the requested url returned error automatically assigning platform ios with version on target runner because no platform was specified please specify a platform for this target in your podfile see automatically assigning platform ios with version on target onesignalnotificationserviceextension because no platform was specified please specify a platform for this target in your podfile see img width alt screen shot at pm src | 0 |
194,894 | 22,281,547,421 | IssuesEvent | 2022-06-11 01:01:07 | temporalio/sdk-go | https://api.github.com/repos/temporalio/sdk-go | closed | github.com/stretchr/testify-v1.7.0: 1 vulnerabilities (highest severity is: 7.5) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/stretchr/testify-v1.7.0</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/3a2b86ebed54b2f01acfa03635867e89913c3bd4">3a2b86ebed54b2f01acfa03635867e89913c3bd4</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-28948](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-28948) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/go-yaml/yaml-496545a6307b2a7d7a710fd516e5e16e8ab62dbc | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-28948</summary>
### Vulnerable Library - <b>github.com/go-yaml/yaml-496545a6307b2a7d7a710fd516e5e16e8ab62dbc</b></p>
<p>YAML support for the Go language.</p>
<p>
Dependency Hierarchy:
- github.com/stretchr/testify-v1.7.0 (Root Library)
- :x: **github.com/go-yaml/yaml-496545a6307b2a7d7a710fd516e5e16e8ab62dbc** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/3a2b86ebed54b2f01acfa03635867e89913c3bd4">3a2b86ebed54b2f01acfa03635867e89913c3bd4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue in the Unmarshal function in Go-Yaml v3 causes the program to crash when attempting to deserialize invalid input.
<p>Publish Date: 2022-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-28948>CVE-2022-28948</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-hp87-p4gw-j4gq">https://github.com/advisories/GHSA-hp87-p4gw-j4gq</a></p>
<p>Release Date: 2022-05-19</p>
<p>Fix Resolution: 3.0.0</p>
</p>
<p></p>
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/go-yaml/yaml","packageVersion":"496545a6307b2a7d7a710fd516e5e16e8ab62dbc","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/stretchr/testify:v1.7.0;github.com/go-yaml/yaml:496545a6307b2a7d7a710fd516e5e16e8ab62dbc","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.0.0","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-28948","vulnerabilityDetails":"An issue in the Unmarshal function in Go-Yaml v3 causes the program to crash when attempting to deserialize invalid input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-28948","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | True | github.com/stretchr/testify-v1.7.0: 1 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/stretchr/testify-v1.7.0</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/3a2b86ebed54b2f01acfa03635867e89913c3bd4">3a2b86ebed54b2f01acfa03635867e89913c3bd4</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-28948](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-28948) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | github.com/go-yaml/yaml-496545a6307b2a7d7a710fd516e5e16e8ab62dbc | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-28948</summary>
### Vulnerable Library - <b>github.com/go-yaml/yaml-496545a6307b2a7d7a710fd516e5e16e8ab62dbc</b></p>
<p>YAML support for the Go language.</p>
<p>
Dependency Hierarchy:
- github.com/stretchr/testify-v1.7.0 (Root Library)
- :x: **github.com/go-yaml/yaml-496545a6307b2a7d7a710fd516e5e16e8ab62dbc** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/temporalio/sdk-go/commit/3a2b86ebed54b2f01acfa03635867e89913c3bd4">3a2b86ebed54b2f01acfa03635867e89913c3bd4</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue in the Unmarshal function in Go-Yaml v3 causes the program to crash when attempting to deserialize invalid input.
<p>Publish Date: 2022-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-28948>CVE-2022-28948</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-hp87-p4gw-j4gq">https://github.com/advisories/GHSA-hp87-p4gw-j4gq</a></p>
<p>Release Date: 2022-05-19</p>
<p>Fix Resolution: 3.0.0</p>
</p>
<p></p>
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/go-yaml/yaml","packageVersion":"496545a6307b2a7d7a710fd516e5e16e8ab62dbc","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"github.com/stretchr/testify:v1.7.0;github.com/go-yaml/yaml:496545a6307b2a7d7a710fd516e5e16e8ab62dbc","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.0.0","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2022-28948","vulnerabilityDetails":"An issue in the Unmarshal function in Go-Yaml v3 causes the program to crash when attempting to deserialize invalid input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-28948","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}]</REMEDIATE> --> | non_main | github com stretchr testify vulnerabilities highest severity is autoclosed vulnerable library github com stretchr testify found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high github com go yaml yaml transitive n a details cve vulnerable library github com go yaml yaml yaml support for the go language dependency hierarchy github com stretchr testify root library x github com go yaml yaml vulnerable library found in head commit a href found in base branch master vulnerability details an issue in the unmarshal function in go yaml causes the program to crash when attempting to deserialize invalid input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution istransitivedependency true dependencytree github com stretchr testify github com go yaml yaml isminimumfixversionavailable true minimumfixversion isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails an issue in the unmarshal function in go yaml causes the program to crash when attempting to deserialize invalid input vulnerabilityurl | 0 |
538 | 3,952,606,356 | IssuesEvent | 2016-04-29 09:35:39 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | opened | Holiday: Give answers for Mother's Day, Father's Day, etc. | Maintainer Input Requested | Following a [suggestion on Twitter](https://twitter.com/daytonlowell/status/725342856852824066), it would be helpful if this triggered on things such as `when is mothers day`. This changes depending on country, however, so would probably need to incorporate locale detection as well.
------
IA Page: http://duck.co/ia/view/holiday
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @sekhavati | True | Holiday: Give answers for Mother's Day, Father's Day, etc. - Following a [suggestion on Twitter](https://twitter.com/daytonlowell/status/725342856852824066), it would be helpful if this triggered on things such as `when is mothers day`. This changes depending on country, however, so would probably need to incorporate locale detection as well.
------
IA Page: http://duck.co/ia/view/holiday
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @sekhavati | main | holiday give answers for mother s day father s day etc following a it would be helpful if this triggered on things such as when is mothers day this changes depending on country however so would probably need to incorporate locale detection as well ia page sekhavati | 1 |
2,719 | 9,595,364,911 | IssuesEvent | 2019-05-09 15:53:28 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | opened | Do not ignore null arguments | Area: analyzer Area: maintainability feasability unclear feature | If there is an method argument provided, and the argument is checked for null and in such case simply a return is done, then there might be an actual bug hidden underneath.
So such check would silently ignore and cloak that bug.
Therefore we need to check for that and report it as an issue.
Code example:
```C#
public void DoSomething(object o)
{
if (o is null) return;
}
``` | True | Do not ignore null arguments - If there is an method argument provided, and the argument is checked for null and in such case simply a return is done, then there might be an actual bug hidden underneath.
So such check would silently ignore and cloak that bug.
Therefore we need to check for that and report it as an issue.
Code example:
```C#
public void DoSomething(object o)
{
if (o is null) return;
}
``` | main | do not ignore null arguments if there is an method argument provided and the argument is checked for null and in such case simply a return is done then there might be an actual bug hidden underneath so such check would silently ignore and cloak that bug therefore we need to check for that and report it as an issue code example c public void dosomething object o if o is null return | 1 |
3,776 | 15,882,302,571 | IssuesEvent | 2021-04-09 15:51:10 | sympy/sympy | https://api.github.com/repos/sympy/sympy | closed | Failing Master build due to Deprecations in NumPy 1.20 | GitHub Actions Maintainability | Looking at the traceback in the failing Travis build in master, there are a lot of deprecation warnings that are causing the build to fail.
It looks like NumPy has updated their data type aliases and deprecated some of the existing ones like np.int, np.complex, etc. (numpy/numpy#14882) in their latest [1.20 release](https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations). Just looking at the no of failures, it looks like we have a lot of instances of these which we need to update
| True | Failing Master build due to Deprecations in NumPy 1.20 - Looking at the traceback in the failing Travis build in master, there are a lot of deprecation warnings that are causing the build to fail.
It looks like NumPy has updated their data type aliases and deprecated some of the existing ones like np.int, np.complex, etc. (numpy/numpy#14882) in their latest [1.20 release](https://numpy.org/devdocs/release/1.20.0-notes.html#deprecations). Just looking at the no of failures, it looks like we have a lot of instances of these which we need to update
| main | failing master build due to deprecations in numpy looking at the traceback in the failing travis build in master there are a lot of deprecation warnings that are causing the build to fail it looks like numpy has updated their data type aliases and deprecated some of the existing ones like np int np complex etc numpy numpy in their latest just looking at the no of failures it looks like we have a lot of instances of these which we need to update | 1 |
4,555 | 23,725,983,160 | IssuesEvent | 2022-08-30 19:39:22 | rustsec/advisory-db | https://api.github.com/repos/rustsec/advisory-db | closed | `dotenv` crate is implicitly unmaintained | Unmaintained | As of May 21st, 2022, https://github.com/dotenv-rs/dotenv 's latest version is 0.15.0, which was published on October 22nd, 2019. And the latest commit is [3c1a77bc95821777e5ceb996c5e0b082f2a3ea38](https://github.com/dotenv-rs/dotenv/commit/3c1a77bc95821777e5ceb996c5e0b082f2a3ea38), which was pushed on Jun 27th, 2020.
On Dec 24th, 2021, someone asked the project status on [Current maintenance state · Issue #74 · dotenv-rs/dotenv](https://github.com/dotenv-rs/dotenv/issues/74) but there's no response from the maintainers.
I'm not sure how long "prolonged period" refers to, but this crate is a candidate for an "unmaintained" crate, I think. At least we should monitor how things are going there. | True | `dotenv` crate is implicitly unmaintained - As of May 21st, 2022, https://github.com/dotenv-rs/dotenv 's latest version is 0.15.0, which was published on October 22nd, 2019. And the latest commit is [3c1a77bc95821777e5ceb996c5e0b082f2a3ea38](https://github.com/dotenv-rs/dotenv/commit/3c1a77bc95821777e5ceb996c5e0b082f2a3ea38), which was pushed on Jun 27th, 2020.
On Dec 24th, 2021, someone asked the project status on [Current maintenance state · Issue #74 · dotenv-rs/dotenv](https://github.com/dotenv-rs/dotenv/issues/74) but there's no response from the maintainers.
I'm not sure how long "prolonged period" refers to, but this crate is a candidate for an "unmaintained" crate, I think. At least we should monitor how things are going there. | main | dotenv crate is implicitly unmaintained as of may s latest version is which was published on october and the latest commit is which was pushed on jun on dec someone asked the project status on but there s no response from the maintainers i m not sure how long prolonged period refers to but this crate is a candidate for an unmaintained crate i think at least we should monitor how things are going there | 1 |
51,094 | 6,147,329,942 | IssuesEvent | 2017-06-27 15:31:01 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Win32Api SetEnvironmentVariableW - Disabled tests | area-System.Runtime.Extensions disabled-test | Re-Enable all tests which are disabled on UAP/UAPAOT with reason substring "SetEnvironmentVariableW" as soon as Win32 api gets available.
https://github.com/dotnet/corefx/blob/master/src/System.Runtime.Extensions/tests/System/Environment.SetEnvironmentVariable.cs
https://github.com/dotnet/corefx/blob/master/src/System.Runtime.Extensions/tests/System/Environment.GetEnvironmentVariable.cs
https://github.com/dotnet/corefx/blob/master/src/System.Runtime.Extensions/tests/System/Environment.ExpandEnvironmentVariables.cs
| 1.0 | Win32Api SetEnvironmentVariableW - Disabled tests - Re-Enable all tests which are disabled on UAP/UAPAOT with reason substring "SetEnvironmentVariableW" as soon as Win32 api gets available.
https://github.com/dotnet/corefx/blob/master/src/System.Runtime.Extensions/tests/System/Environment.SetEnvironmentVariable.cs
https://github.com/dotnet/corefx/blob/master/src/System.Runtime.Extensions/tests/System/Environment.GetEnvironmentVariable.cs
https://github.com/dotnet/corefx/blob/master/src/System.Runtime.Extensions/tests/System/Environment.ExpandEnvironmentVariables.cs
| non_main | setenvironmentvariablew disabled tests re enable all tests which are disabled on uap uapaot with reason substring setenvironmentvariablew as soon as api gets available | 0 |
322,114 | 9,813,139,638 | IssuesEvent | 2019-06-13 07:12:09 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.esic.in - desktop site instead of mobile site | browser-firefox engine-gecko priority-normal type-connection-error-unknown | <!-- @browser: Firefox 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:67.0) Gecko/20100101 Firefox/67.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.esic.in/
**Browser / Version**: Firefox 67.0
**Operating System**: Windows 10
**Tested Another Browser**: Unknown
**Problem type**: Desktop site instead of mobile site
**Description**: its not open
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/6/86aea0fb-b235-4547-a061-791119c4ef98.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190529130856</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: release</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.esic.in - desktop site instead of mobile site - <!-- @browser: Firefox 67.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:67.0) Gecko/20100101 Firefox/67.0 -->
<!-- @reported_with: desktop-reporter -->
**URL**: https://www.esic.in/
**Browser / Version**: Firefox 67.0
**Operating System**: Windows 10
**Tested Another Browser**: Unknown
**Problem type**: Desktop site instead of mobile site
**Description**: its not open
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/6/86aea0fb-b235-4547-a061-791119c4ef98.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190529130856</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: release</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | desktop site instead of mobile site url browser version firefox operating system windows tested another browser unknown problem type desktop site instead of mobile site description its not open steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel release from with ❤️ | 0 |
2,895 | 10,319,654,184 | IssuesEvent | 2019-08-30 18:08:22 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | Nagios module and Application to join backdrop-contrib | Maintainer application Port in progress | I ported the Nagios module (https://github.com/MegaphoneJon/nagios/tree/1.x-1.x) to Backdrop. I only had a single site running it, but I shared it with Palante Tech, who confirms it works on their Backdrop sites as well. I'd like to submit it to the contrib repo. | True | Nagios module and Application to join backdrop-contrib - I ported the Nagios module (https://github.com/MegaphoneJon/nagios/tree/1.x-1.x) to Backdrop. I only had a single site running it, but I shared it with Palante Tech, who confirms it works on their Backdrop sites as well. I'd like to submit it to the contrib repo. | main | nagios module and application to join backdrop contrib i ported the nagios module to backdrop i only had a single site running it but i shared it with palante tech who confirms it works on their backdrop sites as well i d like to submit it to the contrib repo | 1 |
1,276 | 5,399,957,917 | IssuesEvent | 2017-02-27 20:47:43 | canadainc/sunnah10 | https://api.github.com/repos/canadainc/sunnah10 | closed | Implement RSS feed generator | enhancement invalid logic maintainability ui usability | Allow generating JSON files that can be used in the various BB10 apps for importing.
| True | Implement RSS feed generator - Allow generating JSON files that can be used in the various BB10 apps for importing.
| main | implement rss feed generator allow generating json files that can be used in the various apps for importing | 1 |
198,054 | 14,959,865,397 | IssuesEvent | 2021-01-27 04:19:38 | nasa/osal | https://api.github.com/repos/nasa/osal | closed | mqueue test program | enhancement unit-test | OSAL should include a simple mqueue test program to validate that the user has the correct settings and permissions to create/open/close/delete mqueues. Often users stumble on mqueue configuration and it is more difficult to diagnose when it's wrapped in the entirety of OSAL/cFS. | 1.0 | mqueue test program - OSAL should include a simple mqueue test program to validate that the user has the correct settings and permissions to create/open/close/delete mqueues. Often users stumble on mqueue configuration and it is more difficult to diagnose when it's wrapped in the entirety of OSAL/cFS. | non_main | mqueue test program osal should include a simple mqueue test program to validate that the user has the correct settings and permissions to create open close delete mqueues often users stumble on mqueue configuration and it is more difficult to diagnose when it s wrapped in the entirety of osal cfs | 0 |
343,665 | 30,682,389,137 | IssuesEvent | 2023-07-26 10:00:22 | iho-ohi/S-101_Portrayal-Catalogue | https://api.github.com/repos/iho-ohi/S-101_Portrayal-Catalogue | closed | New symbol for Berth features with categoryOfCargo = 7 (Dangerous or Hazardous) - ENCWG7-5.3_2022 [PSWG #112] | enhancement test PC 1.1.0 | SPEC at: https://github.com/S-101-Portrayal-subWG/Working-Documents/issues/112#issuecomment-1381193928 | 1.0 | New symbol for Berth features with categoryOfCargo = 7 (Dangerous or Hazardous) - ENCWG7-5.3_2022 [PSWG #112] - SPEC at: https://github.com/S-101-Portrayal-subWG/Working-Documents/issues/112#issuecomment-1381193928 | non_main | new symbol for berth features with categoryofcargo dangerous or hazardous spec at | 0 |
2,552 | 8,687,417,498 | IssuesEvent | 2018-12-03 13:42:50 | pbrisbin/bugsnag-haskell | https://api.github.com/repos/pbrisbin/bugsnag-haskell | opened | Integration testing | enhancement help wanted maintainability | We lack any good way to test things end-to-end: e.g. to assert on the result of some `notify(With)` call given some scenario. This makes it hard to get a regression test on #31 for example, which led to us not really fixing it.
I'm thinking about something like sticking an optional `IORef` in Settings that, when present, gets the Events appended to it instead of actually send to Bugsnag. This could even be useful to end-users who want to test Bugsnag-related paths in their own applications. | True | Integration testing - We lack any good way to test things end-to-end: e.g. to assert on the result of some `notify(With)` call given some scenario. This makes it hard to get a regression test on #31 for example, which led to us not really fixing it.
I'm thinking about something like sticking an optional `IORef` in Settings that, when present, gets the Events appended to it instead of actually send to Bugsnag. This could even be useful to end-users who want to test Bugsnag-related paths in their own applications. | main | integration testing we lack any good way to test things end to end e g to assert on the result of some notify with call given some scenario this makes it hard to get a regression test on for example which led to us not really fixing it i m thinking about something like sticking an optional ioref in settings that when present gets the events appended to it instead of actually send to bugsnag this could even be useful to end users who want to test bugsnag related paths in their own applications | 1 |
991 | 3,268,128,405 | IssuesEvent | 2015-10-23 09:31:19 | peter992233/CSCI342Project | https://api.github.com/repos/peter992233/CSCI342Project | opened | Add Proper Level System (With Difficulty) | Core Requirement | Extend on the level system by adding scaling difficulty such as shooting rate, speed of enemies and a multiplied score for each level they pass | 1.0 | Add Proper Level System (With Difficulty) - Extend on the level system by adding scaling difficulty such as shooting rate, speed of enemies and a multiplied score for each level they pass | non_main | add proper level system with difficulty extend on the level system by adding scaling difficulty such as shooting rate speed of enemies and a multiplied score for each level they pass | 0 |
388,019 | 26,748,978,092 | IssuesEvent | 2023-01-30 18:01:26 | WordPress/Advanced-administration-handbook | https://api.github.com/repos/WordPress/Advanced-administration-handbook | opened | Update page: Upgrading WordPress | documentation enhancement help wanted | File: [upgrade/upgrading.md](https://github.com/WordPress/Advanced-administration-handbook/blob/main/upgrade/upgrading.md)
This page needs a general review and update.
Needs to have some different parts. One, the simple update via the Admin panel, Two, the manual update via FTP.
Furthermore, probably, check the upgrade via WP Toolkit, or refer to it.
Plus, refer to [Upgrading (very old) WordPress](https://make.wordpress.org/hosting/handbook/upgrading/), maybe [moving all this content from the Hosting Handbook (in Markdown)](https://github.com/WordPress/hosting-handbook/blob/main/upgrading.md).
If you add documentation from another WordPress.org page, indicate it in the Changelog or in the comments of this issue.
### To-Do
- [ ] General review and updating
- [ ] Review all the process, both simple (admin) and complex (FTP / SQL)
- [ ] Upgrading (very old) WordPress | 1.0 | Update page: Upgrading WordPress - File: [upgrade/upgrading.md](https://github.com/WordPress/Advanced-administration-handbook/blob/main/upgrade/upgrading.md)
This page needs a general review and update.
Needs to have some different parts. One, the simple update via the Admin panel, Two, the manual update via FTP.
Furthermore, probably, check the upgrade via WP Toolkit, or refer to it.
Plus, refer to [Upgrading (very old) WordPress](https://make.wordpress.org/hosting/handbook/upgrading/), maybe [moving all this content from the Hosting Handbook (in Markdown)](https://github.com/WordPress/hosting-handbook/blob/main/upgrading.md).
If you add documentation from another WordPress.org page, indicate it in the Changelog or in the comments of this issue.
### To-Do
- [ ] General review and updating
- [ ] Review all the process, both simple (admin) and complex (FTP / SQL)
- [ ] Upgrading (very old) WordPress | non_main | update page upgrading wordpress file this page needs a general review and update needs to have some different parts one the simple update via the admin panel two the manual update via ftp furthermore probably check the upgrade via wp toolkit or refer to it plus refer to maybe if you add documentation from another wordpress org page indicate it in the changelog or in the comments of this issue to do general review and updating review all the process both simple admin and complex ftp sql upgrading very old wordpress | 0 |
2,873 | 10,276,031,332 | IssuesEvent | 2019-08-24 13:45:14 | arcticicestudio/arctic | https://api.github.com/repos/arcticicestudio/arctic | closed | lint-staged | context-workflow scope-dx scope-maintainability scope-quality type-feature | <p align="center"><img src="https://user-images.githubusercontent.com/7836623/63638143-c84d4280-c684-11e9-93cf-98662c6c0168.png" width="25%" /></p>
Integrate [lint-staged][gh-lint-staged] to run linters against staged Git files to prevent to add code that violates any style guide into the code base.
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/63638144-c84d4280-c684-11e9-8ba1-1cec576a8fdb.gif" width="80%" /></p>
### Configuration
The configuration file `lint-staged.config.js` will be placed in the project root and includes the command that should be run for matching file extensions (globs). It will include at least the three following entries with the same order as listed here:
1. `prettier --list-different` - Run Prettier (#32) against `*.{js,json,md,mdx,ts,tsx,yml}` to ensure all files are formatted correctly. The `--list-different` prints the found files that are not conform to the Prettier configuration.
2. `eslint` - Run ESLint (#30) against `*.{js,ts,tsx}` to ensure all TypeScript and JavaScript files are compliant to the style guide after being formatted with Prettier.
3. `remark --no-stdout` - Run remark-lint (#27) against `*.md` to ensure all Markdown files are compliant to the style guide. The `--no-stdout` flag suppresses the output of the parsed file content.
## Tasks
- [x] Install [lint-staged][npm-lint-staged] package.
- [x] Implement `lint-staged.config.js` configuration file.
[gh-lint-staged]: https://github.com/okonet/lint-staged
[npm-lint-staged]: https://www.npmjs.com/package/lint-staged
| True | lint-staged - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/63638143-c84d4280-c684-11e9-93cf-98662c6c0168.png" width="25%" /></p>
Integrate [lint-staged][gh-lint-staged] to run linters against staged Git files to prevent to add code that violates any style guide into the code base.
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/63638144-c84d4280-c684-11e9-8ba1-1cec576a8fdb.gif" width="80%" /></p>
### Configuration
The configuration file `lint-staged.config.js` will be placed in the project root and includes the command that should be run for matching file extensions (globs). It will include at least the three following entries with the same order as listed here:
1. `prettier --list-different` - Run Prettier (#32) against `*.{js,json,md,mdx,ts,tsx,yml}` to ensure all files are formatted correctly. The `--list-different` prints the found files that are not conform to the Prettier configuration.
2. `eslint` - Run ESLint (#30) against `*.{js,ts,tsx}` to ensure all TypeScript and JavaScript files are compliant to the style guide after being formatted with Prettier.
3. `remark --no-stdout` - Run remark-lint (#27) against `*.md` to ensure all Markdown files are compliant to the style guide. The `--no-stdout` flag suppresses the output of the parsed file content.
## Tasks
- [x] Install [lint-staged][npm-lint-staged] package.
- [x] Implement `lint-staged.config.js` configuration file.
[gh-lint-staged]: https://github.com/okonet/lint-staged
[npm-lint-staged]: https://www.npmjs.com/package/lint-staged
| main | lint staged integrate to run linters against staged git files to prevent to add code that violates any style guide into the code base configuration the configuration file lint staged config js will be placed in the project root and includes the command that should be run for matching file extensions globs it will include at least the three following entries with the same order as listed here prettier list different run prettier against js json md mdx ts tsx yml to ensure all files are formatted correctly the list different prints the found files that are not conform to the prettier configuration eslint run eslint against js ts tsx to ensure all typescript and javascript files are compliant to the style guide after being formatted with prettier remark no stdout run remark lint against md to ensure all markdown files are compliant to the style guide the no stdout flag suppresses the output of the parsed file content tasks install package implement lint staged config js configuration file | 1 |
5,079 | 25,979,343,578 | IssuesEvent | 2022-12-19 17:20:46 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | Lambda Versioning Issues | area/resource/function type/feature contributors/good-first-issue maintainer/need-response | Hello Team,
This is regarding the issue which I faced during lambda versioning.
We used AutoPublishAlias in SAM template for versioning/aliasing of lambdas which has a restriction that we cannot provide custom description to versions. Intent is to pass commit id as a description as a correlation id between lambda version and commits which generated the versions.
Resource AWS::Lambda::Version is also not proving to be any help, as on the second execution I cloudformation is failing with the message that "Update to resource type AWS::Lambda::Version is not supported.". Going through several blogs I came to know that for any new version to be published we need to define an addition AWS::Lambda::Version resource. I assume we cannot perform update to Lambda Version as it is immutable resource and we have to create new version each time.
AutoPublishAlias to accept a description will help to put a relational mapping between the lambda version and the stack which created the stack. Or if you can map commit id and display it in lambda console as Tags that will also suffice the purpose.
Many Thanks. | True | Lambda Versioning Issues - Hello Team,
This is regarding the issue which I faced during lambda versioning.
We used AutoPublishAlias in SAM template for versioning/aliasing of lambdas which has a restriction that we cannot provide custom description to versions. Intent is to pass commit id as a description as a correlation id between lambda version and commits which generated the versions.
Resource AWS::Lambda::Version is also not proving to be any help, as on the second execution I cloudformation is failing with the message that "Update to resource type AWS::Lambda::Version is not supported.". Going through several blogs I came to know that for any new version to be published we need to define an addition AWS::Lambda::Version resource. I assume we cannot perform update to Lambda Version as it is immutable resource and we have to create new version each time.
AutoPublishAlias to accept a description will help to put a relational mapping between the lambda version and the stack which created the stack. Or if you can map commit id and display it in lambda console as Tags that will also suffice the purpose.
Many Thanks. | main | lambda versioning issues hello team this is regarding the issue which i faced during lambda versioning we used autopublishalias in sam template for versioning aliasing of lambdas which has a restriction that we cannot provide custom description to versions intent is to pass commit id as a description as a correlation id between lambda version and commits which generated the versions resource aws lambda version is also not proving to be any help as on the second execution i cloudformation is failing with the message that update to resource type aws lambda version is not supported going through several blogs i came to know that for any new version to be published we need to define an addition aws lambda version resource i assume we cannot perform update to lambda version as it is immutable resource and we have to create new version each time autopublishalias to accept a description will help to put a relational mapping between the lambda version and the stack which created the stack or if you can map commit id and display it in lambda console as tags that will also suffice the purpose many thanks | 1 |
3,870 | 17,111,618,510 | IssuesEvent | 2021-07-10 12:32:07 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | closed | return statements should be preceded by a blank line | Area: analyzer Area: maintainability feature | If a return statement is the only statement on a line of code, then it should be preceded by a blank line. This allows to easily spot them. | True | return statements should be preceded by a blank line - If a return statement is the only statement on a line of code, then it should be preceded by a blank line. This allows to easily spot them. | main | return statements should be preceded by a blank line if a return statement is the only statement on a line of code then it should be preceded by a blank line this allows to easily spot them | 1 |
3,142 | 12,056,605,572 | IssuesEvent | 2020-04-15 14:41:28 | arcticicestudio/igloo | https://api.github.com/repos/arcticicestudio/igloo | opened | “taskwarrior“ & “timewarrior“ snowblock decommission | scope-maintainability snowblock-taskwarrior snowblock-timewarrior type-task | Related to #248
---
Both _snowblocks_ for [Taskwarrior][] and [Timewarrior][] are not required anymore since they have been replaced with my own custom 💙 [Go][] application that is currently private/closed source, bur planned to be open sourced later on.
Both tools are great and provide a lot of features, but it's kind of an overload and I missed the possibility to integrate the data and API into my other Go applications as well as web-based projects with a quite more modern _techstack_ (_Protocol Buffers_, _NATS Messaging_, _React_ SPA etc.).
Therefore the _snowblocks_ will be removed while the data is still available through the [_Git_ repository history/logs][git-docs-hist].
[git-docs-hist]: https://git-scm.com/book/en/v2/Git-Basics-Viewing-the-Commit-History
[go]: https://go.dev
[taskwarrior]: https://taskwarrior.org
[timewarrior]: https://timewarrior.net | True | “taskwarrior“ & “timewarrior“ snowblock decommission - Related to #248
---
Both _snowblocks_ for [Taskwarrior][] and [Timewarrior][] are not required anymore since they have been replaced with my own custom 💙 [Go][] application that is currently private/closed source, bur planned to be open sourced later on.
Both tools are great and provide a lot of features, but it's kind of an overload and I missed the possibility to integrate the data and API into my other Go applications as well as web-based projects with a quite more modern _techstack_ (_Protocol Buffers_, _NATS Messaging_, _React_ SPA etc.).
Therefore the _snowblocks_ will be removed while the data is still available through the [_Git_ repository history/logs][git-docs-hist].
[git-docs-hist]: https://git-scm.com/book/en/v2/Git-Basics-Viewing-the-Commit-History
[go]: https://go.dev
[taskwarrior]: https://taskwarrior.org
[timewarrior]: https://timewarrior.net | main | “taskwarrior“ “timewarrior“ snowblock decommission related to both snowblocks for and are not required anymore since they have been replaced with my own custom 💙 application that is currently private closed source bur planned to be open sourced later on both tools are great and provide a lot of features but it s kind of an overload and i missed the possibility to integrate the data and api into my other go applications as well as web based projects with a quite more modern techstack protocol buffers nats messaging react spa etc therefore the snowblocks will be removed while the data is still available through the | 1 |
76,337 | 15,495,927,271 | IssuesEvent | 2021-03-11 01:44:57 | rgordon95/github-search-redux-thunk | https://api.github.com/repos/rgordon95/github-search-redux-thunk | opened | CVE-2020-7608 (Medium) detected in yargs-parser-10.1.0.tgz, yargs-parser-11.1.1.tgz | security vulnerability | ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-10.1.0.tgz</b>, <b>yargs-parser-11.1.1.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p>
<p>Path to dependency file: /github-search-redux-thunk/package.json</p>
<p>Path to vulnerable library: github-search-redux-thunk/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- webpack-dev-server-3.2.1.tgz
- yargs-12.0.2.tgz
- :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-11.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p>
<p>Path to dependency file: /github-search-redux-thunk/package.json</p>
<p>Path to vulnerable library: github-search-redux-thunk/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- jest-24.7.1.tgz
- jest-cli-24.8.0.tgz
- yargs-12.0.5.tgz
- :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7608 (Medium) detected in yargs-parser-10.1.0.tgz, yargs-parser-11.1.1.tgz - ## CVE-2020-7608 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>yargs-parser-10.1.0.tgz</b>, <b>yargs-parser-11.1.1.tgz</b></p></summary>
<p>
<details><summary><b>yargs-parser-10.1.0.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-10.1.0.tgz</a></p>
<p>Path to dependency file: /github-search-redux-thunk/package.json</p>
<p>Path to vulnerable library: github-search-redux-thunk/node_modules/webpack-dev-server/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- webpack-dev-server-3.2.1.tgz
- yargs-12.0.2.tgz
- :x: **yargs-parser-10.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>yargs-parser-11.1.1.tgz</b></p></summary>
<p>the mighty option parser used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz">https://registry.npmjs.org/yargs-parser/-/yargs-parser-11.1.1.tgz</a></p>
<p>Path to dependency file: /github-search-redux-thunk/package.json</p>
<p>Path to vulnerable library: github-search-redux-thunk/node_modules/yargs-parser/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-3.0.1.tgz (Root Library)
- jest-24.7.1.tgz
- jest-cli-24.8.0.tgz
- yargs-12.0.5.tgz
- :x: **yargs-parser-11.1.1.tgz** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
yargs-parser could be tricked into adding or modifying properties of Object.prototype using a "__proto__" payload.
<p>Publish Date: 2020-03-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608>CVE-2020-7608</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7608</a></p>
<p>Release Date: 2020-03-16</p>
<p>Fix Resolution: v18.1.1;13.1.2;15.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in yargs parser tgz yargs parser tgz cve medium severity vulnerability vulnerable libraries yargs parser tgz yargs parser tgz yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file github search redux thunk package json path to vulnerable library github search redux thunk node modules webpack dev server node modules yargs parser package json dependency hierarchy react scripts tgz root library webpack dev server tgz yargs tgz x yargs parser tgz vulnerable library yargs parser tgz the mighty option parser used by yargs library home page a href path to dependency file github search redux thunk package json path to vulnerable library github search redux thunk node modules yargs parser package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz yargs tgz x yargs parser tgz vulnerable library vulnerability details yargs parser could be tricked into adding or modifying properties of object prototype using a proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
113,670 | 24,469,477,317 | IssuesEvent | 2022-10-07 18:16:17 | WordPress/openverse-catalog | https://api.github.com/repos/WordPress/openverse-catalog | opened | Consider using `url` as field name in provider scripts | 🟨 priority: medium ✨ goal: improvement 💻 aspect: code | ## Current Situation
<!-- Describe the part of the code you think should improve -->
Currently in our provider scripts, we use a different field name per media type for the url field -- so, `image_url` and `audio_url`. In the actual database, these all just map to `url`, and so we have to manually convert the field name in each of the associated media stores ([example in ImageStore](https://github.com/WordPress/openverse-catalog/blob/main/openverse_catalog/dags/common/storage/image.py#L159)).
## Suggested Improvement
<!-- Describe your proposed change -->
Unless there's a reason it must be done this way, I'd like to just use `url` everywhere.
## Benefit
<!-- Describe the benefit of the change (E.g., increase test coverage, reduce running time, etc.) -->
* We would be able to remove the MediaStore logic to convert the field name
* It would be **much easier** to document what fields are needed in provider scripts, and what they map to in the actual database.
* It would be **much easier** for providers that consume media of multiple types to deal with URLs. See [here an example](https://github.com/WordPress/openverse-catalog/blob/main/openverse_catalog/dags/providers/provider_api_scripts/wikimedia_commons.py#L184) of Wikimedia needing to set the `url`, then later pop it off the data and rename it depending on the media type, _only for it to just get changed back in the media store_
* It would be **much easier** to define types for `ImageData` and `AudioData`.
## Additional context
<!-- Add any other context suggestion here. -->
I'm unaware of the reasoning for the current implementation, so it's possible there's a critical reason why these fields are renamed and we should be sure to test thoroughly.
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
| 1.0 | Consider using `url` as field name in provider scripts - ## Current Situation
<!-- Describe the part of the code you think should improve -->
Currently in our provider scripts, we use a different field name per media type for the url field -- so, `image_url` and `audio_url`. In the actual database, these all just map to `url`, and so we have to manually convert the field name in each of the associated media stores ([example in ImageStore](https://github.com/WordPress/openverse-catalog/blob/main/openverse_catalog/dags/common/storage/image.py#L159)).
## Suggested Improvement
<!-- Describe your proposed change -->
Unless there's a reason it must be done this way, I'd like to just use `url` everywhere.
## Benefit
<!-- Describe the benefit of the change (E.g., increase test coverage, reduce running time, etc.) -->
* We would be able to remove the MediaStore logic to convert the field name
* It would be **much easier** to document what fields are needed in provider scripts, and what they map to in the actual database.
* It would be **much easier** for providers that consume media of multiple types to deal with URLs. See [here an example](https://github.com/WordPress/openverse-catalog/blob/main/openverse_catalog/dags/providers/provider_api_scripts/wikimedia_commons.py#L184) of Wikimedia needing to set the `url`, then later pop it off the data and rename it depending on the media type, _only for it to just get changed back in the media store_
* It would be **much easier** to define types for `ImageData` and `AudioData`.
## Additional context
<!-- Add any other context suggestion here. -->
I'm unaware of the reasoning for the current implementation, so it's possible there's a critical reason why these fields are renamed and we should be sure to test thoroughly.
## Implementation
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in implementing this feature.
| non_main | consider using url as field name in provider scripts current situation currently in our provider scripts we use a different field name per media type for the url field so image url and audio url in the actual database these all just map to url and so we have to manually convert the field name in each of the associated media stores suggested improvement unless there s a reason it must be done this way i d like to just use url everywhere benefit we would be able to remove the mediastore logic to convert the field name it would be much easier to document what fields are needed in provider scripts and what they map to in the actual database it would be much easier for providers that consume media of multiple types to deal with urls see of wikimedia needing to set the url then later pop it off the data and rename it depending on the media type only for it to just get changed back in the media store it would be much easier to define types for imagedata and audiodata additional context i m unaware of the reasoning for the current implementation so it s possible there s a critical reason why these fields are renamed and we should be sure to test thoroughly implementation 🙋 i would be interested in implementing this feature | 0 |
22,338 | 3,953,759,332 | IssuesEvent | 2016-04-29 14:37:11 | sass/libsass | https://api.github.com/repos/sass/libsass | closed | Extending inner of :not can lead to invalid css | Bug - @extends Bug - Confirmed Dev - Test Written | Extracted from https://github.com/sass/libsass/issues/2051
```scss
:not(.thing) {
color: red;
}
:not(.bar) {
@extend .thing;
background: blue;
}
```
[Produces] [1]:
```css
:not(.thing):not(:not(.bar)) {
color: red; }
:not(.bar) {
background: blue; }
```
But should be:
```css
:not(.thing) {
color: red; }
:not(.bar) {
background: blue; }
```
I wonder if ruby sass actually just removes the invalidly nested `:not` selector.
Otherwise from what I know about extend, the result in theory doesn't look incorrect!?
[1]: http://libsass.ocbnet.ch/srcmap/#Om5vdCgudGhpbmcpIHsKICAgIGNvbG9yOiByZWQ7Cn0KOm5vdCguYmFyKSB7CiAgICBAZXh0ZW5kIC50aGluZzsKICAgIGJhY2tncm91bmQ6IGJsdWU7Cn0=
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/33403039-extending-inner-of-not-can-lead-to-invalid-css?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | Extending inner of :not can lead to invalid css - Extracted from https://github.com/sass/libsass/issues/2051
```scss
:not(.thing) {
color: red;
}
:not(.bar) {
@extend .thing;
background: blue;
}
```
[Produces] [1]:
```css
:not(.thing):not(:not(.bar)) {
color: red; }
:not(.bar) {
background: blue; }
```
But should be:
```css
:not(.thing) {
color: red; }
:not(.bar) {
background: blue; }
```
I wonder if ruby sass actually just removes the invalidly nested `:not` selector.
Otherwise from what I know about extend, the result in theory doesn't look incorrect!?
[1]: http://libsass.ocbnet.ch/srcmap/#Om5vdCgudGhpbmcpIHsKICAgIGNvbG9yOiByZWQ7Cn0KOm5vdCguYmFyKSB7CiAgICBAZXh0ZW5kIC50aGluZzsKICAgIGJhY2tncm91bmQ6IGJsdWU7Cn0=
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/33403039-extending-inner-of-not-can-lead-to-invalid-css?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F283068&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_main | extending inner of not can lead to invalid css extracted from scss not thing color red not bar extend thing background blue css not thing not not bar color red not bar background blue but should be css not thing color red not bar background blue i wonder if ruby sass actually just removes the invalidly nested not selector otherwise from what i know about extend the result in theory doesn t look incorrect want to back this issue we accept bounties via | 0 |
24,262 | 3,941,468,811 | IssuesEvent | 2016-04-27 07:52:49 | JulienGenoud/renderscript-examples | https://api.github.com/repos/JulienGenoud/renderscript-examples | closed | Page Curl is not working in 4.1 | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Downloaded pagecurl project
2. Imported to Eclipse ->API level 16
3. Created AVD with option GPU on
4. Run the application
What is the expected output? What do you see instead?
Expected
Should show the Pagecurl but
Actual
No rendering,Black screen
What version of the product are you using? On what operating system?
Windows 7 32bit,Android Emulator
Please provide any additional information below.
```
Original issue reported on code.google.com by `kuvetha...@gmail.com` on 17 Jul 2012 at 6:17 | 1.0 | Page Curl is not working in 4.1 - ```
What steps will reproduce the problem?
1. Downloaded pagecurl project
2. Imported to Eclipse ->API level 16
3. Created AVD with option GPU on
4. Run the application
What is the expected output? What do you see instead?
Expected
Should show the Pagecurl but
Actual
No rendering,Black screen
What version of the product are you using? On what operating system?
Windows 7 32bit,Android Emulator
Please provide any additional information below.
```
Original issue reported on code.google.com by `kuvetha...@gmail.com` on 17 Jul 2012 at 6:17 | non_main | page curl is not working in what steps will reproduce the problem downloaded pagecurl project imported to eclipse api level created avd with option gpu on run the application what is the expected output what do you see instead expected should show the pagecurl but actual no rendering black screen what version of the product are you using on what operating system windows android emulator please provide any additional information below original issue reported on code google com by kuvetha gmail com on jul at | 0 |
40,427 | 5,216,419,962 | IssuesEvent | 2017-01-26 10:16:14 | AeroScripts/QuestieDev | https://api.github.com/repos/AeroScripts/QuestieDev | closed | Error message on login v3.69 | by design hotfix resolved | Here is a screenshot of the error i get.

I get this every time i log into my character or every time i /reload.
This is a new character, only first quest is accepted and i haven't completed it yet... I click Yes and after reload it's there again! Never goes away...
The first time i launch the addon after deleting the SavedVariables i don't get the message.
| 1.0 | Error message on login v3.69 - Here is a screenshot of the error i get.

I get this every time i log into my character or every time i /reload.
This is a new character, only first quest is accepted and i haven't completed it yet... I click Yes and after reload it's there again! Never goes away...
The first time i launch the addon after deleting the SavedVariables i don't get the message.
| non_main | error message on login here is a screenshot of the error i get i get this every time i log into my character or every time i reload this is a new character only first quest is accepted and i haven t completed it yet i click yes and after reload it s there again never goes away the first time i launch the addon after deleting the savedvariables i don t get the message | 0 |
783,331 | 27,526,564,274 | IssuesEvent | 2023-03-06 18:30:08 | PardeeCenterDU/IFs-Issues-Tracking | https://api.github.com/repos/PardeeCenterDU/IFs-Issues-Tracking | opened | MFPCORS and IGDPR | New Issue / Not Under Review Priority 1 Modeling - Economic Module | Description:
When IGDPR is used to calculate MFPCORS in the first year, IGDPR is used as the target for all sectors. This leads to lower growth than expected due to agriculture and energy.
With agriculture and energy being computed in their submodules, this led to lower growth than expected in high-growth lower income economies with substantial agricultural sectors (like China) during the validation work.
Numeric Example:
IGDPR = 10
Agricultural Growth Rate = 3
Energy Growth Rate = 6
All Sectoral Shares = 1/6
Calculated Growth Rate = (3 + 6 + 10 + 10 + 10 + 10) / 6 = 8.1
"Ideal" Growth Rates = (3 + 6 + (4 * 10x)) /6 = 10; x = 1.275
Existing Processes:
_"First Year Compute Sectoral Growth Rates"_, adjusts sectoral growth rates to account for slower agricultural growth. For low-income countries, twice TGRYL. Else, just TGRYL.
_"CorrectReduction"_, shifts the correction factor from agricultural sectors to others.
Temporary Workaround:
1. Calculate VADD for each sector, sum for computed GDP
2. Calculate expected GDP based on IGDPR
3. Compare computed and expected GDP
4. Adjust sectoral correction factors for sectors 3 through 6 (see "Ideal" Growth Rate in Numeric Example) | 1.0 | MFPCORS and IGDPR - Description:
When IGDPR is used to calculate MFPCORS in the first year, IGDPR is used as the target for all sectors. This leads to lower growth than expected due to agriculture and energy.
With agriculture and energy being computed in their submodules, this led to lower growth than expected in high-growth lower income economies with substantial agricultural sectors (like China) during the validation work.
Numeric Example:
IGDPR = 10
Agricultural Growth Rate = 3
Energy Growth Rate = 6
All Sectoral Shares = 1/6
Calculated Growth Rate = (3 + 6 + 10 + 10 + 10 + 10) / 6 = 8.1
"Ideal" Growth Rates = (3 + 6 + (4 * 10x)) /6 = 10; x = 1.275
Existing Processes:
_"First Year Compute Sectoral Growth Rates"_, adjusts sectoral growth rates to account for slower agricultural growth. For low-income countries, twice TGRYL. Else, just TGRYL.
_"CorrectReduction"_, shifts the correction factor from agricultural sectors to others.
Temporary Workaround:
1. Calculate VADD for each sector, sum for computed GDP
2. Calculate expected GDP based on IGDPR
3. Compare computed and expected GDP
4. Adjust sectoral correction factors for sectors 3 through 6 (see "Ideal" Growth Rate in Numeric Example) | non_main | mfpcors and igdpr description when igdpr is used to calculate mfpcors in the first year igdpr is used as the target for all sectors this leads to lower growth than expected due to agriculture and energy with agriculture and energy being computed in their submodules this led to lower growth than expected in high growth lower income economies with substantial agricultural sectors like china during the validation work numeric example igdpr agricultural growth rate energy growth rate all sectoral shares calculated growth rate ideal growth rates x existing processes first year compute sectoral growth rates adjusts sectoral growth rates to account for slower agricultural growth for low income countries twice tgryl else just tgryl correctreduction shifts the correction factor from agricultural sectors to others temporary workaround calculate vadd for each sector sum for computed gdp calculate expected gdp based on igdpr compare computed and expected gdp adjust sectoral correction factors for sectors through see ideal growth rate in numeric example | 0 |
221,940 | 17,030,687,401 | IssuesEvent | 2021-07-04 13:55:08 | Kartonagnick/mygtest | https://api.github.com/repos/Kartonagnick/mygtest | closed | pre | documentation | #9-pre
======
Удаляем PRE-версии.
План работ:
- [x] обрабатываем реализацию.
- [x] обрабатываем тесты.
- [x] обрабатываем документацию.
| 1.0 | pre - #9-pre
======
Удаляем PRE-версии.
План работ:
- [x] обрабатываем реализацию.
- [x] обрабатываем тесты.
- [x] обрабатываем документацию.
| non_main | pre pre удаляем pre версии план работ обрабатываем реализацию обрабатываем тесты обрабатываем документацию | 0 |
4,862 | 25,013,380,283 | IssuesEvent | 2022-11-03 16:48:59 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Use bidirectional substring matching within the Record Selector fuzzy search algorithm | type: enhancement work: backend status: ready restricted: maintainers | ## Current behavior
- The search_fuzzy algorithm awards points to records that contain the search string. If I search for "Jenn", a stored value of "Jennifer" will match.
## Desired behavior
- I want the matching to work the other way around too. If I search for "Jennifer", I want stored values like "Jenn" to be awarded _some_ points as well.

## Implementation
- We might want to play with the point values. Perhaps we bump up all the existing point values and bring in these new match types with lower values.
| True | Use bidirectional substring matching within the Record Selector fuzzy search algorithm - ## Current behavior
- The search_fuzzy algorithm awards points to records that contain the search string. If I search for "Jenn", a stored value of "Jennifer" will match.
## Desired behavior
- I want the matching to work the other way around too. If I search for "Jennifer", I want stored values like "Jenn" to be awarded _some_ points as well.

## Implementation
- We might want to play with the point values. Perhaps we bump up all the existing point values and bring in these new match types with lower values.
| main | use bidirectional substring matching within the record selector fuzzy search algorithm current behavior the search fuzzy algorithm awards points to records that contain the search string if i search for jenn a stored value of jennifer will match desired behavior i want the matching to work the other way around too if i search for jennifer i want stored values like jenn to be awarded some points as well implementation we might want to play with the point values perhaps we bump up all the existing point values and bring in these new match types with lower values | 1 |
51,786 | 13,648,272,329 | IssuesEvent | 2020-09-26 08:18:01 | srivatsamarichi/tailspin-spacegame | https://api.github.com/repos/srivatsamarichi/tailspin-spacegame | closed | CVE-2018-20822 (Medium) detected in opennmsopennms-source-22.0.1-1 | bug security vulnerability | ## CVE-2018-20822 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-22.0.1-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/tailspin-spacegame/commit/062dd21a77b97bb3d05571621534cadd053fff2a">062dd21a77b97bb3d05571621534cadd053fff2a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>tailspin-spacegame/node_modules/node-sass/src/libsass/src/ast.hpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20822>CVE-2018-20822</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0;node-sass - 4.13.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-20822 (Medium) detected in opennmsopennms-source-22.0.1-1 - ## CVE-2018-20822 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opennmsopennms-source-22.0.1-1</b></p></summary>
<p>
<p>A Java based fault and performance management system</p>
<p>Library home page: <a href=https://sourceforge.net/projects/opennms/>https://sourceforge.net/projects/opennms/</a></p>
<p>Found in HEAD commit: <a href="https://github.com/srivatsamarichi/tailspin-spacegame/commit/062dd21a77b97bb3d05571621534cadd053fff2a">062dd21a77b97bb3d05571621534cadd053fff2a</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>tailspin-spacegame/node_modules/node-sass/src/libsass/src/ast.hpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
LibSass 3.5.4 allows attackers to cause a denial-of-service (uncontrolled recursion in Sass::Complex_Selector::perform in ast.hpp and Sass::Inspect::operator in inspect.cpp).
<p>Publish Date: 2019-04-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20822>CVE-2018-20822</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20822</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0;node-sass - 4.13.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in opennmsopennms source cve medium severity vulnerability vulnerable library opennmsopennms source a java based fault and performance management system library home page a href found in head commit a href found in base branch master vulnerable source files tailspin spacegame node modules node sass src libsass src ast hpp vulnerability details libsass allows attackers to cause a denial of service uncontrolled recursion in sass complex selector perform in ast hpp and sass inspect operator in inspect cpp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass node sass step up your open source security game with whitesource | 0 |
766 | 4,366,359,147 | IssuesEvent | 2016-08-03 14:09:00 | duckduckgo/zeroclickinfo-fathead | https://api.github.com/repos/duckduckgo/zeroclickinfo-fathead | closed | Python: unify python2 and python3 docs | Maintainer Input Requested | **Background**
We have coverage of docs.python.org for Python3 but none for Python2. Even though Python2 is considered a legacy version, it still has a lot of use.
**Objective**
Unify our python docs fathead to cover both Python3 and Python2.
The exact implementation is up to whoever develops this but I think the general idea is to:
- Default to python3 docs in cases where both python3 & 2 are identical
- In cases where they're different show the Python3 docs but include an infobox link to the Python2 docs.
How I think this could work technically:
- update parse.py to parse both Python3 and Python2 docs
- Generate separate output.txt files for each
- Go through both output files comparing keys and abstracts
- If the Python3 version is identical to the Python2 just keep the Python3 entry
- If they're different we can default that search to Python3 but show an infobox that links to the Python2 version.
---
IA Page: https://duck.co/ia/view/python | True | Python: unify python2 and python3 docs - **Background**
We have coverage of docs.python.org for Python3 but none for Python2. Even though Python2 is considered a legacy version, it still has a lot of use.
**Objective**
Unify our python docs fathead to cover both Python3 and Python2.
The exact implementation is up to whoever develops this but I think the general idea is to:
- Default to python3 docs in cases where both python3 & 2 are identical
- In cases where they're different show the Python3 docs but include an infobox link to the Python2 docs.
How I think this could work technically:
- update parse.py to parse both Python3 and Python2 docs
- Generate separate output.txt files for each
- Go through both output files comparing keys and abstracts
- If the Python3 version is identical to the Python2 just keep the Python3 entry
- If they're different we can default that search to Python3 but show an infobox that links to the Python2 version.
---
IA Page: https://duck.co/ia/view/python | main | python unify and docs background we have coverage of docs python org for but none for even though is considered a legacy version it still has a lot of use objective unify our python docs fathead to cover both and the exact implementation is up to whoever develops this but i think the general idea is to default to docs in cases where both are identical in cases where they re different show the docs but include an infobox link to the docs how i think this could work technically update parse py to parse both and docs generate separate output txt files for each go through both output files comparing keys and abstracts if the version is identical to the just keep the entry if they re different we can default that search to but show an infobox that links to the version ia page | 1 |
1,784 | 6,575,850,595 | IssuesEvent | 2017-09-11 17:34:17 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Validation with visudo does not work for lineinfile if | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
lineinfile
##### ANSIBLE VERSION
```
ansible 2.1.1.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
```
CentOS Linux release 7.2.1511 (Core)
Linux 3.10.0-327.18.2.el7.x86_64 #1 SMP Thu May 12 11:03:55 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
```
##### SUMMARY
When trying to create or modify a file in _/etc/sudoers.d_ by using the `lineinfile` module the validation with visudo fails because a temporary file is not found.
http://docs.ansible.com/ansible/lineinfile_module.html
##### STEPS TO REPRODUCE
```
- name: Setup sudoers permissions
lineinfile: dest=/etc/sudoers.d/icinga2
create=yes
state=present
line='icinga ALL=(ALL) NOPASSWD:/usr/bin/find'
validate='visudo -cf %s'
```
##### EXPECTED RESULTS
A file created under _/etc/sudoers.d/icinga2_ with the content `icinga ALL=(ALL) NOPASSWD:/usr/bin/find`which passed validation.
##### ACTUAL RESULTS
```
FAILED! => {"changed": false, "cmd": "visudo -cf /tmp/tmpSBsM5A", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2
```
| True | Validation with visudo does not work for lineinfile if - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
lineinfile
##### ANSIBLE VERSION
```
ansible 2.1.1.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
```
CentOS Linux release 7.2.1511 (Core)
Linux 3.10.0-327.18.2.el7.x86_64 #1 SMP Thu May 12 11:03:55 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
```
##### SUMMARY
When trying to create or modify a file in _/etc/sudoers.d_ by using the `lineinfile` module the validation with visudo fails because a temporary file is not found.
http://docs.ansible.com/ansible/lineinfile_module.html
##### STEPS TO REPRODUCE
```
- name: Setup sudoers permissions
lineinfile: dest=/etc/sudoers.d/icinga2
create=yes
state=present
line='icinga ALL=(ALL) NOPASSWD:/usr/bin/find'
validate='visudo -cf %s'
```
##### EXPECTED RESULTS
A file created under _/etc/sudoers.d/icinga2_ with the content `icinga ALL=(ALL) NOPASSWD:/usr/bin/find`which passed validation.
##### ACTUAL RESULTS
```
FAILED! => {"changed": false, "cmd": "visudo -cf /tmp/tmpSBsM5A", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2
```
| main | validation with visudo does not work for lineinfile if issue type bug report component name lineinfile ansible version ansible configuration n a os environment centos linux release core linux smp thu may utc gnu linux summary when trying to create or modify a file in etc sudoers d by using the lineinfile module the validation with visudo fails because a temporary file is not found steps to reproduce name setup sudoers permissions lineinfile dest etc sudoers d create yes state present line icinga all all nopasswd usr bin find validate visudo cf s expected results a file created under etc sudoers d with the content icinga all all nopasswd usr bin find which passed validation actual results failed changed false cmd visudo cf tmp failed true msg no such file or directory rc | 1 |
65,284 | 6,954,326,090 | IssuesEvent | 2017-12-07 00:47:59 | equella/Equella | https://api.github.com/repos/equella/Equella | closed | 6.5 Beta contribution wizard - fileAttachment size restriction. | bug Ready for 6.5 GA Testing | Once turned on, it restricts all file sizes, not just the size limit. Tested both with the new drag and drop and file select.
| 1.0 | 6.5 Beta contribution wizard - fileAttachment size restriction. - Once turned on, it restricts all file sizes, not just the size limit. Tested both with the new drag and drop and file select.
| non_main | beta contribution wizard fileattachment size restriction once turned on it restricts all file sizes not just the size limit tested both with the new drag and drop and file select | 0 |
363 | 3,343,606,733 | IssuesEvent | 2015-11-15 16:58:13 | EMGroup/js-eden | https://api.github.com/repos/EMGroup/js-eden | closed | Cyclic dependency checks | Construit maintainer | Either these checks have been broken in "handcraftedparser" branch or they are inadequate. | True | Cyclic dependency checks - Either these checks have been broken in "handcraftedparser" branch or they are inadequate. | main | cyclic dependency checks either these checks have been broken in handcraftedparser branch or they are inadequate | 1 |
204,021 | 7,079,346,842 | IssuesEvent | 2018-01-10 09:13:25 | telerik/kendo-ui-core | https://api.github.com/repos/telerik/kendo-ui-core | closed | AutoCompleteFor client side validation from DataAnnotations does not work | Bug C: AutoComplete Kendo2 Priority 3 S: Wrappers (ASP.NET Core) SEV: Medium | ### Bug report
### Reproduction of the problem
In ASP.NET Core The AutoCompleteFor does not respect the [Required] DataAnnotation of the Model.
Use the following files to create a sample form which has DropDownListFor and AutoCompleteFor and only DropDownListFor is validated
[AutoCompleteFor View,Model,Controller.zip](https://github.com/telerik/kendo-ui-core/files/613602/AutoCompleteFor.View.Model.Controller.zip)
| 1.0 | AutoCompleteFor client side validation from DataAnnotations does not work - ### Bug report
### Reproduction of the problem
In ASP.NET Core The AutoCompleteFor does not respect the [Required] DataAnnotation of the Model.
Use the following files to create a sample form which has DropDownListFor and AutoCompleteFor and only DropDownListFor is validated
[AutoCompleteFor View,Model,Controller.zip](https://github.com/telerik/kendo-ui-core/files/613602/AutoCompleteFor.View.Model.Controller.zip)
| non_main | autocompletefor client side validation from dataannotations does not work bug report reproduction of the problem in asp net core the autocompletefor does not respect the dataannotation of the model use the following files to create a sample form which has dropdownlistfor and autocompletefor and only dropdownlistfor is validated | 0 |
772 | 4,381,215,059 | IssuesEvent | 2016-08-06 03:36:16 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | ec2_snapshot_facts errors on boto3_tag_list_to_ansible_dict | aws bug_report cloud waiting_on_maintainer | ##### ISSUE TYPE
Bug Report
##### COMPONENT NAME
```
ec2_snapshot_facts
```
##### ANSIBLE VERSION
```
ansible 2.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
no ANSIBLE_* env vars
standard settings, boto setup correctly
```
##### OS / ENVIRONMENT
```
mac el capitan
```
##### SUMMARY
```
ec2_snapshot_facts errors when I seem to be using it correctly
```
##### STEPS TO REPRODUCE
```
ansible localhost -i ec2.py -m ec2_snapshot_facts -a "owner_ids=self" -vvvv
```
##### EXPECTED RESULTS
```
I expected it would return with the snapshot results
```
##### ACTUAL RESULTS
```
Loaded callback minimal of type stdout, v2.0
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: user
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593 `" )'
<127.0.0.1> PUT /var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/tmpV4c2wp TO /Users/user/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593/ec2_snapshot_facts
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/local/opt/python/bin/python2.7 /Users/user/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593/ec2_snapshot_facts; rm -rf "/Users/user/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593/" > /dev/null 2>&1'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py", line 226, in <module>
main()
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py", line 220, in main
list_ec2_snapshots(connection, module)
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py", line 187, in list_ec2_snapshots
snapshot['tags'] = boto3_tag_list_to_ansible_dict(snapshot['tags'])
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_modlib.zip/ansible/module_utils/ec2.py", line 367, in boto3_tag_list_to_ansible_dict
KeyError: 'Value'
localhost | FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "ec2_snapshot_facts"
},
"module_stderr": "Traceback (most recent call last):\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py\", line 226, in <module>\n main()\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py\", line 220, in main\n list_ec2_snapshots(connection, module)\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py\", line 187, in list_ec2_snapshots\n snapshot['tags'] = boto3_tag_list_to_ansible_dict(snapshot['tags'])\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_modlib.zip/ansible/module_utils/ec2.py\", line 367, in boto3_tag_list_to_ansible_dict\nKeyError: 'Value'\n",
"module_stdout": "",
"msg": "MODULE FAILURE",
"parsed": false
}
``` | True | ec2_snapshot_facts errors on boto3_tag_list_to_ansible_dict - ##### ISSUE TYPE
Bug Report
##### COMPONENT NAME
```
ec2_snapshot_facts
```
##### ANSIBLE VERSION
```
ansible 2.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
no ANSIBLE_* env vars
standard settings, boto setup correctly
```
##### OS / ENVIRONMENT
```
mac el capitan
```
##### SUMMARY
```
ec2_snapshot_facts errors when I seem to be using it correctly
```
##### STEPS TO REPRODUCE
```
ansible localhost -i ec2.py -m ec2_snapshot_facts -a "owner_ids=self" -vvvv
```
##### EXPECTED RESULTS
```
I expected it would return with the snapshot results
```
##### ACTUAL RESULTS
```
Loaded callback minimal of type stdout, v2.0
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: user
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593 `" )'
<127.0.0.1> PUT /var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/tmpV4c2wp TO /Users/user/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593/ec2_snapshot_facts
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/local/opt/python/bin/python2.7 /Users/user/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593/ec2_snapshot_facts; rm -rf "/Users/user/.ansible/tmp/ansible-tmp-1461889080.94-119372330089593/" > /dev/null 2>&1'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py", line 226, in <module>
main()
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py", line 220, in main
list_ec2_snapshots(connection, module)
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py", line 187, in list_ec2_snapshots
snapshot['tags'] = boto3_tag_list_to_ansible_dict(snapshot['tags'])
File "/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_modlib.zip/ansible/module_utils/ec2.py", line 367, in boto3_tag_list_to_ansible_dict
KeyError: 'Value'
localhost | FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "ec2_snapshot_facts"
},
"module_stderr": "Traceback (most recent call last):\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py\", line 226, in <module>\n main()\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py\", line 220, in main\n list_ec2_snapshots(connection, module)\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_module_ec2_snapshot_facts.py\", line 187, in list_ec2_snapshots\n snapshot['tags'] = boto3_tag_list_to_ansible_dict(snapshot['tags'])\n File \"/var/folders/1c/nn7tpsh97cs174pwq6dmhf1r0000gn/T/ansible_RZp_PC/ansible_modlib.zip/ansible/module_utils/ec2.py\", line 367, in boto3_tag_list_to_ansible_dict\nKeyError: 'Value'\n",
"module_stdout": "",
"msg": "MODULE FAILURE",
"parsed": false
}
``` | main | snapshot facts errors on tag list to ansible dict issue type bug report component name snapshot facts ansible version ansible config file configured module search path default w o overrides configuration no ansible env vars standard settings boto setup correctly os environment mac el capitan summary snapshot facts errors when i seem to be using it correctly steps to reproduce ansible localhost i py m snapshot facts a owner ids self vvvv expected results i expected it would return with the snapshot results actual results loaded callback minimal of type stdout establish local connection for user user exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp put var folders t to users user ansible tmp ansible tmp snapshot facts exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr local opt python bin users user ansible tmp ansible tmp snapshot facts rm rf users user ansible tmp ansible tmp dev null an exception occurred during task execution the full traceback is traceback most recent call last file var folders t ansible rzp pc ansible module snapshot facts py line in main file var folders t ansible rzp pc ansible module snapshot facts py line in main list snapshots connection module file var folders t ansible rzp pc ansible module snapshot facts py line in list snapshots snapshot tag list to ansible dict snapshot file var folders t ansible rzp pc ansible modlib zip ansible module utils py line in tag list to ansible dict keyerror value localhost failed changed false failed true invocation module name snapshot facts module stderr traceback most recent call last n file var folders t ansible rzp pc ansible module snapshot facts py line in n main n file var folders t ansible rzp pc ansible module snapshot facts py line in main n list snapshots connection module n file var folders t ansible rzp pc ansible module snapshot facts py line in list snapshots n snapshot tag list to ansible dict snapshot n file var folders t ansible rzp pc ansible modlib zip ansible module utils py line in tag list to ansible dict nkeyerror value n module stdout msg module failure parsed false | 1 |
350 | 3,252,232,192 | IssuesEvent | 2015-10-19 14:03:34 | Homebrew/homebrew | https://api.github.com/repos/Homebrew/homebrew | closed | List outdated formulae separately in `brew update` | features maintainer feedback | Recent updates have attempted to highlight outdated brews to users by [adding coloured highlights](https://github.com/Homebrew/homebrew/pull/44335), which [didn't play too well](https://github.com/Homebrew/homebrew/issues/45028) with some terminal colour schemes.
Currently, any installed formulae that're affected by `brew update` are displayed in bold with ` (installed)` appended. While a lot easier on the eyes than white-against-yellow, it adds clutter to the feedback - and some users probably aren't that quick to notice the bold letters.
I suggest simply listing affected formulae after the updates:

Ignore what's listed in the example, I had to improvise with makeshift feedback, since all my brews are currently up-to-date. Heh.
Thoughts? | True | List outdated formulae separately in `brew update` - Recent updates have attempted to highlight outdated brews to users by [adding coloured highlights](https://github.com/Homebrew/homebrew/pull/44335), which [didn't play too well](https://github.com/Homebrew/homebrew/issues/45028) with some terminal colour schemes.
Currently, any installed formulae that're affected by `brew update` are displayed in bold with ` (installed)` appended. While a lot easier on the eyes than white-against-yellow, it adds clutter to the feedback - and some users probably aren't that quick to notice the bold letters.
I suggest simply listing affected formulae after the updates:

Ignore what's listed in the example, I had to improvise with makeshift feedback, since all my brews are currently up-to-date. Heh.
Thoughts? | main | list outdated formulae separately in brew update recent updates have attempted to highlight outdated brews to users by which with some terminal colour schemes currently any installed formulae that re affected by brew update are displayed in bold with installed appended while a lot easier on the eyes than white against yellow it adds clutter to the feedback and some users probably aren t that quick to notice the bold letters i suggest simply listing affected formulae after the updates ignore what s listed in the example i had to improvise with makeshift feedback since all my brews are currently up to date heh thoughts | 1 |
5,454 | 27,291,051,580 | IssuesEvent | 2023-02-23 16:37:56 | informalsystems/hermes | https://api.github.com/repos/informalsystems/hermes | opened | Investigate dependence on (and prepare for changes to) DeliverTX Events and Log | I: logic I: dependencies O: maintainability | <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Word of caution: poorly thought-out proposals may be rejected
v without deliberation
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
SDK is planning a refactor of `DeliverTX` responses: https://github.com/cosmos/cosmos-sdk/issues/15017.
I suspect Hermes depends on this surface, see below for example:
https://github.com/informalsystems/hermes/blob/097c5dfd9fc73fde0a337cd00d1e151abe5dbcaf/crates/relayer/src/event/rpc.rs#L117
<!-- Short description of the proposed feature -->
<!-- Apply relevant labels to indicate:
- (WHY) The purpose or objective of this proposal with "O" labels
- (WHICH) The part of the system this proposal relates to (use "E" for external or "I" for internal levels)
- (HOW) If any administrative considerations should be taken into account (use "A" labels)
This will help us prioritize and categorize your proposal more effectively
-->
Suggested actions: allocate time to
- engage with the SDK team to inform their new design/refactoring, because any change has the potential to simplify (or break) the relayer
- prepare for breaking changes
- inform the design for relayer-next based on the refactoring that the SDK is doing
## Problem Definition
<!-- Why do we need this feature?
What problems may be addressed by introducing this feature?
What benefits does IBC-rs stand to gain by including this feature?
Are there any disadvantages of including this feature? -->
## Proposal
<!-- Detailed description of requirements of implementation -->
## Acceptance Criteria
<!-- What's the definition of "done" for this issue? -->
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate milestone (priority) applied
- [ ] Appropriate contributors tagged
- [ ] Contributor assigned/self-assigned
| True | Investigate dependence on (and prepare for changes to) DeliverTX Events and Log - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Word of caution: poorly thought-out proposals may be rejected
v without deliberation
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
SDK is planning a refactor of `DeliverTX` responses: https://github.com/cosmos/cosmos-sdk/issues/15017.
I suspect Hermes depends on this surface, see below for example:
https://github.com/informalsystems/hermes/blob/097c5dfd9fc73fde0a337cd00d1e151abe5dbcaf/crates/relayer/src/event/rpc.rs#L117
<!-- Short description of the proposed feature -->
<!-- Apply relevant labels to indicate:
- (WHY) The purpose or objective of this proposal with "O" labels
- (WHICH) The part of the system this proposal relates to (use "E" for external or "I" for internal levels)
- (HOW) If any administrative considerations should be taken into account (use "A" labels)
This will help us prioritize and categorize your proposal more effectively
-->
Suggested actions: allocate time to
- engage with the SDK team to inform their new design/refactoring, because any change has the potential to simplify (or break) the relayer
- prepare for breaking changes
- inform the design for relayer-next based on the refactoring that the SDK is doing
## Problem Definition
<!-- Why do we need this feature?
What problems may be addressed by introducing this feature?
What benefits does IBC-rs stand to gain by including this feature?
Are there any disadvantages of including this feature? -->
## Proposal
<!-- Detailed description of requirements of implementation -->
## Acceptance Criteria
<!-- What's the definition of "done" for this issue? -->
____
#### For Admin Use
- [ ] Not duplicate issue
- [ ] Appropriate labels applied
- [ ] Appropriate milestone (priority) applied
- [ ] Appropriate contributors tagged
- [ ] Contributor assigned/self-assigned
| main | investigate dependence on and prepare for changes to delivertx events and log ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v word of caution poorly thought out proposals may be rejected v without deliberation ☺ summary sdk is planning a refactor of delivertx responses i suspect hermes depends on this surface see below for example apply relevant labels to indicate why the purpose or objective of this proposal with o labels which the part of the system this proposal relates to use e for external or i for internal levels how if any administrative considerations should be taken into account use a labels this will help us prioritize and categorize your proposal more effectively suggested actions allocate time to engage with the sdk team to inform their new design refactoring because any change has the potential to simplify or break the relayer prepare for breaking changes inform the design for relayer next based on the refactoring that the sdk is doing problem definition why do we need this feature what problems may be addressed by introducing this feature what benefits does ibc rs stand to gain by including this feature are there any disadvantages of including this feature proposal acceptance criteria for admin use not duplicate issue appropriate labels applied appropriate milestone priority applied appropriate contributors tagged contributor assigned self assigned | 1 |
2,069 | 6,997,234,202 | IssuesEvent | 2017-12-16 12:10:06 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | closed | URL audit issues | awaiting maintainer feedback bug core | https://github.com/caskroom/homebrew-cask/pull/41733
Downgrades to HTTP mirrors.
```
audit for qt-creator: failed
- The URL http://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg should use HTTPS rather than HTTP
Error: audit failed for 1 cask: qt-creator
```
```
Link: <http://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg.meta4>; rel=describedby; type="application/metalink4+xml"
Link: <http://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg.torrent>; rel=describedby; type="application/x-bittorrent"
Link: <http://master.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=1; geo=fi
Link: <http://ftp.fau.de/qtproject/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=2; geo=de
Link: <http://qt-mirror.dannhauer.de/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=3; geo=de
Link: <http://mirrors.dotsrc.org/qtproject/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=4; geo=dk
Link: <http://mirror.netcologne.de/qtproject/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=5; geo=de
Location: http://master.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg
```
```
URL: https://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg
Redirected: http://master.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg
```
____
https://github.com/caskroom/homebrew-cask/pull/41724
Downloads then fails.
```
==> 1 modified cask: anka-flow
==> Downloading https://d1efqjhnhbvc57.cloudfront.net/AnkaFlow-1.1.1.79.pkg
==> Verifying checksum for Cask anka-flow
audit for anka-flow: failed
- The URL https://d1efqjhnhbvc57.cloudfront.net/AnkaFlow-1.1.1.79.pkg is not reachable (HTTP status code 403)
Error: audit failed for 1 cask: anka-flow
```
____
https://github.com/caskroom/homebrew-versions/pull/4928
https://www.opera.com/computer/beta is reachable.
```
==> Verifying checksum for Cask opera-developer
audit for opera-developer: failed
- The URL https://www.opera.com/computer/beta is not reachable
Error: audit failed for 1 cask: opera-developer
```
___
https://github.com/caskroom/homebrew-fonts/pull/1386
SVN font Casks
```
audit for font-abeezee: failed
- The URL https://github.com/google/fonts/trunk/ofl/abeezee is not reachable (HTTP status code 404)
audit for font-advent-pro: failed
- The URL https://github.com/google/fonts/trunk/ofl/adventpro is not reachable (HTTP status code 404)
audit for font-allan: failed
- The URL https://github.com/google/fonts/trunk/ofl/allan is not reachable (HTTP status code 404)
```
____
`brew cask audit --debug` `0ad` and `adobe-photoshop441` both hang at `==> Auditing URL format` with no further output. Both return HTTP 200. | True | URL audit issues - https://github.com/caskroom/homebrew-cask/pull/41733
Downgrades to HTTP mirrors.
```
audit for qt-creator: failed
- The URL http://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg should use HTTPS rather than HTTP
Error: audit failed for 1 cask: qt-creator
```
```
Link: <http://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg.meta4>; rel=describedby; type="application/metalink4+xml"
Link: <http://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg.torrent>; rel=describedby; type="application/x-bittorrent"
Link: <http://master.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=1; geo=fi
Link: <http://ftp.fau.de/qtproject/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=2; geo=de
Link: <http://qt-mirror.dannhauer.de/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=3; geo=de
Link: <http://mirrors.dotsrc.org/qtproject/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=4; geo=dk
Link: <http://mirror.netcologne.de/qtproject/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg>; rel=duplicate; pri=5; geo=de
Location: http://master.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg
```
```
URL: https://download.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg
Redirected: http://master.qt.io/official_releases/qtcreator/4.5/4.5.0/qt-creator-opensource-mac-x86_64-4.5.0.dmg
```
____
https://github.com/caskroom/homebrew-cask/pull/41724
Downloads then fails.
```
==> 1 modified cask: anka-flow
==> Downloading https://d1efqjhnhbvc57.cloudfront.net/AnkaFlow-1.1.1.79.pkg
==> Verifying checksum for Cask anka-flow
audit for anka-flow: failed
- The URL https://d1efqjhnhbvc57.cloudfront.net/AnkaFlow-1.1.1.79.pkg is not reachable (HTTP status code 403)
Error: audit failed for 1 cask: anka-flow
```
____
https://github.com/caskroom/homebrew-versions/pull/4928
https://www.opera.com/computer/beta is reachable.
```
==> Verifying checksum for Cask opera-developer
audit for opera-developer: failed
- The URL https://www.opera.com/computer/beta is not reachable
Error: audit failed for 1 cask: opera-developer
```
___
https://github.com/caskroom/homebrew-fonts/pull/1386
SVN font Casks
```
audit for font-abeezee: failed
- The URL https://github.com/google/fonts/trunk/ofl/abeezee is not reachable (HTTP status code 404)
audit for font-advent-pro: failed
- The URL https://github.com/google/fonts/trunk/ofl/adventpro is not reachable (HTTP status code 404)
audit for font-allan: failed
- The URL https://github.com/google/fonts/trunk/ofl/allan is not reachable (HTTP status code 404)
```
____
`brew cask audit --debug` `0ad` and `adobe-photoshop441` both hang at `==> Auditing URL format` with no further output. Both return HTTP 200. | main | url audit issues downgrades to http mirrors audit for qt creator failed the url should use https rather than http error audit failed for cask qt creator link rel describedby type application xml link rel describedby type application x bittorrent link rel duplicate pri geo fi link rel duplicate pri geo de link rel duplicate pri geo de link rel duplicate pri geo dk link rel duplicate pri geo de location url redirected downloads then fails modified cask anka flow downloading verifying checksum for cask anka flow audit for anka flow failed the url is not reachable http status code error audit failed for cask anka flow is reachable verifying checksum for cask opera developer audit for opera developer failed the url is not reachable error audit failed for cask opera developer svn font casks audit for font abeezee failed the url is not reachable http status code audit for font advent pro failed the url is not reachable http status code audit for font allan failed the url is not reachable http status code brew cask audit debug and adobe both hang at auditing url format with no further output both return http | 1 |
2,671 | 9,198,339,296 | IssuesEvent | 2019-03-07 12:21:47 | Inovesa/Inovesa | https://api.github.com/repos/Inovesa/Inovesa | closed | Use FFTW++ to interface FFTW or remove Array.h | maintainability question | Currently, in the sequential code path FFTW is used directly using its C bindings. It could be worth switching to [FFTW++](https://github.com/dealias/fftwpp). | True | Use FFTW++ to interface FFTW or remove Array.h - Currently, in the sequential code path FFTW is used directly using its C bindings. It could be worth switching to [FFTW++](https://github.com/dealias/fftwpp). | main | use fftw to interface fftw or remove array h currently in the sequential code path fftw is used directly using its c bindings it could be worth switching to | 1 |
3,785 | 16,020,568,760 | IssuesEvent | 2021-04-20 22:21:15 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | Application to join: IP Anonymize | Maintainer application Port complete | Hello and welcome to the contrib application process! We're happy to have you :)
## Please note these 3 requirements for new contrib projects:
- [x] Include a README.md file containing license and maintainer information.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md
- [x] Include a LICENSE.txt file.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt.
- [x] If porting a Drupal 7 project, Maintain the Git history from Drupal.
## Please provide the following information:
**The name of your module, theme, or layout**
IP Anonymization
**(Optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project**
https://www.drupal.org/project/ip_anon/issues/3196175
**Post a link to your new Backdrop project under your own GitHub account (option #1)**
https://github.com/djzwerg/ip_anon
**OR (option #2) If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits**
**OR (option #3) If you do not intend to contribute code, but would like to update documentation, manage issue queues, etc, please tag an existing contrib group member so they can post their recommendation**
@jenlampton @klonos
**If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)**
YES
**If you have chosen option #3 above, do you agree to undergo this same maintainer application process again, should you decide to contribute code in the future?**
YES
<!-- Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. -->
<!-- Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
| True | Application to join: IP Anonymize - Hello and welcome to the contrib application process! We're happy to have you :)
## Please note these 3 requirements for new contrib projects:
- [x] Include a README.md file containing license and maintainer information.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md
- [x] Include a LICENSE.txt file.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt.
- [x] If porting a Drupal 7 project, Maintain the Git history from Drupal.
## Please provide the following information:
**The name of your module, theme, or layout**
IP Anonymization
**(Optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project**
https://www.drupal.org/project/ip_anon/issues/3196175
**Post a link to your new Backdrop project under your own GitHub account (option #1)**
https://github.com/djzwerg/ip_anon
**OR (option #2) If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits**
**OR (option #3) If you do not intend to contribute code, but would like to update documentation, manage issue queues, etc, please tag an existing contrib group member so they can post their recommendation**
@jenlampton @klonos
**If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)**
YES
**If you have chosen option #3 above, do you agree to undergo this same maintainer application process again, should you decide to contribute code in the future?**
YES
<!-- Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. -->
<!-- Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
| main | application to join ip anonymize hello and welcome to the contrib application process we re happy to have you please note these requirements for new contrib projects include a readme md file containing license and maintainer information you can use this example include a license txt file you can use this example if porting a drupal project maintain the git history from drupal please provide the following information the name of your module theme or layout ip anonymization optional post a link here to an issue in the drupal org queue notifying the drupal maintainers that you are working on a backdrop port of their project post a link to your new backdrop project under your own github account option or option if you have already contributed code to backdrop core or contrib projects please provide links to pull requests or commits or option if you do not intend to contribute code but would like to update documentation manage issue queues etc please tag an existing contrib group member so they can post their recommendation jenlampton klonos if you have chosen option or above do you agree to the yes if you have chosen option above do you agree to undergo this same maintainer application process again should you decide to contribute code in the future yes | 1 |
4,981 | 25,572,342,197 | IssuesEvent | 2022-11-30 18:46:45 | bazelbuild/intellij | https://api.github.com/repos/bazelbuild/intellij | closed | UI components repeated on titlebar when NewUI is enabled, on 2022.3 EAP | type: bug P1 product: IntelliJ awaiting-maintainer | ### Description of the bug:
Please refer to the attached screenshot.
Tried on
- Non-Bazel project on 2022.3 EAP NewUI enabled
- Bazel project on 2022.3 EAP NewUI disabled
- Bazel project on 2022.2 NewUI enabled
They are all good.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
_No response_
### Which Intellij IDE are you using? Please provide the specific version.
IntelliJ IDEA 2022.3 Beta (Ultimate Edition) Build #IU-223.7571.4
### What programming languages and tools are you using? Please provide specific versions.
_No response_
### What Bazel plugin version are you using?
2022.11.07.0.1-api-version-223
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?

| True | UI components repeated on titlebar when NewUI is enabled, on 2022.3 EAP - ### Description of the bug:
Please refer to the attached screenshot.
Tried on
- Non-Bazel project on 2022.3 EAP NewUI enabled
- Bazel project on 2022.3 EAP NewUI disabled
- Bazel project on 2022.2 NewUI enabled
They are all good.
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
_No response_
### Which Intellij IDE are you using? Please provide the specific version.
IntelliJ IDEA 2022.3 Beta (Ultimate Edition) Build #IU-223.7571.4
### What programming languages and tools are you using? Please provide specific versions.
_No response_
### What Bazel plugin version are you using?
2022.11.07.0.1-api-version-223
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?

| main | ui components repeated on titlebar when newui is enabled on eap description of the bug please refer to the attached screenshot tried on non bazel project on eap newui enabled bazel project on eap newui disabled bazel project on newui enabled they are all good what s the simplest easiest way to reproduce this bug please provide a minimal example if possible no response which intellij ide are you using please provide the specific version intellij idea beta ultimate edition build iu what programming languages and tools are you using please provide specific versions no response what bazel plugin version are you using api version have you found anything relevant by searching the web no response any other information logs or outputs that you want to share | 1 |
300,445 | 22,678,860,327 | IssuesEvent | 2022-07-04 08:05:05 | mercedes-benz/sechub | https://api.github.com/repos/mercedes-benz/sechub | closed | Add short information about PDS security products inside documentation | documentation pds pds-solutions | ## Situation
In `02_security_products.adoc` there is no PDS solution listed.
## Wanted
Documentation shall contain those information as well.
## Solution
Document | 1.0 | Add short information about PDS security products inside documentation - ## Situation
In `02_security_products.adoc` there is no PDS solution listed.
## Wanted
Documentation shall contain those information as well.
## Solution
Document | non_main | add short information about pds security products inside documentation situation in security products adoc there is no pds solution listed wanted documentation shall contain those information as well solution document | 0 |
3,505 | 13,683,660,137 | IssuesEvent | 2020-09-30 02:31:56 | amyjko/faculty | https://api.github.com/repos/amyjko/faculty | closed | Specify the CV data structure | maintainability | It's currently just noted as an object in the schema, but there's so much more. | True | Specify the CV data structure - It's currently just noted as an object in the schema, but there's so much more. | main | specify the cv data structure it s currently just noted as an object in the schema but there s so much more | 1 |
13,767 | 10,027,450,012 | IssuesEvent | 2019-07-17 09:15:22 | katalon-studio/katalon-studio | https://api.github.com/repos/katalon-studio/katalon-studio | closed | Unable to find alert box object in the object tree | blocker bug support-service | Thank you for helping the above issue. Please kindly find the below information.
To register device, please play any video after login the application. An account can only register 5 devices (browser does not count).
Test case : click on menu button -> click on "N" (the pink circle) -> click on "Manage" -> scroll to "Devices" -> click on "Edit" button -> click on "X" button -> click on "Confirm" button in the alert box
I can't find my alert box related objects in the object tree. I have click on "capture object" button several times but the objects still invisible. This is a web view from an iOS native app. | 1.0 | Unable to find alert box object in the object tree - Thank you for helping the above issue. Please kindly find the below information.
To register device, please play any video after login the application. An account can only register 5 devices (browser does not count).
Test case : click on menu button -> click on "N" (the pink circle) -> click on "Manage" -> scroll to "Devices" -> click on "Edit" button -> click on "X" button -> click on "Confirm" button in the alert box
I can't find my alert box related objects in the object tree. I have click on "capture object" button several times but the objects still invisible. This is a web view from an iOS native app. | non_main | unable to find alert box object in the object tree thank you for helping the above issue please kindly find the below information to register device please play any video after login the application an account can only register devices browser does not count test case click on menu button click on n the pink circle click on manage scroll to devices click on edit button click on x button click on confirm button in the alert box i can t find my alert box related objects in the object tree i have click on capture object button several times but the objects still invisible this is a web view from an ios native app | 0 |
117,795 | 4,727,723,650 | IssuesEvent | 2016-10-18 14:15:01 | pixelhumain/communecter | https://api.github.com/repos/pixelhumain/communecter | closed | throw new CTKException => un peu partout | priority 1 question | Retourne des 500 qui stoppent certains process ! On ne devrait pas les enlever? | 1.0 | throw new CTKException => un peu partout - Retourne des 500 qui stoppent certains process ! On ne devrait pas les enlever? | non_main | throw new ctkexception un peu partout retourne des qui stoppent certains process on ne devrait pas les enlever | 0 |
966 | 4,707,894,653 | IssuesEvent | 2016-10-13 21:31:10 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | No changes happening under vm_extra_config | affects_2.2 bug_report cloud vmware waiting_on_maintainer | ##### ISSUE TYPE
Bug Report
##### COMPONENT NAME
vsphere_guest module
##### ANSIBLE VERSION
N/A
##### SUMMARY
Hi
I've configured the below playbook which works fine, except for one thing....nothing is happening under the vm_extra_config section?
Is there something that I have done wrong or is this a bug?
---
- hosts: 127.0.0.1
connection: local
user: root
sudo: false
gather_facts: false
serial: 1
vars:
vcenter_hostname: xxx.xxx.xxx
esxhost: xxx.xxx.xxx.xxx
datastore: UK-xxxx
network: Web Servers
vmtemplate: WIN2K12R2-TEMPLATE
vmcluster: UK-CLUSTER
username: xxxxxxxxx
password: xxxxxxxx
folder: Labs
notes: Created by Ansible
tasks:
- name: Create VM from template
vsphere_guest:
vcenter_hostname: "{{ vcenter_hostname }}"
username: "{{ username }}"
password: "{{ password }}"
guest: "UK-ANSIBLE-TEST0{{ name }}"
from_template: yes
template_src: "{{ vmtemplate }}"
cluster: "{{ vmcluster }}"
resource_pool: "/Resources"
vm_extra_config:
notes: "{{ notes }}"
folder: "{{ folder }}"
esxi:
datacenter: UK
hostname: "{{ esxhost }}"
If I run a playbook to setup a VM from scratch (not using a template), then the vm_extra_config works. It creates the VM in the folder that I specified.
What does this mean? The vm_extra_config does not work with templates?
I did see this page: https://groups.google.com/forum/?utm_medium=email&utm_source=footer#!searchin/ansible-project/vm_extra_config/ansible-project/9Ki7tPqW2i0/vz0swUJ6CAAJ
Any ideas? | True | No changes happening under vm_extra_config - ##### ISSUE TYPE
Bug Report
##### COMPONENT NAME
vsphere_guest module
##### ANSIBLE VERSION
N/A
##### SUMMARY
Hi
I've configured the below playbook which works fine, except for one thing....nothing is happening under the vm_extra_config section?
Is there something that I have done wrong or is this a bug?
---
- hosts: 127.0.0.1
connection: local
user: root
sudo: false
gather_facts: false
serial: 1
vars:
vcenter_hostname: xxx.xxx.xxx
esxhost: xxx.xxx.xxx.xxx
datastore: UK-xxxx
network: Web Servers
vmtemplate: WIN2K12R2-TEMPLATE
vmcluster: UK-CLUSTER
username: xxxxxxxxx
password: xxxxxxxx
folder: Labs
notes: Created by Ansible
tasks:
- name: Create VM from template
vsphere_guest:
vcenter_hostname: "{{ vcenter_hostname }}"
username: "{{ username }}"
password: "{{ password }}"
guest: "UK-ANSIBLE-TEST0{{ name }}"
from_template: yes
template_src: "{{ vmtemplate }}"
cluster: "{{ vmcluster }}"
resource_pool: "/Resources"
vm_extra_config:
notes: "{{ notes }}"
folder: "{{ folder }}"
esxi:
datacenter: UK
hostname: "{{ esxhost }}"
If I run a playbook to setup a VM from scratch (not using a template), then the vm_extra_config works. It creates the VM in the folder that I specified.
What does this mean? The vm_extra_config does not work with templates?
I did see this page: https://groups.google.com/forum/?utm_medium=email&utm_source=footer#!searchin/ansible-project/vm_extra_config/ansible-project/9Ki7tPqW2i0/vz0swUJ6CAAJ
Any ideas? | main | no changes happening under vm extra config issue type bug report component name vsphere guest module ansible version n a summary hi i ve configured the below playbook which works fine except for one thing nothing is happening under the vm extra config section is there something that i have done wrong or is this a bug hosts connection local user root sudo false gather facts false serial vars vcenter hostname xxx xxx xxx esxhost xxx xxx xxx xxx datastore uk xxxx network web servers vmtemplate template vmcluster uk cluster username xxxxxxxxx password xxxxxxxx folder labs notes created by ansible tasks name create vm from template vsphere guest vcenter hostname vcenter hostname username username password password guest uk ansible name from template yes template src vmtemplate cluster vmcluster resource pool resources vm extra config notes notes folder folder esxi datacenter uk hostname esxhost if i run a playbook to setup a vm from scratch not using a template then the vm extra config works it creates the vm in the folder that i specified what does this mean the vm extra config does not work with templates i did see this page any ideas | 1 |
2,948 | 10,589,299,244 | IssuesEvent | 2019-10-09 05:38:36 | ityonemo/zigler | https://api.github.com/repos/ityonemo/zigler | closed | make sensible defaults for which zig compilation mode are possible | maintainability | and make it so that you can :config it. | True | make sensible defaults for which zig compilation mode are possible - and make it so that you can :config it. | main | make sensible defaults for which zig compilation mode are possible and make it so that you can config it | 1 |
241,113 | 26,256,646,501 | IssuesEvent | 2023-01-06 01:44:40 | belialNZ86/version-control-system | https://api.github.com/repos/belialNZ86/version-control-system | opened | WS-2021-0152 (High) detected in color-string-0.3.0.tgz | security vulnerability | ## WS-2021-0152 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>color-string-0.3.0.tgz</b></p></summary>
<p>Parser and generator for CSS color strings</p>
<p>Library home page: <a href="https://registry.npmjs.org/color-string/-/color-string-0.3.0.tgz">https://registry.npmjs.org/color-string/-/color-string-0.3.0.tgz</a></p>
<p>Path to dependency file: /version-control-system/package.json</p>
<p>Path to vulnerable library: /node_modules/color-string/package.json</p>
<p>
Dependency Hierarchy:
- css-loader-0.28.11.tgz (Root Library)
- cssnano-3.10.0.tgz
- postcss-colormin-2.2.2.tgz
- colormin-1.1.2.tgz
- color-0.11.4.tgz
- :x: **color-string-0.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular Expression Denial of Service (ReDoS) was found in color-string before 1.5.5.
<p>Publish Date: 2021-03-12
<p>URL: <a href=https://github.com/Qix-/color-string/commit/0789e21284c33d89ebc4ab4ca6f759b9375ac9d3>WS-2021-0152</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-03-12</p>
<p>Fix Resolution (color-string): 1.5.5</p>
<p>Direct dependency fix Resolution (css-loader): 1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2021-0152 (High) detected in color-string-0.3.0.tgz - ## WS-2021-0152 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>color-string-0.3.0.tgz</b></p></summary>
<p>Parser and generator for CSS color strings</p>
<p>Library home page: <a href="https://registry.npmjs.org/color-string/-/color-string-0.3.0.tgz">https://registry.npmjs.org/color-string/-/color-string-0.3.0.tgz</a></p>
<p>Path to dependency file: /version-control-system/package.json</p>
<p>Path to vulnerable library: /node_modules/color-string/package.json</p>
<p>
Dependency Hierarchy:
- css-loader-0.28.11.tgz (Root Library)
- cssnano-3.10.0.tgz
- postcss-colormin-2.2.2.tgz
- colormin-1.1.2.tgz
- color-0.11.4.tgz
- :x: **color-string-0.3.0.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular Expression Denial of Service (ReDoS) was found in color-string before 1.5.5.
<p>Publish Date: 2021-03-12
<p>URL: <a href=https://github.com/Qix-/color-string/commit/0789e21284c33d89ebc4ab4ca6f759b9375ac9d3>WS-2021-0152</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2021-03-12</p>
<p>Fix Resolution (color-string): 1.5.5</p>
<p>Direct dependency fix Resolution (css-loader): 1.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | ws high detected in color string tgz ws high severity vulnerability vulnerable library color string tgz parser and generator for css color strings library home page a href path to dependency file version control system package json path to vulnerable library node modules color string package json dependency hierarchy css loader tgz root library cssnano tgz postcss colormin tgz colormin tgz color tgz x color string tgz vulnerable library vulnerability details regular expression denial of service redos was found in color string before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution color string direct dependency fix resolution css loader step up your open source security game with mend | 0 |
4,467 | 23,279,972,709 | IssuesEvent | 2022-08-05 11:02:26 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | opened | Contrib Group Application: | Maintainer application | Hello and welcome to the contrib application process! We're happy to have you :)
**Please indicate how you intend to help the Backdrop community by joining this group**
* Option 1: I would like to contribute a project
* Option 2: I would like to maintain a project, but have nothing to contribute at this time
* Option 3: I would like to update documentation and/or triage issue queues
<!-- example: Option 1 -->
## Based on your selection above, please provide the following information:
**(option 1) The name of your module, theme, or layout**
<!-- example: Forum Access -->
## (option 1) Please note these 3 requirements for new contrib projects:
- [ ] Include a README.md file containing license and maintainer information.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md
- [ ] Include a LICENSE.txt file.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt.
- [ ] If porting a Drupal 7 project, Maintain the Git history from Drupal.
**(option 1 -- optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project**
<!-- example: https://www.drupal.org/project/forum_access/issues/3070491 -->
**Post a link to your new Backdrop project under your own GitHub account (option 1)**
<!-- example: https://github.com/jenlampton/forum_access -->
**(option 2) If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits**
https://github.com/backdrop/backdrop/pulls?q=is%3Apr+author%3Ahimanshu007-creator+is%3Aclose
**(option 3) If you do not intend to contribute code, but would like to update documentation, manage issue queues, etc, please tag an existing contrib group member so they can post their recommendation**
<!-- example: @jenlampton -->
**If you have chosen option 2 or 1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)**
YES/no
**If you have chosen option 3 above, do you agree to undergo this same maintainer application process again, should you decide to contribute code in the future?**
YES/no
<!-- (option 1) Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. -->
<!-- (option 1) Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
| True | Contrib Group Application: - Hello and welcome to the contrib application process! We're happy to have you :)
**Please indicate how you intend to help the Backdrop community by joining this group**
* Option 1: I would like to contribute a project
* Option 2: I would like to maintain a project, but have nothing to contribute at this time
* Option 3: I would like to update documentation and/or triage issue queues
<!-- example: Option 1 -->
## Based on your selection above, please provide the following information:
**(option 1) The name of your module, theme, or layout**
<!-- example: Forum Access -->
## (option 1) Please note these 3 requirements for new contrib projects:
- [ ] Include a README.md file containing license and maintainer information.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md
- [ ] Include a LICENSE.txt file.
You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt.
- [ ] If porting a Drupal 7 project, Maintain the Git history from Drupal.
**(option 1 -- optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project**
<!-- example: https://www.drupal.org/project/forum_access/issues/3070491 -->
**Post a link to your new Backdrop project under your own GitHub account (option 1)**
<!-- example: https://github.com/jenlampton/forum_access -->
**(option 2) If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits**
https://github.com/backdrop/backdrop/pulls?q=is%3Apr+author%3Ahimanshu007-creator+is%3Aclose
**(option 3) If you do not intend to contribute code, but would like to update documentation, manage issue queues, etc, please tag an existing contrib group member so they can post their recommendation**
<!-- example: @jenlampton -->
**If you have chosen option 2 or 1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)**
YES/no
**If you have chosen option 3 above, do you agree to undergo this same maintainer application process again, should you decide to contribute code in the future?**
YES/no
<!-- (option 1) Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. -->
<!-- (option 1) Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
| main | contrib group application hello and welcome to the contrib application process we re happy to have you please indicate how you intend to help the backdrop community by joining this group option i would like to contribute a project option i would like to maintain a project but have nothing to contribute at this time option i would like to update documentation and or triage issue queues based on your selection above please provide the following information option the name of your module theme or layout option please note these requirements for new contrib projects include a readme md file containing license and maintainer information you can use this example include a license txt file you can use this example if porting a drupal project maintain the git history from drupal option optional post a link here to an issue in the drupal org queue notifying the drupal maintainers that you are working on a backdrop port of their project post a link to your new backdrop project under your own github account option option if you have already contributed code to backdrop core or contrib projects please provide links to pull requests or commits option if you do not intend to contribute code but would like to update documentation manage issue queues etc please tag an existing contrib group member so they can post their recommendation if you have chosen option or above do you agree to the yes no if you have chosen option above do you agree to undergo this same maintainer application process again should you decide to contribute code in the future yes no | 1 |
3,766 | 15,827,563,921 | IssuesEvent | 2021-04-06 08:50:56 | arcticicestudio/styleguide-javascript | https://api.github.com/repos/arcticicestudio/styleguide-javascript | closed | Migrate to `@arcticicestudio/remark-preset-lint` | context-techstack context-workflow scope-compatibility scope-maintainability scope-quality type-improvement | The currently used `remark-preset-lint-arcticicestudio` package has been deprecated [during the migration into a monorepo][1] and replaced by the [new `@arcticicestudio/remark-preset-lint` package][2]. This package also [introduces support for remark `13.0.0`][3] which comes with some new features and great improvements.
To migrate to the new package the currently `remark-preset-lint-arcticicestudio` preset will be replaced by the new `@arcticicestudio/remark-preset-lint` preset in the `.remarkrc.js` file.
[1]: https://github.com/arcticicestudio/styleguide-markdown/issues/10
[2]: https://github.com/arcticicestudio/styleguide-markdown/tree/main/packages/%40arcticicestudio/remark-preset-lint
[3]: https://github.com/arcticicestudio/styleguide-markdown/issues/28
| True | Migrate to `@arcticicestudio/remark-preset-lint` - The currently used `remark-preset-lint-arcticicestudio` package has been deprecated [during the migration into a monorepo][1] and replaced by the [new `@arcticicestudio/remark-preset-lint` package][2]. This package also [introduces support for remark `13.0.0`][3] which comes with some new features and great improvements.
To migrate to the new package the currently `remark-preset-lint-arcticicestudio` preset will be replaced by the new `@arcticicestudio/remark-preset-lint` preset in the `.remarkrc.js` file.
[1]: https://github.com/arcticicestudio/styleguide-markdown/issues/10
[2]: https://github.com/arcticicestudio/styleguide-markdown/tree/main/packages/%40arcticicestudio/remark-preset-lint
[3]: https://github.com/arcticicestudio/styleguide-markdown/issues/28
| main | migrate to arcticicestudio remark preset lint the currently used remark preset lint arcticicestudio package has been deprecated and replaced by the this package also which comes with some new features and great improvements to migrate to the new package the currently remark preset lint arcticicestudio preset will be replaced by the new arcticicestudio remark preset lint preset in the remarkrc js file | 1 |
1,874 | 6,577,499,684 | IssuesEvent | 2017-09-12 01:20:34 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | docker module has wrong/misleading error handling during container creation | affects_2.0 bug_report cloud docker waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
docker module
##### ANSIBLE VERSION
```
ansible 2.0.0.2
```
and current devel
##### CONFIGURATION
standard
##### OS / ENVIRONMENT
Ubuntu 14.04
##### SUMMARY
When docker module fails to start a container for any reason it tries to fix it by pulling the image from hub and start again.
But if the image is local, module will fail with pull error instead of actual start error.
##### STEPS TO REPRODUCE
1. Create local image with `docker build` or `docker commit` and tag it as `my/test:latest`
2. Try to start container giving wrong network name:
```
docker:
name: 'test'
image: 'my/test:latest'
state: 'restarted'
net: 'bad_network_name'
```
##### EXPECTED RESULTS
```
fatal: [test-host]: FAILED! => {"changed": false, "failed": true, "msg": "Docker API Error: network bad_network_name not found"}
```
##### ACTUAL RESULTS
```
fatal: [test-host]: FAILED! => {"changed": false, "changes": ["{\"status\":\"Pulling repository docker.io/my/test\"}\r\n", "{\"errorDetail\":{\"message\":\"Error: image my/test not found\"},\"error\":\"Error: image my/test not found\"}\r\n"], "failed": true, "msg": "Unrecognized status from pull.", "status": ""}
```
[Part of code in question](https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker.py#L1658) – try/except block with do_create.
Why do we try to pull the image if we get 404 response code?
As per [docker api docs](https://docs.docker.com/engine/reference/api/docker_remote_api_v1.22) response codes for `containers/create` endpoint are too general to make such decisions:
- 201 – no error
- 404 – no such container
- 406 – impossible to attach (container not running)
- 500 – server error
| True | docker module has wrong/misleading error handling during container creation - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
docker module
##### ANSIBLE VERSION
```
ansible 2.0.0.2
```
and current devel
##### CONFIGURATION
standard
##### OS / ENVIRONMENT
Ubuntu 14.04
##### SUMMARY
When docker module fails to start a container for any reason it tries to fix it by pulling the image from hub and start again.
But if the image is local, module will fail with pull error instead of actual start error.
##### STEPS TO REPRODUCE
1. Create local image with `docker build` or `docker commit` and tag it as `my/test:latest`
2. Try to start container giving wrong network name:
```
docker:
name: 'test'
image: 'my/test:latest'
state: 'restarted'
net: 'bad_network_name'
```
##### EXPECTED RESULTS
```
fatal: [test-host]: FAILED! => {"changed": false, "failed": true, "msg": "Docker API Error: network bad_network_name not found"}
```
##### ACTUAL RESULTS
```
fatal: [test-host]: FAILED! => {"changed": false, "changes": ["{\"status\":\"Pulling repository docker.io/my/test\"}\r\n", "{\"errorDetail\":{\"message\":\"Error: image my/test not found\"},\"error\":\"Error: image my/test not found\"}\r\n"], "failed": true, "msg": "Unrecognized status from pull.", "status": ""}
```
[Part of code in question](https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker.py#L1658) – try/except block with do_create.
Why do we try to pull the image if we get 404 response code?
As per [docker api docs](https://docs.docker.com/engine/reference/api/docker_remote_api_v1.22) response codes for `containers/create` endpoint are too general to make such decisions:
- 201 – no error
- 404 – no such container
- 406 – impossible to attach (container not running)
- 500 – server error
| main | docker module has wrong misleading error handling during container creation issue type bug report component name docker module ansible version ansible and current devel configuration standard os environment ubuntu summary when docker module fails to start a container for any reason it tries to fix it by pulling the image from hub and start again but if the image is local module will fail with pull error instead of actual start error steps to reproduce create local image with docker build or docker commit and tag it as my test latest try to start container giving wrong network name docker name test image my test latest state restarted net bad network name expected results fatal failed changed false failed true msg docker api error network bad network name not found actual results fatal failed changed false changes failed true msg unrecognized status from pull status – try except block with do create why do we try to pull the image if we get response code as per response codes for containers create endpoint are too general to make such decisions – no error – no such container – impossible to attach container not running – server error | 1 |
921 | 4,622,220,856 | IssuesEvent | 2016-09-27 06:31:25 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | docker_image often fails to push with docker-py 1.10.x | affects_2.2 bug_report cloud docker in progress waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
docker_image module
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 07a76bece1) last updated 2016/09/16 16:50:58 (GMT -700)
lib/ansible/modules/core: (detached HEAD 488f082761) last updated 2016/09/16 16:51:07 (GMT -700)
lib/ansible/modules/extras: (detached HEAD 24da3602c6) last updated 2016/09/16 16:51:07 (GMT -700)
config file = /Users/rmendes/github/roles/roles-docker/docker_new_image/tests/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
Our inventory is dictionary based so:
```
hash_behaviour=merge
```
This is on a completely rebuilt laptop with one virtualenv and no docker-py at the system level.
```
pip freeze
-e git+https://github.com/ansible/ansible.git@07a76bece15d2568d7ea76e77266190652a0beec#egg=ansible
backports.ssl-match-hostname==3.5.0.1
cffi==1.8.3
cryptography==1.5
-e git+https://github.com/docker/docker-py.git@6b7a828400f46ea81374bc5764d8aa81bf38f6f7#egg=docker_py
docker-pycreds==0.2.1
enum34==1.1.6
idna==2.1
ipaddress==1.0.17
Jinja2==2.8
MarkupSafe==0.23
paramiko==2.0.2
py2-ipaddress==3.4.1
pyasn1==0.1.9
pycparser==2.14
pycrypto==2.6.1
PyYAML==3.12
requests==2.10.0
six==1.10.0
websocket-client==0.37.0
```
##### OS / ENVIRONMENT
OS X El Capitan
##### SUMMARY
Image pushes often fail with what looks like a JSON handling issue.
##### STEPS TO REPRODUCE
```
---
- name: push image
hosts: localhost
connection: local
gather_facts: False
tasks:
- name: push new image
docker_image:
name: "test-image-3"
repository: "127.0.0.1:5000/test-image-3"
tag: "role-test"
pull: False
push: True
state: present
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
Image is successfully pushed.
##### ACTUAL RESULTS
I discovered this in a role I was testing. 4/5 times I ran the role tests I got a failure like the one shown. 1/5 it worked like it should.
```
fatal: [test-image-3]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"api_version": null,
"archive_path": null,
"buildargs": null,
"cacert_path": null,
"cert_path": null,
"container_limits": null,
"debug": false,
"docker_host": null,
"dockerfile": null,
"filter_logger": false,
"force": false,
"http_timeout": null,
"key_path": null,
"load_path": null,
"name": "test-image-3",
"nocache": "False",
"path": null,
"pull": false,
"push": true,
"repository": "127.0.0.1:5000/test-image-3",
"rm": true,
"ssl_version": null,
"state": "present",
"tag": "role-test",
"timeout": null,
"tls": null,
"tls_hostname": null,
"tls_verify": null,
"use_tls": "no"
},
"module_name": "ilmn_docker_image"
},
"msg": "Error pushing image 127.0.0.1:5000/test-image-3: Extra data: line 2 column 1 - line 3 column 1 (char 64 - 128)"
```
I thought this was a docker-py issue, so I reported it there today. @shin worked with me to show it is ultimately an Ansible issue. That thread is here - docker/docker-py#1222.
Here is what he reported:
Oh, I figured out why you're seeing the issue: https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker_image.py#L428
Ansible does the decoding of data chunks itself, so it doesn't rely on our JSON parsing code, causing the issue at their level when the API sometimes sends multiple chunks at a time. I'm afraid this is something you'll have to report there, as there's little to be done on our end.
| True | docker_image often fails to push with docker-py 1.10.x - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
docker_image module
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 07a76bece1) last updated 2016/09/16 16:50:58 (GMT -700)
lib/ansible/modules/core: (detached HEAD 488f082761) last updated 2016/09/16 16:51:07 (GMT -700)
lib/ansible/modules/extras: (detached HEAD 24da3602c6) last updated 2016/09/16 16:51:07 (GMT -700)
config file = /Users/rmendes/github/roles/roles-docker/docker_new_image/tests/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
Our inventory is dictionary based so:
```
hash_behaviour=merge
```
This is on a completely rebuilt laptop with one virtualenv and no docker-py at the system level.
```
pip freeze
-e git+https://github.com/ansible/ansible.git@07a76bece15d2568d7ea76e77266190652a0beec#egg=ansible
backports.ssl-match-hostname==3.5.0.1
cffi==1.8.3
cryptography==1.5
-e git+https://github.com/docker/docker-py.git@6b7a828400f46ea81374bc5764d8aa81bf38f6f7#egg=docker_py
docker-pycreds==0.2.1
enum34==1.1.6
idna==2.1
ipaddress==1.0.17
Jinja2==2.8
MarkupSafe==0.23
paramiko==2.0.2
py2-ipaddress==3.4.1
pyasn1==0.1.9
pycparser==2.14
pycrypto==2.6.1
PyYAML==3.12
requests==2.10.0
six==1.10.0
websocket-client==0.37.0
```
##### OS / ENVIRONMENT
OS X El Capitan
##### SUMMARY
Image pushes often fail with what looks like a JSON handling issue.
##### STEPS TO REPRODUCE
```
---
- name: push image
hosts: localhost
connection: local
gather_facts: False
tasks:
- name: push new image
docker_image:
name: "test-image-3"
repository: "127.0.0.1:5000/test-image-3"
tag: "role-test"
pull: False
push: True
state: present
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
Image is successfully pushed.
##### ACTUAL RESULTS
I discovered this in a role I was testing. 4/5 times I ran the role tests I got a failure like the one shown. 1/5 it worked like it should.
```
fatal: [test-image-3]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"api_version": null,
"archive_path": null,
"buildargs": null,
"cacert_path": null,
"cert_path": null,
"container_limits": null,
"debug": false,
"docker_host": null,
"dockerfile": null,
"filter_logger": false,
"force": false,
"http_timeout": null,
"key_path": null,
"load_path": null,
"name": "test-image-3",
"nocache": "False",
"path": null,
"pull": false,
"push": true,
"repository": "127.0.0.1:5000/test-image-3",
"rm": true,
"ssl_version": null,
"state": "present",
"tag": "role-test",
"timeout": null,
"tls": null,
"tls_hostname": null,
"tls_verify": null,
"use_tls": "no"
},
"module_name": "ilmn_docker_image"
},
"msg": "Error pushing image 127.0.0.1:5000/test-image-3: Extra data: line 2 column 1 - line 3 column 1 (char 64 - 128)"
```
I thought this was a docker-py issue, so I reported it there today. @shin worked with me to show it is ultimately an Ansible issue. That thread is here - docker/docker-py#1222.
Here is what he reported:
Oh, I figured out why you're seeing the issue: https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker_image.py#L428
Ansible does the decoding of data chunks itself, so it doesn't rely on our JSON parsing code, causing the issue at their level when the API sometimes sends multiple chunks at a time. I'm afraid this is something you'll have to report there, as there's little to be done on our end.
| main | docker image often fails to push with docker py x issue type bug report component name docker image module ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file users rmendes github roles roles docker docker new image tests ansible cfg configured module search path default w o overrides configuration our inventory is dictionary based so hash behaviour merge this is on a completely rebuilt laptop with one virtualenv and no docker py at the system level pip freeze e git backports ssl match hostname cffi cryptography e git docker pycreds idna ipaddress markupsafe paramiko ipaddress pycparser pycrypto pyyaml requests six websocket client os environment os x el capitan summary image pushes often fail with what looks like a json handling issue steps to reproduce name push image hosts localhost connection local gather facts false tasks name push new image docker image name test image repository test image tag role test pull false push true state present expected results image is successfully pushed actual results i discovered this in a role i was testing times i ran the role tests i got a failure like the one shown it worked like it should fatal failed changed false failed true invocation module args api version null archive path null buildargs null cacert path null cert path null container limits null debug false docker host null dockerfile null filter logger false force false http timeout null key path null load path null name test image nocache false path null pull false push true repository test image rm true ssl version null state present tag role test timeout null tls null tls hostname null tls verify null use tls no module name ilmn docker image msg error pushing image test image extra data line column line column char i thought this was a docker py issue so i reported it there today shin worked with me to show it is ultimately an ansible issue that thread is here docker docker py here is what he reported oh i figured out why you re seeing the issue ansible does the decoding of data chunks itself so it doesn t rely on our json parsing code causing the issue at their level when the api sometimes sends multiple chunks at a time i m afraid this is something you ll have to report there as there s little to be done on our end | 1 |
832,365 | 32,077,537,433 | IssuesEvent | 2023-09-25 12:02:36 | googleapis/google-cloud-ruby | https://api.github.com/repos/googleapis/google-cloud-ruby | closed | [Nightly CI Failures] Failures detected for google-cloud-bigquery-data_policies | type: bug priority: p1 nightly failure | At 2023-09-09 09:37:43 UTC, detected failures in google-cloud-bigquery-data_policies for: test.
The CI logs can be found [here](https://github.com/googleapis/google-cloud-ruby/actions/runs/6129868852)
report_key_04c55a34a16afa461887c6706b11c026 | 1.0 | [Nightly CI Failures] Failures detected for google-cloud-bigquery-data_policies - At 2023-09-09 09:37:43 UTC, detected failures in google-cloud-bigquery-data_policies for: test.
The CI logs can be found [here](https://github.com/googleapis/google-cloud-ruby/actions/runs/6129868852)
report_key_04c55a34a16afa461887c6706b11c026 | non_main | failures detected for google cloud bigquery data policies at utc detected failures in google cloud bigquery data policies for test the ci logs can be found report key | 0 |
422,796 | 12,287,484,192 | IssuesEvent | 2020-05-09 12:25:20 | googleapis/elixir-google-api | https://api.github.com/repos/googleapis/elixir-google-api | opened | Synthesis failed for GamesManagement | autosynth failure priority: p1 type: bug | Hello! Autosynth couldn't regenerate GamesManagement. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-09 05:13:31 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
2020-05-09 05:13:31,352 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
Switched to branch 'autosynth-gamesmanagement'
2020-05-09 05:13:32 [INFO] Running synthtool
2020-05-09 05:13:32,975 autosynth > Running synthtool
2020-05-09 05:13:32 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/games_management/synth.metadata', 'synth.py', '--']
2020-05-09 05:13:32,975 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/games_management/synth.metadata', 'synth.py', '--']
2020-05-09 05:13:33,184 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-gamesmanagement
nothing to commit, working tree clean
2020-05-09 05:13:33,490 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
2020-05-09 05:13:33,949 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh GamesManagement
2020-05-09 05:13:37,536 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__
write(self.metadata_file_path)
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write
with open(outfile, "w") as fh:
FileNotFoundError: [Errno 2] No such file or directory: 'clients/games_management/synth.metadata'
2020-05-09 05:13:37 [ERROR] Synthesis failed
2020-05-09 05:13:37,563 autosynth > Synthesis failed
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main
).synthesize(base_synth_log_path)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/games_management/synth.metadata', 'synth.py', '--', 'GamesManagement']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
| 1.0 | Synthesis failed for GamesManagement - Hello! Autosynth couldn't regenerate GamesManagement. :broken_heart:
Here's the output from running `synth.py`:
```
2020-05-09 05:13:31 [INFO] logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
2020-05-09 05:13:31,352 autosynth > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/elixir-google-api
Switched to branch 'autosynth-gamesmanagement'
2020-05-09 05:13:32 [INFO] Running synthtool
2020-05-09 05:13:32,975 autosynth > Running synthtool
2020-05-09 05:13:32 [INFO] ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/games_management/synth.metadata', 'synth.py', '--']
2020-05-09 05:13:32,975 autosynth > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/games_management/synth.metadata', 'synth.py', '--']
2020-05-09 05:13:33,184 synthtool > Executing /home/kbuilder/.cache/synthtool/elixir-google-api/synth.py.
On branch autosynth-gamesmanagement
nothing to commit, working tree clean
2020-05-09 05:13:33,490 synthtool > Cloning https://github.com/googleapis/elixir-google-api.git.
2020-05-09 05:13:33,949 synthtool > Running: docker run --rm -v/home/kbuilder/.cache/synthtool/elixir-google-api:/workspace -v/var/run/docker.sock:/var/run/docker.sock -e USER_GROUP=1000:1000 -w /workspace gcr.io/cloud-devrel-public-resources/elixir19 scripts/generate_client.sh GamesManagement
2020-05-09 05:13:37,536 synthtool > No files in sources /home/kbuilder/.cache/synthtool/elixir-google-api/clients were copied. Does the source contain files?
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 102, in <module>
main()
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/tmpfs/src/github/synthtool/env/lib/python3.6/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/tmpfs/src/github/synthtool/synthtool/__main__.py", line 94, in main
spec.loader.exec_module(synth_module) # type: ignore
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 180, in __exit__
write(self.metadata_file_path)
File "/tmpfs/src/github/synthtool/synthtool/metadata.py", line 112, in write
with open(outfile, "w") as fh:
FileNotFoundError: [Errno 2] No such file or directory: 'clients/games_management/synth.metadata'
2020-05-09 05:13:37 [ERROR] Synthesis failed
2020-05-09 05:13:37,563 autosynth > Synthesis failed
Traceback (most recent call last):
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 599, in <module>
main()
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 471, in main
return _inner_main(temp_dir)
File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 549, in _inner_main
).synthesize(base_synth_log_path)
File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 118, in synthesize
synth_proc.check_returncode() # Raise an exception.
File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/subprocess.py", line 389, in check_returncode
self.stderr)
subprocess.CalledProcessError: Command '['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/games_management/synth.metadata', 'synth.py', '--', 'GamesManagement']' returned non-zero exit status 1.
```
Google internal developers can see the full log [here](https://sponge/11ff3741-9158-4831-8681-fff828f77e1a).
| non_main | synthesis failed for gamesmanagement hello autosynth couldn t regenerate gamesmanagement broken heart here s the output from running synth py logs will be written to tmpfs src github synthtool logs googleapis elixir google api autosynth logs will be written to tmpfs src github synthtool logs googleapis elixir google api switched to branch autosynth gamesmanagement running synthtool autosynth running synthtool autosynth synthtool executing home kbuilder cache synthtool elixir google api synth py on branch autosynth gamesmanagement nothing to commit working tree clean synthtool cloning synthtool running docker run rm v home kbuilder cache synthtool elixir google api workspace v var run docker sock var run docker sock e user group w workspace gcr io cloud devrel public resources scripts generate client sh gamesmanagement synthtool no files in sources home kbuilder cache synthtool elixir google api clients were copied does the source contain files traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool synthtool main py line in main file tmpfs src github synthtool env lib site packages click core py line in call return self main args kwargs file tmpfs src github synthtool env lib site packages click core py line in main rv self invoke ctx file tmpfs src github synthtool env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src github synthtool env lib site packages click core py line in invoke return callback args kwargs file tmpfs src github synthtool synthtool main py line in main spec loader exec module synth module type ignore file tmpfs src github synthtool synthtool metadata py line in exit write self metadata file path file tmpfs src github synthtool synthtool metadata py line in write with open outfile w as fh filenotfounderror no such file or directory clients games management synth metadata synthesis failed autosynth synthesis failed traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize synth proc check returncode raise an exception file home kbuilder pyenv versions lib subprocess py line in check returncode self stderr subprocess calledprocesserror command returned non zero exit status google internal developers can see the full log | 0 |
1,901 | 6,577,555,579 | IssuesEvent | 2017-09-12 01:44:04 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | ec2_asg: Usage of replace_all_instances errors out if no instances exist yet | affects_2.1 aws bug_report cloud waiting_on_maintainer | ##### Issue Type:
- Bug Report
##### Plugin Name:
ec2_asg
##### Ansible Version:
```
ansible 2.1.0
config file = /home/wholroyd/source/automation/ansible/ansible.cfg
configured module search path = ./library
```
##### Ansible Configuration:
Error is strictly related to module logic.
##### Environment:
Fedora 23 server with everything running locally
##### Summary:
When attempting to standup an ASG and attempting to use the 'replace_all_instances' flag set to true, the module fails because it can't find any running instances. I'd like to be able to make this script re-runnable so that each time it's run, I'm able to change the related launch configuration. Not all ASGs have any running instances (think of two ASGs used in A/B testing with one empty), especially at creation time.
##### Steps To Reproduce:
Create a launch configuration.
Create an asg with `replace_all_instances: true` referencing the launch configuration
``` yaml
- name: Ensure the existance of the new launch configuration
ec2_lc:
region: "{{ region }}"
profile: "{{ account }}"
state: present
name: "{{ vpc_name }}-Routing-{{ timestamp }}"
image_id: "{{ cloud_ami.results[0].ami_id }}"
instance_type: "t2.small"
instance_profile_name: "{{ account }}-routing-monitor"
key_name: "{{ account }}_{{ region }}"
assign_public_ip: yes
security_groups: "{{ existing_groups_list }}"
user_data: "{{ lookup('file', 'roles/account_infrastructure/files/infrastructure-routing.sh') }}"
- name: "Verify the virtual private cloud subnet to use"
ec2_vpc_subnet_facts:
region: "{{ region }}"
profile: "{{ account }}"
filters:
"tag:Name" : "{{ vpc_name }}-Core"
"availability-zone" : "{{ datacenter_regions_zones[region][0] }}"
register: account_cloud_vpc_subnets_a
- debug: var=account_cloud_vpc_subnets_a
- name: Ensure the existence of the new Auto Scale Group in zone A
ec2_asg:
region: "{{ region }}"
profile: "{{ account }}"
state: present
name: "{{ vpc_name }}-Routing-{{ datacenter_regions_zones[region][0] }}"
min_size: 0
max_size: 1
desired_capacity: 1
availability_zones: [ "{{ datacenter_regions_zones[region][0] }}" ]
vpc_zone_identifier: "{{ account_cloud_vpc_subnets_a.subnets[0].id }}"
launch_config_name: "{{ vpc_name }}-Routing-{{ timestamp }}"
wait_for_instances: false
replace_all_instances: true
```
##### Expected Results:
1. Creation of ASG without any issue.
2. Modification of ASG and instances replaced (if they exist)
##### Actual Results:
The module fails to create the ASG because it doesn't hold any instances yet.
```
ESTABLISH LOCAL CONNECTION FOR USER: wholroyd
localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855 `" )'
localhost PUT /tmp/tmp9W7ZbD TO /home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg
localhost EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg; rm -rf "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/" > /dev/null 2>&1'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg", line 3104, in <module>
main()
File "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg", line 3098, in main
replace_changed, asg_properties=replace(connection, module)
File "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg", line 2832, in replace
instances = props['instances']
KeyError: 'instances'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "ec2_asg"}, "parsed": false}
to retry, use: --limit @account.retry
```
| True | ec2_asg: Usage of replace_all_instances errors out if no instances exist yet - ##### Issue Type:
- Bug Report
##### Plugin Name:
ec2_asg
##### Ansible Version:
```
ansible 2.1.0
config file = /home/wholroyd/source/automation/ansible/ansible.cfg
configured module search path = ./library
```
##### Ansible Configuration:
Error is strictly related to module logic.
##### Environment:
Fedora 23 server with everything running locally
##### Summary:
When attempting to standup an ASG and attempting to use the 'replace_all_instances' flag set to true, the module fails because it can't find any running instances. I'd like to be able to make this script re-runnable so that each time it's run, I'm able to change the related launch configuration. Not all ASGs have any running instances (think of two ASGs used in A/B testing with one empty), especially at creation time.
##### Steps To Reproduce:
Create a launch configuration.
Create an asg with `replace_all_instances: true` referencing the launch configuration
``` yaml
- name: Ensure the existance of the new launch configuration
ec2_lc:
region: "{{ region }}"
profile: "{{ account }}"
state: present
name: "{{ vpc_name }}-Routing-{{ timestamp }}"
image_id: "{{ cloud_ami.results[0].ami_id }}"
instance_type: "t2.small"
instance_profile_name: "{{ account }}-routing-monitor"
key_name: "{{ account }}_{{ region }}"
assign_public_ip: yes
security_groups: "{{ existing_groups_list }}"
user_data: "{{ lookup('file', 'roles/account_infrastructure/files/infrastructure-routing.sh') }}"
- name: "Verify the virtual private cloud subnet to use"
ec2_vpc_subnet_facts:
region: "{{ region }}"
profile: "{{ account }}"
filters:
"tag:Name" : "{{ vpc_name }}-Core"
"availability-zone" : "{{ datacenter_regions_zones[region][0] }}"
register: account_cloud_vpc_subnets_a
- debug: var=account_cloud_vpc_subnets_a
- name: Ensure the existence of the new Auto Scale Group in zone A
ec2_asg:
region: "{{ region }}"
profile: "{{ account }}"
state: present
name: "{{ vpc_name }}-Routing-{{ datacenter_regions_zones[region][0] }}"
min_size: 0
max_size: 1
desired_capacity: 1
availability_zones: [ "{{ datacenter_regions_zones[region][0] }}" ]
vpc_zone_identifier: "{{ account_cloud_vpc_subnets_a.subnets[0].id }}"
launch_config_name: "{{ vpc_name }}-Routing-{{ timestamp }}"
wait_for_instances: false
replace_all_instances: true
```
##### Expected Results:
1. Creation of ASG without any issue.
2. Modification of ASG and instances replaced (if they exist)
##### Actual Results:
The module fails to create the ASG because it doesn't hold any instances yet.
```
ESTABLISH LOCAL CONNECTION FOR USER: wholroyd
localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855 `" )'
localhost PUT /tmp/tmp9W7ZbD TO /home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg
localhost EXEC /bin/sh -c 'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg; rm -rf "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/" > /dev/null 2>&1'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg", line 3104, in <module>
main()
File "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg", line 3098, in main
replace_changed, asg_properties=replace(connection, module)
File "/home/wholroyd/.ansible/tmp/ansible-tmp-1456849907.05-42662666435855/ec2_asg", line 2832, in replace
instances = props['instances']
KeyError: 'instances'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "ec2_asg"}, "parsed": false}
to retry, use: --limit @account.retry
```
| main | asg usage of replace all instances errors out if no instances exist yet issue type bug report plugin name asg ansible version ansible config file home wholroyd source automation ansible ansible cfg configured module search path library ansible configuration error is strictly related to module logic environment fedora server with everything running locally summary when attempting to standup an asg and attempting to use the replace all instances flag set to true the module fails because it can t find any running instances i d like to be able to make this script re runnable so that each time it s run i m able to change the related launch configuration not all asgs have any running instances think of two asgs used in a b testing with one empty especially at creation time steps to reproduce create a launch configuration create an asg with replace all instances true referencing the launch configuration yaml name ensure the existance of the new launch configuration lc region region profile account state present name vpc name routing timestamp image id cloud ami results ami id instance type small instance profile name account routing monitor key name account region assign public ip yes security groups existing groups list user data lookup file roles account infrastructure files infrastructure routing sh name verify the virtual private cloud subnet to use vpc subnet facts region region profile account filters tag name vpc name core availability zone datacenter regions zones register account cloud vpc subnets a debug var account cloud vpc subnets a name ensure the existence of the new auto scale group in zone a asg region region profile account state present name vpc name routing datacenter regions zones min size max size desired capacity availability zones vpc zone identifier account cloud vpc subnets a subnets id launch config name vpc name routing timestamp wait for instances false replace all instances true expected results creation of asg without any issue modification of asg and instances replaced if they exist actual results the module fails to create the asg because it doesn t hold any instances yet establish local connection for user wholroyd localhost exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp localhost put tmp to home wholroyd ansible tmp ansible tmp asg localhost exec bin sh c lang c lc all c lc messages c usr bin python home wholroyd ansible tmp ansible tmp asg rm rf home wholroyd ansible tmp ansible tmp dev null an exception occurred during task execution the full traceback is traceback most recent call last file home wholroyd ansible tmp ansible tmp asg line in main file home wholroyd ansible tmp ansible tmp asg line in main replace changed asg properties replace connection module file home wholroyd ansible tmp ansible tmp asg line in replace instances props keyerror instances fatal failed changed false failed true invocation module name asg parsed false to retry use limit account retry | 1 |
246,495 | 7,895,377,431 | IssuesEvent | 2018-06-29 02:53:11 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Update Libsim documentation to indicate that Libsim VISIT_CSG_XFORM isn't implemented. | Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 3 - Major Irritation Support Group: Any bug version: 2.8.2 | A user sent e-mail to visit-users asking how to use VISIT_CSG_XFORM. It turns out it isn't implemented, so we should update the documentation to indicate it isn't yet implemented.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 04/22/2015 06:44 pm
Original update: 08/27/2015 01:51 pm
Ticket number: 2239 | 1.0 | Update Libsim documentation to indicate that Libsim VISIT_CSG_XFORM isn't implemented. - A user sent e-mail to visit-users asking how to use VISIT_CSG_XFORM. It turns out it isn't implemented, so we should update the documentation to indicate it isn't yet implemented.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Eric Brugger
Original creation: 04/22/2015 06:44 pm
Original update: 08/27/2015 01:51 pm
Ticket number: 2239 | non_main | update libsim documentation to indicate that libsim visit csg xform isn t implemented a user sent e mail to visit users asking how to use visit csg xform it turns out it isn t implemented so we should update the documentation to indicate it isn t yet implemented redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author eric brugger original creation pm original update pm ticket number | 0 |
2,493 | 8,650,854,520 | IssuesEvent | 2018-11-27 00:17:52 | Microsoft/DirectXMesh | https://api.github.com/repos/Microsoft/DirectXMesh | closed | Publish a NuGet packge with DX12 support for Win32 desktop | maintainence | The NuGet package ``DirectXMesh_Uwp`` includes DirectX 12 support side-by-side with DirectX 11, but the ``directxmesh_desktop_2015`` only supports DirectX 11 for Windows 7 support.
I should publish a ``DirectXMesh_desktop_win10`` package that includes the DirectX 12 support for desktop apps that require Windows 10. | True | Publish a NuGet packge with DX12 support for Win32 desktop - The NuGet package ``DirectXMesh_Uwp`` includes DirectX 12 support side-by-side with DirectX 11, but the ``directxmesh_desktop_2015`` only supports DirectX 11 for Windows 7 support.
I should publish a ``DirectXMesh_desktop_win10`` package that includes the DirectX 12 support for desktop apps that require Windows 10. | main | publish a nuget packge with support for desktop the nuget package directxmesh uwp includes directx support side by side with directx but the directxmesh desktop only supports directx for windows support i should publish a directxmesh desktop package that includes the directx support for desktop apps that require windows | 1 |
1,921 | 6,586,382,279 | IssuesEvent | 2017-09-13 17:03:14 | duckduckgo/zeroclickinfo-fathead | https://api.github.com/repos/duckduckgo/zeroclickinfo-fathead | closed | MDN JavaScript: use keywords in the OpenSearch description documents for Categories | Improvement Maintainer Input Requested Programming Mission Status: Needs Community Review Status: Work In Progress Topic: JavaScript | Depends on #256 and #257
This is an idea @bfmags had!
> MDN supports [OpenSearch](http://www.opensearch.org/Home), meaning that every page has an [OpenSearch description document](http://www.opensearch.org/Specifications/OpenSearch/1.1#Examples) with keywords, etc.
This means we can use the keywords contained in there to programmatically generate categories for this Fathead.
Read the docs to learn more about [Fatheads](http://docs.duckduckhack.com/resources/fathead-overview.html).
Please see the forum topic [Improve MDN JavaScript Fathead](https://forum.duckduckhack.com/t/improve-mdn-javascript-fathead/155) for more details.
Thanks!
---
IA Page: https://duck.co/ia/view/mdnjs
maintainer: @ericedens
Forum topic: [Improve MDN JavaScript Fathead](https://forum.duckduckhack.com/t/improve-mdn-javascript-fathead/155)
This issue is part of the [Programming Mission](https://forum.duckduckhack.com/t/duckduckhack-programming-mission-overview/53): help us improve the results for [JavaScript related searches](https://forum.duckduckhack.com/t/javascript-search-overview/94)! | True | MDN JavaScript: use keywords in the OpenSearch description documents for Categories - Depends on #256 and #257
This is an idea @bfmags had!
> MDN supports [OpenSearch](http://www.opensearch.org/Home), meaning that every page has an [OpenSearch description document](http://www.opensearch.org/Specifications/OpenSearch/1.1#Examples) with keywords, etc.
This means we can use the keywords contained in there to programmatically generate categories for this Fathead.
Read the docs to learn more about [Fatheads](http://docs.duckduckhack.com/resources/fathead-overview.html).
Please see the forum topic [Improve MDN JavaScript Fathead](https://forum.duckduckhack.com/t/improve-mdn-javascript-fathead/155) for more details.
Thanks!
---
IA Page: https://duck.co/ia/view/mdnjs
maintainer: @ericedens
Forum topic: [Improve MDN JavaScript Fathead](https://forum.duckduckhack.com/t/improve-mdn-javascript-fathead/155)
This issue is part of the [Programming Mission](https://forum.duckduckhack.com/t/duckduckhack-programming-mission-overview/53): help us improve the results for [JavaScript related searches](https://forum.duckduckhack.com/t/javascript-search-overview/94)! | main | mdn javascript use keywords in the opensearch description documents for categories depends on and this is an idea bfmags had mdn supports meaning that every page has an with keywords etc this means we can use the keywords contained in there to programmatically generate categories for this fathead read the docs to learn more about please see the forum topic for more details thanks ia page maintainer ericedens forum topic this issue is part of the help us improve the results for | 1 |
3,433 | 13,207,119,735 | IssuesEvent | 2020-08-14 22:00:59 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [MultiSelect]: options aria-selected attribute always false | component: multi-select package: react status: waiting for maintainer response 💬 type: bug 🐛 | ## What package(s) are you using?
<!--
Add an x in one of the options below, for example:
- [x] package name
-->
- [ ] `carbon-components`
- [x] `carbon-components-react`
## Detailed description
> Describe in detail the issue you're having.
The `MultiSelect` component options are not changing their `aria-selected` attribute regardless of the checked status of the option
> Is this issue related to a specific component?
Yes, the `MultiSelect`
> What did you expect to happen? What happened instead? What would you like to
> see changed?
I expect that the `aria-selected` attribute on the `MultiSelect` options is set to `"true"` when the option is selected and `"false"` when the option is not selected.
> What browser are you working in?
Chrome Version 84
> What version of the Carbon Design System are you using?
7.17.0
> What offering/product do you work on? Any pressing ship or release dates we
> should be aware of?
IBM Data Explorer
## Steps to reproduce the issue
1. Go to any of the [`MultiSelect` component stories](https://react.carbondesignsystem.com/?path=/story/multiselect--default)
2. Click to expand the `MultiSelect` to display options
3. Inspect one of the options in the browser dev tools
4. Notice the `aria-selected` attribute on the option element
5. Toggle the selected state of that option by click in on it
6. Notice that the `aria-selected` attribute on the option is always set to `"false"` regardless of the selected state
## Additional information
 | True | [MultiSelect]: options aria-selected attribute always false - ## What package(s) are you using?
<!--
Add an x in one of the options below, for example:
- [x] package name
-->
- [ ] `carbon-components`
- [x] `carbon-components-react`
## Detailed description
> Describe in detail the issue you're having.
The `MultiSelect` component options are not changing their `aria-selected` attribute regardless of the checked status of the option
> Is this issue related to a specific component?
Yes, the `MultiSelect`
> What did you expect to happen? What happened instead? What would you like to
> see changed?
I expect that the `aria-selected` attribute on the `MultiSelect` options is set to `"true"` when the option is selected and `"false"` when the option is not selected.
> What browser are you working in?
Chrome Version 84
> What version of the Carbon Design System are you using?
7.17.0
> What offering/product do you work on? Any pressing ship or release dates we
> should be aware of?
IBM Data Explorer
## Steps to reproduce the issue
1. Go to any of the [`MultiSelect` component stories](https://react.carbondesignsystem.com/?path=/story/multiselect--default)
2. Click to expand the `MultiSelect` to display options
3. Inspect one of the options in the browser dev tools
4. Notice the `aria-selected` attribute on the option element
5. Toggle the selected state of that option by click in on it
6. Notice that the `aria-selected` attribute on the option is always set to `"false"` regardless of the selected state
## Additional information
 | main | options aria selected attribute always false what package s are you using add an x in one of the options below for example package name carbon components carbon components react detailed description describe in detail the issue you re having the multiselect component options are not changing their aria selected attribute regardless of the checked status of the option is this issue related to a specific component yes the multiselect what did you expect to happen what happened instead what would you like to see changed i expect that the aria selected attribute on the multiselect options is set to true when the option is selected and false when the option is not selected what browser are you working in chrome version what version of the carbon design system are you using what offering product do you work on any pressing ship or release dates we should be aware of ibm data explorer steps to reproduce the issue go to any of the click to expand the multiselect to display options inspect one of the options in the browser dev tools notice the aria selected attribute on the option element toggle the selected state of that option by click in on it notice that the aria selected attribute on the option is always set to false regardless of the selected state additional information | 1 |
404,242 | 27,455,558,094 | IssuesEvent | 2023-03-02 21:02:21 | pluralsight/pando | https://api.github.com/repos/pluralsight/pando | opened | [Docs?]: Add Popover iconOptions | documentation | ### Latest version
- [X] I have checked the latest version
### Summary 💡
#1135 added iconOptions to the popover output and needs to be added to the documentation.
(Please wait for docs rework to be completed)
### Motivation 🔦
_No response_ | 1.0 | [Docs?]: Add Popover iconOptions - ### Latest version
- [X] I have checked the latest version
### Summary 💡
#1135 added iconOptions to the popover output and needs to be added to the documentation.
(Please wait for docs rework to be completed)
### Motivation 🔦
_No response_ | non_main | add popover iconoptions latest version i have checked the latest version summary 💡 added iconoptions to the popover output and needs to be added to the documentation please wait for docs rework to be completed motivation 🔦 no response | 0 |
2,437 | 8,634,729,893 | IssuesEvent | 2018-11-22 18:05:45 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | reopened | Namespace names of tests should not contain "Test" | Area: analyzer Area: maintainability Area: naming feature | Tests should be located in the same namespace as their class under test because both belong (logically) together. | True | Namespace names of tests should not contain "Test" - Tests should be located in the same namespace as their class under test because both belong (logically) together. | main | namespace names of tests should not contain test tests should be located in the same namespace as their class under test because both belong logically together | 1 |
185 | 2,795,643,132 | IssuesEvent | 2015-05-11 23:24:49 | daemonraco/toobasic | https://api.github.com/repos/daemonraco/toobasic | closed | Database Structure Maintainer | Database Structure Maintainer enhancement next version | ## What to do
* Create a generic specification for database structure.
* Specification files must not override themselves, instead they must extend themselves into a complete spec.
* Create a mechanism to maintain a database structure updated based on the specification.
* Create non existent tables.
* Create non existent columns.
* Create non existent indexes.
* Remove unneeded tables (if allowed by configuration).
* Remove unneeded columns (if allowed by configuration).
* Remove unneeded indexes.
* Update table properties.
* Update column properties.
* Update index structure.
* `.htaccess` must hide these files.
* It should run only if the site is not flagged as installed.
* spec files must be JSON.
| True | Database Structure Maintainer - ## What to do
* Create a generic specification for database structure.
* Specification files must not override themselves, instead they must extend themselves into a complete spec.
* Create a mechanism to maintain a database structure updated based on the specification.
* Create non existent tables.
* Create non existent columns.
* Create non existent indexes.
* Remove unneeded tables (if allowed by configuration).
* Remove unneeded columns (if allowed by configuration).
* Remove unneeded indexes.
* Update table properties.
* Update column properties.
* Update index structure.
* `.htaccess` must hide these files.
* It should run only if the site is not flagged as installed.
* spec files must be JSON.
| main | database structure maintainer what to do create a generic specification for database structure specification files must not override themselves instead they must extend themselves into a complete spec create a mechanism to maintain a database structure updated based on the specification create non existent tables create non existent columns create non existent indexes remove unneeded tables if allowed by configuration remove unneeded columns if allowed by configuration remove unneeded indexes update table properties update column properties update index structure htaccess must hide these files it should run only if the site is not flagged as installed spec files must be json | 1 |
333,984 | 24,398,746,912 | IssuesEvent | 2022-10-04 22:05:42 | tendermint/spn | https://api.github.com/repos/tendermint/spn | opened | Specify a standard for `metadata` for chain | documentation launch | Determine a standard for the `metadata` format of a chain.
`metadata` represents generic bytes but the Ignite CLI and Launchpad will consider a standardized JSON format for their usage
The `metadata` must allow:
- Specify the version of Ignite CLI used to generate the genesis from Ignite information
- Specify the source URL and Hash to get information to show about a chain on the launchpad
| 1.0 | Specify a standard for `metadata` for chain - Determine a standard for the `metadata` format of a chain.
`metadata` represents generic bytes but the Ignite CLI and Launchpad will consider a standardized JSON format for their usage
The `metadata` must allow:
- Specify the version of Ignite CLI used to generate the genesis from Ignite information
- Specify the source URL and Hash to get information to show about a chain on the launchpad
| non_main | specify a standard for metadata for chain determine a standard for the metadata format of a chain metadata represents generic bytes but the ignite cli and launchpad will consider a standardized json format for their usage the metadata must allow specify the version of ignite cli used to generate the genesis from ignite information specify the source url and hash to get information to show about a chain on the launchpad | 0 |
279,142 | 30,702,454,858 | IssuesEvent | 2023-07-27 01:31:35 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | closed | CVE-2017-1000363 (High) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed | Mend: dependency security vulnerability | ## CVE-2017-1000363 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
<p>Linux kernel stable tree mirror</p>
<p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/aae4c2fa46027fd4c477372871df090c6b94f3f1">aae4c2fa46027fd4c477372871df090c6b94f3f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/lp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Linux drivers/char/lp.c Out-of-Bounds Write. Due to a missing bounds check, and the fact that parport_ptr integer is static, a 'secure boot' kernel command line adversary (can happen due to bootloader vulns, e.g. Google Nexus 6's CVE-2016-10277, where due to a vulnerability the adversary has partial control over the command line) can overflow the parport_nr array in the following code, by appending many (>LP_NO) 'lp=none' arguments to the command line.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-1000363>CVE-2017-1000363</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-1000363">https://www.linuxkernelcves.com/cves/CVE-2017-1000363</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: v4.12-rc2,v3.16.46,v3.18.55,v3.2.91,v4.1.41,v4.11.3,v4.4.70,v4.9.30</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-1000363 (High) detected in linux179e72b561d3d331c850e1a5779688d7a7de5246 - autoclosed - ## CVE-2017-1000363 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux179e72b561d3d331c850e1a5779688d7a7de5246</b></p></summary>
<p>
<p>Linux kernel stable tree mirror</p>
<p>Library home page: <a href=https://github.com/gregkh/linux.git>https://github.com/gregkh/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.1.15/commit/aae4c2fa46027fd4c477372871df090c6b94f3f1">aae4c2fa46027fd4c477372871df090c6b94f3f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/lp.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Linux drivers/char/lp.c Out-of-Bounds Write. Due to a missing bounds check, and the fact that parport_ptr integer is static, a 'secure boot' kernel command line adversary (can happen due to bootloader vulns, e.g. Google Nexus 6's CVE-2016-10277, where due to a vulnerability the adversary has partial control over the command line) can overflow the parport_nr array in the following code, by appending many (>LP_NO) 'lp=none' arguments to the command line.
<p>Publish Date: 2017-07-17
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-1000363>CVE-2017-1000363</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2017-1000363">https://www.linuxkernelcves.com/cves/CVE-2017-1000363</a></p>
<p>Release Date: 2017-07-17</p>
<p>Fix Resolution: v4.12-rc2,v3.16.46,v3.18.55,v3.2.91,v4.1.41,v4.11.3,v4.4.70,v4.9.30</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in autoclosed cve high severity vulnerability vulnerable library linux kernel stable tree mirror library home page a href found in head commit a href found in base branch master vulnerable source files drivers char lp c vulnerability details linux drivers char lp c out of bounds write due to a missing bounds check and the fact that parport ptr integer is static a secure boot kernel command line adversary can happen due to bootloader vulns e g google nexus s cve where due to a vulnerability the adversary has partial control over the command line can overflow the parport nr array in the following code by appending many lp no lp none arguments to the command line publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
848 | 4,506,693,372 | IssuesEvent | 2016-09-02 05:44:10 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | replace.py is failing with an UnboundLocalError | bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
replace
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel d99c58ee55) last updated 2016/09/01 10:10:05 (GMT -700)
lib/ansible/modules/core: (detached HEAD 7e79c59d38) last updated 2016/09/01 10:10:05 (GMT -700)
lib/ansible/modules/extras: (detached HEAD e8a5442345) last updated 2016/09/01 10:10:05 (GMT -700)
config file = /Users/jgrigonis/projects/omicia_ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
OS X controller
centos 6 target
##### SUMMARY
Seems like a bad commit
5728ef89f0764be9066fc1bf0fbbf7785e60f4cb
##### STEPS TO REPRODUCE
```
- name: fix ctypes file
replace:
dest: '/usr/local/lib/python2.7/ctypes/__init__.py'
regexp: '^( CFUNCTYPE.c_int..lambda: None.)'
replace: ' # CFUNCTYPE(c_int)(lambda: None)'
when: init.stat.exists == True
```
##### EXPECTED RESULTS
Do a replacement
##### ACTUAL RESULTS
```
{"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_Kl6JDv/ansible_module_replace.py\", line 179, in <module>\r\n main()\r\n File \"/tmp/ansible_Kl6JDv/ansible_module_replace.py\", line 173, in main\r\n module.exit_json(changed=changed, msg=msg, diff=diff)\r\nUnboundLocalError: local variable 'diff' referenced before assignment\r\n", "msg": "MODULE FAILURE"}
```
| True | replace.py is failing with an UnboundLocalError - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
replace
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel d99c58ee55) last updated 2016/09/01 10:10:05 (GMT -700)
lib/ansible/modules/core: (detached HEAD 7e79c59d38) last updated 2016/09/01 10:10:05 (GMT -700)
lib/ansible/modules/extras: (detached HEAD e8a5442345) last updated 2016/09/01 10:10:05 (GMT -700)
config file = /Users/jgrigonis/projects/omicia_ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
OS X controller
centos 6 target
##### SUMMARY
Seems like a bad commit
5728ef89f0764be9066fc1bf0fbbf7785e60f4cb
##### STEPS TO REPRODUCE
```
- name: fix ctypes file
replace:
dest: '/usr/local/lib/python2.7/ctypes/__init__.py'
regexp: '^( CFUNCTYPE.c_int..lambda: None.)'
replace: ' # CFUNCTYPE(c_int)(lambda: None)'
when: init.stat.exists == True
```
##### EXPECTED RESULTS
Do a replacement
##### ACTUAL RESULTS
```
{"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_Kl6JDv/ansible_module_replace.py\", line 179, in <module>\r\n main()\r\n File \"/tmp/ansible_Kl6JDv/ansible_module_replace.py\", line 173, in main\r\n module.exit_json(changed=changed, msg=msg, diff=diff)\r\nUnboundLocalError: local variable 'diff' referenced before assignment\r\n", "msg": "MODULE FAILURE"}
```
| main | replace py is failing with an unboundlocalerror issue type bug report component name replace ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file users jgrigonis projects omicia ansible ansible cfg configured module search path default w o overrides os environment os x controller centos target summary seems like a bad commit steps to reproduce name fix ctypes file replace dest usr local lib ctypes init py regexp cfunctype c int lambda none replace cfunctype c int lambda none when init stat exists true expected results do a replacement actual results changed false failed true module stderr module stdout traceback most recent call last r n file tmp ansible ansible module replace py line in r n main r n file tmp ansible ansible module replace py line in main r n module exit json changed changed msg msg diff diff r nunboundlocalerror local variable diff referenced before assignment r n msg module failure | 1 |
698,712 | 23,989,924,403 | IssuesEvent | 2022-09-13 23:24:08 | Thorium-Sim/thorium | https://api.github.com/repos/Thorium-Sim/thorium | opened | CAN'T SAVE FLIGHTS - HALP! | type/bug priority/high | ### Requested By: Bracken Funk
### Priority: High
### Version: 3.7.0
When I go to save a flight I get an internal server error (a white screen that says that) and I desperately need to be able to save my flights. Could you take a look at this?
### Steps to Reproduce
Try to export a flight. Can't. | 1.0 | CAN'T SAVE FLIGHTS - HALP! - ### Requested By: Bracken Funk
### Priority: High
### Version: 3.7.0
When I go to save a flight I get an internal server error (a white screen that says that) and I desperately need to be able to save my flights. Could you take a look at this?
### Steps to Reproduce
Try to export a flight. Can't. | non_main | can t save flights halp requested by bracken funk priority high version when i go to save a flight i get an internal server error a white screen that says that and i desperately need to be able to save my flights could you take a look at this steps to reproduce try to export a flight can t | 0 |
137,557 | 12,758,986,155 | IssuesEvent | 2020-06-29 04:19:47 | ocaml/ocaml | https://api.github.com/repos/ocaml/ocaml | closed | [> {typexpr} as 't ] yields a syntax error with yacc syntax | Stale bug documentation | **Original bug ID:** 3957
**Reporter:** alexbaretta
**Status:** acknowledged (set by @damiendoligez on 2006-03-29T14:34:19Z)
**Resolution:** open
**Priority:** normal
**Severity:** minor
**Category:** documentation
**Related to:** #3835
## Bug description
root@alex:~# ledit ocaml
Objective Caml version 3.09.1+dev5 (2005-12-05)
# type 'a foo = [ `Foo of 'a ];;
type 'a foo = [ `Foo of 'a ]
# let x : [> 'x foo as 'x] option = None;;
Syntax error
# #load "camlp4o.cma";;
Camlp4 Parsing version 3.09.1+dev5 (2005-12-05)
# let x : [> 'x foo as 'x] option = None;;
val x : [> ('a foo as 'a) foo ] option = None
## Additional information
The problem is hardly significant, as the camlp4 based parsers easily allow to overcome it.
| 1.0 | [> {typexpr} as 't ] yields a syntax error with yacc syntax - **Original bug ID:** 3957
**Reporter:** alexbaretta
**Status:** acknowledged (set by @damiendoligez on 2006-03-29T14:34:19Z)
**Resolution:** open
**Priority:** normal
**Severity:** minor
**Category:** documentation
**Related to:** #3835
## Bug description
root@alex:~# ledit ocaml
Objective Caml version 3.09.1+dev5 (2005-12-05)
# type 'a foo = [ `Foo of 'a ];;
type 'a foo = [ `Foo of 'a ]
# let x : [> 'x foo as 'x] option = None;;
Syntax error
# #load "camlp4o.cma";;
Camlp4 Parsing version 3.09.1+dev5 (2005-12-05)
# let x : [> 'x foo as 'x] option = None;;
val x : [> ('a foo as 'a) foo ] option = None
## Additional information
The problem is hardly significant, as the camlp4 based parsers easily allow to overcome it.
| non_main | yields a syntax error with yacc syntax original bug id reporter alexbaretta status acknowledged set by damiendoligez on resolution open priority normal severity minor category documentation related to bug description root alex ledit ocaml objective caml version type a foo type a foo let x option none syntax error load cma parsing version let x option none val x option none additional information the problem is hardly significant as the based parsers easily allow to overcome it | 0 |
4,933 | 25,351,094,641 | IssuesEvent | 2022-11-19 19:38:23 | Lissy93/dashy | https://api.github.com/repos/Lissy93/dashy | reopened | [BUG] Rename conf.yml to config.yml | 🐛 Bug 👤 Awaiting Maintainer Response | <h3 dir="auto">Environment</h3>
<p dir="auto">Self-Hosted (Docker)</p>
<h3 dir="auto">System</h3>
<p dir="auto">Firefox 107</p>
<h3 dir="auto">Version</h3>
<p dir="auto">2.1.1</p>
<h3 dir="auto">Describe the problem</h3>
<p dir="auto">If followed the guide using the YAML file, and using the template conf.yml found in the public folder of the repo, code-server will report the following error:</p>
<p dir="auto"><code class="notranslate">Property sections is not allowed</code><br>
<code class="notranslate">Property pageInfo is not allowed</code><br>
<code class="notranslate">Property appConfig is not allowed</code></p>
<p dir="auto">This is due to conf.yml being some sort of keyword in code-server/YAML and expecting certain structure: <a href="https://stackoverflow.com/questions/69493762/visual-studio-code-showing-error-property-stages-is-not-allowed-property-param" rel="nofollow">https://stackoverflow.com/questions/69493762/visual-studio-code-showing-error-property-stages-is-not-allowed-property-param</a></p>
<p dir="auto">Changing the filename to config.yml or something else will make it code-server detect it correctly</p>
<h3 dir="auto">Additional info</h3>
<p dir="auto"><em>No response</em></p>
<h3 dir="auto">Please tick the boxes</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You have explained the issue clearly, and included all relevant info</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You are using a <a href="https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions">supported</a> version of Dashy</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked that this <a href="https://github.com/Lissy93/dashy/issues?q=is%3Aissue">issue hasn't already been raised</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked the <a href="https://github.com/Lissy93/dashy/tree/master/docs#readme">docs</a> and <a href="https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting">troubleshooting</a> guide <a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/4dc60d0234fddf1d6e2f5f9306cde840c6325ef0762e8e73f6e9a0ead0b5989e/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f496d706f7274616e74212d463630393445"><img src="https://camo.githubusercontent.com/4dc60d0234fddf1d6e2f5f9306cde840c6325ef0762e8e73f6e9a0ead0b5989e/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f496d706f7274616e74212d463630393445" alt="important" data-canonical-src="https://img.shields.io/badge/Important!-F6094E" style="max-width: 100%;"></a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You agree to the <a href="https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct">code of conduct</a></li>
</ul>
<!-- unfurl begin -->
<blockquote> <img src="https://cdn.sstatic.net/Sites/stackoverflow/Img/apple-touch-icon@2.png?v=73d79a89bded" width="48" align="right" /> <div> <strong> <a href="https://stackoverflow.com/questions/69493762/visual-studio-code-showing-error-property-stages-is-not-allowed-property-param">Visual Studio Code showing error: Property Stages is not allowed, Property parameters is not allowed</a> </strong> </div> <div>I created reusable templates from Azure DevOps yaml pipeline. The build.yaml you see below is one such template.
When I push code, the pipeline successfully runs on Azure DevOps without complaining. </div> </blockquote>
<!-- unfurl end --> | True | [BUG] Rename conf.yml to config.yml - <h3 dir="auto">Environment</h3>
<p dir="auto">Self-Hosted (Docker)</p>
<h3 dir="auto">System</h3>
<p dir="auto">Firefox 107</p>
<h3 dir="auto">Version</h3>
<p dir="auto">2.1.1</p>
<h3 dir="auto">Describe the problem</h3>
<p dir="auto">If followed the guide using the YAML file, and using the template conf.yml found in the public folder of the repo, code-server will report the following error:</p>
<p dir="auto"><code class="notranslate">Property sections is not allowed</code><br>
<code class="notranslate">Property pageInfo is not allowed</code><br>
<code class="notranslate">Property appConfig is not allowed</code></p>
<p dir="auto">This is due to conf.yml being some sort of keyword in code-server/YAML and expecting certain structure: <a href="https://stackoverflow.com/questions/69493762/visual-studio-code-showing-error-property-stages-is-not-allowed-property-param" rel="nofollow">https://stackoverflow.com/questions/69493762/visual-studio-code-showing-error-property-stages-is-not-allowed-property-param</a></p>
<p dir="auto">Changing the filename to config.yml or something else will make it code-server detect it correctly</p>
<h3 dir="auto">Additional info</h3>
<p dir="auto"><em>No response</em></p>
<h3 dir="auto">Please tick the boxes</h3>
<ul class="contains-task-list">
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You have explained the issue clearly, and included all relevant info</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You are using a <a href="https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions">supported</a> version of Dashy</li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked that this <a href="https://github.com/Lissy93/dashy/issues?q=is%3Aissue">issue hasn't already been raised</a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked the <a href="https://github.com/Lissy93/dashy/tree/master/docs#readme">docs</a> and <a href="https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting">troubleshooting</a> guide <a target="_blank" rel="noopener noreferrer nofollow" href="https://camo.githubusercontent.com/4dc60d0234fddf1d6e2f5f9306cde840c6325ef0762e8e73f6e9a0ead0b5989e/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f496d706f7274616e74212d463630393445"><img src="https://camo.githubusercontent.com/4dc60d0234fddf1d6e2f5f9306cde840c6325ef0762e8e73f6e9a0ead0b5989e/68747470733a2f2f696d672e736869656c64732e696f2f62616467652f496d706f7274616e74212d463630393445" alt="important" data-canonical-src="https://img.shields.io/badge/Important!-F6094E" style="max-width: 100%;"></a></li>
<li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You agree to the <a href="https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct">code of conduct</a></li>
</ul>
<!-- unfurl begin -->
<blockquote> <img src="https://cdn.sstatic.net/Sites/stackoverflow/Img/apple-touch-icon@2.png?v=73d79a89bded" width="48" align="right" /> <div> <strong> <a href="https://stackoverflow.com/questions/69493762/visual-studio-code-showing-error-property-stages-is-not-allowed-property-param">Visual Studio Code showing error: Property Stages is not allowed, Property parameters is not allowed</a> </strong> </div> <div>I created reusable templates from Azure DevOps yaml pipeline. The build.yaml you see below is one such template.
When I push code, the pipeline successfully runs on Azure DevOps without complaining. </div> </blockquote>
<!-- unfurl end --> | main | rename conf yml to config yml environment self hosted docker system firefox version describe the problem if followed the guide using the yaml file and using the template conf yml found in the public folder of the repo code server will report the following error property sections is not allowed property pageinfo is not allowed property appconfig is not allowed this is due to conf yml being some sort of keyword in code server yaml and expecting certain structure changing the filename to config yml or something else will make it code server detect it correctly additional info no response please tick the boxes you have explained the issue clearly and included all relevant info you are using a you ve checked that this you ve checked the you agree to the i created reusable templates from azure devops yaml pipeline the build yaml you see below is one such template when i push code the pipeline successfully runs on azure devops without complaining | 1 |
2,477 | 8,639,912,635 | IssuesEvent | 2018-11-23 22:38:51 | F5OEO/rpitx | https://api.github.com/repos/F5OEO/rpitx | closed | rpitx power out & tcp socket freq change | V1 related (not maintained) | What is the actual impdance & output power of rpitx ?
is there any planning to add frequency replacement feature when running through tcp socket like gqrx ?
This is my experiment measuring power with ad8307 powermeter (DIY) https://www.youtube.com/watch?v=juT-2LdAdYg
Thanks
YD1GPW | True | rpitx power out & tcp socket freq change - What is the actual impdance & output power of rpitx ?
is there any planning to add frequency replacement feature when running through tcp socket like gqrx ?
This is my experiment measuring power with ad8307 powermeter (DIY) https://www.youtube.com/watch?v=juT-2LdAdYg
Thanks
YD1GPW | main | rpitx power out tcp socket freq change what is the actual impdance output power of rpitx is there any planning to add frequency replacement feature when running through tcp socket like gqrx this is my experiment measuring power with powermeter diy thanks | 1 |
97,139 | 8,649,370,069 | IssuesEvent | 2018-11-26 19:12:34 | WebliniaERP/webliniaerp-web | https://api.github.com/repos/WebliniaERP/webliniaerp-web | closed | Incluir soma de peso de produtos no Romaneio de Entrega | 1-melhoria 2-prioridade baixa 3- EL SHADDAI GOURMET test | Para facilitar a alocação dos volumes necessários para expedição e transporte, o cliente deseja visualizar no Romaneio de Entrega uma coluna com a informação do peso total (Bruto e Líquido) de todos os produtos | 1.0 | Incluir soma de peso de produtos no Romaneio de Entrega - Para facilitar a alocação dos volumes necessários para expedição e transporte, o cliente deseja visualizar no Romaneio de Entrega uma coluna com a informação do peso total (Bruto e Líquido) de todos os produtos | non_main | incluir soma de peso de produtos no romaneio de entrega para facilitar a alocação dos volumes necessários para expedição e transporte o cliente deseja visualizar no romaneio de entrega uma coluna com a informação do peso total bruto e líquido de todos os produtos | 0 |
1,295 | 5,518,021,290 | IssuesEvent | 2017-03-18 04:05:03 | OpenLightingProject/ola | https://api.github.com/repos/OpenLightingProject/ola | closed | build failures with gcc7 | bug Difficulty-Easy Language-C++ Maintainability OpSys-Linux | Hi,
I received Debian bug [853583](https://bugs.debian.org/853583) today, which claims that ola fails to build with GCC7.
It's not urgent (gcc7 won't be made the default until after the stretch release), but you might want to look into it. The bug report also contains instructions on how to install gcc7 from experimental on a Debian unstable system, so you can try building things. | True | build failures with gcc7 - Hi,
I received Debian bug [853583](https://bugs.debian.org/853583) today, which claims that ola fails to build with GCC7.
It's not urgent (gcc7 won't be made the default until after the stretch release), but you might want to look into it. The bug report also contains instructions on how to install gcc7 from experimental on a Debian unstable system, so you can try building things. | main | build failures with hi i received debian bug today which claims that ola fails to build with it s not urgent won t be made the default until after the stretch release but you might want to look into it the bug report also contains instructions on how to install from experimental on a debian unstable system so you can try building things | 1 |
584,863 | 17,465,915,158 | IssuesEvent | 2021-08-06 16:47:31 | wazuh/wazuh-documentation | https://api.github.com/repos/wazuh/wazuh-documentation | opened | gnupg2 dependency in Wazuh sever all-in-one step-by step installation should be replaced | installation guide priority: high type: bug | The following script
```bash
apt-get install apt-transport-https zip unzip lsb-release curl gnupg2
```
can give the following output:
```output
Reading package lists... Done
Building dependency tree
Reading state information... Done
Note, selecting 'apt' instead of 'apt-transport-https'
Package gnupg2 is not available, but is referred to by another package.
This may mean that the package is missing, has been obsoleted, or
is only available from another source
However the following packages replace it:
gpgv gpgsm gnupg-l10n gnupg dirmngr
E: Package 'gnupg2' has no installation candidate
```
This package needs to be replaced. | 1.0 | gnupg2 dependency in Wazuh sever all-in-one step-by step installation should be replaced - The following script
```bash
apt-get install apt-transport-https zip unzip lsb-release curl gnupg2
```
can give the following output:
```output
Reading package lists... Done
Building dependency tree
Reading state information... Done
Note, selecting 'apt' instead of 'apt-transport-https'
Package gnupg2 is not available, but is referred to by another package.
This may mean that the package is missing, has been obsoleted, or
is only available from another source
However the following packages replace it:
gpgv gpgsm gnupg-l10n gnupg dirmngr
E: Package 'gnupg2' has no installation candidate
```
This package needs to be replaced. | non_main | dependency in wazuh sever all in one step by step installation should be replaced the following script bash apt get install apt transport https zip unzip lsb release curl can give the following output output reading package lists done building dependency tree reading state information done note selecting apt instead of apt transport https package is not available but is referred to by another package this may mean that the package is missing has been obsoleted or is only available from another source however the following packages replace it gpgv gpgsm gnupg gnupg dirmngr e package has no installation candidate this package needs to be replaced | 0 |
546,967 | 16,022,802,438 | IssuesEvent | 2021-04-21 03:53:18 | naev/naev | https://api.github.com/repos/naev/naev | closed | Add more diversity to the lineup of NPC portraits | Priority-Low Type-Enhancement | I've been thinking lately about how, erm, white Naev's NPC lineup is. I can't work on this of course (I'm no good at drawing), but it occurs to me that the collection of NPC portraits should be diversified.
Adding more NPCs with different ethnic backgrounds would not only be a nice gesture to people of color, it would also make Naev's lineup of NPCs more varied.
Obviously no-name NPCs should get this treatment, but I would say that unique portraits should get this treatment too. Currently we just have a single black unique NPC; I think more unique NPCs should be people of color (and with a greater variety of apparent ethnic backgrounds).
I would also note that in resolving this issue, we should take care not to let our unconscious biases seep in, such as by portraying black people as criminals or some such thing (which actually seems to have seeped into the lineup of pirate NPC portraits, so working on that might be worthwhile). | 1.0 | Add more diversity to the lineup of NPC portraits - I've been thinking lately about how, erm, white Naev's NPC lineup is. I can't work on this of course (I'm no good at drawing), but it occurs to me that the collection of NPC portraits should be diversified.
Adding more NPCs with different ethnic backgrounds would not only be a nice gesture to people of color, it would also make Naev's lineup of NPCs more varied.
Obviously no-name NPCs should get this treatment, but I would say that unique portraits should get this treatment too. Currently we just have a single black unique NPC; I think more unique NPCs should be people of color (and with a greater variety of apparent ethnic backgrounds).
I would also note that in resolving this issue, we should take care not to let our unconscious biases seep in, such as by portraying black people as criminals or some such thing (which actually seems to have seeped into the lineup of pirate NPC portraits, so working on that might be worthwhile). | non_main | add more diversity to the lineup of npc portraits i ve been thinking lately about how erm white naev s npc lineup is i can t work on this of course i m no good at drawing but it occurs to me that the collection of npc portraits should be diversified adding more npcs with different ethnic backgrounds would not only be a nice gesture to people of color it would also make naev s lineup of npcs more varied obviously no name npcs should get this treatment but i would say that unique portraits should get this treatment too currently we just have a single black unique npc i think more unique npcs should be people of color and with a greater variety of apparent ethnic backgrounds i would also note that in resolving this issue we should take care not to let our unconscious biases seep in such as by portraying black people as criminals or some such thing which actually seems to have seeped into the lineup of pirate npc portraits so working on that might be worthwhile | 0 |
3,231 | 12,368,706,321 | IssuesEvent | 2020-05-18 14:13:29 | Kashdeya/Tiny-Progressions | https://api.github.com/repos/Kashdeya/Tiny-Progressions | closed | Using Pouch to Reproduce Items Infinitely | Version not Maintainted | First I put the items I want to brush into the pouch and keep the bag interface open. Then I throw the bag out with the Q key. I take the items out of the interface. Then I pick up the bag. You will find another item in the bag. | True | Using Pouch to Reproduce Items Infinitely - First I put the items I want to brush into the pouch and keep the bag interface open. Then I throw the bag out with the Q key. I take the items out of the interface. Then I pick up the bag. You will find another item in the bag. | main | using pouch to reproduce items infinitely first i put the items i want to brush into the pouch and keep the bag interface open then i throw the bag out with the q key i take the items out of the interface then i pick up the bag you will find another item in the bag | 1 |
561,975 | 16,628,668,146 | IssuesEvent | 2021-06-03 13:02:29 | GoogleCloudPlatform/python-docs-samples | https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples | opened | memorystore.redis.cloud_run_deployment.e2e_test: test_end_to_end failed | flakybot: issue priority: p1 type: bug | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 04087bcd872947e56d965c991ef60ac01b2c5c07
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1b0b403b-6b92-419e-86c3-a86b5776814d), [Sponge](http://sponge2/1b0b403b-6b92-419e-86c3-a86b5776814d)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/memorystore/redis/cloud_run_deployment/e2e_test.py", line 181, in test_end_to_end
response = request.urlopen(req)
File "/usr/local/lib/python3.9/urllib/request.py", line 214, in urlopen
return opener.open(url, data, timeout)
File "/usr/local/lib/python3.9/urllib/request.py", line 523, in open
response = meth(req, response)
File "/usr/local/lib/python3.9/urllib/request.py", line 632, in http_response
response = self.parent.error(
File "/usr/local/lib/python3.9/urllib/request.py", line 561, in error
return self._call_chain(*args)
File "/usr/local/lib/python3.9/urllib/request.py", line 494, in _call_chain
result = func(*args)
File "/usr/local/lib/python3.9/urllib/request.py", line 641, in http_error_default
raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 500: Internal Server Error</pre></details> | 1.0 | memorystore.redis.cloud_run_deployment.e2e_test: test_end_to_end failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 04087bcd872947e56d965c991ef60ac01b2c5c07
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/1b0b403b-6b92-419e-86c3-a86b5776814d), [Sponge](http://sponge2/1b0b403b-6b92-419e-86c3-a86b5776814d)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/memorystore/redis/cloud_run_deployment/e2e_test.py", line 181, in test_end_to_end
response = request.urlopen(req)
File "/usr/local/lib/python3.9/urllib/request.py", line 214, in urlopen
return opener.open(url, data, timeout)
File "/usr/local/lib/python3.9/urllib/request.py", line 523, in open
response = meth(req, response)
File "/usr/local/lib/python3.9/urllib/request.py", line 632, in http_response
response = self.parent.error(
File "/usr/local/lib/python3.9/urllib/request.py", line 561, in error
return self._call_chain(*args)
File "/usr/local/lib/python3.9/urllib/request.py", line 494, in _call_chain
result = func(*args)
File "/usr/local/lib/python3.9/urllib/request.py", line 641, in http_error_default
raise HTTPError(req.full_url, code, msg, hdrs, fp)
urllib.error.HTTPError: HTTP Error 500: Internal Server Error</pre></details> | non_main | memorystore redis cloud run deployment test test end to end failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace memorystore redis cloud run deployment test py line in test end to end response request urlopen req file usr local lib urllib request py line in urlopen return opener open url data timeout file usr local lib urllib request py line in open response meth req response file usr local lib urllib request py line in http response response self parent error file usr local lib urllib request py line in error return self call chain args file usr local lib urllib request py line in call chain result func args file usr local lib urllib request py line in http error default raise httperror req full url code msg hdrs fp urllib error httperror http error internal server error | 0 |
4,725 | 24,380,930,686 | IssuesEvent | 2022-10-04 07:46:05 | rustsec/advisory-db | https://api.github.com/repos/rustsec/advisory-db | closed | `badge` is unmaintained | Unmaintained | The [`badge`](https://crates.io/crates/badge) crate is unmaintained and will not receive further updates, as the [code has been removed from the repository](https://github.com/rust-lang/docs.rs/commit/94f3bba6815412bc4672621c4690a93e656486c7).
It is no longer used by the authors and therefore will not receive any updates: https://github.com/rust-lang/docs.rs/issues/1813#issuecomment-1232875809 | True | `badge` is unmaintained - The [`badge`](https://crates.io/crates/badge) crate is unmaintained and will not receive further updates, as the [code has been removed from the repository](https://github.com/rust-lang/docs.rs/commit/94f3bba6815412bc4672621c4690a93e656486c7).
It is no longer used by the authors and therefore will not receive any updates: https://github.com/rust-lang/docs.rs/issues/1813#issuecomment-1232875809 | main | badge is unmaintained the crate is unmaintained and will not receive further updates as the it is no longer used by the authors and therefore will not receive any updates | 1 |
799 | 4,417,051,180 | IssuesEvent | 2016-08-15 01:43:36 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | replace module only replaces last match | bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
replace
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 3c65c03a67) last updated 2016/08/14 16:26:39 (GMT -500)
lib/ansible/modules/core: (detached HEAD decb2ec9fa) last updated 2016/08/14 16:27:00 (GMT -500)
lib/ansible/modules/extras: (detached HEAD 61d5fe148c) last updated 2016/08/14 16:27:13 (GMT -500)
config file = /home/nipsy/.ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
[defaults]
hostfile = ~/.ansible/hosts
host_key_checking = False
[ssh_connection]
ssh_args = -o ControlMaster=auto -o ControlPersist=8h -o Compression=yes
control_path = /dev/shm/%%r@%%h:%%p
scp_if_ssh=True
```
##### OS / ENVIRONMENT
Debian/testing (packaged ansible (currently 2.1.1.0) behaves the same as most recent dev branch)
##### SUMMARY
The replace module is only changing the last line matched by the supplied regex, instead of all matches.
##### STEPS TO REPRODUCE
Create *test.list*
```
deb http://us.archive.ubuntu.com/ubuntu lucid main restricted
deb-src http://us.archive.ubuntu.com/ubuntu lucid main restricted
```
Create *test.yaml*
```
---
- hosts: all
tasks:
# fix outdated Ubuntu repos
- name: fix outdated Ubuntu repos
replace: dest=/home/nipsy/tmp/test.list regexp='^([^#]+)us\.archive\.ubuntu\.com(.*)$' replace='\1old-releases.ubuntu.com\2' backup=yes
```
Run:
```
$ ansible-playbook -vi localhost, test.yaml
Using /home/nipsy/.ansible.cfg as config file
PLAY [all] *********************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [fix outdated Ubuntu repos] ***********************************************
changed: [localhost] => {"changed": true, "msg": "1 replacements made"}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
##### EXPECTED RESULTS
```
$ cat test.list
deb http://old-releases.ubuntu.com/ubuntu lucid main restricted
deb-src http://old-releases.ubuntu.com/ubuntu lucid main restricted
```
##### ACTUAL RESULTS
```
$ cat test.list
deb http://us.archive.ubuntu.com/ubuntu lucid main restricted
deb-src http://old-releases.ubuntu.com/ubuntu lucid main restricted
``` | True | replace module only replaces last match - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
replace
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel 3c65c03a67) last updated 2016/08/14 16:26:39 (GMT -500)
lib/ansible/modules/core: (detached HEAD decb2ec9fa) last updated 2016/08/14 16:27:00 (GMT -500)
lib/ansible/modules/extras: (detached HEAD 61d5fe148c) last updated 2016/08/14 16:27:13 (GMT -500)
config file = /home/nipsy/.ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
[defaults]
hostfile = ~/.ansible/hosts
host_key_checking = False
[ssh_connection]
ssh_args = -o ControlMaster=auto -o ControlPersist=8h -o Compression=yes
control_path = /dev/shm/%%r@%%h:%%p
scp_if_ssh=True
```
##### OS / ENVIRONMENT
Debian/testing (packaged ansible (currently 2.1.1.0) behaves the same as most recent dev branch)
##### SUMMARY
The replace module is only changing the last line matched by the supplied regex, instead of all matches.
##### STEPS TO REPRODUCE
Create *test.list*
```
deb http://us.archive.ubuntu.com/ubuntu lucid main restricted
deb-src http://us.archive.ubuntu.com/ubuntu lucid main restricted
```
Create *test.yaml*
```
---
- hosts: all
tasks:
# fix outdated Ubuntu repos
- name: fix outdated Ubuntu repos
replace: dest=/home/nipsy/tmp/test.list regexp='^([^#]+)us\.archive\.ubuntu\.com(.*)$' replace='\1old-releases.ubuntu.com\2' backup=yes
```
Run:
```
$ ansible-playbook -vi localhost, test.yaml
Using /home/nipsy/.ansible.cfg as config file
PLAY [all] *********************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [fix outdated Ubuntu repos] ***********************************************
changed: [localhost] => {"changed": true, "msg": "1 replacements made"}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
##### EXPECTED RESULTS
```
$ cat test.list
deb http://old-releases.ubuntu.com/ubuntu lucid main restricted
deb-src http://old-releases.ubuntu.com/ubuntu lucid main restricted
```
##### ACTUAL RESULTS
```
$ cat test.list
deb http://us.archive.ubuntu.com/ubuntu lucid main restricted
deb-src http://old-releases.ubuntu.com/ubuntu lucid main restricted
``` | main | replace module only replaces last match issue type bug report component name replace ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file home nipsy ansible cfg configured module search path default w o overrides configuration hostfile ansible hosts host key checking false ssh args o controlmaster auto o controlpersist o compression yes control path dev shm r h p scp if ssh true os environment debian testing packaged ansible currently behaves the same as most recent dev branch summary the replace module is only changing the last line matched by the supplied regex instead of all matches steps to reproduce create test list deb lucid main restricted deb src lucid main restricted create test yaml hosts all tasks fix outdated ubuntu repos name fix outdated ubuntu repos replace dest home nipsy tmp test list regexp us archive ubuntu com replace releases ubuntu com backup yes run ansible playbook vi localhost test yaml using home nipsy ansible cfg as config file play task ok task changed changed true msg replacements made play recap localhost ok changed unreachable failed expected results cat test list deb lucid main restricted deb src lucid main restricted actual results cat test list deb lucid main restricted deb src lucid main restricted | 1 |
482 | 3,756,796,691 | IssuesEvent | 2016-03-13 15:50:31 | espeak-ng/espeak-ng | https://api.github.com/repos/espeak-ng/espeak-ng | closed | Move the current voice definition files to language definition files. | feature languages maintainability resolved/fixed | The voice files are closer to language files. These languages should be organised by the closest ISO 639-5 language family code (e.g. `voices/europe/cy` moving to `language/cel/cy` -- Celtic/Welsh).
Additionally, the languages should be BCP 47 compliant (e.g. `en-GB-scotland` instead of `en-sc`). Where extensions are needed, the private use tags from Cainteoir Text-to-Speech should be used. These should be described in a privateuse.dat file included in the `espeak-ng` project. | True | Move the current voice definition files to language definition files. - The voice files are closer to language files. These languages should be organised by the closest ISO 639-5 language family code (e.g. `voices/europe/cy` moving to `language/cel/cy` -- Celtic/Welsh).
Additionally, the languages should be BCP 47 compliant (e.g. `en-GB-scotland` instead of `en-sc`). Where extensions are needed, the private use tags from Cainteoir Text-to-Speech should be used. These should be described in a privateuse.dat file included in the `espeak-ng` project. | main | move the current voice definition files to language definition files the voice files are closer to language files these languages should be organised by the closest iso language family code e g voices europe cy moving to language cel cy celtic welsh additionally the languages should be bcp compliant e g en gb scotland instead of en sc where extensions are needed the private use tags from cainteoir text to speech should be used these should be described in a privateuse dat file included in the espeak ng project | 1 |
367,914 | 10,863,099,818 | IssuesEvent | 2019-11-14 14:32:48 | ubtue/DatenProbleme | https://api.github.com/repos/ubtue/DatenProbleme | opened | ISSN 1552-3594 Criminal Justice and Behavior 935 mkri doppelt, Online First | high priority | Die Verdoppelung des Tags 935 mkri wurde bereits bei den anderen getesteten Zeitschriften vermerkt. Ich gehe davon aus, dass das global gelöst wird und ich es nicht jedes mal wieder erwähnen muss. Wenn nicht, bitte Rückmeldung.
Auch hier wieder Online First Artikel. Da diese jedoch beim Weiterverabeiten entfernt werden, ist dies kein Problem.
Bedenken habe ich bei den Rezensionen. Leider gibt es in dem getesteten Heft und auch unter den Online First keine. Sie sind auch selten, kommen aber in anderen Heften vor. Z. B. in 46.2019,9
https://journals.sagepub.com/toc/cjbb/46/9
Wenn keine Gelegenheit besteht, ein solches Heft zu testen, müssen wir es auf gut Glück wagen. | 1.0 | ISSN 1552-3594 Criminal Justice and Behavior 935 mkri doppelt, Online First - Die Verdoppelung des Tags 935 mkri wurde bereits bei den anderen getesteten Zeitschriften vermerkt. Ich gehe davon aus, dass das global gelöst wird und ich es nicht jedes mal wieder erwähnen muss. Wenn nicht, bitte Rückmeldung.
Auch hier wieder Online First Artikel. Da diese jedoch beim Weiterverabeiten entfernt werden, ist dies kein Problem.
Bedenken habe ich bei den Rezensionen. Leider gibt es in dem getesteten Heft und auch unter den Online First keine. Sie sind auch selten, kommen aber in anderen Heften vor. Z. B. in 46.2019,9
https://journals.sagepub.com/toc/cjbb/46/9
Wenn keine Gelegenheit besteht, ein solches Heft zu testen, müssen wir es auf gut Glück wagen. | non_main | issn criminal justice and behavior mkri doppelt online first die verdoppelung des tags mkri wurde bereits bei den anderen getesteten zeitschriften vermerkt ich gehe davon aus dass das global gelöst wird und ich es nicht jedes mal wieder erwähnen muss wenn nicht bitte rückmeldung auch hier wieder online first artikel da diese jedoch beim weiterverabeiten entfernt werden ist dies kein problem bedenken habe ich bei den rezensionen leider gibt es in dem getesteten heft und auch unter den online first keine sie sind auch selten kommen aber in anderen heften vor z b in wenn keine gelegenheit besteht ein solches heft zu testen müssen wir es auf gut glück wagen | 0 |
96,051 | 3,963,799,425 | IssuesEvent | 2016-05-02 21:44:26 | GalliumOS/galliumos-distro | https://api.github.com/repos/GalliumOS/galliumos-distro | opened | Update artwork in repo and packages | priority:high | https://github.com/GalliumOS/galliumos-artwork needs to be updated with contents of https://github.com/GalliumOS/galliumos-distro/files/198163/GalliumOSLogos.zip
Some of the new assets need to go into various packages.
* default desktop background (be sure to preserve the current as an option)
* greeter background
* Xorg splash screen
* GRUB background (this is done in the image builder code, not an installable pkg)
* Other places?
| 1.0 | Update artwork in repo and packages - https://github.com/GalliumOS/galliumos-artwork needs to be updated with contents of https://github.com/GalliumOS/galliumos-distro/files/198163/GalliumOSLogos.zip
Some of the new assets need to go into various packages.
* default desktop background (be sure to preserve the current as an option)
* greeter background
* Xorg splash screen
* GRUB background (this is done in the image builder code, not an installable pkg)
* Other places?
| non_main | update artwork in repo and packages needs to be updated with contents of some of the new assets need to go into various packages default desktop background be sure to preserve the current as an option greeter background xorg splash screen grub background this is done in the image builder code not an installable pkg other places | 0 |
550 | 3,984,984,933 | IssuesEvent | 2016-05-07 15:32:38 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | opened | Travis: Failing Builds | awaiting maintainer feedback travis | Looks like Travis builds are still failing - 10.11 builds exit properly but still show up as a failure.
Additionally, caching is not working properly.
Made a few changes to `.travis.yml` as directed by support, but doesn't seem to have the desired effect. | True | Travis: Failing Builds - Looks like Travis builds are still failing - 10.11 builds exit properly but still show up as a failure.
Additionally, caching is not working properly.
Made a few changes to `.travis.yml` as directed by support, but doesn't seem to have the desired effect. | main | travis failing builds looks like travis builds are still failing builds exit properly but still show up as a failure additionally caching is not working properly made a few changes to travis yml as directed by support but doesn t seem to have the desired effect | 1 |
1,503 | 6,516,095,463 | IssuesEvent | 2017-08-27 02:17:17 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | closed | uninstall launchctl runs without sudo | awaiting maintainer feedback | #### General troubleshooting steps
- [x] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue.
- [X] None of the templates was appropriate for my issue, or I’m not sure.
- [X] I ran `brew update-reset && brew update` and retried my command.
- [X] I ran `brew doctor`, fixed as many issues as possible and retried my command.
- [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md).
#### Description of issue
- Installed docker through `brew cask install docker`.
- fiddled around with docker, then decided to remove it.
- launched `brew cask uninstall docker`, got an error.
#### Output of your command with `--verbose --debug`
```shell
# brew cask uninstall docker --verbose --debug
==> Uninstalling Cask docker
==> Uninstalling Cask docker
==> Un-installing artifacts
==> Determining which artifacts are present in Cask docker
==> 3 artifact/s defined
#<Hbc::Artifact::Uninstall:0x007fc15a898eb0>
#<Hbc::Artifact::App:0x007fc15a898e10>
#<Hbc::Artifact::Zap:0x007fc15a898ca8>
==> Un-installing artifact of class Hbc::Artifact::Uninstall
==> Running uninstall process for docker; your password may be necessary
==> Removing launchctl service com.docker.helper
==> Executing: ["/bin/launchctl", "list", "com.docker.helper"]
==> Executing: ["/bin/launchctl", "unload", "-w", "--", "com.docker.helper"]
==> /Users/joe/Library/Containers/com.docker.helper/Data: Path not searched for services
==> /Users/joe/Library/Containers/com.docker.helper/Container.plist: Invalid property list
==> Executing: ["/bin/rm", "-f", "--", "com.docker.helper"]
==> rm: com.docker.helper: is a directory
Error: Command failed to execute!
==> Failed command:
/bin/rm -f -- com.docker.helper
==> Standard Output of failed command:
==> Standard Error of failed command:
rm: com.docker.helper: is a directory
==> Exit status of failed command:
#<Process::Status: pid 67320 exit 1>
Error: Uninstall incomplete.
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:12:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/abstract_command.rb:35:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:98:in `run_command'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:168:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run'
/usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask'
/usr/local/Homebrew/Library/Homebrew/brew.rb:93:in `<main>'
Error: Kernel.exit
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `exit'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `rescue in run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:156:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run'
/usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask'
/usr/local/Homebrew/Library/Homebrew/brew.rb:93:in `<main>'
```
#### Output of `brew cask doctor`
```shell
brew cask doctor
==> Homebrew-Cask Version
Homebrew-Cask 1.2.3-121-gb552908490
caskroom/homebrew-cask (git revision b79d50; last commit 2017-07-02)
==> Homebrew-Cask Install Location
<NONE>
==> Homebrew-Cask Staging Location
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads
~/Library/Caches/Homebrew/Cask
==> Homebrew-Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3653 casks)
==> Contents of $LOAD_PATH
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Environment Variables
LANG="fr_FR.UTF-8"
LC_CTYPE="fr_FR.UTF-8"
PATH="/usr/local/opt/ruby/bin:/usr/local/bin:/usr/local/sbin:/usr/bin:/bin:/usr/sbin:/sbin:~/.antigen/bundles/robbyrussell/oh-my-zsh/lib:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/git:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/python:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/osx:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/pip:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/lein:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/command-not-found:~/.antigen/bundles/zsh-users/zsh-syntax-highlighting:/usr/local/Homebrew/Library/Homebrew/shims/scm"
SHELL="/usr/local/bin/zsh"
```
Would be happy to help if needed! | True | uninstall launchctl runs without sudo - #### General troubleshooting steps
- [x] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue.
- [X] None of the templates was appropriate for my issue, or I’m not sure.
- [X] I ran `brew update-reset && brew update` and retried my command.
- [X] I ran `brew doctor`, fixed as many issues as possible and retried my command.
- [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md).
#### Description of issue
- Installed docker through `brew cask install docker`.
- fiddled around with docker, then decided to remove it.
- launched `brew cask uninstall docker`, got an error.
#### Output of your command with `--verbose --debug`
```shell
# brew cask uninstall docker --verbose --debug
==> Uninstalling Cask docker
==> Uninstalling Cask docker
==> Un-installing artifacts
==> Determining which artifacts are present in Cask docker
==> 3 artifact/s defined
#<Hbc::Artifact::Uninstall:0x007fc15a898eb0>
#<Hbc::Artifact::App:0x007fc15a898e10>
#<Hbc::Artifact::Zap:0x007fc15a898ca8>
==> Un-installing artifact of class Hbc::Artifact::Uninstall
==> Running uninstall process for docker; your password may be necessary
==> Removing launchctl service com.docker.helper
==> Executing: ["/bin/launchctl", "list", "com.docker.helper"]
==> Executing: ["/bin/launchctl", "unload", "-w", "--", "com.docker.helper"]
==> /Users/joe/Library/Containers/com.docker.helper/Data: Path not searched for services
==> /Users/joe/Library/Containers/com.docker.helper/Container.plist: Invalid property list
==> Executing: ["/bin/rm", "-f", "--", "com.docker.helper"]
==> rm: com.docker.helper: is a directory
Error: Command failed to execute!
==> Failed command:
/bin/rm -f -- com.docker.helper
==> Standard Output of failed command:
==> Standard Error of failed command:
rm: com.docker.helper: is a directory
==> Exit status of failed command:
#<Process::Status: pid 67320 exit 1>
Error: Uninstall incomplete.
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/uninstall.rb:12:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/abstract_command.rb:35:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:98:in `run_command'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:168:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run'
/usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask'
/usr/local/Homebrew/Library/Homebrew/brew.rb:93:in `<main>'
Error: Kernel.exit
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `exit'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:173:in `rescue in run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:156:in `run'
/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:132:in `run'
/usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask'
/usr/local/Homebrew/Library/Homebrew/brew.rb:93:in `<main>'
```
#### Output of `brew cask doctor`
```shell
brew cask doctor
==> Homebrew-Cask Version
Homebrew-Cask 1.2.3-121-gb552908490
caskroom/homebrew-cask (git revision b79d50; last commit 2017-07-02)
==> Homebrew-Cask Install Location
<NONE>
==> Homebrew-Cask Staging Location
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads
~/Library/Caches/Homebrew/Cask
==> Homebrew-Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3653 casks)
==> Contents of $LOAD_PATH
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Environment Variables
LANG="fr_FR.UTF-8"
LC_CTYPE="fr_FR.UTF-8"
PATH="/usr/local/opt/ruby/bin:/usr/local/bin:/usr/local/sbin:/usr/bin:/bin:/usr/sbin:/sbin:~/.antigen/bundles/robbyrussell/oh-my-zsh/lib:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/git:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/python:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/osx:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/pip:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/lein:~/.antigen/bundles/robbyrussell/oh-my-zsh/plugins/command-not-found:~/.antigen/bundles/zsh-users/zsh-syntax-highlighting:/usr/local/Homebrew/Library/Homebrew/shims/scm"
SHELL="/usr/local/bin/zsh"
```
Would be happy to help if needed! | main | uninstall launchctl runs without sudo general troubleshooting steps i have checked the instructions for or before opening the issue none of the templates was appropriate for my issue or i’m not sure i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i understand that description of issue installed docker through brew cask install docker fiddled around with docker then decided to remove it launched brew cask uninstall docker got an error output of your command with verbose debug shell brew cask uninstall docker verbose debug uninstalling cask docker uninstalling cask docker un installing artifacts determining which artifacts are present in cask docker artifact s defined un installing artifact of class hbc artifact uninstall running uninstall process for docker your password may be necessary removing launchctl service com docker helper executing executing users joe library containers com docker helper data path not searched for services users joe library containers com docker helper container plist invalid property list executing rm com docker helper is a directory error command failed to execute failed command bin rm f com docker helper standard output of failed command standard error of failed command rm com docker helper is a directory exit status of failed command error uninstall incomplete usr local homebrew library homebrew cask lib hbc cli uninstall rb in run usr local homebrew library homebrew cask lib hbc cli abstract command rb in run usr local homebrew library homebrew cask lib hbc cli rb in run command usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in error kernel exit usr local homebrew library homebrew cask lib hbc cli rb in exit usr local homebrew library homebrew cask lib hbc cli rb in rescue in run usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cask lib hbc cli rb in run usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in output of brew cask doctor shell brew cask doctor homebrew cask version homebrew cask caskroom homebrew cask git revision last commit homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask cached downloads library caches homebrew cask homebrew cask taps usr local homebrew library taps caskroom homebrew cask casks contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal environment variables lang fr fr utf lc ctype fr fr utf path usr local opt ruby bin usr local bin usr local sbin usr bin bin usr sbin sbin antigen bundles robbyrussell oh my zsh lib antigen bundles robbyrussell oh my zsh plugins git antigen bundles robbyrussell oh my zsh plugins python antigen bundles robbyrussell oh my zsh plugins osx antigen bundles robbyrussell oh my zsh plugins pip antigen bundles robbyrussell oh my zsh plugins lein antigen bundles robbyrussell oh my zsh plugins command not found antigen bundles zsh users zsh syntax highlighting usr local homebrew library homebrew shims scm shell usr local bin zsh would be happy to help if needed | 1 |
4,439 | 23,060,460,372 | IssuesEvent | 2022-07-25 09:27:30 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | make MDAnalysisTests conda package "noarch" instead of arch/os specific | maintainability Continuous Integration | ## Is your feature request related to a problem? ##
The MDAnalysisTests conda package at https://github.com/conda-forge/mdanalysistests-feedstock is created as a OS and architecture-dependent package even though it only contains Python code. This leads to delays in updating the package (e.g., for new archs such as osx-arm64).
## Describe the solution you'd like ##
Publish as a **noarch** package.
## Describe alternatives you've considered ##
Keep as is for now, it's working right now.
## Additional context ##
* initial discussion https://github.com/conda-forge/mdanalysistests-feedstock/pull/31
* PR https://github.com/conda-forge/mdanalysistests-feedstock/pull/33 implementing the noarch package | True | make MDAnalysisTests conda package "noarch" instead of arch/os specific - ## Is your feature request related to a problem? ##
The MDAnalysisTests conda package at https://github.com/conda-forge/mdanalysistests-feedstock is created as a OS and architecture-dependent package even though it only contains Python code. This leads to delays in updating the package (e.g., for new archs such as osx-arm64).
## Describe the solution you'd like ##
Publish as a **noarch** package.
## Describe alternatives you've considered ##
Keep as is for now, it's working right now.
## Additional context ##
* initial discussion https://github.com/conda-forge/mdanalysistests-feedstock/pull/31
* PR https://github.com/conda-forge/mdanalysistests-feedstock/pull/33 implementing the noarch package | main | make mdanalysistests conda package noarch instead of arch os specific is your feature request related to a problem the mdanalysistests conda package at is created as a os and architecture dependent package even though it only contains python code this leads to delays in updating the package e g for new archs such as osx describe the solution you d like publish as a noarch package describe alternatives you ve considered keep as is for now it s working right now additional context initial discussion pr implementing the noarch package | 1 |
4,554 | 11,348,404,413 | IssuesEvent | 2020-01-24 00:16:26 | TerriaJS/terriajs | https://api.github.com/repos/TerriaJS/terriajs | closed | Changing coord presentation fails in mobx | New Model Architecture T-Bug | Clicking on the coord bar in a mobx app doesn't do anything. Throws an error related to `toggleUseProjection` method.

| 1.0 | Changing coord presentation fails in mobx - Clicking on the coord bar in a mobx app doesn't do anything. Throws an error related to `toggleUseProjection` method.

| non_main | changing coord presentation fails in mobx clicking on the coord bar in a mobx app doesn t do anything throws an error related to toggleuseprojection method | 0 |
4,700 | 24,257,063,100 | IssuesEvent | 2022-09-27 18:45:26 | coq-community/manifesto | https://api.github.com/repos/coq-community/manifesto | opened | Volunteer interim maintainers needed for VsCoq | change-maintainer | The Coq core team and Coq-community are looking for volunteer interim maintainers of the [VsCoq](github.com/coq-community/vscoq) project.
VsCoq is an extension of the [Visual Studio Code](https://code.visualstudio.com/) (VS Code) editor to support Coq source files and interaction with Coq. In a recent survey of Coq users, around [one third of respondents](https://coq.discourse.group/t/coq-community-survey-2022-results-part-ii/1746#coq-user-interfaces-7) reported that they use VS Code for Coq.
As described in [the public roadmap](https://github.com/coq-community/vscoq/wiki/VsCoq-1.0-Roadmap), VsCoq developers are currently focusing on building a new simplified architecture for VsCoq. This means that there are no longer any resources available for maintaining the current codebase.
VsCoq is an open source project available under the MIT license and developed in the [Coq-community organization](https://github.com/coq-community/manifesto) on GitHub. VsCoq is written in TypeScript and uses an XML-based protocol to communicate with Coq. Interim maintainers are expected to implement bugfixes, respond to pull requests and issues on GitHub, and release new versions of VsCoq on the [Visual Studio Marketplace](https://marketplace.visualstudio.com/) and [Open VSX](https://open-vsx.org/).
An important goal of interim maintenance is to provide a good experience for Coq users relying on VS Code for Coq while the new VsCoq architecture (and support for this architecture in Coq itself) is being developed and tested. However, interim maintainers should be aware that substantial parts of the current codebase are due to be replaced as the roadmap is implemented.
During their tenure, interim maintainers will be considered part of the [Coq Team](https://coq.inria.fr/coq-team.html) and credited for their work in release notes for Coq releases, for example [on Zenodo](https://doi.org/10.5281/zenodo.1003420).
Please respond to this GitHub issue with your motivation, and summary of relevant experience, for becoming an interim maintainer of VsCoq. The interim maintainer(s) will be selected from the issue responders by the Coq core team and Coq-community owners. Those not selected will still be encouraged to contribute to VsCoq in collaboration with the new maintainer(s). | True | Volunteer interim maintainers needed for VsCoq - The Coq core team and Coq-community are looking for volunteer interim maintainers of the [VsCoq](github.com/coq-community/vscoq) project.
VsCoq is an extension of the [Visual Studio Code](https://code.visualstudio.com/) (VS Code) editor to support Coq source files and interaction with Coq. In a recent survey of Coq users, around [one third of respondents](https://coq.discourse.group/t/coq-community-survey-2022-results-part-ii/1746#coq-user-interfaces-7) reported that they use VS Code for Coq.
As described in [the public roadmap](https://github.com/coq-community/vscoq/wiki/VsCoq-1.0-Roadmap), VsCoq developers are currently focusing on building a new simplified architecture for VsCoq. This means that there are no longer any resources available for maintaining the current codebase.
VsCoq is an open source project available under the MIT license and developed in the [Coq-community organization](https://github.com/coq-community/manifesto) on GitHub. VsCoq is written in TypeScript and uses an XML-based protocol to communicate with Coq. Interim maintainers are expected to implement bugfixes, respond to pull requests and issues on GitHub, and release new versions of VsCoq on the [Visual Studio Marketplace](https://marketplace.visualstudio.com/) and [Open VSX](https://open-vsx.org/).
An important goal of interim maintenance is to provide a good experience for Coq users relying on VS Code for Coq while the new VsCoq architecture (and support for this architecture in Coq itself) is being developed and tested. However, interim maintainers should be aware that substantial parts of the current codebase are due to be replaced as the roadmap is implemented.
During their tenure, interim maintainers will be considered part of the [Coq Team](https://coq.inria.fr/coq-team.html) and credited for their work in release notes for Coq releases, for example [on Zenodo](https://doi.org/10.5281/zenodo.1003420).
Please respond to this GitHub issue with your motivation, and summary of relevant experience, for becoming an interim maintainer of VsCoq. The interim maintainer(s) will be selected from the issue responders by the Coq core team and Coq-community owners. Those not selected will still be encouraged to contribute to VsCoq in collaboration with the new maintainer(s). | main | volunteer interim maintainers needed for vscoq the coq core team and coq community are looking for volunteer interim maintainers of the github com coq community vscoq project vscoq is an extension of the vs code editor to support coq source files and interaction with coq in a recent survey of coq users around reported that they use vs code for coq as described in vscoq developers are currently focusing on building a new simplified architecture for vscoq this means that there are no longer any resources available for maintaining the current codebase vscoq is an open source project available under the mit license and developed in the on github vscoq is written in typescript and uses an xml based protocol to communicate with coq interim maintainers are expected to implement bugfixes respond to pull requests and issues on github and release new versions of vscoq on the and an important goal of interim maintenance is to provide a good experience for coq users relying on vs code for coq while the new vscoq architecture and support for this architecture in coq itself is being developed and tested however interim maintainers should be aware that substantial parts of the current codebase are due to be replaced as the roadmap is implemented during their tenure interim maintainers will be considered part of the and credited for their work in release notes for coq releases for example please respond to this github issue with your motivation and summary of relevant experience for becoming an interim maintainer of vscoq the interim maintainer s will be selected from the issue responders by the coq core team and coq community owners those not selected will still be encouraged to contribute to vscoq in collaboration with the new maintainer s | 1 |
2,534 | 8,657,431,698 | IssuesEvent | 2018-11-27 21:18:37 | Kapeli/Dash-User-Contributions | https://api.github.com/repos/Kapeli/Dash-User-Contributions | closed | DraftJS Docset maintainer needed | needs maintainer | I can no longer have time to maintain this docset and I am looking for additional contributors to assist. My repo is located at https://github.com/epitaphmike/draft-js-dash. If this is something you are interested in helping with please reach out. Thank you.
| True | DraftJS Docset maintainer needed - I can no longer have time to maintain this docset and I am looking for additional contributors to assist. My repo is located at https://github.com/epitaphmike/draft-js-dash. If this is something you are interested in helping with please reach out. Thank you.
| main | draftjs docset maintainer needed i can no longer have time to maintain this docset and i am looking for additional contributors to assist my repo is located at if this is something you are interested in helping with please reach out thank you | 1 |
168,613 | 26,672,897,362 | IssuesEvent | 2023-01-26 11:55:58 | VikeLabs/vikelabs.ca | https://api.github.com/repos/VikeLabs/vikelabs.ca | closed | Create Index Page Light Mode Prototype | Design Old | ### User Story:
As a designer I would like to create the light mode prototypes for the index page so that development can begin for this page.
### Details:
The goal for this ticket is to apply the light mode style guide to the base high fidelity prototype. A page has been added to the High Fidelity Prototyopes files in figma called "Index - Light HF Prototype". The prototype should be added and finalized here.
https://www.figma.com/file/lCVetT1xV8rlqwqc3Bq63B/High-Fidelity-Prototypes?node-id=1%3A2
### Acceptance Criteria:
- [ ] The Index Page's light mode high fidelity prototype has been finished. | 1.0 | Create Index Page Light Mode Prototype - ### User Story:
As a designer I would like to create the light mode prototypes for the index page so that development can begin for this page.
### Details:
The goal for this ticket is to apply the light mode style guide to the base high fidelity prototype. A page has been added to the High Fidelity Prototyopes files in figma called "Index - Light HF Prototype". The prototype should be added and finalized here.
https://www.figma.com/file/lCVetT1xV8rlqwqc3Bq63B/High-Fidelity-Prototypes?node-id=1%3A2
### Acceptance Criteria:
- [ ] The Index Page's light mode high fidelity prototype has been finished. | non_main | create index page light mode prototype user story as a designer i would like to create the light mode prototypes for the index page so that development can begin for this page details the goal for this ticket is to apply the light mode style guide to the base high fidelity prototype a page has been added to the high fidelity prototyopes files in figma called index light hf prototype the prototype should be added and finalized here acceptance criteria the index page s light mode high fidelity prototype has been finished | 0 |
1,215 | 5,197,285,303 | IssuesEvent | 2017-01-23 15:16:33 | Particular/PBot | https://api.github.com/repos/Particular/PBot | closed | Add "Withdrawn: Moved" label to an issue when moving it with pbot | Impact: M Size: S Tag: Maintainer Prio Type: Feature Withdrawn: Won't Fix | **Issue by [particularbot](https://github.com/particularbot)** _1/11/2016 11:00:01 AM +00:00_ _Originally opened as https://github.com/Particular/Collaboration/issues/260_
---
**Issue by [adamralph](https://github.com/adamralph)** _12/23/2015 11:32:56 AM +00:00_ _Originally opened as https://github.com/Particular/PBot/issues/64_
---
Taskforce:
The label should be created if it doesn't exist.
| True | Add "Withdrawn: Moved" label to an issue when moving it with pbot - **Issue by [particularbot](https://github.com/particularbot)** _1/11/2016 11:00:01 AM +00:00_ _Originally opened as https://github.com/Particular/Collaboration/issues/260_
---
**Issue by [adamralph](https://github.com/adamralph)** _12/23/2015 11:32:56 AM +00:00_ _Originally opened as https://github.com/Particular/PBot/issues/64_
---
Taskforce:
The label should be created if it doesn't exist.
| main | add withdrawn moved label to an issue when moving it with pbot issue by am originally opened as issue by am originally opened as taskforce the label should be created if it doesn t exist | 1 |
1,965 | 6,694,165,479 | IssuesEvent | 2017-10-10 00:02:11 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | closed | Timer: Recognise 'hour timer' etc as a request for a 1 hour timer | Maintainer Approved Status: On Hold Suggestion | When the singular form of a unit of time is searched, that should be recognised as a timer.
For example:
`hour timer` -> `1 hour timer`
`minute timer` -> `1 minute timer`
---
IA Page: http://duck.co/ia/view/timer
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mattr555
| True | Timer: Recognise 'hour timer' etc as a request for a 1 hour timer - When the singular form of a unit of time is searched, that should be recognised as a timer.
For example:
`hour timer` -> `1 hour timer`
`minute timer` -> `1 minute timer`
---
IA Page: http://duck.co/ia/view/timer
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mattr555
| main | timer recognise hour timer etc as a request for a hour timer when the singular form of a unit of time is searched that should be recognised as a timer for example hour timer hour timer minute timer minute timer ia page | 1 |
4,121 | 19,539,451,332 | IssuesEvent | 2021-12-31 16:31:14 | asclepias/asclepias-broker | https://api.github.com/repos/asclepias/asclepias-broker | closed | Data Collection: Identify issues that come through real-world ingestion to the broker | Monitoring and Maintainence | As current version of the broker is running, determine what the issues in ingestion and harvesting occur. | True | Data Collection: Identify issues that come through real-world ingestion to the broker - As current version of the broker is running, determine what the issues in ingestion and harvesting occur. | main | data collection identify issues that come through real world ingestion to the broker as current version of the broker is running determine what the issues in ingestion and harvesting occur | 1 |
1,595 | 6,572,379,632 | IssuesEvent | 2017-09-11 01:51:48 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | allow s3_bucket to append tags | affects_2.1 aws cloud feature_idea waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
s3_bucket
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.2.0 (stable-2.1 60a6c83abc) last updated 2016/09/26 10:28:28 (GMT -400)
lib/ansible/modules/core: (detached HEAD eef7b779a8) last updated 2016/09/26 10:35:50 (GMT -400)
lib/ansible/modules/extras: (detached HEAD cfbab14aa8) last updated 2016/09/26 10:35:50 (GMT -400)
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
None
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Mac OS X 10.12
##### SUMMARY
<!--- Explain the problem briefly -->
I have tags already applied to a s3 bucket. When I run a script to tag s3 buckets with additional tags, it removes all existing tags.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
---
- hosts: localhost
tasks:
- name: Tag S3 buckets
s3_bucket: {name: "{{ item }}", tags:{Application: "S3"}}
with_items:
- policy-test-rwetzelberger
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Check for s3 tag "Application" with value "S3". Found "Application" tag. no changes.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
Performed check against "Application" tag. Removed additional, existing tag that was on the bucket named "Platform" with value "GIO"
<!--- Paste verbatim command output between quotes below -->
```
ansible-playbook s3-tag.yml -vvvv
No config file found; using defaults
Loaded callback default of type stdout, v2.0
PLAYBOOK: s3-tag.yml ***********************************************************
1 plays in s3-tag.yml
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: rwetzelberger
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974 `" && echo ansible-tmp-1474905837.02-145423988674974="` echo $HOME/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974 `" ) && sleep 0'
<127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpk6N_jf TO /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/setup
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/ /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/setup && sleep 0'
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /Library/Frameworks/Python.framework/Versions/2.7/Resources/Python.app/Contents/MacOS/Python /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/setup; rm -rf "/Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/" > /dev/null 2>&1 && sleep 0'
ok: [localhost]
TASK [Tag S3 buckets] **********************************************************
task path: /Users/rwetzelberger/fork-ansible-playbooks/Gio/tasks/s3-tag.yml:5
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: rwetzelberger
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933 `" && echo ansible-tmp-1474906102.44-21483261800933="` echo $HOME/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933 `" ) && sleep 0'
<127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpWWUkBg TO /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/s3_bucket
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/ /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/s3_bucket && sleep 0'
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /Library/Frameworks/Python.framework/Versions/2.7/Resources/Python.app/Contents/MacOS/Python /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/s3_bucket; rm -rf "/Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/" > /dev/null 2>&1 && sleep 0'
changed: [localhost] => (item=policy-test-rwetzelberger) => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "force": false, "name": "policy-test-rwetzelberger", "policy": null, "profile": null, "region": null, "requester_pays": false, "s3_url": null, "security_token": null, "state": "present", "tags": {"Application": "S3"}, "validate_certs": true, "versioning": false}, "module_name": "s3_bucket"}, "item": "policy-test-rwetzelberger", "name": "policy-test-rwetzelberger", "policy": null, "requester_pays": false, "tags": {"Application": "S3"}, "versioning": {"MfaDelete": "Disabled", "Versioning": "Suspended"}}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
| True | allow s3_bucket to append tags - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
s3_bucket
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.2.0 (stable-2.1 60a6c83abc) last updated 2016/09/26 10:28:28 (GMT -400)
lib/ansible/modules/core: (detached HEAD eef7b779a8) last updated 2016/09/26 10:35:50 (GMT -400)
lib/ansible/modules/extras: (detached HEAD cfbab14aa8) last updated 2016/09/26 10:35:50 (GMT -400)
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
None
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Mac OS X 10.12
##### SUMMARY
<!--- Explain the problem briefly -->
I have tags already applied to a s3 bucket. When I run a script to tag s3 buckets with additional tags, it removes all existing tags.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
---
- hosts: localhost
tasks:
- name: Tag S3 buckets
s3_bucket: {name: "{{ item }}", tags:{Application: "S3"}}
with_items:
- policy-test-rwetzelberger
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Check for s3 tag "Application" with value "S3". Found "Application" tag. no changes.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
Performed check against "Application" tag. Removed additional, existing tag that was on the bucket named "Platform" with value "GIO"
<!--- Paste verbatim command output between quotes below -->
```
ansible-playbook s3-tag.yml -vvvv
No config file found; using defaults
Loaded callback default of type stdout, v2.0
PLAYBOOK: s3-tag.yml ***********************************************************
1 plays in s3-tag.yml
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: rwetzelberger
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974 `" && echo ansible-tmp-1474905837.02-145423988674974="` echo $HOME/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974 `" ) && sleep 0'
<127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpk6N_jf TO /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/setup
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/ /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/setup && sleep 0'
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /Library/Frameworks/Python.framework/Versions/2.7/Resources/Python.app/Contents/MacOS/Python /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/setup; rm -rf "/Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474905837.02-145423988674974/" > /dev/null 2>&1 && sleep 0'
ok: [localhost]
TASK [Tag S3 buckets] **********************************************************
task path: /Users/rwetzelberger/fork-ansible-playbooks/Gio/tasks/s3-tag.yml:5
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: rwetzelberger
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933 `" && echo ansible-tmp-1474906102.44-21483261800933="` echo $HOME/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933 `" ) && sleep 0'
<127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpWWUkBg TO /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/s3_bucket
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/ /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/s3_bucket && sleep 0'
<127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /Library/Frameworks/Python.framework/Versions/2.7/Resources/Python.app/Contents/MacOS/Python /Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/s3_bucket; rm -rf "/Users/rwetzelberger/.ansible/tmp/ansible-tmp-1474906102.44-21483261800933/" > /dev/null 2>&1 && sleep 0'
changed: [localhost] => (item=policy-test-rwetzelberger) => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "force": false, "name": "policy-test-rwetzelberger", "policy": null, "profile": null, "region": null, "requester_pays": false, "s3_url": null, "security_token": null, "state": "present", "tags": {"Application": "S3"}, "validate_certs": true, "versioning": false}, "module_name": "s3_bucket"}, "item": "policy-test-rwetzelberger", "name": "policy-test-rwetzelberger", "policy": null, "requester_pays": false, "tags": {"Application": "S3"}, "versioning": {"MfaDelete": "Disabled", "Versioning": "Suspended"}}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
| main | allow bucket to append tags issue type bug report component name bucket ansible version ansible stable last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables none os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific mac os x summary i have tags already applied to a bucket when i run a script to tag buckets with additional tags it removes all existing tags steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used hosts localhost tasks name tag buckets bucket name item tags application with items policy test rwetzelberger expected results check for tag application with value found application tag no changes actual results performed check against application tag removed additional existing tag that was on the bucket named platform with value gio ansible playbook tag yml vvvv no config file found using defaults loaded callback default of type stdout playbook tag yml plays in tag yml play task establish local connection for user rwetzelberger exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders t jf to users rwetzelberger ansible tmp ansible tmp setup exec bin sh c chmod u x users rwetzelberger ansible tmp ansible tmp users rwetzelberger ansible tmp ansible tmp setup sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf library frameworks python framework versions resources python app contents macos python users rwetzelberger ansible tmp ansible tmp setup rm rf users rwetzelberger ansible tmp ansible tmp dev null sleep ok task task path users rwetzelberger fork ansible playbooks gio tasks tag yml establish local connection for user rwetzelberger exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders t tmpwwukbg to users rwetzelberger ansible tmp ansible tmp bucket exec bin sh c chmod u x users rwetzelberger ansible tmp ansible tmp users rwetzelberger ansible tmp ansible tmp bucket sleep exec bin sh c lang en us utf lc all en us utf lc messages en us utf library frameworks python framework versions resources python app contents macos python users rwetzelberger ansible tmp ansible tmp bucket rm rf users rwetzelberger ansible tmp ansible tmp dev null sleep changed item policy test rwetzelberger changed true invocation module args aws access key null aws secret key null url null force false name policy test rwetzelberger policy null profile null region null requester pays false url null security token null state present tags application validate certs true versioning false module name bucket item policy test rwetzelberger name policy test rwetzelberger policy null requester pays false tags application versioning mfadelete disabled versioning suspended play recap localhost ok changed unreachable failed | 1 |
3,527 | 13,883,391,161 | IssuesEvent | 2020-10-18 11:46:39 | grey-software/org | https://api.github.com/repos/grey-software/org | opened | 🥅 Initiative: Raise awareness in Pakistan's Universities | Domain: User Experience Role: Maintainer Role: Product Owner | ### Motivation 🏁
<!--
A clear and concise motivation for this initiative? How will this help execute the vision of the org?
-->
Since I will be operating out of Pakistan for the next few months, I hope to establish our presence in Pakistan's tech universities. The aim is to raise awareness and recruit Pakistani students interested in open source software engineering.
### Initiative Overview 👁️🗨️
<!--
A clear and concise description of what the initiative is.
-->
The purpose of this initiative is to establish Grey Software in universities other than the University of Toronto Mississauga. This could range from having one or two Pakistani students be regular Grrey Software contributors to establishing a campus ambassador's program.
**Implementation Details 🛠️ **
<!--- Please share a plan to help realize this initiative -->
Here are some first steps that come to mind:
- Email Pakistani universities like NUST, FAST, GIK, UET, LUMS to start the conversation
- Look into setting up talks at these universities
- Research Pakistani universities that have open-source programs
### Impact 💥
Grey Software will be established in more than one country, and my dream of dividing my time between Pakistan and Canada will be closer to reality. | True | 🥅 Initiative: Raise awareness in Pakistan's Universities - ### Motivation 🏁
<!--
A clear and concise motivation for this initiative? How will this help execute the vision of the org?
-->
Since I will be operating out of Pakistan for the next few months, I hope to establish our presence in Pakistan's tech universities. The aim is to raise awareness and recruit Pakistani students interested in open source software engineering.
### Initiative Overview 👁️🗨️
<!--
A clear and concise description of what the initiative is.
-->
The purpose of this initiative is to establish Grey Software in universities other than the University of Toronto Mississauga. This could range from having one or two Pakistani students be regular Grrey Software contributors to establishing a campus ambassador's program.
**Implementation Details 🛠️ **
<!--- Please share a plan to help realize this initiative -->
Here are some first steps that come to mind:
- Email Pakistani universities like NUST, FAST, GIK, UET, LUMS to start the conversation
- Look into setting up talks at these universities
- Research Pakistani universities that have open-source programs
### Impact 💥
Grey Software will be established in more than one country, and my dream of dividing my time between Pakistan and Canada will be closer to reality. | main | 🥅 initiative raise awareness in pakistan s universities motivation 🏁 a clear and concise motivation for this initiative how will this help execute the vision of the org since i will be operating out of pakistan for the next few months i hope to establish our presence in pakistan s tech universities the aim is to raise awareness and recruit pakistani students interested in open source software engineering initiative overview 👁️🗨️ a clear and concise description of what the initiative is the purpose of this initiative is to establish grey software in universities other than the university of toronto mississauga this could range from having one or two pakistani students be regular grrey software contributors to establishing a campus ambassador s program implementation details 🛠️ here are some first steps that come to mind email pakistani universities like nust fast gik uet lums to start the conversation look into setting up talks at these universities research pakistani universities that have open source programs impact 💥 grey software will be established in more than one country and my dream of dividing my time between pakistan and canada will be closer to reality | 1 |
5,638 | 28,365,312,307 | IssuesEvent | 2023-04-12 13:32:56 | beyarkay/eskom-calendar | https://api.github.com/repos/beyarkay/eskom-calendar | closed | Missing area schedule / Bugged area: Brackenhurst | waiting-on-maintainer missing-area-schedule | **What area(s) couldn't you find on [eskomcalendar.co.za](https://eskomcalendar.co.za/ec)?**
Please also give the province/municipality, our beautiful country has a surprising number of places that are named the same as each other. If you know what your area is named on EskomSePush, including that also helps a lot.
Brackenhurst (1) | City of Ekurhuleni
**Where did you hear about [eskomcalendar.co.za](https://eskomcalendar.co.za/ec)?**
This really helps us figure out what's working!
**Any other information**
If you've got any other info you think might be helpful, feel free to leave it here
This should be part of the gauteng-ekurhuleni-block-1.ics, but the file under releases seems to have the wrong time for the area. It shows loadshedding will be at 18:00-20:00 at Stage 3 on Sunday the 9th of March. However, that's wrong, as other apps shows loadshedding between 20:00-22:00 at Stage 2.
| True | Missing area schedule / Bugged area: Brackenhurst - **What area(s) couldn't you find on [eskomcalendar.co.za](https://eskomcalendar.co.za/ec)?**
Please also give the province/municipality, our beautiful country has a surprising number of places that are named the same as each other. If you know what your area is named on EskomSePush, including that also helps a lot.
Brackenhurst (1) | City of Ekurhuleni
**Where did you hear about [eskomcalendar.co.za](https://eskomcalendar.co.za/ec)?**
This really helps us figure out what's working!
**Any other information**
If you've got any other info you think might be helpful, feel free to leave it here
This should be part of the gauteng-ekurhuleni-block-1.ics, but the file under releases seems to have the wrong time for the area. It shows loadshedding will be at 18:00-20:00 at Stage 3 on Sunday the 9th of March. However, that's wrong, as other apps shows loadshedding between 20:00-22:00 at Stage 2.
| main | missing area schedule bugged area brackenhurst what area s couldn t you find on please also give the province municipality our beautiful country has a surprising number of places that are named the same as each other if you know what your area is named on eskomsepush including that also helps a lot brackenhurst city of ekurhuleni where did you hear about this really helps us figure out what s working any other information if you ve got any other info you think might be helpful feel free to leave it here this should be part of the gauteng ekurhuleni block ics but the file under releases seems to have the wrong time for the area it shows loadshedding will be at at stage on sunday the of march however that s wrong as other apps shows loadshedding between at stage | 1 |
352,642 | 25,076,597,487 | IssuesEvent | 2022-11-07 15:56:12 | bounswe574-G3/bounswe574-2022 | https://api.github.com/repos/bounswe574-G3/bounswe574-2022 | closed | Requirements list for content relation | documentation requirements | Requirements for the content relation features will be written and added to the wiki page | 1.0 | Requirements list for content relation - Requirements for the content relation features will be written and added to the wiki page | non_main | requirements list for content relation requirements for the content relation features will be written and added to the wiki page | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.