Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
137,629
18,755,114,447
IssuesEvent
2021-11-05 09:44:20
Dima2022/node-jose
https://api.github.com/repos/Dima2022/node-jose
opened
CVE-2021-33623 (High) detected in trim-newlines-1.0.0.tgz
security vulnerability
## CVE-2021-33623 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p></summary> <p>Trim newlines from the start and/or end of a string</p> <p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p> <p>Path to dependency file: node-jose/package.json</p> <p>Path to vulnerable library: node-jose/node_modules/trim-newlines/package.json</p> <p> Dependency Hierarchy: - karma-coverage-1.1.2.tgz (Root Library) - dateformat-1.0.12.tgz - meow-3.7.0.tgz - :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/node-jose/commit/402fef2668e2cefe13cada8c1eb68c22d93a6837">402fef2668e2cefe13cada8c1eb68c22d93a6837</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623>CVE-2021-33623</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: trim-newlines - 3.0.1, 4.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"trim-newlines","packageVersion":"1.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma-coverage:1.1.2;dateformat:1.0.12;meow:3.7.0;trim-newlines:1.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"trim-newlines - 3.0.1, 4.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33623","vulnerabilityDetails":"The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-33623 (High) detected in trim-newlines-1.0.0.tgz - ## CVE-2021-33623 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-newlines-1.0.0.tgz</b></p></summary> <p>Trim newlines from the start and/or end of a string</p> <p>Library home page: <a href="https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz">https://registry.npmjs.org/trim-newlines/-/trim-newlines-1.0.0.tgz</a></p> <p>Path to dependency file: node-jose/package.json</p> <p>Path to vulnerable library: node-jose/node_modules/trim-newlines/package.json</p> <p> Dependency Hierarchy: - karma-coverage-1.1.2.tgz (Root Library) - dateformat-1.0.12.tgz - meow-3.7.0.tgz - :x: **trim-newlines-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Dima2022/node-jose/commit/402fef2668e2cefe13cada8c1eb68c22d93a6837">402fef2668e2cefe13cada8c1eb68c22d93a6837</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method. <p>Publish Date: 2021-05-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623>CVE-2021-33623</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-33623</a></p> <p>Release Date: 2021-05-28</p> <p>Fix Resolution: trim-newlines - 3.0.1, 4.0.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"trim-newlines","packageVersion":"1.0.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"karma-coverage:1.1.2;dateformat:1.0.12;meow:3.7.0;trim-newlines:1.0.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"trim-newlines - 3.0.1, 4.0.1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-33623","vulnerabilityDetails":"The trim-newlines package before 3.0.1 and 4.x before 4.0.1 for Node.js has an issue related to regular expression denial-of-service (ReDoS) for the .end() method.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
cve high detected in trim newlines tgz cve high severity vulnerability vulnerable library trim newlines tgz trim newlines from the start and or end of a string library home page a href path to dependency file node jose package json path to vulnerable library node jose node modules trim newlines package json dependency hierarchy karma coverage tgz root library dateformat tgz meow tgz x trim newlines tgz vulnerable library found in head commit a href found in base branch master vulnerability details the trim newlines package before and x before for node js has an issue related to regular expression denial of service redos for the end method publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution trim newlines isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree karma coverage dateformat meow trim newlines isminimumfixversionavailable true minimumfixversion trim newlines basebranches vulnerabilityidentifier cve vulnerabilitydetails the trim newlines package before and x before for node js has an issue related to regular expression denial of service redos for the end method vulnerabilityurl
0
4,539
23,621,312,120
IssuesEvent
2022-08-24 20:52:10
aws/aws-sam-build-images
https://api.github.com/repos/aws/aws-sam-build-images
closed
What happened with aws-sam-cli-build-image-nodejs16.x
maintainer/need-response stage/needs-triage type/bug
### Description: Some month ago we was deploying our function using nodejs16 image version. Now it's not available anymore. The repository still contains the Dockerfile for building this image, but it looks like it's excluded to be pushed.
True
What happened with aws-sam-cli-build-image-nodejs16.x - ### Description: Some month ago we was deploying our function using nodejs16 image version. Now it's not available anymore. The repository still contains the Dockerfile for building this image, but it looks like it's excluded to be pushed.
main
what happened with aws sam cli build image x description some month ago we was deploying our function using image version now it s not available anymore the repository still contains the dockerfile for building this image but it looks like it s excluded to be pushed
1
78,787
9,795,148,242
IssuesEvent
2019-06-11 02:20:56
flutter/flutter
https://api.github.com/repos/flutter/flutter
reopened
Let scrollbars avoid obstructing slivers and media query paddings
f: cupertino f: material design f: scrolling framework severe: new feature
Consider a CustomScrollView with a SliverAppBar and a SliverList and the user wants a scrollbar showing the position inside the SliverList's contents only. Let the SliverAppBar report SliverGeometry describing its obstructing and max obstructing extends. Let that get aggregated by the scrollable and bubbled up with a scroll notification so the scrollbar can start below the obstructing area.
1.0
Let scrollbars avoid obstructing slivers and media query paddings - Consider a CustomScrollView with a SliverAppBar and a SliverList and the user wants a scrollbar showing the position inside the SliverList's contents only. Let the SliverAppBar report SliverGeometry describing its obstructing and max obstructing extends. Let that get aggregated by the scrollable and bubbled up with a scroll notification so the scrollbar can start below the obstructing area.
non_main
let scrollbars avoid obstructing slivers and media query paddings consider a customscrollview with a sliverappbar and a sliverlist and the user wants a scrollbar showing the position inside the sliverlist s contents only let the sliverappbar report slivergeometry describing its obstructing and max obstructing extends let that get aggregated by the scrollable and bubbled up with a scroll notification so the scrollbar can start below the obstructing area
0
158,609
24,864,761,647
IssuesEvent
2022-10-27 11:01:52
WordPress/wporg-showcase-2022
https://api.github.com/repos/WordPress/wporg-showcase-2022
reopened
Add wporg submenu
Template: Single Template: Front Page Template: Archive Template: Submit Template: Submit Assets Need Design
We need all pages to display this menu: <img width="498" alt="Screen Shot 2022-10-12 at 1 46 39 PM" src="https://user-images.githubusercontent.com/1657336/195252365-bd4bdcee-42d4-4ba5-a16e-533077dd7d13.png"> **Menu Items** - Submit a site
1.0
Add wporg submenu - We need all pages to display this menu: <img width="498" alt="Screen Shot 2022-10-12 at 1 46 39 PM" src="https://user-images.githubusercontent.com/1657336/195252365-bd4bdcee-42d4-4ba5-a16e-533077dd7d13.png"> **Menu Items** - Submit a site
non_main
add wporg submenu we need all pages to display this menu img width alt screen shot at pm src menu items submit a site
0
23,626
16,475,511,497
IssuesEvent
2021-05-24 04:42:19
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Error in generate apsimx file
bug interface/infrastructure
I use the master branch of ApsimX repository. Steps to reproduce this error. * Open the wheat.apsimx under Tests\Validation\Wheat * Right click the Experiment `Callington` under folder `NPIField2019` to generate *.apsimx * Open the generated file `CallingtonTOS6CvSunbri.apsimx` and run simulations Then I get an error about `GetData` which is related with some information is not stored in the generated apsimx. In the wheat.apsimx file, `TOS 6` specified the `[Sowing].Script.SowDate` and `[Sowing].Script.Treat` ![image](https://user-images.githubusercontent.com/1657381/116646215-e0797700-a9ba-11eb-9fa1-346b55bc6357.png) However, these two attributes are not exported into generated apsimx file. ![image](https://user-images.githubusercontent.com/1657381/116646332-19b1e700-a9bb-11eb-8a8a-78d14294b36a.png)
1.0
Error in generate apsimx file - I use the master branch of ApsimX repository. Steps to reproduce this error. * Open the wheat.apsimx under Tests\Validation\Wheat * Right click the Experiment `Callington` under folder `NPIField2019` to generate *.apsimx * Open the generated file `CallingtonTOS6CvSunbri.apsimx` and run simulations Then I get an error about `GetData` which is related with some information is not stored in the generated apsimx. In the wheat.apsimx file, `TOS 6` specified the `[Sowing].Script.SowDate` and `[Sowing].Script.Treat` ![image](https://user-images.githubusercontent.com/1657381/116646215-e0797700-a9ba-11eb-9fa1-346b55bc6357.png) However, these two attributes are not exported into generated apsimx file. ![image](https://user-images.githubusercontent.com/1657381/116646332-19b1e700-a9bb-11eb-8a8a-78d14294b36a.png)
non_main
error in generate apsimx file i use the master branch of apsimx repository steps to reproduce this error open the wheat apsimx under tests validation wheat right click the experiment callington under folder to generate apsimx open the generated file apsimx and run simulations then i get an error about getdata which is related with some information is not stored in the generated apsimx in the wheat apsimx file tos specified the script sowdate and script treat however these two attributes are not exported into generated apsimx file
0
2,845
10,219,571,266
IssuesEvent
2019-08-15 18:56:07
arcticicestudio/styleguide-javascript
https://api.github.com/repos/arcticicestudio/styleguide-javascript
closed
Git ignore and attribute pattern
context-workflow scope-maintainability scope-quality type-task
<p align="center"><img src="https://upload.wikimedia.org/wikipedia/commons/e/e0/Git-logo.svg" width="20%" /></p> > Epic: #8 Update the [`.gitattributes`][a] and [`.gitignore`][i] configuration files to use the latest pattern. [a]: https://git-scm.com/docs/gitattributes [i]: https://git-scm.com/docs/gitignore
True
Git ignore and attribute pattern - <p align="center"><img src="https://upload.wikimedia.org/wikipedia/commons/e/e0/Git-logo.svg" width="20%" /></p> > Epic: #8 Update the [`.gitattributes`][a] and [`.gitignore`][i] configuration files to use the latest pattern. [a]: https://git-scm.com/docs/gitattributes [i]: https://git-scm.com/docs/gitignore
main
git ignore and attribute pattern epic update the and configuration files to use the latest pattern
1
2,698
9,436,330,164
IssuesEvent
2019-04-13 05:28:06
invertase/react-native-firebase
https://api.github.com/repos/invertase/react-native-firebase
closed
[Proposal][WIP] Storage improvements
Hacktoberfest await-maintainer-feedback docs help-wanted ios js storage 🐞 bug 👁investigate 👉 await-user-feedback 🤖 android
Hi all 👋 `storage()` has fallen slightly behind recently and we plan on improving the module for a 5.x.x release which includes improving it to ensure that it provides the all the functionality in an easy to use and reliable manner with better documentation and bug fixes. This issue is a placeholder that will be updated as we have more details on what will be supported and how this will be structured in the API. Secondly, it's also here to show that we're aware of the issues that have already been raised and will be addressing them as part of this proposal. Any historic or new issues will be closed and redirected here to track all the issues that need addressing in one place. New features to add: - [ ] Multi-bucket support - RNFB internals re-written to support this JS side in https://github.com/invertase/react-native-firebase/commit/7632da1809554cc5314361ec36c5ca88db6a8fa6 - just needs storage work done now. - [ ] `StorageTask`; and support for resumable uploads/downloads - [ ] `cancel()` - [ ] `resume()` - [ ] `pause()` --- Loving `react-native-firebase` and the support we provide? Please consider supporting us with any of the below: - 👉 Back financially via [Open Collective](https://opencollective.com/react-native-firebase/donate) - 👉 Follow [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter - 👉 Star this repo on GitHub ⭐️
True
[Proposal][WIP] Storage improvements - Hi all 👋 `storage()` has fallen slightly behind recently and we plan on improving the module for a 5.x.x release which includes improving it to ensure that it provides the all the functionality in an easy to use and reliable manner with better documentation and bug fixes. This issue is a placeholder that will be updated as we have more details on what will be supported and how this will be structured in the API. Secondly, it's also here to show that we're aware of the issues that have already been raised and will be addressing them as part of this proposal. Any historic or new issues will be closed and redirected here to track all the issues that need addressing in one place. New features to add: - [ ] Multi-bucket support - RNFB internals re-written to support this JS side in https://github.com/invertase/react-native-firebase/commit/7632da1809554cc5314361ec36c5ca88db6a8fa6 - just needs storage work done now. - [ ] `StorageTask`; and support for resumable uploads/downloads - [ ] `cancel()` - [ ] `resume()` - [ ] `pause()` --- Loving `react-native-firebase` and the support we provide? Please consider supporting us with any of the below: - 👉 Back financially via [Open Collective](https://opencollective.com/react-native-firebase/donate) - 👉 Follow [`React Native Firebase`](https://twitter.com/rnfirebase) and [`Invertase`](https://twitter.com/invertaseio) on Twitter - 👉 Star this repo on GitHub ⭐️
main
storage improvements hi all 👋 storage has fallen slightly behind recently and we plan on improving the module for a x x release which includes improving it to ensure that it provides the all the functionality in an easy to use and reliable manner with better documentation and bug fixes this issue is a placeholder that will be updated as we have more details on what will be supported and how this will be structured in the api secondly it s also here to show that we re aware of the issues that have already been raised and will be addressing them as part of this proposal any historic or new issues will be closed and redirected here to track all the issues that need addressing in one place new features to add multi bucket support rnfb internals re written to support this js side in just needs storage work done now storagetask and support for resumable uploads downloads cancel resume pause loving react native firebase and the support we provide please consider supporting us with any of the below 👉 back financially via 👉 follow and on twitter 👉 star this repo on github ⭐️
1
40,430
20,832,471,994
IssuesEvent
2022-03-19 17:19:18
artichoke/artichoke
https://api.github.com/repos/artichoke/artichoke
closed
Add a `EncodedString::utf8` constructor
E-easy A-ruby-core A-performance C-quality
As a nit, I'd rather add a `EncodedString::utf8` constructor, since we pay to branch on the encoding in `EncodedString::new` when we don't need to. _Originally posted by @lopopolo in https://github.com/artichoke/artichoke/pull/1678#r820325590_
True
Add a `EncodedString::utf8` constructor - As a nit, I'd rather add a `EncodedString::utf8` constructor, since we pay to branch on the encoding in `EncodedString::new` when we don't need to. _Originally posted by @lopopolo in https://github.com/artichoke/artichoke/pull/1678#r820325590_
non_main
add a encodedstring constructor as a nit i d rather add a encodedstring constructor since we pay to branch on the encoding in encodedstring new when we don t need to originally posted by lopopolo in
0
147,327
23,200,068,854
IssuesEvent
2022-08-01 20:28:02
dart-lang/dartdoc
https://api.github.com/repos/dart-lang/dartdoc
closed
Allow fold out sections for more detailed/advanced docs
enhancement P3 web-design
From an [old dartdoc discussion](https://docs.google.com/document/d/1tndhthlM9jFls1kA5YvPj-LMiaNhp1uSAnRStpxB370/edit?ts=5a134f5e#heading=h.4tkns0pv6mna): ---snip--- We should also use sections inside member comments. I'm proposing that dartdoc recognizes Advanced sections and folds them by default. That is, a markdown section Advanced in a normal dartdoc comment is not visible by default, but can only be looked at by clicking on a '+' button. ``` /// Converts a double to its decimal string-representation. /// /// # Advanced /// The conversion algorithm must be accurate and follow the internal identity /// requirement as specified in [Steele](http:// …). /// /// If the output number requires more than 5 digits, it is expressed as an /// exponential number. /// ....String toString() { … } ``` ---snip---
1.0
Allow fold out sections for more detailed/advanced docs - From an [old dartdoc discussion](https://docs.google.com/document/d/1tndhthlM9jFls1kA5YvPj-LMiaNhp1uSAnRStpxB370/edit?ts=5a134f5e#heading=h.4tkns0pv6mna): ---snip--- We should also use sections inside member comments. I'm proposing that dartdoc recognizes Advanced sections and folds them by default. That is, a markdown section Advanced in a normal dartdoc comment is not visible by default, but can only be looked at by clicking on a '+' button. ``` /// Converts a double to its decimal string-representation. /// /// # Advanced /// The conversion algorithm must be accurate and follow the internal identity /// requirement as specified in [Steele](http:// …). /// /// If the output number requires more than 5 digits, it is expressed as an /// exponential number. /// ....String toString() { … } ``` ---snip---
non_main
allow fold out sections for more detailed advanced docs from an snip we should also use sections inside member comments i m proposing that dartdoc recognizes advanced sections and folds them by default that is a markdown section advanced in a normal dartdoc comment is not visible by default but can only be looked at by clicking on a button converts a double to its decimal string representation advanced the conversion algorithm must be accurate and follow the internal identity requirement as specified in http … if the output number requires more than digits it is expressed as an exponential number string tostring … snip
0
323,229
23,939,243,803
IssuesEvent
2022-09-11 17:17:49
supabase/supabase
https://api.github.com/repos/supabase/supabase
opened
resetPasswordForEmail() Example does not work.
documentation
# Improve documentation ## Link [Docs Link](https://supabase.com/docs/reference/javascript/next/auth-resetpasswordforemail) ## Describe the problem The code example in the docs has`options` as argument passed down to the function which can't work as `options: { redirectTo: "" }` is no valid object without curly brackets. Like so `{ options: { redirectTo: "" }}`. After looking deeper at this issue i found out to skip the option key completely. So the docs should be improved like written below under **Describe the improvement**. ```js const { error, data } = await supabase.auth.resetPasswordForEmail(email, options: { redirectTo: 'https://example.com/update-password', }) ``` ## Describe the improvement Remove the `options:` key. from args. To prevent users from debugging and looking for the actual object to pass. ```js const { error, data } = await supabase.auth.resetPasswordForEmail(email, { redirectTo: 'https://example.com/update-password', }) ```
1.0
resetPasswordForEmail() Example does not work. - # Improve documentation ## Link [Docs Link](https://supabase.com/docs/reference/javascript/next/auth-resetpasswordforemail) ## Describe the problem The code example in the docs has`options` as argument passed down to the function which can't work as `options: { redirectTo: "" }` is no valid object without curly brackets. Like so `{ options: { redirectTo: "" }}`. After looking deeper at this issue i found out to skip the option key completely. So the docs should be improved like written below under **Describe the improvement**. ```js const { error, data } = await supabase.auth.resetPasswordForEmail(email, options: { redirectTo: 'https://example.com/update-password', }) ``` ## Describe the improvement Remove the `options:` key. from args. To prevent users from debugging and looking for the actual object to pass. ```js const { error, data } = await supabase.auth.resetPasswordForEmail(email, { redirectTo: 'https://example.com/update-password', }) ```
non_main
resetpasswordforemail example does not work improve documentation link describe the problem the code example in the docs has options as argument passed down to the function which can t work as options redirectto is no valid object without curly brackets like so options redirectto after looking deeper at this issue i found out to skip the option key completely so the docs should be improved like written below under describe the improvement js const error data await supabase auth resetpasswordforemail email options redirectto describe the improvement remove the options key from args to prevent users from debugging and looking for the actual object to pass js const error data await supabase auth resetpasswordforemail email redirectto
0
933
4,644,111,853
IssuesEvent
2016-09-30 15:25:17
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Will ec2_asg support "default_cooldown" and "termination_policies" options?
affects_2.0 aws cloud feature_idea waiting_on_maintainer
When trying to use both "termination_policies" and "default_cooldown" it fails as there is no support for them even though they are listed in the ASG_ATTRIBUTES string and also supported by boto downstream. Will these get added to the module?
True
Will ec2_asg support "default_cooldown" and "termination_policies" options? - When trying to use both "termination_policies" and "default_cooldown" it fails as there is no support for them even though they are listed in the ASG_ATTRIBUTES string and also supported by boto downstream. Will these get added to the module?
main
will asg support default cooldown and termination policies options when trying to use both termination policies and default cooldown it fails as there is no support for them even though they are listed in the asg attributes string and also supported by boto downstream will these get added to the module
1
1,916
6,577,706,408
IssuesEvent
2017-09-12 02:45:01
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
cloud/docker: recreates stopped named containers with state=started
affects_2.0 bug_report cloud docker waiting_on_maintainer
##### Issue Type: Bug Report ##### Plugin Name: docker ##### Ansible Version: ``` ansible 2.0.0.2 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: Docker module recreates stopped named container with state=started ##### Steps To Reproduce: Task description ``` - name: Start container docker: image: debian name: lab pull: missing detach: yes net: bridge tty: yes command: sleep infinity state: started ``` Output from ansible-paybook -vv ``` TASK [start-container : Start container] *************************************** changed: [localhost] => {"ansible_facts": {"docker_containers": [{"AppArmorProfile": "", "Args": ["infinity"], "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["sleep", "infinity"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "75bb43da9afb", "Image": "android", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": true, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-02-21T15:57:18.464570878Z", "Driver": "btrfs", "ExecIDs": null, "GraphDriver": {"Data": null, "Name": "btrfs"}, "HostConfig": {"Binds": null, "BlkioDeviceReadBps": null, "BlkioDeviceReadIOps": null, "BlkioDeviceWriteBps": null, "BlkioDeviceWriteIOps": null, "BlkioWeight": 0, "BlkioWeightDevice": null, "CapAdd": null, "CapDrop": null, "CgroupParent": "", "ConsoleSize": [0, 0], "ContainerIDFile": "", "CpuPeriod": 0, "CpuQuota": 0, "CpuShares": 0, "CpusetCpus": "", "CpusetMems": "", "Devices": null, "Dns": null, "DnsOptions": null, "DnsSearch": null, "ExtraHosts": null, "GroupAdd": null, "IpcMode": "", "Isolation": "", "KernelMemory": 0, "Links": null, "LogConfig": {"Config": {}, "Type": "journald"}, "Memory": 0, "MemoryReservation": 0, "MemorySwap": 0, "MemorySwappiness": -1, "NetworkMode": "bridge", "OomKillDisable": false, "OomScoreAdj": 0, "PidMode": "", "PidsLimit": 0, "PortBindings": null, "Privileged": false, "PublishAllPorts": false, "ReadonlyRootfs": false, "RestartPolicy": {"MaximumRetryCount": 0, "Name": ""}, "SecurityOpt": null, "ShmSize": 67108864, "UTSMode": "", "Ulimits": null, "VolumeDriver": "", "VolumesFrom": null}, "HostnamePath": "/var/lib/docker/containers/75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436/hostname", "HostsPath": "/var/lib/docker/containers/75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436/hosts", "Id": "75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436", "Image": "sha256:484a6a69ac0c4f06b7bff344a36745414fe57024c07ab3c90d8146b835256008", "LogPath": "", "MountLabel": "", "Mounts": [], "Name": "/lab", "NetworkSettings": {"Bridge": "", "EndpointID": "a929802ab259e1992472d2ac0383f82ca82ca19f901f565a6dcc1ccd3072f4ce", "Gateway": "172.17.0.1", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "HairpinMode": false, "IPAddress": "172.17.0.2", "IPPrefixLen": 16, "IPv6Gateway": "", "LinkLocalIPv6Address": "", "LinkLocalIPv6PrefixLen": 0, "MacAddress": "02:42:ac:11:00:02", "Networks": {"bridge": {"Aliases": null, "EndpointID": "a929802ab259e1992472d2ac0383f82ca82ca19f901f565a6dcc1ccd3072f4ce", "Gateway": "172.17.0.1", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "IPAMConfig": null, "IPAddress": "172.17.0.2", "IPPrefixLen": 16, "IPv6Gateway": "", "Links": null, "MacAddress": "02:42:ac:11:00:02", "NetworkID": "6a6c469a29d9f7d80a25d6df2e158d1f4c4ead81a463fe4ac13b20a60280ce6d"}}, "Ports": {}, "SandboxID": "6ccf812bac48ad68b87808dcc05dad84cf41054a21d03f48e2eb1dcba45414a7", "SandboxKey": "/var/run/docker/netns/6ccf812bac48", "SecondaryIPAddresses": null, "SecondaryIPv6Addresses": null}, "Path": "sleep", "ProcessLabel": "", "ResolvConfPath": "/var/lib/docker/containers/75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436/resolv.conf", "RestartCount": 0, "State": {"Dead": false, "Error": "", "ExitCode": 0, "FinishedAt": "0001-01-01T00:00:00Z", "OOMKilled": false, "Paused": false, "Pid": 16620, "Restarting": false, "Running": true, "StartedAt": "2016-02-21T15:57:18.649681499Z", "Status": "running"}}]}, "changed": true, "msg": "removed 1 container, started 1 container, created 1 container.", "reload_reasons": null, "summary": {"created": 1, "killed": 0, "pulled": 0, "removed": 1, "restarted": 0, "started": 1, "stopped": 0}} ``` ##### Expected Results: Named container restarts with last saved state ##### Actual Results: Named container was recreated from image Looks like the problem caused by this code: https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker.py#L1690
True
cloud/docker: recreates stopped named containers with state=started - ##### Issue Type: Bug Report ##### Plugin Name: docker ##### Ansible Version: ``` ansible 2.0.0.2 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: N/A ##### Environment: N/A ##### Summary: Docker module recreates stopped named container with state=started ##### Steps To Reproduce: Task description ``` - name: Start container docker: image: debian name: lab pull: missing detach: yes net: bridge tty: yes command: sleep infinity state: started ``` Output from ansible-paybook -vv ``` TASK [start-container : Start container] *************************************** changed: [localhost] => {"ansible_facts": {"docker_containers": [{"AppArmorProfile": "", "Args": ["infinity"], "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["sleep", "infinity"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "75bb43da9afb", "Image": "android", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": true, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-02-21T15:57:18.464570878Z", "Driver": "btrfs", "ExecIDs": null, "GraphDriver": {"Data": null, "Name": "btrfs"}, "HostConfig": {"Binds": null, "BlkioDeviceReadBps": null, "BlkioDeviceReadIOps": null, "BlkioDeviceWriteBps": null, "BlkioDeviceWriteIOps": null, "BlkioWeight": 0, "BlkioWeightDevice": null, "CapAdd": null, "CapDrop": null, "CgroupParent": "", "ConsoleSize": [0, 0], "ContainerIDFile": "", "CpuPeriod": 0, "CpuQuota": 0, "CpuShares": 0, "CpusetCpus": "", "CpusetMems": "", "Devices": null, "Dns": null, "DnsOptions": null, "DnsSearch": null, "ExtraHosts": null, "GroupAdd": null, "IpcMode": "", "Isolation": "", "KernelMemory": 0, "Links": null, "LogConfig": {"Config": {}, "Type": "journald"}, "Memory": 0, "MemoryReservation": 0, "MemorySwap": 0, "MemorySwappiness": -1, "NetworkMode": "bridge", "OomKillDisable": false, "OomScoreAdj": 0, "PidMode": "", "PidsLimit": 0, "PortBindings": null, "Privileged": false, "PublishAllPorts": false, "ReadonlyRootfs": false, "RestartPolicy": {"MaximumRetryCount": 0, "Name": ""}, "SecurityOpt": null, "ShmSize": 67108864, "UTSMode": "", "Ulimits": null, "VolumeDriver": "", "VolumesFrom": null}, "HostnamePath": "/var/lib/docker/containers/75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436/hostname", "HostsPath": "/var/lib/docker/containers/75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436/hosts", "Id": "75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436", "Image": "sha256:484a6a69ac0c4f06b7bff344a36745414fe57024c07ab3c90d8146b835256008", "LogPath": "", "MountLabel": "", "Mounts": [], "Name": "/lab", "NetworkSettings": {"Bridge": "", "EndpointID": "a929802ab259e1992472d2ac0383f82ca82ca19f901f565a6dcc1ccd3072f4ce", "Gateway": "172.17.0.1", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "HairpinMode": false, "IPAddress": "172.17.0.2", "IPPrefixLen": 16, "IPv6Gateway": "", "LinkLocalIPv6Address": "", "LinkLocalIPv6PrefixLen": 0, "MacAddress": "02:42:ac:11:00:02", "Networks": {"bridge": {"Aliases": null, "EndpointID": "a929802ab259e1992472d2ac0383f82ca82ca19f901f565a6dcc1ccd3072f4ce", "Gateway": "172.17.0.1", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "IPAMConfig": null, "IPAddress": "172.17.0.2", "IPPrefixLen": 16, "IPv6Gateway": "", "Links": null, "MacAddress": "02:42:ac:11:00:02", "NetworkID": "6a6c469a29d9f7d80a25d6df2e158d1f4c4ead81a463fe4ac13b20a60280ce6d"}}, "Ports": {}, "SandboxID": "6ccf812bac48ad68b87808dcc05dad84cf41054a21d03f48e2eb1dcba45414a7", "SandboxKey": "/var/run/docker/netns/6ccf812bac48", "SecondaryIPAddresses": null, "SecondaryIPv6Addresses": null}, "Path": "sleep", "ProcessLabel": "", "ResolvConfPath": "/var/lib/docker/containers/75bb43da9afb967ee60c095f3dd5f03183f80b3973ccb75d62d08678f0de6436/resolv.conf", "RestartCount": 0, "State": {"Dead": false, "Error": "", "ExitCode": 0, "FinishedAt": "0001-01-01T00:00:00Z", "OOMKilled": false, "Paused": false, "Pid": 16620, "Restarting": false, "Running": true, "StartedAt": "2016-02-21T15:57:18.649681499Z", "Status": "running"}}]}, "changed": true, "msg": "removed 1 container, started 1 container, created 1 container.", "reload_reasons": null, "summary": {"created": 1, "killed": 0, "pulled": 0, "removed": 1, "restarted": 0, "started": 1, "stopped": 0}} ``` ##### Expected Results: Named container restarts with last saved state ##### Actual Results: Named container was recreated from image Looks like the problem caused by this code: https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker.py#L1690
main
cloud docker recreates stopped named containers with state started issue type bug report plugin name docker ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides ansible configuration n a environment n a summary docker module recreates stopped named container with state started steps to reproduce task description name start container docker image debian name lab pull missing detach yes net bridge tty yes command sleep infinity state started output from ansible paybook vv task changed ansible facts docker containers config attachstderr false attachstdin false attachstdout false cmd domainname entrypoint null env hostname image android labels onbuild null openstdin false stdinonce false tty true user volumes null workingdir created driver btrfs execids null graphdriver data null name btrfs hostconfig binds null blkiodevicereadbps null blkiodevicereadiops null blkiodevicewritebps null blkiodevicewriteiops null blkioweight blkioweightdevice null capadd null capdrop null cgroupparent consolesize containeridfile cpuperiod cpuquota cpushares cpusetcpus cpusetmems devices null dns null dnsoptions null dnssearch null extrahosts null groupadd null ipcmode isolation kernelmemory links null logconfig config type journald memory memoryreservation memoryswap memoryswappiness networkmode bridge oomkilldisable false oomscoreadj pidmode pidslimit portbindings null privileged false publishallports false readonlyrootfs false restartpolicy maximumretrycount name securityopt null shmsize utsmode ulimits null volumedriver volumesfrom null hostnamepath var lib docker containers hostname hostspath var lib docker containers hosts id image logpath mountlabel mounts name lab networksettings bridge endpointid gateway hairpinmode false ipaddress ipprefixlen macaddress ac networks bridge aliases null endpointid gateway ipamconfig null ipaddress ipprefixlen links null macaddress ac networkid ports sandboxid sandboxkey var run docker netns secondaryipaddresses null null path sleep processlabel resolvconfpath var lib docker containers resolv conf restartcount state dead false error exitcode finishedat oomkilled false paused false pid restarting false running true startedat status running changed true msg removed container started container created container reload reasons null summary created killed pulled removed restarted started stopped expected results named container restarts with last saved state actual results named container was recreated from image looks like the problem caused by this code
1
30,882
25,141,834,881
IssuesEvent
2022-11-10 00:03:11
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.IO.FileNotFoundException thrown when using System.Text.Json in mixed .Net version solution. Assembly versions do not match.
question area-Infrastructure-libraries no-recent-activity needs-author-action
### Description We have a fairly large code base which consist of different .Net version projects. Mostly .Net6.0 and .NetStandard2.0 but a few .Net Framework 4.7.2 projects as well. We are migrating all of these to .Net6.0 but it will still take a lot of time before everything is done. The issue we are facing is with System.Text.Json. We have installed the 6.0.5 nuget package to the projects that require it but since couple of them are .Net Framework projects, it seems that they get a different version of the System.Text.Json assembly. For other project types the assembly version is 6.0.0.0 but for the legacy framework it is 6.0.0.5. This causes runtime error: "System.IO.FileNotFoundException: Could not load file or assembly 'System.Text.Json, Version=6.0.0.5, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified." when the framework project is referenced by a .Net6.0 project (yes, I know it's not ideal but at the time being a necessary evil). ### Reproduction Steps Create a .Net6.0 and a .Net4.7.2 projects. Add a System.Text.Json reference to the 4.7.2 project and reference that project from the .Net6.0 project. Run the .Net6.0 and it'll throw an System.IO.FileNotFoundException Here is a sample solution for this: [JsonReferenceIssue.zip](https://github.com/dotnet/runtime/files/9111329/JsonReferenceIssue.zip) ### Expected behavior I would expect the NuGet for System.Text.Json 6.0.5 to have the same assembly version for different .Net versions. ### Actual behavior Runtime error: "System.IO.FileNotFoundException: Could not load file or assembly 'System.Text.Json, Version=6.0.0.5, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified." ### Regression? _No response_ ### Known Workarounds If I downgrade the projects to use the NuGet 6.0.0 this works as expected. In the sample solution I can also workaround the issue by adding the following snippet to the Program.cs but I don't feel comfortable of trying to do this in our production code. ``` AppDomain.CurrentDomain.AssemblyResolve += CurrentDomain_AssemblyResolve; static System.Reflection.Assembly? CurrentDomain_AssemblyResolve(object? sender, ResolveEventArgs args) { if (args.Name == "System.Text.Json, Version=6.0.0.5, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51") { return typeof(System.Text.Json.JsonSerializer).Assembly; } return null; } ``` ### Configuration Partial output of `dotnet --info`: ``` .NET SDK (reflecting any global.json): Version: 6.0.301 Commit: 43f9b18481 Runtime Environment: OS Name: Windows OS Version: 10.0.19044 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\6.0.301\ Host (useful for support): Version: 6.0.6 Commit: 7cca709db2 ``` ### Other information I'm not sure if the different System.Text.Json assembly version numbers between the .Net Framework and .Net Core versions is deliberate. If yes, I would like to know how to circumvent this issue.
1.0
System.IO.FileNotFoundException thrown when using System.Text.Json in mixed .Net version solution. Assembly versions do not match. - ### Description We have a fairly large code base which consist of different .Net version projects. Mostly .Net6.0 and .NetStandard2.0 but a few .Net Framework 4.7.2 projects as well. We are migrating all of these to .Net6.0 but it will still take a lot of time before everything is done. The issue we are facing is with System.Text.Json. We have installed the 6.0.5 nuget package to the projects that require it but since couple of them are .Net Framework projects, it seems that they get a different version of the System.Text.Json assembly. For other project types the assembly version is 6.0.0.0 but for the legacy framework it is 6.0.0.5. This causes runtime error: "System.IO.FileNotFoundException: Could not load file or assembly 'System.Text.Json, Version=6.0.0.5, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified." when the framework project is referenced by a .Net6.0 project (yes, I know it's not ideal but at the time being a necessary evil). ### Reproduction Steps Create a .Net6.0 and a .Net4.7.2 projects. Add a System.Text.Json reference to the 4.7.2 project and reference that project from the .Net6.0 project. Run the .Net6.0 and it'll throw an System.IO.FileNotFoundException Here is a sample solution for this: [JsonReferenceIssue.zip](https://github.com/dotnet/runtime/files/9111329/JsonReferenceIssue.zip) ### Expected behavior I would expect the NuGet for System.Text.Json 6.0.5 to have the same assembly version for different .Net versions. ### Actual behavior Runtime error: "System.IO.FileNotFoundException: Could not load file or assembly 'System.Text.Json, Version=6.0.0.5, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51'. The system cannot find the file specified." ### Regression? _No response_ ### Known Workarounds If I downgrade the projects to use the NuGet 6.0.0 this works as expected. In the sample solution I can also workaround the issue by adding the following snippet to the Program.cs but I don't feel comfortable of trying to do this in our production code. ``` AppDomain.CurrentDomain.AssemblyResolve += CurrentDomain_AssemblyResolve; static System.Reflection.Assembly? CurrentDomain_AssemblyResolve(object? sender, ResolveEventArgs args) { if (args.Name == "System.Text.Json, Version=6.0.0.5, Culture=neutral, PublicKeyToken=cc7b13ffcd2ddd51") { return typeof(System.Text.Json.JsonSerializer).Assembly; } return null; } ``` ### Configuration Partial output of `dotnet --info`: ``` .NET SDK (reflecting any global.json): Version: 6.0.301 Commit: 43f9b18481 Runtime Environment: OS Name: Windows OS Version: 10.0.19044 OS Platform: Windows RID: win10-x64 Base Path: C:\Program Files\dotnet\sdk\6.0.301\ Host (useful for support): Version: 6.0.6 Commit: 7cca709db2 ``` ### Other information I'm not sure if the different System.Text.Json assembly version numbers between the .Net Framework and .Net Core versions is deliberate. If yes, I would like to know how to circumvent this issue.
non_main
system io filenotfoundexception thrown when using system text json in mixed net version solution assembly versions do not match description we have a fairly large code base which consist of different net version projects mostly and but a few net framework projects as well we are migrating all of these to but it will still take a lot of time before everything is done the issue we are facing is with system text json we have installed the nuget package to the projects that require it but since couple of them are net framework projects it seems that they get a different version of the system text json assembly for other project types the assembly version is but for the legacy framework it is this causes runtime error system io filenotfoundexception could not load file or assembly system text json version culture neutral publickeytoken the system cannot find the file specified when the framework project is referenced by a project yes i know it s not ideal but at the time being a necessary evil reproduction steps create a and a projects add a system text json reference to the project and reference that project from the project run the and it ll throw an system io filenotfoundexception here is a sample solution for this expected behavior i would expect the nuget for system text json to have the same assembly version for different net versions actual behavior runtime error system io filenotfoundexception could not load file or assembly system text json version culture neutral publickeytoken the system cannot find the file specified regression no response known workarounds if i downgrade the projects to use the nuget this works as expected in the sample solution i can also workaround the issue by adding the following snippet to the program cs but i don t feel comfortable of trying to do this in our production code appdomain currentdomain assemblyresolve currentdomain assemblyresolve static system reflection assembly currentdomain assemblyresolve object sender resolveeventargs args if args name system text json version culture neutral publickeytoken return typeof system text json jsonserializer assembly return null configuration partial output of dotnet info net sdk reflecting any global json version commit runtime environment os name windows os version os platform windows rid base path c program files dotnet sdk host useful for support version commit other information i m not sure if the different system text json assembly version numbers between the net framework and net core versions is deliberate if yes i would like to know how to circumvent this issue
0
3,528
13,884,113,119
IssuesEvent
2020-10-18 14:52:43
grey-software/org
https://api.github.com/repos/grey-software/org
opened
🥅 Initiative: Grey Software Report Card
Domain: User Experience Role: Maintainer Role: Product Owner
### Motivation 🏁 Students need to show employers real-world software experience in order to stand out. A report card that outlines a student's open-source experience at Grey Software can allow them to stand apart from the competition. ### Initiative Overview 👁️‍🗨️ <!-- A clear and concise description of what the initiative is. --> Every contributor to Grey Software will be able to generate an authentic report card by entering their username into grey.software/report-card Contributors will be able to share a PDF of this report on websites like LinkedIn. A link-able version of their report card will also be available at grey.software/report-card/GITHUB_USERNAME if they choose to. **Implementation Details 🛠️** <!--- Please share a plan to help realize this initiative --> We'll add a page to our website under /report-card that will have an input box for Github usernames. Once users enter their usernames and click "Generate", our backend will scrape Github and query its API to generate a PDF of their report card. If they choose to (via a checkbox), they can host their report card publicly on our website. Additional Notes: - Students will have to authenticate with Guthub in order to generate a verified report card for their username - A user's insights for a repo will have to be scraped ### Impact 💥 Students struggling to demonstrate real-world software experience will now be able to show their open-source work in a professional manner. ### Describe alternatives you've considered 🔍 Students could link their contributor cards from the Github insights page for every repo they have contributed to. ### Additional details ℹ️ N/A
True
🥅 Initiative: Grey Software Report Card - ### Motivation 🏁 Students need to show employers real-world software experience in order to stand out. A report card that outlines a student's open-source experience at Grey Software can allow them to stand apart from the competition. ### Initiative Overview 👁️‍🗨️ <!-- A clear and concise description of what the initiative is. --> Every contributor to Grey Software will be able to generate an authentic report card by entering their username into grey.software/report-card Contributors will be able to share a PDF of this report on websites like LinkedIn. A link-able version of their report card will also be available at grey.software/report-card/GITHUB_USERNAME if they choose to. **Implementation Details 🛠️** <!--- Please share a plan to help realize this initiative --> We'll add a page to our website under /report-card that will have an input box for Github usernames. Once users enter their usernames and click "Generate", our backend will scrape Github and query its API to generate a PDF of their report card. If they choose to (via a checkbox), they can host their report card publicly on our website. Additional Notes: - Students will have to authenticate with Guthub in order to generate a verified report card for their username - A user's insights for a repo will have to be scraped ### Impact 💥 Students struggling to demonstrate real-world software experience will now be able to show their open-source work in a professional manner. ### Describe alternatives you've considered 🔍 Students could link their contributor cards from the Github insights page for every repo they have contributed to. ### Additional details ℹ️ N/A
main
🥅 initiative grey software report card motivation 🏁 students need to show employers real world software experience in order to stand out a report card that outlines a student s open source experience at grey software can allow them to stand apart from the competition initiative overview 👁️‍🗨️ a clear and concise description of what the initiative is every contributor to grey software will be able to generate an authentic report card by entering their username into grey software report card contributors will be able to share a pdf of this report on websites like linkedin a link able version of their report card will also be available at grey software report card github username if they choose to implementation details 🛠️ we ll add a page to our website under report card that will have an input box for github usernames once users enter their usernames and click generate our backend will scrape github and query its api to generate a pdf of their report card if they choose to via a checkbox they can host their report card publicly on our website additional notes students will have to authenticate with guthub in order to generate a verified report card for their username a user s insights for a repo will have to be scraped impact 💥 students struggling to demonstrate real world software experience will now be able to show their open source work in a professional manner describe alternatives you ve considered 🔍 students could link their contributor cards from the github insights page for every repo they have contributed to additional details ℹ️ n a
1
2,527
8,655,460,713
IssuesEvent
2018-11-27 16:00:36
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
init.d script for init.d-based systems
unmaintained
Hello there! It's kinda very long ago that I have actually made an init.d script, especially having forgotten the LSB tags that go at the top... But as I run my machine as a headless home-server, I would like to have QCMA start up, when it starts. It's an old Mac Mini running Ubuntu 18.04 now, so it would easily take scripts by the `init.d` format. Currently, I start `qcma_cli` via `screen` and just detach from the screen. This might be a viable method to really use for "daemonizing" it in a quick way. Also, I could not seem to really find any CLI switches for configuring the search folders for media... Any plans on tidying up the CLI a little? Kind regards, Ingwie
True
init.d script for init.d-based systems - Hello there! It's kinda very long ago that I have actually made an init.d script, especially having forgotten the LSB tags that go at the top... But as I run my machine as a headless home-server, I would like to have QCMA start up, when it starts. It's an old Mac Mini running Ubuntu 18.04 now, so it would easily take scripts by the `init.d` format. Currently, I start `qcma_cli` via `screen` and just detach from the screen. This might be a viable method to really use for "daemonizing" it in a quick way. Also, I could not seem to really find any CLI switches for configuring the search folders for media... Any plans on tidying up the CLI a little? Kind regards, Ingwie
main
init d script for init d based systems hello there it s kinda very long ago that i have actually made an init d script especially having forgotten the lsb tags that go at the top but as i run my machine as a headless home server i would like to have qcma start up when it starts it s an old mac mini running ubuntu now so it would easily take scripts by the init d format currently i start qcma cli via screen and just detach from the screen this might be a viable method to really use for daemonizing it in a quick way also i could not seem to really find any cli switches for configuring the search folders for media any plans on tidying up the cli a little kind regards ingwie
1
28,211
13,595,518,809
IssuesEvent
2020-09-22 03:20:50
layer5io/meshery
https://api.github.com/repos/layer5io/meshery
closed
Complete nighthawk integration with meshery modules
area/performance kind/enhancement
**Current Behavior** <!-- A brief description of what the problem is. (e.g. I need to be able to...) --> Nighthawk Server Interface has already been added but UI, Mesheryctl, SMPConfig & Other configurations needs to be exposed. **Desired Behavior** <!-- A brief description of the enhancement. --> Complete Nighthawk API Integration with all the modules of Meshery ---
True
Complete nighthawk integration with meshery modules - **Current Behavior** <!-- A brief description of what the problem is. (e.g. I need to be able to...) --> Nighthawk Server Interface has already been added but UI, Mesheryctl, SMPConfig & Other configurations needs to be exposed. **Desired Behavior** <!-- A brief description of the enhancement. --> Complete Nighthawk API Integration with all the modules of Meshery ---
non_main
complete nighthawk integration with meshery modules current behavior nighthawk server interface has already been added but ui mesheryctl smpconfig other configurations needs to be exposed desired behavior complete nighthawk api integration with all the modules of meshery
0
3,720
15,382,341,792
IssuesEvent
2021-03-03 00:26:19
Homebrew/homebrew-cask
https://api.github.com/repos/Homebrew/homebrew-cask
closed
No mountable filesystems when my cask extracts an app from a .dmg.zip file
awaiting maintainer feedback stale
Hi all, I have a problem with writing a cask for an application often used by my colleagues. The upstream app is supplied in a zipped DMG file (suffixed by .dmg.zip). I noticed several available casks that work simply and perfectly from such a .dmg.zip file (e.g. [creepy.rb](https://github.com/Homebrew/homebrew-cask/blob/master/Casks/creepy.rb)) I wrote something similar: ``` cask "macghostview" do version "6.1" sha256 "b20270fffea09bb4ca6c19fa41a18e8ef0071402a4b8ba17927f3da3c6f848d3" url "https://www.math.tamu.edu/~tkiffe/tex/programs/MacGhostView#{version.no_dots}.dmg.zip" appcast "https://www.math.tamu.edu/~tkiffe/macghostview.html" name "MacGhostView" desc "Application for previewing Postscript and encapsulated Postscript files" homepage "https://www.math.tamu.edu/~tkiffe/macghostview.html" app "MacGhostView.app" end ``` Surprinsingly, after downloading from the URL, `brew install --cask` fails when mounting the downloaded result, ``` Error: Failure while executing; `hdiutil attach -plist -nobrowse -readonly -mountrandom /var/folders/9p/ztbq9wsn6195z4c5d25_7b340000gq/T/d20201220-71972-19mlccu /var/folders/9p/ztbq9wsn6195z4c5d25_7b340000gq/T/d20201220-71972-19mlccu/MacGhostView61.cdr` exited with 1. Here's the output: hdiutil: attach failed - no mountable filesystems ``` Don't know if the `MacGhostView61.cdr` in the failure output above (and not .dmg as it is inside the zip) can be a clue to explain the failure? Nevertheless: - downloading and mounting a .dmg.zip file works in other cask scripts (mentioned above) - rewriting my cask to download the dmg file (manually unzipped with macOS Finder integrated function (Archive Utility), and uploaded to my own server) mounts perfectly. I tried to play a little with `container type:`and `container nested:` (mentioned [here](https://github.com/Homebrew/homebrew-cask/blob/master/doc/cask_language_reference/all_stanzas.md)) in my cask, but to no avail. Have you any idea for the mount failure when directly using the .dmp.zip file? Would there be limitations in zip-format support in cask subroutines different from the Finder function? Any help welcome, best regards.
True
No mountable filesystems when my cask extracts an app from a .dmg.zip file - Hi all, I have a problem with writing a cask for an application often used by my colleagues. The upstream app is supplied in a zipped DMG file (suffixed by .dmg.zip). I noticed several available casks that work simply and perfectly from such a .dmg.zip file (e.g. [creepy.rb](https://github.com/Homebrew/homebrew-cask/blob/master/Casks/creepy.rb)) I wrote something similar: ``` cask "macghostview" do version "6.1" sha256 "b20270fffea09bb4ca6c19fa41a18e8ef0071402a4b8ba17927f3da3c6f848d3" url "https://www.math.tamu.edu/~tkiffe/tex/programs/MacGhostView#{version.no_dots}.dmg.zip" appcast "https://www.math.tamu.edu/~tkiffe/macghostview.html" name "MacGhostView" desc "Application for previewing Postscript and encapsulated Postscript files" homepage "https://www.math.tamu.edu/~tkiffe/macghostview.html" app "MacGhostView.app" end ``` Surprinsingly, after downloading from the URL, `brew install --cask` fails when mounting the downloaded result, ``` Error: Failure while executing; `hdiutil attach -plist -nobrowse -readonly -mountrandom /var/folders/9p/ztbq9wsn6195z4c5d25_7b340000gq/T/d20201220-71972-19mlccu /var/folders/9p/ztbq9wsn6195z4c5d25_7b340000gq/T/d20201220-71972-19mlccu/MacGhostView61.cdr` exited with 1. Here's the output: hdiutil: attach failed - no mountable filesystems ``` Don't know if the `MacGhostView61.cdr` in the failure output above (and not .dmg as it is inside the zip) can be a clue to explain the failure? Nevertheless: - downloading and mounting a .dmg.zip file works in other cask scripts (mentioned above) - rewriting my cask to download the dmg file (manually unzipped with macOS Finder integrated function (Archive Utility), and uploaded to my own server) mounts perfectly. I tried to play a little with `container type:`and `container nested:` (mentioned [here](https://github.com/Homebrew/homebrew-cask/blob/master/doc/cask_language_reference/all_stanzas.md)) in my cask, but to no avail. Have you any idea for the mount failure when directly using the .dmp.zip file? Would there be limitations in zip-format support in cask subroutines different from the Finder function? Any help welcome, best regards.
main
no mountable filesystems when my cask extracts an app from a dmg zip file hi all i have a problem with writing a cask for an application often used by my colleagues the upstream app is supplied in a zipped dmg file suffixed by dmg zip i noticed several available casks that work simply and perfectly from such a dmg zip file e g i wrote something similar cask macghostview do version url appcast name macghostview desc application for previewing postscript and encapsulated postscript files homepage app macghostview app end surprinsingly after downloading from the url brew install cask fails when mounting the downloaded result error failure while executing hdiutil attach plist nobrowse readonly mountrandom var folders t var folders t cdr exited with here s the output hdiutil attach failed no mountable filesystems don t know if the cdr in the failure output above and not dmg as it is inside the zip can be a clue to explain the failure nevertheless downloading and mounting a dmg zip file works in other cask scripts mentioned above rewriting my cask to download the dmg file manually unzipped with macos finder integrated function archive utility and uploaded to my own server mounts perfectly i tried to play a little with container type and container nested mentioned in my cask but to no avail have you any idea for the mount failure when directly using the dmp zip file would there be limitations in zip format support in cask subroutines different from the finder function any help welcome best regards
1
2,512
8,655,459,981
IssuesEvent
2018-11-27 16:00:32
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
PSTV Doesn't Detect QCMA
unmaintained
Hello, I am running QCMA on Debian Stretch ARM via Crouton on a Chromebook. I have installed VitaMTP with QCMA. Everything appears to be running, I have the icon on my panel, but I can't get my Vita to see my device. I have turned them both off and on, made sure to start QCMA before my PSTV, and even tried reinstalling. It doesn't seem to work. Naturally, I assumed it was something funky with my computer, as I have an odd setup, but the verbose (http://pastebin.com/SgJYKwgD) seems rather normal. Any assistance is greatly appreciated.
True
PSTV Doesn't Detect QCMA - Hello, I am running QCMA on Debian Stretch ARM via Crouton on a Chromebook. I have installed VitaMTP with QCMA. Everything appears to be running, I have the icon on my panel, but I can't get my Vita to see my device. I have turned them both off and on, made sure to start QCMA before my PSTV, and even tried reinstalling. It doesn't seem to work. Naturally, I assumed it was something funky with my computer, as I have an odd setup, but the verbose (http://pastebin.com/SgJYKwgD) seems rather normal. Any assistance is greatly appreciated.
main
pstv doesn t detect qcma hello i am running qcma on debian stretch arm via crouton on a chromebook i have installed vitamtp with qcma everything appears to be running i have the icon on my panel but i can t get my vita to see my device i have turned them both off and on made sure to start qcma before my pstv and even tried reinstalling it doesn t seem to work naturally i assumed it was something funky with my computer as i have an odd setup but the verbose seems rather normal any assistance is greatly appreciated
1
41,624
5,345,827,309
IssuesEvent
2017-02-17 17:59:10
18F/calc
https://api.github.com/repos/18F/calc
closed
Use bugs for status indicators
in progress priority: low skill: design
For the user's own price lists, bugs like the WDS uses for their "alpha" and "beta" component indicators might be really helpful: https://standards.usa.gov/labels/
1.0
Use bugs for status indicators - For the user's own price lists, bugs like the WDS uses for their "alpha" and "beta" component indicators might be really helpful: https://standards.usa.gov/labels/
non_main
use bugs for status indicators for the user s own price lists bugs like the wds uses for their alpha and beta component indicators might be really helpful
0
270,455
20,601,260,510
IssuesEvent
2022-03-06 09:44:13
Kira272921/solidity-quickstart
https://api.github.com/repos/Kira272921/solidity-quickstart
closed
📝docs: improve the `README.md`
documentation
Improve the [`README.md`](https://github.com/Kira272921/solidity-quickstart/blob/main/README.md) file by creating a table of contents for all the markdown guide files and the example solidity files present in the [`contract`](https://github.com/Kira272921/solidity-quickstart/tree/main/contracts) folder.
1.0
📝docs: improve the `README.md` - Improve the [`README.md`](https://github.com/Kira272921/solidity-quickstart/blob/main/README.md) file by creating a table of contents for all the markdown guide files and the example solidity files present in the [`contract`](https://github.com/Kira272921/solidity-quickstart/tree/main/contracts) folder.
non_main
📝docs improve the readme md improve the file by creating a table of contents for all the markdown guide files and the example solidity files present in the folder
0
58,401
24,439,278,147
IssuesEvent
2022-10-06 13:37:17
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
.NET not supported by OpenAI?
cognitive-services/svc triaged assigned-to-author product-question Pri1
I expected to see at least C# since GitHub CoPilot does support it. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 25c15e01-3578-c4ae-afbc-32c7172173d0 * Version Independent ID: 0d20fad8-0e13-2d85-64e5-31861fabdcdd * Content: [Azure OpenAI Engines - Azure OpenAI](https://docs.microsoft.com/en-us/azure/cognitive-services/openai/concepts/engines) * Content Source: [articles/cognitive-services/openai/concepts/engines.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/cognitive-services/openai/concepts/engines.md) * Service: **cognitive-services** * GitHub Login: @mrbullwinkle * Microsoft Alias: **mbullwin**
1.0
.NET not supported by OpenAI? - I expected to see at least C# since GitHub CoPilot does support it. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 25c15e01-3578-c4ae-afbc-32c7172173d0 * Version Independent ID: 0d20fad8-0e13-2d85-64e5-31861fabdcdd * Content: [Azure OpenAI Engines - Azure OpenAI](https://docs.microsoft.com/en-us/azure/cognitive-services/openai/concepts/engines) * Content Source: [articles/cognitive-services/openai/concepts/engines.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/cognitive-services/openai/concepts/engines.md) * Service: **cognitive-services** * GitHub Login: @mrbullwinkle * Microsoft Alias: **mbullwin**
non_main
net not supported by openai i expected to see at least c since github copilot does support it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id afbc version independent id content content source service cognitive services github login mrbullwinkle microsoft alias mbullwin
0
2,847
10,219,571,377
IssuesEvent
2019-08-15 18:56:08
arcticicestudio/styleguide-javascript
https://api.github.com/repos/arcticicestudio/styleguide-javascript
closed
Prettier
context-workflow scope-dx scope-maintainability scope-quality type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/48644231-4556d780-e9e2-11e8-862e-e8ce630fd0ba.png" width="30%" /></p> > Epic: #8 Integrate [Prettier][], the opinionated code formatter with support for many languages and integrations with most editors. It ensures that all outputted code conforms to a consistent style. ### Configuration This is one of the main features of Prettier: It already provides the best and recommended style configurations of-out-the-box™. The only option we will change is the [print width][prettier-docs-pwidth]. It is set to 80 by default which not up-to-date for modern screens (might only be relevant when working in terminals only like e.g. with Vim). It'll be changed to 120 used by all of Arctic Ice Studio's style guides. The `prettier.config.js` configuration file will be placed in the project root as well as the `.prettierignore` file to also define ignore pattern. ### NPM script/task To allow to format all sources a `format:pretty` npm script/task will be added to be included in the main `format` script flow. ### False-Positives To ensure incorrect examples of the style guide won't be fixed by Prettier, the affected lines must be excluded from Prettier by [adding the `<!-- prettier-ignore -->` handle for HTML][p-d-ign]. Note that this might trigger `remark-lint` when added right above a code block (`no-missing-blank-lines`). This can also be fixed by [adding the `<!--lint disable no-missing-blank-lines-->` handle][rml-ign] as well. ## Tasks - [x] Install [prettier][npm-prettier] packages. - [x] Implement `prettier.config.js` configuration file. - [x] Implement `.prettierignore` ignore pattern file. - [x] Implement NPM `format:pretty` script/task. - [x] Format current code base for the first time and fix possible style guide violations using the configured linters of the project. - [x] Ensure compatibility with Prettier (#11) by adding required _ignore_/_disable_ handles for Prettier and `remark-lint`. [npm-prettier]: https://www.npmjs.com/package/prettier [prettier-docs-pwidth]: https://prettier.io/docs/en/options.html#print-width [prettier]: https://prettier.io [p-d-ign]: https://prettier.io/docs/en/ignore.html#html [rml-ign]: https://github.com/remarkjs/remark-lint#configuring-remark-lint
True
Prettier - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48644231-4556d780-e9e2-11e8-862e-e8ce630fd0ba.png" width="30%" /></p> > Epic: #8 Integrate [Prettier][], the opinionated code formatter with support for many languages and integrations with most editors. It ensures that all outputted code conforms to a consistent style. ### Configuration This is one of the main features of Prettier: It already provides the best and recommended style configurations of-out-the-box™. The only option we will change is the [print width][prettier-docs-pwidth]. It is set to 80 by default which not up-to-date for modern screens (might only be relevant when working in terminals only like e.g. with Vim). It'll be changed to 120 used by all of Arctic Ice Studio's style guides. The `prettier.config.js` configuration file will be placed in the project root as well as the `.prettierignore` file to also define ignore pattern. ### NPM script/task To allow to format all sources a `format:pretty` npm script/task will be added to be included in the main `format` script flow. ### False-Positives To ensure incorrect examples of the style guide won't be fixed by Prettier, the affected lines must be excluded from Prettier by [adding the `<!-- prettier-ignore -->` handle for HTML][p-d-ign]. Note that this might trigger `remark-lint` when added right above a code block (`no-missing-blank-lines`). This can also be fixed by [adding the `<!--lint disable no-missing-blank-lines-->` handle][rml-ign] as well. ## Tasks - [x] Install [prettier][npm-prettier] packages. - [x] Implement `prettier.config.js` configuration file. - [x] Implement `.prettierignore` ignore pattern file. - [x] Implement NPM `format:pretty` script/task. - [x] Format current code base for the first time and fix possible style guide violations using the configured linters of the project. - [x] Ensure compatibility with Prettier (#11) by adding required _ignore_/_disable_ handles for Prettier and `remark-lint`. [npm-prettier]: https://www.npmjs.com/package/prettier [prettier-docs-pwidth]: https://prettier.io/docs/en/options.html#print-width [prettier]: https://prettier.io [p-d-ign]: https://prettier.io/docs/en/ignore.html#html [rml-ign]: https://github.com/remarkjs/remark-lint#configuring-remark-lint
main
prettier epic integrate the opinionated code formatter with support for many languages and integrations with most editors it ensures that all outputted code conforms to a consistent style configuration this is one of the main features of prettier it already provides the best and recommended style configurations of out the box™ the only option we will change is the it is set to by default which not up to date for modern screens might only be relevant when working in terminals only like e g with vim it ll be changed to used by all of arctic ice studio s style guides the prettier config js configuration file will be placed in the project root as well as the prettierignore file to also define ignore pattern npm script task to allow to format all sources a format pretty npm script task will be added to be included in the main format script flow false positives to ensure incorrect examples of the style guide won t be fixed by prettier the affected lines must be excluded from prettier by note that this might trigger remark lint when added right above a code block no missing blank lines this can also be fixed by as well tasks install packages implement prettier config js configuration file implement prettierignore ignore pattern file implement npm format pretty script task format current code base for the first time and fix possible style guide violations using the configured linters of the project ensure compatibility with prettier by adding required ignore disable handles for prettier and remark lint
1
211,842
16,460,133,609
IssuesEvent
2021-05-21 17:39:10
brotkrueml/typo3-matomo-widgets
https://api.github.com/repos/brotkrueml/typo3-matomo-widgets
opened
Add possibility to refresh widgets
documentation feature
In TYPO3 v11.3 a new feature was introduced to refresh the widget content: https://docs.typo3.org/c/typo3/cms-core/master/en-us/Changelog/master/Feature-93210-PossibilityToRefreshDashboardWidgets.html As this would be also a nice feature for the Matomo widgets, it should be available for TYPO3 v11. In TYPO3 v10 this option has no effect. - [ ] The refresh action button is available for all available widgets with the exception of "Link to Matomo": - [ ] When the refresh action button is hit, the cache is bypassed and the metrics for this widget updated. - [ ] The documentation is adjusted. - [ ] An entry to the changelog is available.
1.0
Add possibility to refresh widgets - In TYPO3 v11.3 a new feature was introduced to refresh the widget content: https://docs.typo3.org/c/typo3/cms-core/master/en-us/Changelog/master/Feature-93210-PossibilityToRefreshDashboardWidgets.html As this would be also a nice feature for the Matomo widgets, it should be available for TYPO3 v11. In TYPO3 v10 this option has no effect. - [ ] The refresh action button is available for all available widgets with the exception of "Link to Matomo": - [ ] When the refresh action button is hit, the cache is bypassed and the metrics for this widget updated. - [ ] The documentation is adjusted. - [ ] An entry to the changelog is available.
non_main
add possibility to refresh widgets in a new feature was introduced to refresh the widget content as this would be also a nice feature for the matomo widgets it should be available for in this option has no effect the refresh action button is available for all available widgets with the exception of link to matomo when the refresh action button is hit the cache is bypassed and the metrics for this widget updated the documentation is adjusted an entry to the changelog is available
0
30,164
2,722,937,603
IssuesEvent
2015-04-14 08:59:09
BlackCodec/tint2
https://api.github.com/repos/BlackCodec/tint2
closed
Unhide panel by click
auto-migrated Component-Panel Priority-Medium Type-Enhancement
``` What version of the product are you using? On what operating system? tint2 0.11+svn2011 amd64 on Debian Sid Please provide any additional information below. in some very cluttered interfaces it's easy to unexpectedly unhide a panel. It's very annoying when I drag something in editor and panel appears exactly over thing where I would like to drop this thing. Will be nice to have option to make panel appear only when use click with mouse over the edge where panel is hidden ``` Original issue reported on code.google.com by `akhil...@gmail.com` on 23 Aug 2012 at 6:11
1.0
Unhide panel by click - ``` What version of the product are you using? On what operating system? tint2 0.11+svn2011 amd64 on Debian Sid Please provide any additional information below. in some very cluttered interfaces it's easy to unexpectedly unhide a panel. It's very annoying when I drag something in editor and panel appears exactly over thing where I would like to drop this thing. Will be nice to have option to make panel appear only when use click with mouse over the edge where panel is hidden ``` Original issue reported on code.google.com by `akhil...@gmail.com` on 23 Aug 2012 at 6:11
non_main
unhide panel by click what version of the product are you using on what operating system on debian sid please provide any additional information below in some very cluttered interfaces it s easy to unexpectedly unhide a panel it s very annoying when i drag something in editor and panel appears exactly over thing where i would like to drop this thing will be nice to have option to make panel appear only when use click with mouse over the edge where panel is hidden original issue reported on code google com by akhil gmail com on aug at
0
882
4,543,471,170
IssuesEvent
2016-09-10 05:00:16
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_container not preserving case for boolean environmental variables
affects_2.1 bug_report cloud docker waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> docker_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> Default configuration is being used. ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Running on Arch Linux, managing Ubuntu ##### SUMMARY <!--- Explain the problem briefly --> When passing an environmental variable to docker using docker_container, a key with a value of _true_ will be passed to Docker as _True_ instead of _true_. Placing _true_ in quotes fixes this issue. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: Restart gitlab docker_container: name: gitlab image: sameersbn/gitlab state: started recreate: yes pull: true ports: - "443:443" - "80:80" - "22:22" links: - "redis:redisio" - "postgresql:postgresql" env: GITLAB_HTTPS: "true" # Works properly GITLAB_HTTPS: true # Passes True to Docker instead of true ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The GITLAB_HTTPS variable should be set to _true_ ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> The GITLAB_HTTPS variable is set to _True_, breaking the container's configuration. <!--- Paste verbatim command output between quotes below --> ``` ```
True
docker_container not preserving case for boolean environmental variables - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> docker_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> Default configuration is being used. ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Running on Arch Linux, managing Ubuntu ##### SUMMARY <!--- Explain the problem briefly --> When passing an environmental variable to docker using docker_container, a key with a value of _true_ will be passed to Docker as _True_ instead of _true_. Placing _true_ in quotes fixes this issue. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: Restart gitlab docker_container: name: gitlab image: sameersbn/gitlab state: started recreate: yes pull: true ports: - "443:443" - "80:80" - "22:22" links: - "redis:redisio" - "postgresql:postgresql" env: GITLAB_HTTPS: "true" # Works properly GITLAB_HTTPS: true # Passes True to Docker instead of true ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The GITLAB_HTTPS variable should be set to _true_ ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> The GITLAB_HTTPS variable is set to _True_, breaking the container's configuration. <!--- Paste verbatim command output between quotes below --> ``` ```
main
docker container not preserving case for boolean environmental variables issue type bug report component name docker container ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables default configuration is being used os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific running on arch linux managing ubuntu summary when passing an environmental variable to docker using docker container a key with a value of true will be passed to docker as true instead of true placing true in quotes fixes this issue steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name restart gitlab docker container name gitlab image sameersbn gitlab state started recreate yes pull true ports links redis redisio postgresql postgresql env gitlab https true works properly gitlab https true passes true to docker instead of true expected results the gitlab https variable should be set to true actual results the gitlab https variable is set to true breaking the container s configuration
1
76,864
9,967,196,254
IssuesEvent
2019-07-08 13:06:26
cekit/cekit
https://api.github.com/repos/cekit/cekit
closed
Extend local develoment guidelines to cover including changes in modules
complexity/medium priority/medium type/documentation
Currently we do not have any information what would be the best practice to modify a module and what would be expected to be rebuilt locally after this change. Sometimes the image needs to be rebuilt entirely and this takes time. We need to explain it and suggest modules orders. It is hard to do since every image is different, but we should be able to put together some general guidelines.
1.0
Extend local develoment guidelines to cover including changes in modules - Currently we do not have any information what would be the best practice to modify a module and what would be expected to be rebuilt locally after this change. Sometimes the image needs to be rebuilt entirely and this takes time. We need to explain it and suggest modules orders. It is hard to do since every image is different, but we should be able to put together some general guidelines.
non_main
extend local develoment guidelines to cover including changes in modules currently we do not have any information what would be the best practice to modify a module and what would be expected to be rebuilt locally after this change sometimes the image needs to be rebuilt entirely and this takes time we need to explain it and suggest modules orders it is hard to do since every image is different but we should be able to put together some general guidelines
0
1,258
5,332,661,671
IssuesEvent
2017-02-15 22:41:29
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Bug report: info command is case sensitive
awaiting maintainer feedback
#### Description of issue Running `brew cask info <Foo>` fails where `brew cask info <foo>` does not. <details><summary>Output of `brew cask <command> --verbose`</summary> ```console $ brew cask search Pandora ==> Exact match pandora $ brew cask info Pandora --verbose Error: Cask 'Pandora' definition is invalid: Bad header line: 'pandora' does not match file name Error: Kernel.exit $ brew cask info pandora --verbose pandora: 2.0.10 https://www.pandora.com/ /usr/local/Caskroom/pandora/2.0.10 (68B) From: https://github.com/caskroom/homebrew-cask/blob/master/Casks/pandora.rb ==> Name Pandora ==> Artifacts Pandora.app (app) ``` </details> <details><summary>Output of `brew doctor`</summary> ``` Your system is ready to brew. ``` </details> <details><summary>Output of `brew cask doctor`</summary> ``` ==> Homebrew-Cask Version Homebrew-Cask 1.1.10 caskroom/homebrew-cask (git revision 7958; last commit 2017-02-14) ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Cached Downloads ~/Library/Caches/Homebrew/Cask (10 files, 420.9M) ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3546 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin16 /Library/Ruby/Site/2.0.0/universal-darwin16 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16 ==> Environment Variables LANG="en_US.UTF-8" PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/Homebrew/Library/Homebrew/cask/cmd:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/bash" ``` </details>
True
Bug report: info command is case sensitive - #### Description of issue Running `brew cask info <Foo>` fails where `brew cask info <foo>` does not. <details><summary>Output of `brew cask <command> --verbose`</summary> ```console $ brew cask search Pandora ==> Exact match pandora $ brew cask info Pandora --verbose Error: Cask 'Pandora' definition is invalid: Bad header line: 'pandora' does not match file name Error: Kernel.exit $ brew cask info pandora --verbose pandora: 2.0.10 https://www.pandora.com/ /usr/local/Caskroom/pandora/2.0.10 (68B) From: https://github.com/caskroom/homebrew-cask/blob/master/Casks/pandora.rb ==> Name Pandora ==> Artifacts Pandora.app (app) ``` </details> <details><summary>Output of `brew doctor`</summary> ``` Your system is ready to brew. ``` </details> <details><summary>Output of `brew cask doctor`</summary> ``` ==> Homebrew-Cask Version Homebrew-Cask 1.1.10 caskroom/homebrew-cask (git revision 7958; last commit 2017-02-14) ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Cached Downloads ~/Library/Caches/Homebrew/Cask (10 files, 420.9M) ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3546 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin16 /Library/Ruby/Site/2.0.0/universal-darwin16 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16 ==> Environment Variables LANG="en_US.UTF-8" PATH="/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/local/Homebrew/Library/Homebrew/cask/cmd:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/bin/bash" ``` </details>
main
bug report info command is case sensitive description of issue running brew cask info fails where brew cask info does not output of brew cask verbose console brew cask search pandora exact match pandora brew cask info pandora verbose error cask pandora definition is invalid bad header line pandora does not match file name error kernel exit brew cask info pandora verbose pandora usr local caskroom pandora from name pandora artifacts pandora app app output of brew doctor your system is ready to brew output of brew cask doctor homebrew cask version homebrew cask caskroom homebrew cask git revision last commit homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask cached downloads library caches homebrew cask files homebrew cask taps usr local homebrew library taps caskroom homebrew cask casks contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal environment variables lang en us utf path usr local sbin usr local bin usr sbin usr bin sbin bin usr local homebrew library homebrew cask cmd usr local homebrew library homebrew shims scm shell bin bash
1
4,707
24,270,829,079
IssuesEvent
2022-09-28 10:07:26
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
SEO | hreflang conflicts within page source code 
engineering Maintain
This set of pages have been flagged with having issues with the format of the HREFLANG source code. However, there is not a clear issue to correct for these pages and may be appearing error due to the URL parameters.. The issues is identifed by having one of the following issues: - Conflicting hreflang and rel=canonical URLs, Conflicting hreflang URLs, No self-referencing hreflang URLs https://docs.google.com/spreadsheets/d/15HwgpxSYc4Zl809kcebAhLfLYXFuIk8ZP-Qvk3yVV8Q/edit#gid=1772715630
True
SEO | hreflang conflicts within page source code  - This set of pages have been flagged with having issues with the format of the HREFLANG source code. However, there is not a clear issue to correct for these pages and may be appearing error due to the URL parameters.. The issues is identifed by having one of the following issues: - Conflicting hreflang and rel=canonical URLs, Conflicting hreflang URLs, No self-referencing hreflang URLs https://docs.google.com/spreadsheets/d/15HwgpxSYc4Zl809kcebAhLfLYXFuIk8ZP-Qvk3yVV8Q/edit#gid=1772715630
main
seo hreflang conflicts within page source code  this set of pages have been flagged with having issues with the format of the hreflang source code however there is not a clear issue to correct for these pages and may be appearing error due to the url parameters the issues is identifed by having one of the following issues conflicting hreflang and rel canonical urls conflicting hreflang urls no self referencing hreflang urls
1
157,736
13,721,888,386
IssuesEvent
2020-10-03 00:44:17
odwdinc/Python-SimConnect
https://api.github.com/repos/odwdinc/Python-SimConnect
closed
Question : FSX support ?
documentation good first issue
Thank you for all the work that has been put in this library. The question if fairly simple, does this library supports interfacing with FSX:SE ? If not, will this be a feature or a no-go ? If yes, is it hard to do ? How could I help ?
1.0
Question : FSX support ? - Thank you for all the work that has been put in this library. The question if fairly simple, does this library supports interfacing with FSX:SE ? If not, will this be a feature or a no-go ? If yes, is it hard to do ? How could I help ?
non_main
question fsx support thank you for all the work that has been put in this library the question if fairly simple does this library supports interfacing with fsx se if not will this be a feature or a no go if yes is it hard to do how could i help
0
312,595
26,873,403,781
IssuesEvent
2023-02-04 18:55:28
MPMG-DCC-UFMG/F01
https://api.github.com/repos/MPMG-DCC-UFMG/F01
closed
Teste de generalizacao para a tag Informações Institucionais - Leis Municipais - Igaratinga
generalization test development template - Betha (26) tag - Informações Institucionais subtag - Leis Municipais
DoD: Realizar o teste de Generalização do validador da tag Informações Institucionais - Leis Municipais para o Município de Igaratinga.
1.0
Teste de generalizacao para a tag Informações Institucionais - Leis Municipais - Igaratinga - DoD: Realizar o teste de Generalização do validador da tag Informações Institucionais - Leis Municipais para o Município de Igaratinga.
non_main
teste de generalizacao para a tag informações institucionais leis municipais igaratinga dod realizar o teste de generalização do validador da tag informações institucionais leis municipais para o município de igaratinga
0
2,956
10,616,383,349
IssuesEvent
2019-10-12 11:17:28
arcticicestudio/snowsaw
https://api.github.com/repos/arcticicestudio/snowsaw
closed
Drop cross-compilation of FreeBSD binary artifacts
context-workflow scope-compatibility scope-maintainability scope-stability type-task
There are currently too many compatibility problems with some dependencies due to native bindings. Therefore only builds for the main OS and architectures (Linux, macOS and Windows) should be compiled for now which covers almost all users.
True
Drop cross-compilation of FreeBSD binary artifacts - There are currently too many compatibility problems with some dependencies due to native bindings. Therefore only builds for the main OS and architectures (Linux, macOS and Windows) should be compiled for now which covers almost all users.
main
drop cross compilation of freebsd binary artifacts there are currently too many compatibility problems with some dependencies due to native bindings therefore only builds for the main os and architectures linux macos and windows should be compiled for now which covers almost all users
1
217,415
16,705,670,468
IssuesEvent
2021-06-09 09:39:34
awslabs/deequ
https://api.github.com/repos/awslabs/deequ
closed
Documentation for core concepts in our library
documentation
Do we need a separate markdown file where we explain all core concepts like analyzers, checks, constraints, metrics, column profiles, metric repositories, check levels etc.? We should also have a list of all analyzers, constraints, anomaly detection methods and constraint suggestion rules we offer in one place. Maybe even a readthedocs-documentation?
1.0
Documentation for core concepts in our library - Do we need a separate markdown file where we explain all core concepts like analyzers, checks, constraints, metrics, column profiles, metric repositories, check levels etc.? We should also have a list of all analyzers, constraints, anomaly detection methods and constraint suggestion rules we offer in one place. Maybe even a readthedocs-documentation?
non_main
documentation for core concepts in our library do we need a separate markdown file where we explain all core concepts like analyzers checks constraints metrics column profiles metric repositories check levels etc we should also have a list of all analyzers constraints anomaly detection methods and constraint suggestion rules we offer in one place maybe even a readthedocs documentation
0
328,256
28,108,304,150
IssuesEvent
2023-03-31 04:01:56
mindspore-lab/mindocr
https://api.github.com/repos/mindspore-lab/mindocr
opened
`build_model` doesn't support pretrained checkpoint loading from the config file
inside-test
`build_model` doesn't support pretrained checkpoint loading from the config file. It expects the pretrained checkpoint path in [kwargs](https://github.com/mindspore-lab/mindocr/blob/52bd1af1d99232e18102760ecceb71f05e61ab1e/mindocr/models/builder.py#L58) instead.
1.0
`build_model` doesn't support pretrained checkpoint loading from the config file - `build_model` doesn't support pretrained checkpoint loading from the config file. It expects the pretrained checkpoint path in [kwargs](https://github.com/mindspore-lab/mindocr/blob/52bd1af1d99232e18102760ecceb71f05e61ab1e/mindocr/models/builder.py#L58) instead.
non_main
build model doesn t support pretrained checkpoint loading from the config file build model doesn t support pretrained checkpoint loading from the config file it expects the pretrained checkpoint path in instead
0
2,988
3,995,617,790
IssuesEvent
2016-05-10 16:02:29
Comcast/traffic_control
https://api.github.com/repos/Comcast/traffic_control
closed
TC: Ansible Playbooks - Kibana
enhancement Infrastructure
Add Ansible playbook for Kibana Acceptance Criteria - Add Kibana Role - Test that playbooks run correctly - Document
1.0
TC: Ansible Playbooks - Kibana - Add Ansible playbook for Kibana Acceptance Criteria - Add Kibana Role - Test that playbooks run correctly - Document
non_main
tc ansible playbooks kibana add ansible playbook for kibana acceptance criteria add kibana role test that playbooks run correctly document
0
768
4,375,656,331
IssuesEvent
2016-08-05 00:57:38
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
nxos_config exception calling module_utils.netcfg.NetworkConfig.difference() with arg 'path'
bug_report networking waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> nxos_config ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 config file = /home/ec2-user/proj-ansible-dev/ansible-bm-statecontrol/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> Complete ansible.cfg follows... ``` $ cat /home/ec2-user/proj-ansible-dev/ansible-bm-statecontrol/ansible.cfg [defaults] # default inventory points to non-prod, just in case... inventory=./inven-lab transport=local gathering=explicit ``` ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Running Ansible from: ``` Red Hat Enterprise Linux Server 7.2 (Maipo) Python 2.7.5 ``` running in a virtualenv Managing a Cisco Nexus switch ##### SUMMARY <!--- Explain the problem briefly --> nxos_config failing with exception msg: "MODULE FAILURE" while calling ```module_utils.netcfg.NetworkConfig.difference()``` nxos_cfg calls difference() method with the 'path' argument. The difference() method does not accept the 'path' argument as of '@80ab80b' ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Run playbook below <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: Basic Verification to Cisco Devices hosts: akldcb-acc-7s pre_tasks: - debug: var=hostvars[inventory_hostname] tasks: - name: "Configure network interfaces" nxos_config: lines: - "no shutdown" parents: ['interface Ethernet1/44'] provider: { host: "akldcb-acc-7s", ssh_keyfile: "/home/ec2-user/.ssh/sys_Networkautomate_rsa", username: "sys_Networkautomate", } ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Configuration is correctly applied to the Nexus switch. The playbook verified as working in Ansible v2.1 ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Failing with exception message "MODULE FAILURE" <!--- Paste verbatim command output between quotes below --> ``` TASK [Configure network interfaces] ******************************************** task path: /home/ec2-user/proj-ansible-dev/ansible-bm-statecontrol/cisco_basic_check.yaml:9 Using module file /home/ec2-user/proj-ansible-dev/env/lib/python2.7/site-packages/ansible/modules/core/network/nxos/nxos_config.py <akldcb-acc-7s> ESTABLISH LOCAL CONNECTION FOR USER: ec2-user <akldcb-acc-7s> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569 `" && echo ansible-tmp-1470310590.63-137000207474569="` echo $HOME/.ansible/tm p/ansible-tmp-1470310590.63-137000207474569 `" ) && sleep 0' <akldcb-acc-7s> PUT /tmp/tmpPoERwO TO /home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569/nxos_config.py <akldcb-acc-7s> EXEC /bin/sh -c 'chmod -R u+x /home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569/ && sleep 0' <akldcb-acc-7s> EXEC /bin/sh -c '/usr/bin/python /home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569/nxos_config.py; rm -rf "/home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000 207474569/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_hzmwpP/ansible_module_nxos_config.py", line 229, in <module> main() File "/tmp/ansible_hzmwpP/ansible_module_nxos_config.py", line 200, in main commands = candidate.difference(config, path=parents, match=match, replace=replace) TypeError: difference() got an unexpected keyword argument 'path' fatal: [akldcb-acc-7s]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "nxos_config" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_hzmwpP/ansible_module_nxos_config.py\", line 229, in <module>\n main()\n File \"/tmp/ansible_hzmwpP/ansible_module_nxos _config.py\", line 200, in main\n commands = candidate.difference(config, path=parents, match=match, replace=replace)\nTypeError: difference() got an unexpected keyword argument 'path'\n", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false } ```
True
nxos_config exception calling module_utils.netcfg.NetworkConfig.difference() with arg 'path' - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> nxos_config ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 config file = /home/ec2-user/proj-ansible-dev/ansible-bm-statecontrol/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> Complete ansible.cfg follows... ``` $ cat /home/ec2-user/proj-ansible-dev/ansible-bm-statecontrol/ansible.cfg [defaults] # default inventory points to non-prod, just in case... inventory=./inven-lab transport=local gathering=explicit ``` ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Running Ansible from: ``` Red Hat Enterprise Linux Server 7.2 (Maipo) Python 2.7.5 ``` running in a virtualenv Managing a Cisco Nexus switch ##### SUMMARY <!--- Explain the problem briefly --> nxos_config failing with exception msg: "MODULE FAILURE" while calling ```module_utils.netcfg.NetworkConfig.difference()``` nxos_cfg calls difference() method with the 'path' argument. The difference() method does not accept the 'path' argument as of '@80ab80b' ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Run playbook below <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: Basic Verification to Cisco Devices hosts: akldcb-acc-7s pre_tasks: - debug: var=hostvars[inventory_hostname] tasks: - name: "Configure network interfaces" nxos_config: lines: - "no shutdown" parents: ['interface Ethernet1/44'] provider: { host: "akldcb-acc-7s", ssh_keyfile: "/home/ec2-user/.ssh/sys_Networkautomate_rsa", username: "sys_Networkautomate", } ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Configuration is correctly applied to the Nexus switch. The playbook verified as working in Ansible v2.1 ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Failing with exception message "MODULE FAILURE" <!--- Paste verbatim command output between quotes below --> ``` TASK [Configure network interfaces] ******************************************** task path: /home/ec2-user/proj-ansible-dev/ansible-bm-statecontrol/cisco_basic_check.yaml:9 Using module file /home/ec2-user/proj-ansible-dev/env/lib/python2.7/site-packages/ansible/modules/core/network/nxos/nxos_config.py <akldcb-acc-7s> ESTABLISH LOCAL CONNECTION FOR USER: ec2-user <akldcb-acc-7s> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569 `" && echo ansible-tmp-1470310590.63-137000207474569="` echo $HOME/.ansible/tm p/ansible-tmp-1470310590.63-137000207474569 `" ) && sleep 0' <akldcb-acc-7s> PUT /tmp/tmpPoERwO TO /home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569/nxos_config.py <akldcb-acc-7s> EXEC /bin/sh -c 'chmod -R u+x /home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569/ && sleep 0' <akldcb-acc-7s> EXEC /bin/sh -c '/usr/bin/python /home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000207474569/nxos_config.py; rm -rf "/home/ec2-user/.ansible/tmp/ansible-tmp-1470310590.63-137000 207474569/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_hzmwpP/ansible_module_nxos_config.py", line 229, in <module> main() File "/tmp/ansible_hzmwpP/ansible_module_nxos_config.py", line 200, in main commands = candidate.difference(config, path=parents, match=match, replace=replace) TypeError: difference() got an unexpected keyword argument 'path' fatal: [akldcb-acc-7s]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "nxos_config" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_hzmwpP/ansible_module_nxos_config.py\", line 229, in <module>\n main()\n File \"/tmp/ansible_hzmwpP/ansible_module_nxos _config.py\", line 200, in main\n commands = candidate.difference(config, path=parents, match=match, replace=replace)\nTypeError: difference() got an unexpected keyword argument 'path'\n", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false } ```
main
nxos config exception calling module utils netcfg networkconfig difference with arg path issue type bug report component name nxos config ansible version ansible config file home user proj ansible dev ansible bm statecontrol ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables complete ansible cfg follows cat home user proj ansible dev ansible bm statecontrol ansible cfg default inventory points to non prod just in case inventory inven lab transport local gathering explicit os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific running ansible from red hat enterprise linux server maipo python running in a virtualenv managing a cisco nexus switch summary nxos config failing with exception msg module failure while calling module utils netcfg networkconfig difference nxos cfg calls difference method with the path argument the difference method does not accept the path argument as of steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used run playbook below name basic verification to cisco devices hosts akldcb acc pre tasks debug var hostvars tasks name configure network interfaces nxos config lines no shutdown parents provider host akldcb acc ssh keyfile home user ssh sys networkautomate rsa username sys networkautomate expected results configuration is correctly applied to the nexus switch the playbook verified as working in ansible actual results failing with exception message module failure task task path home user proj ansible dev ansible bm statecontrol cisco basic check yaml using module file home user proj ansible dev env lib site packages ansible modules core network nxos nxos config py establish local connection for user user exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tm p ansible tmp sleep put tmp tmppoerwo to home user ansible tmp ansible tmp nxos config py exec bin sh c chmod r u x home user ansible tmp ansible tmp sleep exec bin sh c usr bin python home user ansible tmp ansible tmp nxos config py rm rf home user ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible hzmwpp ansible module nxos config py line in main file tmp ansible hzmwpp ansible module nxos config py line in main commands candidate difference config path parents match match replace replace typeerror difference got an unexpected keyword argument path fatal failed changed false failed true invocation module name nxos config module stderr traceback most recent call last n file tmp ansible hzmwpp ansible module nxos config py line in n main n file tmp ansible hzmwpp ansible module nxos config py line in main n commands candidate difference config path parents match match replace replace ntypeerror difference got an unexpected keyword argument path n module stdout msg module failure parsed false
1
2,709
9,531,849,462
IssuesEvent
2019-04-29 17:01:47
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
Cannot connect to network
unmaintained
i get a "cannot connect to the network" notification after using qcma once This is remedied by restarting qcma but it happens again. I am on archlinux compiled from AUR
True
Cannot connect to network - i get a "cannot connect to the network" notification after using qcma once This is remedied by restarting qcma but it happens again. I am on archlinux compiled from AUR
main
cannot connect to network i get a cannot connect to the network notification after using qcma once this is remedied by restarting qcma but it happens again i am on archlinux compiled from aur
1
39,034
8,568,556,516
IssuesEvent
2018-11-10 22:46:57
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
[LLVM 7] use funnel shift intrinsic to implement rotates
A-LLVM C-enhancement WG-codegen
https://reviews.llvm.org/rL337221 introduced new intrinsics in LLVM IR that make it easier to guarantee that integer rotates will make it to the backend intact and can be lowered to a hardware rotate instruction (instead of possibly being split up into shifts and bitwise ors). Since we already expose rotates as methods on integer types, we could just change the implementation of those methods to lower to these IR intrinsics. As soon as we're on an LLVM version that supports these intrinsics, that is, but I wanted to file this issue before I forget about them. cc @eddyb @nox
1.0
[LLVM 7] use funnel shift intrinsic to implement rotates - https://reviews.llvm.org/rL337221 introduced new intrinsics in LLVM IR that make it easier to guarantee that integer rotates will make it to the backend intact and can be lowered to a hardware rotate instruction (instead of possibly being split up into shifts and bitwise ors). Since we already expose rotates as methods on integer types, we could just change the implementation of those methods to lower to these IR intrinsics. As soon as we're on an LLVM version that supports these intrinsics, that is, but I wanted to file this issue before I forget about them. cc @eddyb @nox
non_main
use funnel shift intrinsic to implement rotates introduced new intrinsics in llvm ir that make it easier to guarantee that integer rotates will make it to the backend intact and can be lowered to a hardware rotate instruction instead of possibly being split up into shifts and bitwise ors since we already expose rotates as methods on integer types we could just change the implementation of those methods to lower to these ir intrinsics as soon as we re on an llvm version that supports these intrinsics that is but i wanted to file this issue before i forget about them cc eddyb nox
0
468,800
13,490,910,292
IssuesEvent
2020-09-11 15:45:11
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
Ability to save resource with thumbnail not allowed
Accepted Priority: Medium bug
## Description <!-- Add here a few sentences describing the bug. --> When the thumbnail with size (over 500Kb) or extensions (.gif) not allowed is added on save window, the warning message is displayed but the resource is saved anyway. ![issue56](https://user-images.githubusercontent.com/56537133/72820359-1fc10000-3c6f-11ea-8549-aa8a11f2c8bd.gif) ## How to reproduce <!-- A list of steps to reproduce the bug --> - click on new map - click on save - change the thumbnail with over 500Kb or a .gif - click on save - go back to the homepage - search the map *Expected Result* <!-- Describe here the expected result --> The map cannot be saved if the thumbnail is not supported and the relative warning message appears *Current Result* <!-- Describe here the current behavior --> The map is saved anyway and it is present on the homepage without any thumbnail - [x] Not browser related <details><summary> <b>Browser info</b> </summary> <!-- If browser related, please compile the following table --> <!-- If your browser is not in the list please add a new row to the table with the version --> (use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users) | Browser Affected | Version | |---|---| |Internet Explorer| | |Edge| | |Chrome| | |Firefox| | |Safari| | </details> ## Other useful information <!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
1.0
Ability to save resource with thumbnail not allowed - ## Description <!-- Add here a few sentences describing the bug. --> When the thumbnail with size (over 500Kb) or extensions (.gif) not allowed is added on save window, the warning message is displayed but the resource is saved anyway. ![issue56](https://user-images.githubusercontent.com/56537133/72820359-1fc10000-3c6f-11ea-8549-aa8a11f2c8bd.gif) ## How to reproduce <!-- A list of steps to reproduce the bug --> - click on new map - click on save - change the thumbnail with over 500Kb or a .gif - click on save - go back to the homepage - search the map *Expected Result* <!-- Describe here the expected result --> The map cannot be saved if the thumbnail is not supported and the relative warning message appears *Current Result* <!-- Describe here the current behavior --> The map is saved anyway and it is present on the homepage without any thumbnail - [x] Not browser related <details><summary> <b>Browser info</b> </summary> <!-- If browser related, please compile the following table --> <!-- If your browser is not in the list please add a new row to the table with the version --> (use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users) | Browser Affected | Version | |---|---| |Internet Explorer| | |Edge| | |Chrome| | |Firefox| | |Safari| | </details> ## Other useful information <!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
non_main
ability to save resource with thumbnail not allowed description when the thumbnail with size over or extensions gif not allowed is added on save window the warning message is displayed but the resource is saved anyway how to reproduce click on new map click on save change the thumbnail with over or a gif click on save go back to the homepage search the map expected result the map cannot be saved if the thumbnail is not supported and the relative warning message appears current result the map is saved anyway and it is present on the homepage without any thumbnail not browser related browser info use this site a href for non expert users browser affected version internet explorer edge chrome firefox safari other useful information
0
1,750
6,574,944,280
IssuesEvent
2017-09-11 14:34:19
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Bug Report: Fetch fails if ansible_ssh_host is localhost
affects_2.1 bug_report waiting_on_maintainer
ISSUE TYPE Bug Report COMPONENT NAME fetch ANSIBLE VERSION 2.1.0.0 CONFIGURATION These are the most relevant config items though I don't know that there is a correlation: ssh_args = -o ControlMaster=auto -o ControlPersist=60s OS / ENVIRONMENT N/A SUMMARY When ansible_ssh_host is set to localhost, fetch says it succeeds, but it never gets the file. It's important that localhost works the same as every other host value for testing purposes. I previously opened an issue for this (https://github.com/ansible/ansible-modules-core/issues/4814) but believe it to be closed in error. The symptoms are the same as another issue in which the user didn't have enough space on the target filesystem, but that is not the root cause in this case. STEPS TO REPRODUCE Use the play below with ansible_ssh_host set to localhost hosts: '{{ hosts }}' gather_facts: False tasks: fetch: src: /tmp/remote_file dest: /tmp/local_file flat: true fail_on_missing: true EXPECTED RESULTS I would expect the behavior to be the same for localhost as it is for every other host. ACTUAL RESULTS Fetch says it succeeds, but verbose output actually shows this error (doesn't matter if the file exists or not). ok: [localhost] => {"changed": false, "file": "/tmp/remote_file", "invocation": {"module_args": {"dest": "/tmp/local_file", "fail_on_missing": true, "flat": true, "src": "/tmp/remote_file"}, "module_name": "fetch"}, "msg": "unable to calculate the checksum of the remote file"}
True
Bug Report: Fetch fails if ansible_ssh_host is localhost - ISSUE TYPE Bug Report COMPONENT NAME fetch ANSIBLE VERSION 2.1.0.0 CONFIGURATION These are the most relevant config items though I don't know that there is a correlation: ssh_args = -o ControlMaster=auto -o ControlPersist=60s OS / ENVIRONMENT N/A SUMMARY When ansible_ssh_host is set to localhost, fetch says it succeeds, but it never gets the file. It's important that localhost works the same as every other host value for testing purposes. I previously opened an issue for this (https://github.com/ansible/ansible-modules-core/issues/4814) but believe it to be closed in error. The symptoms are the same as another issue in which the user didn't have enough space on the target filesystem, but that is not the root cause in this case. STEPS TO REPRODUCE Use the play below with ansible_ssh_host set to localhost hosts: '{{ hosts }}' gather_facts: False tasks: fetch: src: /tmp/remote_file dest: /tmp/local_file flat: true fail_on_missing: true EXPECTED RESULTS I would expect the behavior to be the same for localhost as it is for every other host. ACTUAL RESULTS Fetch says it succeeds, but verbose output actually shows this error (doesn't matter if the file exists or not). ok: [localhost] => {"changed": false, "file": "/tmp/remote_file", "invocation": {"module_args": {"dest": "/tmp/local_file", "fail_on_missing": true, "flat": true, "src": "/tmp/remote_file"}, "module_name": "fetch"}, "msg": "unable to calculate the checksum of the remote file"}
main
bug report fetch fails if ansible ssh host is localhost issue type bug report component name fetch ansible version configuration these are the most relevant config items though i don t know that there is a correlation ssh args o controlmaster auto o controlpersist os environment n a summary when ansible ssh host is set to localhost fetch says it succeeds but it never gets the file it s important that localhost works the same as every other host value for testing purposes i previously opened an issue for this but believe it to be closed in error the symptoms are the same as another issue in which the user didn t have enough space on the target filesystem but that is not the root cause in this case steps to reproduce use the play below with ansible ssh host set to localhost hosts hosts gather facts false tasks fetch src tmp remote file dest tmp local file flat true fail on missing true expected results i would expect the behavior to be the same for localhost as it is for every other host actual results fetch says it succeeds but verbose output actually shows this error doesn t matter if the file exists or not ok changed false file tmp remote file invocation module args dest tmp local file fail on missing true flat true src tmp remote file module name fetch msg unable to calculate the checksum of the remote file
1
179,708
30,286,224,103
IssuesEvent
2023-07-08 18:10:53
nextcloud/server
https://api.github.com/repos/nextcloud/server
closed
No printer icon present in official logos
enhancement design 0. Needs triage
<!-- Thanks for reporting issues back to Nextcloud! Note: This is the **issue tracker of Nextcloud**, please do NOT use this to get answers to your questions or get help for fixing your installation. This is a place to report bugs to developers, after your server has been debugged. You can find help debugging your system on our home user forums: https://help.nextcloud.com or, if you use Nextcloud in a large organization, ask our engineers on https://portal.nextcloud.com. See also https://nextcloud.com/support for support options. Nextcloud is an open source project backed by Nextcloud GmbH. Most of our volunteers are home users and thus primarily care about issues that affect home users. Our paid engineers prioritize issues of our customers. If you are neither a home user nor a customer, consider paying somebody to fix your issue, do it yourself or become a customer. Guidelines for submitting issues: * Please search the existing issues first, it's likely that your issue was already reported or even fixed. - Go to https://github.com/nextcloud and type any word in the top search/command bar. You probably see something like "We couldn’t find any repositories matching ..." then click "Issues" in the left navigation. - You can also filter by appending e. g. "state:open" to the search string. - More info on search syntax within github: https://help.github.com/articles/searching-issues * This repository https://github.com/nextcloud/server/issues is *only* for issues within the Nextcloud Server code. This also includes the apps: files, encryption, external storage, sharing, deleted files, versions, LDAP, and WebDAV Auth * SECURITY: Report any potential security bug to us via our HackerOne page (https://hackerone.com/nextcloud) following our security policy (https://nextcloud.com/security/) instead of filing an issue in our bug tracker. * The issues in other components should be reported in their respective repositories: You will find them in our GitHub Organization (https://github.com/nextcloud/) --> <!--- Please keep this note for other contributors --> ### How to use GitHub * Please use the 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to show that you are interested into the same feature. * Please don't comment if you have no relevant information to add. It's just extra noise for everyone subscribed to this issue. * Subscribe to receive notifications on status change and new comments. **Is your feature request related to a problem? Please describe.** Sort of. I am working on some issues on a NC app and found that there is no official icon for a printer. It would be very nice, if you could provide a logo with a printer/printout or something like that. In [the dev manual](https://docs.nextcloud.com/server/22/developer_manual/design/icons.html) no printer icon is present. Reference to the issue: https://github.com/nextcloud/cookbook/issues/728 **Describe the solution you'd like** A logo visually compatible with the other logos ready to be used by app developers. **Describe alternatives you've considered** We could create and add our own logo. The risk is that if other apps start to allow for printing, there is a mixture of logos in the world. If the chances are low that any other apps start printing and we should design something on ourself, feel free to tell us. We will do this then.
1.0
No printer icon present in official logos - <!-- Thanks for reporting issues back to Nextcloud! Note: This is the **issue tracker of Nextcloud**, please do NOT use this to get answers to your questions or get help for fixing your installation. This is a place to report bugs to developers, after your server has been debugged. You can find help debugging your system on our home user forums: https://help.nextcloud.com or, if you use Nextcloud in a large organization, ask our engineers on https://portal.nextcloud.com. See also https://nextcloud.com/support for support options. Nextcloud is an open source project backed by Nextcloud GmbH. Most of our volunteers are home users and thus primarily care about issues that affect home users. Our paid engineers prioritize issues of our customers. If you are neither a home user nor a customer, consider paying somebody to fix your issue, do it yourself or become a customer. Guidelines for submitting issues: * Please search the existing issues first, it's likely that your issue was already reported or even fixed. - Go to https://github.com/nextcloud and type any word in the top search/command bar. You probably see something like "We couldn’t find any repositories matching ..." then click "Issues" in the left navigation. - You can also filter by appending e. g. "state:open" to the search string. - More info on search syntax within github: https://help.github.com/articles/searching-issues * This repository https://github.com/nextcloud/server/issues is *only* for issues within the Nextcloud Server code. This also includes the apps: files, encryption, external storage, sharing, deleted files, versions, LDAP, and WebDAV Auth * SECURITY: Report any potential security bug to us via our HackerOne page (https://hackerone.com/nextcloud) following our security policy (https://nextcloud.com/security/) instead of filing an issue in our bug tracker. * The issues in other components should be reported in their respective repositories: You will find them in our GitHub Organization (https://github.com/nextcloud/) --> <!--- Please keep this note for other contributors --> ### How to use GitHub * Please use the 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to show that you are interested into the same feature. * Please don't comment if you have no relevant information to add. It's just extra noise for everyone subscribed to this issue. * Subscribe to receive notifications on status change and new comments. **Is your feature request related to a problem? Please describe.** Sort of. I am working on some issues on a NC app and found that there is no official icon for a printer. It would be very nice, if you could provide a logo with a printer/printout or something like that. In [the dev manual](https://docs.nextcloud.com/server/22/developer_manual/design/icons.html) no printer icon is present. Reference to the issue: https://github.com/nextcloud/cookbook/issues/728 **Describe the solution you'd like** A logo visually compatible with the other logos ready to be used by app developers. **Describe alternatives you've considered** We could create and add our own logo. The risk is that if other apps start to allow for printing, there is a mixture of logos in the world. If the chances are low that any other apps start printing and we should design something on ourself, feel free to tell us. We will do this then.
non_main
no printer icon present in official logos thanks for reporting issues back to nextcloud note this is the issue tracker of nextcloud please do not use this to get answers to your questions or get help for fixing your installation this is a place to report bugs to developers after your server has been debugged you can find help debugging your system on our home user forums or if you use nextcloud in a large organization ask our engineers on see also for support options nextcloud is an open source project backed by nextcloud gmbh most of our volunteers are home users and thus primarily care about issues that affect home users our paid engineers prioritize issues of our customers if you are neither a home user nor a customer consider paying somebody to fix your issue do it yourself or become a customer guidelines for submitting issues please search the existing issues first it s likely that your issue was already reported or even fixed go to and type any word in the top search command bar you probably see something like we couldn’t find any repositories matching then click issues in the left navigation you can also filter by appending e g state open to the search string more info on search syntax within github this repository is only for issues within the nextcloud server code this also includes the apps files encryption external storage sharing deleted files versions ldap and webdav auth security report any potential security bug to us via our hackerone page following our security policy instead of filing an issue in our bug tracker the issues in other components should be reported in their respective repositories you will find them in our github organization how to use github please use the 👍 to show that you are interested into the same feature please don t comment if you have no relevant information to add it s just extra noise for everyone subscribed to this issue subscribe to receive notifications on status change and new comments is your feature request related to a problem please describe sort of i am working on some issues on a nc app and found that there is no official icon for a printer it would be very nice if you could provide a logo with a printer printout or something like that in no printer icon is present reference to the issue describe the solution you d like a logo visually compatible with the other logos ready to be used by app developers describe alternatives you ve considered we could create and add our own logo the risk is that if other apps start to allow for printing there is a mixture of logos in the world if the chances are low that any other apps start printing and we should design something on ourself feel free to tell us we will do this then
0
7,128
10,276,342,225
IssuesEvent
2019-08-24 16:28:06
xethya/framework
https://api.github.com/repos/xethya/framework
opened
[publish process] Publish packages as soon as they're merged into master
publish process
We could use something like this: https://github.com/counterfactual/monorepo/blob/master/.circleci/config.yml#L116-L136 https://github.com/counterfactual/monorepo/blob/master/package.json#L24
1.0
[publish process] Publish packages as soon as they're merged into master - We could use something like this: https://github.com/counterfactual/monorepo/blob/master/.circleci/config.yml#L116-L136 https://github.com/counterfactual/monorepo/blob/master/package.json#L24
non_main
publish packages as soon as they re merged into master we could use something like this
0
5,696
30,009,495,764
IssuesEvent
2023-06-26 14:25:32
conbench/conbench
https://api.github.com/repos/conbench/conbench
opened
webapp: skip API layer re-serialization in compare/<> pages
maintainability
It's pretty much impossible to change UI presentation around here: https://github.com/conbench/conbench/blob/f1815a0c99211b1df2407f5f96fe069e0d5ddf89/conbench/app/compare.py#L116 because these objects: ``` comparisons=comparisons, regressions=regressions, improvements=improvements, baseline_id=baseline_id, contender_id=contender_id, baseline=baseline, contender=contender, baseline_run=baseline_run, contender_run=contender_run, ``` are coming from the API layer, and are kind of intransparent. Of course we could modify those on the fly for different UI presentation, but we have taken a better direction elsewhere, where we start feeding more first-class/well-defined objects into the template engine: https://github.com/conbench/conbench/issues/968 https://github.com/conbench/conbench/issues/890
True
webapp: skip API layer re-serialization in compare/<> pages - It's pretty much impossible to change UI presentation around here: https://github.com/conbench/conbench/blob/f1815a0c99211b1df2407f5f96fe069e0d5ddf89/conbench/app/compare.py#L116 because these objects: ``` comparisons=comparisons, regressions=regressions, improvements=improvements, baseline_id=baseline_id, contender_id=contender_id, baseline=baseline, contender=contender, baseline_run=baseline_run, contender_run=contender_run, ``` are coming from the API layer, and are kind of intransparent. Of course we could modify those on the fly for different UI presentation, but we have taken a better direction elsewhere, where we start feeding more first-class/well-defined objects into the template engine: https://github.com/conbench/conbench/issues/968 https://github.com/conbench/conbench/issues/890
main
webapp skip api layer re serialization in compare pages it s pretty much impossible to change ui presentation around here because these objects comparisons comparisons regressions regressions improvements improvements baseline id baseline id contender id contender id baseline baseline contender contender baseline run baseline run contender run contender run are coming from the api layer and are kind of intransparent of course we could modify those on the fly for different ui presentation but we have taken a better direction elsewhere where we start feeding more first class well defined objects into the template engine
1
2,455
8,639,879,730
IssuesEvent
2018-11-23 22:19:43
F5OEO/rpitx
https://api.github.com/repos/F5OEO/rpitx
closed
rpitx process always running in the background
V1 related (not maintained)
Hi, some time ago I wrote about opening my garage door with rpitx… (https://github.com/F5OEO/rpitx/issues/50) Now, for several weeks the Raspberry Pi (3) lives in my garage, running 24/7. As I wrote in the other thread, a triggered script runs the rpitx command when I want to open the door. However, it seems as if the rpitx process isn’t properly ended when transmission ends, since the process is running on and on in the background and so the Pi consumes much more power, as it would in a real idle state… ![bildschirmfoto 2017-04-27 um 14 33 08](https://cloud.githubusercontent.com/assets/19333006/25483808/aa537ba2-2b57-11e7-8497-95244d5de526.png) What can I do about this? Is it a bug or is the rpitx process supposed to run forever when started once? Thanks so much in advance!
True
rpitx process always running in the background - Hi, some time ago I wrote about opening my garage door with rpitx… (https://github.com/F5OEO/rpitx/issues/50) Now, for several weeks the Raspberry Pi (3) lives in my garage, running 24/7. As I wrote in the other thread, a triggered script runs the rpitx command when I want to open the door. However, it seems as if the rpitx process isn’t properly ended when transmission ends, since the process is running on and on in the background and so the Pi consumes much more power, as it would in a real idle state… ![bildschirmfoto 2017-04-27 um 14 33 08](https://cloud.githubusercontent.com/assets/19333006/25483808/aa537ba2-2b57-11e7-8497-95244d5de526.png) What can I do about this? Is it a bug or is the rpitx process supposed to run forever when started once? Thanks so much in advance!
main
rpitx process always running in the background hi some time ago i wrote about opening my garage door with rpitx… now for several weeks the raspberry pi lives in my garage running as i wrote in the other thread a triggered script runs the rpitx command when i want to open the door however it seems as if the rpitx process isn’t properly ended when transmission ends since the process is running on and on in the background and so the pi consumes much more power as it would in a real idle state… what can i do about this is it a bug or is the rpitx process supposed to run forever when started once thanks so much in advance
1
101,825
21,790,748,013
IssuesEvent
2022-05-14 21:32:07
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
opened
[Unstable 0.18.1.0] Midround joining client got a update for a item which doesn't exist
Bug Code Networking
### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? Hosting a multiplayer campaign and a new client joined in middle of a round and each time they tried to spawn in they got a missing entity error. Had to wait until we reached the next level until they were able to join. Server [event_error_log_server_Renko_2109.log](https://github.com/Regalis11/Barotrauma/files/8693927/event_error_log_server_Renko_2109.log) [event_error_log_server_Renko_2108.log](https://github.com/Regalis11/Barotrauma/files/8693929/event_error_log_server_Renko_2108.log) Client [event_error_log_client_Renko_2109.log](https://github.com/Regalis11/Barotrauma/files/8693930/event_error_log_client_Renko_2109.log) [event_error_log_client_Renko_2108.log](https://github.com/Regalis11/Barotrauma/files/8693931/event_error_log_client_Renko_2108.log) ### Reproduction steps _No response_ ### Bug prevalence Happens every now and then ### Version 0.18.1.0 (Unstable) ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
1.0
[Unstable 0.18.1.0] Midround joining client got a update for a item which doesn't exist - ### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? Hosting a multiplayer campaign and a new client joined in middle of a round and each time they tried to spawn in they got a missing entity error. Had to wait until we reached the next level until they were able to join. Server [event_error_log_server_Renko_2109.log](https://github.com/Regalis11/Barotrauma/files/8693927/event_error_log_server_Renko_2109.log) [event_error_log_server_Renko_2108.log](https://github.com/Regalis11/Barotrauma/files/8693929/event_error_log_server_Renko_2108.log) Client [event_error_log_client_Renko_2109.log](https://github.com/Regalis11/Barotrauma/files/8693930/event_error_log_client_Renko_2109.log) [event_error_log_client_Renko_2108.log](https://github.com/Regalis11/Barotrauma/files/8693931/event_error_log_client_Renko_2108.log) ### Reproduction steps _No response_ ### Bug prevalence Happens every now and then ### Version 0.18.1.0 (Unstable) ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
non_main
midround joining client got a update for a item which doesn t exist disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened hosting a multiplayer campaign and a new client joined in middle of a round and each time they tried to spawn in they got a missing entity error had to wait until we reached the next level until they were able to join server client reproduction steps no response bug prevalence happens every now and then version unstable no response which operating system did you encounter this bug on windows relevant error messages and crash reports no response
0
1,479
6,415,441,167
IssuesEvent
2017-08-08 12:49:04
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
Support multiple NICs in vmware_vswitch.py module
affects_2.2 cloud feature_idea vmware waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> vmware_vswitch.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ansible 2.2.0.0 Python 2.7.12Python 2.7.12 pyvmomi (6.0.0.2016.6) ##### SUMMARY <!--- Explain the problem briefly --> Can you please add feature to add multiple NICs in vSwitch?
True
Support multiple NICs in vmware_vswitch.py module - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> vmware_vswitch.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ansible 2.2.0.0 Python 2.7.12Python 2.7.12 pyvmomi (6.0.0.2016.6) ##### SUMMARY <!--- Explain the problem briefly --> Can you please add feature to add multiple NICs in vSwitch?
main
support multiple nics in vmware vswitch py module issue type feature idea component name vmware vswitch py ansible version ansible python pyvmomi summary can you please add feature to add multiple nics in vswitch
1
66,707
27,556,151,857
IssuesEvent
2023-03-07 18:05:09
ovh/terraform-provider-ovh
https://api.github.com/repos/ovh/terraform-provider-ovh
closed
Unable to complete operation on subnet during Terraform destroy
Product: Managed Kubernetes Service Product: Network Status: Missing inputs
Hi, I'm running Terraform 1.38 that has created kubernetes cluster on a private network and subnet. However it is consistently failing at the final resources. The error message is Error: calling DELETE /cloud/project/id/network/private/netId/subnet/subnetId: "Unable to complete operation on subnet subnetId: One or more ports have an IP allocation from this subnet." terraform -v Terraform v1.3.8 on windows_amd64 + provider registry.terraform.io/hashicorp/helm v2.9.0 + provider registry.terraform.io/ovh/ovh v0.27.0 steps to produce 1. terraform apply create kubernetes cluster on private network and subnet. 2. terraform destroy. I beleve the ovh provider has a timing issue. It should probably retry with expotential back-off or is returning before the kubernetes cluster is actually deleted.
1.0
Unable to complete operation on subnet during Terraform destroy - Hi, I'm running Terraform 1.38 that has created kubernetes cluster on a private network and subnet. However it is consistently failing at the final resources. The error message is Error: calling DELETE /cloud/project/id/network/private/netId/subnet/subnetId: "Unable to complete operation on subnet subnetId: One or more ports have an IP allocation from this subnet." terraform -v Terraform v1.3.8 on windows_amd64 + provider registry.terraform.io/hashicorp/helm v2.9.0 + provider registry.terraform.io/ovh/ovh v0.27.0 steps to produce 1. terraform apply create kubernetes cluster on private network and subnet. 2. terraform destroy. I beleve the ovh provider has a timing issue. It should probably retry with expotential back-off or is returning before the kubernetes cluster is actually deleted.
non_main
unable to complete operation on subnet during terraform destroy hi i m running terraform that has created kubernetes cluster on a private network and subnet however it is consistently failing at the final resources the error message is error calling delete cloud project id network private netid subnet subnetid unable to complete operation on subnet subnetid one or more ports have an ip allocation from this subnet terraform v terraform on windows provider registry terraform io hashicorp helm provider registry terraform io ovh ovh steps to produce terraform apply create kubernetes cluster on private network and subnet terraform destroy i beleve the ovh provider has a timing issue it should probably retry with expotential back off or is returning before the kubernetes cluster is actually deleted
0
1,793
6,575,891,994
IssuesEvent
2017-09-11 17:43:58
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
pip always tell changed for some packages
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME pip ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION ``` [defaults] forks = 100 vault_password_file = ~/.vault.password [ssh_connection] pipelining = False ``` ##### OS / ENVIRONMENT ``` $ uname -a Linux dev4 3.16.0-4-amd64 #1 SMP Debian 3.16.7-ckt11-1+deb8u3 (2015-08-04) x86_64 GNU/Linux $ lsb_release -a No LSB modules are available. Distributor ID: Debian Description: Debian GNU/Linux 8.6 (jessie) Release: 8.6 Codename: jessie ``` ##### SUMMARY When installing python modules, some of them (always the same) keep reporting "changed". ##### STEPS TO REPRODUCE Not sure if reproductible but here is my playbook: ``` [...] - name: Install some python3 modules pip: name={{item}} state=latest executable=pip3 with_items: - aiohttp - asyncio - flake8 - jinja2 - pep8 - polib - pyflakes [...] ``` ##### EXPECTED RESULTS When the python module is already up to date, I'd like to get an "OK" instead of a changed. I'd like something like: ``` { "changed": false, "cmd": "/usr/bin/pip3 install -U pyflakes", "invocation": { "module_args": { "chdir": null, "editable": true, "executable": "pip3", "extra_args": null, "name": [ "pyflakes" ], "requirements": null, "state": "latest", "umask": null, "use_mirrors": true, "version": null, "virtualenv": null, "virtualenv_command": "vritualenv", "virtualenv_python": null, "virtualenv_site_packages": false } }, "name": [ "pyflakes" ], "requirements": null, "state": "latest", "stderr": "", "stdout": "Requirement already up-to-date: pyflakes in /usr/local/lib/python3.5/dist-packages\n", "version": null, "virtualenv": null } ``` ##### ACTUAL RESULTS Ansible tells it changed every time, I'm getting: ``` changed: [dev4] => (item=pyflakes) => { "changed": true, "cmd": "/usr/bin/pip3 install -U pyflakes", "invocation": { "module_args": { "chdir": null, "editable": true, "executable": "pip3", "extra_args": null, "name": "pyflakes", "requirements": null, "state": "latest", "umask": null, "use_mirrors": true, "version": null, "virtualenv": null, "virtualenv_command": "virtualenv", "virtualenv_python": null, "virtualenv_site_packages": false }, "module_name": "pip" }, "item": "pyflakes", "name": "pyflakes", "requirements": null, "state": "latest", "stderr": "", "stdout": "Collecting pyflakes\n Using cached pyflakes-1.3.0-py2.py3-none-any.whl\nInstalling collected packages: pyflakes\n Found existing installation: pyflakes 1.2.3\n Uninstalling pyflakes-1.2.3:\n Successfully uninstalled pyflakes-1.2.3\nSuccessfully installed pyflakes-1.3.0\n", "stdout_lines": [ "Collecting pyflakes", " Using cached pyflakes-1.3.0-py2.py3-none-any.whl", "Installing collected packages: pyflakes", " Found existing installation: pyflakes 1.2.3", " Uninstalling pyflakes-1.2.3:", " Successfully uninstalled pyflakes-1.2.3", "Successfully installed pyflakes-1.3.0" ], "version": null, "virtualenv": null } ``` The strange thing is that I'm receiving the 'changed: true' when running from ansible, and I'm receiving the 'changed: false' when running from `test-module` to try to understand myself what is broken: ``` ~julien/ansible/hacking/test-module -m /usr/local/lib/python2.7/dist-packages/ansible/modules/core/packaging/language/pip.py -a 'editable=true executable=pip3 name=pyflakes state=latest use_mirrors=true virtualenv_command=vritualenv virtualenv_site_packages=false' ``` So I'm unable to reproduce it, there may be a slight difference between the environment used by ansible and mine, but I can't easily spot it. For the sake of completness, I searched for the infamous pyflakes-1.2.3, without success: ``` # find /usr/ -name *pyflakes* /usr/local/bin/pyflakes /usr/local/lib/python2.7/dist-packages/pyflakes-1.3.0.dist-info /usr/local/lib/python2.7/dist-packages/pyflakes /usr/local/lib/python2.7/dist-packages/pyflakes/scripts/pyflakes.py /usr/local/lib/python2.7/dist-packages/pyflakes/scripts/pyflakes.pyc /usr/local/lib/python3.5/dist-packages/flake8/plugins/pyflakes.py /usr/local/lib/python3.5/dist-packages/flake8/plugins/__pycache__/pyflakes.cpython-35.pyc /usr/local/lib/python3.5/dist-packages/pyflakes-1.3.0.dist-info /usr/local/lib/python3.5/dist-packages/pyflakes /usr/local/lib/python3.5/dist-packages/pyflakes/scripts/pyflakes.py /usr/local/lib/python3.5/dist-packages/pyflakes/scripts/__pycache__/pyflakes.cpython-35.pyc ``` So I'm unable to understand where this is failing, where the pyflakes-1.2.3 is found, I'm out of ideas from here. I'd like to use strace but stracing ansible-playbook won't be of any help (it's ssh-ing on the machine, I won't be able to see what the process is remotely doing, only commucation between local and remote) and as I can't reproduce it with test-module I'm stuck. Obviously: ``` # pyflakes --version 1.3.0 ``` Any idea ?
True
pip always tell changed for some packages - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME pip ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION ``` [defaults] forks = 100 vault_password_file = ~/.vault.password [ssh_connection] pipelining = False ``` ##### OS / ENVIRONMENT ``` $ uname -a Linux dev4 3.16.0-4-amd64 #1 SMP Debian 3.16.7-ckt11-1+deb8u3 (2015-08-04) x86_64 GNU/Linux $ lsb_release -a No LSB modules are available. Distributor ID: Debian Description: Debian GNU/Linux 8.6 (jessie) Release: 8.6 Codename: jessie ``` ##### SUMMARY When installing python modules, some of them (always the same) keep reporting "changed". ##### STEPS TO REPRODUCE Not sure if reproductible but here is my playbook: ``` [...] - name: Install some python3 modules pip: name={{item}} state=latest executable=pip3 with_items: - aiohttp - asyncio - flake8 - jinja2 - pep8 - polib - pyflakes [...] ``` ##### EXPECTED RESULTS When the python module is already up to date, I'd like to get an "OK" instead of a changed. I'd like something like: ``` { "changed": false, "cmd": "/usr/bin/pip3 install -U pyflakes", "invocation": { "module_args": { "chdir": null, "editable": true, "executable": "pip3", "extra_args": null, "name": [ "pyflakes" ], "requirements": null, "state": "latest", "umask": null, "use_mirrors": true, "version": null, "virtualenv": null, "virtualenv_command": "vritualenv", "virtualenv_python": null, "virtualenv_site_packages": false } }, "name": [ "pyflakes" ], "requirements": null, "state": "latest", "stderr": "", "stdout": "Requirement already up-to-date: pyflakes in /usr/local/lib/python3.5/dist-packages\n", "version": null, "virtualenv": null } ``` ##### ACTUAL RESULTS Ansible tells it changed every time, I'm getting: ``` changed: [dev4] => (item=pyflakes) => { "changed": true, "cmd": "/usr/bin/pip3 install -U pyflakes", "invocation": { "module_args": { "chdir": null, "editable": true, "executable": "pip3", "extra_args": null, "name": "pyflakes", "requirements": null, "state": "latest", "umask": null, "use_mirrors": true, "version": null, "virtualenv": null, "virtualenv_command": "virtualenv", "virtualenv_python": null, "virtualenv_site_packages": false }, "module_name": "pip" }, "item": "pyflakes", "name": "pyflakes", "requirements": null, "state": "latest", "stderr": "", "stdout": "Collecting pyflakes\n Using cached pyflakes-1.3.0-py2.py3-none-any.whl\nInstalling collected packages: pyflakes\n Found existing installation: pyflakes 1.2.3\n Uninstalling pyflakes-1.2.3:\n Successfully uninstalled pyflakes-1.2.3\nSuccessfully installed pyflakes-1.3.0\n", "stdout_lines": [ "Collecting pyflakes", " Using cached pyflakes-1.3.0-py2.py3-none-any.whl", "Installing collected packages: pyflakes", " Found existing installation: pyflakes 1.2.3", " Uninstalling pyflakes-1.2.3:", " Successfully uninstalled pyflakes-1.2.3", "Successfully installed pyflakes-1.3.0" ], "version": null, "virtualenv": null } ``` The strange thing is that I'm receiving the 'changed: true' when running from ansible, and I'm receiving the 'changed: false' when running from `test-module` to try to understand myself what is broken: ``` ~julien/ansible/hacking/test-module -m /usr/local/lib/python2.7/dist-packages/ansible/modules/core/packaging/language/pip.py -a 'editable=true executable=pip3 name=pyflakes state=latest use_mirrors=true virtualenv_command=vritualenv virtualenv_site_packages=false' ``` So I'm unable to reproduce it, there may be a slight difference between the environment used by ansible and mine, but I can't easily spot it. For the sake of completness, I searched for the infamous pyflakes-1.2.3, without success: ``` # find /usr/ -name *pyflakes* /usr/local/bin/pyflakes /usr/local/lib/python2.7/dist-packages/pyflakes-1.3.0.dist-info /usr/local/lib/python2.7/dist-packages/pyflakes /usr/local/lib/python2.7/dist-packages/pyflakes/scripts/pyflakes.py /usr/local/lib/python2.7/dist-packages/pyflakes/scripts/pyflakes.pyc /usr/local/lib/python3.5/dist-packages/flake8/plugins/pyflakes.py /usr/local/lib/python3.5/dist-packages/flake8/plugins/__pycache__/pyflakes.cpython-35.pyc /usr/local/lib/python3.5/dist-packages/pyflakes-1.3.0.dist-info /usr/local/lib/python3.5/dist-packages/pyflakes /usr/local/lib/python3.5/dist-packages/pyflakes/scripts/pyflakes.py /usr/local/lib/python3.5/dist-packages/pyflakes/scripts/__pycache__/pyflakes.cpython-35.pyc ``` So I'm unable to understand where this is failing, where the pyflakes-1.2.3 is found, I'm out of ideas from here. I'd like to use strace but stracing ansible-playbook won't be of any help (it's ssh-ing on the machine, I won't be able to see what the process is remotely doing, only commucation between local and remote) and as I can't reproduce it with test-module I'm stuck. Obviously: ``` # pyflakes --version 1.3.0 ``` Any idea ?
main
pip always tell changed for some packages issue type bug report component name pip ansible version ansible config file configured module search path default w o overrides configuration forks vault password file vault password pipelining false os environment uname a linux smp debian gnu linux lsb release a no lsb modules are available distributor id debian description debian gnu linux jessie release codename jessie summary when installing python modules some of them always the same keep reporting changed steps to reproduce not sure if reproductible but here is my playbook name install some modules pip name item state latest executable with items aiohttp asyncio polib pyflakes expected results when the python module is already up to date i d like to get an ok instead of a changed i d like something like changed false cmd usr bin install u pyflakes invocation module args chdir null editable true executable extra args null name pyflakes requirements null state latest umask null use mirrors true version null virtualenv null virtualenv command vritualenv virtualenv python null virtualenv site packages false name pyflakes requirements null state latest stderr stdout requirement already up to date pyflakes in usr local lib dist packages n version null virtualenv null actual results ansible tells it changed every time i m getting changed item pyflakes changed true cmd usr bin install u pyflakes invocation module args chdir null editable true executable extra args null name pyflakes requirements null state latest umask null use mirrors true version null virtualenv null virtualenv command virtualenv virtualenv python null virtualenv site packages false module name pip item pyflakes name pyflakes requirements null state latest stderr stdout collecting pyflakes n using cached pyflakes none any whl ninstalling collected packages pyflakes n found existing installation pyflakes n uninstalling pyflakes n successfully uninstalled pyflakes nsuccessfully installed pyflakes n stdout lines collecting pyflakes using cached pyflakes none any whl installing collected packages pyflakes found existing installation pyflakes uninstalling pyflakes successfully uninstalled pyflakes successfully installed pyflakes version null virtualenv null the strange thing is that i m receiving the changed true when running from ansible and i m receiving the changed false when running from test module to try to understand myself what is broken julien ansible hacking test module m usr local lib dist packages ansible modules core packaging language pip py a editable true executable name pyflakes state latest use mirrors true virtualenv command vritualenv virtualenv site packages false so i m unable to reproduce it there may be a slight difference between the environment used by ansible and mine but i can t easily spot it for the sake of completness i searched for the infamous pyflakes without success find usr name pyflakes usr local bin pyflakes usr local lib dist packages pyflakes dist info usr local lib dist packages pyflakes usr local lib dist packages pyflakes scripts pyflakes py usr local lib dist packages pyflakes scripts pyflakes pyc usr local lib dist packages plugins pyflakes py usr local lib dist packages plugins pycache pyflakes cpython pyc usr local lib dist packages pyflakes dist info usr local lib dist packages pyflakes usr local lib dist packages pyflakes scripts pyflakes py usr local lib dist packages pyflakes scripts pycache pyflakes cpython pyc so i m unable to understand where this is failing where the pyflakes is found i m out of ideas from here i d like to use strace but stracing ansible playbook won t be of any help it s ssh ing on the machine i won t be able to see what the process is remotely doing only commucation between local and remote and as i can t reproduce it with test module i m stuck obviously pyflakes version any idea
1
149,978
5,732,370,989
IssuesEvent
2017-04-21 14:42:36
lucymonie/tic-tac-toe
https://api.github.com/repos/lucymonie/tic-tac-toe
opened
Built-in console.logs
chore priority-2 T25m
Add a function for rendering content so it's easy to change the output as required
1.0
Built-in console.logs - Add a function for rendering content so it's easy to change the output as required
non_main
built in console logs add a function for rendering content so it s easy to change the output as required
0
856
4,518,724,812
IssuesEvent
2016-09-06 00:54:56
Particular/ServiceInsight
https://api.github.com/repos/Particular/ServiceInsight
reopened
Sequence Diagram misplaces Handlers
Tag: Maintainer Prio Type: Bug
## Symptoms The Sequence Diagram misplaces the handler boxes in occasions and that results the message to appear out of order. ![screen1](https://cloud.githubusercontent.com/assets/8087733/13812222/2e7c7faa-eb7b-11e5-89ae-070987e5703e.png) ## Who's affected This update is recommended for all users of ServiceInsight who rely on Sequence Diagram as it draws a more accurate sequence of messages in certain situations.
True
Sequence Diagram misplaces Handlers - ## Symptoms The Sequence Diagram misplaces the handler boxes in occasions and that results the message to appear out of order. ![screen1](https://cloud.githubusercontent.com/assets/8087733/13812222/2e7c7faa-eb7b-11e5-89ae-070987e5703e.png) ## Who's affected This update is recommended for all users of ServiceInsight who rely on Sequence Diagram as it draws a more accurate sequence of messages in certain situations.
main
sequence diagram misplaces handlers symptoms the sequence diagram misplaces the handler boxes in occasions and that results the message to appear out of order who s affected this update is recommended for all users of serviceinsight who rely on sequence diagram as it draws a more accurate sequence of messages in certain situations
1
3,788
16,087,595,431
IssuesEvent
2021-04-26 13:12:04
ipfs/pinning-services-api-spec
https://api.github.com/repos/ipfs/pinning-services-api-spec
closed
Finalizing MVP API Spec for IPFS WebUI integration
P0 dif/expert effort/days kind/maintenance need/community-input need/maintainers-input
### About This issue tracks overall finalization status of this spec from the perspective of being ready for stakeholders to start implementation of basic functionality. cc @jacobheun @pooja @jessicaschilling ### Stakeholders - @jbenet as BDFL - IPFS Core Impl. WG implementing API client in go-ipfs and js-ipfs - IPFS GUI Team implementing UI in WebUI / IPFS Desktop app - Pinning Services implementing the API ### Remaining Issues While mostly ready, we need to resolve these issues before API can be fully implemented: - [x] Clear Milestone issues: https://github.com/ipfs/pinning-services-api-spec/milestone/1 - [ ] (descoped) <del>Additional endpoint for GUI apps that enables user onboarding without copying secrets as noted in https://github.com/ipfs/pinning-services-api-spec/issues/6#issuecomment-656404675 </del> ### Stakeholder Sign-offs - [x] @jbenet 👉 review scheduled - [x] IPFS Cluster / https://github.com/ipfs/notes/issues/378 (@lanzafame) - [ ] IPFS Core Impl. WG - [ ] IPFS GUI Team implementing UI in WebUI / IPFS Desktop app - [ ] IPFS Pinning Services - PS list TBD
True
Finalizing MVP API Spec for IPFS WebUI integration - ### About This issue tracks overall finalization status of this spec from the perspective of being ready for stakeholders to start implementation of basic functionality. cc @jacobheun @pooja @jessicaschilling ### Stakeholders - @jbenet as BDFL - IPFS Core Impl. WG implementing API client in go-ipfs and js-ipfs - IPFS GUI Team implementing UI in WebUI / IPFS Desktop app - Pinning Services implementing the API ### Remaining Issues While mostly ready, we need to resolve these issues before API can be fully implemented: - [x] Clear Milestone issues: https://github.com/ipfs/pinning-services-api-spec/milestone/1 - [ ] (descoped) <del>Additional endpoint for GUI apps that enables user onboarding without copying secrets as noted in https://github.com/ipfs/pinning-services-api-spec/issues/6#issuecomment-656404675 </del> ### Stakeholder Sign-offs - [x] @jbenet 👉 review scheduled - [x] IPFS Cluster / https://github.com/ipfs/notes/issues/378 (@lanzafame) - [ ] IPFS Core Impl. WG - [ ] IPFS GUI Team implementing UI in WebUI / IPFS Desktop app - [ ] IPFS Pinning Services - PS list TBD
main
finalizing mvp api spec for ipfs webui integration about this issue tracks overall finalization status of this spec from the perspective of being ready for stakeholders to start implementation of basic functionality cc jacobheun pooja jessicaschilling stakeholders jbenet as bdfl ipfs core impl wg implementing api client in go ipfs and js ipfs ipfs gui team implementing ui in webui ipfs desktop app pinning services implementing the api remaining issues while mostly ready we need to resolve these issues before api can be fully implemented clear milestone issues descoped additional endpoint for gui apps that enables user onboarding without copying secrets as noted in stakeholder sign offs jbenet 👉 review scheduled ipfs cluster lanzafame ipfs core impl wg ipfs gui team implementing ui in webui ipfs desktop app ipfs pinning services ps list tbd
1
480,786
13,866,938,601
IssuesEvent
2020-10-16 07:41:14
AY2021S1-CS2113-T13-1/tp
https://api.github.com/repos/AY2021S1-CS2113-T13-1/tp
reopened
Fix color support for windows command prompt
priority.High severity.Medium
Current printing of color is through ANSI escape code which is not supported by windows command prompt
1.0
Fix color support for windows command prompt - Current printing of color is through ANSI escape code which is not supported by windows command prompt
non_main
fix color support for windows command prompt current printing of color is through ansi escape code which is not supported by windows command prompt
0
2,181
7,660,035,524
IssuesEvent
2018-05-11 09:03:54
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
closed
Do not throw ArgumentException on descendants for parameterless methods
Area: analyzer Area: maintainability feature in progress
We should report a warning that a method that has no parameters is not allowed to throw an `ArgumentException` or any of its subclasses. The reason is that as there is no parameter aka. argument on the method, there is no need to throw an `ArgumentException`.
True
Do not throw ArgumentException on descendants for parameterless methods - We should report a warning that a method that has no parameters is not allowed to throw an `ArgumentException` or any of its subclasses. The reason is that as there is no parameter aka. argument on the method, there is no need to throw an `ArgumentException`.
main
do not throw argumentexception on descendants for parameterless methods we should report a warning that a method that has no parameters is not allowed to throw an argumentexception or any of its subclasses the reason is that as there is no parameter aka argument on the method there is no need to throw an argumentexception
1
28,152
2,700,056,438
IssuesEvent
2015-04-03 22:02:55
CenterForOpenScience/osf.io
https://api.github.com/repos/CenterForOpenScience/osf.io
closed
[production] Users with inconsistent `date_confirmed` and `date_last_login` fields
2 - Ready Bug: Production Priority - Low
# Expected All users who have logged in have a non-null value of `date_confirmed` # Observed In production, 18 users have non-null values of `date_last_login` but null values of `date_confirmed`. Based on a user report to @mckidwell, these users can't reset their passwords (but I verified manually that it's still possible to log in). To find users with inconsistent data, run: ```python User.find(Q('date_confirmed', 'eq', None) & Q('date_last_login', 'ne', None)) ``` To fix, we would need to manually set the `date_confirmed` fields of the affected users. All affected users created their accounts in September or October of 2014, so this likely reflects a bug that's since been fixed, although I'm not sure what it might have been.
1.0
[production] Users with inconsistent `date_confirmed` and `date_last_login` fields - # Expected All users who have logged in have a non-null value of `date_confirmed` # Observed In production, 18 users have non-null values of `date_last_login` but null values of `date_confirmed`. Based on a user report to @mckidwell, these users can't reset their passwords (but I verified manually that it's still possible to log in). To find users with inconsistent data, run: ```python User.find(Q('date_confirmed', 'eq', None) & Q('date_last_login', 'ne', None)) ``` To fix, we would need to manually set the `date_confirmed` fields of the affected users. All affected users created their accounts in September or October of 2014, so this likely reflects a bug that's since been fixed, although I'm not sure what it might have been.
non_main
users with inconsistent date confirmed and date last login fields expected all users who have logged in have a non null value of date confirmed observed in production users have non null values of date last login but null values of date confirmed based on a user report to mckidwell these users can t reset their passwords but i verified manually that it s still possible to log in to find users with inconsistent data run python user find q date confirmed eq none q date last login ne none to fix we would need to manually set the date confirmed fields of the affected users all affected users created their accounts in september or october of so this likely reflects a bug that s since been fixed although i m not sure what it might have been
0
552,869
16,329,791,813
IssuesEvent
2021-05-12 07:47:33
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
ENOSYS has ambiguous meaning.
Enhancement priority: low
In the errno.h file, we can find: ``` #define ENOSYS 71 /* Function not implemented */ ``` While checkpatch gives following information: ``` WARNING:ENOSYS: ENOSYS means 'invalid syscall nr' and nothing else ``` i.e. I wanted to use it in the following way: ``` __syscall int uart_configure(struct device *dev, struct uart_config *cfg); static inline int _impl_uart_configure(struct device *dev, struct uart_config *cfg) { const struct uart_driver_api *api = (const struct uart_driver_api *)dev->driver_api; if (api->configure) { return api->configure(dev, cfg); } return -ENOSYS; } ``` Where -ENOSYS If driver has not implemented this function -ENOTSUP If device does not support given configuration. Which information if true? errno.h file or checkpatch?
1.0
ENOSYS has ambiguous meaning. - In the errno.h file, we can find: ``` #define ENOSYS 71 /* Function not implemented */ ``` While checkpatch gives following information: ``` WARNING:ENOSYS: ENOSYS means 'invalid syscall nr' and nothing else ``` i.e. I wanted to use it in the following way: ``` __syscall int uart_configure(struct device *dev, struct uart_config *cfg); static inline int _impl_uart_configure(struct device *dev, struct uart_config *cfg) { const struct uart_driver_api *api = (const struct uart_driver_api *)dev->driver_api; if (api->configure) { return api->configure(dev, cfg); } return -ENOSYS; } ``` Where -ENOSYS If driver has not implemented this function -ENOTSUP If device does not support given configuration. Which information if true? errno.h file or checkpatch?
non_main
enosys has ambiguous meaning in the errno h file we can find define enosys function not implemented while checkpatch gives following information warning enosys enosys means invalid syscall nr and nothing else i e i wanted to use it in the following way syscall int uart configure struct device dev struct uart config cfg static inline int impl uart configure struct device dev struct uart config cfg const struct uart driver api api const struct uart driver api dev driver api if api configure return api configure dev cfg return enosys where enosys if driver has not implemented this function enotsup if device does not support given configuration which information if true errno h file or checkpatch
0
3,430
13,189,183,296
IssuesEvent
2020-08-13 07:59:55
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Make Wikibase schema classes properly immutable
bug maintainability wikidata
### To Reproduce The classes that constitute the wikibase schema (in `org.openrefine.wikidata.schema`) are meant to be immutable. But many of the getters in these classes return lists, which are not immutable in Java. This means that it is possible to modify these lists after getting them from the schema classes, which should be forbidden. This can lead to subtle bugs, see #3060. ### Current Results Getters of Wikibase schema classes return modifiable lists. ### Expected Behavior The returned lists should be unmodifiable. Wrapping with `Collections.immutableList()` should be enough (it is only an unmodifiable wrapper around a list, without actually copying the list).
True
Make Wikibase schema classes properly immutable - ### To Reproduce The classes that constitute the wikibase schema (in `org.openrefine.wikidata.schema`) are meant to be immutable. But many of the getters in these classes return lists, which are not immutable in Java. This means that it is possible to modify these lists after getting them from the schema classes, which should be forbidden. This can lead to subtle bugs, see #3060. ### Current Results Getters of Wikibase schema classes return modifiable lists. ### Expected Behavior The returned lists should be unmodifiable. Wrapping with `Collections.immutableList()` should be enough (it is only an unmodifiable wrapper around a list, without actually copying the list).
main
make wikibase schema classes properly immutable to reproduce the classes that constitute the wikibase schema in org openrefine wikidata schema are meant to be immutable but many of the getters in these classes return lists which are not immutable in java this means that it is possible to modify these lists after getting them from the schema classes which should be forbidden this can lead to subtle bugs see current results getters of wikibase schema classes return modifiable lists expected behavior the returned lists should be unmodifiable wrapping with collections immutablelist should be enough it is only an unmodifiable wrapper around a list without actually copying the list
1
1,673
6,574,094,040
IssuesEvent
2017-09-11 11:27:33
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_image: unable to deal with image IDs
affects_2.2 bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - `docker_image` ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /home/schwarz/code/infrastructure/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Debian GNU/Linux ##### SUMMARY `docker` allows addressing images by ID. Ansible should do the same. Otherwise it's impossible to delete an unnamed image. ##### STEPS TO REPRODUCE ``` sh $ docker pull alpine $ docker inspect --format={{.Id}} alpine sha256:baa5d63471ead618ff91ddfacf1e2c81bf0612bfeb1daf00eb0843a41fbfade3 $ ansible -m docker_image -a 'name=sha256:baa5d63471ead618ff91ddfacf1e2c81bf0612bfeb1daf00eb0843a41fbfade3 state=absent' localhost ``` ##### EXPECTED RESULTS The output should be the same as from `ansible -m docker_image -a 'name=alpine state=absent' localhost`. ``` localhost | SUCCESS => { "actions": [ "Removed image sha256:baa5d63471ead618ff91ddfacf1e2c81bf0612bfeb1daf00eb0843a41fbfade3" ], "changed": true, "image": { "state": "Deleted" } } ``` ##### ACTUAL RESULTS Instead no image is deleted. ``` localhost | SUCCESS => { "actions": [], "changed": false, "image": {} } ```
True
docker_image: unable to deal with image IDs - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - `docker_image` ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /home/schwarz/code/infrastructure/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Debian GNU/Linux ##### SUMMARY `docker` allows addressing images by ID. Ansible should do the same. Otherwise it's impossible to delete an unnamed image. ##### STEPS TO REPRODUCE ``` sh $ docker pull alpine $ docker inspect --format={{.Id}} alpine sha256:baa5d63471ead618ff91ddfacf1e2c81bf0612bfeb1daf00eb0843a41fbfade3 $ ansible -m docker_image -a 'name=sha256:baa5d63471ead618ff91ddfacf1e2c81bf0612bfeb1daf00eb0843a41fbfade3 state=absent' localhost ``` ##### EXPECTED RESULTS The output should be the same as from `ansible -m docker_image -a 'name=alpine state=absent' localhost`. ``` localhost | SUCCESS => { "actions": [ "Removed image sha256:baa5d63471ead618ff91ddfacf1e2c81bf0612bfeb1daf00eb0843a41fbfade3" ], "changed": true, "image": { "state": "Deleted" } } ``` ##### ACTUAL RESULTS Instead no image is deleted. ``` localhost | SUCCESS => { "actions": [], "changed": false, "image": {} } ```
main
docker image unable to deal with image ids issue type bug report component name docker image ansible version ansible config file home schwarz code infrastructure ansible cfg configured module search path default w o overrides configuration n a os environment debian gnu linux summary docker allows addressing images by id ansible should do the same otherwise it s impossible to delete an unnamed image steps to reproduce sh docker pull alpine docker inspect format id alpine ansible m docker image a name state absent localhost expected results the output should be the same as from ansible m docker image a name alpine state absent localhost localhost success actions removed image changed true image state deleted actual results instead no image is deleted localhost success actions changed false image
1
269,136
8,432,453,007
IssuesEvent
2018-10-17 02:08:22
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.youtube.com - site is not usable
browser-firefox priority-critical
<!-- @browser: Firefox 63.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:63.0) Gecko/20100101 Firefox/63.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.youtube.com/?pbjreload=10 **Browser / Version**: Firefox 63.0 **Operating System**: Windows 10 **Tested Another Browser**: Unknown **Problem type**: Site is not usable **Description**: the youtube site can be reached or just blackout **Steps to Reproduce**: just open and then like all black that i can see [![Screenshot Description](https://webcompat.com/uploads/2018/10/71b6b480-5c56-49e2-a379-fac1fac75510-thumb.jpg)](https://webcompat.com/uploads/2018/10/71b6b480-5c56-49e2-a379-fac1fac75510.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>buildID: 20181011200118</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.youtube.com - site is not usable - <!-- @browser: Firefox 63.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:63.0) Gecko/20100101 Firefox/63.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.youtube.com/?pbjreload=10 **Browser / Version**: Firefox 63.0 **Operating System**: Windows 10 **Tested Another Browser**: Unknown **Problem type**: Site is not usable **Description**: the youtube site can be reached or just blackout **Steps to Reproduce**: just open and then like all black that i can see [![Screenshot Description](https://webcompat.com/uploads/2018/10/71b6b480-5c56-49e2-a379-fac1fac75510-thumb.jpg)](https://webcompat.com/uploads/2018/10/71b6b480-5c56-49e2-a379-fac1fac75510.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>buildID: 20181011200118</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.all: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>channel: beta</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
site is not usable url browser version firefox operating system windows tested another browser unknown problem type site is not usable description the youtube site can be reached or just blackout steps to reproduce just open and then like all black that i can see browser configuration mixed active content blocked false buildid tracking content blocked false gfx webrender blob images true gfx webrender all false mixed passive content blocked false gfx webrender enabled false image mem shared true channel beta from with ❤️
0
310,478
26,718,318,549
IssuesEvent
2023-01-28 20:25:48
TestIntegrations/TestForwarding
https://api.github.com/repos/TestIntegrations/TestForwarding
opened
Bug 33
forwardeddddddTest ddw
# :clipboard: Bug Details >Bug 33 key | value --|-- Reported At | 2023-01-28 20:25:40 UTC Email | shussein@instabug.com Categories | Report a bug, sssss Tags | forwardeddddddTest, ddw App Version | 1.0 (4) Session Duration | 771 Device | Simulator, iOS 15.5 Display | 414x736 (@3x) ## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/birdy-demo-app/beta/bugs/9295?utm_source=github&utm_medium=integrations) :point_left: ___ # :chart_with_downwards_trend: Session Profiler Here is what the app was doing right before the bug was reported: Key | Value --|-- CPU Load | 15.5% Used Memory | 100.0% - 0.31/0.31 GB Used Storage | 95.9% - 223.79/233.47 GB Connectivity | WiFi Battery | 100% - unplugged Orientation | portrait Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/birdy-demo-app/beta/bugs/9295?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :bust_in_silhouette: User Info ### User Attributes ``` Age: 18 Logged in: True ``` ___ # :mag_right: Logs ### User Steps Here are the last 10 steps done by the user right before the bug was reported: ``` 20:25:32 Tap in Floating Button of type IBGInvocationFloatingView in ViewController 20:23:51 Tap in Floating Button of type IBGInvocationFloatingView in ViewController 20:13:41 Tap in Floating Button of type IBGInvocationFloatingView in ViewController 20:12:41 Top View: ViewController 20:12:41 Application: DidBecomeActive 20:12:41 Application: SceneDidActivate 20:12:41 Application: WillEnterForeground 20:12:41 Application: SceneWillConnect ``` Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/birdy-demo-app/beta/bugs/9295?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left: ___ # :camera: Images [![image attachment](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/19538279/3cd83712c2332ac1b1c598c6e22b6da8_one_bug_thumb/28705118/2023012810253462347715.jpg?Expires=4830611147&Signature=PfXr~-ANywh8uSat92JVSQTosdzss1Jw6bCIdtFwviBwasv5DdCld1cVkT4ajvRkhuszOuYbY0NgVSwHS1ndBHN1K-WPLXPQp1ZmOLzcnqsNqv1bEuQzQRzaC3b0HqPx9c9JRgvL-doUqOLqtQKUfD9oFzVbU7PpQBfi-bErHl3iPCmjgCKKWrIOlhUo8-wr0um21X3Q5o66aUPCd2-agJDGq-cszjfqnukz0UyQr2DTqNMkeylzrEfWJC3BWFGhWM0ipm5BIDq8g-80KcP9-HnvoHGkEz80IiJFfVuJ7dKwFCfmfu4j0nXVqD0HBfE19Stkg5yTb6unkpd04a3t2Q__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/19538279/3cd83712c2332ac1b1c598c6e22b6da8_original/28705118/2023012810253462347715.jpg?Expires=4830611147&Signature=oy1uj~CXQ90gb3Hr5fbRuHRK6Fc-4CnE9EdwUFQTkjeIsis4lL52rYERFN2NgjxZ0Uvbk5DnUTFhKjNhiFB20k3wL1Ex04jcQJXII645675k1M4zOvIp0euvSP68yhKL0pwndY5U6tDJZc7g0eR7JUEomVNeXRLK9~Ts1c-CnKbMAT~bIywaAsqIPPOkLRzshUif2nxnQWg-kWD3ye7-XeZ-8BNRk0FJKX2ob2MF4G42-NNRMjnFb-KegJM8Cp~Q2kuan0a2VaTGmBQQH9XywySl40yrG~m63QNbgoEbvQsUj8fTb52GO4Em9hF4MXs3JL9GWlWqW9m14J0iAWHcmA__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ) ___ # :warning: Looking for More Details? 1. **Network Log**: we are unable to capture your network requests automatically. If you are using AFNetworking or Alamofire, [**check the details mentioned here**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations#section-requests-not-appearing-in-logs). 2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations). 3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations). 4. **Console Log**: when enabled you will see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
1.0
Bug 33 - # :clipboard: Bug Details >Bug 33 key | value --|-- Reported At | 2023-01-28 20:25:40 UTC Email | shussein@instabug.com Categories | Report a bug, sssss Tags | forwardeddddddTest, ddw App Version | 1.0 (4) Session Duration | 771 Device | Simulator, iOS 15.5 Display | 414x736 (@3x) ## :point_right: [View Full Bug Report on Instabug](https://dashboard.instabug.com/applications/birdy-demo-app/beta/bugs/9295?utm_source=github&utm_medium=integrations) :point_left: ___ # :chart_with_downwards_trend: Session Profiler Here is what the app was doing right before the bug was reported: Key | Value --|-- CPU Load | 15.5% Used Memory | 100.0% - 0.31/0.31 GB Used Storage | 95.9% - 223.79/233.47 GB Connectivity | WiFi Battery | 100% - unplugged Orientation | portrait Find all the changes that happened in the parameters mentioned above during the last 60 seconds before the bug was reported here: :point_right: **[View Full Session Profiler](https://dashboard.instabug.com/applications/birdy-demo-app/beta/bugs/9295?show-session-profiler=true&utm_source=github&utm_medium=integrations)** :point_left: ___ # :bust_in_silhouette: User Info ### User Attributes ``` Age: 18 Logged in: True ``` ___ # :mag_right: Logs ### User Steps Here are the last 10 steps done by the user right before the bug was reported: ``` 20:25:32 Tap in Floating Button of type IBGInvocationFloatingView in ViewController 20:23:51 Tap in Floating Button of type IBGInvocationFloatingView in ViewController 20:13:41 Tap in Floating Button of type IBGInvocationFloatingView in ViewController 20:12:41 Top View: ViewController 20:12:41 Application: DidBecomeActive 20:12:41 Application: SceneDidActivate 20:12:41 Application: WillEnterForeground 20:12:41 Application: SceneWillConnect ``` Find all the user steps done by the user throughout the session here: :point_right: **[View All User Steps](https://dashboard.instabug.com/applications/birdy-demo-app/beta/bugs/9295?show-logs=user_steps&utm_source=github&utm_medium=integrations)** :point_left: ___ # :camera: Images [![image attachment](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/19538279/3cd83712c2332ac1b1c598c6e22b6da8_one_bug_thumb/28705118/2023012810253462347715.jpg?Expires=4830611147&Signature=PfXr~-ANywh8uSat92JVSQTosdzss1Jw6bCIdtFwviBwasv5DdCld1cVkT4ajvRkhuszOuYbY0NgVSwHS1ndBHN1K-WPLXPQp1ZmOLzcnqsNqv1bEuQzQRzaC3b0HqPx9c9JRgvL-doUqOLqtQKUfD9oFzVbU7PpQBfi-bErHl3iPCmjgCKKWrIOlhUo8-wr0um21X3Q5o66aUPCd2-agJDGq-cszjfqnukz0UyQr2DTqNMkeylzrEfWJC3BWFGhWM0ipm5BIDq8g-80KcP9-HnvoHGkEz80IiJFfVuJ7dKwFCfmfu4j0nXVqD0HBfE19Stkg5yTb6unkpd04a3t2Q__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ)](https://d38gnqwzxziyyy.cloudfront.net/attachments/bugs/19538279/3cd83712c2332ac1b1c598c6e22b6da8_original/28705118/2023012810253462347715.jpg?Expires=4830611147&Signature=oy1uj~CXQ90gb3Hr5fbRuHRK6Fc-4CnE9EdwUFQTkjeIsis4lL52rYERFN2NgjxZ0Uvbk5DnUTFhKjNhiFB20k3wL1Ex04jcQJXII645675k1M4zOvIp0euvSP68yhKL0pwndY5U6tDJZc7g0eR7JUEomVNeXRLK9~Ts1c-CnKbMAT~bIywaAsqIPPOkLRzshUif2nxnQWg-kWD3ye7-XeZ-8BNRk0FJKX2ob2MF4G42-NNRMjnFb-KegJM8Cp~Q2kuan0a2VaTGmBQQH9XywySl40yrG~m63QNbgoEbvQsUj8fTb52GO4Em9hF4MXs3JL9GWlWqW9m14J0iAWHcmA__&Key-Pair-Id=APKAIXAG65U6UUX7JAQQ) ___ # :warning: Looking for More Details? 1. **Network Log**: we are unable to capture your network requests automatically. If you are using AFNetworking or Alamofire, [**check the details mentioned here**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations#section-requests-not-appearing-in-logs). 2. **User Events**: start capturing custom User Events to send them along with each report. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations). 3. **Instabug Log**: start adding Instabug logs to see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations). 4. **Console Log**: when enabled you will see them right inside each report you receive. [**Find all the details in the docs**](https://docs.instabug.com/docs/ios-logging?utm_source=github&utm_medium=integrations).
non_main
bug clipboard bug details bug key value reported at utc email shussein instabug com categories report a bug sssss tags forwardeddddddtest ddw app version session duration device simulator ios display point right point left chart with downwards trend session profiler here is what the app was doing right before the bug was reported key value cpu load used memory gb used storage gb connectivity wifi battery unplugged orientation portrait find all the changes that happened in the parameters mentioned above during the last seconds before the bug was reported here point right point left bust in silhouette user info user attributes age logged in true mag right logs user steps here are the last steps done by the user right before the bug was reported tap in floating button of type ibginvocationfloatingview in viewcontroller tap in floating button of type ibginvocationfloatingview in viewcontroller tap in floating button of type ibginvocationfloatingview in viewcontroller top view viewcontroller application didbecomeactive application scenedidactivate application willenterforeground application scenewillconnect find all the user steps done by the user throughout the session here point right point left camera images warning looking for more details network log we are unable to capture your network requests automatically if you are using afnetworking or alamofire user events start capturing custom user events to send them along with each report instabug log start adding instabug logs to see them right inside each report you receive console log when enabled you will see them right inside each report you receive
0
71,027
8,614,570,222
IssuesEvent
2018-11-19 17:51:59
teeworlds/teeworlds
https://api.github.com/repos/teeworlds/teeworlds
closed
SUR/LMS gametypes have confusing names
design discussion
> I still personally consider the naming scheme of the elimination modes to be a mistake - just call it "elimination" or EL. Then team elimination is TEL. It's a lot more consistent than the current lms/sur naming scheme - I for one have no recollection of which one of those is a team mode and which one isn't. I would really like to see this changed. [_Deku on the forums_](https://www.teeworlds.com/forum/viewtopic.php?pid=121817#p121817) ---- It might be a bit late but I believe being consistent with the DM/TDM naming scheme would yield something much more intuitive. Could this be looked into?
1.0
SUR/LMS gametypes have confusing names - > I still personally consider the naming scheme of the elimination modes to be a mistake - just call it "elimination" or EL. Then team elimination is TEL. It's a lot more consistent than the current lms/sur naming scheme - I for one have no recollection of which one of those is a team mode and which one isn't. I would really like to see this changed. [_Deku on the forums_](https://www.teeworlds.com/forum/viewtopic.php?pid=121817#p121817) ---- It might be a bit late but I believe being consistent with the DM/TDM naming scheme would yield something much more intuitive. Could this be looked into?
non_main
sur lms gametypes have confusing names i still personally consider the naming scheme of the elimination modes to be a mistake just call it elimination or el then team elimination is tel it s a lot more consistent than the current lms sur naming scheme i for one have no recollection of which one of those is a team mode and which one isn t i would really like to see this changed it might be a bit late but i believe being consistent with the dm tdm naming scheme would yield something much more intuitive could this be looked into
0
38,599
12,561,296,493
IssuesEvent
2020-06-08 01:02:58
yaeljacobs67/cncjs
https://api.github.com/repos/yaeljacobs67/cncjs
opened
CVE-2020-7661 (Medium) detected in url-regex-3.2.0.tgz
security vulnerability
## CVE-2020-7661 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-regex-3.2.0.tgz</b></p></summary> <p>Regular expression for matching URLs</p> <p>Library home page: <a href="https://registry.npmjs.org/url-regex/-/url-regex-3.2.0.tgz">https://registry.npmjs.org/url-regex/-/url-regex-3.2.0.tgz</a></p> <p>Path to dependency file: /cncjs/package.json</p> <p>Path to vulnerable library: /tmp/git/cncjs/node_modules/url-regex/package.json</p> <p> Dependency Hierarchy: - jimp-0.2.28.tgz (Root Library) - :x: **url-regex-3.2.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> all versions of url-regex are vulnerable to Regular Expression Denial of Service. An attacker providing a very long string in String.test can cause a Denial of Service. <p>Publish Date: 2020-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7661>CVE-2020-7661</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"url-regex","packageVersion":"3.2.0","isTransitiveDependency":true,"dependencyTree":"jimp:0.2.28;url-regex:3.2.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-7661","vulnerabilityDetails":"all versions of url-regex are vulnerable to Regular Expression Denial of Service. An attacker providing a very long string in String.test can cause a Denial of Service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7661","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7661 (Medium) detected in url-regex-3.2.0.tgz - ## CVE-2020-7661 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-regex-3.2.0.tgz</b></p></summary> <p>Regular expression for matching URLs</p> <p>Library home page: <a href="https://registry.npmjs.org/url-regex/-/url-regex-3.2.0.tgz">https://registry.npmjs.org/url-regex/-/url-regex-3.2.0.tgz</a></p> <p>Path to dependency file: /cncjs/package.json</p> <p>Path to vulnerable library: /tmp/git/cncjs/node_modules/url-regex/package.json</p> <p> Dependency Hierarchy: - jimp-0.2.28.tgz (Root Library) - :x: **url-regex-3.2.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> all versions of url-regex are vulnerable to Regular Expression Denial of Service. An attacker providing a very long string in String.test can cause a Denial of Service. <p>Publish Date: 2020-06-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7661>CVE-2020-7661</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"url-regex","packageVersion":"3.2.0","isTransitiveDependency":true,"dependencyTree":"jimp:0.2.28;url-regex:3.2.0","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-7661","vulnerabilityDetails":"all versions of url-regex are vulnerable to Regular Expression Denial of Service. An attacker providing a very long string in String.test can cause a Denial of Service.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7661","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in url regex tgz cve medium severity vulnerability vulnerable library url regex tgz regular expression for matching urls library home page a href path to dependency file cncjs package json path to vulnerable library tmp git cncjs node modules url regex package json dependency hierarchy jimp tgz root library x url regex tgz vulnerable library vulnerability details all versions of url regex are vulnerable to regular expression denial of service an attacker providing a very long string in string test can cause a denial of service publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails all versions of url regex are vulnerable to regular expression denial of service an attacker providing a very long string in string test can cause a denial of service vulnerabilityurl
0
37,135
15,180,962,418
IssuesEvent
2021-02-15 01:54:24
Geonovum/disgeo-arch
https://api.github.com/repos/Geonovum/disgeo-arch
closed
5.2.3.1 Afgeleide opslag
Component Opslag In Behandeling In behandeling - voorstel servicelayering Service layering
In de tekst staat: > Ten behoeve van afname van gegevens en informatie is naar verwachting in de technische uitwerking afgeleide opslag nodig. Dit is geen zelfstandige component, maar een onderdeel van Afname van gegevens en informatie. Geen zelfstandige component? Vraag is of dit een juiste keuze is, kan immers gaan om enorme hoeveelheden te verrijken en verrijkte gegevens. En verder: > Uitgangspunt voor deze vereisten is dat het koppelvlak tussen de componenten Opslag en Afgeleide Opslag een intern koppelvlak is waarvoor geen vereisten gelden m.b.t. het gebruik van open, leveranciersonafhankelijke standaarden en technologieën. Nee, omwille van ontkoppeling en portabiliteit ook hiervoor altijd de beschikbare data (technische) services gebruiken (eat your own dogfood).
2.0
5.2.3.1 Afgeleide opslag - In de tekst staat: > Ten behoeve van afname van gegevens en informatie is naar verwachting in de technische uitwerking afgeleide opslag nodig. Dit is geen zelfstandige component, maar een onderdeel van Afname van gegevens en informatie. Geen zelfstandige component? Vraag is of dit een juiste keuze is, kan immers gaan om enorme hoeveelheden te verrijken en verrijkte gegevens. En verder: > Uitgangspunt voor deze vereisten is dat het koppelvlak tussen de componenten Opslag en Afgeleide Opslag een intern koppelvlak is waarvoor geen vereisten gelden m.b.t. het gebruik van open, leveranciersonafhankelijke standaarden en technologieën. Nee, omwille van ontkoppeling en portabiliteit ook hiervoor altijd de beschikbare data (technische) services gebruiken (eat your own dogfood).
non_main
afgeleide opslag in de tekst staat ten behoeve van afname van gegevens en informatie is naar verwachting in de technische uitwerking afgeleide opslag nodig dit is geen zelfstandige component maar een onderdeel van afname van gegevens en informatie geen zelfstandige component vraag is of dit een juiste keuze is kan immers gaan om enorme hoeveelheden te verrijken en verrijkte gegevens en verder uitgangspunt voor deze vereisten is dat het koppelvlak tussen de componenten opslag en afgeleide opslag een intern koppelvlak is waarvoor geen vereisten gelden m b t het gebruik van open leveranciersonafhankelijke standaarden en technologieën nee omwille van ontkoppeling en portabiliteit ook hiervoor altijd de beschikbare data technische services gebruiken eat your own dogfood
0
1,298
5,541,677,688
IssuesEvent
2017-03-22 13:28:49
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
2.1.1: ec2_vpc does not update/change routing table tags
affects_2.1 aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION [defaults] inventory = ~/inventory/production host_key_checking = False log_path = ~/ansible.log [ssh_connection] pipelining = True ##### OS / ENVIRONMENT N/A ##### SUMMARY Before with Ansible 1.9.6 and the patch mentioned here: http://grokbase.com/t/gg/ansible-project/149pfx39dg/tagging-ec2-vpc-route-tables-and-gateways#20150304gdcdptnuteldhgubbx3m2wfuie , after changing the routing table tag in my playbook or on the AWS console, the routing table tags got changed to the value of my playbook on rerun. In Ansible 2.1.1 the initial tags are set, but not updated on a rerun after changing it in my playbook or on the AWS console. ##### STEPS TO REPRODUCE run playbook with for instance: ``` route_tables: - subnets: - "10.0.1.0/24" - "10.0.2.0/24" routes: - dest: 0.0.0.0/0 gw: igw resource_tags: { "Name" : dmz" } ``` Now change the tags value 'dmz' to 'public' in the playbook or on the AWS console and re-run playbook. See that the tag value does not get changed ##### EXPECTED RESULTS I expect the tag value to correspond with the playbook after running it. ##### ACTUAL RESULTS Tag value did not get updated.
True
2.1.1: ec2_vpc does not update/change routing table tags - ##### ISSUE TYPE - Bug Report ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION [defaults] inventory = ~/inventory/production host_key_checking = False log_path = ~/ansible.log [ssh_connection] pipelining = True ##### OS / ENVIRONMENT N/A ##### SUMMARY Before with Ansible 1.9.6 and the patch mentioned here: http://grokbase.com/t/gg/ansible-project/149pfx39dg/tagging-ec2-vpc-route-tables-and-gateways#20150304gdcdptnuteldhgubbx3m2wfuie , after changing the routing table tag in my playbook or on the AWS console, the routing table tags got changed to the value of my playbook on rerun. In Ansible 2.1.1 the initial tags are set, but not updated on a rerun after changing it in my playbook or on the AWS console. ##### STEPS TO REPRODUCE run playbook with for instance: ``` route_tables: - subnets: - "10.0.1.0/24" - "10.0.2.0/24" routes: - dest: 0.0.0.0/0 gw: igw resource_tags: { "Name" : dmz" } ``` Now change the tags value 'dmz' to 'public' in the playbook or on the AWS console and re-run playbook. See that the tag value does not get changed ##### EXPECTED RESULTS I expect the tag value to correspond with the playbook after running it. ##### ACTUAL RESULTS Tag value did not get updated.
main
vpc does not update change routing table tags issue type bug report ansible version ansible configuration inventory inventory production host key checking false log path ansible log pipelining true os environment n a summary before with ansible and the patch mentioned here after changing the routing table tag in my playbook or on the aws console the routing table tags got changed to the value of my playbook on rerun in ansible the initial tags are set but not updated on a rerun after changing it in my playbook or on the aws console steps to reproduce run playbook with for instance route tables subnets routes dest gw igw resource tags name dmz now change the tags value dmz to public in the playbook or on the aws console and re run playbook see that the tag value does not get changed expected results i expect the tag value to correspond with the playbook after running it actual results tag value did not get updated
1
4,063
19,010,483,033
IssuesEvent
2021-11-23 08:44:36
involve-gmbh/upptime
https://api.github.com/repos/involve-gmbh/upptime
closed
Wartungsfenster
maintainance
<!-- start: 2021-11-11T17:30:00.220Z+01:00 end: 2021-11-11T23:55:00.220Z+01:00 expectedDown: homepage, share-pic, erema-group, primeconcept -->
True
Wartungsfenster - <!-- start: 2021-11-11T17:30:00.220Z+01:00 end: 2021-11-11T23:55:00.220Z+01:00 expectedDown: homepage, share-pic, erema-group, primeconcept -->
main
wartungsfenster start end expecteddown homepage share pic erema group primeconcept
1
285,706
31,155,416,272
IssuesEvent
2023-08-16 12:51:11
amaybaum-local/WEI-3070d
https://api.github.com/repos/amaybaum-local/WEI-3070d
opened
bootstrap-3.3.7.min.js: 5 vulnerabilities (highest severity is: 6.5)
Mend: dependency security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (bootstrap version) | Remediation Possible** | Reachability | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | --- | | [WS-2018-0021](https://github.com/twbs/bootstrap/issues/20184) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | bootstrap-3.3.7.min.js | Direct | 3.4.0 | &#10060;| | | [CVE-2019-8331](https://www.mend.io/vulnerability-database/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | &#10060;| | | [CVE-2018-14040](https://www.mend.io/vulnerability-database/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | &#10060;| | | [CVE-2018-20677](https://www.mend.io/vulnerability-database/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | &#10060;| | | [CVE-2018-14042](https://www.mend.io/vulnerability-database/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | &#10060;| | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2018-0021</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> XSS in data-target in bootstrap (3.3.7 and before) <p>Publish Date: 2017-09-29 <p>URL: <a href=https://github.com/twbs/bootstrap/issues/20184>WS-2018-0021</a></p> </p> <p></p> ### CVSS 2 Score Details (<b>6.5</b>) <p> Base Score Metrics not available</p> </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-06-12</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2018-14040</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14040>CVE-2018-14040</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2018-20677</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-20677>CVE-2018-20677</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2018-14042</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14042>CVE-2018-14042</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details>
True
bootstrap-3.3.7.min.js: 5 vulnerabilities (highest severity is: 6.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p></summary> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (bootstrap version) | Remediation Possible** | Reachability | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | --- | | [WS-2018-0021](https://github.com/twbs/bootstrap/issues/20184) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.5 | bootstrap-3.3.7.min.js | Direct | 3.4.0 | &#10060;| | | [CVE-2019-8331](https://www.mend.io/vulnerability-database/CVE-2019-8331) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1 | &#10060;| | | [CVE-2018-14040](https://www.mend.io/vulnerability-database/CVE-2018-14040) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0 | &#10060;| | | [CVE-2018-20677](https://www.mend.io/vulnerability-database/CVE-2018-20677) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0 | &#10060;| | | [CVE-2018-14042](https://www.mend.io/vulnerability-database/CVE-2018-14042) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Medium | 6.1 | bootstrap-3.3.7.min.js | Direct | org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0 | &#10060;| | <p>**In some cases, Remediation PR cannot be created automatically for a vulnerability despite the availability of remediation</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> WS-2018-0021</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> XSS in data-target in bootstrap (3.3.7 and before) <p>Publish Date: 2017-09-29 <p>URL: <a href=https://github.com/twbs/bootstrap/issues/20184>WS-2018-0021</a></p> </p> <p></p> ### CVSS 2 Score Details (<b>6.5</b>) <p> Base Score Metrics not available</p> </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-06-12</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2019-8331</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.1 and 4.3.x before 4.3.1, XSS is possible in the tooltip or popover data-template attribute. <p>Publish Date: 2019-02-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-8331>CVE-2019-8331</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2019-02-20</p> <p>Fix Resolution: bootstrap - 3.4.1,4.3.1;bootstrap-sass - 3.4.1,4.3.1</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2018-14040</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the collapse data-parent attribute. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14040>CVE-2018-14040</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2,org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2018-20677</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 3.4.0, XSS is possible in the affix configuration target property. <p>Publish Date: 2019-01-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-20677>CVE-2018-20677</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-20677</a></p> <p>Release Date: 2019-01-09</p> <p>Fix Resolution: Bootstrap - v3.4.0;NorDroN.AngularTemplate - 0.1.6;Dynamic.NET.Express.ProjectTemplates - 0.8.0;dotnetng.template - 1.0.0.4;ZNxtApp.Core.Module.Theme - 1.0.9-Beta;JMeter - 5.0.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> CVE-2018-14042</summary> ### Vulnerable Library - <b>bootstrap-3.3.7.min.js</b></p> <p>The most popular front-end framework for developing responsive, mobile first projects on the web.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js">https://cdnjs.cloudflare.com/ajax/libs/twitter-bootstrap/3.3.7/js/bootstrap.min.js</a></p> <p>Path to dependency file: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p>Path to vulnerable library: /src/main/webapp/WEB-INF/views/userList.jsp</p> <p> Dependency Hierarchy: - :x: **bootstrap-3.3.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/amaybaum-local/WEI-3070d/commit/f112c003b572d6e068d8c27d7eca7e428d44d1a9">f112c003b572d6e068d8c27d7eca7e428d44d1a9</a></p> <p>Found in base branch: <b>vp-rem</b></p> </p> <p></p> ### Vulnerability Details <p> In Bootstrap before 4.1.2, XSS is possible in the data-container property of tooltip. <p>Publish Date: 2018-07-13 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-14042>CVE-2018-14042</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2018-07-13</p> <p>Fix Resolution: org.webjars.npm:bootstrap:4.1.2.org.webjars:bootstrap:3.4.0</p> </p> <p></p> </details>
non_main
bootstrap min js vulnerabilities highest severity is vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main webapp web inf views userlist jsp path to vulnerable library src main webapp web inf views userlist jsp found in head commit a href vulnerabilities cve severity cvss dependency type fixed in bootstrap version remediation possible reachability medium bootstrap min js direct medium bootstrap min js direct bootstrap bootstrap sass medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap medium bootstrap min js direct bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter medium bootstrap min js direct org webjars npm bootstrap org webjars bootstrap in some cases remediation pr cannot be created automatically for a vulnerability despite the availability of remediation details ws vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main webapp web inf views userlist jsp path to vulnerable library src main webapp web inf views userlist jsp dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch vp rem vulnerability details xss in data target in bootstrap and before publish date url a href cvss score details base score metrics not available suggested fix type upgrade version release date fix resolution cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main webapp web inf views userlist jsp path to vulnerable library src main webapp web inf views userlist jsp dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch vp rem vulnerability details in bootstrap before and x before xss is possible in the tooltip or popover data template attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution bootstrap bootstrap sass cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main webapp web inf views userlist jsp path to vulnerable library src main webapp web inf views userlist jsp dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch vp rem vulnerability details in bootstrap before xss is possible in the collapse data parent attribute publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main webapp web inf views userlist jsp path to vulnerable library src main webapp web inf views userlist jsp dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch vp rem vulnerability details in bootstrap before xss is possible in the affix configuration target property publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bootstrap nordron angulartemplate dynamic net express projecttemplates dotnetng template znxtapp core module theme beta jmeter cve vulnerable library bootstrap min js the most popular front end framework for developing responsive mobile first projects on the web library home page a href path to dependency file src main webapp web inf views userlist jsp path to vulnerable library src main webapp web inf views userlist jsp dependency hierarchy x bootstrap min js vulnerable library found in head commit a href found in base branch vp rem vulnerability details in bootstrap before xss is possible in the data container property of tooltip publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution org webjars npm bootstrap org webjars bootstrap
0
269,295
23,434,975,478
IssuesEvent
2022-08-15 08:48:34
SaphieNyako/Quest_Giver
https://api.github.com/repos/SaphieNyako/Quest_Giver
closed
Quest villagers don't trade with me after I finished his questline
bug testing required
Title, I tried this with the signed trading scrolls, it worked on regular villagers only. Expected Behavior: After finishing the quest villager's questline, they will lose their exclamation/question mark, and perform regular trading function.
1.0
Quest villagers don't trade with me after I finished his questline - Title, I tried this with the signed trading scrolls, it worked on regular villagers only. Expected Behavior: After finishing the quest villager's questline, they will lose their exclamation/question mark, and perform regular trading function.
non_main
quest villagers don t trade with me after i finished his questline title i tried this with the signed trading scrolls it worked on regular villagers only expected behavior after finishing the quest villager s questline they will lose their exclamation question mark and perform regular trading function
0
76,750
3,491,884,368
IssuesEvent
2016-01-04 17:40:04
Metaswitch/clearwater-etcd
https://api.github.com/repos/Metaswitch/clearwater-etcd
closed
Auto-generated code should be clearer about alarm numbers
cat:diagnostics cat:easy low-priority
We use two different alarm severities - in the code we support alarmModelState but we raise alarms using ituAlarmPerceivedSeverity (see the mapping in the ituAlarmEntry in https://tools.ietf.org/html/rfc3877#section-5.4 and conversion at https://github.com/Metaswitch/clearwater-snmp-handlers/blob/master/alarm_table_defs.cpp#L57). We should put a comment in the auto-generated code explaining this (in https://github.com/Metaswitch/python-common/blob/dev/metaswitch/common/alarms_parser.py and https://github.com/Metaswitch/cpp-common/blob/master/src/alarm_header.cpp)
1.0
Auto-generated code should be clearer about alarm numbers - We use two different alarm severities - in the code we support alarmModelState but we raise alarms using ituAlarmPerceivedSeverity (see the mapping in the ituAlarmEntry in https://tools.ietf.org/html/rfc3877#section-5.4 and conversion at https://github.com/Metaswitch/clearwater-snmp-handlers/blob/master/alarm_table_defs.cpp#L57). We should put a comment in the auto-generated code explaining this (in https://github.com/Metaswitch/python-common/blob/dev/metaswitch/common/alarms_parser.py and https://github.com/Metaswitch/cpp-common/blob/master/src/alarm_header.cpp)
non_main
auto generated code should be clearer about alarm numbers we use two different alarm severities in the code we support alarmmodelstate but we raise alarms using itualarmperceivedseverity see the mapping in the itualarmentry in and conversion at we should put a comment in the auto generated code explaining this in and
0
927
4,630,629,166
IssuesEvent
2016-09-28 13:25:10
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
s3_bucket versioning parameter is required
affects_2.1 aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME s3_bucket ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- --> NONE ##### OS / ENVIRONMENT <!--- --> Mac OS X 10.11.6 ##### SUMMARY <!--- Explain the problem briefly --> When running a play to perform a change other than changing the versioning of the bucket, Ansible will assume that versioning needs to be turned off (default is no). In all documentation, this parameter is marked as not required. But if I perform an iteration over many s3 buckets to add something as simple as a tag, some of which need to have versioning enabled (for Cross Region Replication), it will suspend versioning on the buckets that have versioning enabled. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` --- - hosts: localhost tasks: - name: Tag S3 buckets s3_bucket: {name: "{{ item }}", tags:{Application: "S3"}} with_items: - policy-test-rwetzelberger ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expect to see that the s3 bucket named "policy-test-rwetzelberger" has a tag verified to have or added called "S3". This bucket had versioning enabled. Versioning that is enabled should not be suspended. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` ansible-playbook -vvvv s3-tag.yml No config file found; using defaults Loaded callback default of type stdout, v2.0 PLAYBOOK: s3-tag.yml *********************************************************** 1 plays in s3-tag.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: <username> <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932 `" && echo ansible-tmp-1471356881.14-189485368299932="` echo $HOME/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpiE06ET TO /Users/<username>/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932/setup <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/<username>/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932/setup; rm -rf "/Users/<username>/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Tag S3 buckets] ********************************************************** task path: /Users/<username>/fork-ansible-playbooks/Gio/tasks/s3-tag.yml:5 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: <username> <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130 `" && echo ansible-tmp-1471356885.59-45084579463130="` echo $HOME/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpVvlVdL TO /Users/<username>/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130/s3_bucket <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/<username>/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130/s3_bucket; rm -rf "/Users/<username>/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => (item=policy-test-rwetzelberger) => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "force": false, "name": "policy-test-rwetzelberger", "policy": null, "profile": null, "region": null, "requester_pays": false, "s3_url": null, "security_token": null, "state": "present", "tags": {"Application": "S3"}, "validate_certs": true, "versioning": false}, "module_name": "s3_bucket"}, "item": "policy-test-rwetzelberger", "name": "policy-test-rwetzelberger", "policy": null, "requester_pays": false, "tags": {"Application": "S3"}, "versioning": {"MfaDelete": "Disabled", "Versioning": "Suspended"}} PLAY RECAP ********************************************************************* localhost : ok=2 changed=1 unreachable=0 failed=0 ```
True
s3_bucket versioning parameter is required - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME s3_bucket ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- --> NONE ##### OS / ENVIRONMENT <!--- --> Mac OS X 10.11.6 ##### SUMMARY <!--- Explain the problem briefly --> When running a play to perform a change other than changing the versioning of the bucket, Ansible will assume that versioning needs to be turned off (default is no). In all documentation, this parameter is marked as not required. But if I perform an iteration over many s3 buckets to add something as simple as a tag, some of which need to have versioning enabled (for Cross Region Replication), it will suspend versioning on the buckets that have versioning enabled. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` --- - hosts: localhost tasks: - name: Tag S3 buckets s3_bucket: {name: "{{ item }}", tags:{Application: "S3"}} with_items: - policy-test-rwetzelberger ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expect to see that the s3 bucket named "policy-test-rwetzelberger" has a tag verified to have or added called "S3". This bucket had versioning enabled. Versioning that is enabled should not be suspended. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` ansible-playbook -vvvv s3-tag.yml No config file found; using defaults Loaded callback default of type stdout, v2.0 PLAYBOOK: s3-tag.yml *********************************************************** 1 plays in s3-tag.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: <username> <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932 `" && echo ansible-tmp-1471356881.14-189485368299932="` echo $HOME/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpiE06ET TO /Users/<username>/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932/setup <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/<username>/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932/setup; rm -rf "/Users/<username>/.ansible/tmp/ansible-tmp-1471356881.14-189485368299932/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Tag S3 buckets] ********************************************************** task path: /Users/<username>/fork-ansible-playbooks/Gio/tasks/s3-tag.yml:5 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: <username> <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130 `" && echo ansible-tmp-1471356885.59-45084579463130="` echo $HOME/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130 `" ) && sleep 0' <127.0.0.1> PUT /var/folders/m2/5vh7dqn90gz834mv6swzzdzmh8qf0x/T/tmpVvlVdL TO /Users/<username>/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130/s3_bucket <127.0.0.1> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/<username>/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130/s3_bucket; rm -rf "/Users/<username>/.ansible/tmp/ansible-tmp-1471356885.59-45084579463130/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => (item=policy-test-rwetzelberger) => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "force": false, "name": "policy-test-rwetzelberger", "policy": null, "profile": null, "region": null, "requester_pays": false, "s3_url": null, "security_token": null, "state": "present", "tags": {"Application": "S3"}, "validate_certs": true, "versioning": false}, "module_name": "s3_bucket"}, "item": "policy-test-rwetzelberger", "name": "policy-test-rwetzelberger", "policy": null, "requester_pays": false, "tags": {"Application": "S3"}, "versioning": {"MfaDelete": "Disabled", "Versioning": "Suspended"}} PLAY RECAP ********************************************************************* localhost : ok=2 changed=1 unreachable=0 failed=0 ```
main
bucket versioning parameter is required issue type bug report component name bucket ansible version ansible config file configured module search path default w o overrides configuration none os environment mac os x summary when running a play to perform a change other than changing the versioning of the bucket ansible will assume that versioning needs to be turned off default is no in all documentation this parameter is marked as not required but if i perform an iteration over many buckets to add something as simple as a tag some of which need to have versioning enabled for cross region replication it will suspend versioning on the buckets that have versioning enabled steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used hosts localhost tasks name tag buckets bucket name item tags application with items policy test rwetzelberger expected results i expect to see that the bucket named policy test rwetzelberger has a tag verified to have or added called this bucket had versioning enabled versioning that is enabled should not be suspended actual results ansible playbook vvvv tag yml no config file found using defaults loaded callback default of type stdout playbook tag yml plays in tag yml play task establish local connection for user exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders t to users ansible tmp ansible tmp setup exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users ansible tmp ansible tmp setup rm rf users ansible tmp ansible tmp dev null sleep ok task task path users fork ansible playbooks gio tasks tag yml establish local connection for user exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put var folders t tmpvvlvdl to users ansible tmp ansible tmp bucket exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users ansible tmp ansible tmp bucket rm rf users ansible tmp ansible tmp dev null sleep changed item policy test rwetzelberger changed true invocation module args aws access key null aws secret key null url null force false name policy test rwetzelberger policy null profile null region null requester pays false url null security token null state present tags application validate certs true versioning false module name bucket item policy test rwetzelberger name policy test rwetzelberger policy null requester pays false tags application versioning mfadelete disabled versioning suspended play recap localhost ok changed unreachable failed
1
36,436
8,109,950,882
IssuesEvent
2018-08-14 09:19:21
publiclab/plots2
https://api.github.com/repos/publiclab/plots2
closed
Show most recently updated people on Search API
review-me rgsoc summer-of-code
Part of #2755 Change the sorting of the results in the API to most recently updated people as discussed in #2925
1.0
Show most recently updated people on Search API - Part of #2755 Change the sorting of the results in the API to most recently updated people as discussed in #2925
non_main
show most recently updated people on search api part of change the sorting of the results in the api to most recently updated people as discussed in
0
1,599
6,572,380,949
IssuesEvent
2017-09-11 01:52:27
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ansible 2.0.0.2 - aws ec2_group - rules_egress not working
affects_2.0 aws bug_report cloud waiting_on_maintainer
## Issue Type Bug Report ## Component Name ec2_group ## Ansible Version 2.0.0.2 ## Environment Ansible 2.0.0.2 Ubuntu 14.03 AWS ## Summary I am trying to apply outbound firewall rules to AWS using the rules_egress: However, when I log into my AWS account and look at the winrdp security group, all I see are the inbound rules and no outbound rules. ### dev-environment.yml ``` # security groups to be created security_groups: - name: winrdp desc: the security group for the winrdp server rules: - proto: tcp from_port: 5986 to_port: 5986 cidr_ip: 0.0.0.0/0 - proto: tcp from_port: 3389 to_port: 3389 cidr_ip: 0.0.0.0/0 rules_egress: - proto: all cidr_ip: 0.0.0.0/0 ``` ### playbook ``` --- - name: Provision ec2 instances based on the environment hosts: localhost connection: local gather_facts: False vars_files: - vars/dev-environment.yml - vars/aws-creds.yml tasks: - name: Create required security groups ec2_group: name: "{{ item.name }}" description: "{{ item.desc }}" rules: "{{ item.rules }}" region: "{{ ec2_region }}" ec2_access_key: "{{ ec2_access_key }}" ec2_secret_key: "{{ ec2_secret_key }}" with_items: security_groups ``` ![screenshot from 2016-02-16 20 31 24](https://cloud.githubusercontent.com/assets/6406166/13097103/63c440a8-d4ec-11e5-8769-2ecc6f1c651e.png)
True
ansible 2.0.0.2 - aws ec2_group - rules_egress not working - ## Issue Type Bug Report ## Component Name ec2_group ## Ansible Version 2.0.0.2 ## Environment Ansible 2.0.0.2 Ubuntu 14.03 AWS ## Summary I am trying to apply outbound firewall rules to AWS using the rules_egress: However, when I log into my AWS account and look at the winrdp security group, all I see are the inbound rules and no outbound rules. ### dev-environment.yml ``` # security groups to be created security_groups: - name: winrdp desc: the security group for the winrdp server rules: - proto: tcp from_port: 5986 to_port: 5986 cidr_ip: 0.0.0.0/0 - proto: tcp from_port: 3389 to_port: 3389 cidr_ip: 0.0.0.0/0 rules_egress: - proto: all cidr_ip: 0.0.0.0/0 ``` ### playbook ``` --- - name: Provision ec2 instances based on the environment hosts: localhost connection: local gather_facts: False vars_files: - vars/dev-environment.yml - vars/aws-creds.yml tasks: - name: Create required security groups ec2_group: name: "{{ item.name }}" description: "{{ item.desc }}" rules: "{{ item.rules }}" region: "{{ ec2_region }}" ec2_access_key: "{{ ec2_access_key }}" ec2_secret_key: "{{ ec2_secret_key }}" with_items: security_groups ``` ![screenshot from 2016-02-16 20 31 24](https://cloud.githubusercontent.com/assets/6406166/13097103/63c440a8-d4ec-11e5-8769-2ecc6f1c651e.png)
main
ansible aws group rules egress not working issue type bug report component name group ansible version environment ansible ubuntu aws summary i am trying to apply outbound firewall rules to aws using the rules egress however when i log into my aws account and look at the winrdp security group all i see are the inbound rules and no outbound rules dev environment yml security groups to be created security groups name winrdp desc the security group for the winrdp server rules proto tcp from port to port cidr ip proto tcp from port to port cidr ip rules egress proto all cidr ip playbook name provision instances based on the environment hosts localhost connection local gather facts false vars files vars dev environment yml vars aws creds yml tasks name create required security groups group name item name description item desc rules item rules region region access key access key secret key secret key with items security groups
1
13,932
8,408,576,060
IssuesEvent
2018-10-12 02:27:44
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Rancher networking is slow
area/networking kind/bug kind/performance version/1.6
There are a very big performance gap in Rancher overlay network. This is my rancher configure: ![image](https://cloud.githubusercontent.com/assets/614797/23178314/e6bc4afa-f8a4-11e6-8a07-4700b1e322c6.png) Bandwidth between ros-02 <-> ros-03 by private ip ``` iperf3 -c 192.168.133.195 Connecting to host 192.168.133.195, port 5201 [ 4] local 10.42.89.41 port 48226 connected to 192.168.133.195 port 5201 [ ID] Interval Transfer Bandwidth Retr Cwnd [ 4] 0.00-1.00 sec 238 MBytes 1.99 Gbits/sec 6390 900 KBytes [ 4] 1.00-2.00 sec 119 MBytes 996 Mbits/sec 3789 494 KBytes [ 4] 2.00-3.00 sec 118 MBytes 986 Mbits/sec 3146 954 KBytes [ 4] 3.00-4.00 sec 125 MBytes 1.05 Gbits/sec 5455 373 KBytes [ 4] 4.00-5.00 sec 110 MBytes 923 Mbits/sec 3194 688 KBytes [ 4] 5.00-6.00 sec 111 MBytes 933 Mbits/sec 7845 790 KBytes [ 4] 6.00-7.00 sec 116 MBytes 975 Mbits/sec 3785 818 KBytes [ 4] 7.00-8.00 sec 119 MBytes 996 Mbits/sec 3756 1.53 MBytes [ 4] 8.00-9.00 sec 125 MBytes 1.05 Gbits/sec 3339 322 KBytes [ 4] 9.00-10.00 sec 119 MBytes 996 Mbits/sec 3276 654 KBytes - - - - - - - - - - - - - - - - - - - - - - - - - [ ID] Interval Transfer Bandwidth Retr [ 4] 0.00-10.00 sec 1.27 GBytes 1.09 Gbits/sec 43975 sender [ 4] 0.00-10.00 sec 1.27 GBytes 1.09 Gbits/sec receiver iperf Done. ``` Bandwidth between ros-02 <-> ros-03 by overlay ip ``` iperf3 -c iperf1 Connecting to host iperf1, port 5201 [ 4] local 10.42.89.41 port 33482 connected to 10.42.122.98 port 5201 [ ID] Interval Transfer Bandwidth Retr Cwnd [ 4] 0.00-1.00 sec 56.5 MBytes 474 Mbits/sec 973 354 KBytes [ 4] 1.00-2.00 sec 39.6 MBytes 332 Mbits/sec 1167 962 KBytes [ 4] 2.00-3.00 sec 40.0 MBytes 335 Mbits/sec 4529 266 KBytes [ 4] 3.00-4.00 sec 40.0 MBytes 336 Mbits/sec 2841 673 KBytes [ 4] 4.00-5.00 sec 40.0 MBytes 335 Mbits/sec 3941 2.63 KBytes [ 4] 5.00-6.00 sec 41.2 MBytes 346 Mbits/sec 2610 2.63 KBytes [ 4] 6.00-7.00 sec 43.8 MBytes 367 Mbits/sec 3605 55.3 KBytes [ 4] 7.00-8.00 sec 41.2 MBytes 346 Mbits/sec 2802 675 KBytes [ 4] 8.00-9.00 sec 42.5 MBytes 357 Mbits/sec 4199 22.4 KBytes [ 4] 9.00-10.00 sec 42.5 MBytes 357 Mbits/sec 3283 51.3 KBytes - - - - - - - - - - - - - - - - - - - - - - - - - [ ID] Interval Transfer Bandwidth Retr [ 4] 0.00-10.00 sec 427 MBytes 358 Mbits/sec 29950 sender [ 4] 0.00-10.00 sec 425 MBytes 356 Mbits/sec receiver iperf Done. ``` I tried both IPSec and VXLAN. Any ideas? --- | Useful | Info | | :-- | :-- | |Versions|Rancher `v1.4.1` Cattle: `v0.176.9` UI: `v1.4.6` | |Access|`localauth` `admin`| |Orchestration|`Cattle`| |Route|`authenticated.project.help`|
True
Rancher networking is slow - There are a very big performance gap in Rancher overlay network. This is my rancher configure: ![image](https://cloud.githubusercontent.com/assets/614797/23178314/e6bc4afa-f8a4-11e6-8a07-4700b1e322c6.png) Bandwidth between ros-02 <-> ros-03 by private ip ``` iperf3 -c 192.168.133.195 Connecting to host 192.168.133.195, port 5201 [ 4] local 10.42.89.41 port 48226 connected to 192.168.133.195 port 5201 [ ID] Interval Transfer Bandwidth Retr Cwnd [ 4] 0.00-1.00 sec 238 MBytes 1.99 Gbits/sec 6390 900 KBytes [ 4] 1.00-2.00 sec 119 MBytes 996 Mbits/sec 3789 494 KBytes [ 4] 2.00-3.00 sec 118 MBytes 986 Mbits/sec 3146 954 KBytes [ 4] 3.00-4.00 sec 125 MBytes 1.05 Gbits/sec 5455 373 KBytes [ 4] 4.00-5.00 sec 110 MBytes 923 Mbits/sec 3194 688 KBytes [ 4] 5.00-6.00 sec 111 MBytes 933 Mbits/sec 7845 790 KBytes [ 4] 6.00-7.00 sec 116 MBytes 975 Mbits/sec 3785 818 KBytes [ 4] 7.00-8.00 sec 119 MBytes 996 Mbits/sec 3756 1.53 MBytes [ 4] 8.00-9.00 sec 125 MBytes 1.05 Gbits/sec 3339 322 KBytes [ 4] 9.00-10.00 sec 119 MBytes 996 Mbits/sec 3276 654 KBytes - - - - - - - - - - - - - - - - - - - - - - - - - [ ID] Interval Transfer Bandwidth Retr [ 4] 0.00-10.00 sec 1.27 GBytes 1.09 Gbits/sec 43975 sender [ 4] 0.00-10.00 sec 1.27 GBytes 1.09 Gbits/sec receiver iperf Done. ``` Bandwidth between ros-02 <-> ros-03 by overlay ip ``` iperf3 -c iperf1 Connecting to host iperf1, port 5201 [ 4] local 10.42.89.41 port 33482 connected to 10.42.122.98 port 5201 [ ID] Interval Transfer Bandwidth Retr Cwnd [ 4] 0.00-1.00 sec 56.5 MBytes 474 Mbits/sec 973 354 KBytes [ 4] 1.00-2.00 sec 39.6 MBytes 332 Mbits/sec 1167 962 KBytes [ 4] 2.00-3.00 sec 40.0 MBytes 335 Mbits/sec 4529 266 KBytes [ 4] 3.00-4.00 sec 40.0 MBytes 336 Mbits/sec 2841 673 KBytes [ 4] 4.00-5.00 sec 40.0 MBytes 335 Mbits/sec 3941 2.63 KBytes [ 4] 5.00-6.00 sec 41.2 MBytes 346 Mbits/sec 2610 2.63 KBytes [ 4] 6.00-7.00 sec 43.8 MBytes 367 Mbits/sec 3605 55.3 KBytes [ 4] 7.00-8.00 sec 41.2 MBytes 346 Mbits/sec 2802 675 KBytes [ 4] 8.00-9.00 sec 42.5 MBytes 357 Mbits/sec 4199 22.4 KBytes [ 4] 9.00-10.00 sec 42.5 MBytes 357 Mbits/sec 3283 51.3 KBytes - - - - - - - - - - - - - - - - - - - - - - - - - [ ID] Interval Transfer Bandwidth Retr [ 4] 0.00-10.00 sec 427 MBytes 358 Mbits/sec 29950 sender [ 4] 0.00-10.00 sec 425 MBytes 356 Mbits/sec receiver iperf Done. ``` I tried both IPSec and VXLAN. Any ideas? --- | Useful | Info | | :-- | :-- | |Versions|Rancher `v1.4.1` Cattle: `v0.176.9` UI: `v1.4.6` | |Access|`localauth` `admin`| |Orchestration|`Cattle`| |Route|`authenticated.project.help`|
non_main
rancher networking is slow there are a very big performance gap in rancher overlay network this is my rancher configure bandwidth between ros ros by private ip c connecting to host port local port connected to port interval transfer bandwidth retr cwnd sec mbytes gbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes gbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec mbytes sec mbytes gbits sec kbytes sec mbytes mbits sec kbytes interval transfer bandwidth retr sec gbytes gbits sec sender sec gbytes gbits sec receiver iperf done bandwidth between ros ros by overlay ip c connecting to host port local port connected to port interval transfer bandwidth retr cwnd sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes sec mbytes mbits sec kbytes interval transfer bandwidth retr sec mbytes mbits sec sender sec mbytes mbits sec receiver iperf done i tried both ipsec and vxlan any ideas useful info versions rancher cattle ui access localauth admin orchestration cattle route authenticated project help
0
385,372
26,633,642,277
IssuesEvent
2023-01-24 19:55:44
jlbeidler/cemconvert
https://api.github.com/repos/jlbeidler/cemconvert
opened
Create documentation
documentation
So far there is little documentation. Need specific documentation for both pieces of the tool including how to run for base and future years.
1.0
Create documentation - So far there is little documentation. Need specific documentation for both pieces of the tool including how to run for base and future years.
non_main
create documentation so far there is little documentation need specific documentation for both pieces of the tool including how to run for base and future years
0
4,608
23,860,630,070
IssuesEvent
2022-09-07 06:39:16
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFP - threema
Status: Available For Maintainer(s)
Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL :https://threema.ch/en/download Direct download URL for the software / installer : https://releases.threema.ch/web-electron/v1/release/Threema-Latest.exe Software summary / short description: messaging app
True
RFP - threema - Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL :https://threema.ch/en/download Direct download URL for the software / installer : https://releases.threema.ch/web-electron/v1/release/Threema-Latest.exe Software summary / short description: messaging app
main
rfp threema checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts with rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description messaging app
1
46,662
11,866,052,983
IssuesEvent
2020-03-26 02:26:26
spack/spack
https://api.github.com/repos/spack/spack
opened
autoconf and other packages on ppc64le
build-error
### Spack version <!-- Add the output to the command below --> ```console [kai@longhorn ~]$ spack --version ``` ### Steps to reproduce the issue ```console [kai@longhorn ~]$ spack spec autoconf Input spec -------------------------------- autoconf Concretized -------------------------------- autoconf@2.69%gcc@7.3.0 arch=linux-rhel7-power9le ^m4@1.4.18%gcc@7.3.0 patches=3877ab548f88597ab2327a2230ee048d2d07ace1062efe81fc92e91b7f39cd00,fc9b61654a3ba1a8d6cd78ce087e7c96366c290bc8d2c299f09828d793b853c8 +sigsegv arch=linux-rhel7-power9le ^libsigsegv@2.12%gcc@7.3.0 arch=linux-rhel7-power9le ^perl@5.30.1%gcc@7.3.0+cpanm+shared+threads arch=linux-rhel7-power9le ^gdbm@1.18.1%gcc@7.3.0 arch=linux-rhel7-power9le ^readline@8.0%gcc@7.3.0 arch=linux-rhel7-power9le ^ncurses@6.2%gcc@7.3.0~symlinks+termlib arch=linux-rhel7-power9le ^pkgconf@1.6.3%gcc@7.3.0 arch=linux-rhel7-power9le [kai@longhorn ~]$ spack install autoconf [...] See build log for details: /tmp/kai/spack-stage/spack-stage-autoconf-2.69-ftyunbfd663jlfj24legpgewbdsjygse/spack-build-out.txt Traceback (most recent call last): File "/home/01537/kai/build/spack/lib/spack/spack/build_environment.py", line 801, in child_process return_value = function() File "/home/01537/kai/build/spack/lib/spack/spack/installer.py", line 1113, in build_process phase(pkg.spec, pkg.prefix) File "/home/01537/kai/build/spack/lib/spack/spack/package.py", line 112, in phase_wrapper callback(instance) File "/home/01537/kai/build/spack/lib/spack/spack/build_systems/autotools.py", line 160, in _do_patch_config_guess raise RuntimeError('Failed to find suitable config.guess') RuntimeError: Failed to find suitable config.guess ``` ### Platform and user environment ```console [kai@longhorn ~]$ uname -a Linux login1.longhorn.tacc.utexas.edu 4.14.0-115.10.1.el7a.ppc64le #1 SMP Wed Jun 26 09:32:17 UTC 2019 ppc64le ppc64le ppc64le GNU/Linux [kai@longhorn ~]$ cat /etc/redhat-release Red Hat Enterprise Linux Server release 7.6 (Maipo) ``` This machine (longhorn) is similar to Summit, except that it has very little software installed. I've tracked the issue down already: When trying to install `autoconf`, spack checks whether the included `config.guess` works. It does not, because it does not recognize what should be `powerpc64le-unknown-linux-gnu`. So spack is looking for a newer `config.guess` to replace the included one. On Summit, it'll find one in `/usr/share/automake-x.yy`, and things work. On this machine, automake is not installed (at least not in the standard location). If the package depended on `automake`, it'd look at the spack-installed automake to find a newer `config.guess`. That doesn't apply here. I don't have access to install anything in `/usr/share`, so the question is, how can I provide an updated `config.guess`? I've hacked around it for now by modifying spack's source so that it searches in my home directory, but clearly that's not a sustainable solution. [This problem isn't really limited to just the autoconf package, either, but any package that ships with an outdated `config.guess`, which doesn't depend on `automake`: `libsodium` was another manifestation of the same issue.] One way to make it possible to work around this problem would be for spack to search for `config.guess` in a user-specified location, though I still don't like it, since the user would still have to go find an appropriate `config.guess`, and point spack to it. I'd much rather have something that works out of the box, but I don't have any good idea on how to get it done.
1.0
autoconf and other packages on ppc64le - ### Spack version <!-- Add the output to the command below --> ```console [kai@longhorn ~]$ spack --version ``` ### Steps to reproduce the issue ```console [kai@longhorn ~]$ spack spec autoconf Input spec -------------------------------- autoconf Concretized -------------------------------- autoconf@2.69%gcc@7.3.0 arch=linux-rhel7-power9le ^m4@1.4.18%gcc@7.3.0 patches=3877ab548f88597ab2327a2230ee048d2d07ace1062efe81fc92e91b7f39cd00,fc9b61654a3ba1a8d6cd78ce087e7c96366c290bc8d2c299f09828d793b853c8 +sigsegv arch=linux-rhel7-power9le ^libsigsegv@2.12%gcc@7.3.0 arch=linux-rhel7-power9le ^perl@5.30.1%gcc@7.3.0+cpanm+shared+threads arch=linux-rhel7-power9le ^gdbm@1.18.1%gcc@7.3.0 arch=linux-rhel7-power9le ^readline@8.0%gcc@7.3.0 arch=linux-rhel7-power9le ^ncurses@6.2%gcc@7.3.0~symlinks+termlib arch=linux-rhel7-power9le ^pkgconf@1.6.3%gcc@7.3.0 arch=linux-rhel7-power9le [kai@longhorn ~]$ spack install autoconf [...] See build log for details: /tmp/kai/spack-stage/spack-stage-autoconf-2.69-ftyunbfd663jlfj24legpgewbdsjygse/spack-build-out.txt Traceback (most recent call last): File "/home/01537/kai/build/spack/lib/spack/spack/build_environment.py", line 801, in child_process return_value = function() File "/home/01537/kai/build/spack/lib/spack/spack/installer.py", line 1113, in build_process phase(pkg.spec, pkg.prefix) File "/home/01537/kai/build/spack/lib/spack/spack/package.py", line 112, in phase_wrapper callback(instance) File "/home/01537/kai/build/spack/lib/spack/spack/build_systems/autotools.py", line 160, in _do_patch_config_guess raise RuntimeError('Failed to find suitable config.guess') RuntimeError: Failed to find suitable config.guess ``` ### Platform and user environment ```console [kai@longhorn ~]$ uname -a Linux login1.longhorn.tacc.utexas.edu 4.14.0-115.10.1.el7a.ppc64le #1 SMP Wed Jun 26 09:32:17 UTC 2019 ppc64le ppc64le ppc64le GNU/Linux [kai@longhorn ~]$ cat /etc/redhat-release Red Hat Enterprise Linux Server release 7.6 (Maipo) ``` This machine (longhorn) is similar to Summit, except that it has very little software installed. I've tracked the issue down already: When trying to install `autoconf`, spack checks whether the included `config.guess` works. It does not, because it does not recognize what should be `powerpc64le-unknown-linux-gnu`. So spack is looking for a newer `config.guess` to replace the included one. On Summit, it'll find one in `/usr/share/automake-x.yy`, and things work. On this machine, automake is not installed (at least not in the standard location). If the package depended on `automake`, it'd look at the spack-installed automake to find a newer `config.guess`. That doesn't apply here. I don't have access to install anything in `/usr/share`, so the question is, how can I provide an updated `config.guess`? I've hacked around it for now by modifying spack's source so that it searches in my home directory, but clearly that's not a sustainable solution. [This problem isn't really limited to just the autoconf package, either, but any package that ships with an outdated `config.guess`, which doesn't depend on `automake`: `libsodium` was another manifestation of the same issue.] One way to make it possible to work around this problem would be for spack to search for `config.guess` in a user-specified location, though I still don't like it, since the user would still have to go find an appropriate `config.guess`, and point spack to it. I'd much rather have something that works out of the box, but I don't have any good idea on how to get it done.
non_main
autoconf and other packages on spack version console spack version steps to reproduce the issue console spack spec autoconf input spec autoconf concretized autoconf gcc arch linux gcc patches sigsegv arch linux libsigsegv gcc arch linux perl gcc cpanm shared threads arch linux gdbm gcc arch linux readline gcc arch linux ncurses gcc symlinks termlib arch linux pkgconf gcc arch linux spack install autoconf see build log for details tmp kai spack stage spack stage autoconf spack build out txt traceback most recent call last file home kai build spack lib spack spack build environment py line in child process return value function file home kai build spack lib spack spack installer py line in build process phase pkg spec pkg prefix file home kai build spack lib spack spack package py line in phase wrapper callback instance file home kai build spack lib spack spack build systems autotools py line in do patch config guess raise runtimeerror failed to find suitable config guess runtimeerror failed to find suitable config guess platform and user environment console uname a linux longhorn tacc utexas edu smp wed jun utc gnu linux cat etc redhat release red hat enterprise linux server release maipo this machine longhorn is similar to summit except that it has very little software installed i ve tracked the issue down already when trying to install autoconf spack checks whether the included config guess works it does not because it does not recognize what should be unknown linux gnu so spack is looking for a newer config guess to replace the included one on summit it ll find one in usr share automake x yy and things work on this machine automake is not installed at least not in the standard location if the package depended on automake it d look at the spack installed automake to find a newer config guess that doesn t apply here i don t have access to install anything in usr share so the question is how can i provide an updated config guess i ve hacked around it for now by modifying spack s source so that it searches in my home directory but clearly that s not a sustainable solution one way to make it possible to work around this problem would be for spack to search for config guess in a user specified location though i still don t like it since the user would still have to go find an appropriate config guess and point spack to it i d much rather have something that works out of the box but i don t have any good idea on how to get it done
0
4,043
18,872,099,629
IssuesEvent
2021-11-13 11:12:33
git-reports/wolfree
https://api.github.com/repos/git-reports/wolfree
closed
does not give the result
bug good first issue state: maintainer successfully reproduce bug state: await patch update from programmer
Submitter: valerio limit n to infinity Surd[n,n]+ Divide[ln\(40)Power[n,2]-1\(41),ln n ] does not give the result
True
does not give the result - Submitter: valerio limit n to infinity Surd[n,n]+ Divide[ln\(40)Power[n,2]-1\(41),ln n ] does not give the result
main
does not give the result submitter valerio limit n to infinity surd divide ln n does not give the result
1
5,564
27,825,512,750
IssuesEvent
2023-03-19 18:14:06
NIAEFEUP/website-niaefeup-backend
https://api.github.com/repos/NIAEFEUP/website-niaefeup-backend
closed
readability: use all-open compiler plugin for simplifying abstract entities
good first issue maintainability dependencies
Instead of explicitly "opening" abstract entity fields like we are doing in the `Activity` model, we could instead rely on [this](https://kotlinlang.org/docs/all-open-plugin.html) to create a custom `AbstractEntity` that does this for us.
True
readability: use all-open compiler plugin for simplifying abstract entities - Instead of explicitly "opening" abstract entity fields like we are doing in the `Activity` model, we could instead rely on [this](https://kotlinlang.org/docs/all-open-plugin.html) to create a custom `AbstractEntity` that does this for us.
main
readability use all open compiler plugin for simplifying abstract entities instead of explicitly opening abstract entity fields like we are doing in the activity model we could instead rely on to create a custom abstractentity that does this for us
1
4,447
23,134,591,797
IssuesEvent
2022-07-28 13:23:38
yast/yast-auth-client
https://api.github.com/repos/yast/yast-auth-client
closed
Remove pam_ldap/pam_nss configuration support
other-maintainer
pam_ldap and nss_ldap have not been maintained since 2014. We should not be allowing new configurations of these modules, as I would like to begin to remove these in openSUSE. Instead all ldap/krb/ad clients should use the supported and secure sssd module by default.
True
Remove pam_ldap/pam_nss configuration support - pam_ldap and nss_ldap have not been maintained since 2014. We should not be allowing new configurations of these modules, as I would like to begin to remove these in openSUSE. Instead all ldap/krb/ad clients should use the supported and secure sssd module by default.
main
remove pam ldap pam nss configuration support pam ldap and nss ldap have not been maintained since we should not be allowing new configurations of these modules as i would like to begin to remove these in opensuse instead all ldap krb ad clients should use the supported and secure sssd module by default
1
99,680
4,059,321,881
IssuesEvent
2016-05-25 09:10:10
jpchanson/TomoBay
https://api.github.com/repos/jpchanson/TomoBay
opened
Royal Mail Security: Header Generator
ACTOR: Application High Priority Sub-Feature
**As an:** application module **I want to:** create a security header that conforms to the Royal Mail API specification https://developer.royalmail.net/api/docdownload/aHR0cHM6Ly9ldS5hcGltLmlibWNsb3VkLmNvbS92MS9wb3J0YWwvYXBpcy81NjBkMTViZjBjZjI5M2Q5MzFhM2NmOGMvdjE4L2RvY3VtZW50cy9STUclMjBTaGlwcGluZyUyMEFQSSUyMCUyOFNPQVAlMjklMjBUZWNobmljYWwlMjBVc2VyJTIwR3VpZGUvZmlsZQ%3D%3D **So That:** API calls can be successfully authenticated
1.0
Royal Mail Security: Header Generator - **As an:** application module **I want to:** create a security header that conforms to the Royal Mail API specification https://developer.royalmail.net/api/docdownload/aHR0cHM6Ly9ldS5hcGltLmlibWNsb3VkLmNvbS92MS9wb3J0YWwvYXBpcy81NjBkMTViZjBjZjI5M2Q5MzFhM2NmOGMvdjE4L2RvY3VtZW50cy9STUclMjBTaGlwcGluZyUyMEFQSSUyMCUyOFNPQVAlMjklMjBUZWNobmljYWwlMjBVc2VyJTIwR3VpZGUvZmlsZQ%3D%3D **So That:** API calls can be successfully authenticated
non_main
royal mail security header generator as an application module i want to create a security header that conforms to the royal mail api specification so that api calls can be successfully authenticated
0
95,300
19,690,573,191
IssuesEvent
2022-01-12 06:15:47
coalton-lang/coalton
https://api.github.com/repos/coalton-lang/coalton
closed
Invalid codegen for calls to functions with nested HKTs when the nested types are the same
bug codegen
*Maintainer Edit: See later in the thread for a simplified test case.* *** ```lisp (declare 5xbigger ((list (list integer)) -> (list (list (list integer))))) (define (5xbigger board) (pipe (range 0 4) (map (fn (addition) (pipe board (map (map (+ addition)))))))) ``` with declare ``` (coalton (pipe 1 make-list make-list 5xbigger)) (((1)) ((2)) ((3)) ((4)) ((5))) ``` without declare ``` (type-of '5xbigger) ∀ :A :B. (FUNCTOR :B) (FUNCTOR :A) ⇒ ((:A (:B INTEGER)) → (LIST (:A (:B INTEGER)))) ``` ``` (coalton (pipe 1 make-list make-list 5xbigger)) #<FUNCTION (COMMON-LISP:LAMBDA (#:G11) :IN COALTON-IMPL/CODEGEN:F3) {10167802DB}> ;; closed over ((1)) ``` Now if I modfy the function like so ``` (define (5xbigger board) (pipe (range 0 4) (map (fn (addition) board)))) ``` Coalton is happy. ``` (type-of '5xbigger) ∀ :A. (:A → (LIST :A)) ``` very interestingly if I add the map-map back in, then recompile, I can produce this error when running the example code: ``` The value (#1=#<COALTON-LIBRARY::CLASS/FUNCTOR {10019D87B3}> #1# #1# #1# #1#) is not of type (OR FUNCTION COALTON-IMPL/CODEGEN::FUNCTION-ENTRY) [Condition of type COMMON-LISP:TYPE-ERROR] ```
1.0
Invalid codegen for calls to functions with nested HKTs when the nested types are the same - *Maintainer Edit: See later in the thread for a simplified test case.* *** ```lisp (declare 5xbigger ((list (list integer)) -> (list (list (list integer))))) (define (5xbigger board) (pipe (range 0 4) (map (fn (addition) (pipe board (map (map (+ addition)))))))) ``` with declare ``` (coalton (pipe 1 make-list make-list 5xbigger)) (((1)) ((2)) ((3)) ((4)) ((5))) ``` without declare ``` (type-of '5xbigger) ∀ :A :B. (FUNCTOR :B) (FUNCTOR :A) ⇒ ((:A (:B INTEGER)) → (LIST (:A (:B INTEGER)))) ``` ``` (coalton (pipe 1 make-list make-list 5xbigger)) #<FUNCTION (COMMON-LISP:LAMBDA (#:G11) :IN COALTON-IMPL/CODEGEN:F3) {10167802DB}> ;; closed over ((1)) ``` Now if I modfy the function like so ``` (define (5xbigger board) (pipe (range 0 4) (map (fn (addition) board)))) ``` Coalton is happy. ``` (type-of '5xbigger) ∀ :A. (:A → (LIST :A)) ``` very interestingly if I add the map-map back in, then recompile, I can produce this error when running the example code: ``` The value (#1=#<COALTON-LIBRARY::CLASS/FUNCTOR {10019D87B3}> #1# #1# #1# #1#) is not of type (OR FUNCTION COALTON-IMPL/CODEGEN::FUNCTION-ENTRY) [Condition of type COMMON-LISP:TYPE-ERROR] ```
non_main
invalid codegen for calls to functions with nested hkts when the nested types are the same maintainer edit see later in the thread for a simplified test case lisp declare list list integer list list list integer define board pipe range map fn addition pipe board map map addition with declare coalton pipe make list make list without declare type of ∀ a b functor b functor a ⇒ a b integer → list a b integer coalton pipe make list make list closed over now if i modfy the function like so define board pipe range map fn addition board coalton is happy type of ∀ a a → list a very interestingly if i add the map map back in then recompile i can produce this error when running the example code the value is not of type or function coalton impl codegen function entry
0
3,493
13,634,716,556
IssuesEvent
2020-09-25 00:41:46
amyjko/faculty
https://api.github.com/repos/amyjko/faculty
closed
Extract paper author name resolution
maintainability
It's currently in the `paper.js` rendering but should be in `model.js`
True
Extract paper author name resolution - It's currently in the `paper.js` rendering but should be in `model.js`
main
extract paper author name resolution it s currently in the paper js rendering but should be in model js
1
3,792
16,110,237,550
IssuesEvent
2021-04-27 20:06:53
svengreb/wand
https://api.github.com/repos/svengreb/wand
opened
Dogfooding: Introduce Mage with wand toolkit
context-techstack context-workflow scope-maintainability scope-quality type-improvement
The project currently only uses _GitHub Action_ workflows for CI but not _Mage_ to automate tasks for itself though. Following the [“dogfooding“ concept][1] _Mage_ will finally be added to the repository, using wand as toolkit through the [`Elder` wand reference][2] implementation. [1]: https://en.wikipedia.org/wiki/Eating_your_own_dog_food [2]: https://pkg.go.dev/github.com/svengreb/wand@v0.5.0/pkg/elder#Elder
True
Dogfooding: Introduce Mage with wand toolkit - The project currently only uses _GitHub Action_ workflows for CI but not _Mage_ to automate tasks for itself though. Following the [“dogfooding“ concept][1] _Mage_ will finally be added to the repository, using wand as toolkit through the [`Elder` wand reference][2] implementation. [1]: https://en.wikipedia.org/wiki/Eating_your_own_dog_food [2]: https://pkg.go.dev/github.com/svengreb/wand@v0.5.0/pkg/elder#Elder
main
dogfooding introduce mage with wand toolkit the project currently only uses github action workflows for ci but not mage to automate tasks for itself though following the mage will finally be added to the repository using wand as toolkit through the implementation
1
617,817
19,405,284,822
IssuesEvent
2021-12-19 22:10:23
OnTopicCMS/OnTopic-Library
https://api.github.com/repos/OnTopicCMS/OnTopic-Library
opened
Mapping: Ensure cache entries are only pulled once
Area: Mapping Severity 0: Nice to have Priority: 3 Type: Improvement Status 2: Scheduled
Currently, when pulling objects from the cache, the cache must be queried twice, due to chaining of `MapAsync()` overloads. This should be avoidable.
1.0
Mapping: Ensure cache entries are only pulled once - Currently, when pulling objects from the cache, the cache must be queried twice, due to chaining of `MapAsync()` overloads. This should be avoidable.
non_main
mapping ensure cache entries are only pulled once currently when pulling objects from the cache the cache must be queried twice due to chaining of mapasync overloads this should be avoidable
0
300,123
9,206,167,615
IssuesEvent
2019-03-08 12:56:19
qissue-bot/QGIS
https://api.github.com/repos/qissue-bot/QGIS
closed
crash when closing project that has an open GRASS mapset
Category: GRASS Component: Affected QGIS version Component: Crashes QGIS or corrupts data Component: Easy fix? Component: Operating System Component: Pull Request or Patch supplied Component: Regression? Component: Resolution Priority: Low Project: QGIS Application Status: Closed Tracker: Bug report
--- Author Name: **Maciej Sieczka -** (Maciej Sieczka -) Original Redmine Issue: 1460, https://issues.qgis.org/issues/1460 Original Assignee: nobody - --- 1. open a GRASS mapset 2. save a project 3. close mapset 4. start a new project - QGIS crashes because of "Uncatched fatal GRASS - MAPSET is not set"
1.0
crash when closing project that has an open GRASS mapset - --- Author Name: **Maciej Sieczka -** (Maciej Sieczka -) Original Redmine Issue: 1460, https://issues.qgis.org/issues/1460 Original Assignee: nobody - --- 1. open a GRASS mapset 2. save a project 3. close mapset 4. start a new project - QGIS crashes because of "Uncatched fatal GRASS - MAPSET is not set"
non_main
crash when closing project that has an open grass mapset author name maciej sieczka maciej sieczka original redmine issue original assignee nobody open a grass mapset save a project close mapset start a new project qgis crashes because of uncatched fatal grass mapset is not set
0
333,108
29,508,146,294
IssuesEvent
2023-06-03 15:21:17
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix jax_numpy_searching_sorting.test_jax_numpy_extract
JAX Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details>
1.0
Fix jax_numpy_searching_sorting.test_jax_numpy_extract - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5164180232/jobs/9302917768" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>Not found</summary> Not found </details>
non_main
fix jax numpy searching sorting test jax numpy extract tensorflow img src torch img src numpy img src jax img src paddle img src not found not found
0
74,859
7,447,624,993
IssuesEvent
2018-03-28 13:08:04
AffiliateWP/AffiliateWP
https://api.github.com/repos/AffiliateWP/AffiliateWP
closed
Introduce affiliate_wp_customers DB table
Has PR enhancement needs testing
PR: #2604 We should introduce a `affiliate_wp_customers` DB table. Possible DB columns: - Customer ID - User ID - Affiliate ID - First Name - Last Name - Email Address (primary/first email used) - Date Created - IP Address (Primary/first used by customer) In addition, we would store additional email addresses and IP addresses used by the customer in affiliate meta. The primary advantage of this new table is allowing the site admin to see the relationships between affiliates and customers. A site admin could see how many customers an affiliate has referred, who they are, and more. Other add-ons such as Lifetime Commissions, Order Details for Affiliates, Signup Referrals could then leverage this data to add additional functionality. Things needed: - [x] New database table - [x] Upgrade routine - [x] Additional email addresses and IP addresses stored in the affiliate meta
1.0
Introduce affiliate_wp_customers DB table - PR: #2604 We should introduce a `affiliate_wp_customers` DB table. Possible DB columns: - Customer ID - User ID - Affiliate ID - First Name - Last Name - Email Address (primary/first email used) - Date Created - IP Address (Primary/first used by customer) In addition, we would store additional email addresses and IP addresses used by the customer in affiliate meta. The primary advantage of this new table is allowing the site admin to see the relationships between affiliates and customers. A site admin could see how many customers an affiliate has referred, who they are, and more. Other add-ons such as Lifetime Commissions, Order Details for Affiliates, Signup Referrals could then leverage this data to add additional functionality. Things needed: - [x] New database table - [x] Upgrade routine - [x] Additional email addresses and IP addresses stored in the affiliate meta
non_main
introduce affiliate wp customers db table pr we should introduce a affiliate wp customers db table possible db columns customer id user id affiliate id first name last name email address primary first email used date created ip address primary first used by customer in addition we would store additional email addresses and ip addresses used by the customer in affiliate meta the primary advantage of this new table is allowing the site admin to see the relationships between affiliates and customers a site admin could see how many customers an affiliate has referred who they are and more other add ons such as lifetime commissions order details for affiliates signup referrals could then leverage this data to add additional functionality things needed new database table upgrade routine additional email addresses and ip addresses stored in the affiliate meta
0
4,851
24,983,183,656
IssuesEvent
2022-11-02 13:19:54
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Remove need for `ActorControl#yieldThread`
kind/toil area/reliability area/maintainability
**Description** Actors currently allow busy looping and holding onto a thread while processing multiple jobs. This is done via a so-called `resubmit` behavior, and callers must call `ActorControl#yieldThread` in order to release the thread, allowing other actors to use it. This behavior is, in my opinion, fundamentally antithetical to the actor model. It's also easy to trigger without realizing. While the worst case is a busy loop, even without looping behavior, it's counter-intuitive for successive run calls to hold onto the thread. I assume it's a performance optimization of some kind (sort of like biased locking), but as it has no upper bound, it can quickly become a performance issue, and there is no real proof that it is optimizing anything at the moment. In some cases, it's obvious that you need to yield - for example, when retrying the same job, most developers will remember to yield (assuming they know about the resubmit behavior in the first place). But in other cases, it becomes unclear that you are, in fact, in a busy loop. I would propose testing its removal, but as this is somewhat risky (due to lack of concurrency test coverage in the scheduler module), not backport it until we're confidant (i.e. after some months of testing in SNAPSHOT).
True
Remove need for `ActorControl#yieldThread` - **Description** Actors currently allow busy looping and holding onto a thread while processing multiple jobs. This is done via a so-called `resubmit` behavior, and callers must call `ActorControl#yieldThread` in order to release the thread, allowing other actors to use it. This behavior is, in my opinion, fundamentally antithetical to the actor model. It's also easy to trigger without realizing. While the worst case is a busy loop, even without looping behavior, it's counter-intuitive for successive run calls to hold onto the thread. I assume it's a performance optimization of some kind (sort of like biased locking), but as it has no upper bound, it can quickly become a performance issue, and there is no real proof that it is optimizing anything at the moment. In some cases, it's obvious that you need to yield - for example, when retrying the same job, most developers will remember to yield (assuming they know about the resubmit behavior in the first place). But in other cases, it becomes unclear that you are, in fact, in a busy loop. I would propose testing its removal, but as this is somewhat risky (due to lack of concurrency test coverage in the scheduler module), not backport it until we're confidant (i.e. after some months of testing in SNAPSHOT).
main
remove need for actorcontrol yieldthread description actors currently allow busy looping and holding onto a thread while processing multiple jobs this is done via a so called resubmit behavior and callers must call actorcontrol yieldthread in order to release the thread allowing other actors to use it this behavior is in my opinion fundamentally antithetical to the actor model it s also easy to trigger without realizing while the worst case is a busy loop even without looping behavior it s counter intuitive for successive run calls to hold onto the thread i assume it s a performance optimization of some kind sort of like biased locking but as it has no upper bound it can quickly become a performance issue and there is no real proof that it is optimizing anything at the moment in some cases it s obvious that you need to yield for example when retrying the same job most developers will remember to yield assuming they know about the resubmit behavior in the first place but in other cases it becomes unclear that you are in fact in a busy loop i would propose testing its removal but as this is somewhat risky due to lack of concurrency test coverage in the scheduler module not backport it until we re confidant i e after some months of testing in snapshot
1
5,021
3,899,289,100
IssuesEvent
2016-04-17 17:01:32
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
13541434: [xcode] misspelled 'self' doesn't give in-code error
classification:ui/usability reproducible:always status:open
#### Description Summary: "self" misspelled does not give an error indication in-code. The messages indication takes you to the build output. Steps to Reproduce: Use 'sef' instead of "self", such as sef.nextCueTimeLabel.text = nextTimeText; Expected Results: You get a red stop-bang in the margin, and an error message blob in the code. Click the red stop-bang to go to the code in question. Actual Results: Nothing in the code. Clicking the red stop-bang takes you to the build output. Regression: I don't recall seeing this in 4.5. But I don't recall having a sef-inflected aphasia like i have been having today. Notes: - Product Version: 4.6 (4H127) Created: 2013-03-30T04:25:48.676513 Originated: 2013-03-30T00:00:00 Open Radar Link: http://www.openradar.me/13541434
True
13541434: [xcode] misspelled 'self' doesn't give in-code error - #### Description Summary: "self" misspelled does not give an error indication in-code. The messages indication takes you to the build output. Steps to Reproduce: Use 'sef' instead of "self", such as sef.nextCueTimeLabel.text = nextTimeText; Expected Results: You get a red stop-bang in the margin, and an error message blob in the code. Click the red stop-bang to go to the code in question. Actual Results: Nothing in the code. Clicking the red stop-bang takes you to the build output. Regression: I don't recall seeing this in 4.5. But I don't recall having a sef-inflected aphasia like i have been having today. Notes: - Product Version: 4.6 (4H127) Created: 2013-03-30T04:25:48.676513 Originated: 2013-03-30T00:00:00 Open Radar Link: http://www.openradar.me/13541434
non_main
misspelled self doesn t give in code error description summary self misspelled does not give an error indication in code the messages indication takes you to the build output steps to reproduce use sef instead of self such as sef nextcuetimelabel text nexttimetext expected results you get a red stop bang in the margin and an error message blob in the code click the red stop bang to go to the code in question actual results nothing in the code clicking the red stop bang takes you to the build output regression i don t recall seeing this in but i don t recall having a sef inflected aphasia like i have been having today notes product version created originated open radar link
0
251,193
18,941,312,844
IssuesEvent
2021-11-18 03:25:33
100Automations/Website
https://api.github.com/repos/100Automations/Website
closed
Assets for website
documentation role: project management size: missing feature: Project Framework
### Overview We will collect assets to put onto website and wiki in this issue ### Action Items - [ ] First Item - [ ] Workshop for Automation https://www.meetup.com/hackforla/events/qlggsrybckbdc/ ### Resources/Instructions [100 Automations Youtube Channel](https://www.youtube.com/channel/UCS_BOwZ5Hj3pGHN0gmx4gAA/?guided_help_flow=5)
1.0
Assets for website - ### Overview We will collect assets to put onto website and wiki in this issue ### Action Items - [ ] First Item - [ ] Workshop for Automation https://www.meetup.com/hackforla/events/qlggsrybckbdc/ ### Resources/Instructions [100 Automations Youtube Channel](https://www.youtube.com/channel/UCS_BOwZ5Hj3pGHN0gmx4gAA/?guided_help_flow=5)
non_main
assets for website overview we will collect assets to put onto website and wiki in this issue action items first item workshop for automation resources instructions
0
703,020
24,144,577,661
IssuesEvent
2022-09-21 17:32:49
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
System.Reflection work planned for .NET 7
Epic area-System.Reflection tracking Priority:3 Cost:XL Team:Libraries
This issue captures the planned work for .NET 7. This list is expected to change throughout the release cycle according to ongoing planning and discussions, with possible additions and subtractions to the scope. ## Planned for .NET 7 ### Features / User stories - ~~[ ] [Developers using reflection invoke should be able to use ref struct](https://github.com/dotnet/runtime/issues/44327)~~ - ~~[ ] [Developers using Reflection should be able to use pointers and function pointers](https://github.com/dotnet/runtime/issues/44327)~~ The following are higher priority reflection issues that are affecting the community and for the most part, do not include features or bugs that have reasonable workarounds. <!-- DispatchProxy - [ ] [System.Reflection.DispatchProxy.Create that accepts parameters to pass to TProxy's constructor](https://github.com/dotnet/runtime/issues/42858)--> ### MetadataLoadContext - [x] [PathAssemblyResolver System.Reflection.MetadataLoadContext should be sealed or allow overrides](https://github.com/dotnet/runtime/issues/36753) ### General cleanup and bugs - [x] [Assembly.CodeBase obsoletion is incomplete](https://github.com/dotnet/runtime/issues/42857) - [x] [PseudoCustomAttribute.GetStructLayoutCustomAttribute returns incorrect data](https://github.com/dotnet/runtime/issues/12480) - [x] [Type.IsPublic returns wrong answer for pointer and byref types](https://github.com/dotnet/runtime/issues/12355) - [x] [Inconsistent GetRuntimeProperties() Behavior](https://github.com/dotnet/runtime/issues/28056) - [x] [Allow GetProperties to order results by declaration order](https://github.com/dotnet/runtime/issues/46272) - [x] [Attribute.GetCustomAttributes returns null when passed an unbound generic type](https://github.com/dotnet/runtime/issues/64335) - [x] [Get info about generic attributes for open generic](https://github.com/dotnet/runtime/issues/64169) - [x] [System.Reflection.Metadata doesn't handle generic attributes](https://github.com/dotnet/runtime/issues/58073) - [x] [Method binding by name with Type.DefaultBinder.BindToMethod fails with specific condition](https://github.com/dotnet/runtime/issues/66237) ### Reflection Emit - [x] [TypeBuilder::CreateTypeInfo() should not throw COMException](https://github.com/dotnet/runtime/issues/43282) - [x] [ModuleBuilder.GetType does not work with ignore case with certain unicode chars on Unix](https://github.com/dotnet/runtime/issues/18231) ### Backlog - ~~[ ] [Support equivalent of AssemblyBuilder.Save to save in-memory IL to an assembly](https://github.com/dotnet/runtime/issues/62956)~~
1.0
System.Reflection work planned for .NET 7 - This issue captures the planned work for .NET 7. This list is expected to change throughout the release cycle according to ongoing planning and discussions, with possible additions and subtractions to the scope. ## Planned for .NET 7 ### Features / User stories - ~~[ ] [Developers using reflection invoke should be able to use ref struct](https://github.com/dotnet/runtime/issues/44327)~~ - ~~[ ] [Developers using Reflection should be able to use pointers and function pointers](https://github.com/dotnet/runtime/issues/44327)~~ The following are higher priority reflection issues that are affecting the community and for the most part, do not include features or bugs that have reasonable workarounds. <!-- DispatchProxy - [ ] [System.Reflection.DispatchProxy.Create that accepts parameters to pass to TProxy's constructor](https://github.com/dotnet/runtime/issues/42858)--> ### MetadataLoadContext - [x] [PathAssemblyResolver System.Reflection.MetadataLoadContext should be sealed or allow overrides](https://github.com/dotnet/runtime/issues/36753) ### General cleanup and bugs - [x] [Assembly.CodeBase obsoletion is incomplete](https://github.com/dotnet/runtime/issues/42857) - [x] [PseudoCustomAttribute.GetStructLayoutCustomAttribute returns incorrect data](https://github.com/dotnet/runtime/issues/12480) - [x] [Type.IsPublic returns wrong answer for pointer and byref types](https://github.com/dotnet/runtime/issues/12355) - [x] [Inconsistent GetRuntimeProperties() Behavior](https://github.com/dotnet/runtime/issues/28056) - [x] [Allow GetProperties to order results by declaration order](https://github.com/dotnet/runtime/issues/46272) - [x] [Attribute.GetCustomAttributes returns null when passed an unbound generic type](https://github.com/dotnet/runtime/issues/64335) - [x] [Get info about generic attributes for open generic](https://github.com/dotnet/runtime/issues/64169) - [x] [System.Reflection.Metadata doesn't handle generic attributes](https://github.com/dotnet/runtime/issues/58073) - [x] [Method binding by name with Type.DefaultBinder.BindToMethod fails with specific condition](https://github.com/dotnet/runtime/issues/66237) ### Reflection Emit - [x] [TypeBuilder::CreateTypeInfo() should not throw COMException](https://github.com/dotnet/runtime/issues/43282) - [x] [ModuleBuilder.GetType does not work with ignore case with certain unicode chars on Unix](https://github.com/dotnet/runtime/issues/18231) ### Backlog - ~~[ ] [Support equivalent of AssemblyBuilder.Save to save in-memory IL to an assembly](https://github.com/dotnet/runtime/issues/62956)~~
non_main
system reflection work planned for net this issue captures the planned work for net this list is expected to change throughout the release cycle according to ongoing planning and discussions with possible additions and subtractions to the scope planned for net features user stories the following are higher priority reflection issues that are affecting the community and for the most part do not include features or bugs that have reasonable workarounds dispatchproxy metadataloadcontext general cleanup and bugs reflection emit backlog
0
3,554
14,137,828,442
IssuesEvent
2020-11-10 07:23:14
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
ComboBox in expand/open mode is fully displayed
status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬 type: question ❓
<!-- Hi there! 👋 Hope everything is going okay using projects from the Carbon Design System. It looks like you might have a question about our work, so we wanted to share a couple resources that you could use if you haven't tried them yet 🙂. If you're an IBMer, we have a couple of Slack channels available across all IBM Workspaces: - #carbon-design-system for questions about the Design System - #carbon-components for questions about component styles - #carbon-react for questions about our React components If these resources don't work out, help us out by filling out a couple of details below! --> ## What package(s) are you using? <!-- Add an x in one of the options below, for example: - [x] package name --> - [ ] `carbon-components` - [x] `carbon-components-react` ## Summary I'm trying to use Combox and right after that I have a SplitPane, the problem is when opening the combo... the list of items shrunken! ![image](https://user-images.githubusercontent.com/26740851/96436036-85e2af80-120e-11eb-93eb-9918c38ae554.png) Tried to fix it by css `bx--list-box__menu` but didn't manage. If I use other dropdown (like Symantic for example) the list is opened as expected over the SplitPane. ## Relevant information Sandbox: https://codesandbox.io/s/lucid-https-65srb?file=/src/App.js **Note**: when I enable the import of carbon-react Combobox the sandbox is corrupted with error: Maximum call stack size exceeded !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! <!-- Provide as much useful information as you can -->
True
ComboBox in expand/open mode is fully displayed - <!-- Hi there! 👋 Hope everything is going okay using projects from the Carbon Design System. It looks like you might have a question about our work, so we wanted to share a couple resources that you could use if you haven't tried them yet 🙂. If you're an IBMer, we have a couple of Slack channels available across all IBM Workspaces: - #carbon-design-system for questions about the Design System - #carbon-components for questions about component styles - #carbon-react for questions about our React components If these resources don't work out, help us out by filling out a couple of details below! --> ## What package(s) are you using? <!-- Add an x in one of the options below, for example: - [x] package name --> - [ ] `carbon-components` - [x] `carbon-components-react` ## Summary I'm trying to use Combox and right after that I have a SplitPane, the problem is when opening the combo... the list of items shrunken! ![image](https://user-images.githubusercontent.com/26740851/96436036-85e2af80-120e-11eb-93eb-9918c38ae554.png) Tried to fix it by css `bx--list-box__menu` but didn't manage. If I use other dropdown (like Symantic for example) the list is opened as expected over the SplitPane. ## Relevant information Sandbox: https://codesandbox.io/s/lucid-https-65srb?file=/src/App.js **Note**: when I enable the import of carbon-react Combobox the sandbox is corrupted with error: Maximum call stack size exceeded !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! <!-- Provide as much useful information as you can -->
main
combobox in expand open mode is fully displayed hi there 👋 hope everything is going okay using projects from the carbon design system it looks like you might have a question about our work so we wanted to share a couple resources that you could use if you haven t tried them yet 🙂 if you re an ibmer we have a couple of slack channels available across all ibm workspaces carbon design system for questions about the design system carbon components for questions about component styles carbon react for questions about our react components if these resources don t work out help us out by filling out a couple of details below what package s are you using add an x in one of the options below for example package name carbon components carbon components react summary i m trying to use combox and right after that i have a splitpane the problem is when opening the combo the list of items shrunken tried to fix it by css bx list box menu but didn t manage if i use other dropdown like symantic for example the list is opened as expected over the splitpane relevant information sandbox note when i enable the import of carbon react combobox the sandbox is corrupted with error maximum call stack size exceeded
1
223,306
17,110,977,042
IssuesEvent
2021-07-10 09:32:23
QPI-kr/admin-server
https://api.github.com/repos/QPI-kr/admin-server
opened
Match Page
documentation
- [ ] 남자매칭 & 여자매칭 버튼으로 구분 - [ ] 섹션 분할(남 / 여) - [ ] 각 유저 클릭시 추천 유저 리스팅(validation 후, 매칭 이력 없고 친구 관계 아니고) - [ ] 각 유저에 상세보기 버튼 추가 및 클릭시 팝업 오픈 - [ ] 두 명의 유저 클릭하면 매칭하기 버튼 활성화, 클릭시 API 전송
1.0
Match Page - - [ ] 남자매칭 & 여자매칭 버튼으로 구분 - [ ] 섹션 분할(남 / 여) - [ ] 각 유저 클릭시 추천 유저 리스팅(validation 후, 매칭 이력 없고 친구 관계 아니고) - [ ] 각 유저에 상세보기 버튼 추가 및 클릭시 팝업 오픈 - [ ] 두 명의 유저 클릭하면 매칭하기 버튼 활성화, 클릭시 API 전송
non_main
match page 남자매칭 여자매칭 버튼으로 구분 섹션 분할 남 여 각 유저 클릭시 추천 유저 리스팅 validation 후 매칭 이력 없고 친구 관계 아니고 각 유저에 상세보기 버튼 추가 및 클릭시 팝업 오픈 두 명의 유저 클릭하면 매칭하기 버튼 활성화 클릭시 api 전송
0
424
2,608,627,452
IssuesEvent
2015-02-26 08:39:31
zotero/zotero
https://api.github.com/repos/zotero/zotero
opened
Vacuum database on idle after a month or two
Performance
For reference: https://wiki.mozilla.org/Firefox/Projects/Places_Vacuum And maybe disable auto-vacuum, which can actually slow things down.
True
Vacuum database on idle after a month or two - For reference: https://wiki.mozilla.org/Firefox/Projects/Places_Vacuum And maybe disable auto-vacuum, which can actually slow things down.
non_main
vacuum database on idle after a month or two for reference and maybe disable auto vacuum which can actually slow things down
0
158,040
20,007,557,787
IssuesEvent
2022-02-01 00:00:46
timf-app-sandbox/t1
https://api.github.com/repos/timf-app-sandbox/t1
opened
CVE-2022-21681 (High) detected in marked-0.3.9.tgz
security vulnerability
## CVE-2022-21681 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.9.tgz</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.3.9.tgz">https://registry.npmjs.org/marked/-/marked-0.3.9.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/marked/package.json</p> <p> Dependency Hierarchy: - :x: **marked-0.3.9.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/timf-app-sandbox/t1/commit/ade63e4d4d7f78775fe064056f81c5da1b53aecf">ade63e4d4d7f78775fe064056f81c5da1b53aecf</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `inline.reflinkSearch` may cause catastrophic backtracking against some strings and lead to a denial of service (DoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources. <p>Publish Date: 2022-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21681>CVE-2022-21681</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-5v2h-r2cx-5xgj">https://github.com/advisories/GHSA-5v2h-r2cx-5xgj</a></p> <p>Release Date: 2022-01-14</p> <p>Fix Resolution: marked - 4.0.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"marked","packageVersion":"0.3.9","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"marked:0.3.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"marked - 4.0.10","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2022-21681","vulnerabilityDetails":"Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `inline.reflinkSearch` may cause catastrophic backtracking against some strings and lead to a denial of service (DoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21681","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2022-21681 (High) detected in marked-0.3.9.tgz - ## CVE-2022-21681 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.9.tgz</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://registry.npmjs.org/marked/-/marked-0.3.9.tgz">https://registry.npmjs.org/marked/-/marked-0.3.9.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/marked/package.json</p> <p> Dependency Hierarchy: - :x: **marked-0.3.9.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/timf-app-sandbox/t1/commit/ade63e4d4d7f78775fe064056f81c5da1b53aecf">ade63e4d4d7f78775fe064056f81c5da1b53aecf</a></p> <p>Found in base branch: <b>dev</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `inline.reflinkSearch` may cause catastrophic backtracking against some strings and lead to a denial of service (DoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources. <p>Publish Date: 2022-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21681>CVE-2022-21681</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-5v2h-r2cx-5xgj">https://github.com/advisories/GHSA-5v2h-r2cx-5xgj</a></p> <p>Release Date: 2022-01-14</p> <p>Fix Resolution: marked - 4.0.10</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"marked","packageVersion":"0.3.9","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"marked:0.3.9","isMinimumFixVersionAvailable":true,"minimumFixVersion":"marked - 4.0.10","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2022-21681","vulnerabilityDetails":"Marked is a markdown parser and compiler. Prior to version 4.0.10, the regular expression `inline.reflinkSearch` may cause catastrophic backtracking against some strings and lead to a denial of service (DoS). Anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected. This issue is patched in version 4.0.10. As a workaround, avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21681","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
cve high detected in marked tgz cve high severity vulnerability vulnerable library marked tgz a markdown parser built for speed library home page a href path to dependency file package json path to vulnerable library node modules marked package json dependency hierarchy x marked tgz vulnerable library found in head commit a href found in base branch dev vulnerability details marked is a markdown parser and compiler prior to version the regular expression inline reflinksearch may cause catastrophic backtracking against some strings and lead to a denial of service dos anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected this issue is patched in version as a workaround avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution marked rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree marked isminimumfixversionavailable true minimumfixversion marked isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails marked is a markdown parser and compiler prior to version the regular expression inline reflinksearch may cause catastrophic backtracking against some strings and lead to a denial of service dos anyone who runs untrusted markdown through a vulnerable version of marked and does not use a worker with a time limit may be affected this issue is patched in version as a workaround avoid running untrusted markdown through marked or run marked on a worker thread and set a reasonable time limit to prevent draining resources vulnerabilityurl
0
1,547
6,572,237,148
IssuesEvent
2017-09-11 00:26:31
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
Problem with unit size description in lvol module
affects_2.0 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report AND - Documentation Report ##### COMPONENT NAME system/lvol.py ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` But the code is present in devel, so it'll still work. ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Any Linux ##### SUMMARY In the section of code where the size of the LV is determined there is the following section (238 to 242) ``` if not '%' in size: # LVCREATE(8) -L --size option unit if size[-1].lower() in 'bskmgtpe': size_unit = size[-1].lower() size = size[0:-1] ``` Line 241 causes the unit to be shifted into lowercase, which causes the volume to be expressed in mebibytes instead of megabytes in all cases (power of 2 instead of base 10). This could cause issues if you are using values reported from tools like lvs, as those tools will properly report values issued in the correct unit (`lvs --unit m` returns different values than `lvs --unit M`). Additionally, it is clearly contrary to the documentation: ``` size: description: - The size of the logical volume, according to lvcreate(8) --size, by default in megabytes or optionally with one of [bBsSkKmMgGtTpPeE] units; or according to lvcreate(8) --extents as a percentage of [VG|PVS|FREE]; Float values must begin with a digit. Resizing using percentage values was not supported prior to 2.1. ``` While removing the str.lower() will cause user playbooks to break all over the place, we could edit the documentation to clearly state what the module actually does (always lowercase m power of 2 mebibytes). It would probably also be a good opportunity to be clarify the unit type in use (when the LV tools documentation was written mebibytes had not yet been ratified as a standard, it calls both power of 2 and base 10 units "megabytes). ##### STEPS TO REPRODUCE With a prexisting LV (you can get size in mebibytes using lvs --unit m) ``` ansible 'somehost' -m lvol -a 'vg=somevg lv=somelv size=51196m' ... somehost | SUCCESS => { "changed": false, "lv": "somelv", "size": 51196, "vg": "somevg" } ... ansible 'somehost' -m lvol -a 'vg=somevg lv=somelv size=51196M' ... somehost | SUCCESS => { "changed": false, "lv": "somelv", "size": 51196, "vg": "somevg" } ``` ##### EXPECTED RESULTS `ansible 'somehost' -m lvol -a 'vg=somevg lv=somelv size=51196M'` should fail as it is technically an attempt to shrink the volume. It does not fail because the code at line 241 in lvol.py silently changes the Megabyte unit to the Mebibyte unit (from base 10 to power of 2). ##### ACTUAL RESULTS Nothing (but it does not work as per documentation)
True
Problem with unit size description in lvol module - ##### ISSUE TYPE - Bug Report AND - Documentation Report ##### COMPONENT NAME system/lvol.py ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` But the code is present in devel, so it'll still work. ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Any Linux ##### SUMMARY In the section of code where the size of the LV is determined there is the following section (238 to 242) ``` if not '%' in size: # LVCREATE(8) -L --size option unit if size[-1].lower() in 'bskmgtpe': size_unit = size[-1].lower() size = size[0:-1] ``` Line 241 causes the unit to be shifted into lowercase, which causes the volume to be expressed in mebibytes instead of megabytes in all cases (power of 2 instead of base 10). This could cause issues if you are using values reported from tools like lvs, as those tools will properly report values issued in the correct unit (`lvs --unit m` returns different values than `lvs --unit M`). Additionally, it is clearly contrary to the documentation: ``` size: description: - The size of the logical volume, according to lvcreate(8) --size, by default in megabytes or optionally with one of [bBsSkKmMgGtTpPeE] units; or according to lvcreate(8) --extents as a percentage of [VG|PVS|FREE]; Float values must begin with a digit. Resizing using percentage values was not supported prior to 2.1. ``` While removing the str.lower() will cause user playbooks to break all over the place, we could edit the documentation to clearly state what the module actually does (always lowercase m power of 2 mebibytes). It would probably also be a good opportunity to be clarify the unit type in use (when the LV tools documentation was written mebibytes had not yet been ratified as a standard, it calls both power of 2 and base 10 units "megabytes). ##### STEPS TO REPRODUCE With a prexisting LV (you can get size in mebibytes using lvs --unit m) ``` ansible 'somehost' -m lvol -a 'vg=somevg lv=somelv size=51196m' ... somehost | SUCCESS => { "changed": false, "lv": "somelv", "size": 51196, "vg": "somevg" } ... ansible 'somehost' -m lvol -a 'vg=somevg lv=somelv size=51196M' ... somehost | SUCCESS => { "changed": false, "lv": "somelv", "size": 51196, "vg": "somevg" } ``` ##### EXPECTED RESULTS `ansible 'somehost' -m lvol -a 'vg=somevg lv=somelv size=51196M'` should fail as it is technically an attempt to shrink the volume. It does not fail because the code at line 241 in lvol.py silently changes the Megabyte unit to the Mebibyte unit (from base 10 to power of 2). ##### ACTUAL RESULTS Nothing (but it does not work as per documentation)
main
problem with unit size description in lvol module issue type bug report and documentation report component name system lvol py ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides but the code is present in devel so it ll still work configuration n a os environment any linux summary in the section of code where the size of the lv is determined there is the following section to if not in size lvcreate l size option unit if size lower in bskmgtpe size unit size lower size size line causes the unit to be shifted into lowercase which causes the volume to be expressed in mebibytes instead of megabytes in all cases power of instead of base this could cause issues if you are using values reported from tools like lvs as those tools will properly report values issued in the correct unit lvs unit m returns different values than lvs unit m additionally it is clearly contrary to the documentation size description the size of the logical volume according to lvcreate size by default in megabytes or optionally with one of units or according to lvcreate extents as a percentage of float values must begin with a digit resizing using percentage values was not supported prior to while removing the str lower will cause user playbooks to break all over the place we could edit the documentation to clearly state what the module actually does always lowercase m power of mebibytes it would probably also be a good opportunity to be clarify the unit type in use when the lv tools documentation was written mebibytes had not yet been ratified as a standard it calls both power of and base units megabytes steps to reproduce with a prexisting lv you can get size in mebibytes using lvs unit m ansible somehost m lvol a vg somevg lv somelv size somehost success changed false lv somelv size vg somevg ansible somehost m lvol a vg somevg lv somelv size somehost success changed false lv somelv size vg somevg expected results ansible somehost m lvol a vg somevg lv somelv size should fail as it is technically an attempt to shrink the volume it does not fail because the code at line in lvol py silently changes the megabyte unit to the mebibyte unit from base to power of actual results nothing but it does not work as per documentation
1
5,730
30,292,232,158
IssuesEvent
2023-07-09 12:27:57
svengreb/wand
https://api.github.com/repos/svengreb/wand
opened
`go run` support for versioned modules (Go 1.17+)
context-api context-pkg scope-compatibility scope-dx scope-maintainability scope-stability type-feature
[As of Go 1.17 the `go run` command can finally run in module-aware mode][1] while not “polluting“ the current module in the working directory, of there is one (`go.mod` file present) 🎉 This finally allows to [run commands _on-the-fly_](https://pkg.go.dev/cmd/go#hdr-Compile_and_run_Go_program) of Go `main` module packages without installing them or without changing dependencies of the current module! To support this feature with _wand_ a new [`task.GoModule`][2] will be implemented in a new [`golang/run`][3] package. It can be run using a [command runner][4] that handles tasks of kind [`KindGoModule`][5] so mainly [`gotool.Runner`][6]. The new [`golang/run.Task`][3] will be customizable through the following functions: - `WithArgs(...string) run.Option` — sets additional arguments to pass to the command. - `WithEnv(map[string]string) run.Option` — sets the task specific environment. - `WithModulePath(string) run.Option` — sets the module import path. - `WithModuleVersion(*semver.Version) run.Option` — sets the module version. Next to the new task the [`gotool.Runner`][6] will be adjusted to a new [`WithCache(bool)`][9] runner option to toggle the usage of the local cache directory in the root directory of the module. The runner will be made “smart“ in the way that it either… - installing the executable through a [`golang.Runner`][8], which runs `go install pkg@version` to [leverage Go 1.16‘s feature](https://github.com/svengreb/wand/issues/89), and execute it afterwards. This is the current default behavior of this runner which will be used when [`WithCache(true)`][9] is used. - pass the task to a [`golang.Runner`][8], using the new [`golang/run`][3] package task, so that it can run `go run pkg@version <args>` instead. This is the new “smart“ behavior of the runner which will be used when [`WithCache(false)`][9] (default) is used. The **new default behavior will be to not use a local cache** so that caching will be a opt-in. This decision was made because native support for running commands _on-the-fly_ should always be preferred to custom logic which is what the local cache directory and [`gotool.Runner`][6] purpose is. > [!warning] Note that the minimum Go version for task runners, the new [`golang/run` task][3] and [the _Elder_ wand][7] will be increased to `1.17.0` since this version initially [introduced `go run` support in module-aware mode][1]! > This will be enforced through a [build constraint](https://pkg.go.dev/cmd/go#hdr-Build_constraints) (`go:build go1.17`). The [`Elder`][7] reference implementation will also adapt to this new feature by… 1. **deprecating the `*elder.Elder.Bootstrap(...string) []error` method**! As of _wand_ version `0.9.0` it will be a no-op and will be removed in version `0.10.0`. To install executables anyway the new `*elder.Elder.CacheExecutables error` method should be used instead. To ensure that the wand is properly initialized and operational the `*elder.Elder.Validate(..task.Runner) []error` method is the way to go. A warning message will be printed when the method is called to ensure that users adapt accordionally. 2. providing a new `*elder.Elder.CacheExecutables(...string) error` method which allows to pass paths of Go modules that should be explicitly installed to the local cache directory. This method is a kind of workaround for the, now deprecated, `*elder.Elder.Bootstrap(...string) []error` method to allows users to still cache command executables locally. 3. changing the signature of the `*elder.Elder.Validate() error` method to `*elder.Elder.Validate(...task.Runner) []error` method which allows users to ensure that the _wand_ is properly initialized and operational. Optionally [command runner][4] can be passed that will be validated while passing nothing will validate all currently supported runners. [1]: https://go.dev/doc/go1.17#go%20run [2]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#GoModule [3]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run [4]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#Runner [5]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#KindGoModule [6]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/gotool#Runner [7]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/elder [8]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang#Runner [9]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run#WithCache
True
`go run` support for versioned modules (Go 1.17+) - [As of Go 1.17 the `go run` command can finally run in module-aware mode][1] while not “polluting“ the current module in the working directory, of there is one (`go.mod` file present) 🎉 This finally allows to [run commands _on-the-fly_](https://pkg.go.dev/cmd/go#hdr-Compile_and_run_Go_program) of Go `main` module packages without installing them or without changing dependencies of the current module! To support this feature with _wand_ a new [`task.GoModule`][2] will be implemented in a new [`golang/run`][3] package. It can be run using a [command runner][4] that handles tasks of kind [`KindGoModule`][5] so mainly [`gotool.Runner`][6]. The new [`golang/run.Task`][3] will be customizable through the following functions: - `WithArgs(...string) run.Option` — sets additional arguments to pass to the command. - `WithEnv(map[string]string) run.Option` — sets the task specific environment. - `WithModulePath(string) run.Option` — sets the module import path. - `WithModuleVersion(*semver.Version) run.Option` — sets the module version. Next to the new task the [`gotool.Runner`][6] will be adjusted to a new [`WithCache(bool)`][9] runner option to toggle the usage of the local cache directory in the root directory of the module. The runner will be made “smart“ in the way that it either… - installing the executable through a [`golang.Runner`][8], which runs `go install pkg@version` to [leverage Go 1.16‘s feature](https://github.com/svengreb/wand/issues/89), and execute it afterwards. This is the current default behavior of this runner which will be used when [`WithCache(true)`][9] is used. - pass the task to a [`golang.Runner`][8], using the new [`golang/run`][3] package task, so that it can run `go run pkg@version <args>` instead. This is the new “smart“ behavior of the runner which will be used when [`WithCache(false)`][9] (default) is used. The **new default behavior will be to not use a local cache** so that caching will be a opt-in. This decision was made because native support for running commands _on-the-fly_ should always be preferred to custom logic which is what the local cache directory and [`gotool.Runner`][6] purpose is. > [!warning] Note that the minimum Go version for task runners, the new [`golang/run` task][3] and [the _Elder_ wand][7] will be increased to `1.17.0` since this version initially [introduced `go run` support in module-aware mode][1]! > This will be enforced through a [build constraint](https://pkg.go.dev/cmd/go#hdr-Build_constraints) (`go:build go1.17`). The [`Elder`][7] reference implementation will also adapt to this new feature by… 1. **deprecating the `*elder.Elder.Bootstrap(...string) []error` method**! As of _wand_ version `0.9.0` it will be a no-op and will be removed in version `0.10.0`. To install executables anyway the new `*elder.Elder.CacheExecutables error` method should be used instead. To ensure that the wand is properly initialized and operational the `*elder.Elder.Validate(..task.Runner) []error` method is the way to go. A warning message will be printed when the method is called to ensure that users adapt accordionally. 2. providing a new `*elder.Elder.CacheExecutables(...string) error` method which allows to pass paths of Go modules that should be explicitly installed to the local cache directory. This method is a kind of workaround for the, now deprecated, `*elder.Elder.Bootstrap(...string) []error` method to allows users to still cache command executables locally. 3. changing the signature of the `*elder.Elder.Validate() error` method to `*elder.Elder.Validate(...task.Runner) []error` method which allows users to ensure that the _wand_ is properly initialized and operational. Optionally [command runner][4] can be passed that will be validated while passing nothing will validate all currently supported runners. [1]: https://go.dev/doc/go1.17#go%20run [2]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#GoModule [3]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run [4]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#Runner [5]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#KindGoModule [6]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/gotool#Runner [7]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/elder [8]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang#Runner [9]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run#WithCache
main
go run support for versioned modules go while not “polluting“ the current module in the working directory of there is one go mod file present 🎉 this finally allows to of go main module packages without installing them or without changing dependencies of the current module to support this feature with wand a new will be implemented in a new package it can be run using a that handles tasks of kind so mainly the new will be customizable through the following functions withargs string run option — sets additional arguments to pass to the command withenv map string run option — sets the task specific environment withmodulepath string run option — sets the module import path withmoduleversion semver version run option — sets the module version next to the new task the will be adjusted to a new runner option to toggle the usage of the local cache directory in the root directory of the module the runner will be made “smart“ in the way that it either… installing the executable through a which runs go install pkg version to and execute it afterwards this is the current default behavior of this runner which will be used when is used pass the task to a using the new package task so that it can run go run pkg version instead this is the new “smart“ behavior of the runner which will be used when default is used the new default behavior will be to not use a local cache so that caching will be a opt in this decision was made because native support for running commands on the fly should always be preferred to custom logic which is what the local cache directory and purpose is note that the minimum go version for task runners the new and will be increased to since this version initially this will be enforced through a go build the reference implementation will also adapt to this new feature by… deprecating the elder elder bootstrap string error method as of wand version it will be a no op and will be removed in version to install executables anyway the new elder elder cacheexecutables error method should be used instead to ensure that the wand is properly initialized and operational the elder elder validate task runner error method is the way to go a warning message will be printed when the method is called to ensure that users adapt accordionally providing a new elder elder cacheexecutables string error method which allows to pass paths of go modules that should be explicitly installed to the local cache directory this method is a kind of workaround for the now deprecated elder elder bootstrap string error method to allows users to still cache command executables locally changing the signature of the elder elder validate error method to elder elder validate task runner error method which allows users to ensure that the wand is properly initialized and operational optionally can be passed that will be validated while passing nothing will validate all currently supported runners
1
53,725
7,853,237,627
IssuesEvent
2018-06-20 16:46:07
alexa/alexa-skills-kit-sdk-for-java
https://api.github.com/repos/alexa/alexa-skills-kit-sdk-for-java
closed
Detailed helloWorld documentation for beginners
documentation
<!-- PLEASE HELP US PROCESS GITHUB ISSUES FASTER BY PROVIDING THE FOLLOWING INFORMATION. ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> ## I'm submitting a... <!-- Check one of the following options with "x" --> <pre><code>[ ] Regression (a behavior that used to work and stopped working in a new release) [ ] Bug report <!-- Please search GitHub for a similar issue or PR before submitting --> [ ] Performance issue [ ] Feature request [X] Documentation issue or request [ ] Other... Please describe: </code></pre> <!--- Provide a general summary of the issue in the Title above --> It seems to me that a lot of the people I see at hackathons developing Alexa skills are able to do so using the "copy-paste" method from the numerous Alexa intro tutorials out there. However, when it came time for them to implement anything out of the ordinary, they didn't have a solid foundational knowledge of concepts like states and intents. I think there should be detailed documentation for at least one of the beginner tutorial programs we provide like the "HelloWorld" program in the ReadMe. Something that goes line by line and explains the concepts clearly as well as the most confusing points for a lot of developers. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> Having a handle on the very fundamentals of developing an Alexa skill and what the tutorial code actually means will help reduce the learning curve for first-time developers. This will also allow developers to create Alexa skills that demonstrate better grasp of the resources available in the Alexa SDK.
1.0
Detailed helloWorld documentation for beginners - <!-- PLEASE HELP US PROCESS GITHUB ISSUES FASTER BY PROVIDING THE FOLLOWING INFORMATION. ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> ## I'm submitting a... <!-- Check one of the following options with "x" --> <pre><code>[ ] Regression (a behavior that used to work and stopped working in a new release) [ ] Bug report <!-- Please search GitHub for a similar issue or PR before submitting --> [ ] Performance issue [ ] Feature request [X] Documentation issue or request [ ] Other... Please describe: </code></pre> <!--- Provide a general summary of the issue in the Title above --> It seems to me that a lot of the people I see at hackathons developing Alexa skills are able to do so using the "copy-paste" method from the numerous Alexa intro tutorials out there. However, when it came time for them to implement anything out of the ordinary, they didn't have a solid foundational knowledge of concepts like states and intents. I think there should be detailed documentation for at least one of the beginner tutorial programs we provide like the "HelloWorld" program in the ReadMe. Something that goes line by line and explains the concepts clearly as well as the most confusing points for a lot of developers. ## Context <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> Having a handle on the very fundamentals of developing an Alexa skill and what the tutorial code actually means will help reduce the learning curve for first-time developers. This will also allow developers to create Alexa skills that demonstrate better grasp of the resources available in the Alexa SDK.
non_main
detailed helloworld documentation for beginners please help us process github issues faster by providing the following information issues missing important information may be closed without investigation i m submitting a regression a behavior that used to work and stopped working in a new release bug report performance issue feature request documentation issue or request other please describe it seems to me that a lot of the people i see at hackathons developing alexa skills are able to do so using the copy paste method from the numerous alexa intro tutorials out there however when it came time for them to implement anything out of the ordinary they didn t have a solid foundational knowledge of concepts like states and intents i think there should be detailed documentation for at least one of the beginner tutorial programs we provide like the helloworld program in the readme something that goes line by line and explains the concepts clearly as well as the most confusing points for a lot of developers context having a handle on the very fundamentals of developing an alexa skill and what the tutorial code actually means will help reduce the learning curve for first time developers this will also allow developers to create alexa skills that demonstrate better grasp of the resources available in the alexa sdk
0
2,827
10,136,146,996
IssuesEvent
2019-08-02 12:13:21
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
opened
Migrate to JavaPackager
mac maintainability windows
We currently use Launch4j and Appbundler to package OpenRefine for Windows and Mac. Appbundler is no longer supported by Oracle and they recommend JavaPackager instead. JavaPackager handles a lot of distribution formats and is also integrated with Maven via this plugin: https://github.com/fvarrui/JavaPackager We should consider migrating to that.
True
Migrate to JavaPackager - We currently use Launch4j and Appbundler to package OpenRefine for Windows and Mac. Appbundler is no longer supported by Oracle and they recommend JavaPackager instead. JavaPackager handles a lot of distribution formats and is also integrated with Maven via this plugin: https://github.com/fvarrui/JavaPackager We should consider migrating to that.
main
migrate to javapackager we currently use and appbundler to package openrefine for windows and mac appbundler is no longer supported by oracle and they recommend javapackager instead javapackager handles a lot of distribution formats and is also integrated with maven via this plugin we should consider migrating to that
1
3,329
12,908,335,216
IssuesEvent
2020-07-15 07:13:31
cncf/sig-contributor-strategy
https://api.github.com/repos/cncf/sig-contributor-strategy
closed
Draft Maintainers Circle Welcome Email
wg/maintainer
- [x] create a welcome email to send to the maintainer lists including an action (example: join the maintainer slack channel) (merged with #19) - [ ] Try and recruit at least one voice per project (this is optional but will be good to have someone/small group to keep in constant contact about contributor related matters) and keep a record somewhere TBD (need clear ownership on this) - [ ] think about what we would want the first meeting to look like or whatever it is that we are delivering. whats a good valuable first step? (need clear ownership on this and discussion)
True
Draft Maintainers Circle Welcome Email - - [x] create a welcome email to send to the maintainer lists including an action (example: join the maintainer slack channel) (merged with #19) - [ ] Try and recruit at least one voice per project (this is optional but will be good to have someone/small group to keep in constant contact about contributor related matters) and keep a record somewhere TBD (need clear ownership on this) - [ ] think about what we would want the first meeting to look like or whatever it is that we are delivering. whats a good valuable first step? (need clear ownership on this and discussion)
main
draft maintainers circle welcome email create a welcome email to send to the maintainer lists including an action example join the maintainer slack channel merged with try and recruit at least one voice per project this is optional but will be good to have someone small group to keep in constant contact about contributor related matters and keep a record somewhere tbd need clear ownership on this think about what we would want the first meeting to look like or whatever it is that we are delivering whats a good valuable first step need clear ownership on this and discussion
1
414,459
12,103,654,692
IssuesEvent
2020-04-20 18:47:25
treehouses/builder
https://api.github.com/repos/treehouses/builder
closed
/etc/systemd/system/dbus-org.bluez.service needs to have -C on first/second boot
priority
```bash BusName=org.bluez ExecStart=/usr/lib/bluetooth/bluetoothd NotifyAccess=main #WatchdogSec=10 #Restart=on-failure ``` This bluetooth service needs to be the same as `templates/bluetooth/hotspot` which correlates to `treehouses bluetooth on`. e.g. `ExecStart=/usr/lib/bluetooth/bluetoothd -C`
1.0
/etc/systemd/system/dbus-org.bluez.service needs to have -C on first/second boot - ```bash BusName=org.bluez ExecStart=/usr/lib/bluetooth/bluetoothd NotifyAccess=main #WatchdogSec=10 #Restart=on-failure ``` This bluetooth service needs to be the same as `templates/bluetooth/hotspot` which correlates to `treehouses bluetooth on`. e.g. `ExecStart=/usr/lib/bluetooth/bluetoothd -C`
non_main
etc systemd system dbus org bluez service needs to have c on first second boot bash busname org bluez execstart usr lib bluetooth bluetoothd notifyaccess main watchdogsec restart on failure this bluetooth service needs to be the same as templates bluetooth hotspot which correlates to treehouses bluetooth on e g execstart usr lib bluetooth bluetoothd c
0
2,725
9,631,088,406
IssuesEvent
2019-05-15 13:32:51
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Migrate out of Codacy
maintainability
Codacy has multiple issues: - their assessment of pull requests is not reliable as their detection of new and closed issues tends to issues that are unrelated to the changes - the assessment takes quite a long time: codacy-bot generally comments on pull requests a long time after the PR is opened, which generates unwelcome notifications. - I haven't managed to turn off the spurious issues that are not useful to us - code coverage is not taken into account in the PR assessment, and is poorly displayed in the codacy UI (if at all) - they tend to have a fairly aggressive marketing and send out a lot of emails, the last one signalling me that they would stop OpenRefine's analysis if I don't log in to Codacy in a few days. I propose that we migrate to a different service, focused on test coverage, such as Coveralls (which I know works well from other projects) or Codecov (which has an interesting diff coverage view).
True
Migrate out of Codacy - Codacy has multiple issues: - their assessment of pull requests is not reliable as their detection of new and closed issues tends to issues that are unrelated to the changes - the assessment takes quite a long time: codacy-bot generally comments on pull requests a long time after the PR is opened, which generates unwelcome notifications. - I haven't managed to turn off the spurious issues that are not useful to us - code coverage is not taken into account in the PR assessment, and is poorly displayed in the codacy UI (if at all) - they tend to have a fairly aggressive marketing and send out a lot of emails, the last one signalling me that they would stop OpenRefine's analysis if I don't log in to Codacy in a few days. I propose that we migrate to a different service, focused on test coverage, such as Coveralls (which I know works well from other projects) or Codecov (which has an interesting diff coverage view).
main
migrate out of codacy codacy has multiple issues their assessment of pull requests is not reliable as their detection of new and closed issues tends to issues that are unrelated to the changes the assessment takes quite a long time codacy bot generally comments on pull requests a long time after the pr is opened which generates unwelcome notifications i haven t managed to turn off the spurious issues that are not useful to us code coverage is not taken into account in the pr assessment and is poorly displayed in the codacy ui if at all they tend to have a fairly aggressive marketing and send out a lot of emails the last one signalling me that they would stop openrefine s analysis if i don t log in to codacy in a few days i propose that we migrate to a different service focused on test coverage such as coveralls which i know works well from other projects or codecov which has an interesting diff coverage view
1
216,354
7,307,128,663
IssuesEvent
2018-02-28 01:08:07
Mechanical-Advantage/RobotCode2018
https://api.github.com/repos/Mechanical-Advantage/RobotCode2018
opened
Test behavior of autonomous when robot encounters immovable object
Priority 1
Make sure that robot autonomous does not get stuck if it hits an immovable object and the wheels won't turn. May require simulating this condition on Everybot since behavior of Revenge's shifting gearbox varies depending on low/high gear setting.
1.0
Test behavior of autonomous when robot encounters immovable object - Make sure that robot autonomous does not get stuck if it hits an immovable object and the wheels won't turn. May require simulating this condition on Everybot since behavior of Revenge's shifting gearbox varies depending on low/high gear setting.
non_main
test behavior of autonomous when robot encounters immovable object make sure that robot autonomous does not get stuck if it hits an immovable object and the wheels won t turn may require simulating this condition on everybot since behavior of revenge s shifting gearbox varies depending on low high gear setting
0