Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
1,463
6,363,107,670
IssuesEvent
2017-07-31 16:17:29
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Conversions: converting cm to feet and inches
Maintainer Approved
When trying to convert centimeters to Imperial units, it does not show inches - only feet. Example: "200 cm = 6.562 feet" instead of the desired feet _and_ inches --- IA Page: http://duck.co/ia/view/conversions [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft
True
Conversions: converting cm to feet and inches - When trying to convert centimeters to Imperial units, it does not show inches - only feet. Example: "200 cm = 6.562 feet" instead of the desired feet _and_ inches --- IA Page: http://duck.co/ia/view/conversions [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft
main
conversions converting cm to feet and inches when trying to convert centimeters to imperial units it does not show inches only feet example cm feet instead of the desired feet and inches ia page mintsoft
1
160,821
6,103,309,800
IssuesEvent
2017-06-20 18:25:47
PredictiveEcology/SpaDES
https://api.github.com/repos/PredictiveEcology/SpaDES
closed
error when module has empty reqdPkgs field
Low priority resolved in branch
The easy workaround is to put something in the field (*e.g.*, `SpaDES`), but we should allow empty package lists in the module metadata.
1.0
error when module has empty reqdPkgs field - The easy workaround is to put something in the field (*e.g.*, `SpaDES`), but we should allow empty package lists in the module metadata.
non_main
error when module has empty reqdpkgs field the easy workaround is to put something in the field e g spades but we should allow empty package lists in the module metadata
0
3,279
12,509,734,510
IssuesEvent
2020-06-02 17:26:36
numbbo/coco
https://api.github.com/repos/numbbo/coco
closed
Naming of datasets
Maintainability Usability
```python import cocopp cocopp.archives.all.find(".*bob/.*BFGS") ``` ``` ['bbob/2009/BFGS_ros_noiseless.tgz', 'bbob/2012/DE-BFGS_voglis_noiseless.tgz', 'bbob/2012/PSO-BFGS_voglis_noiseless.tgz', 'bbob/2014-others/BFGS-scipy-Baudis.tgz', 'bbob/2014-others/L-BFGS-B-scipy-Baudis.tgz', 'bbob/2018/BFGS-M-17.tgz', 'bbob/2018/BFGS-P-09.tgz', 'bbob/2018/BFGS-P-Instances.tgz', 'bbob/2018/BFGS-P-range.tgz', 'bbob/2018/BFGS-P-StPt.tgz'] ``` I think we should get back to use the original naming convention which gives algorithm name _and_ author of the data set.
True
Naming of datasets - ```python import cocopp cocopp.archives.all.find(".*bob/.*BFGS") ``` ``` ['bbob/2009/BFGS_ros_noiseless.tgz', 'bbob/2012/DE-BFGS_voglis_noiseless.tgz', 'bbob/2012/PSO-BFGS_voglis_noiseless.tgz', 'bbob/2014-others/BFGS-scipy-Baudis.tgz', 'bbob/2014-others/L-BFGS-B-scipy-Baudis.tgz', 'bbob/2018/BFGS-M-17.tgz', 'bbob/2018/BFGS-P-09.tgz', 'bbob/2018/BFGS-P-Instances.tgz', 'bbob/2018/BFGS-P-range.tgz', 'bbob/2018/BFGS-P-StPt.tgz'] ``` I think we should get back to use the original naming convention which gives algorithm name _and_ author of the data set.
main
naming of datasets python import cocopp cocopp archives all find bob bfgs bbob bfgs ros noiseless tgz bbob de bfgs voglis noiseless tgz bbob pso bfgs voglis noiseless tgz bbob others bfgs scipy baudis tgz bbob others l bfgs b scipy baudis tgz bbob bfgs m tgz bbob bfgs p tgz bbob bfgs p instances tgz bbob bfgs p range tgz bbob bfgs p stpt tgz i think we should get back to use the original naming convention which gives algorithm name and author of the data set
1
3,447
13,212,224,579
IssuesEvent
2020-08-16 05:31:52
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Add support for outputs only in terraform module
affects_2.9 bot_closed cloud collection collection:community.general feature has_pr module needs_collection_redirect needs_maintainer needs_triage support:community
<!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Allow to gather outputs only from terraform module. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> terraform ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> Sometimes we just want to gather information from Terraform, without need to run a full "plan". Should be nice to have this shortcut. <!--- Paste example playbooks or commands between quotes below --> ```yaml terraform: state: outputs project_path: '{{ project_dir }}' register: terraform ``` <!--- HINT: You can also paste gist.github.com links for larger files -->
True
Add support for outputs only in terraform module - <!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Allow to gather outputs only from terraform module. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> terraform ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> Sometimes we just want to gather information from Terraform, without need to run a full "plan". Should be nice to have this shortcut. <!--- Paste example playbooks or commands between quotes below --> ```yaml terraform: state: outputs project_path: '{{ project_dir }}' register: terraform ``` <!--- HINT: You can also paste gist.github.com links for larger files -->
main
add support for outputs only in terraform module summary allow to gather outputs only from terraform module issue type feature idea component name terraform additional information sometimes we just want to gather information from terraform without need to run a full plan should be nice to have this shortcut yaml terraform state outputs project path project dir register terraform
1
38,755
15,788,924,656
IssuesEvent
2021-04-01 21:39:06
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
opened
Feature Request: Support for S3 Batch Operations
new-resource service/s3control
<!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a πŸ‘ [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * aws_s3control_job ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "aws_s3control_job" "test" { operation { lambda_invoke { function_arn = "" } } report { bucket = "" enabled = true format = "" prefix = "" report_scope = "" } client_request_token = "" manifest { location { etag = "" object_arn = "" object_version_id = "" } spec { fields = [] format = "" } description = "" priority = "" role_arn "" tags = {} } } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/ ---> - https://aws.amazon.com/blogs/aws/new-amazon-s3-batch-operations/ - https://docs.aws.amazon.com/AmazonS3/latest/API/API_control_CreateJob.html
1.0
Feature Request: Support for S3 Batch Operations - <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a πŸ‘ [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Description <!--- Please leave a helpful description of the feature request here. ---> ### New or Affected Resource(s) <!--- Please list the new or affected resources and data sources. ---> * aws_s3control_job ### Potential Terraform Configuration <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "aws_s3control_job" "test" { operation { lambda_invoke { function_arn = "" } } report { bucket = "" enabled = true format = "" prefix = "" report_scope = "" } client_request_token = "" manifest { location { etag = "" object_arn = "" object_version_id = "" } spec { fields = [] format = "" } description = "" priority = "" role_arn "" tags = {} } } ``` ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example: * https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/ ---> - https://aws.amazon.com/blogs/aws/new-amazon-s3-batch-operations/ - https://docs.aws.amazon.com/AmazonS3/latest/API/API_control_CreateJob.html
non_main
feature request support for batch operations community note please vote on this issue by adding a πŸ‘ to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description new or affected resource s aws job potential terraform configuration hcl resource aws job test operation lambda invoke function arn report bucket enabled true format prefix report scope client request token manifest location etag object arn object version id spec fields format description priority role arn tags references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example
0
3,252
12,402,316,182
IssuesEvent
2020-05-21 11:43:28
ocaml/opam-repository
https://api.github.com/repos/ocaml/opam-repository
closed
ocaml-inifiles - missing cmx file
Stale needs maintainer action no maintainer
Since opam package doesn't refer to any real source repository or a maintainer, filling issue here: ``` File "_none_", line 1: Warning 58: no cmx file was found in path for module Inifiles, and its interface was not compiled with -opaque ```
True
ocaml-inifiles - missing cmx file - Since opam package doesn't refer to any real source repository or a maintainer, filling issue here: ``` File "_none_", line 1: Warning 58: no cmx file was found in path for module Inifiles, and its interface was not compiled with -opaque ```
main
ocaml inifiles missing cmx file since opam package doesn t refer to any real source repository or a maintainer filling issue here file none line warning no cmx file was found in path for module inifiles and its interface was not compiled with opaque
1
11,495
2,652,259,196
IssuesEvent
2015-03-16 16:23:58
JoseExposito/touchegg
https://api.github.com/repos/JoseExposito/touchegg
closed
Touchegg doesn't work on Probook 4540s
auto-migrated Type-Defect
``` So, Im using a hp probook 4540s on archlinux and I cannot make touchegg work, I've described my problem on Arch forum here :https://bbs.archlinux.org/viewtopic.php?id=167419 Any chance that I can know how can I make touchegg recognize the gestures ? But I think my situation is quite popular so maybe I left out something at the configuration. If anyone could tell me a brief guide of what I have to do since the installation, that would be appreciated. Thanks in advance :D ``` Original issue reported on code.google.com by `elvenson...@gmail.com` on 30 Jul 2013 at 11:56
1.0
Touchegg doesn't work on Probook 4540s - ``` So, Im using a hp probook 4540s on archlinux and I cannot make touchegg work, I've described my problem on Arch forum here :https://bbs.archlinux.org/viewtopic.php?id=167419 Any chance that I can know how can I make touchegg recognize the gestures ? But I think my situation is quite popular so maybe I left out something at the configuration. If anyone could tell me a brief guide of what I have to do since the installation, that would be appreciated. Thanks in advance :D ``` Original issue reported on code.google.com by `elvenson...@gmail.com` on 30 Jul 2013 at 11:56
non_main
touchegg doesn t work on probook so im using a hp probook on archlinux and i cannot make touchegg work i ve described my problem on arch forum here any chance that i can know how can i make touchegg recognize the gestures but i think my situation is quite popular so maybe i left out something at the configuration if anyone could tell me a brief guide of what i have to do since the installation that would be appreciated thanks in advance d original issue reported on code google com by elvenson gmail com on jul at
0
168
2,716,257,169
IssuesEvent
2015-04-10 17:58:35
Kapeli/Dash-User-Contributions
https://api.github.com/repos/Kapeli/Dash-User-Contributions
closed
Atom docset outdated
needs maintainer
I'm the contributor of the Atom.io docset, it's quite outdated now and I don't have the time to keep up with the changes. Perhaps it might be best to remove this docset for now and I'll take another look at it once they reach a 1.0 version and hopefully the changes to the API slow down a little. Also happy for anyone else to take over this if they have time/interest.
True
Atom docset outdated - I'm the contributor of the Atom.io docset, it's quite outdated now and I don't have the time to keep up with the changes. Perhaps it might be best to remove this docset for now and I'll take another look at it once they reach a 1.0 version and hopefully the changes to the API slow down a little. Also happy for anyone else to take over this if they have time/interest.
main
atom docset outdated i m the contributor of the atom io docset it s quite outdated now and i don t have the time to keep up with the changes perhaps it might be best to remove this docset for now and i ll take another look at it once they reach a version and hopefully the changes to the api slow down a little also happy for anyone else to take over this if they have time interest
1
3,815
16,587,641,377
IssuesEvent
2021-06-01 00:40:33
microsoft/DirectXTK
https://api.github.com/repos/microsoft/DirectXTK
closed
CMakeLists.txt needs updated to support XAudio2Redist
audio maintainence
To support Windows 7, *DirectX Tool Kit for Audio* needs to use the ``Microsoft.XAudio2.Redist`` NuGet. To date, the CMakeLists.txt only supported building for XAudio 2.9 with Windows 10 and XAudio 2.8 with Windows 8.x. This package is now supported by [vcpkg](https://github.com/microsoft/vcpkg/pull/17174/files).
True
CMakeLists.txt needs updated to support XAudio2Redist - To support Windows 7, *DirectX Tool Kit for Audio* needs to use the ``Microsoft.XAudio2.Redist`` NuGet. To date, the CMakeLists.txt only supported building for XAudio 2.9 with Windows 10 and XAudio 2.8 with Windows 8.x. This package is now supported by [vcpkg](https://github.com/microsoft/vcpkg/pull/17174/files).
main
cmakelists txt needs updated to support to support windows directx tool kit for audio needs to use the microsoft redist nuget to date the cmakelists txt only supported building for xaudio with windows and xaudio with windows x this package is now supported by
1
71,741
23,780,464,905
IssuesEvent
2022-09-02 03:47:55
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
An error occurred during software installation
T: defect P: normal
(moose) zhangyu@AZY:/mnt/c/For_MOOSE/projects/moose/test$ make -j 4.3 Command 'make' not found, but can be installed with: sudo apt install make # version 4.3-4.1build1, or sudo apt install make-guile # version 4.3-4.1build1 (moose) zhangyu@AZY:/mnt/c/For_MOOSE/projects/moose/test$ ![Snipaste_2022-09-02_11-44-52](https://user-images.githubusercontent.com/53269552/188055230-7b2fea7c-cfa1-43ba-9683-8e3e463a755f.png)
1.0
An error occurred during software installation - (moose) zhangyu@AZY:/mnt/c/For_MOOSE/projects/moose/test$ make -j 4.3 Command 'make' not found, but can be installed with: sudo apt install make # version 4.3-4.1build1, or sudo apt install make-guile # version 4.3-4.1build1 (moose) zhangyu@AZY:/mnt/c/For_MOOSE/projects/moose/test$ ![Snipaste_2022-09-02_11-44-52](https://user-images.githubusercontent.com/53269552/188055230-7b2fea7c-cfa1-43ba-9683-8e3e463a755f.png)
non_main
an error occurred during software installation moose zhangyu azy mnt c for moose projects moose test make j command make not found but can be installed with sudo apt install make version or sudo apt install make guile version moose zhangyu azy mnt c for moose projects moose test
0
4,587
23,817,644,721
IssuesEvent
2022-09-05 08:21:15
tgbot-collection/ytdlbot
https://api.github.com/repos/tgbot-collection/ytdlbot
closed
Upload downloaded file before downloading entire playlist and private video bug
enhancement not-maintained
First off, thanks for this amazing bot. my issues: 1) when downloading a playlist contating for example 50 videos, currently bot downloads all 50 videos first, then starts uploading one by one. If possible, Instead if it is made to upload each video first before beginning next download of second video, space consumption in local drive is minimised and will bw hwlpful for those with limited drive space (like me) 2) If there is a private or deleted video in a playlist, all the downloads stop. instead if there is an ability to skip those videos and download and upload others, it will be more useful. Thanks.
True
Upload downloaded file before downloading entire playlist and private video bug - First off, thanks for this amazing bot. my issues: 1) when downloading a playlist contating for example 50 videos, currently bot downloads all 50 videos first, then starts uploading one by one. If possible, Instead if it is made to upload each video first before beginning next download of second video, space consumption in local drive is minimised and will bw hwlpful for those with limited drive space (like me) 2) If there is a private or deleted video in a playlist, all the downloads stop. instead if there is an ability to skip those videos and download and upload others, it will be more useful. Thanks.
main
upload downloaded file before downloading entire playlist and private video bug first off thanks for this amazing bot my issues when downloading a playlist contating for example videos currently bot downloads all videos first then starts uploading one by one if possible instead if it is made to upload each video first before beginning next download of second video space consumption in local drive is minimised and will bw hwlpful for those with limited drive space like me if there is a private or deleted video in a playlist all the downloads stop instead if there is an ability to skip those videos and download and upload others it will be more useful thanks
1
4,131
19,596,997,385
IssuesEvent
2022-01-05 19:08:30
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
Progress Indicator Animations
type: enhancement πŸ’‘ role: ux 🍿 role: visual 🎨 proposal: open status: waiting for maintainer response πŸ’¬
### Summary Requesting to enhance the progress indicator component with animations showing the blue line progressing as steps are completed and the icons change to help the user better understand their progress as they complete forms. Also, requesting a slight change to the "optional text" under each text to instead a show a summary of what the user selected in that step. This would require motion design and development work. ### Justification This would reduce the cognitive load of the user and help them better make connections between the steps. The addition of the "summary text" could also provide a "summary" step for our users - allowing them to review progress made without going back and taking additional steps. ### Desired UX and success metrics <!--alex disable failure--> When a user completes a step, the progress line would animate to show their progress. At the same time, the summary text would fade in as well as the blue step indicator would transform to the check mark. Here's a mock up of how this could work: ![Horizontal Progress Tracker](https://user-images.githubusercontent.com/73298410/125118741-bed0f600-e0b5-11eb-9807-f0cdd2b52e94.gif) ![Progress Tracker 01](https://user-images.githubusercontent.com/73298410/125118831-df994b80-e0b5-11eb-93d0-294be96f5dd8.gif) Success/failure could be measured through user testing the current implementation of this component with the proposed changes OR how current forms work without vs with and user's preferences. <!--alex enable failure--> ### "Must have" functionality - Blue line/icons animating Some additional functionality to consider: - Creating anchor links in the indicator so users can easily revisit steps - Showing error states in the progress indicator steps when users try to complete the form or miss a section ### Specific timeline issues / requests <!--alex disable period--> no specific time line <!--alex enable period--> ### Available extra resources What resources do you have to assist this effort? I (@lisa-j-kaiser) am happy to help with any design/animation and @kristastarr (krista.starr@ibm.com) is working on development of this component to contribute back.
True
Progress Indicator Animations - ### Summary Requesting to enhance the progress indicator component with animations showing the blue line progressing as steps are completed and the icons change to help the user better understand their progress as they complete forms. Also, requesting a slight change to the "optional text" under each text to instead a show a summary of what the user selected in that step. This would require motion design and development work. ### Justification This would reduce the cognitive load of the user and help them better make connections between the steps. The addition of the "summary text" could also provide a "summary" step for our users - allowing them to review progress made without going back and taking additional steps. ### Desired UX and success metrics <!--alex disable failure--> When a user completes a step, the progress line would animate to show their progress. At the same time, the summary text would fade in as well as the blue step indicator would transform to the check mark. Here's a mock up of how this could work: ![Horizontal Progress Tracker](https://user-images.githubusercontent.com/73298410/125118741-bed0f600-e0b5-11eb-9807-f0cdd2b52e94.gif) ![Progress Tracker 01](https://user-images.githubusercontent.com/73298410/125118831-df994b80-e0b5-11eb-93d0-294be96f5dd8.gif) Success/failure could be measured through user testing the current implementation of this component with the proposed changes OR how current forms work without vs with and user's preferences. <!--alex enable failure--> ### "Must have" functionality - Blue line/icons animating Some additional functionality to consider: - Creating anchor links in the indicator so users can easily revisit steps - Showing error states in the progress indicator steps when users try to complete the form or miss a section ### Specific timeline issues / requests <!--alex disable period--> no specific time line <!--alex enable period--> ### Available extra resources What resources do you have to assist this effort? I (@lisa-j-kaiser) am happy to help with any design/animation and @kristastarr (krista.starr@ibm.com) is working on development of this component to contribute back.
main
progress indicator animations summary requesting to enhance the progress indicator component with animations showing the blue line progressing as steps are completed and the icons change to help the user better understand their progress as they complete forms also requesting a slight change to the optional text under each text to instead a show a summary of what the user selected in that step this would require motion design and development work justification this would reduce the cognitive load of the user and help them better make connections between the steps the addition of the summary text could also provide a summary step for our users allowing them to review progress made without going back and taking additional steps desired ux and success metrics when a user completes a step the progress line would animate to show their progress at the same time the summary text would fade in as well as the blue step indicator would transform to the check mark here s a mock up of how this could work success failure could be measured through user testing the current implementation of this component with the proposed changes or how current forms work without vs with and user s preferences must have functionality blue line icons animating some additional functionality to consider creating anchor links in the indicator so users can easily revisit steps showing error states in the progress indicator steps when users try to complete the form or miss a section specific timeline issues requests no specific time line available extra resources what resources do you have to assist this effort i lisa j kaiser am happy to help with any design animation and kristastarr krista starr ibm com is working on development of this component to contribute back
1
4,578
23,776,534,288
IssuesEvent
2022-09-01 21:35:39
SethMMorton/natsort
https://api.github.com/repos/SethMMorton/natsort
closed
Loosen types and type checking
bug maintainability
When I added typing stubs to `natsort`, I went ALL OUT. I over-constrained the system. I need to pull back a bit. I observed this when I tried to use `natsort` in another project and it failed type checking when it shouldn't have.
True
Loosen types and type checking - When I added typing stubs to `natsort`, I went ALL OUT. I over-constrained the system. I need to pull back a bit. I observed this when I tried to use `natsort` in another project and it failed type checking when it shouldn't have.
main
loosen types and type checking when i added typing stubs to natsort i went all out i over constrained the system i need to pull back a bit i observed this when i tried to use natsort in another project and it failed type checking when it shouldn t have
1
5,875
31,886,061,852
IssuesEvent
2023-09-17 00:20:40
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Heretic Sacrifice template should use the lazyloading template system
Maintainability/Hinders improvements
## Reproduction: Free refactor GBP: The way heretics load the sacrifice z level (`/datum/map_template/heretic_sacrifice_level`) was conceived before lazyloading templates was added, effectively making it a proto-lazyloaded template that doesn't use the system. It should be refactored to use the system and loaded with the dynamic ruleset execution rather than now, which ties it into the sacrifice knowledge being "researched".
True
Heretic Sacrifice template should use the lazyloading template system - ## Reproduction: Free refactor GBP: The way heretics load the sacrifice z level (`/datum/map_template/heretic_sacrifice_level`) was conceived before lazyloading templates was added, effectively making it a proto-lazyloaded template that doesn't use the system. It should be refactored to use the system and loaded with the dynamic ruleset execution rather than now, which ties it into the sacrifice knowledge being "researched".
main
heretic sacrifice template should use the lazyloading template system reproduction free refactor gbp the way heretics load the sacrifice z level datum map template heretic sacrifice level was conceived before lazyloading templates was added effectively making it a proto lazyloaded template that doesn t use the system it should be refactored to use the system and loaded with the dynamic ruleset execution rather than now which ties it into the sacrifice knowledge being researched
1
231,120
25,490,399,895
IssuesEvent
2022-11-27 01:09:09
NirShaharabani/Vital
https://api.github.com/repos/NirShaharabani/Vital
opened
CVE-2018-3717 (Medium) detected in connect-2.6.0.tgz
security vulnerability
## CVE-2018-3717 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>connect-2.6.0.tgz</b></p></summary> <p>High performance middleware framework</p> <p>Library home page: <a href="https://registry.npmjs.org/connect/-/connect-2.6.0.tgz">https://registry.npmjs.org/connect/-/connect-2.6.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/connect/package.json</p> <p> Dependency Hierarchy: - express-3.0.0.tgz (Root Library) - :x: **connect-2.6.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> connect node module before 2.14.0 suffers from a Cross-Site Scripting (XSS) vulnerability due to a lack of validation of file in directory.js middleware. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-3717>CVE-2018-3717</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3717">https://nvd.nist.gov/vuln/detail/CVE-2018-3717</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution (connect): 2.14.0</p> <p>Direct dependency fix Resolution (express): 3.5.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2018-3717 (Medium) detected in connect-2.6.0.tgz - ## CVE-2018-3717 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>connect-2.6.0.tgz</b></p></summary> <p>High performance middleware framework</p> <p>Library home page: <a href="https://registry.npmjs.org/connect/-/connect-2.6.0.tgz">https://registry.npmjs.org/connect/-/connect-2.6.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/connect/package.json</p> <p> Dependency Hierarchy: - express-3.0.0.tgz (Root Library) - :x: **connect-2.6.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> connect node module before 2.14.0 suffers from a Cross-Site Scripting (XSS) vulnerability due to a lack of validation of file in directory.js middleware. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-3717>CVE-2018-3717</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3717">https://nvd.nist.gov/vuln/detail/CVE-2018-3717</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution (connect): 2.14.0</p> <p>Direct dependency fix Resolution (express): 3.5.0</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_main
cve medium detected in connect tgz cve medium severity vulnerability vulnerable library connect tgz high performance middleware framework library home page a href path to dependency file package json path to vulnerable library node modules connect package json dependency hierarchy express tgz root library x connect tgz vulnerable library found in base branch main vulnerability details connect node module before suffers from a cross site scripting xss vulnerability due to a lack of validation of file in directory js middleware publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution connect direct dependency fix resolution express check this box to open an automated fix pr
0
17
2,515,203,608
IssuesEvent
2015-01-15 17:03:57
simplesamlphp/simplesamlphp
https://api.github.com/repos/simplesamlphp/simplesamlphp
opened
Move the SimpleSAML_Auth_LDAP class
enhancement low maintainability
This is actually more a collection of LDAP-related utilities, so move it to `lib/SimpleSAML/Utils/LDAP.php`. This of course needs refactoring everything using this class.
True
Move the SimpleSAML_Auth_LDAP class - This is actually more a collection of LDAP-related utilities, so move it to `lib/SimpleSAML/Utils/LDAP.php`. This of course needs refactoring everything using this class.
main
move the simplesaml auth ldap class this is actually more a collection of ldap related utilities so move it to lib simplesaml utils ldap php this of course needs refactoring everything using this class
1
210,573
23,754,875,373
IssuesEvent
2022-09-01 01:24:15
SelvaMiru/WebGoat
https://api.github.com/repos/SelvaMiru/WebGoat
opened
CVE-2022-36033 (Medium) detected in jsoup-1.14.2.jar
security vulnerability
## CVE-2022-36033 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsoup-1.14.2.jar</b></p></summary> <p>jsoup is a Java library for working with real-world HTML. It provides a very convenient API for fetching URLs and extracting and manipulating data, using the best of HTML5 DOM methods and CSS selectors. jsoup implements the WHATWG HTML5 specification, and parses HTML to the same DOM as modern browsers do.</p> <p>Library home page: <a href="https://jsoup.org/">https://jsoup.org/</a></p> <p>Path to dependency file: /webgoat-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.14.2/jsoup-1.14.2.jar,/m2/repository/org/jsoup/jsoup/1.14.2/jsoup-1.14.2.jar</p> <p> Dependency Hierarchy: - :x: **jsoup-1.14.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SelvaMiru/WebGoat/commit/5b9ec46378504fb88d70e4275b259a70c871daae">5b9ec46378504fb88d70e4275b259a70c871daae</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jsoup is a Java HTML parser, built for HTML editing, cleaning, scraping, and cross-site scripting (XSS) safety. jsoup may incorrectly sanitize HTML including `javascript:` URL expressions, which could allow XSS attacks when a reader subsequently clicks that link. If the non-default `SafeList.preserveRelativeLinks` option is enabled, HTML including `javascript:` URLs that have been crafted with control characters will not be sanitized. If the site that this HTML is published on does not set a Content Security Policy, an XSS attack is then possible. This issue is patched in jsoup 1.15.3. Users should upgrade to this version. Additionally, as the unsanitized input may have been persisted, old content should be cleaned again using the updated version. To remediate this issue without immediately upgrading: - disable `SafeList.preserveRelativeLinks`, which will rewrite input URLs as absolute URLs - ensure an appropriate [Content Security Policy](https://developer.mozilla.org/en-US/docs/Web/HTTP/CSP) is defined. (This should be used regardless of upgrading, as a defence-in-depth best practice.) <p>Publish Date: 2022-08-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36033>CVE-2022-36033</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369">https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369</a></p> <p>Release Date: 2022-08-29</p> <p>Fix Resolution: 1.15.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-36033 (Medium) detected in jsoup-1.14.2.jar - ## CVE-2022-36033 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsoup-1.14.2.jar</b></p></summary> <p>jsoup is a Java library for working with real-world HTML. It provides a very convenient API for fetching URLs and extracting and manipulating data, using the best of HTML5 DOM methods and CSS selectors. jsoup implements the WHATWG HTML5 specification, and parses HTML to the same DOM as modern browsers do.</p> <p>Library home page: <a href="https://jsoup.org/">https://jsoup.org/</a></p> <p>Path to dependency file: /webgoat-server/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/jsoup/jsoup/1.14.2/jsoup-1.14.2.jar,/m2/repository/org/jsoup/jsoup/1.14.2/jsoup-1.14.2.jar</p> <p> Dependency Hierarchy: - :x: **jsoup-1.14.2.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SelvaMiru/WebGoat/commit/5b9ec46378504fb88d70e4275b259a70c871daae">5b9ec46378504fb88d70e4275b259a70c871daae</a></p> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jsoup is a Java HTML parser, built for HTML editing, cleaning, scraping, and cross-site scripting (XSS) safety. jsoup may incorrectly sanitize HTML including `javascript:` URL expressions, which could allow XSS attacks when a reader subsequently clicks that link. If the non-default `SafeList.preserveRelativeLinks` option is enabled, HTML including `javascript:` URLs that have been crafted with control characters will not be sanitized. If the site that this HTML is published on does not set a Content Security Policy, an XSS attack is then possible. This issue is patched in jsoup 1.15.3. Users should upgrade to this version. Additionally, as the unsanitized input may have been persisted, old content should be cleaned again using the updated version. To remediate this issue without immediately upgrading: - disable `SafeList.preserveRelativeLinks`, which will rewrite input URLs as absolute URLs - ensure an appropriate [Content Security Policy](https://developer.mozilla.org/en-US/docs/Web/HTTP/CSP) is defined. (This should be used regardless of upgrading, as a defence-in-depth best practice.) <p>Publish Date: 2022-08-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-36033>CVE-2022-36033</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369">https://github.com/jhy/jsoup/security/advisories/GHSA-gp7f-rwcx-9369</a></p> <p>Release Date: 2022-08-29</p> <p>Fix Resolution: 1.15.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in jsoup jar cve medium severity vulnerability vulnerable library jsoup jar jsoup is a java library for working with real world html it provides a very convenient api for fetching urls and extracting and manipulating data using the best of dom methods and css selectors jsoup implements the whatwg specification and parses html to the same dom as modern browsers do library home page a href path to dependency file webgoat server pom xml path to vulnerable library home wss scanner repository org jsoup jsoup jsoup jar repository org jsoup jsoup jsoup jar dependency hierarchy x jsoup jar vulnerable library found in head commit a href found in base branch develop vulnerability details jsoup is a java html parser built for html editing cleaning scraping and cross site scripting xss safety jsoup may incorrectly sanitize html including javascript url expressions which could allow xss attacks when a reader subsequently clicks that link if the non default safelist preserverelativelinks option is enabled html including javascript urls that have been crafted with control characters will not be sanitized if the site that this html is published on does not set a content security policy an xss attack is then possible this issue is patched in jsoup users should upgrade to this version additionally as the unsanitized input may have been persisted old content should be cleaned again using the updated version to remediate this issue without immediately upgrading disable safelist preserverelativelinks which will rewrite input urls as absolute urls ensure an appropriate is defined this should be used regardless of upgrading as a defence in depth best practice publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
1,644
6,572,668,206
IssuesEvent
2017-09-11 04:14:52
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
support for zabbix maintenance to modify the period
affects_2.1 feature_idea waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME - monitoring/zabbix_maintenance.py ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- N/A --> ##### OS / ENVIRONMENT <!--- N/A --> ##### SUMMARY The module creates or deletes a maintenance periode in zabbix. But it would be nice if the role would also give the option to update a role. sometimes you want to deploy a system with a role name that is the same as another system that was deployed ##### STEPS TO REPRODUCE <!--- N/A -->
True
support for zabbix maintenance to modify the period - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME - monitoring/zabbix_maintenance.py ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### CONFIGURATION <!--- N/A --> ##### OS / ENVIRONMENT <!--- N/A --> ##### SUMMARY The module creates or deletes a maintenance periode in zabbix. But it would be nice if the role would also give the option to update a role. sometimes you want to deploy a system with a role name that is the same as another system that was deployed ##### STEPS TO REPRODUCE <!--- N/A -->
main
support for zabbix maintenance to modify the period issue type feature idea component name monitoring zabbix maintenance py ansible version ansible configuration n a os environment n a summary the module creates or deletes a maintenance periode in zabbix but it would be nice if the role would also give the option to update a role sometimes you want to deploy a system with a role name that is the same as another system that was deployed steps to reproduce n a
1
2,604
8,838,821,819
IssuesEvent
2019-01-05 21:36:40
lrozenblyum/chess
https://api.github.com/repos/lrozenblyum/chess
closed
Refactoring: 0.4
maintainability
Based on #240 and also IDE inspections: refactor the code. This task has no particular limit, however it's a free refactoring for 0.4. In order to make it more manageable, let's set up goals: 1) 78 -> 65 issues in SonarCloud 2) IDE: at least 1 inspection issue fixed
True
Refactoring: 0.4 - Based on #240 and also IDE inspections: refactor the code. This task has no particular limit, however it's a free refactoring for 0.4. In order to make it more manageable, let's set up goals: 1) 78 -> 65 issues in SonarCloud 2) IDE: at least 1 inspection issue fixed
main
refactoring based on and also ide inspections refactor the code this task has no particular limit however it s a free refactoring for in order to make it more manageable let s set up goals issues in sonarcloud ide at least inspection issue fixed
1
3,396
13,170,558,886
IssuesEvent
2020-08-11 15:19:42
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
reopened
[DataTable]: Disabled batch action buttons are visually too prominent
component: data-table role: dev πŸ€– status: needs triage πŸ•΅οΈβ€β™€οΈ status: waiting for maintainer response πŸ’¬ type: bug πŸ›
## Detailed description > Describe in detail the issue you're having. When a button in the data table batch action bar is disabled, it draws the user's attention more than the actions they are able to perform. Also it looks a bit out-of-place in general and might be confusing to the user since the background color is close to the "selected row" background color. > Is this issue related to a specific component? `DataTable.TableBatchAction` > What version of the Carbon Design System are you using? `carbon-components@10.16.0` `carbon-components-react@7.16.0` ## Steps to reproduce the issue 1. Open demo: https://6t41j.csb.app/ 2. Select at least one row ## Additional information ![image](https://user-images.githubusercontent.com/28265588/88920619-6ff7cb00-d26d-11ea-86d9-de9226d3f5b7.png) It seems there is also an inconsistency in the text color compared to the sketch kit, although that's not the issue I think. πŸ˜„ ![image](https://user-images.githubusercontent.com/28265588/88920658-80a84100-d26d-11ea-99ee-231c5c83054c.png)
True
[DataTable]: Disabled batch action buttons are visually too prominent - ## Detailed description > Describe in detail the issue you're having. When a button in the data table batch action bar is disabled, it draws the user's attention more than the actions they are able to perform. Also it looks a bit out-of-place in general and might be confusing to the user since the background color is close to the "selected row" background color. > Is this issue related to a specific component? `DataTable.TableBatchAction` > What version of the Carbon Design System are you using? `carbon-components@10.16.0` `carbon-components-react@7.16.0` ## Steps to reproduce the issue 1. Open demo: https://6t41j.csb.app/ 2. Select at least one row ## Additional information ![image](https://user-images.githubusercontent.com/28265588/88920619-6ff7cb00-d26d-11ea-86d9-de9226d3f5b7.png) It seems there is also an inconsistency in the text color compared to the sketch kit, although that's not the issue I think. πŸ˜„ ![image](https://user-images.githubusercontent.com/28265588/88920658-80a84100-d26d-11ea-99ee-231c5c83054c.png)
main
disabled batch action buttons are visually too prominent detailed description describe in detail the issue you re having when a button in the data table batch action bar is disabled it draws the user s attention more than the actions they are able to perform also it looks a bit out of place in general and might be confusing to the user since the background color is close to the selected row background color is this issue related to a specific component datatable tablebatchaction what version of the carbon design system are you using carbon components carbon components react steps to reproduce the issue open demo select at least one row additional information it seems there is also an inconsistency in the text color compared to the sketch kit although that s not the issue i think πŸ˜„
1
6,911
2,824,213,057
IssuesEvent
2015-05-21 13:38:37
PSMGGamesSS2015/PSMG_SS_2015_Antichamber
https://api.github.com/repos/PSMGGamesSS2015/PSMG_SS_2015_Antichamber
opened
Image descriptions Level 1 - 8
Design
We need ideas and sketches for the image descriptions for level 1 - 8. Reference: http://striker503.de/wordpress/wp-content/uploads/2013/04/AntiChamber-2.jpg @FelixKalley & @xcezz
1.0
Image descriptions Level 1 - 8 - We need ideas and sketches for the image descriptions for level 1 - 8. Reference: http://striker503.de/wordpress/wp-content/uploads/2013/04/AntiChamber-2.jpg @FelixKalley & @xcezz
non_main
image descriptions level we need ideas and sketches for the image descriptions for level reference felixkalley xcezz
0
3,875
17,146,685,954
IssuesEvent
2021-07-13 15:16:53
chaoss/website
https://api.github.com/repos/chaoss/website
closed
Change the sponsor prospectus link on CHAOSScon NA 2021 website
Maintainer Task
The link currently goes to the github repo. It would be better to directly download the file when someone clicks: https://chaoss.github.io/website/CHAOSScon/2021NA/CHAOSSconNA21_funding_prospectus.pdf
True
Change the sponsor prospectus link on CHAOSScon NA 2021 website - The link currently goes to the github repo. It would be better to directly download the file when someone clicks: https://chaoss.github.io/website/CHAOSScon/2021NA/CHAOSSconNA21_funding_prospectus.pdf
main
change the sponsor prospectus link on chaosscon na website the link currently goes to the github repo it would be better to directly download the file when someone clicks
1
101,088
8,774,140,337
IssuesEvent
2018-12-18 18:54:53
ORNL-CEES/DataTransferKit
https://api.github.com/repos/ORNL-CEES/DataTransferKit
reopened
Add performance regression testing
Testing
Now that we have spent a lot of time optimizing the search, we should monitor the performance to make sure that a future commit does not slow downs the code.
1.0
Add performance regression testing - Now that we have spent a lot of time optimizing the search, we should monitor the performance to make sure that a future commit does not slow downs the code.
non_main
add performance regression testing now that we have spent a lot of time optimizing the search we should monitor the performance to make sure that a future commit does not slow downs the code
0
5,196
26,415,597,493
IssuesEvent
2023-01-13 15:43:23
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
opened
SPIKE | PNI Refactoring
engineering maintain
Let's identify where we can improve the readability and performance of PNI from a code standpoint. The goal of this ticket is to identify our largest problem areas concerning PNI. Later (not part of this ticket), we will take these problem areas and break them down into actionable tickets. Timebox TBD
True
SPIKE | PNI Refactoring - Let's identify where we can improve the readability and performance of PNI from a code standpoint. The goal of this ticket is to identify our largest problem areas concerning PNI. Later (not part of this ticket), we will take these problem areas and break them down into actionable tickets. Timebox TBD
main
spike pni refactoring let s identify where we can improve the readability and performance of pni from a code standpoint the goal of this ticket is to identify our largest problem areas concerning pni later not part of this ticket we will take these problem areas and break them down into actionable tickets timebox tbd
1
267,079
28,494,025,107
IssuesEvent
2023-04-18 13:08:01
OpenLiberty/open-liberty
https://api.github.com/repos/OpenLiberty/open-liberty
closed
nbf should not be validated in a back channel logout logout_token
bug team:Security SSO
Testing back channel logout and have a logout token that contains an nbf value that is in the future. The validation of the token should ignore this value (as it is not required by the spec). The runtime is currently validating it and should not. nbf = 2056387597 [4/10/23, 12:44:21:647 CDT] 00000041 rity.openidconnect.backchannellogout.BackchannelLogoutHelper E CWWKS1541E: The back-channel logout request sent to [/oidcclient/backchannel_logout/clientSignHS256] encountered an error. CWWKS1543E: The logout token in the back-channel logout request cannot be validated: org.jose4j.jwt.consumer.InvalidJwtException: JWT (claims->{"iss":"https://localhost:8920/oidc/endpoint/OidcConfigSample_hs","sub":"testuser","sid":"cZP1fFZf1j6SaPwdaGnF","aud":"clientSignHS256","iat":1681148661,"jti":"G45Fl-dYK03jEktW5jFXuQ","events":{"http://schemas.openid.net/event/backchannel-logout":{}},"nbf":2056387597}) rejected due to invalid claims. Additional details: [[6] The JWT is not yet valid as the evaluation time NumericDate{1681148661 -> Apr 10, 2023, 12:44:21 PM CDT} is before the Not Before (nbf=NumericDate{2056387597 -> Mar 1, 2035, 12:46:37 PM CST}) claim time.] When the runtime is updated, the test "LogoutTokenValidationTests_include_ignored_claims" should be updated - the changes needed are marked with "TODO"s.
True
nbf should not be validated in a back channel logout logout_token - Testing back channel logout and have a logout token that contains an nbf value that is in the future. The validation of the token should ignore this value (as it is not required by the spec). The runtime is currently validating it and should not. nbf = 2056387597 [4/10/23, 12:44:21:647 CDT] 00000041 rity.openidconnect.backchannellogout.BackchannelLogoutHelper E CWWKS1541E: The back-channel logout request sent to [/oidcclient/backchannel_logout/clientSignHS256] encountered an error. CWWKS1543E: The logout token in the back-channel logout request cannot be validated: org.jose4j.jwt.consumer.InvalidJwtException: JWT (claims->{"iss":"https://localhost:8920/oidc/endpoint/OidcConfigSample_hs","sub":"testuser","sid":"cZP1fFZf1j6SaPwdaGnF","aud":"clientSignHS256","iat":1681148661,"jti":"G45Fl-dYK03jEktW5jFXuQ","events":{"http://schemas.openid.net/event/backchannel-logout":{}},"nbf":2056387597}) rejected due to invalid claims. Additional details: [[6] The JWT is not yet valid as the evaluation time NumericDate{1681148661 -> Apr 10, 2023, 12:44:21 PM CDT} is before the Not Before (nbf=NumericDate{2056387597 -> Mar 1, 2035, 12:46:37 PM CST}) claim time.] When the runtime is updated, the test "LogoutTokenValidationTests_include_ignored_claims" should be updated - the changes needed are marked with "TODO"s.
non_main
nbf should not be validated in a back channel logout logout token testing back channel logout and have a logout token that contains an nbf value that is in the future the validation of the token should ignore this value as it is not required by the spec the runtime is currently validating it and should not nbf rity openidconnect backchannellogout backchannellogouthelper e the back channel logout request sent to encountered an error the logout token in the back channel logout request cannot be validated org jwt consumer invalidjwtexception jwt claims iss rejected due to invalid claims additional details the jwt is not yet valid as the evaluation time numericdate apr pm cdt is before the not before nbf numericdate mar pm cst claim time when the runtime is updated the test logouttokenvalidationtests include ignored claims should be updated the changes needed are marked with todo s
0
1,486
6,424,513,133
IssuesEvent
2017-08-09 13:41:31
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
vmware extra module vmware_vm_facts.py output result
affects_2.1 cloud docs_report vmware waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Documentation Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> ansible-modules-extras/cloud/vmware/vmware_vm_facts.py ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.0.0 ``` ##### CONFIGURATION ANSIBLE_HOSTS=~/ansible/hostname/hosts ANSIBLE_HOST_KEY_CHECKING=False ANSIBLE_SSH_ARGS='' ANSIBLE_TRANSPORT=ssh ANSIBLE_SCP_IF_SSH=y ANSIBLE_LIBRARY=~/ansible/modules/ansible-modules-extras ##### OS / ENVIRONMENT Fedora release 22 (Twenty Two) ##### SUMMARY <!--- Explain the problem briefly --> Need more explain on how to get the facts from module output. Suggest add completed example in doc. ``` - name: Gather all registered virtual machines local_action: module: vmware_vm_facts hostname: esxi_or_vcenter_ip_or_hostname username: username password: password + register : result + - debug : var=result ``` ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` tasks: - name: get all vmware_vm_facts vmware_vm_facts: hostname: "{{vcenter_hostname}}" username: "{{username}}" password: "{{password}}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS ``` TASK [get all vmware_vm_facts] ************************************************* ok: [localhost] "result": { "changed": false, "virtual_machines": { "vm01": { "guest_fullname": "Red Hat Enterprise Linux 7", "ip_address": "10.10.10.10", "power_state": "poweredOn" ... ... } } } ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> PLAY [Gather all registered virtual machines] ********************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [get all vmware_vm_facts] ************************************************* ok: [localhost] <!--- Paste verbatim command output between quotes below --> ``` ```
True
vmware extra module vmware_vm_facts.py output result - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Documentation Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> ansible-modules-extras/cloud/vmware/vmware_vm_facts.py ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.0.0 ``` ##### CONFIGURATION ANSIBLE_HOSTS=~/ansible/hostname/hosts ANSIBLE_HOST_KEY_CHECKING=False ANSIBLE_SSH_ARGS='' ANSIBLE_TRANSPORT=ssh ANSIBLE_SCP_IF_SSH=y ANSIBLE_LIBRARY=~/ansible/modules/ansible-modules-extras ##### OS / ENVIRONMENT Fedora release 22 (Twenty Two) ##### SUMMARY <!--- Explain the problem briefly --> Need more explain on how to get the facts from module output. Suggest add completed example in doc. ``` - name: Gather all registered virtual machines local_action: module: vmware_vm_facts hostname: esxi_or_vcenter_ip_or_hostname username: username password: password + register : result + - debug : var=result ``` ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` tasks: - name: get all vmware_vm_facts vmware_vm_facts: hostname: "{{vcenter_hostname}}" username: "{{username}}" password: "{{password}}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS ``` TASK [get all vmware_vm_facts] ************************************************* ok: [localhost] "result": { "changed": false, "virtual_machines": { "vm01": { "guest_fullname": "Red Hat Enterprise Linux 7", "ip_address": "10.10.10.10", "power_state": "poweredOn" ... ... } } } ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> PLAY [Gather all registered virtual machines] ********************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [get all vmware_vm_facts] ************************************************* ok: [localhost] <!--- Paste verbatim command output between quotes below --> ``` ```
main
vmware extra module vmware vm facts py output result issue type documentation report component name ansible modules extras cloud vmware vmware vm facts py ansible version ansible configuration ansible hosts ansible hostname hosts ansible host key checking false ansible ssh args ansible transport ssh ansible scp if ssh y ansible library ansible modules ansible modules extras os environment fedora release twenty two summary need more explain on how to get the facts from module output suggest add completed example in doc name gather all registered virtual machines local action module vmware vm facts hostname esxi or vcenter ip or hostname username username password password register result debug var result steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used tasks name get all vmware vm facts vmware vm facts hostname vcenter hostname username username password password expected results task ok result changed false virtual machines guest fullname red hat enterprise linux ip address power state poweredon actual results play task ok task ok
1
5,172
26,344,036,728
IssuesEvent
2023-01-10 20:15:46
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
sam package and sam build refer to different locations for layers
type/bug maintainer/need-response
<!-- Before reporting a new issue, make sure we don't have any duplicates already open or closed by searching the issues list. If there is a duplicate, re-open or add a comment to the existing issue instead of creating a new one. If you are reporting a bug, make sure to include relevant information asked below to help with debugging. ## GENERAL HELP QUESTIONS ## Github Issues is for bug reports and feature requests. If you have general support questions, the following locations are a good place: - Post a question in StackOverflow with "aws-sam" tag --> **Description:** While working with SAM to build a serverless app using both functions and layers locally appears to refer to different locations depending on if I am building or deploying. <!-- Briefly describe the problem you are facing --> It appears that when you do `sam build` it copies your `CodeUri` content to a tmp dir and exposes `$(ARTIFACTS_DIR)`. Which is a pointer to `$PWD/.aws-sam/build/<logical-resource-id>` . However, when you run `sam package` it bundles from `CodeUri` My project has the following structure: ```bash .aws-sam/build/ Lambda/ # my lambda code is staged here LayerOne/Makefile LayerOne/requirements.txt LayerOne/python # all staged modules are located here ``` **Steps to reproduce the issue:** Python version: ```bash $ pipenv run python -VVV Loading .env environment variables… Python 3.7.6 (default, Feb 26 2020, 20:54:15) [GCC 7.3.1 20180712 (Red Hat 7.3.1-6)] ``` OS Version: ```bash $ uname -a Linux swfa 4.19.76-linuxkit aws/serverless-application-model#1 SMP Tue May 26 11:42:35 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux $ cat /etc/os-release NAME="Amazon Linux" VERSION="2" ID="amzn" ID_LIKE="centos rhel fedora" VERSION_ID="2" PRETTY_NAME="Amazon Linux 2" ANSI_COLOR="0;33" CPE_NAME="cpe:2.3:o:amazon:amazon_linux:2" HOME_URL="https://amazonlinux.com/" ``` SAM version: ```bash $ sam version SAM CLI, version 0.52.0 ``` 1. Create a layer based on instructions from [here](https://docs.aws.amazon.com/serverless-application-model/latest/developerguide/building-layers.html) 2. Use a `CodeUri` of `Layer1` with a config similar too: SAM Template: ```yaml LayerOne: Type: "AWS::Serverless::LayerVersion" Properties: LayerName: layer_one CompatibleRuntimes: - python3.7 ContentUri: LayerOne/ Description: Common code for Lambdas LicenseInfo: MIT RetentionPolicy: Retain Metadata: BuildMethod: makefile ``` 3. Define a `Makefile`: ```Makefile requirements.txt: pipenv lock -r > ./requirements.txt build-LayerOne: requirements.txt $(info [+] Installing '3rd party' dependencies...) pipenv run pip3 install \ --isolated \ --disable-pip-version-check \ -Ur requirements.txt -t $(ARTIFACTS_DIR)/python cp -f ./requirements.txt $(ARTIFACTS_DIR)/python/ ``` 3. Build: ```bash sam build LayerOne ``` 4. Package: ```bash sam package \ --template-file template.yaml \ --output-template-file packaged.yaml \ --s3-bucket ${BUCKET} \ --s3-prefix ${BUCKET_PREFIX} ``` 5. Deploy ``` sam deploy \ --no-fail-on-empty-changeset \ --template-file packaged.yaml \ --stack-name ${STACK_NAME} \ --s3-bucket ${BUCKET} \ --s3-prefix stacks \ --capabilities CAPABILITY_IAM CAPABILITY_AUTO_EXPAND ``` **Observed result:** When you download the layer, the only files that will be bundled will be `requirements.txt` because the pip install targeted the `$(ARTIFACTS_DIR)` and did not stage to `CodeUri/python` **Expected result:** The layer should contain the entire contents of `$(ARTIFACTS_DIR)` 6. Revise your `Makefile` Makefile: ``` build-LayerOne: configure ls -latr > $(ARTIFACTS_DIR)/log pwd >> $(ARTIFACTS_DIR)/log cp -fa ./python $(ARTIFACTS_DIR)/ cp -f ./requirements.txt $(ARTIFACTS_DIR)/python/ requirements.txt: pipenv lock -r > ./requirements.txt configure: ./requirements.txt $(info [+] Installing '3rd party' dependencies...) pipenv run pip3 install \ --isolated \ --disable-pip-version-check \ -Ur requirements.txt -t ./python ``` 7. Re-run steps 3, 4, & 5 **Observed result:** When you download the layer, all files that are in `CodeUri/python` will be visible.
True
sam package and sam build refer to different locations for layers - <!-- Before reporting a new issue, make sure we don't have any duplicates already open or closed by searching the issues list. If there is a duplicate, re-open or add a comment to the existing issue instead of creating a new one. If you are reporting a bug, make sure to include relevant information asked below to help with debugging. ## GENERAL HELP QUESTIONS ## Github Issues is for bug reports and feature requests. If you have general support questions, the following locations are a good place: - Post a question in StackOverflow with "aws-sam" tag --> **Description:** While working with SAM to build a serverless app using both functions and layers locally appears to refer to different locations depending on if I am building or deploying. <!-- Briefly describe the problem you are facing --> It appears that when you do `sam build` it copies your `CodeUri` content to a tmp dir and exposes `$(ARTIFACTS_DIR)`. Which is a pointer to `$PWD/.aws-sam/build/<logical-resource-id>` . However, when you run `sam package` it bundles from `CodeUri` My project has the following structure: ```bash .aws-sam/build/ Lambda/ # my lambda code is staged here LayerOne/Makefile LayerOne/requirements.txt LayerOne/python # all staged modules are located here ``` **Steps to reproduce the issue:** Python version: ```bash $ pipenv run python -VVV Loading .env environment variables… Python 3.7.6 (default, Feb 26 2020, 20:54:15) [GCC 7.3.1 20180712 (Red Hat 7.3.1-6)] ``` OS Version: ```bash $ uname -a Linux swfa 4.19.76-linuxkit aws/serverless-application-model#1 SMP Tue May 26 11:42:35 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux $ cat /etc/os-release NAME="Amazon Linux" VERSION="2" ID="amzn" ID_LIKE="centos rhel fedora" VERSION_ID="2" PRETTY_NAME="Amazon Linux 2" ANSI_COLOR="0;33" CPE_NAME="cpe:2.3:o:amazon:amazon_linux:2" HOME_URL="https://amazonlinux.com/" ``` SAM version: ```bash $ sam version SAM CLI, version 0.52.0 ``` 1. Create a layer based on instructions from [here](https://docs.aws.amazon.com/serverless-application-model/latest/developerguide/building-layers.html) 2. Use a `CodeUri` of `Layer1` with a config similar too: SAM Template: ```yaml LayerOne: Type: "AWS::Serverless::LayerVersion" Properties: LayerName: layer_one CompatibleRuntimes: - python3.7 ContentUri: LayerOne/ Description: Common code for Lambdas LicenseInfo: MIT RetentionPolicy: Retain Metadata: BuildMethod: makefile ``` 3. Define a `Makefile`: ```Makefile requirements.txt: pipenv lock -r > ./requirements.txt build-LayerOne: requirements.txt $(info [+] Installing '3rd party' dependencies...) pipenv run pip3 install \ --isolated \ --disable-pip-version-check \ -Ur requirements.txt -t $(ARTIFACTS_DIR)/python cp -f ./requirements.txt $(ARTIFACTS_DIR)/python/ ``` 3. Build: ```bash sam build LayerOne ``` 4. Package: ```bash sam package \ --template-file template.yaml \ --output-template-file packaged.yaml \ --s3-bucket ${BUCKET} \ --s3-prefix ${BUCKET_PREFIX} ``` 5. Deploy ``` sam deploy \ --no-fail-on-empty-changeset \ --template-file packaged.yaml \ --stack-name ${STACK_NAME} \ --s3-bucket ${BUCKET} \ --s3-prefix stacks \ --capabilities CAPABILITY_IAM CAPABILITY_AUTO_EXPAND ``` **Observed result:** When you download the layer, the only files that will be bundled will be `requirements.txt` because the pip install targeted the `$(ARTIFACTS_DIR)` and did not stage to `CodeUri/python` **Expected result:** The layer should contain the entire contents of `$(ARTIFACTS_DIR)` 6. Revise your `Makefile` Makefile: ``` build-LayerOne: configure ls -latr > $(ARTIFACTS_DIR)/log pwd >> $(ARTIFACTS_DIR)/log cp -fa ./python $(ARTIFACTS_DIR)/ cp -f ./requirements.txt $(ARTIFACTS_DIR)/python/ requirements.txt: pipenv lock -r > ./requirements.txt configure: ./requirements.txt $(info [+] Installing '3rd party' dependencies...) pipenv run pip3 install \ --isolated \ --disable-pip-version-check \ -Ur requirements.txt -t ./python ``` 7. Re-run steps 3, 4, & 5 **Observed result:** When you download the layer, all files that are in `CodeUri/python` will be visible.
main
sam package and sam build refer to different locations for layers before reporting a new issue make sure we don t have any duplicates already open or closed by searching the issues list if there is a duplicate re open or add a comment to the existing issue instead of creating a new one if you are reporting a bug make sure to include relevant information asked below to help with debugging general help questions github issues is for bug reports and feature requests if you have general support questions the following locations are a good place post a question in stackoverflow with aws sam tag description while working with sam to build a serverless app using both functions and layers locally appears to refer to different locations depending on if i am building or deploying it appears that when you do sam build it copies your codeuri content to a tmp dir and exposes artifacts dir which is a pointer to pwd aws sam build however when you run sam package it bundles from codeuri my project has the following structure bash aws sam build lambda my lambda code is staged here layerone makefile layerone requirements txt layerone python all staged modules are located here steps to reproduce the issue python version bash pipenv run python vvv loading env environment variables… python default feb os version bash uname a linux swfa linuxkit aws serverless application model smp tue may utc gnu linux cat etc os release name amazon linux version id amzn id like centos rhel fedora version id pretty name amazon linux ansi color cpe name cpe o amazon amazon linux home url sam version bash sam version sam cli version create a layer based on instructions from use a codeuri of with a config similar too sam template yaml layerone type aws serverless layerversion properties layername layer one compatibleruntimes contenturi layerone description common code for lambdas licenseinfo mit retentionpolicy retain metadata buildmethod makefile define a makefile makefile requirements txt pipenv lock r requirements txt build layerone requirements txt info installing party dependencies pipenv run install isolated disable pip version check ur requirements txt t artifacts dir python cp f requirements txt artifacts dir python build bash sam build layerone package bash sam package template file template yaml output template file packaged yaml bucket bucket prefix bucket prefix deploy sam deploy no fail on empty changeset template file packaged yaml stack name stack name bucket bucket prefix stacks capabilities capability iam capability auto expand observed result when you download the layer the only files that will be bundled will be requirements txt because the pip install targeted the artifacts dir and did not stage to codeuri python expected result the layer should contain the entire contents of artifacts dir revise your makefile makefile build layerone configure ls latr artifacts dir log pwd artifacts dir log cp fa python artifacts dir cp f requirements txt artifacts dir python requirements txt pipenv lock r requirements txt configure requirements txt info installing party dependencies pipenv run install isolated disable pip version check ur requirements txt t python re run steps observed result when you download the layer all files that are in codeuri python will be visible
1
4,080
19,269,015,227
IssuesEvent
2021-12-10 01:32:56
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Self-contained linux pacakge should reside in /opt generally
type/feature area/installation maintainer/need-response
This issue lies in between a bug and a feature request. Thank you for making a standalone Linux installer without Homebrew via https://github.com/aws/aws-sam-cli/pull/2597. As a convention, self-contained packages go to /opt and /opt/bin [1]. By default, this patch installs aws-sam to /usr/local. However, those directories are usually managed by package managers like apt and yum. I do not recommend providing .deb or .rpm packages just because /usr/local is deemed to host aws-sam files. Because, packages there reference libs and configs from /usr/lib and /usr/share. https://github.com/aws/aws-sam-cli/blob/05895aa651b55e07ae5b585d932d85262982d1bc/installer/assets/install#L65-L66 You may also consider providing a Snap or Flatpak package (not sure if it's possible because binaries are to be sandboxed) Again, the above applies to only mainstream Linux distributions. Distributions such as Nix have a different filesystem hierarchy standard [2]. Respekt :heart: to you for providing the -i and -b options for custom installation destination directories. Reference: [1] https://unix.stackexchange.com/questions/11544/what-is-the-difference-between-opt-and-usr-loca [2] https://tldp.org/LDP/Linux-Filesystem-Hierarchy/html/opt.html
True
Self-contained linux pacakge should reside in /opt generally - This issue lies in between a bug and a feature request. Thank you for making a standalone Linux installer without Homebrew via https://github.com/aws/aws-sam-cli/pull/2597. As a convention, self-contained packages go to /opt and /opt/bin [1]. By default, this patch installs aws-sam to /usr/local. However, those directories are usually managed by package managers like apt and yum. I do not recommend providing .deb or .rpm packages just because /usr/local is deemed to host aws-sam files. Because, packages there reference libs and configs from /usr/lib and /usr/share. https://github.com/aws/aws-sam-cli/blob/05895aa651b55e07ae5b585d932d85262982d1bc/installer/assets/install#L65-L66 You may also consider providing a Snap or Flatpak package (not sure if it's possible because binaries are to be sandboxed) Again, the above applies to only mainstream Linux distributions. Distributions such as Nix have a different filesystem hierarchy standard [2]. Respekt :heart: to you for providing the -i and -b options for custom installation destination directories. Reference: [1] https://unix.stackexchange.com/questions/11544/what-is-the-difference-between-opt-and-usr-loca [2] https://tldp.org/LDP/Linux-Filesystem-Hierarchy/html/opt.html
main
self contained linux pacakge should reside in opt generally this issue lies in between a bug and a feature request thank you for making a standalone linux installer without homebrew via as a convention self contained packages go to opt and opt bin by default this patch installs aws sam to usr local however those directories are usually managed by package managers like apt and yum i do not recommend providing deb or rpm packages just because usr local is deemed to host aws sam files because packages there reference libs and configs from usr lib and usr share you may also consider providing a snap or flatpak package not sure if it s possible because binaries are to be sandboxed again the above applies to only mainstream linux distributions distributions such as nix have a different filesystem hierarchy standard respekt heart to you for providing the i and b options for custom installation destination directories reference
1
1,402
6,025,462,176
IssuesEvent
2017-06-08 08:46:11
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_iis_webapplication: overriding upgrade support
affects_2.1 feature_idea waiting_on_maintainer windows
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> win_iis_webapplication ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION no changes ##### OS / ENVIRONMENT running from Ubuntu 14.04 running from: Description: Ubuntu 14.04.4 LTS managing: windows 2012R2 ##### SUMMARY the module's script (ps1) assumes old webapplication physical path is present during the script runtime. hence, if it isn't the script fails with confusing error message If you wish to force deploy everything, i.e. delete files under IIS's feet, and create everything anew and reconfig - then the old instance would fail to find its physical path. this the script fails. I think force upgrade is a very useful use case, esp. in cloud situations. It should be supported. ##### STEPS TO REPRODUCE delete webapplications old physical path, try deploying over new physical path. behold: ``` failed: [ec2-x-x-x-x.compute-1.amazonaws.com] (item={u'key1': u'val1', u'application_pool': u'appPool1', u'name': u'myApp', u'site': u'mySite'}) => {"failed": true, "invocation": {"module_name": "win_iis_webapplication"}, "item": {"application_pool": "appPool1", "key2": "myKey", "name": "myApp", "site": "mySite"}, "msg": "The property 'FullName' cannot be found on this object. Verify that the property exists."} ``` I think a keyword "force" needs to be introduced. If it is "yes", then the script should not unconditionally try and access `$application.PhysicalPath,` but do it in condition, otherwise setting app_folder to be `$env:TEMP` <!--- Paste example playbooks or commands between quotes below --> ``` # new keyword is added: force: true ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expected the webapp to be reconfigured ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> The script failed, with unclear message, it took me several hours to understand what happens, manually copying ps1 file and running it locally. <!--- Paste verbatim command output between quotes below --> ``` ```
True
win_iis_webapplication: overriding upgrade support - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> win_iis_webapplication ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION no changes ##### OS / ENVIRONMENT running from Ubuntu 14.04 running from: Description: Ubuntu 14.04.4 LTS managing: windows 2012R2 ##### SUMMARY the module's script (ps1) assumes old webapplication physical path is present during the script runtime. hence, if it isn't the script fails with confusing error message If you wish to force deploy everything, i.e. delete files under IIS's feet, and create everything anew and reconfig - then the old instance would fail to find its physical path. this the script fails. I think force upgrade is a very useful use case, esp. in cloud situations. It should be supported. ##### STEPS TO REPRODUCE delete webapplications old physical path, try deploying over new physical path. behold: ``` failed: [ec2-x-x-x-x.compute-1.amazonaws.com] (item={u'key1': u'val1', u'application_pool': u'appPool1', u'name': u'myApp', u'site': u'mySite'}) => {"failed": true, "invocation": {"module_name": "win_iis_webapplication"}, "item": {"application_pool": "appPool1", "key2": "myKey", "name": "myApp", "site": "mySite"}, "msg": "The property 'FullName' cannot be found on this object. Verify that the property exists."} ``` I think a keyword "force" needs to be introduced. If it is "yes", then the script should not unconditionally try and access `$application.PhysicalPath,` but do it in condition, otherwise setting app_folder to be `$env:TEMP` <!--- Paste example playbooks or commands between quotes below --> ``` # new keyword is added: force: true ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expected the webapp to be reconfigured ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> The script failed, with unclear message, it took me several hours to understand what happens, manually copying ps1 file and running it locally. <!--- Paste verbatim command output between quotes below --> ``` ```
main
win iis webapplication overriding upgrade support issue type feature idea component name win iis webapplication ansible version ansible config file configured module search path default w o overrides configuration no changes os environment running from ubuntu running from description ubuntu lts managing windows summary the module s script assumes old webapplication physical path is present during the script runtime hence if it isn t the script fails with confusing error message if you wish to force deploy everything i e delete files under iis s feet and create everything anew and reconfig then the old instance would fail to find its physical path this the script fails i think force upgrade is a very useful use case esp in cloud situations it should be supported steps to reproduce delete webapplications old physical path try deploying over new physical path behold failed item u u u application pool u u name u myapp u site u mysite failed true invocation module name win iis webapplication item application pool mykey name myapp site mysite msg the property fullname cannot be found on this object verify that the property exists i think a keyword force needs to be introduced if it is yes then the script should not unconditionally try and access application physicalpath but do it in condition otherwise setting app folder to be env temp new keyword is added force true expected results i expected the webapp to be reconfigured actual results the script failed with unclear message it took me several hours to understand what happens manually copying file and running it locally
1
3,916
17,538,056,428
IssuesEvent
2021-08-12 08:47:49
synthesized-io/fairlens
https://api.github.com/repos/synthesized-io/fairlens
closed
Add "open in colab" badge
category:repository-maintainance status:triage
**Is your feature request related to a problem? Please describe.** We should have a badge that links to colab. Either to the SDK notebook (when fairlens is added) or to the notebook in this repo) See here: https://colab.research.google.com/github/googlecolab/colabtools/blob/master/notebooks/colab-github-demo.ipynb **Describe the solution you'd like** <!--- A clear and concise description of what you want to happen. ---> **Describe alternatives you've considered** <!--- A clear and concise description of any alternative solutions or features you've considered. ---> **Additional context** <!--- Add any other context or screenshots about the feature request here. --->
True
Add "open in colab" badge - **Is your feature request related to a problem? Please describe.** We should have a badge that links to colab. Either to the SDK notebook (when fairlens is added) or to the notebook in this repo) See here: https://colab.research.google.com/github/googlecolab/colabtools/blob/master/notebooks/colab-github-demo.ipynb **Describe the solution you'd like** <!--- A clear and concise description of what you want to happen. ---> **Describe alternatives you've considered** <!--- A clear and concise description of any alternative solutions or features you've considered. ---> **Additional context** <!--- Add any other context or screenshots about the feature request here. --->
main
add open in colab badge is your feature request related to a problem please describe we should have a badge that links to colab either to the sdk notebook when fairlens is added or to the notebook in this repo see here describe the solution you d like describe alternatives you ve considered additional context
1
4,118
19,533,314,528
IssuesEvent
2021-12-30 22:01:00
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
Application to join: a8w4
Maintainer application
Hello and welcome to the contrib application process! We're happy to have you :) ## Please note these 3 requirements for new contrib projects: - [x] N/A Include a README.md file containing license and maintainer information. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md - [x] N/A Include a LICENSE.txt file. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt. - [x] N/A If porting a Drupal 7 project, Maintain the Git history from Drupal. ## Please provide the following information: **If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits** - https://github.com/backdrop-contrib/pdf_to_image/pull/7 - https://github.com/backdrop-contrib/pdf_to_image/pull/5 - https://github.com/backdrop-contrib/pdf_to_image/pull/6 **If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)** YES
True
Application to join: a8w4 - Hello and welcome to the contrib application process! We're happy to have you :) ## Please note these 3 requirements for new contrib projects: - [x] N/A Include a README.md file containing license and maintainer information. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md - [x] N/A Include a LICENSE.txt file. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt. - [x] N/A If porting a Drupal 7 project, Maintain the Git history from Drupal. ## Please provide the following information: **If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits** - https://github.com/backdrop-contrib/pdf_to_image/pull/7 - https://github.com/backdrop-contrib/pdf_to_image/pull/5 - https://github.com/backdrop-contrib/pdf_to_image/pull/6 **If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)** YES
main
application to join hello and welcome to the contrib application process we re happy to have you please note these requirements for new contrib projects n a include a readme md file containing license and maintainer information you can use this example n a include a license txt file you can use this example n a if porting a drupal project maintain the git history from drupal please provide the following information if you have already contributed code to backdrop core or contrib projects please provide links to pull requests or commits if you have chosen option or above do you agree to the yes
1
4,547
23,695,732,574
IssuesEvent
2022-08-29 14:30:24
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Remove LogStream writers from Engine
kind/toil area/maintainability
**Description** Part of #9600 In the current state the TypedProcessors (which might become the Engine later) and other entities, like [JobTimeoutTrigger](https://github.com/camunda/zeebe/blob/2d4901c0e516b7fa8b1cf3408b06708c1644ca57/engine/src/main/java/io/camunda/zeebe/engine/processing/job/JobTimeoutTrigger.java#L79-L82), DeploymentDistributor etc. have knowledge about how to write Records to the LogStream abstraction. Ideally, they shouldn't care about that detail. In the end, it would be great if the Engine can get something in and produce something out. For that we have to do some pre-work like removing the actual LogStreamBatchWriters usage and reduce related interfaces. In the POC #9602 we split up the implementation of the `TypedStreamWriterImpl` so that it just writes into a wrapped buffer. This allowed to pre-claim that buffer in the start, and initialize the Engine with that Writer. This means we can reduce the dependency to the LogStream (helps when we split the Engine and the StreamPlatform). Later we renamed the Writers to something like Builder, but this is discussable. See `Writers` in https://github.com/camunda/zeebe/issues/9602#issuecomment-1172029736 **Todo:** - [x] #9780 - [x] Copy the content from `TypedStreamWriterImpl` to a new class - [x] Remove the LogStreamBatch usage from that class, write directly into a pre-claimed buffer - [x] ~Add a new method to the LogStreamBatch~ - [x] ~Write tests for `LogStreamBatch`~ - [x] made the canWriteEvent method public on the LogStreamBatchWriter in order to use it in the RecordBatch - [x] https://github.com/camunda/zeebe/issues/10001 - [x] Replace with RecordBatch - [x] Use RecordBatch everywhere - [x] #9838 - [x] https://github.com/camunda/zeebe/issues/9998 - [x] Find a good new name for the Writers Interface, including Command-, Rejection-, StateWriter. In the POC #9602, we called it Builders, because wanted in the end build a list of Records. But maybe we can also use a different name here. - [x] Create a new Result Interface / Class which can be returned by the Engine - [x] The Result can for simplicity return a BufferWriter like we did in the POC #9602 (we can improve that later), but we can also just return a list of Records if you find a good way for it - [x] Result should be used by Processing and ScheduledTasks to return Results - [x] Engine should return the result - :warning: here we need the Stream Processing changes first https://github.com/camunda/zeebe/issues/9725 - [x] Use result in platform, write the records to the LogStreamBatchWriter - [x] **Bonus:** The result contains only the list of the records (instead of the general Buffer or BufferWriter) - [x] **Bonus2** The serialization is done in the StreamPlatform on writing to the LogStreamBatchWriter - [x] We might need to adjust some tests
True
Remove LogStream writers from Engine - **Description** Part of #9600 In the current state the TypedProcessors (which might become the Engine later) and other entities, like [JobTimeoutTrigger](https://github.com/camunda/zeebe/blob/2d4901c0e516b7fa8b1cf3408b06708c1644ca57/engine/src/main/java/io/camunda/zeebe/engine/processing/job/JobTimeoutTrigger.java#L79-L82), DeploymentDistributor etc. have knowledge about how to write Records to the LogStream abstraction. Ideally, they shouldn't care about that detail. In the end, it would be great if the Engine can get something in and produce something out. For that we have to do some pre-work like removing the actual LogStreamBatchWriters usage and reduce related interfaces. In the POC #9602 we split up the implementation of the `TypedStreamWriterImpl` so that it just writes into a wrapped buffer. This allowed to pre-claim that buffer in the start, and initialize the Engine with that Writer. This means we can reduce the dependency to the LogStream (helps when we split the Engine and the StreamPlatform). Later we renamed the Writers to something like Builder, but this is discussable. See `Writers` in https://github.com/camunda/zeebe/issues/9602#issuecomment-1172029736 **Todo:** - [x] #9780 - [x] Copy the content from `TypedStreamWriterImpl` to a new class - [x] Remove the LogStreamBatch usage from that class, write directly into a pre-claimed buffer - [x] ~Add a new method to the LogStreamBatch~ - [x] ~Write tests for `LogStreamBatch`~ - [x] made the canWriteEvent method public on the LogStreamBatchWriter in order to use it in the RecordBatch - [x] https://github.com/camunda/zeebe/issues/10001 - [x] Replace with RecordBatch - [x] Use RecordBatch everywhere - [x] #9838 - [x] https://github.com/camunda/zeebe/issues/9998 - [x] Find a good new name for the Writers Interface, including Command-, Rejection-, StateWriter. In the POC #9602, we called it Builders, because wanted in the end build a list of Records. But maybe we can also use a different name here. - [x] Create a new Result Interface / Class which can be returned by the Engine - [x] The Result can for simplicity return a BufferWriter like we did in the POC #9602 (we can improve that later), but we can also just return a list of Records if you find a good way for it - [x] Result should be used by Processing and ScheduledTasks to return Results - [x] Engine should return the result - :warning: here we need the Stream Processing changes first https://github.com/camunda/zeebe/issues/9725 - [x] Use result in platform, write the records to the LogStreamBatchWriter - [x] **Bonus:** The result contains only the list of the records (instead of the general Buffer or BufferWriter) - [x] **Bonus2** The serialization is done in the StreamPlatform on writing to the LogStreamBatchWriter - [x] We might need to adjust some tests
main
remove logstream writers from engine description part of in the current state the typedprocessors which might become the engine later and other entities like deploymentdistributor etc have knowledge about how to write records to the logstream abstraction ideally they shouldn t care about that detail in the end it would be great if the engine can get something in and produce something out for that we have to do some pre work like removing the actual logstreambatchwriters usage and reduce related interfaces in the poc we split up the implementation of the typedstreamwriterimpl so that it just writes into a wrapped buffer this allowed to pre claim that buffer in the start and initialize the engine with that writer this means we can reduce the dependency to the logstream helps when we split the engine and the streamplatform later we renamed the writers to something like builder but this is discussable see writers in todo copy the content from typedstreamwriterimpl to a new class remove the logstreambatch usage from that class write directly into a pre claimed buffer add a new method to the logstreambatch write tests for logstreambatch made the canwriteevent method public on the logstreambatchwriter in order to use it in the recordbatch replace with recordbatch use recordbatch everywhere find a good new name for the writers interface including command rejection statewriter in the poc we called it builders because wanted in the end build a list of records but maybe we can also use a different name here create a new result interface class which can be returned by the engine the result can for simplicity return a bufferwriter like we did in the poc we can improve that later but we can also just return a list of records if you find a good way for it result should be used by processing and scheduledtasks to return results engine should return the result warning here we need the stream processing changes first use result in platform write the records to the logstreambatchwriter bonus the result contains only the list of the records instead of the general buffer or bufferwriter the serialization is done in the streamplatform on writing to the logstreambatchwriter we might need to adjust some tests
1
15,226
3,933,341,165
IssuesEvent
2016-04-25 18:47:14
driskell/log-courier
https://api.github.com/repos/driskell/log-courier
closed
lc-admin utility outdated
documentation roadmap
Hi, seems that the information in https://github.com/driskell/log-courier/blob/master/docs/AdministrationUtility.md is outdated. There's no thing as "status", the supported commands are: ``` > help Available commands: prospector [status | files [id]] Get information on prospector state and running harvesters publisher [status | endpoints [id]] Get information on connectivity and endpoints version Get the remote version exit Exit ``` The `help` also doesn't show a "reload" command, which is supported.
1.0
lc-admin utility outdated - Hi, seems that the information in https://github.com/driskell/log-courier/blob/master/docs/AdministrationUtility.md is outdated. There's no thing as "status", the supported commands are: ``` > help Available commands: prospector [status | files [id]] Get information on prospector state and running harvesters publisher [status | endpoints [id]] Get information on connectivity and endpoints version Get the remote version exit Exit ``` The `help` also doesn't show a "reload" command, which is supported.
non_main
lc admin utility outdated hi seems that the information in is outdated there s no thing as status the supported commands are help available commands prospector get information on prospector state and running harvesters publisher get information on connectivity and endpoints version get the remote version exit exit the help also doesn t show a reload command which is supported
0
3,102
5,255,133,752
IssuesEvent
2017-02-02 14:51:59
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Stack Reports Up to Date but Custom Catalog has Higher Versions
area/catalog-service
**Rancher Versions:** Server: 1.3.2 **Docker Version:** 1.10.3 **OS and where are the hosts located? (cloud, bare metal, etc):** Ubuntu 14.04 LTS, AWS EC2 **Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)** single node rancher with external host agent **Environment Type: (Cattle/Kubernetes/Swarm/Mesos)** Cattle After upgrading to rancher server version 1.3.2, a few of our existing stacks that had available upgrades now report "up to date". If you browse to the catalog there are higher versions available, however if you click on the "up to date" link it brings you to an upgrade page with only the current version in the drop down. Thoughts?
1.0
Stack Reports Up to Date but Custom Catalog has Higher Versions - **Rancher Versions:** Server: 1.3.2 **Docker Version:** 1.10.3 **OS and where are the hosts located? (cloud, bare metal, etc):** Ubuntu 14.04 LTS, AWS EC2 **Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)** single node rancher with external host agent **Environment Type: (Cattle/Kubernetes/Swarm/Mesos)** Cattle After upgrading to rancher server version 1.3.2, a few of our existing stacks that had available upgrades now report "up to date". If you browse to the catalog there are higher versions available, however if you click on the "up to date" link it brings you to an upgrade page with only the current version in the drop down. Thoughts?
non_main
stack reports up to date but custom catalog has higher versions rancher versions server docker version os and where are the hosts located cloud bare metal etc ubuntu lts aws setup details single node rancher vs ha rancher internal db vs external db single node rancher with external host agent environment type cattle kubernetes swarm mesos cattle after upgrading to rancher server version a few of our existing stacks that had available upgrades now report up to date if you browse to the catalog there are higher versions available however if you click on the up to date link it brings you to an upgrade page with only the current version in the drop down thoughts
0
163,043
6,188,880,176
IssuesEvent
2017-07-04 11:24:39
chartjs/Chart.js
https://api.github.com/repos/chartjs/Chart.js
closed
[BUG] array in pointBorderWidth
Category: Enhancement Help wanted Priority: p2
## Expected Behavior Array of numbers in pointBorderWidth (using line type chart) change the width of border for each point of chart. ## Current Behavior Array with more than one number in pointBorderWidth (using line type chart) are ignored and chart use borderWidth instead. ## Possible Solution Function getPointBorderWidth only check if value is not NaN. no behaviour if value is array. ## Context I need every point in line chart have different border width. ## Environment * Chart.js version: 2.5.0
1.0
[BUG] array in pointBorderWidth - ## Expected Behavior Array of numbers in pointBorderWidth (using line type chart) change the width of border for each point of chart. ## Current Behavior Array with more than one number in pointBorderWidth (using line type chart) are ignored and chart use borderWidth instead. ## Possible Solution Function getPointBorderWidth only check if value is not NaN. no behaviour if value is array. ## Context I need every point in line chart have different border width. ## Environment * Chart.js version: 2.5.0
non_main
array in pointborderwidth expected behavior array of numbers in pointborderwidth using line type chart change the width of border for each point of chart current behavior array with more than one number in pointborderwidth using line type chart are ignored and chart use borderwidth instead possible solution function getpointborderwidth only check if value is not nan no behaviour if value is array context i need every point in line chart have different border width environment chart js version
0
410,185
11,984,576,503
IssuesEvent
2020-04-07 16:04:11
OpenEnergyDashboard/OED
https://api.github.com/repos/OpenEnergyDashboard/OED
closed
make compare graphs compatible with fast pt
p-high-priority
Fast PT does not support short time interval comparison graphs. To fix this we need to - [ ] Augment back-end to get sub-daily data needed. The first step will be to use the raw data since the number of points will not be great with the current comparison graphs. - [ ] The current comparison graphs are tied to bar charts. These will be decoupled and comparison will now use the new back-end. - [ ] The back-end will be enhanced so it used Fast PT data for daily segments and then add in the extra sub-day data from the raw data.
1.0
make compare graphs compatible with fast pt - Fast PT does not support short time interval comparison graphs. To fix this we need to - [ ] Augment back-end to get sub-daily data needed. The first step will be to use the raw data since the number of points will not be great with the current comparison graphs. - [ ] The current comparison graphs are tied to bar charts. These will be decoupled and comparison will now use the new back-end. - [ ] The back-end will be enhanced so it used Fast PT data for daily segments and then add in the extra sub-day data from the raw data.
non_main
make compare graphs compatible with fast pt fast pt does not support short time interval comparison graphs to fix this we need to augment back end to get sub daily data needed the first step will be to use the raw data since the number of points will not be great with the current comparison graphs the current comparison graphs are tied to bar charts these will be decoupled and comparison will now use the new back end the back end will be enhanced so it used fast pt data for daily segments and then add in the extra sub day data from the raw data
0
110,833
16,990,991,989
IssuesEvent
2021-06-30 20:26:57
joshnewton31080/angular
https://api.github.com/repos/joshnewton31080/angular
opened
WS-2017-0113 (High) detected in angular-1.5.5.js, angular-1.5.3.js
security vulnerability
## WS-2017-0113 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.5.5.js</b>, <b>angular-1.5.3.js</b></p></summary> <p> <details><summary><b>angular-1.5.5.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.5/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.5/angular.js</a></p> <p>Path to dependency file: angular/aio/content/examples/upgrade-phonecat-2-hybrid/index.html</p> <p>Path to vulnerable library: angular/aio/content/examples/upgrade-phonecat-2-hybrid/index.html,angular/aio/content/examples/upgrade-phonecat-1-typescript/app/index.html</p> <p> Dependency Hierarchy: - :x: **angular-1.5.5.js** (Vulnerable Library) </details> <details><summary><b>angular-1.5.3.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.3/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.3/angular.js</a></p> <p>Path to dependency file: angular/aio/content/examples/upgrade-module/src/index-ajs-to-a-providers.html</p> <p>Path to vulnerable library: angular/aio/content/examples/upgrade-module/src/index-ajs-to-a-providers.html</p> <p> Dependency Hierarchy: - :x: **angular-1.5.3.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/joshnewton31080/angular/commit/0754f95d8686bc67f2d9e82ca6b2652dc6fd0bf3">0754f95d8686bc67f2d9e82ca6b2652dc6fd0bf3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> angular.js is vulnerable to XSS. This happens since an attacker can load Angular from the extension, and Angular's auto-bootstrapping can be used to bypass the victim site's CSP protection. <p>Publish Date: 2016-11-02 <p>URL: <a href=https://github.com/angular/angular.js/commit/0ff10e1b56c6b7c4ac465e35c96a5886e294bac5>WS-2017-0113</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/pull/15346">https://github.com/angular/angular.js/pull/15346</a></p> <p>Release Date: 2016-11-02</p> <p>Fix Resolution: angular - 1.6.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"angular.js","packageVersion":"1.5.5","packageFilePaths":["/aio/content/examples/upgrade-phonecat-2-hybrid/index.html","/aio/content/examples/upgrade-phonecat-1-typescript/app/index.html"],"isTransitiveDependency":false,"dependencyTree":"angular.js:1.5.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.6.0"},{"packageType":"JavaScript","packageName":"angular.js","packageVersion":"1.5.3","packageFilePaths":["/aio/content/examples/upgrade-module/src/index-ajs-to-a-providers.html"],"isTransitiveDependency":false,"dependencyTree":"angular.js:1.5.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.6.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0113","vulnerabilityDetails":"angular.js is vulnerable to XSS. This happens since an attacker can load Angular from the extension, and Angular\u0027s auto-bootstrapping can be used to bypass the victim site\u0027s CSP protection.","vulnerabilityUrl":"https://github.com/angular/angular.js/commit/0ff10e1b56c6b7c4ac465e35c96a5886e294bac5","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2017-0113 (High) detected in angular-1.5.5.js, angular-1.5.3.js - ## WS-2017-0113 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>angular-1.5.5.js</b>, <b>angular-1.5.3.js</b></p></summary> <p> <details><summary><b>angular-1.5.5.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.5/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.5/angular.js</a></p> <p>Path to dependency file: angular/aio/content/examples/upgrade-phonecat-2-hybrid/index.html</p> <p>Path to vulnerable library: angular/aio/content/examples/upgrade-phonecat-2-hybrid/index.html,angular/aio/content/examples/upgrade-phonecat-1-typescript/app/index.html</p> <p> Dependency Hierarchy: - :x: **angular-1.5.5.js** (Vulnerable Library) </details> <details><summary><b>angular-1.5.3.js</b></p></summary> <p>AngularJS is an MVC framework for building web applications. The core features include HTML enhanced with custom component and data-binding capabilities, dependency injection and strong focus on simplicity, testability, maintainability and boiler-plate reduction.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.3/angular.js">https://cdnjs.cloudflare.com/ajax/libs/angular.js/1.5.3/angular.js</a></p> <p>Path to dependency file: angular/aio/content/examples/upgrade-module/src/index-ajs-to-a-providers.html</p> <p>Path to vulnerable library: angular/aio/content/examples/upgrade-module/src/index-ajs-to-a-providers.html</p> <p> Dependency Hierarchy: - :x: **angular-1.5.3.js** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/joshnewton31080/angular/commit/0754f95d8686bc67f2d9e82ca6b2652dc6fd0bf3">0754f95d8686bc67f2d9e82ca6b2652dc6fd0bf3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> angular.js is vulnerable to XSS. This happens since an attacker can load Angular from the extension, and Angular's auto-bootstrapping can be used to bypass the victim site's CSP protection. <p>Publish Date: 2016-11-02 <p>URL: <a href=https://github.com/angular/angular.js/commit/0ff10e1b56c6b7c4ac465e35c96a5886e294bac5>WS-2017-0113</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/angular/angular.js/pull/15346">https://github.com/angular/angular.js/pull/15346</a></p> <p>Release Date: 2016-11-02</p> <p>Fix Resolution: angular - 1.6.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"angular.js","packageVersion":"1.5.5","packageFilePaths":["/aio/content/examples/upgrade-phonecat-2-hybrid/index.html","/aio/content/examples/upgrade-phonecat-1-typescript/app/index.html"],"isTransitiveDependency":false,"dependencyTree":"angular.js:1.5.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.6.0"},{"packageType":"JavaScript","packageName":"angular.js","packageVersion":"1.5.3","packageFilePaths":["/aio/content/examples/upgrade-module/src/index-ajs-to-a-providers.html"],"isTransitiveDependency":false,"dependencyTree":"angular.js:1.5.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"angular - 1.6.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-0113","vulnerabilityDetails":"angular.js is vulnerable to XSS. This happens since an attacker can load Angular from the extension, and Angular\u0027s auto-bootstrapping can be used to bypass the victim site\u0027s CSP protection.","vulnerabilityUrl":"https://github.com/angular/angular.js/commit/0ff10e1b56c6b7c4ac465e35c96a5886e294bac5","cvss3Severity":"high","cvss3Score":"8.6","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
ws high detected in angular js angular js ws high severity vulnerability vulnerable libraries angular js angular js angular js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to dependency file angular aio content examples upgrade phonecat hybrid index html path to vulnerable library angular aio content examples upgrade phonecat hybrid index html angular aio content examples upgrade phonecat typescript app index html dependency hierarchy x angular js vulnerable library angular js angularjs is an mvc framework for building web applications the core features include html enhanced with custom component and data binding capabilities dependency injection and strong focus on simplicity testability maintainability and boiler plate reduction library home page a href path to dependency file angular aio content examples upgrade module src index ajs to a providers html path to vulnerable library angular aio content examples upgrade module src index ajs to a providers html dependency hierarchy x angular js vulnerable library found in head commit a href found in base branch master vulnerability details angular js is vulnerable to xss this happens since an attacker can load angular from the extension and angular s auto bootstrapping can be used to bypass the victim site s csp protection publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution angular isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree angular js isminimumfixversionavailable true minimumfixversion angular packagetype javascript packagename angular js packageversion packagefilepaths istransitivedependency false dependencytree angular js isminimumfixversionavailable true minimumfixversion angular basebranches vulnerabilityidentifier ws vulnerabilitydetails angular js is vulnerable to xss this happens since an attacker can load angular from the extension and angular auto bootstrapping can be used to bypass the victim site csp protection vulnerabilityurl
0
1,137
4,998,875,105
IssuesEvent
2016-12-09 21:19:22
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2_group requires "description" to delete SG
affects_2.1 aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME ec2_group ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> ##### SUMMARY `ec2_group` requires `description` parameter to be specified to delete a security group. This doesn't make sense, IMHO, and it seems to contradict to [what Boto requires](http://boto.cloudhackers.com/en/latest/ref/ec2.html#boto.ec2.connection.EC2Connection.delete_security_group). ##### STEPS TO REPRODUCE Tried to execute the following task: <!--- Paste example playbooks or commands between quotes below --> ``` - name: "Destroy Security Group for FE trusted ELB" ec2_group: state: absent region: "{{ region }}" name: "{{ owner }}_sg_{{ env }}_fe_elb_trst" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS Expected that the above task would execute without error. ##### ACTUAL RESULTS The above task raised an error: <!--- Paste verbatim command output between quotes below --> ``` missing required arguments: description ```
True
ec2_group requires "description" to delete SG - ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME ec2_group ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> ##### SUMMARY `ec2_group` requires `description` parameter to be specified to delete a security group. This doesn't make sense, IMHO, and it seems to contradict to [what Boto requires](http://boto.cloudhackers.com/en/latest/ref/ec2.html#boto.ec2.connection.EC2Connection.delete_security_group). ##### STEPS TO REPRODUCE Tried to execute the following task: <!--- Paste example playbooks or commands between quotes below --> ``` - name: "Destroy Security Group for FE trusted ELB" ec2_group: state: absent region: "{{ region }}" name: "{{ owner }}_sg_{{ env }}_fe_elb_trst" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS Expected that the above task would execute without error. ##### ACTUAL RESULTS The above task raised an error: <!--- Paste verbatim command output between quotes below --> ``` missing required arguments: description ```
main
group requires description to delete sg issue type bug report component name group ansible version ansible config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say β€œn a” for anything that is not platform specific summary group requires description parameter to be specified to delete a security group this doesn t make sense imho and it seems to contradict to steps to reproduce tried to execute the following task name destroy security group for fe trusted elb group state absent region region name owner sg env fe elb trst expected results expected that the above task would execute without error actual results the above task raised an error missing required arguments description
1
5,876
31,934,444,193
IssuesEvent
2023-09-19 09:32:27
dbt-labs/docs.getdbt.com
https://api.github.com/repos/dbt-labs/docs.getdbt.com
opened
Add AWS Glue adapter to trusted list
adapter maintainers
### Link to adapter repo https://github.com/aws-samples/dbt-glue ### Contact Details mehdmldj@amazon.com ### Which of these best describes you? I work for the vendor on top of which the dbt adapter functions ### Please agree to the each of the following - [X] I am a maintainer of the adapter being submited for Trusted status - [X] I have read both the [Trusted adapters](https://docs.getdbt.com/docs/trusted-adapters) and [Building a Trusted Adapter](https://docs.getdbt.com/guides/dbt-ecosystem/adapter-development/8-building-a-trusted-adapter) pages. - [X] I believe that the adapter currently meets the expectations given above - [X] I will ensure this adapter stays in compliance with the guidelines - [X] I understand that dbt Labs reserves the right to remove an adapter from the trusted adapter list at any time, should any of the below guidelines not be met ### What icon should be used? <img width="272" alt="image" src="https://github.com/dbt-labs/docs.getdbt.com/assets/38988556/73d76a3f-fd2f-45bc-962a-9f54a755a1fa">
True
Add AWS Glue adapter to trusted list - ### Link to adapter repo https://github.com/aws-samples/dbt-glue ### Contact Details mehdmldj@amazon.com ### Which of these best describes you? I work for the vendor on top of which the dbt adapter functions ### Please agree to the each of the following - [X] I am a maintainer of the adapter being submited for Trusted status - [X] I have read both the [Trusted adapters](https://docs.getdbt.com/docs/trusted-adapters) and [Building a Trusted Adapter](https://docs.getdbt.com/guides/dbt-ecosystem/adapter-development/8-building-a-trusted-adapter) pages. - [X] I believe that the adapter currently meets the expectations given above - [X] I will ensure this adapter stays in compliance with the guidelines - [X] I understand that dbt Labs reserves the right to remove an adapter from the trusted adapter list at any time, should any of the below guidelines not be met ### What icon should be used? <img width="272" alt="image" src="https://github.com/dbt-labs/docs.getdbt.com/assets/38988556/73d76a3f-fd2f-45bc-962a-9f54a755a1fa">
main
add aws glue adapter to trusted list link to adapter repo contact details mehdmldj amazon com which of these best describes you i work for the vendor on top of which the dbt adapter functions please agree to the each of the following i am a maintainer of the adapter being submited for trusted status i have read both the and pages i believe that the adapter currently meets the expectations given above i will ensure this adapter stays in compliance with the guidelines i understand that dbt labs reserves the right to remove an adapter from the trusted adapter list at any time should any of the below guidelines not be met what icon should be used img width alt image src
1
1,907
6,577,567,202
IssuesEvent
2017-09-12 01:48:46
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Add support for StepScaling policies to ec2_scaling_policy module
affects_1.9 aws bug_report cloud feature_idea waiting_on_maintainer
##### Issue Type: - Feature Idea ##### Plugin Name: ec2_scaling_policy ##### Ansible Version: ``` ansible 1.9.3 configured module search path = library ``` ##### Environment: N/A ##### Summary: Seems like ansible does not allow me to create Step Scaling policies. Using the `ec2_scaling_policy` module always results in a simple policy and I'm not sure if I'm misunderstanding the parameters in order to create a StepScaling policy as described here: http://docs.aws.amazon.com/AutoScaling/latest/DeveloperGuide/as-scale-based-on-demand.html (Step Scaling Policies section) You can also find documentation on the AWS CLI docs here: http://docs.aws.amazon.com/cli/latest/reference/autoscaling/put-scaling-policy.html (describes in the `--policy-type` parameter) ##### Steps To Reproduce: Simply try to create a scaling policy with the needed alarm metric and notice how AWS shows a simple scaling policy on the ASG. ##### Expected Results: N/A ##### Actual Results: N/A
True
Add support for StepScaling policies to ec2_scaling_policy module - ##### Issue Type: - Feature Idea ##### Plugin Name: ec2_scaling_policy ##### Ansible Version: ``` ansible 1.9.3 configured module search path = library ``` ##### Environment: N/A ##### Summary: Seems like ansible does not allow me to create Step Scaling policies. Using the `ec2_scaling_policy` module always results in a simple policy and I'm not sure if I'm misunderstanding the parameters in order to create a StepScaling policy as described here: http://docs.aws.amazon.com/AutoScaling/latest/DeveloperGuide/as-scale-based-on-demand.html (Step Scaling Policies section) You can also find documentation on the AWS CLI docs here: http://docs.aws.amazon.com/cli/latest/reference/autoscaling/put-scaling-policy.html (describes in the `--policy-type` parameter) ##### Steps To Reproduce: Simply try to create a scaling policy with the needed alarm metric and notice how AWS shows a simple scaling policy on the ASG. ##### Expected Results: N/A ##### Actual Results: N/A
main
add support for stepscaling policies to scaling policy module issue type feature idea plugin name scaling policy ansible version ansible configured module search path library environment n a summary seems like ansible does not allow me to create step scaling policies using the scaling policy module always results in a simple policy and i m not sure if i m misunderstanding the parameters in order to create a stepscaling policy as described here step scaling policies section you can also find documentation on the aws cli docs here describes in the policy type parameter steps to reproduce simply try to create a scaling policy with the needed alarm metric and notice how aws shows a simple scaling policy on the asg expected results n a actual results n a
1
48,705
5,967,567,371
IssuesEvent
2017-05-30 16:12:18
LLNL/spack
https://api.github.com/repos/LLNL/spack
closed
Python 3 testing
python3 tests
We have a lot of unit tests for Spack, and we use Travis to test several versions of Python 3. But these tests never seem to catch anything. After we added Python 3 support, commands like `spack fetch`, `spack checksum`, and `spack info` didn't work at all for Python 3, but all of our unit tests passed. It occurred to me that perhaps we are not actually testing Python 3 at all. When we run `spack`, isn't it picking up the first `python` in the `PATH`, usually for the system `python2` installation? @alalazo @tgamblin
1.0
Python 3 testing - We have a lot of unit tests for Spack, and we use Travis to test several versions of Python 3. But these tests never seem to catch anything. After we added Python 3 support, commands like `spack fetch`, `spack checksum`, and `spack info` didn't work at all for Python 3, but all of our unit tests passed. It occurred to me that perhaps we are not actually testing Python 3 at all. When we run `spack`, isn't it picking up the first `python` in the `PATH`, usually for the system `python2` installation? @alalazo @tgamblin
non_main
python testing we have a lot of unit tests for spack and we use travis to test several versions of python but these tests never seem to catch anything after we added python support commands like spack fetch spack checksum and spack info didn t work at all for python but all of our unit tests passed it occurred to me that perhaps we are not actually testing python at all when we run spack isn t it picking up the first python in the path usually for the system installation alalazo tgamblin
0
3,899
17,358,803,344
IssuesEvent
2021-07-29 17:33:17
imdhemy/laravel-in-app-purchases
https://api.github.com/repos/imdhemy/laravel-in-app-purchases
closed
Routes are not working with laravel 8
bug maintainer_replied
How to fix the issue with routes? it shows in `php artisan route:list` but when I make a post request it fallback to home route. it matches the route method i.e POST or GET but does not reach to its controller's method.
True
Routes are not working with laravel 8 - How to fix the issue with routes? it shows in `php artisan route:list` but when I make a post request it fallback to home route. it matches the route method i.e POST or GET but does not reach to its controller's method.
main
routes are not working with laravel how to fix the issue with routes it shows in php artisan route list but when i make a post request it fallback to home route it matches the route method i e post or get but does not reach to its controller s method
1
5,343
26,959,532,378
IssuesEvent
2023-02-08 17:07:26
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Support Ref for CodeUri property.
type/feature area/intrinsics maintainer/need-response
We are using sam template to describe our cloud formation stack deploy for Api gateway with lambdas. What we have now in our SAM template for one lambda is: #CodeUri: ./artifacts/AbcProject.zip CodeUri: ./src/providers/AbcProject/bin/Debug/netcoreapp2.1 So when we deploy to AWS - we need to use CodeUri: ./artifacts/AbcProject.zip and when we run api locally we use: CodeUri: ./src/providers/AbcProject/bin/Debug/netcoreapp2.1 Thats why we need to comment/uncomments codeUri depending on whether we deploy or run locally. Is there any way to use conditional CodeUri? Expected result - use conditional CodeUri.
True
Support Ref for CodeUri property. - We are using sam template to describe our cloud formation stack deploy for Api gateway with lambdas. What we have now in our SAM template for one lambda is: #CodeUri: ./artifacts/AbcProject.zip CodeUri: ./src/providers/AbcProject/bin/Debug/netcoreapp2.1 So when we deploy to AWS - we need to use CodeUri: ./artifacts/AbcProject.zip and when we run api locally we use: CodeUri: ./src/providers/AbcProject/bin/Debug/netcoreapp2.1 Thats why we need to comment/uncomments codeUri depending on whether we deploy or run locally. Is there any way to use conditional CodeUri? Expected result - use conditional CodeUri.
main
support ref for codeuri property we are using sam template to describe our cloud formation stack deploy for api gateway with lambdas what we have now in our sam template for one lambda is codeuri artifacts abcproject zip codeuri src providers abcproject bin debug so when we deploy to aws we need to use codeuri artifacts abcproject zip and when we run api locally we use codeuri src providers abcproject bin debug thats why we need to comment uncomments codeuri depending on whether we deploy or run locally is there any way to use conditional codeuri expected result use conditional codeuri
1
1,349
5,790,605,844
IssuesEvent
2017-05-02 01:20:59
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Ansible 2.0.2.0 and later break Webmin port test
affects_2.0 bug_report waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> uri module ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = /home/ansible/library/:/usr/share/ansible/library/ ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ``` [defaults] # some basic default values... hostfile = /etc/ansible/hosts library = ~/library/:/usr/share/ansible/library/ remote_tmp = ~/.ansible/tmp pattern = * forks = 5 poll_interval = 15 remote_user = root sudo_user = root #ask_sudo_pass = True #ask_pass = True transport = smart remote_port = 22 module_lang = C ``` ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> Local: Ansible is running on a TurnKey GNU/Linux Ansible appliance version 14.1 (based on Debian Jessie) Remote: LXC is running on a TurnKey GNU/Linux LXC appliance version 14.1. Chifflier's ansible-lxc-ssh plugin is used to connect to containers https://github.com/chifflier/ansible-lxc-ssh ##### SUMMARY <!--- Explain the problem briefly --> Ansible versions 2.0.2.0 and later break a simple task used test a Webmin port. The task has worked from versions 1.9 through 2.0.1.0. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> The failing task is part of a playbook designed to create an LXC container running a specified TurnKey appliance, and then run a series of tests that verify the appliance is functional. The failing task uses the uri module to connect to the container's Webmin port (12321). It is considered successful if one of the 'good' response codes is returned. <!--- Paste example playbooks or commands between quotes below --> Ansible task that fails ``` vars: ... good_codes: [200, 201, 202, 300, 301, 302, 303, 304, 307, 308, 400, 401, 403] tasks: ... - name: Test appliance webmin tags: test uri: url: "https://{{container}}:12321/" validate_certs: no status_code: "{{good_codes}}" register: webmin ignore_errors: yes until: webmin|success delay: 10 retries: 1 when: app not in ['tkldev'] ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Results from version 2.0.1.0 (passed) ``` TASK [Test appliance webmin] *************************************************** task path: /home/ansible/playbooks/webmin-test.yml:139 <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r -tt lxc '/bin/sh -c '"'"'mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122 `" && echo "` echo ~/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122 `"'"'"'' <lxc> PUT /tmp/tmpJw5Mg0 TO /root/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122/uri <lxc> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r '[lxc]' <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r -tt lxc '/bin/sh -c '"'"'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122/uri; rm -rf "/root/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122/" > /dev/null 2>&1'"'"'' ok: [lxc] => {"cache_control": "no-store, no-cache, must-revalidate, post-check=0, pre-check=0", "changed": false, "connection": "close", "content_location": "https://drupal8-natbr0-container:12321/", "content_security_policy": "script-src 'self' 'unsafe-inline' 'unsafe-eval'; frame-src 'self'; child-src 'self'", "content_type": "text/html; Charset=UTF-8", "date": "Mon, 8 Aug 2016 03:06:44 GMT", "expires": "Thu, 1 Jan 1970 00:00:00 GMT", "invocation": {"module_args": {"backup": null, "body": null, "body_format": "raw", "content": null, "creates": null, "delimiter": null, "dest": null, "directory_mode": null, "follow": false, "follow_redirects": "safe", "force": null, "force_basic_auth": false, "group": null, "method": "GET", "mode": null, "owner": null, "password": null, "regexp": null, "remote_src": null, "removes": null, "return_content": false, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": null, "status_code": [200, 201, 202, 300, 301, 302, 303, 304, 307, 308, 400, 401, 403], "timeout": 30, "url": "https://drupal8-natbr0-container:12321/", "user": null, "validate_certs": false}, "module_name": "uri"}, "pragma": "no-cache", "redirected": false, "server": "MiniServ/1.780", "set_cookie": "testing=1; path=/; secure; httpOnly; httpOnly", "status": 200, "x_frame_options": "SAMEORIGIN"} ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> Results from version 2.0.2.0 ``` TASK [Test appliance webmin] *************************************************** task path: /home/ansible/playbooks/webmin-test.yml:139 <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r lxc '/bin/sh -c '"'"'mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525 `" && echo "` echo ~/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525 `"'"'"'' <lxc> PUT /tmp/tmppg6gtg TO /root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri <lxc> SSH: EXEC sftp -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r '[lxc]' <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r -tt lxc '/bin/sh -c '"'"'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri; rm -rf "/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/" > /dev/null 2>&1'"'"'' fatal: [lxc]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "uri"}, "module_stderr": "OpenSSH_6.7p1 Debian-5+deb8u2, OpenSSL 1.0.1k 8 Jan 2015\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 15131\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to lxc closed.\r\n", "module_stdout": "Traceback (most recent call last):\r\n File \"/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri\", line 3310, in <module>\r\n main()\r\n File \"/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri\", line 407, in main\r\n dict_headers, socket_timeout)\r\n File \"/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri\", line 334, in uri\r\n content = resp.read()\r\n File \"/usr/lib/python2.7/socket.py\", line 351, in read\r\n data = self._sock.recv(rbufsize)\r\n File \"/usr/lib/python2.7/httplib.py\", line 573, in read\r\n s = self.fp.read(amt)\r\n File \"/usr/lib/python2.7/socket.py\", line 380, in read\r\n data = self._sock.recv(left)\r\n File \"/usr/lib/python2.7/ssl.py\", line 714, in recv\r\n return self.read(buflen)\r\n File \"/usr/lib/python2.7/ssl.py\", line 608, in read\r\n v = self._sslobj.read(len or 1024)\r\nssl.SSLError: ('The read operation timed out',)\r\n", "msg": "MODULE FAILURE", "parsed": false} ...ignoring ``` ##### COMMENTS I think the problem may be related to issue #3437 "Discrepancy in uri module behavior between 2.0.1.0 and 2.0.2.0", although I tried the fix suggested for #3437 and it did not help. The problem seems unique to Webmin, as several tests of other ports continue to function.
True
Ansible 2.0.2.0 and later break Webmin port test - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> uri module ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = /home/ansible/library/:/usr/share/ansible/library/ ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ``` [defaults] # some basic default values... hostfile = /etc/ansible/hosts library = ~/library/:/usr/share/ansible/library/ remote_tmp = ~/.ansible/tmp pattern = * forks = 5 poll_interval = 15 remote_user = root sudo_user = root #ask_sudo_pass = True #ask_pass = True transport = smart remote_port = 22 module_lang = C ``` ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> Local: Ansible is running on a TurnKey GNU/Linux Ansible appliance version 14.1 (based on Debian Jessie) Remote: LXC is running on a TurnKey GNU/Linux LXC appliance version 14.1. Chifflier's ansible-lxc-ssh plugin is used to connect to containers https://github.com/chifflier/ansible-lxc-ssh ##### SUMMARY <!--- Explain the problem briefly --> Ansible versions 2.0.2.0 and later break a simple task used test a Webmin port. The task has worked from versions 1.9 through 2.0.1.0. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> The failing task is part of a playbook designed to create an LXC container running a specified TurnKey appliance, and then run a series of tests that verify the appliance is functional. The failing task uses the uri module to connect to the container's Webmin port (12321). It is considered successful if one of the 'good' response codes is returned. <!--- Paste example playbooks or commands between quotes below --> Ansible task that fails ``` vars: ... good_codes: [200, 201, 202, 300, 301, 302, 303, 304, 307, 308, 400, 401, 403] tasks: ... - name: Test appliance webmin tags: test uri: url: "https://{{container}}:12321/" validate_certs: no status_code: "{{good_codes}}" register: webmin ignore_errors: yes until: webmin|success delay: 10 retries: 1 when: app not in ['tkldev'] ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Results from version 2.0.1.0 (passed) ``` TASK [Test appliance webmin] *************************************************** task path: /home/ansible/playbooks/webmin-test.yml:139 <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r -tt lxc '/bin/sh -c '"'"'mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122 `" && echo "` echo ~/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122 `"'"'"'' <lxc> PUT /tmp/tmpJw5Mg0 TO /root/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122/uri <lxc> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r '[lxc]' <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r -tt lxc '/bin/sh -c '"'"'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122/uri; rm -rf "/root/.ansible/tmp/ansible-tmp-1470625603.84-183242562950122/" > /dev/null 2>&1'"'"'' ok: [lxc] => {"cache_control": "no-store, no-cache, must-revalidate, post-check=0, pre-check=0", "changed": false, "connection": "close", "content_location": "https://drupal8-natbr0-container:12321/", "content_security_policy": "script-src 'self' 'unsafe-inline' 'unsafe-eval'; frame-src 'self'; child-src 'self'", "content_type": "text/html; Charset=UTF-8", "date": "Mon, 8 Aug 2016 03:06:44 GMT", "expires": "Thu, 1 Jan 1970 00:00:00 GMT", "invocation": {"module_args": {"backup": null, "body": null, "body_format": "raw", "content": null, "creates": null, "delimiter": null, "dest": null, "directory_mode": null, "follow": false, "follow_redirects": "safe", "force": null, "force_basic_auth": false, "group": null, "method": "GET", "mode": null, "owner": null, "password": null, "regexp": null, "remote_src": null, "removes": null, "return_content": false, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": null, "status_code": [200, 201, 202, 300, 301, 302, 303, 304, 307, 308, 400, 401, 403], "timeout": 30, "url": "https://drupal8-natbr0-container:12321/", "user": null, "validate_certs": false}, "module_name": "uri"}, "pragma": "no-cache", "redirected": false, "server": "MiniServ/1.780", "set_cookie": "testing=1; path=/; secure; httpOnly; httpOnly", "status": 200, "x_frame_options": "SAMEORIGIN"} ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> Results from version 2.0.2.0 ``` TASK [Test appliance webmin] *************************************************** task path: /home/ansible/playbooks/webmin-test.yml:139 <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r lxc '/bin/sh -c '"'"'mkdir -p "` echo ~/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525 `" && echo "` echo ~/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525 `"'"'"'' <lxc> PUT /tmp/tmppg6gtg TO /root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri <lxc> SSH: EXEC sftp -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r '[lxc]' <lxc> ESTABLISH SSH CONNECTION FOR USER: root <lxc> SSH: EXEC ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=22 -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o ControlPath=/home/ansible/.ansible/cp/ansible-ssh-%h-%p-%r -tt lxc '/bin/sh -c '"'"'LANG=C LC_ALL=C LC_MESSAGES=C /usr/bin/python /root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri; rm -rf "/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/" > /dev/null 2>&1'"'"'' fatal: [lxc]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "uri"}, "module_stderr": "OpenSSH_6.7p1 Debian-5+deb8u2, OpenSSL 1.0.1k 8 Jan 2015\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 15131\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to lxc closed.\r\n", "module_stdout": "Traceback (most recent call last):\r\n File \"/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri\", line 3310, in <module>\r\n main()\r\n File \"/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri\", line 407, in main\r\n dict_headers, socket_timeout)\r\n File \"/root/.ansible/tmp/ansible-tmp-1470679357.66-240132650115525/uri\", line 334, in uri\r\n content = resp.read()\r\n File \"/usr/lib/python2.7/socket.py\", line 351, in read\r\n data = self._sock.recv(rbufsize)\r\n File \"/usr/lib/python2.7/httplib.py\", line 573, in read\r\n s = self.fp.read(amt)\r\n File \"/usr/lib/python2.7/socket.py\", line 380, in read\r\n data = self._sock.recv(left)\r\n File \"/usr/lib/python2.7/ssl.py\", line 714, in recv\r\n return self.read(buflen)\r\n File \"/usr/lib/python2.7/ssl.py\", line 608, in read\r\n v = self._sslobj.read(len or 1024)\r\nssl.SSLError: ('The read operation timed out',)\r\n", "msg": "MODULE FAILURE", "parsed": false} ...ignoring ``` ##### COMMENTS I think the problem may be related to issue #3437 "Discrepancy in uri module behavior between 2.0.1.0 and 2.0.2.0", although I tried the fix suggested for #3437 and it did not help. The problem seems unique to Webmin, as several tests of other ports continue to function.
main
ansible and later break webmin port test issue type bug report component name uri module ansible version ansible config file etc ansible ansible cfg configured module search path home ansible library usr share ansible library configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables some basic default values hostfile etc ansible hosts library library usr share ansible library remote tmp ansible tmp pattern forks poll interval remote user root sudo user root ask sudo pass true ask pass true transport smart remote port module lang c os environment mention the os you are running ansible from and the os you are managing or say β€œn a” for anything that is not platform specific local ansible is running on a turnkey gnu linux ansible appliance version based on debian jessie remote lxc is running on a turnkey gnu linux lxc appliance version chifflier s ansible lxc ssh plugin is used to connect to containers summary ansible versions and later break a simple task used test a webmin port the task has worked from versions through steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used the failing task is part of a playbook designed to create an lxc container running a specified turnkey appliance and then run a series of tests that verify the appliance is functional the failing task uses the uri module to connect to the container s webmin port it is considered successful if one of the good response codes is returned ansible task that fails vars good codes tasks name test appliance webmin tags test uri url validate certs no status code good codes register webmin ignore errors yes until webmin success delay retries when app not in expected results results from version passed task task path home ansible playbooks webmin test yml establish ssh connection for user root ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home ansible ansible cp ansible ssh h p r tt lxc bin sh c mkdir p echo ansible tmp ansible tmp echo echo ansible tmp ansible tmp put tmp to root ansible tmp ansible tmp uri ssh exec sftp b c o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home ansible ansible cp ansible ssh h p r establish ssh connection for user root ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home ansible ansible cp ansible ssh h p r tt lxc bin sh c lang c lc all c lc messages c usr bin python root ansible tmp ansible tmp uri rm rf root ansible tmp ansible tmp dev null ok cache control no store no cache must revalidate post check pre check changed false connection close content location content security policy script src self unsafe inline unsafe eval frame src self child src self content type text html charset utf date mon aug gmt expires thu jan gmt invocation module args backup null body null body format raw content null creates null delimiter null dest null directory mode null follow false follow redirects safe force null force basic auth false group null method get mode null owner null password null regexp null remote src null removes null return content false selevel null serole null setype null seuser null src null status code timeout url user null validate certs false module name uri pragma no cache redirected false server miniserv set cookie testing path secure httponly httponly status x frame options sameorigin actual results results from version task task path home ansible playbooks webmin test yml establish ssh connection for user root ssh exec ssh c vvv o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home ansible ansible cp ansible ssh h p r lxc bin sh c mkdir p echo ansible tmp ansible tmp echo echo ansible tmp ansible tmp put tmp to root ansible tmp ansible tmp uri ssh exec sftp b c vvv o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home ansible ansible cp ansible ssh h p r establish ssh connection for user root ssh exec ssh c vvv o controlmaster auto o controlpersist o stricthostkeychecking no o port o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user root o connecttimeout o controlpath home ansible ansible cp ansible ssh h p r tt lxc bin sh c lang c lc all c lc messages c usr bin python root ansible tmp ansible tmp uri rm rf root ansible tmp ansible tmp dev null fatal failed changed false failed true invocation module name uri module stderr openssh debian openssl jan r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r mux client read packet read header failed broken pipe r received exit status from master r nshared connection to lxc closed r n module stdout traceback most recent call last r n file root ansible tmp ansible tmp uri line in r n main r n file root ansible tmp ansible tmp uri line in main r n dict headers socket timeout r n file root ansible tmp ansible tmp uri line in uri r n content resp read r n file usr lib socket py line in read r n data self sock recv rbufsize r n file usr lib httplib py line in read r n s self fp read amt r n file usr lib socket py line in read r n data self sock recv left r n file usr lib ssl py line in recv r n return self read buflen r n file usr lib ssl py line in read r n v self sslobj read len or r nssl sslerror the read operation timed out r n msg module failure parsed false ignoring comments i think the problem may be related to issue discrepancy in uri module behavior between and although i tried the fix suggested for and it did not help the problem seems unique to webmin as several tests of other ports continue to function
1
127,220
27,008,045,776
IssuesEvent
2023-02-10 13:19:15
vadimcn/vscode-lldb
https://api.github.com/repos/vadimcn/vscode-lldb
closed
vs-code URL does not open debugging session if URL is opened from inside VSCode
cause:VSCode
OS: macOS, remotely connected to Ubuntu 22.04 VSCode version: Version: 1.70.2; Commit: e4503b30fc78200f846c62cf8091b76ff5547662; Date: 2022-08-16T05:36:37.829Z CodeLLDB version: v1.7.4 Compiler: not relevant Debuggee: not relevant **Context** I want to use vscode-URLs to directly drop into a debugging session. My script can be found at https://gist.github.com/vogelsgesang/d803377b8b143164961e08e58f5186c0 **Expected behavior** When cmd-clicking this link from inside VSCode, I would expect to be dropped into the debugging session <img width="1287" alt="Screenshot 2022-08-22 at 10 25 59" src="https://user-images.githubusercontent.com/6820896/185875277-3f40bd9a-ccdc-4ac5-8a66-d7a07bd137c9.png"> **Actual behavior** Instead, the link gets opened in VS Code's search bar <img width="904" alt="Screenshot 2022-08-22 at 10 27 17" src="https://user-images.githubusercontent.com/6820896/185875587-cd4839e3-cf70-45ef-a3c4-977a237847c6.png"> **Additional information** I am sure the link itself is fine, because when I copy it from the console and paste it into Chrome's address bar, then VSCode opens my debugging session as expected.
1.0
vs-code URL does not open debugging session if URL is opened from inside VSCode - OS: macOS, remotely connected to Ubuntu 22.04 VSCode version: Version: 1.70.2; Commit: e4503b30fc78200f846c62cf8091b76ff5547662; Date: 2022-08-16T05:36:37.829Z CodeLLDB version: v1.7.4 Compiler: not relevant Debuggee: not relevant **Context** I want to use vscode-URLs to directly drop into a debugging session. My script can be found at https://gist.github.com/vogelsgesang/d803377b8b143164961e08e58f5186c0 **Expected behavior** When cmd-clicking this link from inside VSCode, I would expect to be dropped into the debugging session <img width="1287" alt="Screenshot 2022-08-22 at 10 25 59" src="https://user-images.githubusercontent.com/6820896/185875277-3f40bd9a-ccdc-4ac5-8a66-d7a07bd137c9.png"> **Actual behavior** Instead, the link gets opened in VS Code's search bar <img width="904" alt="Screenshot 2022-08-22 at 10 27 17" src="https://user-images.githubusercontent.com/6820896/185875587-cd4839e3-cf70-45ef-a3c4-977a237847c6.png"> **Additional information** I am sure the link itself is fine, because when I copy it from the console and paste it into Chrome's address bar, then VSCode opens my debugging session as expected.
non_main
vs code url does not open debugging session if url is opened from inside vscode os macos remotely connected to ubuntu vscode version version commit date codelldb version compiler not relevant debuggee not relevant context i want to use vscode urls to directly drop into a debugging session my script can be found at expected behavior when cmd clicking this link from inside vscode i would expect to be dropped into the debugging session img width alt screenshot at src actual behavior instead the link gets opened in vs code s search bar img width alt screenshot at src additional information i am sure the link itself is fine because when i copy it from the console and paste it into chrome s address bar then vscode opens my debugging session as expected
0
113,322
24,399,418,498
IssuesEvent
2022-10-04 22:59:55
fwouts/previewjs
https://api.github.com/repos/fwouts/previewjs
closed
Components with forwardRef are not rendered
bug fix merged vscode fix shipped
### Describe the bug Trying to a component build with forwardRef in Preview.js results in various errors messages. ### Reproduction (Using my example Repo as reference here) 1. Open `index.tsx` 2. Open Preview for `Input` component 3. See error in previewjs log ### Preview.js version v1.13.0 ### Framework React 18.2.0 ### System Info ```shell System: OS: Linux 5.15 Ubuntu 22.04.1 LTS 22.04.1 LTS (Jammy Jellyfish) CPU: (12) x64 Intel(R) Core(TM) i7-8750H CPU @ 2.20GHz Memory: 22.85 GB / 30.82 GB Container: Yes Shell: 5.8.1 - /bin/zsh Binaries: Node: 18.7.0 - ~/.nvm/versions/node/v18.7.0/bin/node Yarn: 1.22.19 - ~/.nvm/versions/node/v18.7.0/bin/yarn npm: 8.19.2 - ~/.nvm/versions/node/v18.7.0/bin/npm IDEs: Nano: 6.2 - /usr/bin/nano VSCode: 1.71.0 - /snap/bin/code Vim: 8.2 - /usr/bin/vim Browsers: Chrome: 105.0.5195.125 Firefox: 104.0.2 ``` ### Used Package Manager npm ### Extension logs (useful for crashes) _No response_ ### Preview logs (useful for rendering errors) ```shell [8:25:15 AM] Warning: Unexpected ref object provided for select. Use either a ref-setter function or React.createRef(). at select at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at SelectField2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26652:11) at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26664:19 at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at FormControl2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:16974:19) at FormControlWrapper (http://localhost:3140/preview/src/components/form-control-wrapper.tsx:10:3) at http://localhost:3140/preview/src/components/form-controls.tsx:116:5 at EnvironmentProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:20429:11) at ColorModeProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:1047:5) at ThemeProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3935:45) at ThemeProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:5515:11) at ChakraProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34899:5) at ChakraProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34923:3) at IntlProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/react-intl.js?v=34ee91d6:4087:43) at PreviewWrapper (http://localhost:3140/preview/src/preview-wrapper.jsx:6:27) at Renderer [8:25:15 AM] TypeError: Cannot add property current, object is not extensible at commitAttachRef (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:16823:27) at commitLayoutEffectOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:16693:17) at commitLayoutMountEffects_complete (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17503:17) at commitLayoutEffects_begin (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17492:15) at commitLayoutEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17444:11) at commitRootImpl (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18848:13) at commitRoot (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18772:13) at finishConcurrentRender (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18301:15) at performConcurrentWorkOnRoot (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18215:15) at workLoop (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:197:42) [8:25:15 AM] TypeError: Cannot add property current, object is not extensible at safelyDetachRef (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:16265:27) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17025:17) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17107:15) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17107:15) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17033:17) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17107:15) [8:25:15 AM] The above error occurred in the <select> component: at select at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at SelectField2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26652:11) at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26664:19 at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at FormControl2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:16974:19) at FormControlWrapper (http://localhost:3140/preview/src/components/form-control-wrapper.tsx:10:3) at http://localhost:3140/preview/src/components/form-controls.tsx:116:5 at EnvironmentProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:20429:11) at ColorModeProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:1047:5) at ThemeProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3935:45) at ThemeProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:5515:11) at ChakraProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34899:5) at ChakraProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34923:3) at IntlProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/react-intl.js?v=34ee91d6:4087:43) at PreviewWrapper (http://localhost:3140/preview/src/preview-wrapper.jsx:6:27) at Renderer Consider adding an error boundary to your tree to customize error handling behavior. Visit https://reactjs.org/link/error-boundaries to learn more about error boundaries. [8:25:15 AM] The above error occurred in the <Renderer> component: at Renderer Consider adding an error boundary to your tree to customize error handling behavior. Visit https://reactjs.org/link/error-boundaries to learn more about error boundaries. ``` ### Repo link (if available) https://github.com/trigo-at/previewjs-forwardref-error (See `Input` component in `index.tsx`) ### Anything else? _No response_
1.0
Components with forwardRef are not rendered - ### Describe the bug Trying to a component build with forwardRef in Preview.js results in various errors messages. ### Reproduction (Using my example Repo as reference here) 1. Open `index.tsx` 2. Open Preview for `Input` component 3. See error in previewjs log ### Preview.js version v1.13.0 ### Framework React 18.2.0 ### System Info ```shell System: OS: Linux 5.15 Ubuntu 22.04.1 LTS 22.04.1 LTS (Jammy Jellyfish) CPU: (12) x64 Intel(R) Core(TM) i7-8750H CPU @ 2.20GHz Memory: 22.85 GB / 30.82 GB Container: Yes Shell: 5.8.1 - /bin/zsh Binaries: Node: 18.7.0 - ~/.nvm/versions/node/v18.7.0/bin/node Yarn: 1.22.19 - ~/.nvm/versions/node/v18.7.0/bin/yarn npm: 8.19.2 - ~/.nvm/versions/node/v18.7.0/bin/npm IDEs: Nano: 6.2 - /usr/bin/nano VSCode: 1.71.0 - /snap/bin/code Vim: 8.2 - /usr/bin/vim Browsers: Chrome: 105.0.5195.125 Firefox: 104.0.2 ``` ### Used Package Manager npm ### Extension logs (useful for crashes) _No response_ ### Preview logs (useful for rendering errors) ```shell [8:25:15 AM] Warning: Unexpected ref object provided for select. Use either a ref-setter function or React.createRef(). at select at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at SelectField2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26652:11) at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26664:19 at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at FormControl2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:16974:19) at FormControlWrapper (http://localhost:3140/preview/src/components/form-control-wrapper.tsx:10:3) at http://localhost:3140/preview/src/components/form-controls.tsx:116:5 at EnvironmentProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:20429:11) at ColorModeProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:1047:5) at ThemeProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3935:45) at ThemeProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:5515:11) at ChakraProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34899:5) at ChakraProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34923:3) at IntlProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/react-intl.js?v=34ee91d6:4087:43) at PreviewWrapper (http://localhost:3140/preview/src/preview-wrapper.jsx:6:27) at Renderer [8:25:15 AM] TypeError: Cannot add property current, object is not extensible at commitAttachRef (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:16823:27) at commitLayoutEffectOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:16693:17) at commitLayoutMountEffects_complete (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17503:17) at commitLayoutEffects_begin (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17492:15) at commitLayoutEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17444:11) at commitRootImpl (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18848:13) at commitRoot (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18772:13) at finishConcurrentRender (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18301:15) at performConcurrentWorkOnRoot (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:18215:15) at workLoop (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:197:42) [8:25:15 AM] TypeError: Cannot add property current, object is not extensible at safelyDetachRef (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:16265:27) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17025:17) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17107:15) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17107:15) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17033:17) at recursivelyTraverseDeletionEffects (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17016:13) at commitDeletionEffectsOnFiber (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-LGNAA2TQ.js?v=34ee91d6:17107:15) [8:25:15 AM] The above error occurred in the <select> component: at select at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at SelectField2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26652:11) at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:26664:19 at div at http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3908:46 at FormControl2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:16974:19) at FormControlWrapper (http://localhost:3140/preview/src/components/form-control-wrapper.tsx:10:3) at http://localhost:3140/preview/src/components/form-controls.tsx:116:5 at EnvironmentProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:20429:11) at ColorModeProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:1047:5) at ThemeProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:3935:45) at ThemeProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:5515:11) at ChakraProvider (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34899:5) at ChakraProvider2 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/chunk-S4HW63PM.js?v=34ee91d6:34923:3) at IntlProvider3 (http://localhost:3140/preview/node_modules/.previewjs/v7.0.0/vite/deps/react-intl.js?v=34ee91d6:4087:43) at PreviewWrapper (http://localhost:3140/preview/src/preview-wrapper.jsx:6:27) at Renderer Consider adding an error boundary to your tree to customize error handling behavior. Visit https://reactjs.org/link/error-boundaries to learn more about error boundaries. [8:25:15 AM] The above error occurred in the <Renderer> component: at Renderer Consider adding an error boundary to your tree to customize error handling behavior. Visit https://reactjs.org/link/error-boundaries to learn more about error boundaries. ``` ### Repo link (if available) https://github.com/trigo-at/previewjs-forwardref-error (See `Input` component in `index.tsx`) ### Anything else? _No response_
non_main
components with forwardref are not rendered describe the bug trying to a component build with forwardref in preview js results in various errors messages reproduction using my example repo as reference here open index tsx open preview for input component see error in previewjs log preview js version framework react system info shell system os linux ubuntu lts lts jammy jellyfish cpu intel r core tm cpu memory gb gb container yes shell bin zsh binaries node nvm versions node bin node yarn nvm versions node bin yarn npm nvm versions node bin npm ides nano usr bin nano vscode snap bin code vim usr bin vim browsers chrome firefox used package manager npm extension logs useful for crashes no response preview logs useful for rendering errors shell warning unexpected ref object provided for select use either a ref setter function or react createref at select at at at div at at at div at at at formcontrolwrapper at at environmentprovider at colormodeprovider at at at chakraprovider at at at previewwrapper at renderer typeerror cannot add property current object is not extensible at commitattachref at commitlayouteffectonfiber at commitlayoutmounteffects complete at commitlayouteffects begin at commitlayouteffects at commitrootimpl at commitroot at finishconcurrentrender at performconcurrentworkonroot at workloop typeerror cannot add property current object is not extensible at safelydetachref at commitdeletioneffectsonfiber at recursivelytraversedeletioneffects at commitdeletioneffectsonfiber at recursivelytraversedeletioneffects at commitdeletioneffectsonfiber at recursivelytraversedeletioneffects at commitdeletioneffectsonfiber at recursivelytraversedeletioneffects at commitdeletioneffectsonfiber the above error occurred in the component at select at at at div at at at div at at at formcontrolwrapper at at environmentprovider at colormodeprovider at at at chakraprovider at at at previewwrapper at renderer consider adding an error boundary to your tree to customize error handling behavior visit to learn more about error boundaries the above error occurred in the component at renderer consider adding an error boundary to your tree to customize error handling behavior visit to learn more about error boundaries repo link if available see input component in index tsx anything else no response
0
5,476
27,363,850,307
IssuesEvent
2023-02-27 17:37:51
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Bug: samconfig.toml values don't override default values in sam template
blocked/more-info-needed blocked/close-if-inactive maintainer/need-followup
### Description: Running a .NET Core 3.1 project containing Lambdas, trying to deploy using sam build and sam deploy. Build is successful, but Deploy fails in AWS because parameter overrides specified in samconfig.toml are not used, even though log for both sam build and sam deploy states that they are found correctly. ### Steps to reproduce: Running ADO pipeline (attached), deploying SAM template (attached) with samconfig.toml. ### Observed result: The template generated by sam build has original default values specified for parameters, and they are the ones used during deployment, even though log says : for sam build: Loading configuration values from [uat.['build'].parameters] (env.command_name.section) in config file at '/workspace/1/s/Main/Source/Services/samconfig.toml' and similar for sam deploy ### Expected result: the template to be deployed with parameters specified in [uat.global.parameters] section in samconfig.toml ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) OS: Lambda in AWS sam --version: Tool 'amazon.lambda.tools' (version '5.6.2') AWS region: us-west-2 Attached are logs for both Build and Deploy steps as well as the original template, generated one and samconfig.toml (in templatesAndSamCfg.zip) [Build-Log.txt](https://github.com/aws/aws-sam-cli/files/10406008/Build-Log.txt) [Deploy-Log.txt](https://github.com/aws/aws-sam-cli/files/10406009/Deploy-Log.txt) [samconfig.zip](https://github.com/aws/aws-sam-cli/files/10406011/samconfig.zip) [templatesAndSamCfg.zip](https://github.com/aws/aws-sam-cli/files/10406018/templatesAndSamCfg.zip)
True
Bug: samconfig.toml values don't override default values in sam template - ### Description: Running a .NET Core 3.1 project containing Lambdas, trying to deploy using sam build and sam deploy. Build is successful, but Deploy fails in AWS because parameter overrides specified in samconfig.toml are not used, even though log for both sam build and sam deploy states that they are found correctly. ### Steps to reproduce: Running ADO pipeline (attached), deploying SAM template (attached) with samconfig.toml. ### Observed result: The template generated by sam build has original default values specified for parameters, and they are the ones used during deployment, even though log says : for sam build: Loading configuration values from [uat.['build'].parameters] (env.command_name.section) in config file at '/workspace/1/s/Main/Source/Services/samconfig.toml' and similar for sam deploy ### Expected result: the template to be deployed with parameters specified in [uat.global.parameters] section in samconfig.toml ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) OS: Lambda in AWS sam --version: Tool 'amazon.lambda.tools' (version '5.6.2') AWS region: us-west-2 Attached are logs for both Build and Deploy steps as well as the original template, generated one and samconfig.toml (in templatesAndSamCfg.zip) [Build-Log.txt](https://github.com/aws/aws-sam-cli/files/10406008/Build-Log.txt) [Deploy-Log.txt](https://github.com/aws/aws-sam-cli/files/10406009/Deploy-Log.txt) [samconfig.zip](https://github.com/aws/aws-sam-cli/files/10406011/samconfig.zip) [templatesAndSamCfg.zip](https://github.com/aws/aws-sam-cli/files/10406018/templatesAndSamCfg.zip)
main
bug samconfig toml values don t override default values in sam template description running a net core project containing lambdas trying to deploy using sam build and sam deploy build is successful but deploy fails in aws because parameter overrides specified in samconfig toml are not used even though log for both sam build and sam deploy states that they are found correctly steps to reproduce running ado pipeline attached deploying sam template attached with samconfig toml observed result the template generated by sam build has original default values specified for parameters and they are the ones used during deployment even though log says for sam build loading configuration values from parameters env command name section in config file at workspace s main source services samconfig toml and similar for sam deploy expected result the template to be deployed with parameters specified in section in samconfig toml additional environment details ex windows mac amazon linux etc os lambda in aws sam version tool amazon lambda tools version aws region us west attached are logs for both build and deploy steps as well as the original template generated one and samconfig toml in templatesandsamcfg zip
1
14,953
2,832,206,518
IssuesEvent
2015-05-25 05:26:40
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Footer moves down with Horizontal Scrolling and Selection
defect
In Internet Explorer 9 (IE9), with a DataTable if we enable Row Selection (single or multiple) and we enable scrollable with a scrollWidth small enough to cause horizontal scrollbar, the footer of the DataTable moves down the page whenever we mouseover the data rows. The more you move the mouse on mouseover, the more it moves down. Mouse movement can be either side-to-side or up-and-down. Clicking on an item in the list will reset the footer back to normal. This is reproducible by taking the PrimeFaces ShowCase application and modifying either ui/datatableRowSelectionSingle.xhtml or ui/datatableRowSelectionMultiple.xhtml by adding the following attributes to the <p:dataTable> tag: scrollable="true" scrollWidth="250" Note that this appears to only occur in IE9. IE8 (or IE8 compatibility mode) and IE10 do not exhibit this behavior.
1.0
Footer moves down with Horizontal Scrolling and Selection - In Internet Explorer 9 (IE9), with a DataTable if we enable Row Selection (single or multiple) and we enable scrollable with a scrollWidth small enough to cause horizontal scrollbar, the footer of the DataTable moves down the page whenever we mouseover the data rows. The more you move the mouse on mouseover, the more it moves down. Mouse movement can be either side-to-side or up-and-down. Clicking on an item in the list will reset the footer back to normal. This is reproducible by taking the PrimeFaces ShowCase application and modifying either ui/datatableRowSelectionSingle.xhtml or ui/datatableRowSelectionMultiple.xhtml by adding the following attributes to the <p:dataTable> tag: scrollable="true" scrollWidth="250" Note that this appears to only occur in IE9. IE8 (or IE8 compatibility mode) and IE10 do not exhibit this behavior.
non_main
footer moves down with horizontal scrolling and selection in internet explorer with a datatable if we enable row selection single or multiple and we enable scrollable with a scrollwidth small enough to cause horizontal scrollbar the footer of the datatable moves down the page whenever we mouseover the data rows the more you move the mouse on mouseover the more it moves down mouse movement can be either side to side or up and down clicking on an item in the list will reset the footer back to normal this is reproducible by taking the primefaces showcase application and modifying either ui datatablerowselectionsingle xhtml or ui datatablerowselectionmultiple xhtml by adding the following attributes to the tag scrollable true scrollwidth note that this appears to only occur in or compatibility mode and do not exhibit this behavior
0
3,059
2,659,925,850
IssuesEvent
2015-03-19 00:37:36
TeamMentor/TM_4_0_Design
https://api.github.com/repos/TeamMentor/TM_4_0_Design
opened
Minor fix to new PoC page filters-with-empty-results.jade
Action:Design Effort: Taster - 4h Type:Bug
@luchia can you fix the minor css issue with the https://github.com/TeamMentor/TM_4_0_Design/blob/Issue_472_Multiple_Filters/source/jade/__poc/filters/filters-with-empty-results.jade page You can see the page in action if you check out the **Issue_472_Multiple_Filters** branch (on both TM_Design and TM_Graph), and open this page http://localhost:1337/poc/filters-with-empty-results/query-847ac915eb71,query-fd50a62b22c4 ![image](https://cloud.githubusercontent.com/assets/656739/6722334/f2e6e7a6-cdcf-11e4-8793-1c1c7a2721bb.png) the prob is the extra spacing on the right-hand side filters. The left-hand side looks good @roman87 this is that PoC page that you asked for (the one that shows the filters with zero results)
1.0
Minor fix to new PoC page filters-with-empty-results.jade - @luchia can you fix the minor css issue with the https://github.com/TeamMentor/TM_4_0_Design/blob/Issue_472_Multiple_Filters/source/jade/__poc/filters/filters-with-empty-results.jade page You can see the page in action if you check out the **Issue_472_Multiple_Filters** branch (on both TM_Design and TM_Graph), and open this page http://localhost:1337/poc/filters-with-empty-results/query-847ac915eb71,query-fd50a62b22c4 ![image](https://cloud.githubusercontent.com/assets/656739/6722334/f2e6e7a6-cdcf-11e4-8793-1c1c7a2721bb.png) the prob is the extra spacing on the right-hand side filters. The left-hand side looks good @roman87 this is that PoC page that you asked for (the one that shows the filters with zero results)
non_main
minor fix to new poc page filters with empty results jade luchia can you fix the minor css issue with the page you can see the page in action if you check out the issue multiple filters branch on both tm design and tm graph and open this page the prob is the extra spacing on the right hand side filters the left hand side looks good this is that poc page that you asked for the one that shows the filters with zero results
0
3,738
15,691,790,993
IssuesEvent
2021-03-25 18:18:07
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
closed
Cleanup: remove unsupported defines or negative-defines: NOLIBC, DCONTEXT_IN_EDI, etc.
Maintainability
There are several defines we can eliminate: DCONTEXT_IN_EDI: Remove all its code. NOLIBC: Remove all the negative code. Xref #4045 on unsupported API define removal Xref #2971 on supported API define removal
True
Cleanup: remove unsupported defines or negative-defines: NOLIBC, DCONTEXT_IN_EDI, etc. - There are several defines we can eliminate: DCONTEXT_IN_EDI: Remove all its code. NOLIBC: Remove all the negative code. Xref #4045 on unsupported API define removal Xref #2971 on supported API define removal
main
cleanup remove unsupported defines or negative defines nolibc dcontext in edi etc there are several defines we can eliminate dcontext in edi remove all its code nolibc remove all the negative code xref on unsupported api define removal xref on supported api define removal
1
2,153
7,479,225,998
IssuesEvent
2018-04-04 14:05:30
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Chemdisp recipes/macros quietly fail to get added if list of chems ends with ; instead of nothing (just amount of last chemical)
Maintainability/Hinders improvements
Issue reported from Round ID: 85939 (/tg/Station Sybil [ENGLISH] [US-EAST] [100% LAG FREE]) Reporting client version: 512
True
Chemdisp recipes/macros quietly fail to get added if list of chems ends with ; instead of nothing (just amount of last chemical) - Issue reported from Round ID: 85939 (/tg/Station Sybil [ENGLISH] [US-EAST] [100% LAG FREE]) Reporting client version: 512
main
chemdisp recipes macros quietly fail to get added if list of chems ends with instead of nothing just amount of last chemical issue reported from round id tg station sybil reporting client version
1
295,846
22,274,413,065
IssuesEvent
2022-06-10 15:12:22
AzureAD/microsoft-identity-web
https://api.github.com/repos/AzureAD/microsoft-identity-web
opened
[Documentation] No example for a background service calling a Web API
documentation
### Documentation related to component Microsoft.Identity.Web ### Please check all that apply - [ ] typo - [ ] documentation doesn't exist - [ ] documentation needs clarification - [ ] error(s) in the example - [x] needs an example ### Description of the issue There is an example for a Web API calling a downstream Web API on the users behalf. However an example for a Background Service running in an ASP.Net Core Host calling a Web API is missing. When a Web API calls another Web API, it can just pass on the token, it received. However in a long running Background Service, the token has to be aquired for the App itself (using client credentials or a certificate) and has to be kept up to date, because it might expire. The token has to be added to all/some (maybe named) HttpClients using IHttpClientFactory. Is there an example, that I'm unable to find, or is this example missing? The closest, I could find is a deamon console app which calls a Web API. However that is not using the ASP.net Core DI ccontainer and does not automatically renew the token when it is about to expire. https://github.com/Azure-Samples/active-directory-dotnetcore-daemon-v2/
1.0
[Documentation] No example for a background service calling a Web API - ### Documentation related to component Microsoft.Identity.Web ### Please check all that apply - [ ] typo - [ ] documentation doesn't exist - [ ] documentation needs clarification - [ ] error(s) in the example - [x] needs an example ### Description of the issue There is an example for a Web API calling a downstream Web API on the users behalf. However an example for a Background Service running in an ASP.Net Core Host calling a Web API is missing. When a Web API calls another Web API, it can just pass on the token, it received. However in a long running Background Service, the token has to be aquired for the App itself (using client credentials or a certificate) and has to be kept up to date, because it might expire. The token has to be added to all/some (maybe named) HttpClients using IHttpClientFactory. Is there an example, that I'm unable to find, or is this example missing? The closest, I could find is a deamon console app which calls a Web API. However that is not using the ASP.net Core DI ccontainer and does not automatically renew the token when it is about to expire. https://github.com/Azure-Samples/active-directory-dotnetcore-daemon-v2/
non_main
no example for a background service calling a web api documentation related to component microsoft identity web please check all that apply typo documentation doesn t exist documentation needs clarification error s in the example needs an example description of the issue there is an example for a web api calling a downstream web api on the users behalf however an example for a background service running in an asp net core host calling a web api is missing when a web api calls another web api it can just pass on the token it received however in a long running background service the token has to be aquired for the app itself using client credentials or a certificate and has to be kept up to date because it might expire the token has to be added to all some maybe named httpclients using ihttpclientfactory is there an example that i m unable to find or is this example missing the closest i could find is a deamon console app which calls a web api however that is not using the asp net core di ccontainer and does not automatically renew the token when it is about to expire
0
1,243
5,308,973,653
IssuesEvent
2017-02-12 03:59:45
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
Set notes and resource_pool on vmware_guest
affects_2.1 cloud feature_idea vmware waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME vmware_guest ##### ANSIBLE VERSION ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ``` ##### OS / ENVIRONMENT CentOS Linux release 7.2.1511 (Core) ##### SUMMARY Set notes and resource_pool ##### STEPS TO REPRODUCE ``` - name: Create VM vmware_guest: validate_certs: False hostname: "{{ vsphere_vcenter_hostname }}" username: "{{ vsphere_vcenter_username }}" password: "{{ vsphere_vcenter_password }}" name: "{{ inventory_hostname }}" state: poweredoff folder: "{{ vsphere_vm_folder }}" disk: - size_gb: 16 type: thin datastore: "{{ vsphere_datastore_name }}" nic: - type: vmxnet3 network: DMZ network_type: dvs datacenter: "{{ vsphere_datacenter_name }}" esxi_hostname: "{{ vsphere_vcenter_password }}" template: "{{ vsphere_template_name }}" wait_for_ip_address: False register: deploy ``` where set notes and resource_pool? ##### EXPECTED RESULTS is set the pool resource and a description for notes. ##### ACTUAL RESULTS data not septate ``` ```
True
Set notes and resource_pool on vmware_guest - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME vmware_guest ##### ANSIBLE VERSION ansible 2.1.1.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ``` ##### OS / ENVIRONMENT CentOS Linux release 7.2.1511 (Core) ##### SUMMARY Set notes and resource_pool ##### STEPS TO REPRODUCE ``` - name: Create VM vmware_guest: validate_certs: False hostname: "{{ vsphere_vcenter_hostname }}" username: "{{ vsphere_vcenter_username }}" password: "{{ vsphere_vcenter_password }}" name: "{{ inventory_hostname }}" state: poweredoff folder: "{{ vsphere_vm_folder }}" disk: - size_gb: 16 type: thin datastore: "{{ vsphere_datastore_name }}" nic: - type: vmxnet3 network: DMZ network_type: dvs datacenter: "{{ vsphere_datacenter_name }}" esxi_hostname: "{{ vsphere_vcenter_password }}" template: "{{ vsphere_template_name }}" wait_for_ip_address: False register: deploy ``` where set notes and resource_pool? ##### EXPECTED RESULTS is set the pool resource and a description for notes. ##### ACTUAL RESULTS data not septate ``` ```
main
set notes and resource pool on vmware guest issue type feature idea component name vmware guest ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides os environment centos linux release core summary set notes and resource pool steps to reproduce name create vm vmware guest validate certs false hostname vsphere vcenter hostname username vsphere vcenter username password vsphere vcenter password name inventory hostname state poweredoff folder vsphere vm folder disk size gb type thin datastore vsphere datastore name nic type network dmz network type dvs datacenter vsphere datacenter name esxi hostname vsphere vcenter password template vsphere template name wait for ip address false register deploy where set notes and resource pool expected results is set the pool resource and a description for notes actual results data not septate
1
1,834
6,577,362,926
IssuesEvent
2017-09-12 00:23:14
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2 module tracebacks on cred failure
affects_2.0 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2 module ##### ANSIBLE VERSION ``` ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = /usr/lib/python2.7/site-packages/awx/plugins/library ``` ##### CONFIGURATION Nothing of note. ##### OS / ENVIRONMENT CentOS 7. ##### SUMMARY ec2 module tracebacks when using an invalidated credential. ##### STEPS TO REPRODUCE 1. Have a basic provisioning task: ``` - hosts: localhost connection: local gather_facts: False vars_files: - group_vars/all tasks: - name: Launch some instances ec2: > access_key="{{ ec2_access_key }}" secret_key="{{ ec2_secret_key }}" keypair="{{ ec2_keypair }}" group="{{ ec2_security_group }}" type="{{ ec2_instance_type }}" image="{{ ec2_image }}" region="{{ ec2_region }}" instance_tags="{'type':'{{ ec2_instance_type }}', 'group':'{{ ec2_security_group }}', 'Name':'demo_''{{ demo_tag_name }}'}" count="{{ ec2_instance_count }}" wait=true register: ec2 ``` 1. Have a AWS key that's been disabled. 2. Run the playbook ##### EXPECTED RESULTS Clean failure with a 'permission denied' error. ##### ACTUAL RESULTS ``` TASK [Launch some instances] *************************************************** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: <Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>0d83c70d-523b-49ab-9a48-0a1d8eca26e6</RequestID></Response> fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last): File \"/var/lib/awx/.ansible/tmp/ansible-tmp-1464811193.49-144253552378267/ec2\", line 3628, in <module> main() File \"/var/lib/awx/.ansible/tmp/ansible-tmp-1464811193.49-144253552378267/ec2\", line 1413, in main (instance_dict_array, new_instance_ids, changed) = create_instances(module, ec2, vpc) File \"/var/lib/awx/.ansible/tmp/ansible-tmp-1464811193.49-144253552378267/ec2\", line 898, in create_instances grp_details = ec2.get_all_security_groups() File \"/var/lib/awx/venv/ansible/lib/python2.7/site-packages/boto/ec2/connection.py\", line 2969, in get_all_security_groups [('item', SecurityGroup)], verb='POST') File \"/var/lib/awx/venv/ansible/lib/python2.7/site-packages/boto/connection.py\", line 1182, in get_list raise self.ResponseError(response.status, response.reason, body) boto.exception.EC2ResponseError: EC2ResponseError: 401 Unauthorized <?xml version=\"1.0\" encoding=\"UTF-8\"?> <Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>0d83c70d-523b-49ab-9a48-0a1d8eca26e6</RequestID></Response> ", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false} ```
True
ec2 module tracebacks on cred failure - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2 module ##### ANSIBLE VERSION ``` ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = /usr/lib/python2.7/site-packages/awx/plugins/library ``` ##### CONFIGURATION Nothing of note. ##### OS / ENVIRONMENT CentOS 7. ##### SUMMARY ec2 module tracebacks when using an invalidated credential. ##### STEPS TO REPRODUCE 1. Have a basic provisioning task: ``` - hosts: localhost connection: local gather_facts: False vars_files: - group_vars/all tasks: - name: Launch some instances ec2: > access_key="{{ ec2_access_key }}" secret_key="{{ ec2_secret_key }}" keypair="{{ ec2_keypair }}" group="{{ ec2_security_group }}" type="{{ ec2_instance_type }}" image="{{ ec2_image }}" region="{{ ec2_region }}" instance_tags="{'type':'{{ ec2_instance_type }}', 'group':'{{ ec2_security_group }}', 'Name':'demo_''{{ demo_tag_name }}'}" count="{{ ec2_instance_count }}" wait=true register: ec2 ``` 1. Have a AWS key that's been disabled. 2. Run the playbook ##### EXPECTED RESULTS Clean failure with a 'permission denied' error. ##### ACTUAL RESULTS ``` TASK [Launch some instances] *************************************************** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: <Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>0d83c70d-523b-49ab-9a48-0a1d8eca26e6</RequestID></Response> fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last): File \"/var/lib/awx/.ansible/tmp/ansible-tmp-1464811193.49-144253552378267/ec2\", line 3628, in <module> main() File \"/var/lib/awx/.ansible/tmp/ansible-tmp-1464811193.49-144253552378267/ec2\", line 1413, in main (instance_dict_array, new_instance_ids, changed) = create_instances(module, ec2, vpc) File \"/var/lib/awx/.ansible/tmp/ansible-tmp-1464811193.49-144253552378267/ec2\", line 898, in create_instances grp_details = ec2.get_all_security_groups() File \"/var/lib/awx/venv/ansible/lib/python2.7/site-packages/boto/ec2/connection.py\", line 2969, in get_all_security_groups [('item', SecurityGroup)], verb='POST') File \"/var/lib/awx/venv/ansible/lib/python2.7/site-packages/boto/connection.py\", line 1182, in get_list raise self.ResponseError(response.status, response.reason, body) boto.exception.EC2ResponseError: EC2ResponseError: 401 Unauthorized <?xml version=\"1.0\" encoding=\"UTF-8\"?> <Response><Errors><Error><Code>AuthFailure</Code><Message>AWS was not able to validate the provided access credentials</Message></Error></Errors><RequestID>0d83c70d-523b-49ab-9a48-0a1d8eca26e6</RequestID></Response> ", "module_stdout": "", "msg": "MODULE FAILURE", "parsed": false} ```
main
module tracebacks on cred failure issue type bug report component name module ansible version ansible config file etc ansible ansible cfg configured module search path usr lib site packages awx plugins library configuration nothing of note os environment centos summary module tracebacks when using an invalidated credential steps to reproduce have a basic provisioning task hosts localhost connection local gather facts false vars files group vars all tasks name launch some instances access key access key secret key secret key keypair keypair group security group type instance type image image region region instance tags type instance type group security group name demo demo tag name count instance count wait true register have a aws key that s been disabled run the playbook expected results clean failure with a permission denied error actual results task an exception occurred during task execution to see the full traceback use vvv the error was authfailure aws was not able to validate the provided access credentials fatal failed changed false failed true module stderr traceback most recent call last file var lib awx ansible tmp ansible tmp line in main file var lib awx ansible tmp ansible tmp line in main instance dict array new instance ids changed create instances module vpc file var lib awx ansible tmp ansible tmp line in create instances grp details get all security groups file var lib awx venv ansible lib site packages boto connection py line in get all security groups verb post file var lib awx venv ansible lib site packages boto connection py line in get list raise self responseerror response status response reason body boto exception unauthorized authfailure aws was not able to validate the provided access credentials module stdout msg module failure parsed false
1
5,005
25,732,911,480
IssuesEvent
2022-12-07 21:49:07
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Question]: Is it possible to still use FilterableMultiSelectClassic?
type: question ❓ status: waiting for maintainer response πŸ’¬
### Question for Carbon Hey Team πŸ‘‹ , I'm trying to use the `FilterableMultiSelectClassic` in the latest release of @carbon/react. Based on this line. it should be still available thought `Multiselect.Filterable`: https://github.com/carbon-design-system/carbon/blob/main/packages/react/src/components/MultiSelect/index.js#L16 But I'm getting `undefined`. Any help is appreciated, thanks in advance! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
True
[Question]: Is it possible to still use FilterableMultiSelectClassic? - ### Question for Carbon Hey Team πŸ‘‹ , I'm trying to use the `FilterableMultiSelectClassic` in the latest release of @carbon/react. Based on this line. it should be still available thought `Multiselect.Filterable`: https://github.com/carbon-design-system/carbon/blob/main/packages/react/src/components/MultiSelect/index.js#L16 But I'm getting `undefined`. Any help is appreciated, thanks in advance! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
main
is it possible to still use filterablemultiselectclassic question for carbon hey team πŸ‘‹ i m trying to use the filterablemultiselectclassic in the latest release of carbon react based on this line it should be still available thought multiselect filterable but i m getting undefined any help is appreciated thanks in advance code of conduct i agree to follow this project s
1
22,696
32,006,392,701
IssuesEvent
2023-09-21 15:04:40
googleapis/sdk-platform-java
https://api.github.com/repos/googleapis/sdk-platform-java
closed
Your .repo-metadata.json file has a problem πŸ€’
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan πŸ“ˆ: * must have required property 'release_level' in gax-java/.repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem πŸ€’ - You have a problem with your .repo-metadata.json file: Result of scan πŸ“ˆ: * must have required property 'release_level' in gax-java/.repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
non_main
your repo metadata json file has a problem πŸ€’ you have a problem with your repo metadata json file result of scan πŸ“ˆ must have required property release level in gax java repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
0
1,855
6,577,401,795
IssuesEvent
2017-09-12 00:39:39
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Feature request: Support instance protection settings for EC2 Auto Scaling Groups in ec2_asg
affects_2.0 aws cloud feature_idea waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME ec2_asg module ##### ANSIBLE VERSION > 2.0 ##### SUMMARY Would like the ec2_asg module to support instance protection settings. https://aws.amazon.com/blogs/aws/new-instance-protection-for-auto-scaling
True
Feature request: Support instance protection settings for EC2 Auto Scaling Groups in ec2_asg - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME ec2_asg module ##### ANSIBLE VERSION > 2.0 ##### SUMMARY Would like the ec2_asg module to support instance protection settings. https://aws.amazon.com/blogs/aws/new-instance-protection-for-auto-scaling
main
feature request support instance protection settings for auto scaling groups in asg issue type feature idea component name asg module ansible version summary would like the asg module to support instance protection settings
1
1,272
5,395,004,314
IssuesEvent
2017-02-27 06:41:51
shawnlaffan/biodiverse
https://api.github.com/repos/shawnlaffan/biodiverse
closed
Add travis CI support
Maintainability
Should be able to adapt from https://github.com/PDLPorters/pdla-stats/blob/master/.travis.yml To do list: - [x] Remove Gtk2::GladeXML from cpanfile - dependency was removed under #413 but Task files were not updated. - [x] Some means of caching gdal in the travis builds, or implement apt-get gdal-dev from ubuntugis-stable (if it exists) - [x] call cpanfile with and without the GUI option (we just tickled one latent bug with it not being called) - probably only needed for only one perl version. - [x] shift (most of) the gdal build steps into a shell script. https://docs.travis-ci.com/user/customizing-the-build#Implementing-Complex-Build-Steps
True
Add travis CI support - Should be able to adapt from https://github.com/PDLPorters/pdla-stats/blob/master/.travis.yml To do list: - [x] Remove Gtk2::GladeXML from cpanfile - dependency was removed under #413 but Task files were not updated. - [x] Some means of caching gdal in the travis builds, or implement apt-get gdal-dev from ubuntugis-stable (if it exists) - [x] call cpanfile with and without the GUI option (we just tickled one latent bug with it not being called) - probably only needed for only one perl version. - [x] shift (most of) the gdal build steps into a shell script. https://docs.travis-ci.com/user/customizing-the-build#Implementing-Complex-Build-Steps
main
add travis ci support should be able to adapt from to do list remove gladexml from cpanfile dependency was removed under but task files were not updated some means of caching gdal in the travis builds or implement apt get gdal dev from ubuntugis stable if it exists call cpanfile with and without the gui option we just tickled one latent bug with it not being called probably only needed for only one perl version shift most of the gdal build steps into a shell script
1
1,457
6,304,450,451
IssuesEvent
2017-07-21 15:55:46
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
meta: reset_connection results in "IndexError: list index out of range"
affects_2.3 bug_report module needs_maintainer python3 support:core
<!--- Verify first that your issue/request is not already reported on GitHub. Also test if the latest release, and master branch are affected too. --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME meta module ##### ANSIBLE VERSION ``` ansible-playbook 2.3.1.0 config file = configured module search path = Default w/o overrides python version = 3.5.2 (default, Nov 17 2016, 17:05:23) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION No configuration. ##### OS / ENVIRONMENT Ubuntu 16.04 with ansible pip installed, running in python3.5. Contacting a VM running CentOS 6 ##### SUMMARY Using `meta: reset_connection` works and does reset the connection, but then later on this IndexError ocurrs. ##### STEPS TO REPRODUCE This playbook shows the problem: ``` - name: Ping them hosts: all tasks: - name: stuff debug: msg: Here is a message - name: reset conn meta: reset_connection ``` ##### EXPECTED RESULTS ansible-playbook should exit 0 without errors ##### ACTUAL RESULTS ~/tmp/inv has this content: ``` spamaps.org ``` My user has SSH agent access to spamaps.org. ``` $ ansible-playbook test.yml -i ~/tmp/inv -vvvvv No config file found; using defaults Loading callback plugin default of type stdout, v2.0 from /usr/local/lib/python3.5/dist-packages/ansible/plugins/callback/__init__.py PLAYBOOK: test.yml ************************************************************************************************************************************************************************************************* 1 plays in test.yml PLAY [Ping them] *************************************************************************************************************************************************************************************************** TASK [Gathering Facts] ********************************************************************************************************************************************************************************************* Using module file /usr/local/lib/python3.5/dist-packages/ansible/modules/system/setup.py <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 spamaps.org '/bin/sh -c '"'"'echo ~ && sleep 0'"'"'' <spamaps.org> (0, b'/home/clint\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug1: Control socket "/home/clint/.ansible/cp/80e720ccb6" does not exist\r\ndebug2: resolving "spamaps.org" port 22\r\ndebug2: ssh_connect_direct: needpriv 0\r\ndebug1: Connecting to spamaps.org [184.105.137.237] port 22.\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug1: fd 3 clearing O_NONBLOCK\r\ndebug1: Connection established.\r\ndebug3: timeout: 9959 ms remain after connect\r\ndebug1: identity file /home/clint/.ssh/id_rsa type 1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_rsa-cert type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_dsa type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_dsa-cert type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ecdsa type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ecdsa-cert type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ed25519 type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ed25519-cert type -1\r\ndebug1: Enabling compatibility mode for protocol 2.0\r\ndebug1: Local version string SSH-2.0-OpenSSH_7.2p2 Ubuntu-4ubuntu2.2\r\ndebug1: Remote protocol version 2.0, remote software version OpenSSH_5.3\r\ndebug1: match: OpenSSH_5.3 pat OpenSSH_5* compat 0x0c000000\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug1: Authenticating to spamaps.org:22 as \'clint\'\r\ndebug3: hostkeys_foreach: reading file "/home/clint/.ssh/known_hosts"\r\ndebug3: record_hostkey: found key type RSA in file /home/clint/.ssh/known_hosts:1\r\ndebug3: load_hostkeys: loaded 1 keys from spamaps.org\r\ndebug3: order_hostkeyalgs: prefer hostkeyalgs: ssh-rsa-cert-v01@openssh.com,rsa-sha2-512,rsa-sha2-256,ssh-rsa\r\ndebug3: send packet: type 20\r\ndebug1: SSH2_MSG_KEXINIT sent\r\ndebug3: receive packet: type 20\r\ndebug1: SSH2_MSG_KEXINIT received\r\ndebug2: local client KEXINIT proposal\r\ndebug2: KEX algorithms: curve25519-sha256@libssh.org,ecdh-sha2-nistp256,ecdh-sha2-nistp384,ecdh-sha2-nistp521,diffie-hellman-group-exchange-sha256,diffie-hellman-group-exchange-sha1,diffie-hellman-group14-sha1,ext-info-c\r\ndebug2: host key algorithms: ssh-rsa-cert-v01@openssh.com,rsa-sha2-512,rsa-sha2-256,ssh-rsa,ecdsa-sha2-nistp256-cert-v01@openssh.com,ecdsa-sha2-nistp384-cert-v01@openssh.com,ecdsa-sha2-nistp521-cert-v01@openssh.com,ssh-ed25519-cert-v01@openssh.com,ecdsa-sha2-nistp256,ecdsa-sha2-nistp384,ecdsa-sha2-nistp521,ssh-ed25519\r\ndebug2: ciphers ctos: chacha20-poly1305@openssh.com,aes128-ctr,aes192-ctr,aes256-ctr,aes128-gcm@openssh.com,aes256-gcm@openssh.com,aes128-cbc,aes192-cbc,aes256-cbc,3des-cbc\r\ndebug2: ciphers stoc: chacha20-poly1305@openssh.com,aes128-ctr,aes192-ctr,aes256-ctr,aes128-gcm@openssh.com,aes256-gcm@openssh.com,aes128-cbc,aes192-cbc,aes256-cbc,3des-cbc\r\ndebug2: MACs ctos: umac-64-etm@openssh.com,umac-128-etm@openssh.com,hmac-sha2-256-etm@openssh.com,hmac-sha2-512-etm@openssh.com,hmac-sha1-etm@openssh.com,umac-64@openssh.com,umac-128@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-sha1\r\ndebug2: MACs stoc: umac-64-etm@openssh.com,umac-128-etm@openssh.com,hmac-sha2-256-etm@openssh.com,hmac-sha2-512-etm@openssh.com,hmac-sha1-etm@openssh.com,umac-64@openssh.com,umac-128@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-sha1\r\ndebug2: compression ctos: zlib@openssh.com,zlib,none\r\ndebug2: compression stoc: zlib@openssh.com,zlib,none\r\ndebug2: languages ctos: \r\ndebug2: languages stoc: \r\ndebug2: first_kex_follows 0 \r\ndebug2: reserved 0 \r\ndebug2: peer server KEXINIT proposal\r\ndebug2: KEX algorithms: diffie-hellman-group-exchange-sha256,diffie-hellman-group-exchange-sha1,diffie-hellman-group14-sha1,diffie-hellman-group1-sha1\r\ndebug2: host key algorithms: ssh-rsa,ssh-dss\r\ndebug2: ciphers ctos: aes128-ctr,aes192-ctr,aes256-ctr,arcfour256,arcfour128,aes128-cbc,3des-cbc,blowfish-cbc,cast128-cbc,aes192-cbc,aes256-cbc,arcfour,rijndael-cbc@lysator.liu.se\r\ndebug2: ciphers stoc: aes128-ctr,aes192-ctr,aes256-ctr,arcfour256,arcfour128,aes128-cbc,3des-cbc,blowfish-cbc,cast128-cbc,aes192-cbc,aes256-cbc,arcfour,rijndael-cbc@lysator.liu.se\r\ndebug2: MACs ctos: hmac-md5,hmac-sha1,umac-64@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-ripemd160,hmac-ripemd160@openssh.com,hmac-sha1-96,hmac-md5-96\r\ndebug2: MACs stoc: hmac-md5,hmac-sha1,umac-64@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-ripemd160,hmac-ripemd160@openssh.com,hmac-sha1-96,hmac-md5-96\r\ndebug2: compression ctos: none,zlib@openssh.com\r\ndebug2: compression stoc: none,zlib@openssh.com\r\ndebug2: languages ctos: \r\ndebug2: languages stoc: \r\ndebug2: first_kex_follows 0 \r\ndebug2: reserved 0 \r\ndebug1: kex: algorithm: diffie-hellman-group-exchange-sha256\r\ndebug1: kex: host key algorithm: ssh-rsa\r\ndebug1: kex: server->client cipher: aes128-ctr MAC: umac-64@openssh.com compression: zlib@openssh.com\r\ndebug1: kex: client->server cipher: aes128-ctr MAC: umac-64@openssh.com compression: zlib@openssh.com\r\ndebug3: send packet: type 34\r\ndebug1: SSH2_MSG_KEX_DH_GEX_REQUEST(2048<3072<8192) sent\r\ndebug3: receive packet: type 31\r\ndebug1: got SSH2_MSG_KEX_DH_GEX_GROUP\r\ndebug2: bits set: 1591/3072\r\ndebug3: send packet: type 32\r\ndebug1: SSH2_MSG_KEX_DH_GEX_INIT sent\r\ndebug3: receive packet: type 33\r\ndebug1: got SSH2_MSG_KEX_DH_GEX_REPLY\r\ndebug1: Server host key: ssh-rsa SHA256:mf9ThSOdsfxUY7DlJSmJL0XOOoKW2ub01frEsbNtVq4\r\ndebug3: hostkeys_foreach: reading file "/home/clint/.ssh/known_hosts"\r\ndebug3: record_hostkey: found key type RSA in file /home/clint/.ssh/known_hosts:1\r\ndebug3: load_hostkeys: loaded 1 keys from spamaps.org\r\ndebug3: hostkeys_foreach: reading file "/home/clint/.ssh/known_hosts"\r\ndebug3: record_hostkey: found key type RSA in file /home/clint/.ssh/known_hosts:2\r\ndebug3: load_hostkeys: loaded 1 keys from 184.105.137.237\r\ndebug1: Host \'spamaps.org\' is known and matches the RSA host key.\r\ndebug1: Found key in /home/clint/.ssh/known_hosts:1\r\ndebug2: bits set: 1560/3072\r\ndebug3: send packet: type 21\r\ndebug2: set_newkeys: mode 1\r\ndebug1: rekey after 4294967296 blocks\r\ndebug1: SSH2_MSG_NEWKEYS sent\r\ndebug1: expecting SSH2_MSG_NEWKEYS\r\ndebug3: receive packet: type 21\r\ndebug2: set_newkeys: mode 0\r\ndebug1: rekey after 4294967296 blocks\r\ndebug1: SSH2_MSG_NEWKEYS received\r\ndebug2: key: /home/clint/.ssh/id_rsa (0x55d803782db0), agent\r\ndebug2: key: clint@clint-ThinkPad-X250 (0x55d803784420), agent\r\ndebug2: key: noadmin@pyrrrat-root (0x55d8037840c0), agent\r\ndebug2: key: incoming-sftp (0x55d8037882a0), agent\r\ndebug2: key: /home/clint/.ssh/id_dsa ((nil))\r\ndebug2: key: /home/clint/.ssh/id_ecdsa ((nil))\r\ndebug2: key: /home/clint/.ssh/id_ed25519 ((nil))\r\ndebug3: send packet: type 5\r\ndebug3: receive packet: type 6\r\ndebug2: service_accept: ssh-userauth\r\ndebug1: SSH2_MSG_SERVICE_ACCEPT received\r\ndebug3: send packet: type 50\r\ndebug3: receive packet: type 51\r\ndebug1: Authentications that can continue: publickey,gssapi-keyex,gssapi-with-mic,password\r\ndebug3: start over, passed a different list publickey,gssapi-keyex,gssapi-with-mic,password\r\ndebug3: preferred gssapi-with-mic,gssapi-keyex,hostbased,publickey\r\ndebug3: authmethod_lookup gssapi-with-mic\r\ndebug3: remaining preferred: gssapi-keyex,hostbased,publickey\r\ndebug3: authmethod_is_enabled gssapi-with-mic\r\ndebug1: Next authentication method: gssapi-with-mic\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\nNo Kerberos credentials available\n\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\nNo Kerberos credentials available\n\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\n\n\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\nNo Kerberos credentials available\n\r\ndebug2: we did not send a packet, disable method\r\ndebug3: authmethod_lookup gssapi-keyex\r\ndebug3: remaining preferred: hostbased,publickey\r\ndebug3: authmethod_is_enabled gssapi-keyex\r\ndebug1: Next authentication method: gssapi-keyex\r\ndebug1: No valid Key exchange context\r\ndebug2: we did not send a packet, disable method\r\ndebug3: authmethod_lookup publickey\r\ndebug3: remaining preferred: ,publickey\r\ndebug3: authmethod_is_enabled publickey\r\ndebug1: Next authentication method: publickey\r\ndebug1: Offering RSA public key: /home/clint/.ssh/id_rsa\r\ndebug3: send_pubkey_test\r\ndebug3: send packet: type 50\r\ndebug2: we sent a publickey packet, wait for reply\r\ndebug3: receive packet: type 60\r\ndebug1: Server accepts key: pkalg ssh-rsa blen 279\r\ndebug2: input_userauth_pk_ok: fp SHA256:qzvF1wyt0Tt+T3Y8kIxwjaK68utj4IsxXzYvjBToPOI\r\ndebug3: sign_and_send_pubkey: RSA SHA256:qzvF1wyt0Tt+T3Y8kIxwjaK68utj4IsxXzYvjBToPOI\r\ndebug3: send packet: type 50\r\ndebug3: receive packet: type 52\r\ndebug1: Enabling compression at level 6.\r\ndebug1: Authentication succeeded (publickey).\r\nAuthenticated to spamaps.org ([184.105.137.237]:22).\r\ndebug1: setting up multiplex master socket\r\ndebug3: muxserver_listen: temporary control path /home/clint/.ansible/cp/80e720ccb6.rvVvdvdxZsu7NdlB\r\ndebug2: fd 5 setting O_NONBLOCK\r\ndebug3: fd 5 is O_NONBLOCK\r\ndebug3: fd 5 is O_NONBLOCK\r\ndebug1: channel 0: new [/home/clint/.ansible/cp/80e720ccb6]\r\ndebug3: muxserver_listen: mux listener channel 0 fd 5\r\ndebug2: fd 3 setting TCP_NODELAY\r\ndebug3: ssh_packet_set_tos: set IP_TOS 0x08\r\ndebug1: control_persist_detach: backgrounding master process\r\ndebug2: control_persist_detach: background process is 16635\r\ndebug2: fd 5 setting O_NONBLOCK\r\ndebug1: forking to background\r\ndebug1: Entering interactive session.\r\ndebug1: pledge: id\r\ndebug2: set_control_persist_exit_time: schedule exit in 60 seconds\r\ndebug1: multiplexing control connection\r\ndebug2: fd 6 setting O_NONBLOCK\r\ndebug3: fd 6 is O_NONBLOCK\r\ndebug1: channel 1: new [mux-control]\r\ndebug3: channel_post_mux_listener: new mux channel 1 fd 6\r\ndebug3: mux_master_read_cb: channel 1: hello sent\r\ndebug2: set_control_persist_exit_time: cancel scheduled exit\r\ndebug3: mux_master_read_cb: channel 1 packet type 0x00000001 len 4\r\ndebug2: process_mux_master_hello: channel 1 slave version 4\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_master_read_cb: channel 1 packet type 0x10000004 len 4\r\ndebug2: process_mux_alive_check: channel 1: alive check\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug3: mux_master_read_cb: channel 1 packet type 0x10000002 len 100\r\ndebug2: process_mux_new_session: channel 1: request tty 0, X 0, agent 0, subsys 0, term "xterm-256color", cmd "/bin/sh -c \'echo ~ && sleep 0\'", env 1\r\ndebug3: process_mux_new_session: got fds stdin 7, stdout 8, stderr 9\r\ndebug2: fd 8 setting O_NONBLOCK\r\ndebug2: fd 9 setting O_NONBLOCK\r\ndebug1: channel 2: new [client-session]\r\ndebug2: process_mux_new_session: channel_new: 2 linked to control channel 1\r\ndebug2: channel 2: send open\r\ndebug3: send packet: type 90\r\ndebug3: receive packet: type 91\r\ndebug2: callback start\r\ndebug2: client_session2_setup: id 2\r\ndebug1: Sending environment.\r\ndebug1: Sending env LANG = en_US.UTF-8\r\ndebug2: channel 2: request env confirm 0\r\ndebug3: send packet: type 98\r\ndebug1: Sending command: /bin/sh -c \'echo ~ && sleep 0\'\r\ndebug2: channel 2: request exec confirm 1\r\ndebug3: send packet: type 98\r\ndebug3: mux_session_confirm: sending success reply\r\ndebug2: callback done\r\ndebug2: channel 2: open confirm rwindow 0 rmax 32768\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug2: channel 2: rcvd adjust 2097152\r\ndebug3: receive packet: type 99\r\ndebug2: channel_input_status_confirm: type 99 id 2\r\ndebug2: exec request accepted on channel 2\r\ndebug3: receive packet: type 98\r\ndebug1: client_input_channel_req: channel 2 rtype exit-status reply 0\r\ndebug3: mux_exit_message: channel 2: exit message, exitval 0\r\ndebug3: receive packet: type 98\r\ndebug1: client_input_channel_req: channel 2 rtype eow@openssh.com reply 0\r\ndebug2: channel 2: rcvd eow\r\ndebug2: channel 2: close_read\r\ndebug2: channel 2: input open -> closed\r\ndebug3: receive packet: type 96\r\ndebug2: channel 2: rcvd eof\r\ndebug2: channel 2: output open -> drain\r\ndebug2: channel 2: obuf empty\r\ndebug2: channel 2: close_write\r\ndebug2: channel 2: output drain -> closed\r\ndebug3: receive packet: type 97\r\ndebug2: channel 2: rcvd close\r\ndebug3: channel 2: will not send data after close\r\ndebug2: channel 2: send close\r\ndebug3: send packet: type 97\r\ndebug2: channel 2: is dead\r\ndebug2: channel 2: gc: notify user\r\ndebug3: mux_master_session_cleanup_cb: entering for channel 2\r\ndebug2: channel 1: rcvd close\r\ndebug2: channel 1: output open -> drain\r\ndebug2: channel 1: close_read\r\ndebug2: channel 1: input open -> closed\r\ndebug2: channel 2: gc: user detached\r\ndebug2: channel 2: is dead\r\ndebug2: channel 2: garbage collecting\r\ndebug1: channel 2: free: client-session, nchannels 3\r\ndebug3: channel 2: status: The following connections are open:\r\n #2 client-session (t4 r0 i3/0 o3/0 fd -1/-1 cc -1)\r\n\r\ndebug2: channel 1: obuf empty\r\ndebug2: channel 1: close_write\r\ndebug2: channel 1: output drain -> closed\r\ndebug2: channel 1: is dead (local)\r\ndebug2: channel 1: gc: notify user\r\ndebug3: mux_master_control_cleanup_cb: entering for channel 1\r\ndebug2: channel 1: gc: user detached\r\ndebug2: channel 1: is dead (local)\r\ndebug2: channel 1: garbage collecting\r\ndebug1: channel 1: free: mux-control, nchannels 2\r\ndebug3: channel 1: status: The following connections are open:\r\n\r\ndebug2: set_control_persist_exit_time: schedule exit in 60 seconds\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 spamaps.org '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351 `" && echo ansible-tmp-1500415665.0875719-218494642853351="` echo /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351 `" ) && sleep 0'"'"'' <spamaps.org> (0, b'ansible-tmp-1500415665.0875719-218494642853351=/home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> PUT /tmp/tmpbaaj5_y1 TO /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set sftp_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC sftp -b - -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 '[spamaps.org]' <spamaps.org> (0, b'sftp> put /tmp/tmpbaaj5_y1 /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug2: Remote version: 3\r\ndebug2: Server supports extension "posix-rename@openssh.com" revision 1\r\ndebug2: Server supports extension "statvfs@openssh.com" revision 2\r\ndebug2: Server supports extension "fstatvfs@openssh.com" revision 2\r\ndebug3: Sent message fd 3 T:16 I:1\r\ndebug3: SSH_FXP_REALPATH . -> /home/clint size 0\r\ndebug3: Looking up /tmp/tmpbaaj5_y1\r\ndebug3: Sent message fd 3 T:17 I:2\r\ndebug3: Received stat reply T:101 I:2\r\ndebug1: Couldn\'t stat remote file: No such file or directory\r\ndebug3: Sent message SSH2_FXP_OPEN I:3 P:/home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py\r\ndebug3: Sent message SSH2_FXP_WRITE I:4 O:0 S:32768\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 4 32768 bytes at 0\r\ndebug3: Sent message SSH2_FXP_WRITE I:5 O:32768 S:32768\r\ndebug3: Sent message SSH2_FXP_WRITE I:6 O:65536 S:32768\r\ndebug3: Sent message SSH2_FXP_WRITE I:7 O:98304 S:5343\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 5 32768 bytes at 32768\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 6 32768 bytes at 65536\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 7 5343 bytes at 98304\r\ndebug3: Sent message SSH2_FXP_CLOSE I:4\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 spamaps.org '/bin/sh -c '"'"'chmod u+x /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/ /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py && sleep 0'"'"'' <spamaps.org> (0, b'', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 -tt spamaps.org '/bin/sh -c '"'"'/usr/bin/python /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py; rm -rf "/home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/" > /dev/null 2>&1 && sleep 0'"'"'' <spamaps.org> (0, b'/tmp/ansible_sKLfd8/ansible_modlib.zip/ansible/module_utils/facts.py:1019: DeprecationWarning: object.__new__() takes no parameters\r\n/tmp/ansible_sKLfd8/ansible_modlib.zip/ansible/module_utils/facts.py:2438: DeprecationWarning: object.__new__() takes no parameters\r\n/tmp/ansible_sKLfd8/ansible_modlib.zip/ansible/module_utils/facts.py:3342: DeprecationWarning: object.__new__() takes no parameters\r\n\r\n{"invocation": {"module_args": {"filter": "*", "gather_subset": ["all"], "fact_path": "/etc/ansible/facts.d", "gather_timeout": 10}}, "changed": false, "ansible_facts": {"ansible_product_serial": "NA", "ansible_form_factor": "NA", "ansible_product_version": "NA", "ansible_fips": false, "ansible_service_mgr": "upstart", "ansible_user_id": "clint", "ansible_user_dir": "/home/clint", "ansible_memtotal_mb": 1459, "ansible_system_capabilities": [""], "ansible_distribution_version": "6.9", "ansible_domain": "srihosting.com", "ansible_date_time": {"weekday_number": "2", "iso8601_basic_short": "20170718T150754", "tz": "PDT", "weeknumber": "29", "hour": "15", "year": "2017", "minute": "07", "tz_offset": "-0700", "month": "07", "epoch": "1500415674", "iso8601_micro": "2017-07-18T22:07:54.378020Z", "weekday": "Tuesday", "time": "15:07:54", "date": "2017-07-18", "iso8601": "2017-07-18T22:07:54Z", "day": "18", "iso8601_basic": "20170718T150754377747", "second": "54"}, "ansible_real_user_id": 500, "ansible_processor_cores": 1, "ansible_virtualization_role": "guest", "ansible_dns": {"nameservers": ["8.8.8.8", "8.8.4.4"]}, "ansible_processor_vcpus": 1, "ansible_bios_version": "NA", "ansible_processor": ["GenuineIntel", "Intel(R) Xeon(R) CPU L5630 @ 2.13GHz"], "ansible_virtualization_type": "xen", "ansible_lo": {"features": {"tx_checksum_ipv4": "off [fixed]", "generic_receive_offload": "on", "tx_checksum_ipv6": "off [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_checksum_unneeded": "off [fixed]", "highdma": "on [fixed]", "tx_lockless": "on [fixed]", "tx_tcp_ecn_segmentation": "on", "tx_gso_robust": "off [fixed]", "tx_checksumming": "on", "vlan_challenged": "on [fixed]", "loopback": "on [fixed]", "fcoe_mtu": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "large_receive_offload": "off [fixed]", "tx_scatter_gather": "on [fixed]", "rx_checksumming": "on [fixed]", "tx_tcp_segmentation": "on", "netns_local": "on [fixed]", "generic_segmentation_offload": "on", "tx_udp_tnl_segmentation": "off [fixed]", "tcp_segmentation_offload": "on", "rx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_tcp6_segmentation": "on", "udp_fragmentation_offload": "on", "scatter_gather": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "rx_vlan_filter": "off [fixed]", "receive_hashing": "off [fixed]", "tx_gre_segmentation": "off [fixed]"}, "mtu": 65536, "device": "lo", "promisc": false, "ipv4": {"broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0", "address": "127.0.0.1"}, "ipv6": [{"scope": "host", "prefix": "128", "address": "::1"}], "active": true, "type": "loopback"}, "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_default_ipv4": {"macaddress": "00:16:3e:1f:ea:f2", "network": "184.105.137.192", "mtu": 1500, "broadcast": "184.105.137.255", "alias": "eth0", "netmask": "255.255.255.192", "address": "184.105.137.237", "interface": "eth0", "type": "ether", "gateway": "184.105.137.193"}, "ansible_swapfree_mb": 0, "ansible_default_ipv6": {}, "ansible_distribution_release": "Final", "ansible_system_vendor": "NA", "ansible_apparmor": {"status": "disabled"}, "ansible_cmdline": {"LANG": "en_US.UTF-8", "rd_NO_LUKS": true, "ro": true, "console": "hvc0", "rd_NO_MD": true, "quiet": true, "rd_LVM_LV": "XenVolGroup/lv_root", "rhgb": true, "KEYTABLE": "us", "crashkernel": "auto", "SYSFONT": "latarcyrheb-sun16", "root": "/dev/mapper/XenVolGroup-lv_root", "rd_NO_DM": true}, "ansible_effective_user_id": 500, "ansible_mounts": [{"uuid": "N/A", "size_total": 31602700288, "mount": "/", "size_available": 5297983488, "fstype": "ext3", "device": "/dev/mapper/XenVolGroup-lv_root", "options": "rw"}, {"uuid": "N/A", "size_total": 101529600, "mount": "/boot", "size_available": 30937088, "fstype": "ext3", "device": "/dev/xvda1", "options": "rw"}], "ansible_selinux": {"status": "disabled"}, "ansible_os_family": "RedHat", "ansible_userspace_architecture": "x86_64", "ansible_product_uuid": "NA", "ansible_product_name": "NA", "ansible_pkg_mgr": "yum", "ansible_memfree_mb": 60, "ansible_devices": {"xvda": {"scheduler_mode": "cfq", "rotational": "0", "vendor": null, "sectors": "94371840", "sas_device_handle": null, "sas_address": null, "host": "", "sectorsize": "512", "removable": "0", "support_discard": "0", "model": null, "partitions": {"xvda1": {"sectorsize": 512, "uuid": "8703c7d1-ebe2-4c8b-8205-4a3ea341e3ce", "sectors": "204800", "start": "2048", "holders": [], "size": "100.00 MB"}, "xvda2": {"sectorsize": 512, "uuid": null, "sectors": "62703692", "start": "206848", "holders": ["XenVolGroup-lv_root"], "size": "29.90 GB"}}, "holders": [], "size": "45.00 GB"}}, "ansible_user_uid": 500, "ansible_memory_mb": {"real": {"total": 1459, "used": 1399, "free": 60}, "swap": {"cached": 0, "total": 0, "free": 0, "used": 0}, "nocache": {"used": 594, "free": 865}}, "ansible_distribution": "CentOS", "ansible_env": {"LANG": "en_US.UTF-8", "TERM": "xterm-256color", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_TTY": "/dev/pts/3", "G_BROKEN_FILENAMES": "1", "SSH_CLIENT": "172.250.7.7 45782 22", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "PWD": "/home/clint", "CVS_RSH": "ssh", "LOGNAME": "clint", "USER": "clint", "MANPATH": "/usr/lib/courier-imap/man:", "PATH": "/usr/lib/courier-imap/bin:/usr/local/bin:/bin:/usr/bin", "MAIL": "/var/mail/clint", "LS_COLORS": "rs=0:di=38;5;27:ln=38;5;51:mh=44;38;5;15:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=05;48;5;232;38;5;15:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;34:*.tar=38;5;9:*.tgz=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.Z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lz=38;5;9:*.xz=38;5;9:*.bz2=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.bz=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.rar=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.axv=38;5;13:*.anx=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.axa=38;5;45:*.oga=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "HOME": "/home/clint", "_": "/usr/bin/python", "SSH_CONNECTION": "172.250.7.7 45782 184.105.137.237 22"}, "ansible_distribution_major_version": "6", "module_setup": true, "ansible_processor_count": 1, "ansible_hostname": "xencbyrum2", "ansible_effective_group_id": 500, "ansible_swaptotal_mb": 0, "ansible_lsb": {"release": "6.9", "major_release": "6", "codename": "Final", "id": "CentOS", "description": "CentOS release 6.9 (Final)"}, "ansible_real_group_id": 500, "ansible_bios_date": "NA", "ansible_all_ipv6_addresses": ["fe80::216:3eff:fe1f:eaf2"], "ansible_interfaces": ["lo", "eth0"], "ansible_uptime_seconds": 14524108, "ansible_machine_id": "8b82be7fd3ea5ff44379c1b60000001a", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAABIwAAAQEAyLKP0RyBHL4TLU5NlD3nzPBa6D/PVvCsgmNASkzFrqKKe52XrREnsI26JUnaSYIR9L7NOrshd99kxcYnxXpV/pLBXgzdae3r4vTDNDUwzAq3vlQXmSw18IN1REe1cGg8E8OawcOxuNDn69Lbll4e4vGriIyWgwTifqjAL2JPd+Orw6DpRYMG3MvDW7CUnF7wr6H1kC4AiviGtckmtFwP+hDTe8Q1C88pq1rfnasdjm9PwWbrJoJmqUq5m3aT3WwW27NJXz6QSgTnAF7UD9EZfU5/ixBdW2S1FcAP28pEQ3BoFRH0LuuadUdN6cLPCNPV07aTZ79Z3vH5KWRopjqLtQ==", "ansible_gather_subset": ["hardware", "network", "virtual"], "ansible_user_gecos": "Clint Byrum", "ansible_system_capabilities_enforced": "True", "ansible_python": {"executable": "/usr/bin/python", "version": {"micro": 6, "major": 2, "releaselevel": "final", "serial": 0, "minor": 6}, "type": "CPython", "has_sslcontext": false, "version_info": [2, 6, 6, "final", 0]}, "ansible_kernel": "2.6.32-642.11.1.el6.x86_64", "ansible_processor_threads_per_core": 1, "ansible_fqdn": "xencbyrum2.srihosting.com", "ansible_user_gid": 500, "ansible_eth0": {"macaddress": "00:16:3e:1f:ea:f2", "features": {"tx_checksum_ipv4": "on", "generic_receive_offload": "on", "tx_checksum_ipv6": "off", "tx_scatter_gather_fraglist": "off [fixed]", "tx_checksum_unneeded": "off", "highdma": "off [fixed]", "tx_lockless": "off [fixed]", "tx_tcp_ecn_segmentation": "off", "tx_gso_robust": "on [fixed]", "tx_checksumming": "on", "vlan_challenged": "off [fixed]", "loopback": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "large_receive_offload": "off [fixed]", "tx_scatter_gather": "on", "rx_checksumming": "off [fixed]", "tx_tcp_segmentation": "on", "netns_local": "off [fixed]", "generic_segmentation_offload": "on", "tx_udp_tnl_segmentation": "off [fixed]", "tcp_segmentation_offload": "on", "rx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_tcp6_segmentation": "off", "udp_fragmentation_offload": "off [fixed]", "scatter_gather": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off", "tx_fcoe_segmentation": "off [fixed]", "rx_vlan_filter": "off [fixed]", "receive_hashing": "off [fixed]", "tx_gre_segmentation": "off [fixed]"}, "pciid": "vif-0", "module": "xen_netfront", "mtu": 1500, "device": "eth0", "promisc": false, "ipv4": {"broadcast": "184.105.137.255", "netmask": "255.255.255.192", "network": "184.105.137.192", "address": "184.105.137.237"}, "ipv6": [{"scope": "link", "prefix": "64", "address": "fe80::216:3eff:fe1f:eaf2"}], "active": true, "type": "ether"}, "ansible_nodename": "xencbyrum2.srihosting.com", "ansible_system": "Linux", "ansible_user_shell": "/bin/bash", "ansible_machine": "x86_64", "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPOlY3CbmfqEerypkFdWJO6fNaz7c4HH4UCAeVQxVrpo+PoSqtTJHhB24gOmxicCcmkXgb6KJP16tWcpV47dfLCfNY078k24yFBRtAfIIXhZSGyXjPdcp3BTTQVkqRyR1ToS572FAgzi6it6wUc0psSXdaOJXdRqtqieFQUnR3Q5AAAAFQDDLXxM4aIcL5sYlSFRbLwCsIGAXQAAAIBv71rbqG53O2b9MKZ0MIGp2vDT+0H32SycdeoWk0YqTqOesjQaKz2OHEdFqcYf0nm8zLyiIuNTR0MLBq1mHt2QUbXi+T4pAxnseV5wLclgZY5Hydg6F5v86ixQjV9W5/unlL4MKcmLL62hGnJIykVHPXhJ2elPykwp3Oj0eurfOwAAAIEA7981Kzo+/PfYYoI2clhvjtxAEolpCMv/EpJ3N504LZ79K4ADrIS7PhI9lb8Bne6UqZo+9eoKm5JAnzvfyArb2jUOrz8BSU4Q0KVZnCaBkz0TtuFyKHcWpmasSIimAzckAtF+LRCzbz/8hJg9/1Gbp8Zq2HQl5O3Hxde+Ry9zdFc=", "ansible_all_ipv4_addresses": ["184.105.137.237"], "ansible_python_version": "2.6.6"}}\r\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to spamaps.org closed.\r\n') ok: [spamaps.org] META: ran handlers TASK [stuff] ******************************************************************************************************************************************************************************************************* task path: /home/clint/test.yml:4 ok: [spamaps.org] => { "msg": "Here is a message" } SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) SSH: PlayContext set ssh_common_args: () SSH: PlayContext set ssh_extra_args: () SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/276d6f8b70) ERROR! Unexpected Exception: list index out of range the full traceback was: Traceback (most recent call last): File "/usr/local/bin/ansible-playbook", line 109, in <module> exit_code = cli.run() File "/usr/local/lib/python3.5/dist-packages/ansible/cli/playbook.py", line 154, in run results = pbex.run() File "/usr/local/lib/python3.5/dist-packages/ansible/executor/playbook_executor.py", line 153, in run result = self._tqm.run(play=play) File "/usr/local/lib/python3.5/dist-packages/ansible/executor/task_queue_manager.py", line 284, in run play_return = strategy.run(iterator, play_context) File "/usr/local/lib/python3.5/dist-packages/ansible/plugins/strategy/linear.py", line 221, in run results.extend(self._execute_meta(task, play_context, iterator, host)) File "/usr/local/lib/python3.5/dist-packages/ansible/plugins/strategy/__init__.py", line 905, in _execute_meta connection.reset() File "/usr/local/lib/python3.5/dist-packages/ansible/plugins/connection/ssh.py", line 816, in reset p = subprocess.Popen(cmd, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) File "/usr/lib/python3.5/subprocess.py", line 947, in __init__ restore_signals, start_new_session) File "/usr/lib/python3.5/subprocess.py", line 1448, in _execute_child executable = args[0] IndexError: list index out of range $ ```
True
meta: reset_connection results in "IndexError: list index out of range" - <!--- Verify first that your issue/request is not already reported on GitHub. Also test if the latest release, and master branch are affected too. --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME meta module ##### ANSIBLE VERSION ``` ansible-playbook 2.3.1.0 config file = configured module search path = Default w/o overrides python version = 3.5.2 (default, Nov 17 2016, 17:05:23) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION No configuration. ##### OS / ENVIRONMENT Ubuntu 16.04 with ansible pip installed, running in python3.5. Contacting a VM running CentOS 6 ##### SUMMARY Using `meta: reset_connection` works and does reset the connection, but then later on this IndexError ocurrs. ##### STEPS TO REPRODUCE This playbook shows the problem: ``` - name: Ping them hosts: all tasks: - name: stuff debug: msg: Here is a message - name: reset conn meta: reset_connection ``` ##### EXPECTED RESULTS ansible-playbook should exit 0 without errors ##### ACTUAL RESULTS ~/tmp/inv has this content: ``` spamaps.org ``` My user has SSH agent access to spamaps.org. ``` $ ansible-playbook test.yml -i ~/tmp/inv -vvvvv No config file found; using defaults Loading callback plugin default of type stdout, v2.0 from /usr/local/lib/python3.5/dist-packages/ansible/plugins/callback/__init__.py PLAYBOOK: test.yml ************************************************************************************************************************************************************************************************* 1 plays in test.yml PLAY [Ping them] *************************************************************************************************************************************************************************************************** TASK [Gathering Facts] ********************************************************************************************************************************************************************************************* Using module file /usr/local/lib/python3.5/dist-packages/ansible/modules/system/setup.py <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 spamaps.org '/bin/sh -c '"'"'echo ~ && sleep 0'"'"'' <spamaps.org> (0, b'/home/clint\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug1: Control socket "/home/clint/.ansible/cp/80e720ccb6" does not exist\r\ndebug2: resolving "spamaps.org" port 22\r\ndebug2: ssh_connect_direct: needpriv 0\r\ndebug1: Connecting to spamaps.org [184.105.137.237] port 22.\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug1: fd 3 clearing O_NONBLOCK\r\ndebug1: Connection established.\r\ndebug3: timeout: 9959 ms remain after connect\r\ndebug1: identity file /home/clint/.ssh/id_rsa type 1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_rsa-cert type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_dsa type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_dsa-cert type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ecdsa type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ecdsa-cert type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ed25519 type -1\r\ndebug1: key_load_public: No such file or directory\r\ndebug1: identity file /home/clint/.ssh/id_ed25519-cert type -1\r\ndebug1: Enabling compatibility mode for protocol 2.0\r\ndebug1: Local version string SSH-2.0-OpenSSH_7.2p2 Ubuntu-4ubuntu2.2\r\ndebug1: Remote protocol version 2.0, remote software version OpenSSH_5.3\r\ndebug1: match: OpenSSH_5.3 pat OpenSSH_5* compat 0x0c000000\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug1: Authenticating to spamaps.org:22 as \'clint\'\r\ndebug3: hostkeys_foreach: reading file "/home/clint/.ssh/known_hosts"\r\ndebug3: record_hostkey: found key type RSA in file /home/clint/.ssh/known_hosts:1\r\ndebug3: load_hostkeys: loaded 1 keys from spamaps.org\r\ndebug3: order_hostkeyalgs: prefer hostkeyalgs: ssh-rsa-cert-v01@openssh.com,rsa-sha2-512,rsa-sha2-256,ssh-rsa\r\ndebug3: send packet: type 20\r\ndebug1: SSH2_MSG_KEXINIT sent\r\ndebug3: receive packet: type 20\r\ndebug1: SSH2_MSG_KEXINIT received\r\ndebug2: local client KEXINIT proposal\r\ndebug2: KEX algorithms: curve25519-sha256@libssh.org,ecdh-sha2-nistp256,ecdh-sha2-nistp384,ecdh-sha2-nistp521,diffie-hellman-group-exchange-sha256,diffie-hellman-group-exchange-sha1,diffie-hellman-group14-sha1,ext-info-c\r\ndebug2: host key algorithms: ssh-rsa-cert-v01@openssh.com,rsa-sha2-512,rsa-sha2-256,ssh-rsa,ecdsa-sha2-nistp256-cert-v01@openssh.com,ecdsa-sha2-nistp384-cert-v01@openssh.com,ecdsa-sha2-nistp521-cert-v01@openssh.com,ssh-ed25519-cert-v01@openssh.com,ecdsa-sha2-nistp256,ecdsa-sha2-nistp384,ecdsa-sha2-nistp521,ssh-ed25519\r\ndebug2: ciphers ctos: chacha20-poly1305@openssh.com,aes128-ctr,aes192-ctr,aes256-ctr,aes128-gcm@openssh.com,aes256-gcm@openssh.com,aes128-cbc,aes192-cbc,aes256-cbc,3des-cbc\r\ndebug2: ciphers stoc: chacha20-poly1305@openssh.com,aes128-ctr,aes192-ctr,aes256-ctr,aes128-gcm@openssh.com,aes256-gcm@openssh.com,aes128-cbc,aes192-cbc,aes256-cbc,3des-cbc\r\ndebug2: MACs ctos: umac-64-etm@openssh.com,umac-128-etm@openssh.com,hmac-sha2-256-etm@openssh.com,hmac-sha2-512-etm@openssh.com,hmac-sha1-etm@openssh.com,umac-64@openssh.com,umac-128@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-sha1\r\ndebug2: MACs stoc: umac-64-etm@openssh.com,umac-128-etm@openssh.com,hmac-sha2-256-etm@openssh.com,hmac-sha2-512-etm@openssh.com,hmac-sha1-etm@openssh.com,umac-64@openssh.com,umac-128@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-sha1\r\ndebug2: compression ctos: zlib@openssh.com,zlib,none\r\ndebug2: compression stoc: zlib@openssh.com,zlib,none\r\ndebug2: languages ctos: \r\ndebug2: languages stoc: \r\ndebug2: first_kex_follows 0 \r\ndebug2: reserved 0 \r\ndebug2: peer server KEXINIT proposal\r\ndebug2: KEX algorithms: diffie-hellman-group-exchange-sha256,diffie-hellman-group-exchange-sha1,diffie-hellman-group14-sha1,diffie-hellman-group1-sha1\r\ndebug2: host key algorithms: ssh-rsa,ssh-dss\r\ndebug2: ciphers ctos: aes128-ctr,aes192-ctr,aes256-ctr,arcfour256,arcfour128,aes128-cbc,3des-cbc,blowfish-cbc,cast128-cbc,aes192-cbc,aes256-cbc,arcfour,rijndael-cbc@lysator.liu.se\r\ndebug2: ciphers stoc: aes128-ctr,aes192-ctr,aes256-ctr,arcfour256,arcfour128,aes128-cbc,3des-cbc,blowfish-cbc,cast128-cbc,aes192-cbc,aes256-cbc,arcfour,rijndael-cbc@lysator.liu.se\r\ndebug2: MACs ctos: hmac-md5,hmac-sha1,umac-64@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-ripemd160,hmac-ripemd160@openssh.com,hmac-sha1-96,hmac-md5-96\r\ndebug2: MACs stoc: hmac-md5,hmac-sha1,umac-64@openssh.com,hmac-sha2-256,hmac-sha2-512,hmac-ripemd160,hmac-ripemd160@openssh.com,hmac-sha1-96,hmac-md5-96\r\ndebug2: compression ctos: none,zlib@openssh.com\r\ndebug2: compression stoc: none,zlib@openssh.com\r\ndebug2: languages ctos: \r\ndebug2: languages stoc: \r\ndebug2: first_kex_follows 0 \r\ndebug2: reserved 0 \r\ndebug1: kex: algorithm: diffie-hellman-group-exchange-sha256\r\ndebug1: kex: host key algorithm: ssh-rsa\r\ndebug1: kex: server->client cipher: aes128-ctr MAC: umac-64@openssh.com compression: zlib@openssh.com\r\ndebug1: kex: client->server cipher: aes128-ctr MAC: umac-64@openssh.com compression: zlib@openssh.com\r\ndebug3: send packet: type 34\r\ndebug1: SSH2_MSG_KEX_DH_GEX_REQUEST(2048<3072<8192) sent\r\ndebug3: receive packet: type 31\r\ndebug1: got SSH2_MSG_KEX_DH_GEX_GROUP\r\ndebug2: bits set: 1591/3072\r\ndebug3: send packet: type 32\r\ndebug1: SSH2_MSG_KEX_DH_GEX_INIT sent\r\ndebug3: receive packet: type 33\r\ndebug1: got SSH2_MSG_KEX_DH_GEX_REPLY\r\ndebug1: Server host key: ssh-rsa SHA256:mf9ThSOdsfxUY7DlJSmJL0XOOoKW2ub01frEsbNtVq4\r\ndebug3: hostkeys_foreach: reading file "/home/clint/.ssh/known_hosts"\r\ndebug3: record_hostkey: found key type RSA in file /home/clint/.ssh/known_hosts:1\r\ndebug3: load_hostkeys: loaded 1 keys from spamaps.org\r\ndebug3: hostkeys_foreach: reading file "/home/clint/.ssh/known_hosts"\r\ndebug3: record_hostkey: found key type RSA in file /home/clint/.ssh/known_hosts:2\r\ndebug3: load_hostkeys: loaded 1 keys from 184.105.137.237\r\ndebug1: Host \'spamaps.org\' is known and matches the RSA host key.\r\ndebug1: Found key in /home/clint/.ssh/known_hosts:1\r\ndebug2: bits set: 1560/3072\r\ndebug3: send packet: type 21\r\ndebug2: set_newkeys: mode 1\r\ndebug1: rekey after 4294967296 blocks\r\ndebug1: SSH2_MSG_NEWKEYS sent\r\ndebug1: expecting SSH2_MSG_NEWKEYS\r\ndebug3: receive packet: type 21\r\ndebug2: set_newkeys: mode 0\r\ndebug1: rekey after 4294967296 blocks\r\ndebug1: SSH2_MSG_NEWKEYS received\r\ndebug2: key: /home/clint/.ssh/id_rsa (0x55d803782db0), agent\r\ndebug2: key: clint@clint-ThinkPad-X250 (0x55d803784420), agent\r\ndebug2: key: noadmin@pyrrrat-root (0x55d8037840c0), agent\r\ndebug2: key: incoming-sftp (0x55d8037882a0), agent\r\ndebug2: key: /home/clint/.ssh/id_dsa ((nil))\r\ndebug2: key: /home/clint/.ssh/id_ecdsa ((nil))\r\ndebug2: key: /home/clint/.ssh/id_ed25519 ((nil))\r\ndebug3: send packet: type 5\r\ndebug3: receive packet: type 6\r\ndebug2: service_accept: ssh-userauth\r\ndebug1: SSH2_MSG_SERVICE_ACCEPT received\r\ndebug3: send packet: type 50\r\ndebug3: receive packet: type 51\r\ndebug1: Authentications that can continue: publickey,gssapi-keyex,gssapi-with-mic,password\r\ndebug3: start over, passed a different list publickey,gssapi-keyex,gssapi-with-mic,password\r\ndebug3: preferred gssapi-with-mic,gssapi-keyex,hostbased,publickey\r\ndebug3: authmethod_lookup gssapi-with-mic\r\ndebug3: remaining preferred: gssapi-keyex,hostbased,publickey\r\ndebug3: authmethod_is_enabled gssapi-with-mic\r\ndebug1: Next authentication method: gssapi-with-mic\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\nNo Kerberos credentials available\n\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\nNo Kerberos credentials available\n\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\n\n\r\ndebug1: Unspecified GSS failure. Minor code may provide more information\nNo Kerberos credentials available\n\r\ndebug2: we did not send a packet, disable method\r\ndebug3: authmethod_lookup gssapi-keyex\r\ndebug3: remaining preferred: hostbased,publickey\r\ndebug3: authmethod_is_enabled gssapi-keyex\r\ndebug1: Next authentication method: gssapi-keyex\r\ndebug1: No valid Key exchange context\r\ndebug2: we did not send a packet, disable method\r\ndebug3: authmethod_lookup publickey\r\ndebug3: remaining preferred: ,publickey\r\ndebug3: authmethod_is_enabled publickey\r\ndebug1: Next authentication method: publickey\r\ndebug1: Offering RSA public key: /home/clint/.ssh/id_rsa\r\ndebug3: send_pubkey_test\r\ndebug3: send packet: type 50\r\ndebug2: we sent a publickey packet, wait for reply\r\ndebug3: receive packet: type 60\r\ndebug1: Server accepts key: pkalg ssh-rsa blen 279\r\ndebug2: input_userauth_pk_ok: fp SHA256:qzvF1wyt0Tt+T3Y8kIxwjaK68utj4IsxXzYvjBToPOI\r\ndebug3: sign_and_send_pubkey: RSA SHA256:qzvF1wyt0Tt+T3Y8kIxwjaK68utj4IsxXzYvjBToPOI\r\ndebug3: send packet: type 50\r\ndebug3: receive packet: type 52\r\ndebug1: Enabling compression at level 6.\r\ndebug1: Authentication succeeded (publickey).\r\nAuthenticated to spamaps.org ([184.105.137.237]:22).\r\ndebug1: setting up multiplex master socket\r\ndebug3: muxserver_listen: temporary control path /home/clint/.ansible/cp/80e720ccb6.rvVvdvdxZsu7NdlB\r\ndebug2: fd 5 setting O_NONBLOCK\r\ndebug3: fd 5 is O_NONBLOCK\r\ndebug3: fd 5 is O_NONBLOCK\r\ndebug1: channel 0: new [/home/clint/.ansible/cp/80e720ccb6]\r\ndebug3: muxserver_listen: mux listener channel 0 fd 5\r\ndebug2: fd 3 setting TCP_NODELAY\r\ndebug3: ssh_packet_set_tos: set IP_TOS 0x08\r\ndebug1: control_persist_detach: backgrounding master process\r\ndebug2: control_persist_detach: background process is 16635\r\ndebug2: fd 5 setting O_NONBLOCK\r\ndebug1: forking to background\r\ndebug1: Entering interactive session.\r\ndebug1: pledge: id\r\ndebug2: set_control_persist_exit_time: schedule exit in 60 seconds\r\ndebug1: multiplexing control connection\r\ndebug2: fd 6 setting O_NONBLOCK\r\ndebug3: fd 6 is O_NONBLOCK\r\ndebug1: channel 1: new [mux-control]\r\ndebug3: channel_post_mux_listener: new mux channel 1 fd 6\r\ndebug3: mux_master_read_cb: channel 1: hello sent\r\ndebug2: set_control_persist_exit_time: cancel scheduled exit\r\ndebug3: mux_master_read_cb: channel 1 packet type 0x00000001 len 4\r\ndebug2: process_mux_master_hello: channel 1 slave version 4\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_master_read_cb: channel 1 packet type 0x10000004 len 4\r\ndebug2: process_mux_alive_check: channel 1: alive check\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug3: mux_master_read_cb: channel 1 packet type 0x10000002 len 100\r\ndebug2: process_mux_new_session: channel 1: request tty 0, X 0, agent 0, subsys 0, term "xterm-256color", cmd "/bin/sh -c \'echo ~ && sleep 0\'", env 1\r\ndebug3: process_mux_new_session: got fds stdin 7, stdout 8, stderr 9\r\ndebug2: fd 8 setting O_NONBLOCK\r\ndebug2: fd 9 setting O_NONBLOCK\r\ndebug1: channel 2: new [client-session]\r\ndebug2: process_mux_new_session: channel_new: 2 linked to control channel 1\r\ndebug2: channel 2: send open\r\ndebug3: send packet: type 90\r\ndebug3: receive packet: type 91\r\ndebug2: callback start\r\ndebug2: client_session2_setup: id 2\r\ndebug1: Sending environment.\r\ndebug1: Sending env LANG = en_US.UTF-8\r\ndebug2: channel 2: request env confirm 0\r\ndebug3: send packet: type 98\r\ndebug1: Sending command: /bin/sh -c \'echo ~ && sleep 0\'\r\ndebug2: channel 2: request exec confirm 1\r\ndebug3: send packet: type 98\r\ndebug3: mux_session_confirm: sending success reply\r\ndebug2: callback done\r\ndebug2: channel 2: open confirm rwindow 0 rmax 32768\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug2: channel 2: rcvd adjust 2097152\r\ndebug3: receive packet: type 99\r\ndebug2: channel_input_status_confirm: type 99 id 2\r\ndebug2: exec request accepted on channel 2\r\ndebug3: receive packet: type 98\r\ndebug1: client_input_channel_req: channel 2 rtype exit-status reply 0\r\ndebug3: mux_exit_message: channel 2: exit message, exitval 0\r\ndebug3: receive packet: type 98\r\ndebug1: client_input_channel_req: channel 2 rtype eow@openssh.com reply 0\r\ndebug2: channel 2: rcvd eow\r\ndebug2: channel 2: close_read\r\ndebug2: channel 2: input open -> closed\r\ndebug3: receive packet: type 96\r\ndebug2: channel 2: rcvd eof\r\ndebug2: channel 2: output open -> drain\r\ndebug2: channel 2: obuf empty\r\ndebug2: channel 2: close_write\r\ndebug2: channel 2: output drain -> closed\r\ndebug3: receive packet: type 97\r\ndebug2: channel 2: rcvd close\r\ndebug3: channel 2: will not send data after close\r\ndebug2: channel 2: send close\r\ndebug3: send packet: type 97\r\ndebug2: channel 2: is dead\r\ndebug2: channel 2: gc: notify user\r\ndebug3: mux_master_session_cleanup_cb: entering for channel 2\r\ndebug2: channel 1: rcvd close\r\ndebug2: channel 1: output open -> drain\r\ndebug2: channel 1: close_read\r\ndebug2: channel 1: input open -> closed\r\ndebug2: channel 2: gc: user detached\r\ndebug2: channel 2: is dead\r\ndebug2: channel 2: garbage collecting\r\ndebug1: channel 2: free: client-session, nchannels 3\r\ndebug3: channel 2: status: The following connections are open:\r\n #2 client-session (t4 r0 i3/0 o3/0 fd -1/-1 cc -1)\r\n\r\ndebug2: channel 1: obuf empty\r\ndebug2: channel 1: close_write\r\ndebug2: channel 1: output drain -> closed\r\ndebug2: channel 1: is dead (local)\r\ndebug2: channel 1: gc: notify user\r\ndebug3: mux_master_control_cleanup_cb: entering for channel 1\r\ndebug2: channel 1: gc: user detached\r\ndebug2: channel 1: is dead (local)\r\ndebug2: channel 1: garbage collecting\r\ndebug1: channel 1: free: mux-control, nchannels 2\r\ndebug3: channel 1: status: The following connections are open:\r\n\r\ndebug2: set_control_persist_exit_time: schedule exit in 60 seconds\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 spamaps.org '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351 `" && echo ansible-tmp-1500415665.0875719-218494642853351="` echo /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351 `" ) && sleep 0'"'"'' <spamaps.org> (0, b'ansible-tmp-1500415665.0875719-218494642853351=/home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> PUT /tmp/tmpbaaj5_y1 TO /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set sftp_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC sftp -b - -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 '[spamaps.org]' <spamaps.org> (0, b'sftp> put /tmp/tmpbaaj5_y1 /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug2: Remote version: 3\r\ndebug2: Server supports extension "posix-rename@openssh.com" revision 1\r\ndebug2: Server supports extension "statvfs@openssh.com" revision 2\r\ndebug2: Server supports extension "fstatvfs@openssh.com" revision 2\r\ndebug3: Sent message fd 3 T:16 I:1\r\ndebug3: SSH_FXP_REALPATH . -> /home/clint size 0\r\ndebug3: Looking up /tmp/tmpbaaj5_y1\r\ndebug3: Sent message fd 3 T:17 I:2\r\ndebug3: Received stat reply T:101 I:2\r\ndebug1: Couldn\'t stat remote file: No such file or directory\r\ndebug3: Sent message SSH2_FXP_OPEN I:3 P:/home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py\r\ndebug3: Sent message SSH2_FXP_WRITE I:4 O:0 S:32768\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 4 32768 bytes at 0\r\ndebug3: Sent message SSH2_FXP_WRITE I:5 O:32768 S:32768\r\ndebug3: Sent message SSH2_FXP_WRITE I:6 O:65536 S:32768\r\ndebug3: Sent message SSH2_FXP_WRITE I:7 O:98304 S:5343\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 5 32768 bytes at 32768\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 6 32768 bytes at 65536\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: In write loop, ack for 7 5343 bytes at 98304\r\ndebug3: Sent message SSH2_FXP_CLOSE I:4\r\ndebug3: SSH2_FXP_STATUS 0\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 spamaps.org '/bin/sh -c '"'"'chmod u+x /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/ /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py && sleep 0'"'"'' <spamaps.org> (0, b'', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\n') <spamaps.org> ESTABLISH SSH CONNECTION FOR USER: None <spamaps.org> SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) <spamaps.org> SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) <spamaps.org> SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) <spamaps.org> SSH: PlayContext set ssh_common_args: () <spamaps.org> SSH: PlayContext set ssh_extra_args: () <spamaps.org> SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/80e720ccb6) <spamaps.org> SSH: EXEC ssh -vvv -C -o ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/clint/.ansible/cp/80e720ccb6 -tt spamaps.org '/bin/sh -c '"'"'/usr/bin/python /home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/setup.py; rm -rf "/home/clint/.ansible/tmp/ansible-tmp-1500415665.0875719-218494642853351/" > /dev/null 2>&1 && sleep 0'"'"'' <spamaps.org> (0, b'/tmp/ansible_sKLfd8/ansible_modlib.zip/ansible/module_utils/facts.py:1019: DeprecationWarning: object.__new__() takes no parameters\r\n/tmp/ansible_sKLfd8/ansible_modlib.zip/ansible/module_utils/facts.py:2438: DeprecationWarning: object.__new__() takes no parameters\r\n/tmp/ansible_sKLfd8/ansible_modlib.zip/ansible/module_utils/facts.py:3342: DeprecationWarning: object.__new__() takes no parameters\r\n\r\n{"invocation": {"module_args": {"filter": "*", "gather_subset": ["all"], "fact_path": "/etc/ansible/facts.d", "gather_timeout": 10}}, "changed": false, "ansible_facts": {"ansible_product_serial": "NA", "ansible_form_factor": "NA", "ansible_product_version": "NA", "ansible_fips": false, "ansible_service_mgr": "upstart", "ansible_user_id": "clint", "ansible_user_dir": "/home/clint", "ansible_memtotal_mb": 1459, "ansible_system_capabilities": [""], "ansible_distribution_version": "6.9", "ansible_domain": "srihosting.com", "ansible_date_time": {"weekday_number": "2", "iso8601_basic_short": "20170718T150754", "tz": "PDT", "weeknumber": "29", "hour": "15", "year": "2017", "minute": "07", "tz_offset": "-0700", "month": "07", "epoch": "1500415674", "iso8601_micro": "2017-07-18T22:07:54.378020Z", "weekday": "Tuesday", "time": "15:07:54", "date": "2017-07-18", "iso8601": "2017-07-18T22:07:54Z", "day": "18", "iso8601_basic": "20170718T150754377747", "second": "54"}, "ansible_real_user_id": 500, "ansible_processor_cores": 1, "ansible_virtualization_role": "guest", "ansible_dns": {"nameservers": ["8.8.8.8", "8.8.4.4"]}, "ansible_processor_vcpus": 1, "ansible_bios_version": "NA", "ansible_processor": ["GenuineIntel", "Intel(R) Xeon(R) CPU L5630 @ 2.13GHz"], "ansible_virtualization_type": "xen", "ansible_lo": {"features": {"tx_checksum_ipv4": "off [fixed]", "generic_receive_offload": "on", "tx_checksum_ipv6": "off [fixed]", "tx_scatter_gather_fraglist": "on [fixed]", "tx_checksum_unneeded": "off [fixed]", "highdma": "on [fixed]", "tx_lockless": "on [fixed]", "tx_tcp_ecn_segmentation": "on", "tx_gso_robust": "off [fixed]", "tx_checksumming": "on", "vlan_challenged": "on [fixed]", "loopback": "on [fixed]", "fcoe_mtu": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "large_receive_offload": "off [fixed]", "tx_scatter_gather": "on [fixed]", "rx_checksumming": "on [fixed]", "tx_tcp_segmentation": "on", "netns_local": "on [fixed]", "generic_segmentation_offload": "on", "tx_udp_tnl_segmentation": "off [fixed]", "tcp_segmentation_offload": "on", "rx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_tcp6_segmentation": "on", "udp_fragmentation_offload": "on", "scatter_gather": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "on [fixed]", "tx_fcoe_segmentation": "off [fixed]", "rx_vlan_filter": "off [fixed]", "receive_hashing": "off [fixed]", "tx_gre_segmentation": "off [fixed]"}, "mtu": 65536, "device": "lo", "promisc": false, "ipv4": {"broadcast": "host", "netmask": "255.0.0.0", "network": "127.0.0.0", "address": "127.0.0.1"}, "ipv6": [{"scope": "host", "prefix": "128", "address": "::1"}], "active": true, "type": "loopback"}, "ansible_userspace_bits": "64", "ansible_architecture": "x86_64", "ansible_default_ipv4": {"macaddress": "00:16:3e:1f:ea:f2", "network": "184.105.137.192", "mtu": 1500, "broadcast": "184.105.137.255", "alias": "eth0", "netmask": "255.255.255.192", "address": "184.105.137.237", "interface": "eth0", "type": "ether", "gateway": "184.105.137.193"}, "ansible_swapfree_mb": 0, "ansible_default_ipv6": {}, "ansible_distribution_release": "Final", "ansible_system_vendor": "NA", "ansible_apparmor": {"status": "disabled"}, "ansible_cmdline": {"LANG": "en_US.UTF-8", "rd_NO_LUKS": true, "ro": true, "console": "hvc0", "rd_NO_MD": true, "quiet": true, "rd_LVM_LV": "XenVolGroup/lv_root", "rhgb": true, "KEYTABLE": "us", "crashkernel": "auto", "SYSFONT": "latarcyrheb-sun16", "root": "/dev/mapper/XenVolGroup-lv_root", "rd_NO_DM": true}, "ansible_effective_user_id": 500, "ansible_mounts": [{"uuid": "N/A", "size_total": 31602700288, "mount": "/", "size_available": 5297983488, "fstype": "ext3", "device": "/dev/mapper/XenVolGroup-lv_root", "options": "rw"}, {"uuid": "N/A", "size_total": 101529600, "mount": "/boot", "size_available": 30937088, "fstype": "ext3", "device": "/dev/xvda1", "options": "rw"}], "ansible_selinux": {"status": "disabled"}, "ansible_os_family": "RedHat", "ansible_userspace_architecture": "x86_64", "ansible_product_uuid": "NA", "ansible_product_name": "NA", "ansible_pkg_mgr": "yum", "ansible_memfree_mb": 60, "ansible_devices": {"xvda": {"scheduler_mode": "cfq", "rotational": "0", "vendor": null, "sectors": "94371840", "sas_device_handle": null, "sas_address": null, "host": "", "sectorsize": "512", "removable": "0", "support_discard": "0", "model": null, "partitions": {"xvda1": {"sectorsize": 512, "uuid": "8703c7d1-ebe2-4c8b-8205-4a3ea341e3ce", "sectors": "204800", "start": "2048", "holders": [], "size": "100.00 MB"}, "xvda2": {"sectorsize": 512, "uuid": null, "sectors": "62703692", "start": "206848", "holders": ["XenVolGroup-lv_root"], "size": "29.90 GB"}}, "holders": [], "size": "45.00 GB"}}, "ansible_user_uid": 500, "ansible_memory_mb": {"real": {"total": 1459, "used": 1399, "free": 60}, "swap": {"cached": 0, "total": 0, "free": 0, "used": 0}, "nocache": {"used": 594, "free": 865}}, "ansible_distribution": "CentOS", "ansible_env": {"LANG": "en_US.UTF-8", "TERM": "xterm-256color", "SHELL": "/bin/bash", "SHLVL": "2", "SSH_TTY": "/dev/pts/3", "G_BROKEN_FILENAMES": "1", "SSH_CLIENT": "172.250.7.7 45782 22", "LESSOPEN": "||/usr/bin/lesspipe.sh %s", "PWD": "/home/clint", "CVS_RSH": "ssh", "LOGNAME": "clint", "USER": "clint", "MANPATH": "/usr/lib/courier-imap/man:", "PATH": "/usr/lib/courier-imap/bin:/usr/local/bin:/bin:/usr/bin", "MAIL": "/var/mail/clint", "LS_COLORS": "rs=0:di=38;5;27:ln=38;5;51:mh=44;38;5;15:pi=40;38;5;11:so=38;5;13:do=38;5;5:bd=48;5;232;38;5;11:cd=48;5;232;38;5;3:or=48;5;232;38;5;9:mi=05;48;5;232;38;5;15:su=48;5;196;38;5;15:sg=48;5;11;38;5;16:ca=48;5;196;38;5;226:tw=48;5;10;38;5;16:ow=48;5;10;38;5;21:st=48;5;21;38;5;15:ex=38;5;34:*.tar=38;5;9:*.tgz=38;5;9:*.arj=38;5;9:*.taz=38;5;9:*.lzh=38;5;9:*.lzma=38;5;9:*.tlz=38;5;9:*.txz=38;5;9:*.zip=38;5;9:*.z=38;5;9:*.Z=38;5;9:*.dz=38;5;9:*.gz=38;5;9:*.lz=38;5;9:*.xz=38;5;9:*.bz2=38;5;9:*.tbz=38;5;9:*.tbz2=38;5;9:*.bz=38;5;9:*.tz=38;5;9:*.deb=38;5;9:*.rpm=38;5;9:*.jar=38;5;9:*.rar=38;5;9:*.ace=38;5;9:*.zoo=38;5;9:*.cpio=38;5;9:*.7z=38;5;9:*.rz=38;5;9:*.jpg=38;5;13:*.jpeg=38;5;13:*.gif=38;5;13:*.bmp=38;5;13:*.pbm=38;5;13:*.pgm=38;5;13:*.ppm=38;5;13:*.tga=38;5;13:*.xbm=38;5;13:*.xpm=38;5;13:*.tif=38;5;13:*.tiff=38;5;13:*.png=38;5;13:*.svg=38;5;13:*.svgz=38;5;13:*.mng=38;5;13:*.pcx=38;5;13:*.mov=38;5;13:*.mpg=38;5;13:*.mpeg=38;5;13:*.m2v=38;5;13:*.mkv=38;5;13:*.ogm=38;5;13:*.mp4=38;5;13:*.m4v=38;5;13:*.mp4v=38;5;13:*.vob=38;5;13:*.qt=38;5;13:*.nuv=38;5;13:*.wmv=38;5;13:*.asf=38;5;13:*.rm=38;5;13:*.rmvb=38;5;13:*.flc=38;5;13:*.avi=38;5;13:*.fli=38;5;13:*.flv=38;5;13:*.gl=38;5;13:*.dl=38;5;13:*.xcf=38;5;13:*.xwd=38;5;13:*.yuv=38;5;13:*.cgm=38;5;13:*.emf=38;5;13:*.axv=38;5;13:*.anx=38;5;13:*.ogv=38;5;13:*.ogx=38;5;13:*.aac=38;5;45:*.au=38;5;45:*.flac=38;5;45:*.mid=38;5;45:*.midi=38;5;45:*.mka=38;5;45:*.mp3=38;5;45:*.mpc=38;5;45:*.ogg=38;5;45:*.ra=38;5;45:*.wav=38;5;45:*.axa=38;5;45:*.oga=38;5;45:*.spx=38;5;45:*.xspf=38;5;45:", "HOME": "/home/clint", "_": "/usr/bin/python", "SSH_CONNECTION": "172.250.7.7 45782 184.105.137.237 22"}, "ansible_distribution_major_version": "6", "module_setup": true, "ansible_processor_count": 1, "ansible_hostname": "xencbyrum2", "ansible_effective_group_id": 500, "ansible_swaptotal_mb": 0, "ansible_lsb": {"release": "6.9", "major_release": "6", "codename": "Final", "id": "CentOS", "description": "CentOS release 6.9 (Final)"}, "ansible_real_group_id": 500, "ansible_bios_date": "NA", "ansible_all_ipv6_addresses": ["fe80::216:3eff:fe1f:eaf2"], "ansible_interfaces": ["lo", "eth0"], "ansible_uptime_seconds": 14524108, "ansible_machine_id": "8b82be7fd3ea5ff44379c1b60000001a", "ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAABIwAAAQEAyLKP0RyBHL4TLU5NlD3nzPBa6D/PVvCsgmNASkzFrqKKe52XrREnsI26JUnaSYIR9L7NOrshd99kxcYnxXpV/pLBXgzdae3r4vTDNDUwzAq3vlQXmSw18IN1REe1cGg8E8OawcOxuNDn69Lbll4e4vGriIyWgwTifqjAL2JPd+Orw6DpRYMG3MvDW7CUnF7wr6H1kC4AiviGtckmtFwP+hDTe8Q1C88pq1rfnasdjm9PwWbrJoJmqUq5m3aT3WwW27NJXz6QSgTnAF7UD9EZfU5/ixBdW2S1FcAP28pEQ3BoFRH0LuuadUdN6cLPCNPV07aTZ79Z3vH5KWRopjqLtQ==", "ansible_gather_subset": ["hardware", "network", "virtual"], "ansible_user_gecos": "Clint Byrum", "ansible_system_capabilities_enforced": "True", "ansible_python": {"executable": "/usr/bin/python", "version": {"micro": 6, "major": 2, "releaselevel": "final", "serial": 0, "minor": 6}, "type": "CPython", "has_sslcontext": false, "version_info": [2, 6, 6, "final", 0]}, "ansible_kernel": "2.6.32-642.11.1.el6.x86_64", "ansible_processor_threads_per_core": 1, "ansible_fqdn": "xencbyrum2.srihosting.com", "ansible_user_gid": 500, "ansible_eth0": {"macaddress": "00:16:3e:1f:ea:f2", "features": {"tx_checksum_ipv4": "on", "generic_receive_offload": "on", "tx_checksum_ipv6": "off", "tx_scatter_gather_fraglist": "off [fixed]", "tx_checksum_unneeded": "off", "highdma": "off [fixed]", "tx_lockless": "off [fixed]", "tx_tcp_ecn_segmentation": "off", "tx_gso_robust": "on [fixed]", "tx_checksumming": "on", "vlan_challenged": "off [fixed]", "loopback": "off [fixed]", "fcoe_mtu": "off [fixed]", "tx_checksum_sctp": "off [fixed]", "large_receive_offload": "off [fixed]", "tx_scatter_gather": "on", "rx_checksumming": "off [fixed]", "tx_tcp_segmentation": "on", "netns_local": "off [fixed]", "generic_segmentation_offload": "on", "tx_udp_tnl_segmentation": "off [fixed]", "tcp_segmentation_offload": "on", "rx_vlan_offload": "off [fixed]", "ntuple_filters": "off [fixed]", "tx_vlan_offload": "off [fixed]", "tx_tcp6_segmentation": "off", "udp_fragmentation_offload": "off [fixed]", "scatter_gather": "on", "tx_checksum_fcoe_crc": "off [fixed]", "tx_checksum_ip_generic": "off", "tx_fcoe_segmentation": "off [fixed]", "rx_vlan_filter": "off [fixed]", "receive_hashing": "off [fixed]", "tx_gre_segmentation": "off [fixed]"}, "pciid": "vif-0", "module": "xen_netfront", "mtu": 1500, "device": "eth0", "promisc": false, "ipv4": {"broadcast": "184.105.137.255", "netmask": "255.255.255.192", "network": "184.105.137.192", "address": "184.105.137.237"}, "ipv6": [{"scope": "link", "prefix": "64", "address": "fe80::216:3eff:fe1f:eaf2"}], "active": true, "type": "ether"}, "ansible_nodename": "xencbyrum2.srihosting.com", "ansible_system": "Linux", "ansible_user_shell": "/bin/bash", "ansible_machine": "x86_64", "ansible_ssh_host_key_dsa_public": "AAAAB3NzaC1kc3MAAACBAPOlY3CbmfqEerypkFdWJO6fNaz7c4HH4UCAeVQxVrpo+PoSqtTJHhB24gOmxicCcmkXgb6KJP16tWcpV47dfLCfNY078k24yFBRtAfIIXhZSGyXjPdcp3BTTQVkqRyR1ToS572FAgzi6it6wUc0psSXdaOJXdRqtqieFQUnR3Q5AAAAFQDDLXxM4aIcL5sYlSFRbLwCsIGAXQAAAIBv71rbqG53O2b9MKZ0MIGp2vDT+0H32SycdeoWk0YqTqOesjQaKz2OHEdFqcYf0nm8zLyiIuNTR0MLBq1mHt2QUbXi+T4pAxnseV5wLclgZY5Hydg6F5v86ixQjV9W5/unlL4MKcmLL62hGnJIykVHPXhJ2elPykwp3Oj0eurfOwAAAIEA7981Kzo+/PfYYoI2clhvjtxAEolpCMv/EpJ3N504LZ79K4ADrIS7PhI9lb8Bne6UqZo+9eoKm5JAnzvfyArb2jUOrz8BSU4Q0KVZnCaBkz0TtuFyKHcWpmasSIimAzckAtF+LRCzbz/8hJg9/1Gbp8Zq2HQl5O3Hxde+Ry9zdFc=", "ansible_all_ipv4_addresses": ["184.105.137.237"], "ansible_python_version": "2.6.6"}}\r\n', b'OpenSSH_7.2p2 Ubuntu-4ubuntu2.2, OpenSSL 1.0.2g 1 Mar 2016\r\ndebug1: Reading configuration data /home/clint/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 16637\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to spamaps.org closed.\r\n') ok: [spamaps.org] META: ran handlers TASK [stuff] ******************************************************************************************************************************************************************************************************* task path: /home/clint/test.yml:4 ok: [spamaps.org] => { "msg": "Here is a message" } SSH: ansible.cfg set ssh_args: (-C)(-o)(ControlMaster=auto)(-o)(ControlPersist=60s) SSH: ansible_password/ansible_ssh_pass not set: (-o)(KbdInteractiveAuthentication=no)(-o)(PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey)(-o)(PasswordAuthentication=no) SSH: ANSIBLE_TIMEOUT/timeout set: (-o)(ConnectTimeout=10) SSH: PlayContext set ssh_common_args: () SSH: PlayContext set ssh_extra_args: () SSH: found only ControlPersist; added ControlPath: (-o)(ControlPath=/home/clint/.ansible/cp/276d6f8b70) ERROR! Unexpected Exception: list index out of range the full traceback was: Traceback (most recent call last): File "/usr/local/bin/ansible-playbook", line 109, in <module> exit_code = cli.run() File "/usr/local/lib/python3.5/dist-packages/ansible/cli/playbook.py", line 154, in run results = pbex.run() File "/usr/local/lib/python3.5/dist-packages/ansible/executor/playbook_executor.py", line 153, in run result = self._tqm.run(play=play) File "/usr/local/lib/python3.5/dist-packages/ansible/executor/task_queue_manager.py", line 284, in run play_return = strategy.run(iterator, play_context) File "/usr/local/lib/python3.5/dist-packages/ansible/plugins/strategy/linear.py", line 221, in run results.extend(self._execute_meta(task, play_context, iterator, host)) File "/usr/local/lib/python3.5/dist-packages/ansible/plugins/strategy/__init__.py", line 905, in _execute_meta connection.reset() File "/usr/local/lib/python3.5/dist-packages/ansible/plugins/connection/ssh.py", line 816, in reset p = subprocess.Popen(cmd, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE) File "/usr/lib/python3.5/subprocess.py", line 947, in __init__ restore_signals, start_new_session) File "/usr/lib/python3.5/subprocess.py", line 1448, in _execute_child executable = args[0] IndexError: list index out of range $ ```
main
meta reset connection results in indexerror list index out of range verify first that your issue request is not already reported on github also test if the latest release and master branch are affected too issue type bug report component name meta module ansible version ansible playbook config file configured module search path default w o overrides python version default nov configuration no configuration os environment ubuntu with ansible pip installed running in contacting a vm running centos summary using meta reset connection works and does reset the connection but then later on this indexerror ocurrs steps to reproduce this playbook shows the problem name ping them hosts all tasks name stuff debug msg here is a message name reset conn meta reset connection expected results ansible playbook should exit without errors actual results tmp inv has this content spamaps org my user has ssh agent access to spamaps org ansible playbook test yml i tmp inv vvvvv no config file found using defaults loading callback plugin default of type stdout from usr local lib dist packages ansible plugins callback init py playbook test yml plays in test yml play task using module file usr local lib dist packages ansible modules system setup py establish ssh connection for user none ssh ansible cfg set ssh args c o controlmaster auto o controlpersist ssh ansible password ansible ssh pass not set o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no ssh ansible timeout timeout set o connecttimeout ssh playcontext set ssh common args ssh playcontext set ssh extra args ssh found only controlpersist added controlpath o controlpath home clint ansible cp ssh exec ssh vvv c o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home clint ansible cp spamaps org bin sh c echo sleep b home clint n b openssh ubuntu openssl mar r reading configuration data home clint ssh config r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r control socket home clint ansible cp does not exist r resolving spamaps org port r ssh connect direct needpriv r connecting to spamaps org port r fd setting o nonblock r fd clearing o nonblock r connection established r timeout ms remain after connect r identity file home clint ssh id rsa type r key load public no such file or directory r identity file home clint ssh id rsa cert type r key load public no such file or directory r identity file home clint ssh id dsa type r key load public no such file or directory r identity file home clint ssh id dsa cert type r key load public no such file or directory r identity file home clint ssh id ecdsa type r key load public no such file or directory r identity file home clint ssh id ecdsa cert type r key load public no such file or directory r identity file home clint ssh id type r key load public no such file or directory r identity file home clint ssh id cert type r enabling compatibility mode for protocol r local version string ssh openssh ubuntu r remote protocol version remote software version openssh r match openssh pat openssh compat r fd setting o nonblock r authenticating to spamaps org as clint r hostkeys foreach reading file home clint ssh known hosts r record hostkey found key type rsa in file home clint ssh known hosts r load hostkeys loaded keys from spamaps org r order hostkeyalgs prefer hostkeyalgs ssh rsa cert openssh com rsa rsa ssh rsa r send packet type r msg kexinit sent r receive packet type r msg kexinit received r local client kexinit proposal r kex algorithms libssh org ecdh ecdh ecdh diffie hellman group exchange diffie hellman group exchange diffie hellman ext info c r host key algorithms ssh rsa cert openssh com rsa rsa ssh rsa ecdsa cert openssh com ecdsa cert openssh com ecdsa cert openssh com ssh cert openssh com ecdsa ecdsa ecdsa ssh r ciphers ctos openssh com ctr ctr ctr gcm openssh com gcm openssh com cbc cbc cbc cbc r ciphers stoc openssh com ctr ctr ctr gcm openssh com gcm openssh com cbc cbc cbc cbc r macs ctos umac etm openssh com umac etm openssh com hmac etm openssh com hmac etm openssh com hmac etm openssh com umac openssh com umac openssh com hmac hmac hmac r macs stoc umac etm openssh com umac etm openssh com hmac etm openssh com hmac etm openssh com hmac etm openssh com umac openssh com umac openssh com hmac hmac hmac r compression ctos zlib openssh com zlib none r compression stoc zlib openssh com zlib none r languages ctos r languages stoc r first kex follows r reserved r peer server kexinit proposal r kex algorithms diffie hellman group exchange diffie hellman group exchange diffie hellman diffie hellman r host key algorithms ssh rsa ssh dss r ciphers ctos ctr ctr ctr cbc cbc blowfish cbc cbc cbc cbc arcfour rijndael cbc lysator liu se r ciphers stoc ctr ctr ctr cbc cbc blowfish cbc cbc cbc cbc arcfour rijndael cbc lysator liu se r macs ctos hmac hmac umac openssh com hmac hmac hmac hmac openssh com hmac hmac r macs stoc hmac hmac umac openssh com hmac hmac hmac hmac openssh com hmac hmac r compression ctos none zlib openssh com r compression stoc none zlib openssh com r languages ctos r languages stoc r first kex follows r reserved r kex algorithm diffie hellman group exchange r kex host key algorithm ssh rsa r kex server client cipher ctr mac umac openssh com compression zlib openssh com r kex client server cipher ctr mac umac openssh com compression zlib openssh com r send packet type r msg kex dh gex request closed r receive packet type r channel rcvd eof r channel output open drain r channel obuf empty r channel close write r channel output drain closed r receive packet type r channel rcvd close r channel will not send data after close r channel send close r send packet type r channel is dead r channel gc notify user r mux master session cleanup cb entering for channel r channel rcvd close r channel output open drain r channel close read r channel input open closed r channel gc user detached r channel is dead r channel garbage collecting r channel free client session nchannels r channel status the following connections are open r n client session fd cc r n r channel obuf empty r channel close write r channel output drain closed r channel is dead local r channel gc notify user r mux master control cleanup cb entering for channel r channel gc user detached r channel is dead local r channel garbage collecting r channel free mux control nchannels r channel status the following connections are open r n r set control persist exit time schedule exit in seconds r mux client read packet read header failed broken pipe r received exit status from master r n establish ssh connection for user none ssh ansible cfg set ssh args c o controlmaster auto o controlpersist ssh ansible password ansible ssh pass not set o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no ssh ansible timeout timeout set o connecttimeout ssh playcontext set ssh common args ssh playcontext set ssh extra args ssh found only controlpersist added controlpath o controlpath home clint ansible cp ssh exec ssh vvv c o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home clint ansible cp spamaps org bin sh c umask mkdir p echo home clint ansible tmp ansible tmp echo ansible tmp echo home clint ansible tmp ansible tmp sleep b ansible tmp home clint ansible tmp ansible tmp n b openssh ubuntu openssl mar r reading configuration data home clint ssh config r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r mux client read packet read header failed broken pipe r received exit status from master r n put tmp to home clint ansible tmp ansible tmp setup py ssh ansible cfg set ssh args c o controlmaster auto o controlpersist ssh ansible password ansible ssh pass not set o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no ssh ansible timeout timeout set o connecttimeout ssh playcontext set ssh common args ssh playcontext set sftp extra args ssh found only controlpersist added controlpath o controlpath home clint ansible cp ssh exec sftp b vvv c o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home clint ansible cp b sftp put tmp home clint ansible tmp ansible tmp setup py n b openssh ubuntu openssl mar r reading configuration data home clint ssh config r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r remote version r server supports extension posix rename openssh com revision r server supports extension statvfs openssh com revision r server supports extension fstatvfs openssh com revision r sent message fd t i r ssh fxp realpath home clint size r looking up tmp r sent message fd t i r received stat reply t i r couldn t stat remote file no such file or directory r sent message fxp open i p home clint ansible tmp ansible tmp setup py r sent message fxp write i o s r fxp status r in write loop ack for bytes at r sent message fxp write i o s r sent message fxp write i o s r sent message fxp write i o s r fxp status r in write loop ack for bytes at r fxp status r in write loop ack for bytes at r fxp status r in write loop ack for bytes at r sent message fxp close i r fxp status r mux client read packet read header failed broken pipe r received exit status from master r n establish ssh connection for user none ssh ansible cfg set ssh args c o controlmaster auto o controlpersist ssh ansible password ansible ssh pass not set o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no ssh ansible timeout timeout set o connecttimeout ssh playcontext set ssh common args ssh playcontext set ssh extra args ssh found only controlpersist added controlpath o controlpath home clint ansible cp ssh exec ssh vvv c o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home clint ansible cp spamaps org bin sh c chmod u x home clint ansible tmp ansible tmp home clint ansible tmp ansible tmp setup py sleep b b openssh ubuntu openssl mar r reading configuration data home clint ssh config r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r mux client read packet read header failed broken pipe r received exit status from master r n establish ssh connection for user none ssh ansible cfg set ssh args c o controlmaster auto o controlpersist ssh ansible password ansible ssh pass not set o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no ssh ansible timeout timeout set o connecttimeout ssh playcontext set ssh common args ssh playcontext set ssh extra args ssh found only controlpersist added controlpath o controlpath home clint ansible cp ssh exec ssh vvv c o controlmaster auto o controlpersist o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home clint ansible cp tt spamaps org bin sh c usr bin python home clint ansible tmp ansible tmp setup py rm rf home clint ansible tmp ansible tmp dev null sleep b tmp ansible ansible modlib zip ansible module utils facts py deprecationwarning object new takes no parameters r n tmp ansible ansible modlib zip ansible module utils facts py deprecationwarning object new takes no parameters r n tmp ansible ansible modlib zip ansible module utils facts py deprecationwarning object new takes no parameters r n r n invocation module args filter gather subset fact path etc ansible facts d gather timeout changed false ansible facts ansible product serial na ansible form factor na ansible product version na ansible fips false ansible service mgr upstart ansible user id clint ansible user dir home clint ansible memtotal mb ansible system capabilities ansible distribution version ansible domain srihosting com ansible date time weekday number basic short tz pdt weeknumber hour year minute tz offset month epoch micro weekday tuesday time date day basic second ansible real user id ansible processor cores ansible virtualization role guest ansible dns nameservers ansible processor vcpus ansible bios version na ansible processor ansible virtualization type xen ansible lo features tx checksum off generic receive offload on tx checksum off tx scatter gather fraglist on tx checksum unneeded off highdma on tx lockless on tx tcp ecn segmentation on tx gso robust off tx checksumming on vlan challenged on loopback on fcoe mtu off tx checksum sctp off large receive offload off tx scatter gather on rx checksumming on tx tcp segmentation on netns local on generic segmentation offload on tx udp tnl segmentation off tcp segmentation offload on rx vlan offload off ntuple filters off tx vlan offload off tx segmentation on udp fragmentation offload on scatter gather on tx checksum fcoe crc off tx checksum ip generic on tx fcoe segmentation off rx vlan filter off receive hashing off tx gre segmentation off mtu device lo promisc false broadcast host netmask network address active true type loopback ansible userspace bits ansible architecture ansible default macaddress ea network mtu broadcast alias netmask address interface type ether gateway ansible swapfree mb ansible default ansible distribution release final ansible system vendor na ansible apparmor status disabled ansible cmdline lang en us utf rd no luks true ro true console rd no md true quiet true rd lvm lv xenvolgroup lv root rhgb true keytable us crashkernel auto sysfont latarcyrheb root dev mapper xenvolgroup lv root rd no dm true ansible effective user id ansible mounts ansible selinux status disabled ansible os family redhat ansible userspace architecture ansible product uuid na ansible product name na ansible pkg mgr yum ansible memfree mb ansible devices xvda scheduler mode cfq rotational vendor null sectors sas device handle null sas address null host sectorsize removable support discard model null partitions sectorsize uuid sectors start holders size mb sectorsize uuid null sectors start holders size gb holders size gb ansible user uid ansible memory mb real total used free swap cached total free used nocache used free ansible distribution centos ansible env lang en us utf term xterm shell bin bash shlvl ssh tty dev pts g broken filenames ssh client lessopen usr bin lesspipe sh s pwd home clint cvs rsh ssh logname clint user clint manpath usr lib courier imap man path usr lib courier imap bin usr local bin bin usr bin mail var mail clint ls colors rs di ln mh pi so do bd cd or mi su sg ca tw ow st ex tar tgz arj taz lzh lzma tlz txz zip z z dz gz lz xz tbz bz tz deb rpm jar rar ace zoo cpio rz jpg jpeg gif bmp pbm pgm ppm tga xbm xpm tif tiff png svg svgz mng pcx mov mpg mpeg mkv ogm vob qt nuv wmv asf rm rmvb flc avi fli flv gl dl xcf xwd yuv cgm emf axv anx ogv ogx aac au flac mid midi mka mpc ogg ra wav axa oga spx xspf home home clint usr bin python ssh connection ansible distribution major version module setup true ansible processor count ansible hostname ansible effective group id ansible swaptotal mb ansible lsb release major release codename final id centos description centos release final ansible real group id ansible bios date na ansible all addresses ansible interfaces ansible uptime seconds ansible machine id ansible ssh host key rsa public ansible gather subset ansible user gecos clint byrum ansible system capabilities enforced true ansible python executable usr bin python version micro major releaselevel final serial minor type cpython has sslcontext false version info ansible kernel ansible processor threads per core ansible fqdn srihosting com ansible user gid ansible macaddress ea features tx checksum on generic receive offload on tx checksum off tx scatter gather fraglist off tx checksum unneeded off highdma off tx lockless off tx tcp ecn segmentation off tx gso robust on tx checksumming on vlan challenged off loopback off fcoe mtu off tx checksum sctp off large receive offload off tx scatter gather on rx checksumming off tx tcp segmentation on netns local off generic segmentation offload on tx udp tnl segmentation off tcp segmentation offload on rx vlan offload off ntuple filters off tx vlan offload off tx segmentation off udp fragmentation offload off scatter gather on tx checksum fcoe crc off tx checksum ip generic off tx fcoe segmentation off rx vlan filter off receive hashing off tx gre segmentation off pciid vif module xen netfront mtu device promisc false broadcast netmask network address active true type ether ansible nodename srihosting com ansible system linux ansible user shell bin bash ansible machine ansible ssh host key dsa public lrczbz ansible all addresses ansible python version r n b openssh ubuntu openssl mar r reading configuration data home clint ssh config r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r mux client read packet read header failed broken pipe r received exit status from master r nshared connection to spamaps org closed r n ok meta ran handlers task task path home clint test yml ok msg here is a message ssh ansible cfg set ssh args c o controlmaster auto o controlpersist ssh ansible password ansible ssh pass not set o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no ssh ansible timeout timeout set o connecttimeout ssh playcontext set ssh common args ssh playcontext set ssh extra args ssh found only controlpersist added controlpath o controlpath home clint ansible cp error unexpected exception list index out of range the full traceback was traceback most recent call last file usr local bin ansible playbook line in exit code cli run file usr local lib dist packages ansible cli playbook py line in run results pbex run file usr local lib dist packages ansible executor playbook executor py line in run result self tqm run play play file usr local lib dist packages ansible executor task queue manager py line in run play return strategy run iterator play context file usr local lib dist packages ansible plugins strategy linear py line in run results extend self execute meta task play context iterator host file usr local lib dist packages ansible plugins strategy init py line in execute meta connection reset file usr local lib dist packages ansible plugins connection ssh py line in reset p subprocess popen cmd stdin subprocess pipe stdout subprocess pipe stderr subprocess pipe file usr lib subprocess py line in init restore signals start new session file usr lib subprocess py line in execute child executable args indexerror list index out of range
1
6,050
2,610,220,080
IssuesEvent
2015-02-26 19:09:55
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ .doc
auto-migrated Priority-Medium Type-Defect
``` '''Π’Π°Π½Π°Π΄ΠΈΠΉ Π‘ΠΌΠΈΡ€Π½ΠΎΠ²''' ΠŸΡ€ΠΈΠ²Π΅Ρ‚ всСм Π½Π΅ подскаТитС Π³Π΄Π΅ ΠΌΠΎΠΆΠ½ΠΎ Π½Π°ΠΉΡ‚ΠΈ .Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ .doc. Π³Π΄Π΅ Ρ‚ΠΎ Π²ΠΈΠ΄Π΅Π» ΡƒΠΆΠ΅ '''ΠΠ»ΡŒΠ±Π΅Ρ€Ρ‚ Казаков''' ΠšΠ°Ρ‡Π°ΠΉ Ρ‚ΡƒΡ‚ http://bit.ly/1cDi5ZT '''АтСист ΠœΠΈΡ…Π°ΠΉΠ»ΠΎΠ²''' ΠŸΡ€ΠΎΡΠΈΡ‚ ввСсти Π½ΠΎΠΌΠ΅Ρ€ ΠΌΠΎΠ±ΠΈΠ»Ρ‹!НС опасно Π»ΠΈ это? '''Π’Π°Ρ€Π»Π°ΠΌ Π§Π΅Ρ€Π½ΠΎΠ²''' НСа всС ΠΎΠΊ Ρƒ мСня Π½ΠΈΡ‡Π΅Π³ΠΎ Π½Π΅ списало '''Π“Π΅Π»ΠΈΠ°Π½ Π›Π°Π²Ρ€Π΅Π½Ρ‚ΡŒΠ΅Π²''' НСа всС ΠΎΠΊ Ρƒ мСня Π½ΠΈΡ‡Π΅Π³ΠΎ Π½Π΅ списало Π˜Π½Ρ„ΠΎΡ€ΠΌΠ°Ρ†ΠΈΡ ΠΎ Ρ„Π°ΠΉΠ»Π΅: Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ .doc Π—Π°Π³Ρ€ΡƒΠΆΠ΅Π½: Π’ этом мСсяцС Π‘ΠΊΠ°Ρ‡Π°Π½ Ρ€Π°Π·: 687 Π Π΅ΠΉΡ‚ΠΈΠ½Π³: 1041 БрСдняя ΡΠΊΠΎΡ€ΠΎΡΡ‚ΡŒ скачивания: 106 ΠŸΠΎΡ…ΠΎΠΆΠΈΡ… Ρ„Π°ΠΉΠ»ΠΎΠ²: 34 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 5:48
1.0
Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ .doc - ``` '''Π’Π°Π½Π°Π΄ΠΈΠΉ Π‘ΠΌΠΈΡ€Π½ΠΎΠ²''' ΠŸΡ€ΠΈΠ²Π΅Ρ‚ всСм Π½Π΅ подскаТитС Π³Π΄Π΅ ΠΌΠΎΠΆΠ½ΠΎ Π½Π°ΠΉΡ‚ΠΈ .Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ .doc. Π³Π΄Π΅ Ρ‚ΠΎ Π²ΠΈΠ΄Π΅Π» ΡƒΠΆΠ΅ '''ΠΠ»ΡŒΠ±Π΅Ρ€Ρ‚ Казаков''' ΠšΠ°Ρ‡Π°ΠΉ Ρ‚ΡƒΡ‚ http://bit.ly/1cDi5ZT '''АтСист ΠœΠΈΡ…Π°ΠΉΠ»ΠΎΠ²''' ΠŸΡ€ΠΎΡΠΈΡ‚ ввСсти Π½ΠΎΠΌΠ΅Ρ€ ΠΌΠΎΠ±ΠΈΠ»Ρ‹!НС опасно Π»ΠΈ это? '''Π’Π°Ρ€Π»Π°ΠΌ Π§Π΅Ρ€Π½ΠΎΠ²''' НСа всС ΠΎΠΊ Ρƒ мСня Π½ΠΈΡ‡Π΅Π³ΠΎ Π½Π΅ списало '''Π“Π΅Π»ΠΈΠ°Π½ Π›Π°Π²Ρ€Π΅Π½Ρ‚ΡŒΠ΅Π²''' НСа всС ΠΎΠΊ Ρƒ мСня Π½ΠΈΡ‡Π΅Π³ΠΎ Π½Π΅ списало Π˜Π½Ρ„ΠΎΡ€ΠΌΠ°Ρ†ΠΈΡ ΠΎ Ρ„Π°ΠΉΠ»Π΅: Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ .doc Π—Π°Π³Ρ€ΡƒΠΆΠ΅Π½: Π’ этом мСсяцС Π‘ΠΊΠ°Ρ‡Π°Π½ Ρ€Π°Π·: 687 Π Π΅ΠΉΡ‚ΠΈΠ½Π³: 1041 БрСдняя ΡΠΊΠΎΡ€ΠΎΡΡ‚ΡŒ скачивания: 106 ΠŸΠΎΡ…ΠΎΠΆΠΈΡ… Ρ„Π°ΠΉΠ»ΠΎΠ²: 34 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 17 Dec 2013 at 5:48
non_main
Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ doc Π²Π°Π½Π°Π΄ΠΈΠΉ смирнов ΠΏΡ€ΠΈΠ²Π΅Ρ‚ всСм Π½Π΅ подскаТитС Π³Π΄Π΅ ΠΌΠΎΠΆΠ½ΠΎ Π½Π°ΠΉΡ‚ΠΈ Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ doc Π³Π΄Π΅ Ρ‚ΠΎ Π²ΠΈΠ΄Π΅Π» ΡƒΠΆΠ΅ Π°Π»ΡŒΠ±Π΅Ρ€Ρ‚ ΠΊΠ°Π·Π°ΠΊΠΎΠ² ΠΊΠ°Ρ‡Π°ΠΉ Ρ‚ΡƒΡ‚ атСист ΠΌΠΈΡ…Π°ΠΉΠ»ΠΎΠ² просит ввСсти Π½ΠΎΠΌΠ΅Ρ€ ΠΌΠΎΠ±ΠΈΠ»Ρ‹ Π½Π΅ опасно Π»ΠΈ это Π²Π°Ρ€Π»Π°ΠΌ Ρ‡Π΅Ρ€Π½ΠΎΠ² Π½Π΅Π° всС ΠΎΠΊ Ρƒ мСня Π½ΠΈΡ‡Π΅Π³ΠΎ Π½Π΅ списало Π³Π΅Π»ΠΈΠ°Π½ Π»Π°Π²Ρ€Π΅Π½Ρ‚ΡŒΠ΅Π² Π½Π΅Π° всС ΠΎΠΊ Ρƒ мСня Π½ΠΈΡ‡Π΅Π³ΠΎ Π½Π΅ списало информация ΠΎ Ρ„Π°ΠΉΠ»Π΅ Π½Π΅ΡƒΠΌΡ‹Π²Π°ΠΊΠΈΠ½ ΠΈΠ²Π°Π½ ΠΏΠ°Π²Π»ΠΎΠ²ΠΈΡ‡ ΠΊΠ½ΠΈΠ³ΠΈ doc Π·Π°Π³Ρ€ΡƒΠΆΠ΅Π½ Π² этом мСсяцС скачан Ρ€Π°Π· Ρ€Π΅ΠΉΡ‚ΠΈΠ½Π³ срСдняя ΡΠΊΠΎΡ€ΠΎΡΡ‚ΡŒ скачивания ΠΏΠΎΡ…ΠΎΠΆΠΈΡ… Ρ„Π°ΠΉΠ»ΠΎΠ² original issue reported on code google com by kondense gmail com on dec at
0
368,253
10,868,929,291
IssuesEvent
2019-11-15 05:50:57
Gravecat/Duskfall
https://api.github.com/repos/Gravecat/Duskfall
opened
Port over video poker minigame from Krasten
enhancement low priority no help needed
Another to-do entry for myself: The video-poker minigame in Krasten is hella cool and should be ported across. For the sake of theme, it can just be reworked a little to be Goblin Poker or something.
1.0
Port over video poker minigame from Krasten - Another to-do entry for myself: The video-poker minigame in Krasten is hella cool and should be ported across. For the sake of theme, it can just be reworked a little to be Goblin Poker or something.
non_main
port over video poker minigame from krasten another to do entry for myself the video poker minigame in krasten is hella cool and should be ported across for the sake of theme it can just be reworked a little to be goblin poker or something
0
5,517
27,577,989,589
IssuesEvent
2023-03-08 14:22:41
cosmos/ibc-rs
https://api.github.com/repos/cosmos/ibc-rs
opened
[Transfer App] Refine return type of `process_recv_packet_execute`
O: maintainability
<!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Problem Statement The current implementation of the `process_recv_packet_execute` in the relay module returns a `Result<ModuleExtras, (ModuleExtras, TokenTransferError)>` type which is not conventional and may cause confusion to future maintainers. ## Proposal It is suggested to update the implementation to return `Result<ModuleExtras, TokenTransferError>` directly, where `TokenTransferError` contains all the necessary error information. If `ModuleExtras` is needed to handle the error case, it can be returned as part of the error variant.
True
[Transfer App] Refine return type of `process_recv_packet_execute` - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺ v ✰ Thanks for opening an issue! ✰ v Before smashing the submit button please review the template. v Please also ensure that this is not a duplicate issue :) ☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > --> ## Problem Statement The current implementation of the `process_recv_packet_execute` in the relay module returns a `Result<ModuleExtras, (ModuleExtras, TokenTransferError)>` type which is not conventional and may cause confusion to future maintainers. ## Proposal It is suggested to update the implementation to return `Result<ModuleExtras, TokenTransferError>` directly, where `TokenTransferError` contains all the necessary error information. If `ModuleExtras` is needed to handle the error case, it can be returned as part of the error variant.
main
refine return type of process recv packet execute ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ problem statement the current implementation of the process recv packet execute in the relay module returns a result type which is not conventional and may cause confusion to future maintainers proposal it is suggested to update the implementation to return result directly where tokentransfererror contains all the necessary error information if moduleextras is needed to handle the error case it can be returned as part of the error variant
1
2,166
7,551,050,416
IssuesEvent
2018-04-18 18:51:00
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
There is no plastitanium window spawner
Maintainability/Hinders improvements
There is no window spawner for plastitanium windows with a grille placed underneath. Very easy to implement, the main issue is that there's no sprite for this in structures_spawners.dmi.
True
There is no plastitanium window spawner - There is no window spawner for plastitanium windows with a grille placed underneath. Very easy to implement, the main issue is that there's no sprite for this in structures_spawners.dmi.
main
there is no plastitanium window spawner there is no window spawner for plastitanium windows with a grille placed underneath very easy to implement the main issue is that there s no sprite for this in structures spawners dmi
1
2,205
7,781,409,387
IssuesEvent
2018-06-06 00:04:02
Microsoft/DirectXTK12
https://api.github.com/repos/Microsoft/DirectXTK12
opened
Remove AlignedNew.h helper
maintainence
VS 2015 supports ``alignas``, but does not support aligned new. VS 2017 in the 15.5 update added support for C++17 aligned new, which when present we could use instead of the ``AlignedNew.h`` helper. See [this speclet](http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2016/p0035r4.html)
True
Remove AlignedNew.h helper - VS 2015 supports ``alignas``, but does not support aligned new. VS 2017 in the 15.5 update added support for C++17 aligned new, which when present we could use instead of the ``AlignedNew.h`` helper. See [this speclet](http://www.open-std.org/jtc1/sc22/wg21/docs/papers/2016/p0035r4.html)
main
remove alignednew h helper vs supports alignas but does not support aligned new vs in the update added support for c aligned new which when present we could use instead of the alignednew h helper see
1
3,922
2,937,997,943
IssuesEvent
2015-07-01 07:51:43
ndomar/megasoft-13
https://api.github.com/repos/ndomar/megasoft-13
closed
C4S5 Versioning
code-verified Component-4 doc-verified Points-13 Priority-Low scenario-verified Status-new
Success: As a designer: I can load a previous version of my created project. I can save the current version of my project. Failure: Not a signed in user, not his project, flashed with error.
1.0
C4S5 Versioning - Success: As a designer: I can load a previous version of my created project. I can save the current version of my project. Failure: Not a signed in user, not his project, flashed with error.
non_main
versioning success as a designer i can load a previous version of my created project i can save the current version of my project failure not a signed in user not his project flashed with error
0
4,326
21,779,668,096
IssuesEvent
2022-05-13 17:20:46
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
npm audit found vulnerabilities
type: bug work: frontend status: blocked restricted: maintainers
Edit: **Blocked because our direct dependencies have not been updated yet (eg., Storybook).** ([discussion](https://matrix.to/#/!UnujZDUxGuMrYdvgTU:matrix.mathesar.org/$lQjgziZVH6pbMYGIKRkEE3TC6ApAV05doQDtuaVc2LM?via=matrix.mathesar.org&via=matrix.org)) ``` # npm audit report ansi-html * Severity: high Uncontrolled Resource Consumption in ansi-html - https://github.com/advisories/GHSA-whgm-jr23-g3j9 fix available via `npm audit fix` node_modules/ansi-html webpack-hot-middleware 2.9.0 - 2.25.0 Depends on vulnerable versions of ansi-html node_modules/webpack-hot-middleware ansi-regex >2.1.1 <5.0.1 Severity: moderate Inefficient Regular Expression Complexity in chalk/ansi-regex - https://github.com/advisories/GHSA-93q8-gq69-wqmw fix available via `npm audit fix` node_modules/ansi-align/node_modules/ansi-regex node_modules/ansi-regex node_modules/sveltedoc-parser/node_modules/ansi-regex node_modules/wide-align/node_modules/ansi-regex strip-ansi 4.0.0 - 5.2.0 Depends on vulnerable versions of ansi-regex node_modules/ansi-align/node_modules/strip-ansi node_modules/sveltedoc-parser/node_modules/string-width/node_modules/strip-ansi node_modules/wide-align/node_modules/strip-ansi string-width 2.1.0 - 4.1.0 Depends on vulnerable versions of strip-ansi node_modules/ansi-align/node_modules/string-width node_modules/sveltedoc-parser/node_modules/string-width node_modules/wide-align/node_modules/string-width ansi-align 3.0.0 Depends on vulnerable versions of string-width node_modules/ansi-align immer <9.0.6 Severity: critical Prototype Pollution in immer - https://github.com/advisories/GHSA-33f9-j839-rf8h fix available via `npm audit fix --force` Will install @storybook/addon-essentials@6.1.21, which is a breaking change node_modules/immer react-dev-utils 6.0.6-next.9b4009d7 - 12.0.0-next.60 Depends on vulnerable versions of immer node_modules/react-dev-utils @storybook/builder-webpack4 * Depends on vulnerable versions of postcss Depends on vulnerable versions of react-dev-utils node_modules/@storybook/builder-webpack4 @storybook/addon-docs >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/core node_modules/@storybook/addon-docs @storybook/addon-essentials >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/addon-docs node_modules/@storybook/addon-essentials @storybook/core-server * Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/manager-webpack4 node_modules/@storybook/core-server @storybook/core >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/core-server node_modules/@storybook/core @storybook/svelte 6.2.0-alpha.0 - 6.4.12 || 6.5.0-alpha.1 - 6.5.0-alpha.5 Depends on vulnerable versions of @storybook/core node_modules/@storybook/svelte @storybook/builder-webpack5 <=6.4.12 || 6.5.0-alpha.1 - 6.5.0-alpha.5 Depends on vulnerable versions of react-dev-utils node_modules/@storybook/builder-webpack5 json-schema <0.4.0 Severity: moderate json-schema is vulnerable to Prototype Pollution - https://github.com/advisories/GHSA-896r-f27r-55mw fix available via `npm audit fix` node_modules/json-schema jsprim 0.3.0 - 1.4.1 || 2.0.0 - 2.0.1 Depends on vulnerable versions of json-schema node_modules/jsprim nanoid <3.1.31 Severity: moderate Exposure of Sensitive Information to an Unauthorized Actor in nanoid - https://github.com/advisories/GHSA-qrpm-p2h7-hrv2 fix available via `npm audit fix` node_modules/nanoid node-fetch <2.6.7 Severity: high node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor - https://github.com/advisories/GHSA-r683-j2x4-v87g fix available via `npm audit fix` node_modules/node-fetch nth-check <2.0.1 Severity: moderate Inefficient Regular Expression Complexity in nth-check - https://github.com/advisories/GHSA-rp65-9cf3-cjxr fix available via `npm audit fix` node_modules/nth-check postcss <8.2.13 Severity: moderate Regular Expression Denial of Service in postcss - https://github.com/advisories/GHSA-566m-qj78-rww5 fix available via `npm audit fix --force` Will install @storybook/addon-essentials@6.1.21, which is a breaking change node_modules/@storybook/builder-webpack4/node_modules/postcss node_modules/autoprefixer/node_modules/postcss node_modules/css-loader/node_modules/postcss node_modules/icss-utils/node_modules/postcss node_modules/postcss-flexbugs-fixes/node_modules/postcss node_modules/postcss-modules-extract-imports/node_modules/postcss node_modules/postcss-modules-local-by-default/node_modules/postcss node_modules/postcss-modules-scope/node_modules/postcss node_modules/postcss-modules-values/node_modules/postcss @storybook/builder-webpack4 * Depends on vulnerable versions of postcss Depends on vulnerable versions of react-dev-utils node_modules/@storybook/builder-webpack4 @storybook/addon-docs >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/core node_modules/@storybook/addon-docs @storybook/addon-essentials >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/addon-docs node_modules/@storybook/addon-essentials @storybook/core-server * Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/manager-webpack4 node_modules/@storybook/core-server @storybook/core >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/core-server node_modules/@storybook/core @storybook/svelte 6.2.0-alpha.0 - 6.4.12 || 6.5.0-alpha.1 - 6.5.0-alpha.5 Depends on vulnerable versions of @storybook/core node_modules/@storybook/svelte autoprefixer 1.0.20131222 - 9.8.8 Depends on vulnerable versions of postcss node_modules/autoprefixer css-loader 0.15.0 - 4.3.0 Depends on vulnerable versions of icss-utils Depends on vulnerable versions of postcss Depends on vulnerable versions of postcss-modules-values node_modules/css-loader @storybook/manager-webpack4 * Depends on vulnerable versions of css-loader node_modules/@storybook/manager-webpack4 icss-utils <=4.1.1 Depends on vulnerable versions of postcss node_modules/icss-utils postcss-modules-local-by-default <=4.0.0-rc.4 Depends on vulnerable versions of icss-utils Depends on vulnerable versions of postcss node_modules/postcss-modules-local-by-default postcss-modules-values <=4.0.0-rc.5 Depends on vulnerable versions of icss-utils Depends on vulnerable versions of postcss node_modules/postcss-modules-values postcss-flexbugs-fixes <=4.2.1 Depends on vulnerable versions of postcss node_modules/postcss-flexbugs-fixes postcss-modules-extract-imports <=2.0.0 Depends on vulnerable versions of postcss node_modules/postcss-modules-extract-imports postcss-modules-scope <=2.2.0 Depends on vulnerable versions of postcss node_modules/postcss-modules-scope prismjs <1.25.0 Severity: moderate Regular Expression Denial of Service in prismjs - https://github.com/advisories/GHSA-hqhp-5p83-hx96 fix available via `npm audit fix` node_modules/prismjs refractor <=3.4.0 || 4.0.0 - 4.1.1 Depends on vulnerable versions of prismjs node_modules/refractor tmpl <1.0.5 Severity: moderate Regular Expression Denial of Service in tmpl - https://github.com/advisories/GHSA-jgrx-mgxx-jf9v fix available via `npm audit fix` node_modules/tmpl 33 vulnerabilities (22 moderate, 3 high, 8 critical) To address issues that do not require attention, run: npm audit fix To address all issues (including breaking changes), run: npm audit fix --force ```
True
npm audit found vulnerabilities - Edit: **Blocked because our direct dependencies have not been updated yet (eg., Storybook).** ([discussion](https://matrix.to/#/!UnujZDUxGuMrYdvgTU:matrix.mathesar.org/$lQjgziZVH6pbMYGIKRkEE3TC6ApAV05doQDtuaVc2LM?via=matrix.mathesar.org&via=matrix.org)) ``` # npm audit report ansi-html * Severity: high Uncontrolled Resource Consumption in ansi-html - https://github.com/advisories/GHSA-whgm-jr23-g3j9 fix available via `npm audit fix` node_modules/ansi-html webpack-hot-middleware 2.9.0 - 2.25.0 Depends on vulnerable versions of ansi-html node_modules/webpack-hot-middleware ansi-regex >2.1.1 <5.0.1 Severity: moderate Inefficient Regular Expression Complexity in chalk/ansi-regex - https://github.com/advisories/GHSA-93q8-gq69-wqmw fix available via `npm audit fix` node_modules/ansi-align/node_modules/ansi-regex node_modules/ansi-regex node_modules/sveltedoc-parser/node_modules/ansi-regex node_modules/wide-align/node_modules/ansi-regex strip-ansi 4.0.0 - 5.2.0 Depends on vulnerable versions of ansi-regex node_modules/ansi-align/node_modules/strip-ansi node_modules/sveltedoc-parser/node_modules/string-width/node_modules/strip-ansi node_modules/wide-align/node_modules/strip-ansi string-width 2.1.0 - 4.1.0 Depends on vulnerable versions of strip-ansi node_modules/ansi-align/node_modules/string-width node_modules/sveltedoc-parser/node_modules/string-width node_modules/wide-align/node_modules/string-width ansi-align 3.0.0 Depends on vulnerable versions of string-width node_modules/ansi-align immer <9.0.6 Severity: critical Prototype Pollution in immer - https://github.com/advisories/GHSA-33f9-j839-rf8h fix available via `npm audit fix --force` Will install @storybook/addon-essentials@6.1.21, which is a breaking change node_modules/immer react-dev-utils 6.0.6-next.9b4009d7 - 12.0.0-next.60 Depends on vulnerable versions of immer node_modules/react-dev-utils @storybook/builder-webpack4 * Depends on vulnerable versions of postcss Depends on vulnerable versions of react-dev-utils node_modules/@storybook/builder-webpack4 @storybook/addon-docs >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/core node_modules/@storybook/addon-docs @storybook/addon-essentials >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/addon-docs node_modules/@storybook/addon-essentials @storybook/core-server * Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/manager-webpack4 node_modules/@storybook/core-server @storybook/core >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/core-server node_modules/@storybook/core @storybook/svelte 6.2.0-alpha.0 - 6.4.12 || 6.5.0-alpha.1 - 6.5.0-alpha.5 Depends on vulnerable versions of @storybook/core node_modules/@storybook/svelte @storybook/builder-webpack5 <=6.4.12 || 6.5.0-alpha.1 - 6.5.0-alpha.5 Depends on vulnerable versions of react-dev-utils node_modules/@storybook/builder-webpack5 json-schema <0.4.0 Severity: moderate json-schema is vulnerable to Prototype Pollution - https://github.com/advisories/GHSA-896r-f27r-55mw fix available via `npm audit fix` node_modules/json-schema jsprim 0.3.0 - 1.4.1 || 2.0.0 - 2.0.1 Depends on vulnerable versions of json-schema node_modules/jsprim nanoid <3.1.31 Severity: moderate Exposure of Sensitive Information to an Unauthorized Actor in nanoid - https://github.com/advisories/GHSA-qrpm-p2h7-hrv2 fix available via `npm audit fix` node_modules/nanoid node-fetch <2.6.7 Severity: high node-fetch is vulnerable to Exposure of Sensitive Information to an Unauthorized Actor - https://github.com/advisories/GHSA-r683-j2x4-v87g fix available via `npm audit fix` node_modules/node-fetch nth-check <2.0.1 Severity: moderate Inefficient Regular Expression Complexity in nth-check - https://github.com/advisories/GHSA-rp65-9cf3-cjxr fix available via `npm audit fix` node_modules/nth-check postcss <8.2.13 Severity: moderate Regular Expression Denial of Service in postcss - https://github.com/advisories/GHSA-566m-qj78-rww5 fix available via `npm audit fix --force` Will install @storybook/addon-essentials@6.1.21, which is a breaking change node_modules/@storybook/builder-webpack4/node_modules/postcss node_modules/autoprefixer/node_modules/postcss node_modules/css-loader/node_modules/postcss node_modules/icss-utils/node_modules/postcss node_modules/postcss-flexbugs-fixes/node_modules/postcss node_modules/postcss-modules-extract-imports/node_modules/postcss node_modules/postcss-modules-local-by-default/node_modules/postcss node_modules/postcss-modules-scope/node_modules/postcss node_modules/postcss-modules-values/node_modules/postcss @storybook/builder-webpack4 * Depends on vulnerable versions of postcss Depends on vulnerable versions of react-dev-utils node_modules/@storybook/builder-webpack4 @storybook/addon-docs >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/core node_modules/@storybook/addon-docs @storybook/addon-essentials >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/addon-docs node_modules/@storybook/addon-essentials @storybook/core-server * Depends on vulnerable versions of @storybook/builder-webpack4 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/manager-webpack4 node_modules/@storybook/core-server @storybook/core >=6.2.0-alpha.0 Depends on vulnerable versions of @storybook/builder-webpack5 Depends on vulnerable versions of @storybook/core-server node_modules/@storybook/core @storybook/svelte 6.2.0-alpha.0 - 6.4.12 || 6.5.0-alpha.1 - 6.5.0-alpha.5 Depends on vulnerable versions of @storybook/core node_modules/@storybook/svelte autoprefixer 1.0.20131222 - 9.8.8 Depends on vulnerable versions of postcss node_modules/autoprefixer css-loader 0.15.0 - 4.3.0 Depends on vulnerable versions of icss-utils Depends on vulnerable versions of postcss Depends on vulnerable versions of postcss-modules-values node_modules/css-loader @storybook/manager-webpack4 * Depends on vulnerable versions of css-loader node_modules/@storybook/manager-webpack4 icss-utils <=4.1.1 Depends on vulnerable versions of postcss node_modules/icss-utils postcss-modules-local-by-default <=4.0.0-rc.4 Depends on vulnerable versions of icss-utils Depends on vulnerable versions of postcss node_modules/postcss-modules-local-by-default postcss-modules-values <=4.0.0-rc.5 Depends on vulnerable versions of icss-utils Depends on vulnerable versions of postcss node_modules/postcss-modules-values postcss-flexbugs-fixes <=4.2.1 Depends on vulnerable versions of postcss node_modules/postcss-flexbugs-fixes postcss-modules-extract-imports <=2.0.0 Depends on vulnerable versions of postcss node_modules/postcss-modules-extract-imports postcss-modules-scope <=2.2.0 Depends on vulnerable versions of postcss node_modules/postcss-modules-scope prismjs <1.25.0 Severity: moderate Regular Expression Denial of Service in prismjs - https://github.com/advisories/GHSA-hqhp-5p83-hx96 fix available via `npm audit fix` node_modules/prismjs refractor <=3.4.0 || 4.0.0 - 4.1.1 Depends on vulnerable versions of prismjs node_modules/refractor tmpl <1.0.5 Severity: moderate Regular Expression Denial of Service in tmpl - https://github.com/advisories/GHSA-jgrx-mgxx-jf9v fix available via `npm audit fix` node_modules/tmpl 33 vulnerabilities (22 moderate, 3 high, 8 critical) To address issues that do not require attention, run: npm audit fix To address all issues (including breaking changes), run: npm audit fix --force ```
main
npm audit found vulnerabilities edit blocked because our direct dependencies have not been updated yet eg storybook npm audit report ansi html severity high uncontrolled resource consumption in ansi html fix available via npm audit fix node modules ansi html webpack hot middleware depends on vulnerable versions of ansi html node modules webpack hot middleware ansi regex severity moderate inefficient regular expression complexity in chalk ansi regex fix available via npm audit fix node modules ansi align node modules ansi regex node modules ansi regex node modules sveltedoc parser node modules ansi regex node modules wide align node modules ansi regex strip ansi depends on vulnerable versions of ansi regex node modules ansi align node modules strip ansi node modules sveltedoc parser node modules string width node modules strip ansi node modules wide align node modules strip ansi string width depends on vulnerable versions of strip ansi node modules ansi align node modules string width node modules sveltedoc parser node modules string width node modules wide align node modules string width ansi align depends on vulnerable versions of string width node modules ansi align immer severity critical prototype pollution in immer fix available via npm audit fix force will install storybook addon essentials which is a breaking change node modules immer react dev utils next next depends on vulnerable versions of immer node modules react dev utils storybook builder depends on vulnerable versions of postcss depends on vulnerable versions of react dev utils node modules storybook builder storybook addon docs alpha depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook core node modules storybook addon docs storybook addon essentials alpha depends on vulnerable versions of storybook addon docs node modules storybook addon essentials storybook core server depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook manager node modules storybook core server storybook core alpha depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook core server node modules storybook core storybook svelte alpha alpha alpha depends on vulnerable versions of storybook core node modules storybook svelte storybook builder alpha alpha depends on vulnerable versions of react dev utils node modules storybook builder json schema severity moderate json schema is vulnerable to prototype pollution fix available via npm audit fix node modules json schema jsprim depends on vulnerable versions of json schema node modules jsprim nanoid severity moderate exposure of sensitive information to an unauthorized actor in nanoid fix available via npm audit fix node modules nanoid node fetch severity high node fetch is vulnerable to exposure of sensitive information to an unauthorized actor fix available via npm audit fix node modules node fetch nth check severity moderate inefficient regular expression complexity in nth check fix available via npm audit fix node modules nth check postcss severity moderate regular expression denial of service in postcss fix available via npm audit fix force will install storybook addon essentials which is a breaking change node modules storybook builder node modules postcss node modules autoprefixer node modules postcss node modules css loader node modules postcss node modules icss utils node modules postcss node modules postcss flexbugs fixes node modules postcss node modules postcss modules extract imports node modules postcss node modules postcss modules local by default node modules postcss node modules postcss modules scope node modules postcss node modules postcss modules values node modules postcss storybook builder depends on vulnerable versions of postcss depends on vulnerable versions of react dev utils node modules storybook builder storybook addon docs alpha depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook core node modules storybook addon docs storybook addon essentials alpha depends on vulnerable versions of storybook addon docs node modules storybook addon essentials storybook core server depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook manager node modules storybook core server storybook core alpha depends on vulnerable versions of storybook builder depends on vulnerable versions of storybook core server node modules storybook core storybook svelte alpha alpha alpha depends on vulnerable versions of storybook core node modules storybook svelte autoprefixer depends on vulnerable versions of postcss node modules autoprefixer css loader depends on vulnerable versions of icss utils depends on vulnerable versions of postcss depends on vulnerable versions of postcss modules values node modules css loader storybook manager depends on vulnerable versions of css loader node modules storybook manager icss utils depends on vulnerable versions of postcss node modules icss utils postcss modules local by default rc depends on vulnerable versions of icss utils depends on vulnerable versions of postcss node modules postcss modules local by default postcss modules values rc depends on vulnerable versions of icss utils depends on vulnerable versions of postcss node modules postcss modules values postcss flexbugs fixes depends on vulnerable versions of postcss node modules postcss flexbugs fixes postcss modules extract imports depends on vulnerable versions of postcss node modules postcss modules extract imports postcss modules scope depends on vulnerable versions of postcss node modules postcss modules scope prismjs severity moderate regular expression denial of service in prismjs fix available via npm audit fix node modules prismjs refractor depends on vulnerable versions of prismjs node modules refractor tmpl severity moderate regular expression denial of service in tmpl fix available via npm audit fix node modules tmpl vulnerabilities moderate high critical to address issues that do not require attention run npm audit fix to address all issues including breaking changes run npm audit fix force
1
436,384
12,550,376,903
IssuesEvent
2020-06-06 10:53:50
googleapis/google-api-java-client-services
https://api.github.com/repos/googleapis/google-api-java-client-services
opened
Synthesis failed for plus
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate plus. :broken_heart: Here's the output from running `synth.py`: ``` 2020-06-06 03:53:43,812 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services 2020-06-06 03:53:44,790 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-06-06 03:53:44,794 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2020-06-06 03:53:44,818 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2020-06-06 03:53:44,822 autosynth [DEBUG] > Running: git config push.default simple 2020-06-06 03:53:44,825 autosynth [DEBUG] > Running: git branch -f autosynth-plus 2020-06-06 03:53:44,828 autosynth [DEBUG] > Running: git checkout autosynth-plus Switched to branch 'autosynth-plus' 2020-06-06 03:53:45,384 autosynth [INFO] > Running synthtool 2020-06-06 03:53:45,384 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/google-api-services-plus/synth.metadata', 'synth.py', '--'] 2020-06-06 03:53:45,387 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/google-api-services-plus/synth.metadata synth.py -- plus tee: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services: Is a directory 2020-06-06 03:53:45,598 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-java-client-services/synth.py. On branch autosynth-plus nothing to commit, working tree clean 2020-06-06 03:53:45,980 synthtool [DEBUG] > Cloning https://github.com/googleapis/discovery-artifact-manager.git. DEBUG:synthtool:Cloning https://github.com/googleapis/discovery-artifact-manager.git. 2020-06-06 03:53:46,374 synthtool [DEBUG] > Cleaning output directory. DEBUG:synthtool:Cleaning output directory. 2020-06-06 03:53:46,376 synthtool [DEBUG] > Installing dependencies. DEBUG:synthtool:Installing dependencies. 2020-06-06 03:53:48,157 synthtool [INFO] > Generating plus v1. INFO:synthtool:Generating plus v1. 2020-06-06 03:53:48,157 synthtool [INFO] > 1.30.1 INFO:synthtool: 1.30.1 Writing json metadata to clients/google-api-services-plus/v1.metadata.json 2020-06-06 03:53:49,110 synthtool [DEBUG] > Wrote metadata to clients/google-api-services-plus/synth.metadata. DEBUG:synthtool:Wrote metadata to clients/google-api-services-plus/synth.metadata. 2020-06-06 03:53:49,152 autosynth [DEBUG] > Running: git clean -fdx Removing .cache/ Removing __pycache__/ Removing clients/google-api-services-plus/synth.metadata Removing generator/.cache/ Removing generator/ez_setup.pyc Removing generator/src/google_apis_client_generator.egg-info/ Removing generator/src/googleapis/__init__.pyc Removing generator/src/googleapis/codegen/__init__.pyc Removing generator/src/googleapis/codegen/api.pyc Removing generator/src/googleapis/codegen/api_exception.pyc Removing generator/src/googleapis/codegen/api_library_generator.pyc Removing generator/src/googleapis/codegen/data_types.pyc Removing generator/src/googleapis/codegen/django_helpers.pyc Removing generator/src/googleapis/codegen/filesys/__init__.pyc Removing generator/src/googleapis/codegen/filesys/files.pyc Removing generator/src/googleapis/codegen/filesys/filesystem_library_package.pyc Removing generator/src/googleapis/codegen/filesys/library_package.pyc Removing generator/src/googleapis/codegen/filesys/package_writer_foundry.pyc Removing generator/src/googleapis/codegen/filesys/tar_library_package.pyc Removing generator/src/googleapis/codegen/filesys/zip_library_package.pyc Removing generator/src/googleapis/codegen/generate_library.pyc Removing generator/src/googleapis/codegen/generator.pyc Removing generator/src/googleapis/codegen/generator_lookup.pyc Removing generator/src/googleapis/codegen/import_definition.pyc Removing generator/src/googleapis/codegen/java_generator.pyc Removing generator/src/googleapis/codegen/java_import_manager.pyc Removing generator/src/googleapis/codegen/language_model.pyc Removing generator/src/googleapis/codegen/schema.pyc Removing generator/src/googleapis/codegen/targets.pyc Removing generator/src/googleapis/codegen/template_helpers.pyc Removing generator/src/googleapis/codegen/template_objects.pyc Removing generator/src/googleapis/codegen/utilities/__init__.pyc Removing generator/src/googleapis/codegen/utilities/convert_size.pyc Removing generator/src/googleapis/codegen/utilities/html_stripper.pyc Removing generator/src/googleapis/codegen/utilities/json_expander.pyc Removing generator/src/googleapis/codegen/utilities/json_with_comments.pyc Removing generator/src/googleapis/codegen/utilities/maven_utils.pyc Removing generator/src/googleapis/codegen/utilities/name_validator.pyc Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 555, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 121, in synthesize with open(log_file_path, "rt") as fp: IsADirectoryError: [Errno 21] Is a directory: '/tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services' ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/a12fce2f-fa64-48b1-abd7-5a9cfc4df713/targets/github%2Fsynthtool;config=default/tests;query=google-api-java-client-services;failed=false).
1.0
Synthesis failed for plus - Hello! Autosynth couldn't regenerate plus. :broken_heart: Here's the output from running `synth.py`: ``` 2020-06-06 03:53:43,812 autosynth [INFO] > logs will be written to: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services 2020-06-06 03:53:44,790 autosynth [DEBUG] > Running: git config --global core.excludesfile /home/kbuilder/.autosynth-gitignore 2020-06-06 03:53:44,794 autosynth [DEBUG] > Running: git config user.name yoshi-automation 2020-06-06 03:53:44,818 autosynth [DEBUG] > Running: git config user.email yoshi-automation@google.com 2020-06-06 03:53:44,822 autosynth [DEBUG] > Running: git config push.default simple 2020-06-06 03:53:44,825 autosynth [DEBUG] > Running: git branch -f autosynth-plus 2020-06-06 03:53:44,828 autosynth [DEBUG] > Running: git checkout autosynth-plus Switched to branch 'autosynth-plus' 2020-06-06 03:53:45,384 autosynth [INFO] > Running synthtool 2020-06-06 03:53:45,384 autosynth [INFO] > ['/tmpfs/src/github/synthtool/env/bin/python3', '-m', 'synthtool', '--metadata', 'clients/google-api-services-plus/synth.metadata', 'synth.py', '--'] 2020-06-06 03:53:45,387 autosynth [DEBUG] > Running: /tmpfs/src/github/synthtool/env/bin/python3 -m synthtool --metadata clients/google-api-services-plus/synth.metadata synth.py -- plus tee: /tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services: Is a directory 2020-06-06 03:53:45,598 synthtool [DEBUG] > Executing /home/kbuilder/.cache/synthtool/google-api-java-client-services/synth.py. On branch autosynth-plus nothing to commit, working tree clean 2020-06-06 03:53:45,980 synthtool [DEBUG] > Cloning https://github.com/googleapis/discovery-artifact-manager.git. DEBUG:synthtool:Cloning https://github.com/googleapis/discovery-artifact-manager.git. 2020-06-06 03:53:46,374 synthtool [DEBUG] > Cleaning output directory. DEBUG:synthtool:Cleaning output directory. 2020-06-06 03:53:46,376 synthtool [DEBUG] > Installing dependencies. DEBUG:synthtool:Installing dependencies. 2020-06-06 03:53:48,157 synthtool [INFO] > Generating plus v1. INFO:synthtool:Generating plus v1. 2020-06-06 03:53:48,157 synthtool [INFO] > 1.30.1 INFO:synthtool: 1.30.1 Writing json metadata to clients/google-api-services-plus/v1.metadata.json 2020-06-06 03:53:49,110 synthtool [DEBUG] > Wrote metadata to clients/google-api-services-plus/synth.metadata. DEBUG:synthtool:Wrote metadata to clients/google-api-services-plus/synth.metadata. 2020-06-06 03:53:49,152 autosynth [DEBUG] > Running: git clean -fdx Removing .cache/ Removing __pycache__/ Removing clients/google-api-services-plus/synth.metadata Removing generator/.cache/ Removing generator/ez_setup.pyc Removing generator/src/google_apis_client_generator.egg-info/ Removing generator/src/googleapis/__init__.pyc Removing generator/src/googleapis/codegen/__init__.pyc Removing generator/src/googleapis/codegen/api.pyc Removing generator/src/googleapis/codegen/api_exception.pyc Removing generator/src/googleapis/codegen/api_library_generator.pyc Removing generator/src/googleapis/codegen/data_types.pyc Removing generator/src/googleapis/codegen/django_helpers.pyc Removing generator/src/googleapis/codegen/filesys/__init__.pyc Removing generator/src/googleapis/codegen/filesys/files.pyc Removing generator/src/googleapis/codegen/filesys/filesystem_library_package.pyc Removing generator/src/googleapis/codegen/filesys/library_package.pyc Removing generator/src/googleapis/codegen/filesys/package_writer_foundry.pyc Removing generator/src/googleapis/codegen/filesys/tar_library_package.pyc Removing generator/src/googleapis/codegen/filesys/zip_library_package.pyc Removing generator/src/googleapis/codegen/generate_library.pyc Removing generator/src/googleapis/codegen/generator.pyc Removing generator/src/googleapis/codegen/generator_lookup.pyc Removing generator/src/googleapis/codegen/import_definition.pyc Removing generator/src/googleapis/codegen/java_generator.pyc Removing generator/src/googleapis/codegen/java_import_manager.pyc Removing generator/src/googleapis/codegen/language_model.pyc Removing generator/src/googleapis/codegen/schema.pyc Removing generator/src/googleapis/codegen/targets.pyc Removing generator/src/googleapis/codegen/template_helpers.pyc Removing generator/src/googleapis/codegen/template_objects.pyc Removing generator/src/googleapis/codegen/utilities/__init__.pyc Removing generator/src/googleapis/codegen/utilities/convert_size.pyc Removing generator/src/googleapis/codegen/utilities/html_stripper.pyc Removing generator/src/googleapis/codegen/utilities/json_expander.pyc Removing generator/src/googleapis/codegen/utilities/json_with_comments.pyc Removing generator/src/googleapis/codegen/utilities/maven_utils.pyc Removing generator/src/googleapis/codegen/utilities/name_validator.pyc Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.9/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 615, in <module> main() File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 476, in main return _inner_main(temp_dir) File "/tmpfs/src/github/synthtool/autosynth/synth.py", line 555, in _inner_main ).synthesize(base_synth_log_path) File "/tmpfs/src/github/synthtool/autosynth/synthesizer.py", line 121, in synthesize with open(log_file_path, "rt") as fp: IsADirectoryError: [Errno 21] Is a directory: '/tmpfs/src/github/synthtool/logs/googleapis/google-api-java-client-services' ``` Google internal developers can see the full log [here](http://sponge2/results/invocations/a12fce2f-fa64-48b1-abd7-5a9cfc4df713/targets/github%2Fsynthtool;config=default/tests;query=google-api-java-client-services;failed=false).
non_main
synthesis failed for plus hello autosynth couldn t regenerate plus broken heart here s the output from running synth py autosynth logs will be written to tmpfs src github synthtool logs googleapis google api java client services autosynth running git config global core excludesfile home kbuilder autosynth gitignore autosynth running git config user name yoshi automation autosynth running git config user email yoshi automation google com autosynth running git config push default simple autosynth running git branch f autosynth plus autosynth running git checkout autosynth plus switched to branch autosynth plus autosynth running synthtool autosynth autosynth running tmpfs src github synthtool env bin m synthtool metadata clients google api services plus synth metadata synth py plus tee tmpfs src github synthtool logs googleapis google api java client services is a directory synthtool executing home kbuilder cache synthtool google api java client services synth py on branch autosynth plus nothing to commit working tree clean synthtool cloning debug synthtool cloning synthtool cleaning output directory debug synthtool cleaning output directory synthtool installing dependencies debug synthtool installing dependencies synthtool generating plus info synthtool generating plus synthtool info synthtool writing json metadata to clients google api services plus metadata json synthtool wrote metadata to clients google api services plus synth metadata debug synthtool wrote metadata to clients google api services plus synth metadata autosynth running git clean fdx removing cache removing pycache removing clients google api services plus synth metadata removing generator cache removing generator ez setup pyc removing generator src google apis client generator egg info removing generator src googleapis init pyc removing generator src googleapis codegen init pyc removing generator src googleapis codegen api pyc removing generator src googleapis codegen api exception pyc removing generator src googleapis codegen api library generator pyc removing generator src googleapis codegen data types pyc removing generator src googleapis codegen django helpers pyc removing generator src googleapis codegen filesys init pyc removing generator src googleapis codegen filesys files pyc removing generator src googleapis codegen filesys filesystem library package pyc removing generator src googleapis codegen filesys library package pyc removing generator src googleapis codegen filesys package writer foundry pyc removing generator src googleapis codegen filesys tar library package pyc removing generator src googleapis codegen filesys zip library package pyc removing generator src googleapis codegen generate library pyc removing generator src googleapis codegen generator pyc removing generator src googleapis codegen generator lookup pyc removing generator src googleapis codegen import definition pyc removing generator src googleapis codegen java generator pyc removing generator src googleapis codegen java import manager pyc removing generator src googleapis codegen language model pyc removing generator src googleapis codegen schema pyc removing generator src googleapis codegen targets pyc removing generator src googleapis codegen template helpers pyc removing generator src googleapis codegen template objects pyc removing generator src googleapis codegen utilities init pyc removing generator src googleapis codegen utilities convert size pyc removing generator src googleapis codegen utilities html stripper pyc removing generator src googleapis codegen utilities json expander pyc removing generator src googleapis codegen utilities json with comments pyc removing generator src googleapis codegen utilities maven utils pyc removing generator src googleapis codegen utilities name validator pyc traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src github synthtool autosynth synth py line in main file tmpfs src github synthtool autosynth synth py line in main return inner main temp dir file tmpfs src github synthtool autosynth synth py line in inner main synthesize base synth log path file tmpfs src github synthtool autosynth synthesizer py line in synthesize with open log file path rt as fp isadirectoryerror is a directory tmpfs src github synthtool logs googleapis google api java client services google internal developers can see the full log
0
1,814
6,577,317,634
IssuesEvent
2017-09-12 00:04:07
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Modifying ec2 security groups in ansible playbook doesn't work for running instances
affects_2.1 aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2 module ##### ANSIBLE VERSION $ ansible --version ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### OS / ENVIRONMENT Ubuntu on Windows 10 ##### SUMMARY Changing the group_id value for a running instance has no effect on it's security groups ##### STEPS TO REPRODUCE Run playbook that creates ec2 instance Add extra group_id to playbook Run playbook again ##### EXPECTED RESULTS ec2 instance is in both security groups ##### ACTUAL RESULTS security group of instance is not changed.
True
Modifying ec2 security groups in ansible playbook doesn't work for running instances - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2 module ##### ANSIBLE VERSION $ ansible --version ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### OS / ENVIRONMENT Ubuntu on Windows 10 ##### SUMMARY Changing the group_id value for a running instance has no effect on it's security groups ##### STEPS TO REPRODUCE Run playbook that creates ec2 instance Add extra group_id to playbook Run playbook again ##### EXPECTED RESULTS ec2 instance is in both security groups ##### ACTUAL RESULTS security group of instance is not changed.
main
modifying security groups in ansible playbook doesn t work for running instances issue type bug report component name module ansible version ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides os environment ubuntu on windows summary changing the group id value for a running instance has no effect on it s security groups steps to reproduce run playbook that creates instance add extra group id to playbook run playbook again expected results instance is in both security groups actual results security group of instance is not changed
1
148,702
19,541,576,619
IssuesEvent
2022-01-01 01:52:50
prodsec-review/juice-shop
https://api.github.com/repos/prodsec-review/juice-shop
closed
CVE-2016-1000237 (Medium) detected in sanitize-html-1.4.2.tgz
security vulnerability Stale
## CVE-2016-1000237 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sanitize-html-1.4.2.tgz</b></p></summary> <p>Clean up user-submitted HTML, preserving whitelisted elements and whitelisted attributes on a per-element basis</p> <p>Library home page: <a href="https://registry.npmjs.org/sanitize-html/-/sanitize-html-1.4.2.tgz">https://registry.npmjs.org/sanitize-html/-/sanitize-html-1.4.2.tgz</a></p> <p>Path to dependency file: juice-shop/package.json</p> <p>Path to vulnerable library: /node_modules/sanitize-html/package.json</p> <p> Dependency Hierarchy: - :x: **sanitize-html-1.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/prodsec-review/juice-shop/commit/026c393c72368c780faf7b4197dea54d7734cfd3">026c393c72368c780faf7b4197dea54d7734cfd3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> sanitize-html before 1.4.3 has XSS. <p>Publish Date: 2020-01-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000237>CVE-2016-1000237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000237">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000237</a></p> <p>Release Date: 2020-01-23</p> <p>Fix Resolution: sanitize-html - 1.4.3</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"sanitize-html","packageVersion":"1.4.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"sanitize-html:1.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"sanitize-html - 1.4.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-1000237","vulnerabilityDetails":"sanitize-html before 1.4.3 has XSS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000237","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-1000237 (Medium) detected in sanitize-html-1.4.2.tgz - ## CVE-2016-1000237 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sanitize-html-1.4.2.tgz</b></p></summary> <p>Clean up user-submitted HTML, preserving whitelisted elements and whitelisted attributes on a per-element basis</p> <p>Library home page: <a href="https://registry.npmjs.org/sanitize-html/-/sanitize-html-1.4.2.tgz">https://registry.npmjs.org/sanitize-html/-/sanitize-html-1.4.2.tgz</a></p> <p>Path to dependency file: juice-shop/package.json</p> <p>Path to vulnerable library: /node_modules/sanitize-html/package.json</p> <p> Dependency Hierarchy: - :x: **sanitize-html-1.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/prodsec-review/juice-shop/commit/026c393c72368c780faf7b4197dea54d7734cfd3">026c393c72368c780faf7b4197dea54d7734cfd3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> sanitize-html before 1.4.3 has XSS. <p>Publish Date: 2020-01-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000237>CVE-2016-1000237</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000237">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-1000237</a></p> <p>Release Date: 2020-01-23</p> <p>Fix Resolution: sanitize-html - 1.4.3</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"sanitize-html","packageVersion":"1.4.2","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"sanitize-html:1.4.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"sanitize-html - 1.4.3","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-1000237","vulnerabilityDetails":"sanitize-html before 1.4.3 has XSS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-1000237","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in sanitize html tgz cve medium severity vulnerability vulnerable library sanitize html tgz clean up user submitted html preserving whitelisted elements and whitelisted attributes on a per element basis library home page a href path to dependency file juice shop package json path to vulnerable library node modules sanitize html package json dependency hierarchy x sanitize html tgz vulnerable library found in head commit a href found in base branch master vulnerability details sanitize html before has xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sanitize html check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree sanitize html isminimumfixversionavailable true minimumfixversion sanitize html isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails sanitize html before has xss vulnerabilityurl
0
4,406
22,633,852,269
IssuesEvent
2022-06-30 16:54:25
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[SHOWCASE] Dipan's Dash
πŸ’― Showcase πŸ‘€ Awaiting Maintainer Response
<h3 dir="auto">Title</h3> <p dir="auto">Dipan's Dashboard</p> <h3 dir="auto">Screenshot</h3> <p dir="auto"><a target="_blank" rel="noopener noreferrer" href="https://user-images.githubusercontent.com/5004441/176696796-522e0233-1956-4aa4-b27f-8a376a369870.png"><img src="https://user-images.githubusercontent.com/5004441/176696796-522e0233-1956-4aa4-b27f-8a376a369870.png" alt="screencapture-dash-dipan-de-2022-06-30-16_00_21" style="max-width: 100%;"></a></p> <h3 dir="auto">Would you like your name/ username included?</h3> <p dir="auto">Yes</p> <h3 dir="auto">Link to your Website/ Profile/ Twitter (optional)</h3> <p dir="auto"><a href="https://dipan.de" rel="nofollow">https://dipan.de</a></p> <h3 dir="auto">Description (Optional)</h3> <p dir="auto">With the option of embedding iframes, it becomes possible to combine and conquer. I have set up Dashy to include graphs from grafana, and set them so that they update automatically. I also have uptime monitors from uptime Kuma, panels from homeasistant are also possible to be included.</p> <p dir="auto">Dashy is awesome!</p> <!-- unfurl begin --> <blockquote> <div> <strong> <a href="https://www.thedesignerd.in/">Web | Dipan Ghosh</a> </strong> </div> <div>Personal Website of Dipan Ghosh. Enjoy nature and travel photography especially flower.</div> </blockquote> <!-- unfurl end -->
True
[SHOWCASE] Dipan's Dash - <h3 dir="auto">Title</h3> <p dir="auto">Dipan's Dashboard</p> <h3 dir="auto">Screenshot</h3> <p dir="auto"><a target="_blank" rel="noopener noreferrer" href="https://user-images.githubusercontent.com/5004441/176696796-522e0233-1956-4aa4-b27f-8a376a369870.png"><img src="https://user-images.githubusercontent.com/5004441/176696796-522e0233-1956-4aa4-b27f-8a376a369870.png" alt="screencapture-dash-dipan-de-2022-06-30-16_00_21" style="max-width: 100%;"></a></p> <h3 dir="auto">Would you like your name/ username included?</h3> <p dir="auto">Yes</p> <h3 dir="auto">Link to your Website/ Profile/ Twitter (optional)</h3> <p dir="auto"><a href="https://dipan.de" rel="nofollow">https://dipan.de</a></p> <h3 dir="auto">Description (Optional)</h3> <p dir="auto">With the option of embedding iframes, it becomes possible to combine and conquer. I have set up Dashy to include graphs from grafana, and set them so that they update automatically. I also have uptime monitors from uptime Kuma, panels from homeasistant are also possible to be included.</p> <p dir="auto">Dashy is awesome!</p> <!-- unfurl begin --> <blockquote> <div> <strong> <a href="https://www.thedesignerd.in/">Web | Dipan Ghosh</a> </strong> </div> <div>Personal Website of Dipan Ghosh. Enjoy nature and travel photography especially flower.</div> </blockquote> <!-- unfurl end -->
main
dipan s dash title dipan s dashboard screenshot would you like your name username included yes link to your website profile twitter optional description optional with the option of embedding iframes it becomes possible to combine and conquer i have set up dashy to include graphs from grafana and set them so that they update automatically i also have uptime monitors from uptime kuma panels from homeasistant are also possible to be included dashy is awesome personal website of dipan ghosh enjoy nature and travel photography especially flower
1
55,855
14,072,217,456
IssuesEvent
2020-11-04 01:13:26
jgeraigery/build-blockchain-insurance-app
https://api.github.com/repos/jgeraigery/build-blockchain-insurance-app
opened
CVE-2018-1109 (High) detected in braces-1.8.5.tgz
security vulnerability
## CVE-2018-1109 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary> <p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p> <p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p> <p>Path to dependency file: build-blockchain-insurance-app/web/package.json</p> <p>Path to vulnerable library: build-blockchain-insurance-app/web/node_modules/braces/package.json</p> <p> Dependency Hierarchy: - babel-cli-6.26.0.tgz (Root Library) - chokidar-1.7.0.tgz - anymatch-1.3.2.tgz - micromatch-2.3.11.tgz - :x: **braces-1.8.5.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Braces before 1.4.2 and 2.17.2 is vulnerable to ReDoS. It used a regular expression (^\{(,+(?:(\{,+\})*),*|,*(?:(\{,+\})*),+)\}) in order to detects empty braces. This can cause an impact of about 10 seconds matching time for data 50K characters long. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1109>CVE-2018-1109</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1547272">https://bugzilla.redhat.com/show_bug.cgi?id=1547272</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.3.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"braces","packageVersion":"1.8.5","isTransitiveDependency":true,"dependencyTree":"babel-cli:6.26.0;chokidar:1.7.0;anymatch:1.3.2;micromatch:2.3.11;braces:1.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.1"}],"vulnerabilityIdentifier":"CVE-2018-1109","vulnerabilityDetails":"Braces before 1.4.2 and 2.17.2 is vulnerable to ReDoS. It used a regular expression (^\\{(,+(?:(\\{,+\\})*),*|,*(?:(\\{,+\\})*),+)\\}) in order to detects empty braces. This can cause an impact of about 10 seconds matching time for data 50K characters long.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1109","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-1109 (High) detected in braces-1.8.5.tgz - ## CVE-2018-1109 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary> <p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p> <p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p> <p>Path to dependency file: build-blockchain-insurance-app/web/package.json</p> <p>Path to vulnerable library: build-blockchain-insurance-app/web/node_modules/braces/package.json</p> <p> Dependency Hierarchy: - babel-cli-6.26.0.tgz (Root Library) - chokidar-1.7.0.tgz - anymatch-1.3.2.tgz - micromatch-2.3.11.tgz - :x: **braces-1.8.5.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Braces before 1.4.2 and 2.17.2 is vulnerable to ReDoS. It used a regular expression (^\{(,+(?:(\{,+\})*),*|,*(?:(\{,+\})*),+)\}) in order to detects empty braces. This can cause an impact of about 10 seconds matching time for data 50K characters long. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1109>CVE-2018-1109</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1547272">https://bugzilla.redhat.com/show_bug.cgi?id=1547272</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: 2.3.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"braces","packageVersion":"1.8.5","isTransitiveDependency":true,"dependencyTree":"babel-cli:6.26.0;chokidar:1.7.0;anymatch:1.3.2;micromatch:2.3.11;braces:1.8.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.3.1"}],"vulnerabilityIdentifier":"CVE-2018-1109","vulnerabilityDetails":"Braces before 1.4.2 and 2.17.2 is vulnerable to ReDoS. It used a regular expression (^\\{(,+(?:(\\{,+\\})*),*|,*(?:(\\{,+\\})*),+)\\}) in order to detects empty braces. This can cause an impact of about 10 seconds matching time for data 50K characters long.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1109","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
cve high detected in braces tgz cve high severity vulnerability vulnerable library braces tgz fastest brace expansion for node js with the most complete support for the bash braces specification library home page a href path to dependency file build blockchain insurance app web package json path to vulnerable library build blockchain insurance app web node modules braces package json dependency hierarchy babel cli tgz root library chokidar tgz anymatch tgz micromatch tgz x braces tgz vulnerable library found in base branch master vulnerability details braces before and is vulnerable to redos it used a regular expression in order to detects empty braces this can cause an impact of about seconds matching time for data characters long publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails braces before and is vulnerable to redos it used a regular expression in order to detects empty braces this can cause an impact of about seconds matching time for data characters long vulnerabilityurl
0
140,867
5,425,361,399
IssuesEvent
2017-03-03 05:46:02
YaleSTC/vesta
https://api.github.com/repos/YaleSTC/vesta
opened
Extract Group#remove_members
complexity: 1 priority: 3 type: enhancement
The `GroupUpdater#remove_users` should really be a method on group.
1.0
Extract Group#remove_members - The `GroupUpdater#remove_users` should really be a method on group.
non_main
extract group remove members the groupupdater remove users should really be a method on group
0
293,423
22,056,860,819
IssuesEvent
2022-05-30 13:37:56
CS-SI/eodag
https://api.github.com/repos/CS-SI/eodag
closed
Avoid trying to order an OFFLINE product
documentation enhancement priority::1
From @remi-braun in https://github.com/CS-SI/eodag/issues/214 > Maybe add the LTA order as a parameter, as it is not needed in every usecase (sometimes, sb just want to download everything he can without triggering any LTA retrieval) LTA: Long Term Archive The `HTTPDownload` plugin (used by creodias, onda, theia, peps, sobloo) can automatically try to order a product if its status is *OFFLINE*, and can download it afterwards. The feature request is to add a way to avoid trying to order an *OFFLINE* product. This could be done in at least two complementary ways: 1. Document the way to filter out *OFFLINE* products from the result of a search. This can be done as follows currently: ```python from eodag.plugins.crunch.filter_property import FilterProperty dag.set_preferred_provider("peps") prods, _ = dag.search(...) online_prods = prods.crunch(FilterProperty({"storageStatus": "ONLINE", "operator": "eq"})) dag.download_all(online_prods) ``` 2. Add a parameter to pass as a kwarg to `EODataAccessGateway.download_all` (not sure it's relevant for `download`), e.g. `skip_offline`, that would be passed to the HTTPPlugin and, if `True`, would stop trying to download this product, which would lead to `download_all` trying to download the next product.
1.0
Avoid trying to order an OFFLINE product - From @remi-braun in https://github.com/CS-SI/eodag/issues/214 > Maybe add the LTA order as a parameter, as it is not needed in every usecase (sometimes, sb just want to download everything he can without triggering any LTA retrieval) LTA: Long Term Archive The `HTTPDownload` plugin (used by creodias, onda, theia, peps, sobloo) can automatically try to order a product if its status is *OFFLINE*, and can download it afterwards. The feature request is to add a way to avoid trying to order an *OFFLINE* product. This could be done in at least two complementary ways: 1. Document the way to filter out *OFFLINE* products from the result of a search. This can be done as follows currently: ```python from eodag.plugins.crunch.filter_property import FilterProperty dag.set_preferred_provider("peps") prods, _ = dag.search(...) online_prods = prods.crunch(FilterProperty({"storageStatus": "ONLINE", "operator": "eq"})) dag.download_all(online_prods) ``` 2. Add a parameter to pass as a kwarg to `EODataAccessGateway.download_all` (not sure it's relevant for `download`), e.g. `skip_offline`, that would be passed to the HTTPPlugin and, if `True`, would stop trying to download this product, which would lead to `download_all` trying to download the next product.
non_main
avoid trying to order an offline product from remi braun in maybe add the lta order as a parameter as it is not needed in every usecase sometimes sb just want to download everything he can without triggering any lta retrieval lta long term archive the httpdownload plugin used by creodias onda theia peps sobloo can automatically try to order a product if its status is offline and can download it afterwards the feature request is to add a way to avoid trying to order an offline product this could be done in at least two complementary ways document the way to filter out offline products from the result of a search this can be done as follows currently python from eodag plugins crunch filter property import filterproperty dag set preferred provider peps prods dag search online prods prods crunch filterproperty storagestatus online operator eq dag download all online prods add a parameter to pass as a kwarg to eodataaccessgateway download all not sure it s relevant for download e g skip offline that would be passed to the httpplugin and if true would stop trying to download this product which would lead to download all trying to download the next product
0
289,257
21,776,593,189
IssuesEvent
2022-05-13 14:21:37
JuliaRobotics/RoME.jl
https://api.github.com/repos/JuliaRobotics/RoME.jl
closed
update Pose2Pose2 TYPEDEF docs with Manifolds math
enhancement documentation manifolds user experience
New CJL Docs [Using Manifolds.jl page has a link to the Pose2Pose2 as easy example](https://juliarobotics.org/Caesar.jl/latest/concepts/using_manifolds/#Using-Manifolds-in-Factors) where and how the manifold math is used. So there should be a math description to look at when you click through to those docs. This one: - https://juliarobotics.org/Caesar.jl/latest/concepts/available_varfacs/#RoME.Pose2Pose2
1.0
update Pose2Pose2 TYPEDEF docs with Manifolds math - New CJL Docs [Using Manifolds.jl page has a link to the Pose2Pose2 as easy example](https://juliarobotics.org/Caesar.jl/latest/concepts/using_manifolds/#Using-Manifolds-in-Factors) where and how the manifold math is used. So there should be a math description to look at when you click through to those docs. This one: - https://juliarobotics.org/Caesar.jl/latest/concepts/available_varfacs/#RoME.Pose2Pose2
non_main
update typedef docs with manifolds math new cjl docs where and how the manifold math is used so there should be a math description to look at when you click through to those docs this one
0
4,409
22,640,292,550
IssuesEvent
2022-07-01 00:48:33
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Simple Prometheus receiver should use HTTP Client Settings
bug good first issue unmaintained comp:simpleprometheus
As seen on #9570, the simple Prometheus receiver should make use of the standard HTTP Client Settings, so that it can take advantage of the standard features that it provides, such as access to the client authenticators.
True
Simple Prometheus receiver should use HTTP Client Settings - As seen on #9570, the simple Prometheus receiver should make use of the standard HTTP Client Settings, so that it can take advantage of the standard features that it provides, such as access to the client authenticators.
main
simple prometheus receiver should use http client settings as seen on the simple prometheus receiver should make use of the standard http client settings so that it can take advantage of the standard features that it provides such as access to the client authenticators
1
53,379
28,109,168,733
IssuesEvent
2023-03-31 05:15:05
nodejs/node
https://api.github.com/repos/nodejs/node
opened
large performance regressions since f51c152
performance whatwg-url regression
Since the PR was landed before the CI benchmark machine was restarted, these regressions were not caught before it was merged. I'm not sure if there are other built-in modules affected, but the URL module is one such user of `toUSVString()`. https://ci.nodejs.org/view/Node.js%20benchmark/job/benchmark-node-micro-benchmarks/1306/consoleFull ``` url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='altspaces' *** -17.43 % Β±1.75% Β±2.33% Β±3.03% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='encodefake' *** -25.28 % Β±2.79% Β±3.72% Β±4.86% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='encodelast' *** -20.54 % Β±2.20% Β±2.93% Β±3.82% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='encodemany' *** -16.30 % Β±2.05% Β±2.74% Β±3.57% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='manyblankpairs' *** -38.49 % Β±4.55% Β±6.11% Β±8.09% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='manypairs' *** -16.03 % Β±2.22% Β±2.96% Β±3.86% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='multicharsep' *** -27.57 % Β±2.15% Β±2.87% Β±3.76% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='multivalue' *** -23.27 % Β±1.78% Β±2.37% Β±3.08% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='multivaluemany' *** -16.69 % Β±1.17% Β±1.56% Β±2.05% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='noencode' *** -26.78 % Β±2.35% Β±3.13% Β±4.09% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='array' *** -25.56 % Β±0.71% Β±0.95% Β±1.24% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='encodelast' *** -40.80 % Β±1.59% Β±2.14% Β±2.81% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='encodemany' *** -41.81 % Β±2.19% Β±2.91% Β±3.79% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='multiprimitives' *** -42.60 % Β±1.83% Β±2.46% Β±3.23% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='noencode' *** -39.99 % Β±1.57% Β±2.11% Β±2.77% url/url-searchparams-creation.js n=1000000 inputType='object' type='array' *** -29.03 % Β±1.94% Β±2.58% Β±3.36% url/url-searchparams-creation.js n=1000000 inputType='object' type='encodelast' *** -34.75 % Β±1.69% Β±2.27% Β±2.99% url/url-searchparams-creation.js n=1000000 inputType='object' type='encodemany' *** -35.23 % Β±1.30% Β±1.74% Β±2.28% url/url-searchparams-creation.js n=1000000 inputType='object' type='multiprimitives' *** -35.42 % Β±2.37% Β±3.17% Β±4.17% url/url-searchparams-creation.js n=1000000 inputType='object' type='noencode' *** -35.35 % Β±1.82% Β±2.43% Β±3.19% url/url-searchparams-creation.js n=1000000 inputType='string' type='array' *** -23.30 % Β±2.24% Β±2.98% Β±3.88% url/url-searchparams-creation.js n=1000000 inputType='string' type='encodelast' *** -26.13 % Β±2.25% Β±3.02% Β±3.98% url/url-searchparams-creation.js n=1000000 inputType='string' type='encodemany' *** -23.93 % Β±1.70% Β±2.27% Β±2.95% url/url-searchparams-creation.js n=1000000 inputType='string' type='multiprimitives' *** -24.05 % Β±1.58% Β±2.10% Β±2.74% url/url-searchparams-creation.js n=1000000 inputType='string' type='noencode' *** -23.97 % Β±1.77% Β±2.35% Β±3.06% url/url-searchparams-read.js n=20000000 param='nonexistent' accessMethod='get' *** -60.10 % Β±1.35% Β±1.80% Β±2.34% url/url-searchparams-read.js n=20000000 param='nonexistent' accessMethod='getAll' *** -57.29 % Β±2.38% Β±3.20% Β±4.22% url/url-searchparams-read.js n=20000000 param='nonexistent' accessMethod='has' *** -58.05 % Β±2.19% Β±2.94% Β±3.87% url/url-searchparams-read.js n=20000000 param='one' accessMethod='get' *** -60.58 % Β±1.58% Β±2.11% Β±2.75% url/url-searchparams-read.js n=20000000 param='one' accessMethod='getAll' *** -39.72 % Β±2.91% Β±3.87% Β±5.04% url/url-searchparams-read.js n=20000000 param='one' accessMethod='has' *** -60.04 % Β±2.66% Β±3.57% Β±4.71% url/url-searchparams-read.js n=20000000 param='three' accessMethod='get' *** -57.71 % Β±1.36% Β±1.81% Β±2.36% url/url-searchparams-read.js n=20000000 param='three' accessMethod='getAll' *** -37.37 % Β±2.63% Β±3.51% Β±4.60% url/url-searchparams-read.js n=20000000 param='three' accessMethod='has' *** -58.09 % Β±2.38% Β±3.19% Β±4.20% url/url-searchparams-read.js n=20000000 param='two' accessMethod='get' *** -58.71 % Β±1.70% Β±2.27% Β±2.95% url/url-searchparams-read.js n=20000000 param='two' accessMethod='getAll' *** -37.93 % Β±2.61% Β±3.47% Β±4.52% url/url-searchparams-read.js n=20000000 param='two' accessMethod='has' *** -57.99 % Β±2.55% Β±3.40% Β±4.44% url/usvstring.js n=50000000 input='nonstring' *** -45.54 % Β±3.05% Β±4.08% Β±5.35% url/usvstring.js n=50000000 input='valid' *** -73.28 % Β±2.34% Β±3.14% Β±4.15% url/usvstring.js n=50000000 input='validsurr' *** -14.33 % Β±2.56% Β±3.42% Β±4.48% ``` /cc @anonrig
True
large performance regressions since f51c152 - Since the PR was landed before the CI benchmark machine was restarted, these regressions were not caught before it was merged. I'm not sure if there are other built-in modules affected, but the URL module is one such user of `toUSVString()`. https://ci.nodejs.org/view/Node.js%20benchmark/job/benchmark-node-micro-benchmarks/1306/consoleFull ``` url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='altspaces' *** -17.43 % Β±1.75% Β±2.33% Β±3.03% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='encodefake' *** -25.28 % Β±2.79% Β±3.72% Β±4.86% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='encodelast' *** -20.54 % Β±2.20% Β±2.93% Β±3.82% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='encodemany' *** -16.30 % Β±2.05% Β±2.74% Β±3.57% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='manyblankpairs' *** -38.49 % Β±4.55% Β±6.11% Β±8.09% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='manypairs' *** -16.03 % Β±2.22% Β±2.96% Β±3.86% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='multicharsep' *** -27.57 % Β±2.15% Β±2.87% Β±3.76% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='multivalue' *** -23.27 % Β±1.78% Β±2.37% Β±3.08% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='multivaluemany' *** -16.69 % Β±1.17% Β±1.56% Β±2.05% url/legacy-vs-whatwg-url-searchparams-parse.js n=1000000 method='whatwg' searchParam='noencode' *** -26.78 % Β±2.35% Β±3.13% Β±4.09% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='array' *** -25.56 % Β±0.71% Β±0.95% Β±1.24% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='encodelast' *** -40.80 % Β±1.59% Β±2.14% Β±2.81% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='encodemany' *** -41.81 % Β±2.19% Β±2.91% Β±3.79% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='multiprimitives' *** -42.60 % Β±1.83% Β±2.46% Β±3.23% url/url-searchparams-creation.js n=1000000 inputType='iterable' type='noencode' *** -39.99 % Β±1.57% Β±2.11% Β±2.77% url/url-searchparams-creation.js n=1000000 inputType='object' type='array' *** -29.03 % Β±1.94% Β±2.58% Β±3.36% url/url-searchparams-creation.js n=1000000 inputType='object' type='encodelast' *** -34.75 % Β±1.69% Β±2.27% Β±2.99% url/url-searchparams-creation.js n=1000000 inputType='object' type='encodemany' *** -35.23 % Β±1.30% Β±1.74% Β±2.28% url/url-searchparams-creation.js n=1000000 inputType='object' type='multiprimitives' *** -35.42 % Β±2.37% Β±3.17% Β±4.17% url/url-searchparams-creation.js n=1000000 inputType='object' type='noencode' *** -35.35 % Β±1.82% Β±2.43% Β±3.19% url/url-searchparams-creation.js n=1000000 inputType='string' type='array' *** -23.30 % Β±2.24% Β±2.98% Β±3.88% url/url-searchparams-creation.js n=1000000 inputType='string' type='encodelast' *** -26.13 % Β±2.25% Β±3.02% Β±3.98% url/url-searchparams-creation.js n=1000000 inputType='string' type='encodemany' *** -23.93 % Β±1.70% Β±2.27% Β±2.95% url/url-searchparams-creation.js n=1000000 inputType='string' type='multiprimitives' *** -24.05 % Β±1.58% Β±2.10% Β±2.74% url/url-searchparams-creation.js n=1000000 inputType='string' type='noencode' *** -23.97 % Β±1.77% Β±2.35% Β±3.06% url/url-searchparams-read.js n=20000000 param='nonexistent' accessMethod='get' *** -60.10 % Β±1.35% Β±1.80% Β±2.34% url/url-searchparams-read.js n=20000000 param='nonexistent' accessMethod='getAll' *** -57.29 % Β±2.38% Β±3.20% Β±4.22% url/url-searchparams-read.js n=20000000 param='nonexistent' accessMethod='has' *** -58.05 % Β±2.19% Β±2.94% Β±3.87% url/url-searchparams-read.js n=20000000 param='one' accessMethod='get' *** -60.58 % Β±1.58% Β±2.11% Β±2.75% url/url-searchparams-read.js n=20000000 param='one' accessMethod='getAll' *** -39.72 % Β±2.91% Β±3.87% Β±5.04% url/url-searchparams-read.js n=20000000 param='one' accessMethod='has' *** -60.04 % Β±2.66% Β±3.57% Β±4.71% url/url-searchparams-read.js n=20000000 param='three' accessMethod='get' *** -57.71 % Β±1.36% Β±1.81% Β±2.36% url/url-searchparams-read.js n=20000000 param='three' accessMethod='getAll' *** -37.37 % Β±2.63% Β±3.51% Β±4.60% url/url-searchparams-read.js n=20000000 param='three' accessMethod='has' *** -58.09 % Β±2.38% Β±3.19% Β±4.20% url/url-searchparams-read.js n=20000000 param='two' accessMethod='get' *** -58.71 % Β±1.70% Β±2.27% Β±2.95% url/url-searchparams-read.js n=20000000 param='two' accessMethod='getAll' *** -37.93 % Β±2.61% Β±3.47% Β±4.52% url/url-searchparams-read.js n=20000000 param='two' accessMethod='has' *** -57.99 % Β±2.55% Β±3.40% Β±4.44% url/usvstring.js n=50000000 input='nonstring' *** -45.54 % Β±3.05% Β±4.08% Β±5.35% url/usvstring.js n=50000000 input='valid' *** -73.28 % Β±2.34% Β±3.14% Β±4.15% url/usvstring.js n=50000000 input='validsurr' *** -14.33 % Β±2.56% Β±3.42% Β±4.48% ``` /cc @anonrig
non_main
large performance regressions since since the pr was landed before the ci benchmark machine was restarted these regressions were not caught before it was merged i m not sure if there are other built in modules affected but the url module is one such user of tousvstring url legacy vs whatwg url searchparams parse js n method whatwg searchparam altspaces Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam encodefake Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam encodelast Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam encodemany Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam manyblankpairs Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam manypairs Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam multicharsep Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam multivalue Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam multivaluemany Β± Β± Β± url legacy vs whatwg url searchparams parse js n method whatwg searchparam noencode Β± Β± Β± url url searchparams creation js n inputtype iterable type array Β± Β± Β± url url searchparams creation js n inputtype iterable type encodelast Β± Β± Β± url url searchparams creation js n inputtype iterable type encodemany Β± Β± Β± url url searchparams creation js n inputtype iterable type multiprimitives Β± Β± Β± url url searchparams creation js n inputtype iterable type noencode Β± Β± Β± url url searchparams creation js n inputtype object type array Β± Β± Β± url url searchparams creation js n inputtype object type encodelast Β± Β± Β± url url searchparams creation js n inputtype object type encodemany Β± Β± Β± url url searchparams creation js n inputtype object type multiprimitives Β± Β± Β± url url searchparams creation js n inputtype object type noencode Β± Β± Β± url url searchparams creation js n inputtype string type array Β± Β± Β± url url searchparams creation js n inputtype string type encodelast Β± Β± Β± url url searchparams creation js n inputtype string type encodemany Β± Β± Β± url url searchparams creation js n inputtype string type multiprimitives Β± Β± Β± url url searchparams creation js n inputtype string type noencode Β± Β± Β± url url searchparams read js n param nonexistent accessmethod get Β± Β± Β± url url searchparams read js n param nonexistent accessmethod getall Β± Β± Β± url url searchparams read js n param nonexistent accessmethod has Β± Β± Β± url url searchparams read js n param one accessmethod get Β± Β± Β± url url searchparams read js n param one accessmethod getall Β± Β± Β± url url searchparams read js n param one accessmethod has Β± Β± Β± url url searchparams read js n param three accessmethod get Β± Β± Β± url url searchparams read js n param three accessmethod getall Β± Β± Β± url url searchparams read js n param three accessmethod has Β± Β± Β± url url searchparams read js n param two accessmethod get Β± Β± Β± url url searchparams read js n param two accessmethod getall Β± Β± Β± url url searchparams read js n param two accessmethod has Β± Β± Β± url usvstring js n input nonstring Β± Β± Β± url usvstring js n input valid Β± Β± Β± url usvstring js n input validsurr Β± Β± Β± cc anonrig
0
435,400
12,535,082,957
IssuesEvent
2020-06-04 20:40:22
microsoftgraph/microsoft-graph-toolkit
https://api.github.com/repos/microsoftgraph/microsoft-graph-toolkit
closed
[BUG] Tasks people picker has unnecessary margin
1.3 Component: tasks Needs: Triage :mag: Priority: 1 State: In Review bug
<!-- Please make sure you are posting an issue pertaining to the Microsoft Graph Toolkit. --> <!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> **Describe the bug** After the redesign of the people picker, there is an unnecessary top margin in the tasks component and looks weird. ![image](https://user-images.githubusercontent.com/711864/83701262-fb324880-a5bd-11ea-8ad7-d0f61786f0f5.png)
1.0
[BUG] Tasks people picker has unnecessary margin - <!-- Please make sure you are posting an issue pertaining to the Microsoft Graph Toolkit. --> <!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> **Describe the bug** After the redesign of the people picker, there is an unnecessary top margin in the tasks component and looks weird. ![image](https://user-images.githubusercontent.com/711864/83701262-fb324880-a5bd-11ea-8ad7-d0f61786f0f5.png)
non_main
tasks people picker has unnecessary margin describe the bug after the redesign of the people picker there is an unnecessary top margin in the tasks component and looks weird
0
1,105
4,981,797,398
IssuesEvent
2016-12-07 09:17:26
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
The DM files often have duplicated or unhelpful names
Maintainability - Hinders improvements - Not a bug
Guns folder is the worst example, where there is energy.dm for energy guns as well as energy.dm for projectiles as well as energy.dm for casings. There is also fourth energy dm elsewhere in the code. Same shit for projectile.dm. I request that someone A) Tell me how to change a dm files name without ruining the file history B) Make it a code standard that you name your file something helpful/not duplicated (for example, larva_powers, slime_powers, handofgod_powers instead of multiple powers.dm, or the dozen life.dm and death.dm files we have)
True
The DM files often have duplicated or unhelpful names - Guns folder is the worst example, where there is energy.dm for energy guns as well as energy.dm for projectiles as well as energy.dm for casings. There is also fourth energy dm elsewhere in the code. Same shit for projectile.dm. I request that someone A) Tell me how to change a dm files name without ruining the file history B) Make it a code standard that you name your file something helpful/not duplicated (for example, larva_powers, slime_powers, handofgod_powers instead of multiple powers.dm, or the dozen life.dm and death.dm files we have)
main
the dm files often have duplicated or unhelpful names guns folder is the worst example where there is energy dm for energy guns as well as energy dm for projectiles as well as energy dm for casings there is also fourth energy dm elsewhere in the code same shit for projectile dm i request that someone a tell me how to change a dm files name without ruining the file history b make it a code standard that you name your file something helpful not duplicated for example larva powers slime powers handofgod powers instead of multiple powers dm or the dozen life dm and death dm files we have
1
260,618
22,635,272,200
IssuesEvent
2022-06-30 18:18:29
briandfoy/mac-propertylist
https://api.github.com/repos/briandfoy/mac-propertylist
closed
Invalid XML in t/array.t (extra closing </array>)
tests
@kulp (working on a SAX) version noted that an extra closing tag is in the XML in t/array.t.
1.0
Invalid XML in t/array.t (extra closing </array>) - @kulp (working on a SAX) version noted that an extra closing tag is in the XML in t/array.t.
non_main
invalid xml in t array t extra closing kulp working on a sax version noted that an extra closing tag is in the xml in t array t
0
25,971
12,810,635,893
IssuesEvent
2020-07-03 19:23:33
hajimehoshi/ebiten
https://api.github.com/repos/hajimehoshi/ebiten
closed
Low performance with multiple Ebiten processes
bug external performance wontfix
@Shnifer reported: > AMD Phenom II X4, Win 7 minimal test case: > https://github.com/Shnifer/magellan/tree/master/tests/performance_test/ > if i run it 3 times fps = 60/3 = 20 fps, and two instances have 60/2 = 30 fps > as if they all have one common 60hz ticker and use it consistently ) > ( in fact image and text are not needed, just a black screen shows the same fps )
True
Low performance with multiple Ebiten processes - @Shnifer reported: > AMD Phenom II X4, Win 7 minimal test case: > https://github.com/Shnifer/magellan/tree/master/tests/performance_test/ > if i run it 3 times fps = 60/3 = 20 fps, and two instances have 60/2 = 30 fps > as if they all have one common 60hz ticker and use it consistently ) > ( in fact image and text are not needed, just a black screen shows the same fps )
non_main
low performance with multiple ebiten processes shnifer reported amd phenom ii win minimal test case if i run it times fps fps and two instances have fps as if they all have one common ticker and use it consistently in fact image and text are not needed just a black screen shows the same fps
0
4,489
23,375,624,273
IssuesEvent
2022-08-11 02:28:07
restqa/restqa
https://api.github.com/repos/restqa/restqa
closed
[Dashboard] Delete file on the editor section
enhancement wontfix pair with maintainer
Hello πŸ‘‹, ### πŸ‘€ Background While using RestQA from command line, we can delete file on our own since we have access to the file systems. However if we are using RestQA on dashboard mode we have limited access to the file system. ### ✌️ What is the actual behavior? Currently there is no way to delete a file from the dashboard. ### πŸ•΅οΈβ€β™€οΈ How to reproduce the current behavior? 1. Install RestQA `npm i -g @restqa/restqa` 2. Initiate a RestQA project `restqa init` 3. Run the dashboard `restqa dashboard` 4. Access to the dashboard from your browser 5. Select Editor from the left menu. ### 🀞 What is the expected behavior? In the editor section from the dashboard we should have a way to interact with the file system in order to delete a feature file. ### 😎 Proposed solution. - [ ] Add a new button on the editor section of the dashboard: `Delete the feature file` - [ ] Only allow the deletion of `.feature` file extension - [ ] Create an new endpoint on the API layer `​DELETE /api​/project​/features​/{path}` - [ ] Do not allow the creation of file if the dashboard is running on read only mode - [ ] Find the best UI approach for the button Cheers.
True
[Dashboard] Delete file on the editor section - Hello πŸ‘‹, ### πŸ‘€ Background While using RestQA from command line, we can delete file on our own since we have access to the file systems. However if we are using RestQA on dashboard mode we have limited access to the file system. ### ✌️ What is the actual behavior? Currently there is no way to delete a file from the dashboard. ### πŸ•΅οΈβ€β™€οΈ How to reproduce the current behavior? 1. Install RestQA `npm i -g @restqa/restqa` 2. Initiate a RestQA project `restqa init` 3. Run the dashboard `restqa dashboard` 4. Access to the dashboard from your browser 5. Select Editor from the left menu. ### 🀞 What is the expected behavior? In the editor section from the dashboard we should have a way to interact with the file system in order to delete a feature file. ### 😎 Proposed solution. - [ ] Add a new button on the editor section of the dashboard: `Delete the feature file` - [ ] Only allow the deletion of `.feature` file extension - [ ] Create an new endpoint on the API layer `​DELETE /api​/project​/features​/{path}` - [ ] Do not allow the creation of file if the dashboard is running on read only mode - [ ] Find the best UI approach for the button Cheers.
main
delete file on the editor section hello πŸ‘‹ πŸ‘€ background while using restqa from command line we can delete file on our own since we have access to the file systems however if we are using restqa on dashboard mode we have limited access to the file system ✌️ what is the actual behavior currently there is no way to delete a file from the dashboard πŸ•΅οΈβ€β™€οΈ how to reproduce the current behavior install restqa npm i g restqa restqa initiate a restqa project restqa init run the dashboard restqa dashboard access to the dashboard from your browser select editor from the left menu 🀞 what is the expected behavior in the editor section from the dashboard we should have a way to interact with the file system in order to delete a feature file 😎 proposed solution add a new button on the editor section of the dashboard delete the feature file only allow the deletion of feature file extension create an new endpoint on the api layer ​delete api​ project​ features​ path do not allow the creation of file if the dashboard is running on read only mode find the best ui approach for the button cheers
1
4,961
25,476,798,253
IssuesEvent
2022-11-25 15:14:00
libp2p/js-libp2p
https://api.github.com/repos/libp2p/js-libp2p
closed
Duplicated addresses in _createDialTarget
kind/bug P1 need/maintainer-input
<!-- Thank you for reporting an issue. This issue tracker is for bugs found within the JavaScript implementation of libp2p. If you are asking a question about how to use libp2p, please ask on https://discuss.libp2p.io Otherwise please fill in as much of the template below as possible. --> - **Version**: 0.39.2 https://github.com/ChainSafe/lodestar/blob/ca27458b59c66c9d3da7dd95322e479f78407269/packages/beacon-node/package.json#L144 - **Platform**: Linux - Severity: High #### Description: Running local devnets with Lodestar nodes couldn't dial with each other due to this error ``` Oct-05 22:08:39.721[rest] error: Req req-1 gossipsubAttack error dial with more addresses than allowed 5 > 4 Error: dial with more addresses than allowed 5 > 4 at DefaultDialer._createDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:264:21) at DefaultDialer._createCancellableDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:217:26) at DefaultDialer.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:179:24) at EventTarget.openConnection (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/index.ts:521:26) at EventTarget.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/libp2p.ts:382:12) ``` I manually called dial submitting a single Multiaddr with `/ip4/127.0.0.1/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw`. But how can there be 5 addresses if I only ever provided a single IP port combo? Then added a console log to print the multiaddr: ``` Trace: [ <Multiaddr 047f000001062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/127.0.0.1/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 047f000001062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/127.0.0.1/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 04c0a80214062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/192.168.2.20/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 04c0a802ef062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/192.168.2.239/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 04c0a802ef062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/192.168.2.239/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw> ] at DefaultDialer._createDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:264:13) at DefaultDialer._createCancellableDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:217:26) at DefaultDialer.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:179:24) at EventTarget.openConnection (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/index.ts:521:26) at EventTarget.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/libp2p.ts:382:12) ``` so: - Why are addresses duplicated? - Why are IPs `192.168.2.20` and `192.168.2.239` added to the targets? #### Steps to reproduce the error: On a local devnet - Start node 1 with dev command - Start node 2 with dev command - Call connectPeer API on node 2 with node 1 identity
True
Duplicated addresses in _createDialTarget - <!-- Thank you for reporting an issue. This issue tracker is for bugs found within the JavaScript implementation of libp2p. If you are asking a question about how to use libp2p, please ask on https://discuss.libp2p.io Otherwise please fill in as much of the template below as possible. --> - **Version**: 0.39.2 https://github.com/ChainSafe/lodestar/blob/ca27458b59c66c9d3da7dd95322e479f78407269/packages/beacon-node/package.json#L144 - **Platform**: Linux - Severity: High #### Description: Running local devnets with Lodestar nodes couldn't dial with each other due to this error ``` Oct-05 22:08:39.721[rest] error: Req req-1 gossipsubAttack error dial with more addresses than allowed 5 > 4 Error: dial with more addresses than allowed 5 > 4 at DefaultDialer._createDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:264:21) at DefaultDialer._createCancellableDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:217:26) at DefaultDialer.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:179:24) at EventTarget.openConnection (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/index.ts:521:26) at EventTarget.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/libp2p.ts:382:12) ``` I manually called dial submitting a single Multiaddr with `/ip4/127.0.0.1/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw`. But how can there be 5 addresses if I only ever provided a single IP port combo? Then added a console log to print the multiaddr: ``` Trace: [ <Multiaddr 047f000001062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/127.0.0.1/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 047f000001062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/127.0.0.1/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 04c0a80214062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/192.168.2.20/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 04c0a802ef062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/192.168.2.239/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw>, <Multiaddr 04c0a802ef062328a5032700250802122102a6e664edcd4b990208e101fe7486282a205d4a88edae4be6d54ead4257bb1f98 - /ip4/192.168.2.239/tcp/9000/p2p/16Uiu2HAm6fBRJ3MWKUpVhscGdPkrUJjB943JfG8hbg41k9soaNpw> ] at DefaultDialer._createDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:264:13) at DefaultDialer._createCancellableDialTarget (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:217:26) at DefaultDialer.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/dialer/index.ts:179:24) at EventTarget.openConnection (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/connection-manager/index.ts:521:26) at EventTarget.dial (file:///home/lion/Code/eth2.0/lodestar-attack/node_modules/libp2p/src/libp2p.ts:382:12) ``` so: - Why are addresses duplicated? - Why are IPs `192.168.2.20` and `192.168.2.239` added to the targets? #### Steps to reproduce the error: On a local devnet - Start node 1 with dev command - Start node 2 with dev command - Call connectPeer API on node 2 with node 1 identity
main
duplicated addresses in createdialtarget thank you for reporting an issue this issue tracker is for bugs found within the javascript implementation of if you are asking a question about how to use please ask on otherwise please fill in as much of the template below as possible version platform linux severity high description running local devnets with lodestar nodes couldn t dial with each other due to this error oct error req req gossipsubattack error dial with more addresses than allowed error dial with more addresses than allowed at defaultdialer createdialtarget file home lion code lodestar attack node modules src connection manager dialer index ts at defaultdialer createcancellabledialtarget file home lion code lodestar attack node modules src connection manager dialer index ts at defaultdialer dial file home lion code lodestar attack node modules src connection manager dialer index ts at eventtarget openconnection file home lion code lodestar attack node modules src connection manager index ts at eventtarget dial file home lion code lodestar attack node modules src ts i manually called dial submitting a single multiaddr with tcp but how can there be addresses if i only ever provided a single ip port combo then added a console log to print the multiaddr trace at defaultdialer createdialtarget file home lion code lodestar attack node modules src connection manager dialer index ts at defaultdialer createcancellabledialtarget file home lion code lodestar attack node modules src connection manager dialer index ts at defaultdialer dial file home lion code lodestar attack node modules src connection manager dialer index ts at eventtarget openconnection file home lion code lodestar attack node modules src connection manager index ts at eventtarget dial file home lion code lodestar attack node modules src ts so why are addresses duplicated why are ips and added to the targets steps to reproduce the error on a local devnet start node with dev command start node with dev command call connectpeer api on node with node identity
1
3,276
12,506,427,436
IssuesEvent
2020-06-02 12:34:15
diofant/diofant
https://api.github.com/repos/diofant/diofant
closed
Port factorization algorithms to use sparse polys
enhancement maintainability polys
That include implementing factor_list method and is_cyclotomic property per [TODO](https://github.com/diofant/diofant/blob/987094a0a22b101e545e291af5ee9f526233f279/diofant/polys/rings.py#L2058-L2059). I.e., porting current factortools and galoistools modules.
True
Port factorization algorithms to use sparse polys - That include implementing factor_list method and is_cyclotomic property per [TODO](https://github.com/diofant/diofant/blob/987094a0a22b101e545e291af5ee9f526233f279/diofant/polys/rings.py#L2058-L2059). I.e., porting current factortools and galoistools modules.
main
port factorization algorithms to use sparse polys that include implementing factor list method and is cyclotomic property per i e porting current factortools and galoistools modules
1
34,591
4,933,371,042
IssuesEvent
2016-11-28 16:11:49
mozilla/fxa-content-server
https://api.github.com/repos/mozilla/fxa-content-server
closed
local Failing OAuth functional test - oauth reset password - reset password, verify in a different browser, from the new browser's P.O.V.
tests waffle:active
Using Selenium 2.53.0 and Fx 46.0.1 ``` Γ— firefox on any platform - oauth reset password - reset password, verify in a different browser, from the new browser's P.O.V. (2.064s) StaleElementReference: [POST http://localhost:4444/wd/hub/session/aefe94f1-0e9f-4ed9-b450-6f1bfcf61f2c/element/30/element / {"using":"css selector","value":"#footer-main"}] Element not found in the cache - perhaps the page has changed since it was looked up For documentation on this error, please visit: http://seleniumhq.org/exceptions/stale_element_reference.html Build info: version: '2.53.0', revision: '35ae25b', time: '2016-03-15 17:00:58' System info: host: 'Shanes-Retina-MBP.local', ip: '10.239.30.167', os.name: 'Mac OS X', os.arch: 'x86_64', os.version: '10.12.1', java.version: '1.8.0_102' Driver info: driver.version: unknown at runRequest <node_modules/intern/node_modules/leadfoot/Session.js:88:40> at <node_modules/intern/node_modules/leadfoot/Session.js:109:39> at new Promise <node_modules/intern/node_modules/dojo/Promise.ts:411:3> at ProxiedSession._post <node_modules/intern/node_modules/leadfoot/Session.js:63:10> at Element._post <node_modules/intern/node_modules/leadfoot/Element.js:23:31> at Element.find <node_modules/intern/node_modules/leadfoot/Element.js:114:15> at Command.<anonymous> <node_modules/intern/node_modules/leadfoot/Command.js:28:40> at <node_modules/intern/node_modules/dojo/Promise.ts:393:15> at run <node_modules/intern/node_modules/dojo/Promise.ts:237:7> at <node_modules/intern/node_modules/dojo/nextTick.ts:44:3> at Command.find <node_modules/intern/node_modules/leadfoot/Command.js:23:10> at Command.prototype.(anonymous function) [as findByCssSelector] <node_modules/intern/node_modules/leadfoot/lib/strategies.js:24:16> at clear123DoneState <tests/functional/lib/helpers.js:128:8> at Command.<anonymous> <tests/functional/lib/helpers.js:61:20> at runCallback <node_modules/intern/node_modules/leadfoot/Command.js:526:31> at Command.<anonymous> <node_modules/intern/node_modules/leadfoot/Command.js:543:11> at <node_modules/intern/node_modules/dojo/Promise.ts:393:15> at run <node_modules/intern/node_modules/dojo/Promise.ts:237:7> at <node_modules/intern/node_modules/dojo/nextTick.ts:44:3> at nextTickCallbackWith0Args <node.js:420:9> at Command.then <node_modules/intern/node_modules/leadfoot/Command.js:542:10> at Command.<anonymous> <tests/functional/lib/helpers.js:59:10> at runCallback <node_modules/intern/node_modules/leadfoot/Command.js:526:31> at Command.<anonymous> <node_modules/intern/node_modules/leadfoot/Command.js:543:11> at <node_modules/intern/node_modules/dojo/Promise.ts:393:15> at run <node_modules/intern/node_modules/dojo/Promise.ts:237:7> at <node_modules/intern/node_modules/dojo/nextTick.ts:44:3> at nextTickCallbackWith0Args <node.js:420:9> at process._tickCallback <node.js:349:13> at Command.then <node_modules/intern/node_modules/leadfoot/Command.js:542:10> at Test.registerSuite.reset password, verify in a different browser, from the new browser's P.O.V. [as test] <tests/functional/oauth_reset_password.js:225:10> at <node_modules/intern/lib/Test.js:181:24> at <node_modules/intern/browser_modules/dojo/Promise.ts:393:15> at runCallbacks <node_modules/intern/browser_modules/dojo/Promise.ts:11:11> at <node_modules/intern/browser_modules/dojo/Promise.ts:317:4> at run <node_modules/intern/browser_modules/dojo/Promise.ts:237:7> at <node_modules/intern/browser_modules/dojo/nextTick.ts:44:3> at nextTickCallbackWith0Args <node.js:420:9> at process._tickCallback <node.js:349:13> ```
1.0
local Failing OAuth functional test - oauth reset password - reset password, verify in a different browser, from the new browser's P.O.V. - Using Selenium 2.53.0 and Fx 46.0.1 ``` Γ— firefox on any platform - oauth reset password - reset password, verify in a different browser, from the new browser's P.O.V. (2.064s) StaleElementReference: [POST http://localhost:4444/wd/hub/session/aefe94f1-0e9f-4ed9-b450-6f1bfcf61f2c/element/30/element / {"using":"css selector","value":"#footer-main"}] Element not found in the cache - perhaps the page has changed since it was looked up For documentation on this error, please visit: http://seleniumhq.org/exceptions/stale_element_reference.html Build info: version: '2.53.0', revision: '35ae25b', time: '2016-03-15 17:00:58' System info: host: 'Shanes-Retina-MBP.local', ip: '10.239.30.167', os.name: 'Mac OS X', os.arch: 'x86_64', os.version: '10.12.1', java.version: '1.8.0_102' Driver info: driver.version: unknown at runRequest <node_modules/intern/node_modules/leadfoot/Session.js:88:40> at <node_modules/intern/node_modules/leadfoot/Session.js:109:39> at new Promise <node_modules/intern/node_modules/dojo/Promise.ts:411:3> at ProxiedSession._post <node_modules/intern/node_modules/leadfoot/Session.js:63:10> at Element._post <node_modules/intern/node_modules/leadfoot/Element.js:23:31> at Element.find <node_modules/intern/node_modules/leadfoot/Element.js:114:15> at Command.<anonymous> <node_modules/intern/node_modules/leadfoot/Command.js:28:40> at <node_modules/intern/node_modules/dojo/Promise.ts:393:15> at run <node_modules/intern/node_modules/dojo/Promise.ts:237:7> at <node_modules/intern/node_modules/dojo/nextTick.ts:44:3> at Command.find <node_modules/intern/node_modules/leadfoot/Command.js:23:10> at Command.prototype.(anonymous function) [as findByCssSelector] <node_modules/intern/node_modules/leadfoot/lib/strategies.js:24:16> at clear123DoneState <tests/functional/lib/helpers.js:128:8> at Command.<anonymous> <tests/functional/lib/helpers.js:61:20> at runCallback <node_modules/intern/node_modules/leadfoot/Command.js:526:31> at Command.<anonymous> <node_modules/intern/node_modules/leadfoot/Command.js:543:11> at <node_modules/intern/node_modules/dojo/Promise.ts:393:15> at run <node_modules/intern/node_modules/dojo/Promise.ts:237:7> at <node_modules/intern/node_modules/dojo/nextTick.ts:44:3> at nextTickCallbackWith0Args <node.js:420:9> at Command.then <node_modules/intern/node_modules/leadfoot/Command.js:542:10> at Command.<anonymous> <tests/functional/lib/helpers.js:59:10> at runCallback <node_modules/intern/node_modules/leadfoot/Command.js:526:31> at Command.<anonymous> <node_modules/intern/node_modules/leadfoot/Command.js:543:11> at <node_modules/intern/node_modules/dojo/Promise.ts:393:15> at run <node_modules/intern/node_modules/dojo/Promise.ts:237:7> at <node_modules/intern/node_modules/dojo/nextTick.ts:44:3> at nextTickCallbackWith0Args <node.js:420:9> at process._tickCallback <node.js:349:13> at Command.then <node_modules/intern/node_modules/leadfoot/Command.js:542:10> at Test.registerSuite.reset password, verify in a different browser, from the new browser's P.O.V. [as test] <tests/functional/oauth_reset_password.js:225:10> at <node_modules/intern/lib/Test.js:181:24> at <node_modules/intern/browser_modules/dojo/Promise.ts:393:15> at runCallbacks <node_modules/intern/browser_modules/dojo/Promise.ts:11:11> at <node_modules/intern/browser_modules/dojo/Promise.ts:317:4> at run <node_modules/intern/browser_modules/dojo/Promise.ts:237:7> at <node_modules/intern/browser_modules/dojo/nextTick.ts:44:3> at nextTickCallbackWith0Args <node.js:420:9> at process._tickCallback <node.js:349:13> ```
non_main
local failing oauth functional test oauth reset password reset password verify in a different browser from the new browser s p o v using selenium and fx Γ— firefox on any platform oauth reset password reset password verify in a different browser from the new browser s p o v staleelementreference element not found in the cache perhaps the page has changed since it was looked up for documentation on this error please visit build info version revision time system info host shanes retina mbp local ip os name mac os x os arch os version java version driver info driver version unknown at runrequest at at new promise at proxiedsession post at element post at element find at command at at run at at command find at command prototype anonymous function at at command at runcallback at command at at run at at at command then at command at runcallback at command at at run at at at process tickcallback at command then at test registersuite reset password verify in a different browser from the new browser s p o v at at at runcallbacks at at run at at at process tickcallback
0
1,794
6,575,901,688
IssuesEvent
2017-09-11 17:46:07
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
sysctl: parameter value check for already defined params
affects_2.3 feature_idea waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> sysctl ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> ##### SUMMARY <!--- Explain the problem briefly --> Multiple roles/plays can modify the sysctl and set different values for the same parameter. My suggestion is to add a check functionality, that - if set to yes - checks if the new value is greater ( or less or equal) than the existing value. If so, it sets it, otherwise skip it. The problem at the moment is that I can't check the value and set it in one task. It doesn't look clean at the moment and is just a workaround. example follows. I know that this can't work for all parameter, because i.e. kernel.sem contains a string of parameters. But for the mayority it should be a benefit. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: check parameter fs.file-max shell: "sysctl -a | grep fs.file-max | grep -o [0-9]*" args: executable: /bin/bash register: fs_file_max - name: set parameter in sysctl fs.file-max sysctl: name="fs.file-max" value=6815744 state=present when: fs_file_max.stdout <= 6815744 # Would be nice to have something like that: - name: set parameter in sysctl fs.file-max sysctl: name="fs.file-max" value=6815744 state=present check=yes check-params: ge # ge - greater equal, le - less equal, e - equal ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` ```
True
sysctl: parameter value check for already defined params - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> sysctl ##### ANSIBLE VERSION <!--- Paste verbatim output from β€œansible --version” between quotes below --> ``` ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say β€œN/A” for anything that is not platform-specific. --> ##### SUMMARY <!--- Explain the problem briefly --> Multiple roles/plays can modify the sysctl and set different values for the same parameter. My suggestion is to add a check functionality, that - if set to yes - checks if the new value is greater ( or less or equal) than the existing value. If so, it sets it, otherwise skip it. The problem at the moment is that I can't check the value and set it in one task. It doesn't look clean at the moment and is just a workaround. example follows. I know that this can't work for all parameter, because i.e. kernel.sem contains a string of parameters. But for the mayority it should be a benefit. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: check parameter fs.file-max shell: "sysctl -a | grep fs.file-max | grep -o [0-9]*" args: executable: /bin/bash register: fs_file_max - name: set parameter in sysctl fs.file-max sysctl: name="fs.file-max" value=6815744 state=present when: fs_file_max.stdout <= 6815744 # Would be nice to have something like that: - name: set parameter in sysctl fs.file-max sysctl: name="fs.file-max" value=6815744 state=present check=yes check-params: ge # ge - greater equal, le - less equal, e - equal ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` ```
main
sysctl parameter value check for already defined params issue type feature idea component name sysctl ansible version configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say β€œn a” for anything that is not platform specific summary multiple roles plays can modify the sysctl and set different values for the same parameter my suggestion is to add a check functionality that if set to yes checks if the new value is greater or less or equal than the existing value if so it sets it otherwise skip it the problem at the moment is that i can t check the value and set it in one task it doesn t look clean at the moment and is just a workaround example follows i know that this can t work for all parameter because i e kernel sem contains a string of parameters but for the mayority it should be a benefit steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name check parameter fs file max shell sysctl a grep fs file max grep o args executable bin bash register fs file max name set parameter in sysctl fs file max sysctl name fs file max value state present when fs file max stdout would be nice to have something like that name set parameter in sysctl fs file max sysctl name fs file max value state present check yes check params ge ge greater equal le less equal e equal expected results actual results
1
111,057
17,009,699,334
IssuesEvent
2021-07-02 01:08:44
tamirdahan/keycloak
https://api.github.com/repos/tamirdahan/keycloak
opened
CVE-2015-9251 (Medium) detected in multiple libraries
security vulnerability
## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.11.1.min.js</b>, <b>jquery-2.2.0.min.js</b>, <b>jquery-1.11.3.min.js</b>, <b>jquery-1.11.0.min.js</b>, <b>jquery-2.1.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.11.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-2.2.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html,keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-2.2.0.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html,keycloak/common/resources/node_modules/bootstrap-select/test.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.3.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.0.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-2.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-slider/dependencies/js/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.1.min.js** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.2.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.3","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-select/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-9251","vulnerabilityDetails":"jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2015-9251 (Medium) detected in multiple libraries - ## CVE-2015-9251 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.11.1.min.js</b>, <b>jquery-2.2.0.min.js</b>, <b>jquery-1.11.3.min.js</b>, <b>jquery-1.11.0.min.js</b>, <b>jquery-2.1.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.11.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/c3/htdocs/samples/element.html,keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-2.2.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.0/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html,keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-2.2.0.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.3.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.3/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html,keycloak/common/resources/node_modules/bootstrap-select/test.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.3.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.0/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/./lib/jquery-1.11.0.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.0.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-2.1.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.1/jquery.min.js</a></p> <p>Path to dependency file: keycloak/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html</p> <p>Path to vulnerable library: keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html,keycloak/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-slider/dependencies/js/jquery.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-2.1.1.min.js** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed. <p>Publish Date: 2018-01-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251>CVE-2015-9251</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-9251">https://nvd.nist.gov/vuln/detail/CVE-2015-9251</a></p> <p>Release Date: 2018-01-18</p> <p>Fix Resolution: jQuery - v3.0.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-touchspin/demo/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/c3/htdocs/samples/element.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/c3/htdocs/samples/element.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-touchspin/demo/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.2.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/docs/playground/index.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/index.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.2.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.3","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/docs/playground/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/bootstrap-select/docs/custom_theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/bootstrap-select/test.html","/themes/target/classes/theme/keycloak/common/resources/node_modules/bootstrap-select/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.0","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/jquery-match-height/test/page/test.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/jquery-match-height/test/page/test.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:1.11.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.1","packageFilePaths":["/themes/src/main/resources/theme/keycloak/common/resources/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html","/themes/src/main/resources/theme/keycloak/common/resources/node_modules/rcue/node_modules/patternfly/node_modules/eonasdan-bootstrap-datetimepicker/docs/theme/base.html"],"isTransitiveDependency":false,"dependencyTree":"jquery:2.1.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jQuery - v3.0.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2015-9251","vulnerabilityDetails":"jQuery before 3.0.0 is vulnerable to Cross-site Scripting (XSS) attacks when a cross-domain Ajax request is performed without the dataType option, causing text/javascript responses to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9251","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules rcue node modules patternfly node modules bootstrap touchspin demo index html path to vulnerable library keycloak common resources node modules rcue node modules patternfly node modules bootstrap touchspin demo index html keycloak common resources node modules rcue node modules patternfly node modules htdocs samples element html keycloak common resources node modules htdocs samples element html keycloak common resources node modules htdocs samples element html keycloak common resources node modules bootstrap touchspin demo index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules rcue node modules patternfly node modules bootstrap select docs docs playground index html path to vulnerable library keycloak common resources node modules rcue node modules patternfly node modules bootstrap select docs docs playground index html keycloak common resources node modules bootstrap select docs docs playground index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules bootstrap select docs docs playground test html path to vulnerable library keycloak common resources node modules bootstrap select docs docs playground test html keycloak common resources node modules rcue node modules patternfly node modules bootstrap select docs custom theme base html keycloak common resources node modules bootstrap select docs custom theme base html keycloak common resources node modules rcue node modules patternfly node modules bootstrap select test html keycloak common resources node modules bootstrap select test html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules jquery match height test page test html path to vulnerable library keycloak common resources node modules jquery match height test page lib jquery min js keycloak common resources node modules rcue node modules patternfly node modules jquery match height test page lib jquery min js dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file keycloak themes src main resources theme keycloak common resources node modules eonasdan bootstrap datetimepicker docs theme base html path to vulnerable library keycloak common resources node modules eonasdan bootstrap datetimepicker docs theme base html keycloak common resources node modules rcue node modules patternfly node modules eonasdan bootstrap datetimepicker docs theme base html keycloak themes target classes theme keycloak common resources node modules bootstrap slider dependencies js jquery min js dependency hierarchy x jquery min js vulnerable library found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery packagetype javascript packagename jquery packageversion packagefilepaths istransitivedependency false dependencytree jquery isminimumfixversionavailable true minimumfixversion jquery basebranches vulnerabilityidentifier cve vulnerabilitydetails jquery before is vulnerable to cross site scripting xss attacks when a cross domain ajax request is performed without the datatype option causing text javascript responses to be executed vulnerabilityurl
0
42,858
2,874,856,617
IssuesEvent
2015-06-09 02:14:45
mcupak/beacon-of-beacons
https://api.github.com/repos/mcupak/beacon-of-beacons
closed
Update beacon names
priority:high severity:minor type:enhancement
The names of the beacons we are using at the moment are not descriptive enough.
1.0
Update beacon names - The names of the beacons we are using at the moment are not descriptive enough.
non_main
update beacon names the names of the beacons we are using at the moment are not descriptive enough
0
703,549
24,165,908,642
IssuesEvent
2022-09-22 15:01:20
NickleDave/songdkl
https://api.github.com/repos/NickleDave/songdkl
closed
pin scikit-learn version to less than / equal to 0.18.2
bug High Priority
scripts in PCB paper use `GMM`, deprecated in version 0.18.2 https://scikit-learn.org/0.19/whats_new.html @dgmets reports that > there are differences in the Likelihoods emitted from GaussianMixture as compared to the previous GMM module. In particular, the Likelihoods are strongly impacted by the covariance type used. This didn't used to be the case. Anyway, I am going to try to run this down... The current Dkl measures are proportional to the previous ones, but not the same. in the meantime we can pin to a version before deprecation. Using 0.18.2 gives a DeperecationWarning that can be annoying when it gets dumped to stdout 150k times, might be worth using a slightly earlier version
1.0
pin scikit-learn version to less than / equal to 0.18.2 - scripts in PCB paper use `GMM`, deprecated in version 0.18.2 https://scikit-learn.org/0.19/whats_new.html @dgmets reports that > there are differences in the Likelihoods emitted from GaussianMixture as compared to the previous GMM module. In particular, the Likelihoods are strongly impacted by the covariance type used. This didn't used to be the case. Anyway, I am going to try to run this down... The current Dkl measures are proportional to the previous ones, but not the same. in the meantime we can pin to a version before deprecation. Using 0.18.2 gives a DeperecationWarning that can be annoying when it gets dumped to stdout 150k times, might be worth using a slightly earlier version
non_main
pin scikit learn version to less than equal to scripts in pcb paper use gmm deprecated in version dgmets reports that there are differences in the likelihoods emitted from gaussianmixture as compared to the previous gmm module in particular the likelihoods are strongly impacted by the covariance type used this didn t used to be the case anyway i am going to try to run this down the current dkl measures are proportional to the previous ones but not the same in the meantime we can pin to a version before deprecation using gives a deperecationwarning that can be annoying when it gets dumped to stdout times might be worth using a slightly earlier version
0
21,928
11,660,539,337
IssuesEvent
2020-03-03 03:41:24
cityofaustin/atd-geospatial
https://api.github.com/repos/cityofaustin/atd-geospatial
closed
Data-Informed PHB Ranking
Epic Service: Geo Type: Enhancement Workgroup: AMD
Email > renee.orr@austintexas.gov Describe the problem. > AMD is working toward a data-based process to identify locations for new PHBs. We would like to investigate the possibility of using existing data developed for Active Trans' Pedestrian Safety Action Plan (PSAP), and revise it to fit our program. Active Trans is also interested in updating this data for the the PSAP. We anticipate needing to coordinate this GIS data with PHB requests in Data Tracker. Also need to discuss the frequency this data can be updated. How soon do you need this? > Flexible β€” An extended timeline is OK Is there anything else we should know? > Would like to have this process defined and in use by November, so we can meet the stated annual December ranking. I request a scoping meeting be scheduled to discuss this request further. Please include Joel Meyer in the meeting, This process is supported by Jen in response to a request from Jim Dale. Request ID: DTS19-100165
1.0
Data-Informed PHB Ranking - Email > renee.orr@austintexas.gov Describe the problem. > AMD is working toward a data-based process to identify locations for new PHBs. We would like to investigate the possibility of using existing data developed for Active Trans' Pedestrian Safety Action Plan (PSAP), and revise it to fit our program. Active Trans is also interested in updating this data for the the PSAP. We anticipate needing to coordinate this GIS data with PHB requests in Data Tracker. Also need to discuss the frequency this data can be updated. How soon do you need this? > Flexible β€” An extended timeline is OK Is there anything else we should know? > Would like to have this process defined and in use by November, so we can meet the stated annual December ranking. I request a scoping meeting be scheduled to discuss this request further. Please include Joel Meyer in the meeting, This process is supported by Jen in response to a request from Jim Dale. Request ID: DTS19-100165
non_main
data informed phb ranking email renee orr austintexas gov describe the problem amd is working toward a data based process to identify locations for new phbs we would like to investigate the possibility of using existing data developed for active trans pedestrian safety action plan psap and revise it to fit our program active trans is also interested in updating this data for the the psap we anticipate needing to coordinate this gis data with phb requests in data tracker also need to discuss the frequency this data can be updated how soon do you need this flexible β€” an extended timeline is ok is there anything else we should know would like to have this process defined and in use by november so we can meet the stated annual december ranking i request a scoping meeting be scheduled to discuss this request further please include joel meyer in the meeting this process is supported by jen in response to a request from jim dale request id
0
5,456
19,661,658,134
IssuesEvent
2022-01-10 17:38:14
carpentries/amy
https://api.github.com/repos/carpentries/amy
closed
Profile Archival: Yearly reminder emails should be sent in bulk.
type: new feature component: email automation
Generalize the bulk email reminders functionality seen with consents, so that yearly reminders can be sent in bulk instead. See this conversation here for additional context: https://github.com/carpentries/amy/pull/2032#discussion_r686477108
1.0
Profile Archival: Yearly reminder emails should be sent in bulk. - Generalize the bulk email reminders functionality seen with consents, so that yearly reminders can be sent in bulk instead. See this conversation here for additional context: https://github.com/carpentries/amy/pull/2032#discussion_r686477108
non_main
profile archival yearly reminder emails should be sent in bulk generalize the bulk email reminders functionality seen with consents so that yearly reminders can be sent in bulk instead see this conversation here for additional context
0
80,916
10,217,537,233
IssuesEvent
2019-08-15 13:54:22
Royal-Navy/standards-toolkit
https://api.github.com/repos/Royal-Navy/standards-toolkit
closed
Fix 404 for /feedback page
bug documentation
**Describe the bug** Navigating to the `/feedback` page on docs.royalnavy.io results in a 404 page **To Reproduce** Steps to reproduce the behavior: 1. Scroll to the footer 2. Click on the Feedback button 3. See error **Expected behavior** Feedback page should either be created or the user should be redirected to the contact page
1.0
Fix 404 for /feedback page - **Describe the bug** Navigating to the `/feedback` page on docs.royalnavy.io results in a 404 page **To Reproduce** Steps to reproduce the behavior: 1. Scroll to the footer 2. Click on the Feedback button 3. See error **Expected behavior** Feedback page should either be created or the user should be redirected to the contact page
non_main
fix for feedback page describe the bug navigating to the feedback page on docs royalnavy io results in a page to reproduce steps to reproduce the behavior scroll to the footer click on the feedback button see error expected behavior feedback page should either be created or the user should be redirected to the contact page
0
66,611
8,034,026,429
IssuesEvent
2018-07-29 13:58:20
GitHawkApp/GitHawk
https://api.github.com/repos/GitHawkApp/GitHawk
closed
First comment of Issue/PR full width
design
I think this would make a better design. The root comment should be full width, while all replies be cards. <details> <summary>Bug Report Dump (Auto-generated)</summary> <pre> Version 1.22.0 (1531698972) Device: iPhone X (iOS 11.4) TestFlight: true </pre> </details>
1.0
First comment of Issue/PR full width - I think this would make a better design. The root comment should be full width, while all replies be cards. <details> <summary>Bug Report Dump (Auto-generated)</summary> <pre> Version 1.22.0 (1531698972) Device: iPhone X (iOS 11.4) TestFlight: true </pre> </details>
non_main
first comment of issue pr full width i think this would make a better design the root comment should be full width while all replies be cards bug report dump auto generated version device iphone x ios testflight true
0
123,582
4,865,360,600
IssuesEvent
2016-11-14 20:33:10
coreos/dex
https://api.github.com/repos/coreos/dex
closed
api: add version endpoint
low hanging fruit priority/P2 v2
something like this ```proto message Version { // semantic version of the server string server // numeric version of the API int api } message VersionReq {} Version(VersionReq) Version ```
1.0
api: add version endpoint - something like this ```proto message Version { // semantic version of the server string server // numeric version of the API int api } message VersionReq {} Version(VersionReq) Version ```
non_main
api add version endpoint something like this proto message version semantic version of the server string server numeric version of the api int api message versionreq version versionreq version
0
106,783
9,186,834,476
IssuesEvent
2019-03-06 00:18:33
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: jepsen-batch2/g2/majority-ring failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/047556f80399b7f70cb36c09345c6cd244210c2b Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen-batch2/g2/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1161867&tab=buildLog ``` The test failed on master: jepsen.go:247,jepsen.go:308,test.go:1214: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1161867-jepsen-batch2:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.142.0.59 -n 10.142.0.32 -n 10.142.0.67 -n 10.142.0.55 -n 10.142.0.62 \ --test g2 --nemesis majority-ring \ > invoke.log 2>&1 \ " returned: stderr: stdout: Error: exit status 255 : exit status 1 ```
2.0
roachtest: jepsen-batch2/g2/majority-ring failed - SHA: https://github.com/cockroachdb/cockroach/commits/047556f80399b7f70cb36c09345c6cd244210c2b Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen-batch2/g2/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1161867&tab=buildLog ``` The test failed on master: jepsen.go:247,jepsen.go:308,test.go:1214: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1161867-jepsen-batch2:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.142.0.59 -n 10.142.0.32 -n 10.142.0.67 -n 10.142.0.55 -n 10.142.0.62 \ --test g2 --nemesis majority-ring \ > invoke.log 2>&1 \ " returned: stderr: stdout: Error: exit status 255 : exit status 1 ```
non_main
roachtest jepsen majority ring failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests jepsen majority ring pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on master jepsen go jepsen go test go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity jepsen bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test nemesis majority ring invoke log returned stderr stdout error exit status exit status
0
1,436
6,224,838,529
IssuesEvent
2017-07-10 15:00:55
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Calculator: Support context-sensitive constants
Maintainer Submitted Status: On Hold
It might be useful to have constants that can vary based on a specified context. For example, the query `(2G * 1.7e3) / c^2` doesn't yield a result as there are 'undefined' constants; but if you did `(2G * 1.7e3) / c^2 (astrophysics)` we could yield a result informing the user that the calculator interpreted: `G` as the gravitational constant, and `c` as the speed of light in a vacuum. So, as a quick mock-up: ![screenshot from 2016-04-24 12 28 01](https://cloud.githubusercontent.com/assets/8598426/14767202/12c2d9ba-0a18-11e6-8112-32c0516b0844.png) /cc @duckduckgo/duckduckhack-contributors Would any of you find this useful? --- IA Page: http://duck.co/ia/view/calculator [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @GuiltyDolphin
True
Calculator: Support context-sensitive constants - It might be useful to have constants that can vary based on a specified context. For example, the query `(2G * 1.7e3) / c^2` doesn't yield a result as there are 'undefined' constants; but if you did `(2G * 1.7e3) / c^2 (astrophysics)` we could yield a result informing the user that the calculator interpreted: `G` as the gravitational constant, and `c` as the speed of light in a vacuum. So, as a quick mock-up: ![screenshot from 2016-04-24 12 28 01](https://cloud.githubusercontent.com/assets/8598426/14767202/12c2d9ba-0a18-11e6-8112-32c0516b0844.png) /cc @duckduckgo/duckduckhack-contributors Would any of you find this useful? --- IA Page: http://duck.co/ia/view/calculator [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @GuiltyDolphin
main
calculator support context sensitive constants it might be useful to have constants that can vary based on a specified context for example the query c doesn t yield a result as there are undefined constants but if you did c astrophysics we could yield a result informing the user that the calculator interpreted g as the gravitational constant and c as the speed of light in a vacuum so as a quick mock up cc duckduckgo duckduckhack contributors would any of you find this useful ia page guiltydolphin
1
256,316
19,406,772,919
IssuesEvent
2021-12-20 02:36:05
timescale/docs
https://api.github.com/repos/timescale/docs
opened
[Docs RFC]
documentation enhancement
# Describe change in content, appearance, or functionality We will be officially releasing a new feature for Timescale cloud called multi-user. Terms Defined: **Project** = The collection of all the services in your account. **Owner** = The owner of the project **User** = Members that are invited to the owner's project This feature allows you to invite other members to to join and collaborate on all the services in your account. The current roles that can be assigned to invited members are "Owner/ Admin" and "User". The only difference between the two roles are that an "User" cannot delete the "Owner/ Admin" from a project. The new functionalities to Timescale cloud's platform are as follows: **A Project Switcher** = Allows you to switch between the projects in your account **Members tab on the left sidebar** = This shows you all the members tied to the project you are currently in/ allows you to invite other members to the project **Invitations tab under your profile icon** = This shows you all your current invitations and you have the ability to accept/ delete them **Invitation emails** = Once someone has invited you to their project, you will receive an email asking you to accept the invitation **Confirmation emails** = Once someone accepts your invitation to join your project, you will receive a confirmation email # Subject matter expert (SME) Brandon (Product), Kim (Backend), Monae (Frontend), Cami (Design) # Deadline Product is to be released in January but we need the docs for the invitation emails. # Any further info The feature is currently on dev under "multi-user" feature flag. It is not connected to the backend yet so only the base functionalities are available. You can't invite other users to your project until the backend connection is complete. You can switch between projects, "accept" and "reject" invitations, and change project name. [Miro](https://miro.com/app/board/o9J_luT1OPI=/) [Figma](https://www.figma.com/file/7l3jQH86XLfDui9szYjqbw/02.02-Cloud-1.0?node-id=839%3A21204) [Design Doc](https://docs.google.com/document/d/1PmJSjTgRrNHVAS6mgydaR16VvU-4cxUqjyI5I1RW4_U/edit?usp=sharing)
1.0
[Docs RFC] - # Describe change in content, appearance, or functionality We will be officially releasing a new feature for Timescale cloud called multi-user. Terms Defined: **Project** = The collection of all the services in your account. **Owner** = The owner of the project **User** = Members that are invited to the owner's project This feature allows you to invite other members to to join and collaborate on all the services in your account. The current roles that can be assigned to invited members are "Owner/ Admin" and "User". The only difference between the two roles are that an "User" cannot delete the "Owner/ Admin" from a project. The new functionalities to Timescale cloud's platform are as follows: **A Project Switcher** = Allows you to switch between the projects in your account **Members tab on the left sidebar** = This shows you all the members tied to the project you are currently in/ allows you to invite other members to the project **Invitations tab under your profile icon** = This shows you all your current invitations and you have the ability to accept/ delete them **Invitation emails** = Once someone has invited you to their project, you will receive an email asking you to accept the invitation **Confirmation emails** = Once someone accepts your invitation to join your project, you will receive a confirmation email # Subject matter expert (SME) Brandon (Product), Kim (Backend), Monae (Frontend), Cami (Design) # Deadline Product is to be released in January but we need the docs for the invitation emails. # Any further info The feature is currently on dev under "multi-user" feature flag. It is not connected to the backend yet so only the base functionalities are available. You can't invite other users to your project until the backend connection is complete. You can switch between projects, "accept" and "reject" invitations, and change project name. [Miro](https://miro.com/app/board/o9J_luT1OPI=/) [Figma](https://www.figma.com/file/7l3jQH86XLfDui9szYjqbw/02.02-Cloud-1.0?node-id=839%3A21204) [Design Doc](https://docs.google.com/document/d/1PmJSjTgRrNHVAS6mgydaR16VvU-4cxUqjyI5I1RW4_U/edit?usp=sharing)
non_main
describe change in content appearance or functionality we will be officially releasing a new feature for timescale cloud called multi user terms defined project the collection of all the services in your account owner the owner of the project user members that are invited to the owner s project this feature allows you to invite other members to to join and collaborate on all the services in your account the current roles that can be assigned to invited members are owner admin and user the only difference between the two roles are that an user cannot delete the owner admin from a project the new functionalities to timescale cloud s platform are as follows a project switcher allows you to switch between the projects in your account members tab on the left sidebar this shows you all the members tied to the project you are currently in allows you to invite other members to the project invitations tab under your profile icon this shows you all your current invitations and you have the ability to accept delete them invitation emails once someone has invited you to their project you will receive an email asking you to accept the invitation confirmation emails once someone accepts your invitation to join your project you will receive a confirmation email subject matter expert sme brandon product kim backend monae frontend cami design deadline product is to be released in january but we need the docs for the invitation emails any further info the feature is currently on dev under multi user feature flag it is not connected to the backend yet so only the base functionalities are available you can t invite other users to your project until the backend connection is complete you can switch between projects accept and reject invitations and change project name
0
5,302
26,779,892,881
IssuesEvent
2023-01-31 20:13:04
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Move all Spring specific integration code to the distribution and QA
kind/toil scope/broker scope/gateway area/maintainability
**Description** Any Spring specific code should go in the distribution module for production code, and in the QA module for tests as well. This means removing Spring extensions and the likes from the `util` module and the likes. The motivation here is to keep our integration with Spring as light as possible, and keep it as our configuration layer and management server only. It should be used to startup the application, but afterwards should not be referred to internally. As the distribution code grows over time, it might mean we will need to split the gateway and broker artifacts, but I think that's something we probably need to do eventually anyway. This will also avoid things like the client module pulling in `util`, which then pulls in Spring, something that is more than unexpected when pulling in our client.
True
Move all Spring specific integration code to the distribution and QA - **Description** Any Spring specific code should go in the distribution module for production code, and in the QA module for tests as well. This means removing Spring extensions and the likes from the `util` module and the likes. The motivation here is to keep our integration with Spring as light as possible, and keep it as our configuration layer and management server only. It should be used to startup the application, but afterwards should not be referred to internally. As the distribution code grows over time, it might mean we will need to split the gateway and broker artifacts, but I think that's something we probably need to do eventually anyway. This will also avoid things like the client module pulling in `util`, which then pulls in Spring, something that is more than unexpected when pulling in our client.
main
move all spring specific integration code to the distribution and qa description any spring specific code should go in the distribution module for production code and in the qa module for tests as well this means removing spring extensions and the likes from the util module and the likes the motivation here is to keep our integration with spring as light as possible and keep it as our configuration layer and management server only it should be used to startup the application but afterwards should not be referred to internally as the distribution code grows over time it might mean we will need to split the gateway and broker artifacts but i think that s something we probably need to do eventually anyway this will also avoid things like the client module pulling in util which then pulls in spring something that is more than unexpected when pulling in our client
1
64,353
14,662,053,065
IssuesEvent
2020-12-29 06:05:15
tamirverthim/NodeGoat
https://api.github.com/repos/tamirverthim/NodeGoat
opened
WS-2018-0069 (High) detected in is-my-json-valid-2.15.0.tgz
security vulnerability
## WS-2018-0069 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/3de6c5862c1fef83d38a1fec17b579f1a5e328fb">3de6c5862c1fef83d38a1fec17b579f1a5e328fb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function. <p>Publish Date: 2018-02-14 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p> <p>Release Date: 2018-01-24</p> <p>Fix Resolution: 1.4.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1"}],"vulnerabilityIdentifier":"WS-2018-0069","vulnerabilityDetails":"Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2018-0069 (High) detected in is-my-json-valid-2.15.0.tgz - ## WS-2018-0069 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>is-my-json-valid-2.15.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tamirverthim/NodeGoat/commit/3de6c5862c1fef83d38a1fec17b579f1a5e328fb">3de6c5862c1fef83d38a1fec17b579f1a5e328fb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function. <p>Publish Date: 2018-02-14 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976>WS-2018-0069</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/572">https://nodesecurity.io/advisories/572</a></p> <p>Release Date: 2018-01-24</p> <p>Fix Resolution: 1.4.1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1"}],"vulnerabilityIdentifier":"WS-2018-0069","vulnerabilityDetails":"Version of is-my-json-valid before 1.4.1 or 2.17.2 are vulnerable to regular expression denial of service (ReDoS) via the email validation function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
ws high detected in is my json valid tgz ws high severity vulnerability vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules npm node modules request node modules har validator node modules is my json valid package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch master vulnerability details version of is my json valid before or are vulnerable to regular expression denial of service redos via the email validation function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails version of is my json valid before or are vulnerable to regular expression denial of service redos via the email validation function vulnerabilityurl
0
130,963
27,798,999,937
IssuesEvent
2023-03-17 14:33:45
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
[API Proposal]: Add analyzer for Environment.SystemDirectory
api-suggestion area-System.Runtime code-analyzer code-fixer
### Background and motivation `Environment.SystemDirectory` is a lot more efficient than `Environment.GetFolderPath(Environment.SpecialFolder.System)`. We should add an analyzer and fixer for this. See * https://github.com/dotnet/runtime/pull/83564#pullrequestreview-1345143358 * https://github.com/dotnet/aspnetcore/issues/47269 **Category**: [Performance](https://learn.microsoft.com/en-us/dotnet/fundamentals/code-analysis/quality-rules/performance-warnings) **Severity = Info** ### API Proposal N/A ### API Usage ```diff - Environment.GetFolderPath(Environment.SpecialFolder.System) + Environment.SystemDirectory ``` ### Alternative Designs _No response_ ### Risks _No response_
2.0
[API Proposal]: Add analyzer for Environment.SystemDirectory - ### Background and motivation `Environment.SystemDirectory` is a lot more efficient than `Environment.GetFolderPath(Environment.SpecialFolder.System)`. We should add an analyzer and fixer for this. See * https://github.com/dotnet/runtime/pull/83564#pullrequestreview-1345143358 * https://github.com/dotnet/aspnetcore/issues/47269 **Category**: [Performance](https://learn.microsoft.com/en-us/dotnet/fundamentals/code-analysis/quality-rules/performance-warnings) **Severity = Info** ### API Proposal N/A ### API Usage ```diff - Environment.GetFolderPath(Environment.SpecialFolder.System) + Environment.SystemDirectory ``` ### Alternative Designs _No response_ ### Risks _No response_
non_main
add analyzer for environment systemdirectory background and motivation environment systemdirectory is a lot more efficient than environment getfolderpath environment specialfolder system we should add an analyzer and fixer for this see category severity info api proposal n a api usage diff environment getfolderpath environment specialfolder system environment systemdirectory alternative designs no response risks no response
0
5,418
27,207,115,197
IssuesEvent
2023-02-20 13:55:33
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Joinable tables endpoint throws 500 when requested by a user with viewer access to schema
type: bug work: backend status: ready restricted: maintainers
## Description * Create a user with viewer access to a schema. * Make a request to the `/api/db/v0/tables/<table_id>/joinable_tables/` endpoint. * Notice a 500. * * Note: This happens only for a user with viewer access. For editor and manager access level users, we get a 200. ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/249/joinable_tables/ Django Version: 3.1.14 Python Version: 3.9.16 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', 'mathesar.middleware.CursorClosedHandlerMiddleware', 'mathesar.middleware.PasswordChangeNeededMiddleware', 'django_userforeignkey.middleware.UserForeignKeyMiddleware', 'django_request_cache.middleware.RequestCacheMiddleware'] Traceback (most recent call last): File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/tables.py", line 85, in joinable_tables table = self.get_object() File "/usr/local/lib/python3.9/site-packages/rest_framework/generics.py", line 96, in get_object obj = get_object_or_404(queryset, **filter_kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/generics.py", line 19, in get_object_or_404 return _get_object_or_404(queryset, *filter_args, **filter_kwargs) File "/usr/local/lib/python3.9/site-packages/django/shortcuts.py", line 76, in get_object_or_404 return queryset.get(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/django/db/models/query.py", line 433, in get raise self.model.MultipleObjectsReturned( Exception Type: MultipleObjectsReturned at /api/db/v0/tables/249/joinable_tables/ Exception Value: get() returned more than one Table -- it returned 2! ```
True
Joinable tables endpoint throws 500 when requested by a user with viewer access to schema - ## Description * Create a user with viewer access to a schema. * Make a request to the `/api/db/v0/tables/<table_id>/joinable_tables/` endpoint. * Notice a 500. * * Note: This happens only for a user with viewer access. For editor and manager access level users, we get a 200. ``` Environment: Request Method: GET Request URL: http://localhost:8000/api/db/v0/tables/249/joinable_tables/ Django Version: 3.1.14 Python Version: 3.9.16 Installed Applications: ['django.contrib.admin', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.messages', 'django.contrib.staticfiles', 'rest_framework', 'django_filters', 'django_property_filter', 'mathesar'] Installed Middleware: ['django.middleware.security.SecurityMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.csrf.CsrfViewMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'django.middleware.clickjacking.XFrameOptionsMiddleware', 'mathesar.middleware.CursorClosedHandlerMiddleware', 'mathesar.middleware.PasswordChangeNeededMiddleware', 'django_userforeignkey.middleware.UserForeignKeyMiddleware', 'django_request_cache.middleware.RequestCacheMiddleware'] Traceback (most recent call last): File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner response = get_response(request) File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view return view_func(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view return self.dispatch(request, *args, **kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch response = self.handle_exception(exc) File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception response = exception_handler(exc, context) File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler raise exc File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch response = handler(request, *args, **kwargs) File "/code/mathesar/api/db/viewsets/tables.py", line 85, in joinable_tables table = self.get_object() File "/usr/local/lib/python3.9/site-packages/rest_framework/generics.py", line 96, in get_object obj = get_object_or_404(queryset, **filter_kwargs) File "/usr/local/lib/python3.9/site-packages/rest_framework/generics.py", line 19, in get_object_or_404 return _get_object_or_404(queryset, *filter_args, **filter_kwargs) File "/usr/local/lib/python3.9/site-packages/django/shortcuts.py", line 76, in get_object_or_404 return queryset.get(*args, **kwargs) File "/usr/local/lib/python3.9/site-packages/django/db/models/query.py", line 433, in get raise self.model.MultipleObjectsReturned( Exception Type: MultipleObjectsReturned at /api/db/v0/tables/249/joinable_tables/ Exception Value: get() returned more than one Table -- it returned 2! ```
main
joinable tables endpoint throws when requested by a user with viewer access to schema description create a user with viewer access to a schema make a request to the api db tables joinable tables endpoint notice a note this happens only for a user with viewer access for editor and manager access level users we get a environment request method get request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware mathesar middleware cursorclosedhandlermiddleware mathesar middleware passwordchangeneededmiddleware django userforeignkey middleware userforeignkeymiddleware django request cache middleware requestcachemiddleware traceback most recent call last file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file code mathesar api db viewsets tables py line in joinable tables table self get object file usr local lib site packages rest framework generics py line in get object obj get object or queryset filter kwargs file usr local lib site packages rest framework generics py line in get object or return get object or queryset filter args filter kwargs file usr local lib site packages django shortcuts py line in get object or return queryset get args kwargs file usr local lib site packages django db models query py line in get raise self model multipleobjectsreturned exception type multipleobjectsreturned at api db tables joinable tables exception value get returned more than one table it returned
1