Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
5,372
26,997,744,683
IssuesEvent
2023-02-10 03:36:36
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
ClassNotFoundException in package that works fine on Lambda
stage/needs-investigation maintainer/need-followup
I have a Lambda + API Gateway deployment working fine, but when I try to invoke the same ZIP file locally I get a `ClassNotFoundException` for the same handler string as the one in the Lambda Console. I am invoking it with: ``` >sam local start-api --template sam.yaml ``` My sam.yaml is: ``` AWSTemplateFormatVersion: '2010-09-09' Transform: AWS::Serverless-2016-10-31 Description: Users API Endpoint Globals: Api: EndpointConfiguration: REGIONAL Resources: UsersFunction: Type: AWS::Serverless::Function Properties: Handler: com.myproject.users.StreamLambdaHandler::handleRequest Runtime: java8 CodeUri: lambda/build/distributions/lambda-1.0.0-SNAPSHOT MemorySize: 320 Policies: AWSLambdaVPCAccessExecutionRole Timeout: 20 Events: GetResource: Type: Api Properties: Path: /users/{proxy+} Method: any Outputs: UsersApi: Description: URL for application Value: !Sub 'https://${ServerlessRestApi}.execute-api.${AWS::Region}.amazonaws.com/Prod/users' Export: Name: UsersApi ``` 1. OS: Windows 2. `sam --version`: 0.17.0
True
ClassNotFoundException in package that works fine on Lambda - I have a Lambda + API Gateway deployment working fine, but when I try to invoke the same ZIP file locally I get a `ClassNotFoundException` for the same handler string as the one in the Lambda Console. I am invoking it with: ``` >sam local start-api --template sam.yaml ``` My sam.yaml is: ``` AWSTemplateFormatVersion: '2010-09-09' Transform: AWS::Serverless-2016-10-31 Description: Users API Endpoint Globals: Api: EndpointConfiguration: REGIONAL Resources: UsersFunction: Type: AWS::Serverless::Function Properties: Handler: com.myproject.users.StreamLambdaHandler::handleRequest Runtime: java8 CodeUri: lambda/build/distributions/lambda-1.0.0-SNAPSHOT MemorySize: 320 Policies: AWSLambdaVPCAccessExecutionRole Timeout: 20 Events: GetResource: Type: Api Properties: Path: /users/{proxy+} Method: any Outputs: UsersApi: Description: URL for application Value: !Sub 'https://${ServerlessRestApi}.execute-api.${AWS::Region}.amazonaws.com/Prod/users' Export: Name: UsersApi ``` 1. OS: Windows 2. `sam --version`: 0.17.0
main
classnotfoundexception in package that works fine on lambda i have a lambda api gateway deployment working fine but when i try to invoke the same zip file locally i get a classnotfoundexception for the same handler string as the one in the lambda console i am invoking it with sam local start api template sam yaml my sam yaml is awstemplateformatversion transform aws serverless description users api endpoint globals api endpointconfiguration regional resources usersfunction type aws serverless function properties handler com myproject users streamlambdahandler handlerequest runtime codeuri lambda build distributions lambda snapshot memorysize policies awslambdavpcaccessexecutionrole timeout events getresource type api properties path users proxy method any outputs usersapi description url for application value sub export name usersapi os windows sam version
1
84,747
7,931,837,822
IssuesEvent
2018-07-07 06:06:05
PlasmaPy/PlasmaPy
https://api.github.com/repos/PlasmaPy/PlasmaPy
closed
Test against Python 3.7
Status: on hold Testing
Python 3.7 is due to be released around 2018 June 15. Once that happens, we'll need to include both Python 3.6 and Python 3.7 in our test suite on Travis CI and AppVeyer. I'm not sure what we'll need to do for CircleCI. We should check out [what's new in Python 3.7](https://docs.python.org/3.7/whatsnew/3.7.html) when we do this too, since any removed features that cause errors should be described there. If our most recent `0.1.*` release doesn't work on Python 3.7, then we'll need to do a patch fix release quickly. Cross-reference: #264
1.0
Test against Python 3.7 - Python 3.7 is due to be released around 2018 June 15. Once that happens, we'll need to include both Python 3.6 and Python 3.7 in our test suite on Travis CI and AppVeyer. I'm not sure what we'll need to do for CircleCI. We should check out [what's new in Python 3.7](https://docs.python.org/3.7/whatsnew/3.7.html) when we do this too, since any removed features that cause errors should be described there. If our most recent `0.1.*` release doesn't work on Python 3.7, then we'll need to do a patch fix release quickly. Cross-reference: #264
non_main
test against python python is due to be released around june once that happens we ll need to include both python and python in our test suite on travis ci and appveyer i m not sure what we ll need to do for circleci we should check out when we do this too since any removed features that cause errors should be described there if our most recent release doesn t work on python then we ll need to do a patch fix release quickly cross reference
0
3,089
11,741,740,244
IssuesEvent
2020-03-11 22:32:01
alacritty/alacritty
https://api.github.com/repos/alacritty/alacritty
closed
Wrong mouse cursor size on Wayland
A - deps B - bug C - waiting on maintainer DS - Wayland
Mouse cursor is small, ~20px. Looks like system cursor settings are not respected. Have HiDPI display. Sway scale setting is default (=1). ```shell GDK_DPI_SCALE=1.66 XCURSOR_SIZE=40 gsettings set org.gnome.desktop.interface cursor-size 40 ``` All other apps (including XWayland ones) show mouse cursor 40px sized. With `WINIT_UNIX_BACKEND=x11` cursor sized properly. **OS:** Arch Linux, kernel 5.4.15 **Version:** 0.4.1 **Compositor:** sway <details> <summary>alacritty -vv</summary> [2020-01-30 16:14] [INFO] Welcome to Alacritty<br> [2020-01-30 16:14] [INFO] Configuration loaded from "/home/user/.config/alacritty/alacritty.yml"<br> [2020-01-30 16:14] [DEBUG] Estimated DPR: 1<br> [2020-01-30 16:14] [DEBUG] Estimated Cell Size: 13 x 29<br> [2020-01-30 16:14] [DEBUG] Estimated Dimensions: None<br> [2020-01-30 16:14] [INFO] Device pixel ratio: 1<br> [2020-01-30 16:14] [INFO] Initializing glyph cache...<br> [2020-01-30 16:14] [INFO] ... finished initializing glyph cache in 0.023989566s<br> [2020-01-30 16:14] [INFO] Cell Size: 13 x 29<br> [2020-01-30 16:14] [INFO] Padding: 0 x 0<br> [2020-01-30 16:14] [INFO] Width: 800, Height: 600<br> [2020-01-30 16:14] [INFO] PTY Dimensions: Line(20) x Column(61)<br> [2020-01-30 16:14] [INFO] Initialisation complete<br> [2020-01-30 16:14] [DEBUG] New num_cols is 73 and num_lines is 36<br> [2020-01-30 16:14] [INFO] Width: 956, Height: 1045 </details>
True
Wrong mouse cursor size on Wayland - Mouse cursor is small, ~20px. Looks like system cursor settings are not respected. Have HiDPI display. Sway scale setting is default (=1). ```shell GDK_DPI_SCALE=1.66 XCURSOR_SIZE=40 gsettings set org.gnome.desktop.interface cursor-size 40 ``` All other apps (including XWayland ones) show mouse cursor 40px sized. With `WINIT_UNIX_BACKEND=x11` cursor sized properly. **OS:** Arch Linux, kernel 5.4.15 **Version:** 0.4.1 **Compositor:** sway <details> <summary>alacritty -vv</summary> [2020-01-30 16:14] [INFO] Welcome to Alacritty<br> [2020-01-30 16:14] [INFO] Configuration loaded from "/home/user/.config/alacritty/alacritty.yml"<br> [2020-01-30 16:14] [DEBUG] Estimated DPR: 1<br> [2020-01-30 16:14] [DEBUG] Estimated Cell Size: 13 x 29<br> [2020-01-30 16:14] [DEBUG] Estimated Dimensions: None<br> [2020-01-30 16:14] [INFO] Device pixel ratio: 1<br> [2020-01-30 16:14] [INFO] Initializing glyph cache...<br> [2020-01-30 16:14] [INFO] ... finished initializing glyph cache in 0.023989566s<br> [2020-01-30 16:14] [INFO] Cell Size: 13 x 29<br> [2020-01-30 16:14] [INFO] Padding: 0 x 0<br> [2020-01-30 16:14] [INFO] Width: 800, Height: 600<br> [2020-01-30 16:14] [INFO] PTY Dimensions: Line(20) x Column(61)<br> [2020-01-30 16:14] [INFO] Initialisation complete<br> [2020-01-30 16:14] [DEBUG] New num_cols is 73 and num_lines is 36<br> [2020-01-30 16:14] [INFO] Width: 956, Height: 1045 </details>
main
wrong mouse cursor size on wayland mouse cursor is small looks like system cursor settings are not respected have hidpi display sway scale setting is default shell gdk dpi scale xcursor size gsettings set org gnome desktop interface cursor size all other apps including xwayland ones show mouse cursor sized with winit unix backend cursor sized properly os arch linux kernel version compositor sway alacritty vv welcome to alacritty configuration loaded from home user config alacritty alacritty yml estimated dpr estimated cell size x estimated dimensions none device pixel ratio initializing glyph cache finished initializing glyph cache in cell size x padding x width height pty dimensions line x column initialisation complete new num cols is and num lines is width height
1
387,810
11,470,664,433
IssuesEvent
2020-02-09 05:12:26
ShabadOS/desktop
https://api.github.com/repos/ShabadOS/desktop
opened
Add customization of display fields for navigator lines
Priority: 2 Medium Scope: Frontend/Controller Status: Confirmed Type: Feature/Enhancement
So a user can see the gurbani, english translation/pronunciation, etc. for each line of the navigator. User request. Personally think it could be useful for live streamers that have a small navigator window in the corner that would also like to follow along. Or for remote control users that have only a phone in an awkward position from the projectors. Related to #225 Not sure if we want one settings tabs to customize all our activities or individual settings tabs per activity. Would be nice if we had headers or such in settings so we could have an "Activities" tab in settings and then different headers for each activity with their respective options inside them.
1.0
Add customization of display fields for navigator lines - So a user can see the gurbani, english translation/pronunciation, etc. for each line of the navigator. User request. Personally think it could be useful for live streamers that have a small navigator window in the corner that would also like to follow along. Or for remote control users that have only a phone in an awkward position from the projectors. Related to #225 Not sure if we want one settings tabs to customize all our activities or individual settings tabs per activity. Would be nice if we had headers or such in settings so we could have an "Activities" tab in settings and then different headers for each activity with their respective options inside them.
non_main
add customization of display fields for navigator lines so a user can see the gurbani english translation pronunciation etc for each line of the navigator user request personally think it could be useful for live streamers that have a small navigator window in the corner that would also like to follow along or for remote control users that have only a phone in an awkward position from the projectors related to not sure if we want one settings tabs to customize all our activities or individual settings tabs per activity would be nice if we had headers or such in settings so we could have an activities tab in settings and then different headers for each activity with their respective options inside them
0
4,008
18,685,468,844
IssuesEvent
2021-11-01 11:52:43
WarenGonzaga/gather.js
https://api.github.com/repos/WarenGonzaga/gather.js
opened
repo banner for GatherJS
maintainers-only todo tweak
I might also need help with graphic design for the repo banner. If you think you can do this let me know by leaving a comment.
True
repo banner for GatherJS - I might also need help with graphic design for the repo banner. If you think you can do this let me know by leaving a comment.
main
repo banner for gatherjs i might also need help with graphic design for the repo banner if you think you can do this let me know by leaving a comment
1
638
4,155,236,297
IssuesEvent
2016-06-16 14:22:18
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
DOTA 2 Heroes: "more at" link not always working
Maintainer Input Requested
When viewing the Queen of Pain IA, the "more at" link in the IA does not link to the directly Queen of Pain overview on dota2.com. It brings you to the general "choose a hero" screen. It links directly to other heros' profiles when viewing their IA and following the "more at" link. Is there anything actionable there? It might just be an issue with dota2.com ------ IA Page: http://duck.co/ia/view/dota2 [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @echosa
True
DOTA 2 Heroes: "more at" link not always working - When viewing the Queen of Pain IA, the "more at" link in the IA does not link to the directly Queen of Pain overview on dota2.com. It brings you to the general "choose a hero" screen. It links directly to other heros' profiles when viewing their IA and following the "more at" link. Is there anything actionable there? It might just be an issue with dota2.com ------ IA Page: http://duck.co/ia/view/dota2 [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @echosa
main
dota heroes more at link not always working when viewing the queen of pain ia the more at link in the ia does not link to the directly queen of pain overview on com it brings you to the general choose a hero screen it links directly to other heros profiles when viewing their ia and following the more at link is there anything actionable there it might just be an issue with com ia page echosa
1
58
2,603,062,071
IssuesEvent
2015-02-24 13:54:39
mranney/node_pcap
https://api.github.com/repos/mranney/node_pcap
closed
Remove examples.
maintainance
A fair number of the bug reports this project gets are about the examples not the lib. Let's remove all examples from this repo and put them in their own repos. Examples of this already being done: https://github.com/mranney/redis_trace https://github.com/mranney/http_trace
True
Remove examples. - A fair number of the bug reports this project gets are about the examples not the lib. Let's remove all examples from this repo and put them in their own repos. Examples of this already being done: https://github.com/mranney/redis_trace https://github.com/mranney/http_trace
main
remove examples a fair number of the bug reports this project gets are about the examples not the lib let s remove all examples from this repo and put them in their own repos examples of this already being done
1
4,240
21,000,078,884
IssuesEvent
2022-03-29 16:35:07
aws/serverless-application-model
https://api.github.com/repos/aws/serverless-application-model
closed
Proposal: Edge (boolean) on Function
type/feature maintainer/need-response
**Description:** If you want to use `AWS::Serverless::Function` for an Lambda@edge, you need to configure a special `AssumeRolePolicyDocument` and an `AutoPublishAlias`. It would be more hany if you just say `Edge: true` on the lambda. **Current example:** ```yaml DownloadRequestTokenFunctionProxy: Type: AWS::Serverless::Function Properties: Handler: index.handler Runtime: nodejs12.x AutoPublishAlias: edge AssumeRolePolicyDocument: Version: 2012-10-17 Statement: - Effect: Allow Principal: Service: - lambda.amazonaws.com - edgelambda.amazonaws.com Action: - sts:AssumeRole CodeUri: . ``` **Future example:** ```yaml DownloadRequestTokenFunctionProxy: Type: AWS::Serverless::Function Properties: Handler: index.handler Runtime: nodejs12.x Edge: true CodeUri: . ``` **Expected result:** Both templates delivering the same result. `Edge: true` builds an `AssumeRolePolicyDocument` for `lambda` / `edgelambda` and sets `AutoPublishAlias` to `edge`. **More possible use cases:** SAM CLI cloud add validation based on it to warn for the use of layers, runtimes not supported by edge, to hight timeouts (> 30 sec)
True
Proposal: Edge (boolean) on Function - **Description:** If you want to use `AWS::Serverless::Function` for an Lambda@edge, you need to configure a special `AssumeRolePolicyDocument` and an `AutoPublishAlias`. It would be more hany if you just say `Edge: true` on the lambda. **Current example:** ```yaml DownloadRequestTokenFunctionProxy: Type: AWS::Serverless::Function Properties: Handler: index.handler Runtime: nodejs12.x AutoPublishAlias: edge AssumeRolePolicyDocument: Version: 2012-10-17 Statement: - Effect: Allow Principal: Service: - lambda.amazonaws.com - edgelambda.amazonaws.com Action: - sts:AssumeRole CodeUri: . ``` **Future example:** ```yaml DownloadRequestTokenFunctionProxy: Type: AWS::Serverless::Function Properties: Handler: index.handler Runtime: nodejs12.x Edge: true CodeUri: . ``` **Expected result:** Both templates delivering the same result. `Edge: true` builds an `AssumeRolePolicyDocument` for `lambda` / `edgelambda` and sets `AutoPublishAlias` to `edge`. **More possible use cases:** SAM CLI cloud add validation based on it to warn for the use of layers, runtimes not supported by edge, to hight timeouts (> 30 sec)
main
proposal edge boolean on function description if you want to use aws serverless function for an lambda edge you need to configure a special assumerolepolicydocument and an autopublishalias it would be more hany if you just say edge true on the lambda current example yaml downloadrequesttokenfunctionproxy type aws serverless function properties handler index handler runtime x autopublishalias edge assumerolepolicydocument version statement effect allow principal service lambda amazonaws com edgelambda amazonaws com action sts assumerole codeuri future example yaml downloadrequesttokenfunctionproxy type aws serverless function properties handler index handler runtime x edge true codeuri expected result both templates delivering the same result edge true builds an assumerolepolicydocument for lambda edgelambda and sets autopublishalias to edge more possible use cases sam cli cloud add validation based on it to warn for the use of layers runtimes not supported by edge to hight timeouts sec
1
334,283
24,412,089,357
IssuesEvent
2022-10-05 13:12:52
nanoporetech/dorado
https://api.github.com/repos/nanoporetech/dorado
closed
Building on mac osx
documentation
I'm going to document what I ended up installing on my m1 macbook pro max to compile dorado: ## 0. Packages installed using brew - cmake - openssl - hdf5 ## 1. Xcode installation and profile switch for being able to use metal You need to download and install Xcode, i would recommend downloading it directly from https://developer.apple.com/download/all/ instead of installing it from the App store as the latter is relatively slow (max 2.4 Mb/sec download) Afterwards, you need to switch `sudo xcode-select --switch /Applications/Xcode.app/Contents/Developer`
1.0
Building on mac osx - I'm going to document what I ended up installing on my m1 macbook pro max to compile dorado: ## 0. Packages installed using brew - cmake - openssl - hdf5 ## 1. Xcode installation and profile switch for being able to use metal You need to download and install Xcode, i would recommend downloading it directly from https://developer.apple.com/download/all/ instead of installing it from the App store as the latter is relatively slow (max 2.4 Mb/sec download) Afterwards, you need to switch `sudo xcode-select --switch /Applications/Xcode.app/Contents/Developer`
non_main
building on mac osx i m going to document what i ended up installing on my macbook pro max to compile dorado packages installed using brew cmake openssl xcode installation and profile switch for being able to use metal you need to download and install xcode i would recommend downloading it directly from instead of installing it from the app store as the latter is relatively slow max mb sec download afterwards you need to switch sudo xcode select switch applications xcode app contents developer
0
976
4,728,066,189
IssuesEvent
2016-10-18 15:04:16
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Homebrew/Homebrew Cask deduplication and naming guidelines
awaiting maintainer feedback discussion
This issue has cropped up in a few places before, and seems to still be unresolved. There are a few apps that have both a Formula and a Cask, and in the interest of deduplication, there's an ongoing effort to remove one in favor of the other. However, it is my strong opinion that favoring one over the other is incorrect, and there are legitimate use cases for having both a Formula, and a Cask for the same application. These include: - Preference of vendor-provided builds/binary - Ability to use pre-compiled builds (instead of building from source) - Differing functionality in the two (ie. one includes a GUI) Therefore, this issue aims to discuss and implement a formal naming scheme for such cases, so that there does not need to be the same discussion every time a user brings up a previously removed Cask for re-inclusion. A starting point for naming: "If a application is not identical or has additional functionality as compared to the homebrew Formula, the Cask should be named with a `-app` or `-gui` suffix." (the token reference would be changed accordingly) Examples/relevant issues: https://github.com/caskroom/homebrew-cask/pull/24232 https://github.com/caskroom/homebrew-cask/pull/23584 https://github.com/caskroom/homebrew-cask/pull/23973 I'd like to specifically invite @vitorgalvao and @MikeMcQuaid into this discussion, as they have been active in this regard, but also other users, if they have thoughts one way or another.
True
Homebrew/Homebrew Cask deduplication and naming guidelines - This issue has cropped up in a few places before, and seems to still be unresolved. There are a few apps that have both a Formula and a Cask, and in the interest of deduplication, there's an ongoing effort to remove one in favor of the other. However, it is my strong opinion that favoring one over the other is incorrect, and there are legitimate use cases for having both a Formula, and a Cask for the same application. These include: - Preference of vendor-provided builds/binary - Ability to use pre-compiled builds (instead of building from source) - Differing functionality in the two (ie. one includes a GUI) Therefore, this issue aims to discuss and implement a formal naming scheme for such cases, so that there does not need to be the same discussion every time a user brings up a previously removed Cask for re-inclusion. A starting point for naming: "If a application is not identical or has additional functionality as compared to the homebrew Formula, the Cask should be named with a `-app` or `-gui` suffix." (the token reference would be changed accordingly) Examples/relevant issues: https://github.com/caskroom/homebrew-cask/pull/24232 https://github.com/caskroom/homebrew-cask/pull/23584 https://github.com/caskroom/homebrew-cask/pull/23973 I'd like to specifically invite @vitorgalvao and @MikeMcQuaid into this discussion, as they have been active in this regard, but also other users, if they have thoughts one way or another.
main
homebrew homebrew cask deduplication and naming guidelines this issue has cropped up in a few places before and seems to still be unresolved there are a few apps that have both a formula and a cask and in the interest of deduplication there s an ongoing effort to remove one in favor of the other however it is my strong opinion that favoring one over the other is incorrect and there are legitimate use cases for having both a formula and a cask for the same application these include preference of vendor provided builds binary ability to use pre compiled builds instead of building from source differing functionality in the two ie one includes a gui therefore this issue aims to discuss and implement a formal naming scheme for such cases so that there does not need to be the same discussion every time a user brings up a previously removed cask for re inclusion a starting point for naming if a application is not identical or has additional functionality as compared to the homebrew formula the cask should be named with a app or gui suffix the token reference would be changed accordingly examples relevant issues i d like to specifically invite vitorgalvao and mikemcquaid into this discussion as they have been active in this regard but also other users if they have thoughts one way or another
1
4,592
23,829,466,846
IssuesEvent
2022-09-05 18:31:20
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFP - ludusavi-x64
Status: Available For Maintainer(s)
Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL :https://github.com/mtkennerly/ludusavi Direct download URL for the software / installer :https://github.com/mtkennerly/ludusavi/releases/download/v0.12.1/ludusavi-v0.12.1-win64.zip Software summary / short description: Ludusavi is a tool for backing up your PC video game save data, written in Rust. It is cross-platform and supports multiple game stores.
True
RFP - ludusavi-x64 - Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL :https://github.com/mtkennerly/ludusavi Direct download URL for the software / installer :https://github.com/mtkennerly/ludusavi/releases/download/v0.12.1/ludusavi-v0.12.1-win64.zip Software summary / short description: Ludusavi is a tool for backing up your PC video game save data, written in Rust. It is cross-platform and supports multiple game stores.
main
rfp ludusavi checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts with rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description ludusavi is a tool for backing up your pc video game save data written in rust it is cross platform and supports multiple game stores
1
665,772
22,329,200,042
IssuesEvent
2022-06-14 13:17:27
yl-ang/NsStayFit
https://api.github.com/repos/yl-ang/NsStayFit
closed
As a NSMen, I want to keep track of my past IPPT results
type.Story priority.Low IPPT
so that I know whether I improve over time.
1.0
As a NSMen, I want to keep track of my past IPPT results - so that I know whether I improve over time.
non_main
as a nsmen i want to keep track of my past ippt results so that i know whether i improve over time
0
435
3,549,641,077
IssuesEvent
2016-01-20 18:49:02
pypiserver/pypiserver
https://api.github.com/repos/pypiserver/pypiserver
opened
Stop supporting legacy Python versions
Maintainance
The plan is to limit compatibility of the following:. python-2 < 2.7 python-3 < 3.3 One purpose is to demotivate the use of really old python releases. Another is to allow using newer syntax when programming it. For older python-2 version, use version `1.1.x` series.
True
Stop supporting legacy Python versions - The plan is to limit compatibility of the following:. python-2 < 2.7 python-3 < 3.3 One purpose is to demotivate the use of really old python releases. Another is to allow using newer syntax when programming it. For older python-2 version, use version `1.1.x` series.
main
stop supporting legacy python versions the plan is to limit compatibility of the following python python one purpose is to demotivate the use of really old python releases another is to allow using newer syntax when programming it for older python version use version x series
1
175,135
14,517,670,066
IssuesEvent
2020-12-13 20:32:00
sdelgadoc/download-tweets-ai-text-gen-plus
https://api.github.com/repos/sdelgadoc/download-tweets-ai-text-gen-plus
closed
Need testers for new reply and sentiment functionality
documentation
I added two new pieces of functionality that others might be interested in using. 1. The code will now collect data that lets you train an AI that can not only generate tweets, but also generate replies to existing tweets. 2. Second, the code can also measure the sentiment of tweets, so you can generate positive/neutral/negative tweets programmatically. I've been testing it with my [Data Journalist](https://twitter.com/datajourno_gpt2/with_replies) twitter bot account, and it's been surprisingly good at responding to some tweets. You can find the code with the new functionality in the ***development*** branch of this repo below: [https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus/tree/development](https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus/tree/development) You can clone the repo with the following command: ```sh git clone -b development https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus.git ``` Although the [repo's README](https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus/tree/development) explains what you need to do in detail to use the new functionality, below is a quick summary: - Install two new Python libraries (textblob: for sentiment analysis, tweepy: for access to the Twitter API) - [Create a Twitter App](https://developer.twitter.com/en/docs/basics/apps/overview) to gain access to the Twitter API, which will be used to traverse reply trees because the twint library doesn't support the functionality - Input your Twitter API credentials into the new `keys.py` file You can then run the script with the command below to generate text that can train an AI to reply, and includes sentiment information: ```sh python3 download_tweets.py <twitter_username> None True False False False 3 reply ``` Let me know if you have any questions. Below I am tagging folks that have used this repo, or that I have helped with issues in the past, in case they are interested: @arceusisafish. @DeFiDude, @wradd, @RalphORama, @minimaxir, @YazzyYaz, @antherknee, @dmccaffrey12, @zhermi, @rewinded, @luisgabriel53, @faizankshaikh
1.0
Need testers for new reply and sentiment functionality - I added two new pieces of functionality that others might be interested in using. 1. The code will now collect data that lets you train an AI that can not only generate tweets, but also generate replies to existing tweets. 2. Second, the code can also measure the sentiment of tweets, so you can generate positive/neutral/negative tweets programmatically. I've been testing it with my [Data Journalist](https://twitter.com/datajourno_gpt2/with_replies) twitter bot account, and it's been surprisingly good at responding to some tweets. You can find the code with the new functionality in the ***development*** branch of this repo below: [https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus/tree/development](https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus/tree/development) You can clone the repo with the following command: ```sh git clone -b development https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus.git ``` Although the [repo's README](https://github.com/sdelgadoc/download-tweets-ai-text-gen-plus/tree/development) explains what you need to do in detail to use the new functionality, below is a quick summary: - Install two new Python libraries (textblob: for sentiment analysis, tweepy: for access to the Twitter API) - [Create a Twitter App](https://developer.twitter.com/en/docs/basics/apps/overview) to gain access to the Twitter API, which will be used to traverse reply trees because the twint library doesn't support the functionality - Input your Twitter API credentials into the new `keys.py` file You can then run the script with the command below to generate text that can train an AI to reply, and includes sentiment information: ```sh python3 download_tweets.py <twitter_username> None True False False False 3 reply ``` Let me know if you have any questions. Below I am tagging folks that have used this repo, or that I have helped with issues in the past, in case they are interested: @arceusisafish. @DeFiDude, @wradd, @RalphORama, @minimaxir, @YazzyYaz, @antherknee, @dmccaffrey12, @zhermi, @rewinded, @luisgabriel53, @faizankshaikh
non_main
need testers for new reply and sentiment functionality i added two new pieces of functionality that others might be interested in using the code will now collect data that lets you train an ai that can not only generate tweets but also generate replies to existing tweets second the code can also measure the sentiment of tweets so you can generate positive neutral negative tweets programmatically i ve been testing it with my twitter bot account and it s been surprisingly good at responding to some tweets you can find the code with the new functionality in the development branch of this repo below you can clone the repo with the following command sh git clone b development although the explains what you need to do in detail to use the new functionality below is a quick summary install two new python libraries textblob for sentiment analysis tweepy for access to the twitter api to gain access to the twitter api which will be used to traverse reply trees because the twint library doesn t support the functionality input your twitter api credentials into the new keys py file you can then run the script with the command below to generate text that can train an ai to reply and includes sentiment information sh download tweets py none true false false false reply let me know if you have any questions below i am tagging folks that have used this repo or that i have helped with issues in the past in case they are interested arceusisafish defidude wradd ralphorama minimaxir yazzyyaz antherknee zhermi rewinded faizankshaikh
0
2,905
10,326,168,722
IssuesEvent
2019-09-02 00:05:21
vostpt/mobile-app
https://api.github.com/repos/vostpt/mobile-app
closed
Flare Splash Screen
Needs Maintainers Help
**Description** Create Flutter Splash Screen **File Location** ``` - presentation |__ ui ``` **Requirements** - "This screen will show an animation with the app logo - Internally, this screen will check in the shared preferences if the user has opened the app previously - If first install, go to tutorial - Else, go to main app screen" **UI** <img width="364" alt="imagem" src="https://user-images.githubusercontent.com/10728633/63039860-1fe8f280-bebc-11e9-911f-6a48d9bf219d.png">
True
Flare Splash Screen - **Description** Create Flutter Splash Screen **File Location** ``` - presentation |__ ui ``` **Requirements** - "This screen will show an animation with the app logo - Internally, this screen will check in the shared preferences if the user has opened the app previously - If first install, go to tutorial - Else, go to main app screen" **UI** <img width="364" alt="imagem" src="https://user-images.githubusercontent.com/10728633/63039860-1fe8f280-bebc-11e9-911f-6a48d9bf219d.png">
main
flare splash screen description create flutter splash screen file location presentation ui requirements this screen will show an animation with the app logo internally this screen will check in the shared preferences if the user has opened the app previously if first install go to tutorial else go to main app screen ui img width alt imagem src
1
221,654
17,024,967,250
IssuesEvent
2021-07-03 09:45:29
Psionica/dual-server
https://api.github.com/repos/Psionica/dual-server
opened
Set up backend documentation
documentation
After the build tests are in place (#3), the backend API will have hopefully settled. A static web page for the docs should be set using GitHub Pages.
1.0
Set up backend documentation - After the build tests are in place (#3), the backend API will have hopefully settled. A static web page for the docs should be set using GitHub Pages.
non_main
set up backend documentation after the build tests are in place the backend api will have hopefully settled a static web page for the docs should be set using github pages
0
3,234
12,368,706,405
IssuesEvent
2020-05-18 14:13:30
Kashdeya/Tiny-Progressions
https://api.github.com/repos/Kashdeya/Tiny-Progressions
closed
Watering Can multiplayer behaviour
Version not Maintainted
When multiple players have watering cans in their inventory cans cannot be activated, and one player sneaking while another right clicks will activate the sneaking players watering can. Have seen this in multiple mod packs including Project Ozone 3 and Sky Factory 4. Running on a LAN World with two players. Both have a watering can. Have done some testing and it seems that when the last player to pick up a watering can tries to activate the can, it activates for a split second, and then deactivates. It appears to happen in less than a frame in some cases as the item flickers with the enchanted 'glow'. If one player activates their watering can while the other is holding it in their hand, both watering cans are activated. If the other player shift right clicks both cans are deactivated. It seems as though there is some sort of global state for the cans that isn't working with multiple players.
True
Watering Can multiplayer behaviour - When multiple players have watering cans in their inventory cans cannot be activated, and one player sneaking while another right clicks will activate the sneaking players watering can. Have seen this in multiple mod packs including Project Ozone 3 and Sky Factory 4. Running on a LAN World with two players. Both have a watering can. Have done some testing and it seems that when the last player to pick up a watering can tries to activate the can, it activates for a split second, and then deactivates. It appears to happen in less than a frame in some cases as the item flickers with the enchanted 'glow'. If one player activates their watering can while the other is holding it in their hand, both watering cans are activated. If the other player shift right clicks both cans are deactivated. It seems as though there is some sort of global state for the cans that isn't working with multiple players.
main
watering can multiplayer behaviour when multiple players have watering cans in their inventory cans cannot be activated and one player sneaking while another right clicks will activate the sneaking players watering can have seen this in multiple mod packs including project ozone and sky factory running on a lan world with two players both have a watering can have done some testing and it seems that when the last player to pick up a watering can tries to activate the can it activates for a split second and then deactivates it appears to happen in less than a frame in some cases as the item flickers with the enchanted glow if one player activates their watering can while the other is holding it in their hand both watering cans are activated if the other player shift right clicks both cans are deactivated it seems as though there is some sort of global state for the cans that isn t working with multiple players
1
430,820
30,201,959,707
IssuesEvent
2023-07-05 06:44:25
duncanamps/box80
https://api.github.com/repos/duncanamps/box80
closed
Add licence conditions to each source file
documentation
Add the GPL V3 header to each source file and place a link on the Help menu to the GNU website.
1.0
Add licence conditions to each source file - Add the GPL V3 header to each source file and place a link on the Help menu to the GNU website.
non_main
add licence conditions to each source file add the gpl header to each source file and place a link on the help menu to the gnu website
0
5,867
31,823,116,701
IssuesEvent
2023-09-14 05:06:45
onebeyond/maintainers
https://api.github.com/repos/onebeyond/maintainers
closed
OpenSSF Scorecard Report Updated!
maintainers-agenda
Hello! There are changes in your OpenSSF Scorecard report. Please review the following changes and take action if necessary. ## Summary There are changes in the following repositories: | Repository | Commit | Score | Score Delta | Report | StepSecurity | | -- | -- | -- | -- | -- | -- | | [onebeyond/license-checker](https://github.com/onebeyond/license-checker) | [aece706](https://github.com/onebeyond/license-checker/commit/aece706c6718cd2278f09fbebb3f536b2657caa9) | 6.4 | -0.3 / [Details](https://kooltheba.github.io/openssf-scorecard-api-visualizer/#/projects/github.com/onebeyond/license-checker/compare/5c68f897f5fd00a13b2ea21ccaa0bf1afcad33ee/aece706c6718cd2278f09fbebb3f536b2657caa9) | [View](https://kooltheba.github.io/openssf-scorecard-api-visualizer/#/projects/github.com/onebeyond/license-checker/commit/aece706c6718cd2278f09fbebb3f536b2657caa9) | [Fix it](https://app.stepsecurity.io/securerepo?repo=onebeyond/license-checker) | _Report generated by [UlisesGascon/openssf-scorecard-monitor](https://github.com/UlisesGascon/openssf-scorecard-monitor)._
True
OpenSSF Scorecard Report Updated! - Hello! There are changes in your OpenSSF Scorecard report. Please review the following changes and take action if necessary. ## Summary There are changes in the following repositories: | Repository | Commit | Score | Score Delta | Report | StepSecurity | | -- | -- | -- | -- | -- | -- | | [onebeyond/license-checker](https://github.com/onebeyond/license-checker) | [aece706](https://github.com/onebeyond/license-checker/commit/aece706c6718cd2278f09fbebb3f536b2657caa9) | 6.4 | -0.3 / [Details](https://kooltheba.github.io/openssf-scorecard-api-visualizer/#/projects/github.com/onebeyond/license-checker/compare/5c68f897f5fd00a13b2ea21ccaa0bf1afcad33ee/aece706c6718cd2278f09fbebb3f536b2657caa9) | [View](https://kooltheba.github.io/openssf-scorecard-api-visualizer/#/projects/github.com/onebeyond/license-checker/commit/aece706c6718cd2278f09fbebb3f536b2657caa9) | [Fix it](https://app.stepsecurity.io/securerepo?repo=onebeyond/license-checker) | _Report generated by [UlisesGascon/openssf-scorecard-monitor](https://github.com/UlisesGascon/openssf-scorecard-monitor)._
main
openssf scorecard report updated hello there are changes in your openssf scorecard report please review the following changes and take action if necessary summary there are changes in the following repositories repository commit score score delta report stepsecurity report generated by
1
345,795
10,373,236,343
IssuesEvent
2019-09-09 06:40:30
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
duckduckgo.com - see bug description
browser-fenix engine-gecko priority-important
<!-- @browser: Firefox Mobile 69.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://duckduckgo.com/?q= **Browser / Version**: Firefox Mobile 69.0 **Operating System**: Android 7.1.2 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Cannot be used as default search engine **Steps to Reproduce**: Cannot be used as default search engine <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
duckduckgo.com - see bug description - <!-- @browser: Firefox Mobile 69.0 --> <!-- @ua_header: Mozilla/5.0 (Android 7.1.2; Mobile; rv:69.0) Gecko/69.0 Firefox/69.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://duckduckgo.com/?q= **Browser / Version**: Firefox Mobile 69.0 **Operating System**: Android 7.1.2 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Cannot be used as default search engine **Steps to Reproduce**: Cannot be used as default search engine <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
duckduckgo com see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description cannot be used as default search engine steps to reproduce cannot be used as default search engine browser configuration none from with ❤️
0
341,532
24,702,903,040
IssuesEvent
2022-10-19 16:35:18
franciellyferreira/design-apis-guide
https://api.github.com/repos/franciellyferreira/design-apis-guide
opened
Melhorar a organização dos arquivos da raiz
documentation
Muitos arquivos na raiz do projeto, necessário organizar melhor e ajustar os links de navegação.
1.0
Melhorar a organização dos arquivos da raiz - Muitos arquivos na raiz do projeto, necessário organizar melhor e ajustar os links de navegação.
non_main
melhorar a organização dos arquivos da raiz muitos arquivos na raiz do projeto necessário organizar melhor e ajustar os links de navegação
0
4,232
20,970,579,709
IssuesEvent
2022-03-28 10:58:54
adda-team/adda
https://api.github.com/repos/adda-team/adda
opened
Enforce formatting
pri-Medium maintainability DevOps
We have description of the code style - https://github.com/adda-team/adda/wiki/CodeStyleGuide. But it is neither comprehensive (covers all cases) nor enforced. It would be great to implement testing for the style in Github Actions - probably some existing standard can be adopted for that. And, naturally, it should be as much close to the existing de-facto style as possible. The first step is to study existing standards for C code, and whether they are readily available in IDEs and automatic tools.
True
Enforce formatting - We have description of the code style - https://github.com/adda-team/adda/wiki/CodeStyleGuide. But it is neither comprehensive (covers all cases) nor enforced. It would be great to implement testing for the style in Github Actions - probably some existing standard can be adopted for that. And, naturally, it should be as much close to the existing de-facto style as possible. The first step is to study existing standards for C code, and whether they are readily available in IDEs and automatic tools.
main
enforce formatting we have description of the code style but it is neither comprehensive covers all cases nor enforced it would be great to implement testing for the style in github actions probably some existing standard can be adopted for that and naturally it should be as much close to the existing de facto style as possible the first step is to study existing standards for c code and whether they are readily available in ides and automatic tools
1
324,694
9,907,708,286
IssuesEvent
2019-06-27 16:25:19
containous/traefik
https://api.github.com/repos/containous/traefik
closed
TLSOptions don't get applied - Traefik v2 with docker provider
area/provider/docker area/tls priority/P2
### Do you want to request a *feature* or report a *bug*? Bug ### Did you try using a 1.7.x configuration for the version 2.0? - [ ] Yes - [X] No ### What did you do? Using docker-compose to manage traefik. Have configured the docker and acme provider. After this a created some more services with docker-compose Part of my config ``` version: '3.7' services: ... web: build: ./web restart: always labels: - traefik.domain=nextcloud.example.com - traefik.http.routers.nextcloud-http.rule=Host(`nextcloud.example.com`) - traefik.http.routers.nextcloud-http.EntryPoints=http - traefik.http.routers.nextcloud-https.rule=Host(`nextcloud.example.com`) - traefik.http.routers.nextcloud-https.EntryPoints=https - traefik.http.routers.nextcloud-https.tls=true - traefik.http.routers.nextcloud-https.tls.options=foo - traefik.http.routers.nextcloud-http.middlewares=nextcloud-redirect - traefik.http.middlewares.nextcloud-redirect.redirectscheme.scheme=https - traefik.http.middlewares.nextcloud-redirect.redirectscheme.permanent=true - traefik.enable=true - traefik.docker.network=web expose: - 80 ... ``` ### What did you expect to see? My tls.options (`foo`) defined with a docker label, should be linked to the tls.options definition in my main configuration file. ### What did you see instead? ``` traefik_1 | time="2019-06-26T19:48:03+02:00" level=debug msg="unknown TLS options: foo@docker" entryPointName=https routerName=nextcloud-https@docker ``` ### Output of `traefik version`: (_What version of Traefik are you using?_) ``` Traefik version 2.0.0-alpha7 built on 2019-06-21T15:59:46Z" ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? My configuration ```toml [Global] sendAnonymousUsage = true [entrypoints] [entrypoints.http] address = ":80" [entrypoints.https] address = ":443" [log] level = "DEBUG" ################################################################ # API and dashboard configuration ################################################################ [api] dashboard = true ################################################################ # Docker configuration backend ################################################################ [providers] [providers.docker] watch = true endpoint = "unix:///var/run/docker.sock" exposedByDefault = false swarmMode = false [acme] email = "mymailexample.com" entryPoint = "https" acmeLogging = true onHostRule = true storage = "/etc/traefik/acme/acme.json" [acme.httpChallenge] entryPoint = "http" [TLSOptions] [TLSOptions.foo] MinVersion = "VersionTLS12" [TLSOptions.default] MinVersion = "VersionTLS13" ``` ### If applicable, please paste the log output in DEBUG level (`--log.level=DEBUG` switch) ```# docker-compose up --force-recreate Recreating traefik_traefik_1 ... done Attaching to traefik_traefik_1 traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Configuration loaded from file: /etc/traefik/traefik.toml" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Traefik version 2.0.0-alpha7 built on 2019-06-21T15:59:46Z" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Static configuration loaded {\"Global\":{\"CheckNewVersion\":true,\"SendAnonymousUsage\":true},\"ServersTransport\":{\"InsecureSkipVerify\":false,\"RootCAs\":null,\"MaxIdleConnsPerHost\":200,\"ForwardingTimeouts\":null},\"EntryPoints\":{\"http\":{\"Address\":\":80\",\"Transport\":{\"LifeCycle\":{\"RequestAcceptGraceTimeout\":0,\"GraceTimeOut\":10000000000},\"RespondingTimeouts\":{\"ReadTimeout\":0,\"WriteTimeout\":0,\"IdleTimeout\":180000000000}},\"ProxyProtocol\":null,\"ForwardedHeaders\":{\"Insecure\":false,\"TrustedIPs\":null}},\"https\":{\"Address\":\":443\",\"Transport\":{\"LifeCycle\":{\"RequestAcceptGraceTimeout\":0,\"GraceTimeOut\":10000000000},\"RespondingTimeouts\":{\"ReadTimeout\":0,\"WriteTimeout\":0,\"IdleTimeout\":180000000000}},\"ProxyProtocol\":null,\"ForwardedHeaders\":{\"Insecure\":false,\"TrustedIPs\":null}},\"traefik\":{\"Address\":\":8080\",\"Transport\":{\"LifeCycle\":{\"RequestAcceptGraceTimeout\":0,\"GraceTimeOut\":10000000000},\"RespondingTimeouts\":{\"ReadTimeout\":0,\"WriteTimeout\":0,\"IdleTimeout\":180000000000}},\"ProxyProtocol\":null,\"ForwardedHeaders\":{\"Insecure\":false,\"TrustedIPs\":null}}},\"Providers\":{\"ProvidersThrottleDuration\":2000000000,\"Docker\":{\"Constraints\":\"\",\"Watch\":true,\"Endpoint\":\"unix:///var/run/docker.sock\",\"DefaultRule\":\"Host(`{{ normalize .Name }}`)\",\"TLS\":null,\"ExposedByDefault\":false,\"UseBindPortIP\":false,\"SwarmMode\":false,\"Network\":\"\",\"SwarmModeRefreshSeconds\":15000000000},\"File\":null,\"Marathon\":null,\"Kubernetes\":null,\"KubernetesCRD\":null,\"Rest\":null,\"Rancher\":null},\"API\":{\"EntryPoint\":\"traefik\",\"Dashboard\":true,\"Debug\":false,\"Statistics\":null,\"Middlewares\":null},\"Metrics\":null,\"Ping\":null,\"Log\":{\"Level\":\"DEBUG\",\"format\":\"common\"},\"AccessLog\":null,\"Tracing\":null,\"HostResolver\":null,\"ACME\":{\"Email\":\"mymailexample.com\",\"ACMELogging\":true,\"CAServer\":\"https://acme-v02.api.letsencrypt.org/directory\",\"Storage\":\"/etc/traefik/acme/acme.json\",\"EntryPoint\":\"https\",\"KeyType\":\"RSA4096\",\"OnHostRule\":true,\"DNSChallenge\":null,\"HTTPChallenge\":{\"EntryPoint\":\"http\"},\"TLSChallenge\":null,\"Domains\":null}}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="\nStats collection is enabled.\nMany thanks for contributing to Traefik's improvement by allowing us to receive anonymous information from your configuration.\nHelp us improve Traefik by leaving this feature on :)\nMore details on: https://docs.traefik.io/basics/#collected-data\n" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No default certificate, generate one" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Starting provider aggregator.ProviderAggregator {}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Start TCP Server" entryPointName=https traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Start TCP Server" entryPointName=http traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Start TCP Server" entryPointName=traefik traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Starting provider *docker.Provider {\"Constraints\":\"\",\"Watch\":true,\"Endpoint\":\"unix:///var/run/docker.sock\",\"DefaultRule\":\"Host(`{{ normalize .Name }}`)\",\"TLS\":null,\"ExposedByDefault\":false,\"UseBindPortIP\":false,\"SwarmMode\":false,\"Network\":\"\",\"SwarmModeRefreshSeconds\":15000000000}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Starting provider *acme.Provider {\"Email\":\"mymailexample.com\",\"ACMELogging\":true,\"CAServer\":\"https://acme-v02.api.letsencrypt.org/directory\",\"Storage\":\"/etc/traefik/acme/acme.json\",\"EntryPoint\":\"https\",\"KeyType\":\"RSA4096\",\"OnHostRule\":true,\"DNSChallenge\":null,\"HTTPChallenge\":{\"EntryPoint\":\"http\"},\"TLSChallenge\":null,\"Domains\":null,\"Store\":{}}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Testing certificate renew..." providerName=acme traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Configuration received from provider ACME: {\"HTTP\":{},\"TCP\":null,\"TLSOptions\":null,\"TLSStores\":null}" providerName=ACME traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Provider connection established with docker 18.09.6 (API 1.39)" providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGaDCCBVCgAwIBAgISA1AklLVG+HfO4s+CamwsTr7fMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) traefik.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGazCCBVOgAwIBAgISA2mdYkBdOp3OL7bFGwyd5E0mMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) grafana.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGYTCCBUmgAwIBAgISA0ZbRXKGzBB7Q9odIqGMY31iMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) www.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGbjCCBVagAwIBAgISA/yY9jN2SEmQKPiVzmExqcrbMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) nextcloud.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No default certificate, generate one" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=traefik_traefik-284c591c83b87d6dff4f62a6d7e6d8a7d77d1a81bc21e2c6e94f319459454626 traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" container=cron_nextcloud-9f71f2f89979a5a351cc84e0c4b21e338da50336f0cef18f0be94421d2901d73 providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=app_nextcloud-809d59097b8afb9570adc7ba039db8bf30412516435bce9aaec03a6569e87aee traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=db_nextcloud-d61b6a8d207022049e3b1e2795c66d9c4fcdc7f954e5c927bc16d99e99645b53 traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" container=redis_nextcloud-617f69090b57aee4704341b363b5a662f2c4bb53015e37e3c1df4259e55c2380 providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=influxdb_influxdb-71ca581651d56c0dff4bac1cad1b1a0b80e580d8cc32aa86114f1d2a439f1fb6 traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" container=tor_tor-a530a71bcc09caf4e6a0fd7a996d5c8dbe42f42af47f9779c1bd256bd9c3e6a1 providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Configuration received from provider docker: {\"HTTP\":{\"routers\":{\"nextcloud-http\":{\"entryPoints\":[\"http\"],\"middlewares\":[\"nextcloud-redirect\"],\"service\":\"web_nextcloud\",\"rule\":\"Host(`nextcloud.example.com`)\"},\"nextcloud-https\":{\"entryPoints\":[\"https\"],\"service\":\"web_nextcloud\",\"rule\":\"Host(`nextcloud.example.com`)\",\"tls\":{\"options\":\"foo\"}}},\"middlewares\":{\"nextcloud-redirect\":{\"redirectScheme\":{\"scheme\":\"https\",\"permanent\":true}}},\"services\":{\"web_nextcloud\":{\"loadbalancer\":{\"servers\":[{\"url\":\"http://172.22.0.2:80\"}],\"passHostHeader\":true}}}},\"TCP\":{},\"TLSOptions\":null,\"TLSStores\":null}" providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) traefik.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) grafana.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) www.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) nextcloud.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No default certificate, generate one" traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" entryPointName=http middlewareName=pipelining middlewareType=Pipelining routerName=nextcloud-http@docker serviceName=web_nextcloud traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating load-balancer" entryPointName=http routerName=nextcloud-http@docker serviceName=web_nextcloud traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating server 0 http://172.22.0.2:80" routerName=nextcloud-http@docker serverName=0 serviceName=web_nextcloud entryPointName=http traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Added outgoing tracing middleware web_nextcloud" middlewareName=tracing middlewareType=TracingForwarder entryPointName=http routerName=nextcloud-http@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" entryPointName=http routerName=nextcloud-http@docker middlewareName=nextcloud-redirect@docker middlewareType=RedirectScheme traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Setting up redirection to https " entryPointName=http routerName=nextcloud-http@docker middlewareName=nextcloud-redirect@docker middlewareType=RedirectScheme traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Adding tracing to middleware" entryPointName=http routerName=nextcloud-http@docker middlewareName=nextcloud-redirect@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery middlewareType=Recovery entryPointName=http traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" routerName=nextcloud-https@docker entryPointName=https serviceName=web_nextcloud middlewareName=pipelining middlewareType=Pipelining traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating load-balancer" serviceName=web_nextcloud routerName=nextcloud-https@docker entryPointName=https traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating server 0 http://172.22.0.2:80" entryPointName=https serverName=0 serviceName=web_nextcloud routerName=nextcloud-https@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Added outgoing tracing middleware web_nextcloud" middlewareName=tracing entryPointName=https routerName=nextcloud-https@docker middlewareType=TracingForwarder traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" entryPointName=https middlewareName=traefik-internal-recovery middlewareType=Recovery traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="unknown TLS options: foo@docker" entryPointName=https routerName=nextcloud-https@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Try to challenge certificate for domain [nextcloud.example.com] founded in HostSNI rule" providerName=acme routerName=nextcloud-https rule="Host(`nextcloud.example.com`)" traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Looking for provided certificate(s) to validate [\"nextcloud.example.com\"]..." providerName=acme routerName=nextcloud-https rule="Host(`nextcloud.example.com`)" traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="No ACME certificate generation required for domains [\"nextcloud.example.com\"]." providerName=acme routerName=nextcloud-https rule="Host(`nextcloud.example.com`)"```
1.0
TLSOptions don't get applied - Traefik v2 with docker provider - ### Do you want to request a *feature* or report a *bug*? Bug ### Did you try using a 1.7.x configuration for the version 2.0? - [ ] Yes - [X] No ### What did you do? Using docker-compose to manage traefik. Have configured the docker and acme provider. After this a created some more services with docker-compose Part of my config ``` version: '3.7' services: ... web: build: ./web restart: always labels: - traefik.domain=nextcloud.example.com - traefik.http.routers.nextcloud-http.rule=Host(`nextcloud.example.com`) - traefik.http.routers.nextcloud-http.EntryPoints=http - traefik.http.routers.nextcloud-https.rule=Host(`nextcloud.example.com`) - traefik.http.routers.nextcloud-https.EntryPoints=https - traefik.http.routers.nextcloud-https.tls=true - traefik.http.routers.nextcloud-https.tls.options=foo - traefik.http.routers.nextcloud-http.middlewares=nextcloud-redirect - traefik.http.middlewares.nextcloud-redirect.redirectscheme.scheme=https - traefik.http.middlewares.nextcloud-redirect.redirectscheme.permanent=true - traefik.enable=true - traefik.docker.network=web expose: - 80 ... ``` ### What did you expect to see? My tls.options (`foo`) defined with a docker label, should be linked to the tls.options definition in my main configuration file. ### What did you see instead? ``` traefik_1 | time="2019-06-26T19:48:03+02:00" level=debug msg="unknown TLS options: foo@docker" entryPointName=https routerName=nextcloud-https@docker ``` ### Output of `traefik version`: (_What version of Traefik are you using?_) ``` Traefik version 2.0.0-alpha7 built on 2019-06-21T15:59:46Z" ``` ### What is your environment & configuration (arguments, toml, provider, platform, ...)? My configuration ```toml [Global] sendAnonymousUsage = true [entrypoints] [entrypoints.http] address = ":80" [entrypoints.https] address = ":443" [log] level = "DEBUG" ################################################################ # API and dashboard configuration ################################################################ [api] dashboard = true ################################################################ # Docker configuration backend ################################################################ [providers] [providers.docker] watch = true endpoint = "unix:///var/run/docker.sock" exposedByDefault = false swarmMode = false [acme] email = "mymailexample.com" entryPoint = "https" acmeLogging = true onHostRule = true storage = "/etc/traefik/acme/acme.json" [acme.httpChallenge] entryPoint = "http" [TLSOptions] [TLSOptions.foo] MinVersion = "VersionTLS12" [TLSOptions.default] MinVersion = "VersionTLS13" ``` ### If applicable, please paste the log output in DEBUG level (`--log.level=DEBUG` switch) ```# docker-compose up --force-recreate Recreating traefik_traefik_1 ... done Attaching to traefik_traefik_1 traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Configuration loaded from file: /etc/traefik/traefik.toml" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Traefik version 2.0.0-alpha7 built on 2019-06-21T15:59:46Z" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Static configuration loaded {\"Global\":{\"CheckNewVersion\":true,\"SendAnonymousUsage\":true},\"ServersTransport\":{\"InsecureSkipVerify\":false,\"RootCAs\":null,\"MaxIdleConnsPerHost\":200,\"ForwardingTimeouts\":null},\"EntryPoints\":{\"http\":{\"Address\":\":80\",\"Transport\":{\"LifeCycle\":{\"RequestAcceptGraceTimeout\":0,\"GraceTimeOut\":10000000000},\"RespondingTimeouts\":{\"ReadTimeout\":0,\"WriteTimeout\":0,\"IdleTimeout\":180000000000}},\"ProxyProtocol\":null,\"ForwardedHeaders\":{\"Insecure\":false,\"TrustedIPs\":null}},\"https\":{\"Address\":\":443\",\"Transport\":{\"LifeCycle\":{\"RequestAcceptGraceTimeout\":0,\"GraceTimeOut\":10000000000},\"RespondingTimeouts\":{\"ReadTimeout\":0,\"WriteTimeout\":0,\"IdleTimeout\":180000000000}},\"ProxyProtocol\":null,\"ForwardedHeaders\":{\"Insecure\":false,\"TrustedIPs\":null}},\"traefik\":{\"Address\":\":8080\",\"Transport\":{\"LifeCycle\":{\"RequestAcceptGraceTimeout\":0,\"GraceTimeOut\":10000000000},\"RespondingTimeouts\":{\"ReadTimeout\":0,\"WriteTimeout\":0,\"IdleTimeout\":180000000000}},\"ProxyProtocol\":null,\"ForwardedHeaders\":{\"Insecure\":false,\"TrustedIPs\":null}}},\"Providers\":{\"ProvidersThrottleDuration\":2000000000,\"Docker\":{\"Constraints\":\"\",\"Watch\":true,\"Endpoint\":\"unix:///var/run/docker.sock\",\"DefaultRule\":\"Host(`{{ normalize .Name }}`)\",\"TLS\":null,\"ExposedByDefault\":false,\"UseBindPortIP\":false,\"SwarmMode\":false,\"Network\":\"\",\"SwarmModeRefreshSeconds\":15000000000},\"File\":null,\"Marathon\":null,\"Kubernetes\":null,\"KubernetesCRD\":null,\"Rest\":null,\"Rancher\":null},\"API\":{\"EntryPoint\":\"traefik\",\"Dashboard\":true,\"Debug\":false,\"Statistics\":null,\"Middlewares\":null},\"Metrics\":null,\"Ping\":null,\"Log\":{\"Level\":\"DEBUG\",\"format\":\"common\"},\"AccessLog\":null,\"Tracing\":null,\"HostResolver\":null,\"ACME\":{\"Email\":\"mymailexample.com\",\"ACMELogging\":true,\"CAServer\":\"https://acme-v02.api.letsencrypt.org/directory\",\"Storage\":\"/etc/traefik/acme/acme.json\",\"EntryPoint\":\"https\",\"KeyType\":\"RSA4096\",\"OnHostRule\":true,\"DNSChallenge\":null,\"HTTPChallenge\":{\"EntryPoint\":\"http\"},\"TLSChallenge\":null,\"Domains\":null}}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="\nStats collection is enabled.\nMany thanks for contributing to Traefik's improvement by allowing us to receive anonymous information from your configuration.\nHelp us improve Traefik by leaving this feature on :)\nMore details on: https://docs.traefik.io/basics/#collected-data\n" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No default certificate, generate one" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Starting provider aggregator.ProviderAggregator {}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Start TCP Server" entryPointName=https traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Start TCP Server" entryPointName=http traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Start TCP Server" entryPointName=traefik traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Starting provider *docker.Provider {\"Constraints\":\"\",\"Watch\":true,\"Endpoint\":\"unix:///var/run/docker.sock\",\"DefaultRule\":\"Host(`{{ normalize .Name }}`)\",\"TLS\":null,\"ExposedByDefault\":false,\"UseBindPortIP\":false,\"SwarmMode\":false,\"Network\":\"\",\"SwarmModeRefreshSeconds\":15000000000}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Starting provider *acme.Provider {\"Email\":\"mymailexample.com\",\"ACMELogging\":true,\"CAServer\":\"https://acme-v02.api.letsencrypt.org/directory\",\"Storage\":\"/etc/traefik/acme/acme.json\",\"EntryPoint\":\"https\",\"KeyType\":\"RSA4096\",\"OnHostRule\":true,\"DNSChallenge\":null,\"HTTPChallenge\":{\"EntryPoint\":\"http\"},\"TLSChallenge\":null,\"Domains\":null,\"Store\":{}}" traefik_1 | time="2019-06-26T20:19:52+02:00" level=info msg="Testing certificate renew..." providerName=acme traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Configuration received from provider ACME: {\"HTTP\":{},\"TCP\":null,\"TLSOptions\":null,\"TLSStores\":null}" providerName=ACME traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Provider connection established with docker 18.09.6 (API 1.39)" providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGaDCCBVCgAwIBAgISA1AklLVG+HfO4s+CamwsTr7fMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) traefik.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGazCCBVOgAwIBAgISA2mdYkBdOp3OL7bFGwyd5E0mMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) grafana.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGYTCCBUmgAwIBAgISA0ZbRXKGzBB7Q9odIqGMY31iMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) www.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No store is defined to add the certificate MIIGbjCCBVagAwIBAgISA/yY9jN2SEmQKPiVzmExqcrbMA0GCS, it will be added to the default store." traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) nextcloud.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No default certificate, generate one" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=traefik_traefik-284c591c83b87d6dff4f62a6d7e6d8a7d77d1a81bc21e2c6e94f319459454626 traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" container=cron_nextcloud-9f71f2f89979a5a351cc84e0c4b21e338da50336f0cef18f0be94421d2901d73 providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=app_nextcloud-809d59097b8afb9570adc7ba039db8bf30412516435bce9aaec03a6569e87aee traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=db_nextcloud-d61b6a8d207022049e3b1e2795c66d9c4fcdc7f954e5c927bc16d99e99645b53 traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" container=redis_nextcloud-617f69090b57aee4704341b363b5a662f2c4bb53015e37e3c1df4259e55c2380 providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" providerName=docker container=influxdb_influxdb-71ca581651d56c0dff4bac1cad1b1a0b80e580d8cc32aa86114f1d2a439f1fb6 traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Filtering disabled container" container=tor_tor-a530a71bcc09caf4e6a0fd7a996d5c8dbe42f42af47f9779c1bd256bd9c3e6a1 providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Configuration received from provider docker: {\"HTTP\":{\"routers\":{\"nextcloud-http\":{\"entryPoints\":[\"http\"],\"middlewares\":[\"nextcloud-redirect\"],\"service\":\"web_nextcloud\",\"rule\":\"Host(`nextcloud.example.com`)\"},\"nextcloud-https\":{\"entryPoints\":[\"https\"],\"service\":\"web_nextcloud\",\"rule\":\"Host(`nextcloud.example.com`)\",\"tls\":{\"options\":\"foo\"}}},\"middlewares\":{\"nextcloud-redirect\":{\"redirectScheme\":{\"scheme\":\"https\",\"permanent\":true}}},\"services\":{\"web_nextcloud\":{\"loadbalancer\":{\"servers\":[{\"url\":\"http://172.22.0.2:80\"}],\"passHostHeader\":true}}}},\"TCP\":{},\"TLSOptions\":null,\"TLSStores\":null}" providerName=docker traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) traefik.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) grafana.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) www.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="Adding certificate for domain(s) nextcloud.example.com" traefik_1 | time="2019-06-26T20:19:52+02:00" level=debug msg="No default certificate, generate one" traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" entryPointName=http middlewareName=pipelining middlewareType=Pipelining routerName=nextcloud-http@docker serviceName=web_nextcloud traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating load-balancer" entryPointName=http routerName=nextcloud-http@docker serviceName=web_nextcloud traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating server 0 http://172.22.0.2:80" routerName=nextcloud-http@docker serverName=0 serviceName=web_nextcloud entryPointName=http traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Added outgoing tracing middleware web_nextcloud" middlewareName=tracing middlewareType=TracingForwarder entryPointName=http routerName=nextcloud-http@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" entryPointName=http routerName=nextcloud-http@docker middlewareName=nextcloud-redirect@docker middlewareType=RedirectScheme traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Setting up redirection to https " entryPointName=http routerName=nextcloud-http@docker middlewareName=nextcloud-redirect@docker middlewareType=RedirectScheme traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Adding tracing to middleware" entryPointName=http routerName=nextcloud-http@docker middlewareName=nextcloud-redirect@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" middlewareName=traefik-internal-recovery middlewareType=Recovery entryPointName=http traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" routerName=nextcloud-https@docker entryPointName=https serviceName=web_nextcloud middlewareName=pipelining middlewareType=Pipelining traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating load-balancer" serviceName=web_nextcloud routerName=nextcloud-https@docker entryPointName=https traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating server 0 http://172.22.0.2:80" entryPointName=https serverName=0 serviceName=web_nextcloud routerName=nextcloud-https@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Added outgoing tracing middleware web_nextcloud" middlewareName=tracing entryPointName=https routerName=nextcloud-https@docker middlewareType=TracingForwarder traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Creating middleware" entryPointName=https middlewareName=traefik-internal-recovery middlewareType=Recovery traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="unknown TLS options: foo@docker" entryPointName=https routerName=nextcloud-https@docker traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Try to challenge certificate for domain [nextcloud.example.com] founded in HostSNI rule" providerName=acme routerName=nextcloud-https rule="Host(`nextcloud.example.com`)" traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="Looking for provided certificate(s) to validate [\"nextcloud.example.com\"]..." providerName=acme routerName=nextcloud-https rule="Host(`nextcloud.example.com`)" traefik_1 | time="2019-06-26T20:19:53+02:00" level=debug msg="No ACME certificate generation required for domains [\"nextcloud.example.com\"]." providerName=acme routerName=nextcloud-https rule="Host(`nextcloud.example.com`)"```
non_main
tlsoptions don t get applied traefik with docker provider do you want to request a feature or report a bug bug did you try using a x configuration for the version yes no what did you do using docker compose to manage traefik have configured the docker and acme provider after this a created some more services with docker compose part of my config version services web build web restart always labels traefik domain nextcloud example com traefik http routers nextcloud http rule host nextcloud example com traefik http routers nextcloud http entrypoints http traefik http routers nextcloud https rule host nextcloud example com traefik http routers nextcloud https entrypoints https traefik http routers nextcloud https tls true traefik http routers nextcloud https tls options foo traefik http routers nextcloud http middlewares nextcloud redirect traefik http middlewares nextcloud redirect redirectscheme scheme https traefik http middlewares nextcloud redirect redirectscheme permanent true traefik enable true traefik docker network web expose what did you expect to see my tls options foo defined with a docker label should be linked to the tls options definition in my main configuration file what did you see instead traefik time level debug msg unknown tls options foo docker entrypointname https routername nextcloud https docker output of traefik version what version of traefik are you using traefik version built on what is your environment configuration arguments toml provider platform my configuration toml sendanonymoususage true address address level debug api and dashboard configuration dashboard true docker configuration backend watch true endpoint unix var run docker sock exposedbydefault false swarmmode false email mymailexample com entrypoint https acmelogging true onhostrule true storage etc traefik acme acme json entrypoint http minversion minversion if applicable please paste the log output in debug level log level debug switch docker compose up force recreate recreating traefik traefik done attaching to traefik traefik traefik time level info msg configuration loaded from file etc traefik traefik toml traefik time level info msg traefik version built on traefik time level debug msg static configuration loaded global checknewversion true sendanonymoususage true serverstransport insecureskipverify false rootcas null maxidleconnsperhost forwardingtimeouts null entrypoints http address transport lifecycle requestacceptgracetimeout gracetimeout respondingtimeouts readtimeout writetimeout idletimeout proxyprotocol null forwardedheaders insecure false trustedips null https address transport lifecycle requestacceptgracetimeout gracetimeout respondingtimeouts readtimeout writetimeout idletimeout proxyprotocol null forwardedheaders insecure false trustedips null traefik address transport lifecycle requestacceptgracetimeout gracetimeout respondingtimeouts readtimeout writetimeout idletimeout proxyprotocol null forwardedheaders insecure false trustedips null providers providersthrottleduration docker constraints watch true endpoint unix var run docker sock defaultrule host normalize name tls null exposedbydefault false usebindportip false swarmmode false network swarmmoderefreshseconds file null marathon null kubernetes null kubernetescrd null rest null rancher null api entrypoint traefik dashboard true debug false statistics null middlewares null metrics null ping null log level debug format common accesslog null tracing null hostresolver null acme email mymailexample com acmelogging true caserver traefik time level info msg nstats collection is enabled nmany thanks for contributing to traefik s improvement by allowing us to receive anonymous information from your configuration nhelp us improve traefik by leaving this feature on nmore details on traefik time level debug msg no default certificate generate one traefik time level info msg starting provider aggregator provideraggregator traefik time level debug msg start tcp server entrypointname https traefik time level debug msg start tcp server entrypointname http traefik time level debug msg start tcp server entrypointname traefik traefik time level info msg starting provider docker provider constraints watch true endpoint unix var run docker sock defaultrule host normalize name tls null exposedbydefault false usebindportip false swarmmode false network swarmmoderefreshseconds traefik time level info msg starting provider acme provider email mymailexample com acmelogging true caserver traefik time level info msg testing certificate renew providername acme traefik time level debug msg configuration received from provider acme http tcp null tlsoptions null tlsstores null providername acme traefik time level debug msg provider connection established with docker api providername docker traefik time level debug msg no store is defined to add the certificate it will be added to the default store traefik time level debug msg adding certificate for domain s traefik example com traefik time level debug msg no store is defined to add the certificate it will be added to the default store traefik time level debug msg adding certificate for domain s grafana example com traefik time level debug msg no store is defined to add the certificate it will be added to the default store traefik time level debug msg adding certificate for domain s traefik time level debug msg no store is defined to add the certificate miigbjccbvagawibagisa it will be added to the default store traefik time level debug msg adding certificate for domain s nextcloud example com traefik time level debug msg no default certificate generate one traefik time level debug msg filtering disabled container providername docker container traefik traefik traefik time level debug msg filtering disabled container container cron nextcloud providername docker traefik time level debug msg filtering disabled container providername docker container app nextcloud traefik time level debug msg filtering disabled container providername docker container db nextcloud traefik time level debug msg filtering disabled container container redis nextcloud providername docker traefik time level debug msg filtering disabled container providername docker container influxdb influxdb traefik time level debug msg filtering disabled container container tor tor providername docker traefik time level debug msg configuration received from provider docker http routers nextcloud http entrypoints middlewares service web nextcloud rule host nextcloud example com nextcloud https entrypoints service web nextcloud rule host nextcloud example com tls options foo middlewares nextcloud redirect redirectscheme scheme https permanent true services web nextcloud loadbalancer servers passhostheader true tcp tlsoptions null tlsstores null providername docker traefik time level debug msg adding certificate for domain s traefik example com traefik time level debug msg adding certificate for domain s grafana example com traefik time level debug msg adding certificate for domain s traefik time level debug msg adding certificate for domain s nextcloud example com traefik time level debug msg no default certificate generate one traefik time level debug msg creating middleware entrypointname http middlewarename pipelining middlewaretype pipelining routername nextcloud http docker servicename web nextcloud traefik time level debug msg creating load balancer entrypointname http routername nextcloud http docker servicename web nextcloud traefik time level debug msg creating server routername nextcloud http docker servername servicename web nextcloud entrypointname http traefik time level debug msg added outgoing tracing middleware web nextcloud middlewarename tracing middlewaretype tracingforwarder entrypointname http routername nextcloud http docker traefik time level debug msg creating middleware entrypointname http routername nextcloud http docker middlewarename nextcloud redirect docker middlewaretype redirectscheme traefik time level debug msg setting up redirection to https entrypointname http routername nextcloud http docker middlewarename nextcloud redirect docker middlewaretype redirectscheme traefik time level debug msg adding tracing to middleware entrypointname http routername nextcloud http docker middlewarename nextcloud redirect docker traefik time level debug msg creating middleware middlewarename traefik internal recovery middlewaretype recovery entrypointname http traefik time level debug msg creating middleware routername nextcloud https docker entrypointname https servicename web nextcloud middlewarename pipelining middlewaretype pipelining traefik time level debug msg creating load balancer servicename web nextcloud routername nextcloud https docker entrypointname https traefik time level debug msg creating server entrypointname https servername servicename web nextcloud routername nextcloud https docker traefik time level debug msg added outgoing tracing middleware web nextcloud middlewarename tracing entrypointname https routername nextcloud https docker middlewaretype tracingforwarder traefik time level debug msg creating middleware entrypointname https middlewarename traefik internal recovery middlewaretype recovery traefik time level debug msg unknown tls options foo docker entrypointname https routername nextcloud https docker traefik time level debug msg try to challenge certificate for domain founded in hostsni rule providername acme routername nextcloud https rule host nextcloud example com traefik time level debug msg looking for provided certificate s to validate providername acme routername nextcloud https rule host nextcloud example com traefik time level debug msg no acme certificate generation required for domains providername acme routername nextcloud https rule host nextcloud example com
0
3,424
13,182,104,664
IssuesEvent
2020-08-12 15:16:22
duo-labs/cloudmapper
https://api.github.com/repos/duo-labs/cloudmapper
closed
Add ability to append security group description
enhancement map unmaintained_functionality
This is a feature request more than anything. It would be great if there was an option to append security group IP descriptions next to their IP address. It would help when we visually look at a map and say hey what is this IP? I suppose this would basically be automating the IP definitions in the config.json. (Assuming descriptions exist for those IPs in EC2) Anyways, wouldn't make sense to enable this by default by an option such as --include-descriptions would be a nice add.
True
Add ability to append security group description - This is a feature request more than anything. It would be great if there was an option to append security group IP descriptions next to their IP address. It would help when we visually look at a map and say hey what is this IP? I suppose this would basically be automating the IP definitions in the config.json. (Assuming descriptions exist for those IPs in EC2) Anyways, wouldn't make sense to enable this by default by an option such as --include-descriptions would be a nice add.
main
add ability to append security group description this is a feature request more than anything it would be great if there was an option to append security group ip descriptions next to their ip address it would help when we visually look at a map and say hey what is this ip i suppose this would basically be automating the ip definitions in the config json assuming descriptions exist for those ips in anyways wouldn t make sense to enable this by default by an option such as include descriptions would be a nice add
1
1,219
5,199,689,645
IssuesEvent
2017-01-23 21:35:05
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
It would be nice for `copy` to optionally show a progress bar during upload
affects_2.0 feature_idea waiting_on_maintainer
##### Issue Type: - Feature Idea ##### Plugin Name: copy ##### Ansible Version: ``` ansible 2.0.1.0 config file = /home/.../ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: ``` $ cat ansible.cfg [defaults] inventory = hosts roles_path = roles/ ``` ##### Environment: N/A, not platform-specific ##### Summary: I have an ansible playbook which uses the `copy` module to upload a 3GB file to managed hosts. This takes a long time, and causes me to wonder whether it's still uploading or got stuck somewhere. It would be nice if the `copy` module had an option for showing a progress indicator while uploading. ##### Steps To Reproduce: How the new feature might be used: ``` - name: Upload Xcode DMG to the host copy: src={{ xcode_dmg }} dest=/Users/{{ ansible_ssh_user }} show_progress_bar=True when: xcode_dl_path.stat.exists == False ``` ##### Expected Results: It would be nice if the playbook output a progress indicator when the progress bar setting was enabled. ##### Actual Results: I read the docs and asked on IRC and daemoniz confirmed that there is not a built-in feature to do this.
True
It would be nice for `copy` to optionally show a progress bar during upload - ##### Issue Type: - Feature Idea ##### Plugin Name: copy ##### Ansible Version: ``` ansible 2.0.1.0 config file = /home/.../ansible.cfg configured module search path = Default w/o overrides ``` ##### Ansible Configuration: ``` $ cat ansible.cfg [defaults] inventory = hosts roles_path = roles/ ``` ##### Environment: N/A, not platform-specific ##### Summary: I have an ansible playbook which uses the `copy` module to upload a 3GB file to managed hosts. This takes a long time, and causes me to wonder whether it's still uploading or got stuck somewhere. It would be nice if the `copy` module had an option for showing a progress indicator while uploading. ##### Steps To Reproduce: How the new feature might be used: ``` - name: Upload Xcode DMG to the host copy: src={{ xcode_dmg }} dest=/Users/{{ ansible_ssh_user }} show_progress_bar=True when: xcode_dl_path.stat.exists == False ``` ##### Expected Results: It would be nice if the playbook output a progress indicator when the progress bar setting was enabled. ##### Actual Results: I read the docs and asked on IRC and daemoniz confirmed that there is not a built-in feature to do this.
main
it would be nice for copy to optionally show a progress bar during upload issue type feature idea plugin name copy ansible version ansible config file home ansible cfg configured module search path default w o overrides ansible configuration cat ansible cfg inventory hosts roles path roles environment n a not platform specific summary i have an ansible playbook which uses the copy module to upload a file to managed hosts this takes a long time and causes me to wonder whether it s still uploading or got stuck somewhere it would be nice if the copy module had an option for showing a progress indicator while uploading steps to reproduce how the new feature might be used name upload xcode dmg to the host copy src xcode dmg dest users ansible ssh user show progress bar true when xcode dl path stat exists false expected results it would be nice if the playbook output a progress indicator when the progress bar setting was enabled actual results i read the docs and asked on irc and daemoniz confirmed that there is not a built in feature to do this
1
4,276
21,509,509,014
IssuesEvent
2022-04-28 01:48:54
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
aws-cdk: sam local error on generated template
type/bug stage/needs-investigation maintainer/need-followup
When I try to run `sam local start-api` with the generated template from aws-cdk, I get the following error when I make an OPTIONS request to the `/time` endpoint. Error: ``` Exception on /time [OPTIONS] Traceback (most recent call last): File "/usr/lib/python3.8/site-packages/flask/app.py", line 2447, in wsgi_app response = self.full_dispatch_request() File "/usr/lib/python3.8/site-packages/flask/app.py", line 1952, in full_dispatch_request rv = self.handle_user_exception(e) File "/usr/lib/python3.8/site-packages/flask/app.py", line 1821, in handle_user_exception reraise(exc_type, exc_value, tb) File "/usr/lib/python3.8/site-packages/flask/_compat.py", line 39, in reraise raise value File "/usr/lib/python3.8/site-packages/flask/app.py", line 1950, in full_dispatch_request rv = self.dispatch_request() File "/usr/lib/python3.8/site-packages/flask/app.py", line 1936, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/usr/lib/python3.8/site-packages/samcli/local/apigw/local_apigw_service.py", line 203, in _request_handler self.lambda_runner.invoke(route.function_name, event, stdout=stdout_stream_writer, stderr=self.stderr) File "/usr/lib/python3.8/site-packages/samcli/commands/local/lib/local_lambda.py", line 84, in invoke function = self.provider.get(function_name) File "/usr/lib/python3.8/site-packages/samcli/lib/providers/sam_function_provider.py", line 65, in get raise ValueError("Function name is required") ValueError: Function name is required ``` The template file: ``` Resources: WorkTrackerBC4B47B0: Type: AWS::DynamoDB::Table Properties: KeySchema: - AttributeName: user KeyType: HASH - AttributeName: timestamp KeyType: RANGE AttributeDefinitions: - AttributeName: user AttributeType: S - AttributeName: timestamp AttributeType: S ProvisionedThroughput: ReadCapacityUnits: 5 WriteCapacityUnits: 5 TableName: WorkTracker UpdateReplacePolicy: Retain DeletionPolicy: Retain Metadata: aws:cdk:path: work-tracker/WorkTracker/Resource WorkTrackerApiA3820895: Type: AWS::ApiGateway::RestApi Properties: Name: WorkTrackerApi Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Resource WorkTrackerApiCloudWatchRole46BF1BDC: Type: AWS::IAM::Role Properties: AssumeRolePolicyDocument: Statement: - Action: sts:AssumeRole Effect: Allow Principal: Service: apigateway.amazonaws.com Version: "2012-10-17" ManagedPolicyArns: - Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - :iam::aws:policy/service-role/AmazonAPIGatewayPushToCloudWatchLogs Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/CloudWatchRole/Resource WorkTrackerApiAccountB5820D17: Type: AWS::ApiGateway::Account Properties: CloudWatchRoleArn: Fn::GetAtt: - WorkTrackerApiCloudWatchRole46BF1BDC - Arn DependsOn: - WorkTrackerApiA3820895 Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Account WorkTrackerApiDeployment667E8F39a4dede785b4b39d382183529b3eb56f9: Type: AWS::ApiGateway::Deployment Properties: RestApiId: Ref: WorkTrackerApiA3820895 Description: Automatically created by the RestApi construct DependsOn: - WorkTrackerApiOPTIONS4CF25EFB - WorkTrackerApitimeGETAC751B15 - WorkTrackerApitimeOPTIONS121CA0A4 - WorkTrackerApitimePOSTB776371B - WorkTrackerApitimeE76CB408 Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Deployment/Resource WorkTrackerApiDeploymentStageprodE0E1DD82: Type: AWS::ApiGateway::Stage Properties: RestApiId: Ref: WorkTrackerApiA3820895 DeploymentId: Ref: WorkTrackerApiDeployment667E8F39a4dede785b4b39d382183529b3eb56f9 MethodSettings: - HttpMethod: "*" ResourcePath: /* ThrottlingBurstLimit: 10 ThrottlingRateLimit: 10 StageName: prod Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/DeploymentStage.prod/Resource WorkTrackerApiOPTIONS4CF25EFB: Type: AWS::ApiGateway::Method Properties: HttpMethod: OPTIONS ResourceId: Fn::GetAtt: - WorkTrackerApiA3820895 - RootResourceId RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: NONE Integration: IntegrationResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: "'Authorization'" method.response.header.Access-Control-Allow-Origin: "'*'" method.response.header.Access-Control-Allow-Methods: "'OPTIONS,GET,PUT,POST,DELETE,PATCH,HEAD'" method.response.header.Access-Control-Allow-Credentials: "'true'" StatusCode: "204" RequestTemplates: application/json: "{ statusCode: 200 }" Type: MOCK MethodResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: true method.response.header.Access-Control-Allow-Origin: true method.response.header.Access-Control-Allow-Methods: true method.response.header.Access-Control-Allow-Credentials: true StatusCode: "204" Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/OPTIONS/Resource WorkTrackerApitimeE76CB408: Type: AWS::ApiGateway::Resource Properties: ParentId: Fn::GetAtt: - WorkTrackerApiA3820895 - RootResourceId PathPart: time RestApiId: Ref: WorkTrackerApiA3820895 Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/Resource WorkTrackerApitimeOPTIONS121CA0A4: Type: AWS::ApiGateway::Method Properties: HttpMethod: OPTIONS ResourceId: Ref: WorkTrackerApitimeE76CB408 RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: NONE Integration: IntegrationResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: "'Authorization'" method.response.header.Access-Control-Allow-Origin: "'*'" method.response.header.Access-Control-Allow-Methods: "'OPTIONS,GET,PUT,POST,DELETE,PATCH,HEAD'" method.response.header.Access-Control-Allow-Credentials: "'true'" StatusCode: "204" RequestTemplates: application/json: "{ statusCode: 200 }" Type: MOCK MethodResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: true method.response.header.Access-Control-Allow-Origin: true method.response.header.Access-Control-Allow-Methods: true method.response.header.Access-Control-Allow-Credentials: true StatusCode: "204" Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/OPTIONS/Resource WorkTrackerApitimeGETApiPermissionworktrackerWorkTrackerApi0D6A680AGETtime5350A293: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - / - Ref: WorkTrackerApiDeploymentStageprodE0E1DD82 - /GET/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/GET/ApiPermission.worktrackerWorkTrackerApi0D6A680A.GET..time WorkTrackerApitimeGETApiPermissionTestworktrackerWorkTrackerApi0D6A680AGETtime05C7716C: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - /test-invoke-stage/GET/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/GET/ApiPermission.Test.worktrackerWorkTrackerApi0D6A680A.GET..time WorkTrackerApitimeGETAC751B15: Type: AWS::ApiGateway::Method Properties: HttpMethod: GET ResourceId: Ref: WorkTrackerApitimeE76CB408 RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: COGNITO_USER_POOLS AuthorizerId: Ref: CognitoAuthorizer Integration: IntegrationHttpMethod: POST Type: AWS_PROXY Uri: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":apigateway:" - Ref: AWS::Region - :lambda:path/2015-03-31/functions/ - Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn - /invocations Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/GET/Resource WorkTrackerApitimePOSTApiPermissionworktrackerWorkTrackerApi0D6A680APOSTtime9B216B47: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - / - Ref: WorkTrackerApiDeploymentStageprodE0E1DD82 - /POST/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/POST/ApiPermission.worktrackerWorkTrackerApi0D6A680A.POST..time WorkTrackerApitimePOSTApiPermissionTestworktrackerWorkTrackerApi0D6A680APOSTtimeA2323B78: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - /test-invoke-stage/POST/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/POST/ApiPermission.Test.worktrackerWorkTrackerApi0D6A680A.POST..time WorkTrackerApitimePOSTB776371B: Type: AWS::ApiGateway::Method Properties: HttpMethod: POST ResourceId: Ref: WorkTrackerApitimeE76CB408 RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: COGNITO_USER_POOLS AuthorizerId: Ref: CognitoAuthorizer Integration: IntegrationHttpMethod: POST Type: AWS_PROXY Uri: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":apigateway:" - Ref: AWS::Region - :lambda:path/2015-03-31/functions/ - Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn - /invocations Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/POST/Resource getTimesOfDayServiceRole7C5D69EC: Type: AWS::IAM::Role Properties: AssumeRolePolicyDocument: Statement: - Action: sts:AssumeRole Effect: Allow Principal: Service: lambda.amazonaws.com Version: "2012-10-17" ManagedPolicyArns: - Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - :iam::aws:policy/service-role/AWSLambdaBasicExecutionRole Metadata: aws:cdk:path: work-tracker/getTimesOfDay/ServiceRole/Resource getTimesOfDayServiceRoleDefaultPolicy371E6EE6: Type: AWS::IAM::Policy Properties: PolicyDocument: Statement: - Action: - dynamodb:BatchGetItem - dynamodb:GetRecords - dynamodb:GetShardIterator - dynamodb:Query - dynamodb:GetItem - dynamodb:Scan Effect: Allow Resource: - Fn::GetAtt: - WorkTrackerBC4B47B0 - Arn - Ref: AWS::NoValue Version: "2012-10-17" PolicyName: getTimesOfDayServiceRoleDefaultPolicy371E6EE6 Roles: - Ref: getTimesOfDayServiceRole7C5D69EC Metadata: aws:cdk:path: work-tracker/getTimesOfDay/ServiceRole/DefaultPolicy/Resource getTimesOfDayC45A06A5: Type: AWS::Lambda::Function Properties: Code: S3Bucket: Ref: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3BucketB50D3E5E S3Key: Fn::Join: - "" - - Fn::Select: - 0 - Fn::Split: - "||" - Ref: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3VersionKey9E270B2B - Fn::Select: - 1 - Fn::Split: - "||" - Ref: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3VersionKey9E270B2B Handler: daily-time.handler Role: Fn::GetAtt: - getTimesOfDayServiceRole7C5D69EC - Arn Runtime: nodejs12.x DependsOn: - getTimesOfDayServiceRoleDefaultPolicy371E6EE6 - getTimesOfDayServiceRole7C5D69EC Metadata: aws:cdk:path: work-tracker/getTimesOfDay/Resource aws:asset:path: /home/sakkaku/Documents/workspaces/myin-work/dist/libs/daily-time aws:asset:property: Code CognitoAuthorizer: Type: AWS::ApiGateway::Authorizer Properties: RestApiId: Ref: WorkTrackerApiA3820895 Type: COGNITO_USER_POOLS IdentitySource: method.request.header.Authorization Name: MainPoolAuthorizer ProviderARNs: [] Metadata: aws:cdk:path: work-tracker/CognitoAuthorizer CDKMetadata: Type: AWS::CDK::Metadata Properties: Modules: aws-cdk=1.65.0,@aws-cdk/assets=1.65.0,@aws-cdk/aws-apigateway=1.65.0,@aws-cdk/aws-applicationautoscaling=1.65.0,@aws-cdk/aws-autoscaling-common=1.65.0,@aws-cdk/aws-cloudwatch=1.65.0,@aws-cdk/aws-codeguruprofiler=1.65.0,@aws-cdk/aws-dynamodb=1.65.0,@aws-cdk/aws-ec2=1.65.0,@aws-cdk/aws-events=1.65.0,@aws-cdk/aws-iam=1.65.0,@aws-cdk/aws-kms=1.65.0,@aws-cdk/aws-lambda=1.65.0,@aws-cdk/aws-logs=1.65.0,@aws-cdk/aws-s3=1.65.0,@aws-cdk/aws-s3-assets=1.65.0,@aws-cdk/aws-sqs=1.65.0,@aws-cdk/aws-ssm=1.65.0,@aws-cdk/cloud-assembly-schema=1.65.0,@aws-cdk/core=1.65.0,@aws-cdk/custom-resources=1.65.0,@aws-cdk/cx-api=1.65.0,@aws-cdk/region-info=1.65.0,jsii-runtime=node.js/v14.10.1 Metadata: aws:cdk:path: work-tracker/CDKMetadata/Default Condition: CDKMetadataAvailable Outputs: WorkTrackerApiEndpoint24C3B084: Value: Fn::Join: - "" - - https:// - Ref: WorkTrackerApiA3820895 - .execute-api. - Ref: AWS::Region - "." - Ref: AWS::URLSuffix - / - Ref: WorkTrackerApiDeploymentStageprodE0E1DD82 - / Parameters: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3BucketB50D3E5E: Type: String Description: S3 bucket for asset "f4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781ba" AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3VersionKey9E270B2B: Type: String Description: S3 key for asset version "f4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781ba" AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baArtifactHashD8875F9F: Type: String Description: Artifact hash for asset "f4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781ba" Conditions: CDKMetadataAvailable: Fn::Or: - Fn::Or: - Fn::Equals: - Ref: AWS::Region - ap-east-1 - Fn::Equals: - Ref: AWS::Region - ap-northeast-1 - Fn::Equals: - Ref: AWS::Region - ap-northeast-2 - Fn::Equals: - Ref: AWS::Region - ap-south-1 - Fn::Equals: - Ref: AWS::Region - ap-southeast-1 - Fn::Equals: - Ref: AWS::Region - ap-southeast-2 - Fn::Equals: - Ref: AWS::Region - ca-central-1 - Fn::Equals: - Ref: AWS::Region - cn-north-1 - Fn::Equals: - Ref: AWS::Region - cn-northwest-1 - Fn::Equals: - Ref: AWS::Region - eu-central-1 - Fn::Or: - Fn::Equals: - Ref: AWS::Region - eu-north-1 - Fn::Equals: - Ref: AWS::Region - eu-west-1 - Fn::Equals: - Ref: AWS::Region - eu-west-2 - Fn::Equals: - Ref: AWS::Region - eu-west-3 - Fn::Equals: - Ref: AWS::Region - me-south-1 - Fn::Equals: - Ref: AWS::Region - sa-east-1 - Fn::Equals: - Ref: AWS::Region - us-east-1 - Fn::Equals: - Ref: AWS::Region - us-east-2 - Fn::Equals: - Ref: AWS::Region - us-west-1 - Fn::Equals: - Ref: AWS::Region - us-west-2 ```
True
aws-cdk: sam local error on generated template - When I try to run `sam local start-api` with the generated template from aws-cdk, I get the following error when I make an OPTIONS request to the `/time` endpoint. Error: ``` Exception on /time [OPTIONS] Traceback (most recent call last): File "/usr/lib/python3.8/site-packages/flask/app.py", line 2447, in wsgi_app response = self.full_dispatch_request() File "/usr/lib/python3.8/site-packages/flask/app.py", line 1952, in full_dispatch_request rv = self.handle_user_exception(e) File "/usr/lib/python3.8/site-packages/flask/app.py", line 1821, in handle_user_exception reraise(exc_type, exc_value, tb) File "/usr/lib/python3.8/site-packages/flask/_compat.py", line 39, in reraise raise value File "/usr/lib/python3.8/site-packages/flask/app.py", line 1950, in full_dispatch_request rv = self.dispatch_request() File "/usr/lib/python3.8/site-packages/flask/app.py", line 1936, in dispatch_request return self.view_functions[rule.endpoint](**req.view_args) File "/usr/lib/python3.8/site-packages/samcli/local/apigw/local_apigw_service.py", line 203, in _request_handler self.lambda_runner.invoke(route.function_name, event, stdout=stdout_stream_writer, stderr=self.stderr) File "/usr/lib/python3.8/site-packages/samcli/commands/local/lib/local_lambda.py", line 84, in invoke function = self.provider.get(function_name) File "/usr/lib/python3.8/site-packages/samcli/lib/providers/sam_function_provider.py", line 65, in get raise ValueError("Function name is required") ValueError: Function name is required ``` The template file: ``` Resources: WorkTrackerBC4B47B0: Type: AWS::DynamoDB::Table Properties: KeySchema: - AttributeName: user KeyType: HASH - AttributeName: timestamp KeyType: RANGE AttributeDefinitions: - AttributeName: user AttributeType: S - AttributeName: timestamp AttributeType: S ProvisionedThroughput: ReadCapacityUnits: 5 WriteCapacityUnits: 5 TableName: WorkTracker UpdateReplacePolicy: Retain DeletionPolicy: Retain Metadata: aws:cdk:path: work-tracker/WorkTracker/Resource WorkTrackerApiA3820895: Type: AWS::ApiGateway::RestApi Properties: Name: WorkTrackerApi Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Resource WorkTrackerApiCloudWatchRole46BF1BDC: Type: AWS::IAM::Role Properties: AssumeRolePolicyDocument: Statement: - Action: sts:AssumeRole Effect: Allow Principal: Service: apigateway.amazonaws.com Version: "2012-10-17" ManagedPolicyArns: - Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - :iam::aws:policy/service-role/AmazonAPIGatewayPushToCloudWatchLogs Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/CloudWatchRole/Resource WorkTrackerApiAccountB5820D17: Type: AWS::ApiGateway::Account Properties: CloudWatchRoleArn: Fn::GetAtt: - WorkTrackerApiCloudWatchRole46BF1BDC - Arn DependsOn: - WorkTrackerApiA3820895 Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Account WorkTrackerApiDeployment667E8F39a4dede785b4b39d382183529b3eb56f9: Type: AWS::ApiGateway::Deployment Properties: RestApiId: Ref: WorkTrackerApiA3820895 Description: Automatically created by the RestApi construct DependsOn: - WorkTrackerApiOPTIONS4CF25EFB - WorkTrackerApitimeGETAC751B15 - WorkTrackerApitimeOPTIONS121CA0A4 - WorkTrackerApitimePOSTB776371B - WorkTrackerApitimeE76CB408 Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Deployment/Resource WorkTrackerApiDeploymentStageprodE0E1DD82: Type: AWS::ApiGateway::Stage Properties: RestApiId: Ref: WorkTrackerApiA3820895 DeploymentId: Ref: WorkTrackerApiDeployment667E8F39a4dede785b4b39d382183529b3eb56f9 MethodSettings: - HttpMethod: "*" ResourcePath: /* ThrottlingBurstLimit: 10 ThrottlingRateLimit: 10 StageName: prod Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/DeploymentStage.prod/Resource WorkTrackerApiOPTIONS4CF25EFB: Type: AWS::ApiGateway::Method Properties: HttpMethod: OPTIONS ResourceId: Fn::GetAtt: - WorkTrackerApiA3820895 - RootResourceId RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: NONE Integration: IntegrationResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: "'Authorization'" method.response.header.Access-Control-Allow-Origin: "'*'" method.response.header.Access-Control-Allow-Methods: "'OPTIONS,GET,PUT,POST,DELETE,PATCH,HEAD'" method.response.header.Access-Control-Allow-Credentials: "'true'" StatusCode: "204" RequestTemplates: application/json: "{ statusCode: 200 }" Type: MOCK MethodResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: true method.response.header.Access-Control-Allow-Origin: true method.response.header.Access-Control-Allow-Methods: true method.response.header.Access-Control-Allow-Credentials: true StatusCode: "204" Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/OPTIONS/Resource WorkTrackerApitimeE76CB408: Type: AWS::ApiGateway::Resource Properties: ParentId: Fn::GetAtt: - WorkTrackerApiA3820895 - RootResourceId PathPart: time RestApiId: Ref: WorkTrackerApiA3820895 Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/Resource WorkTrackerApitimeOPTIONS121CA0A4: Type: AWS::ApiGateway::Method Properties: HttpMethod: OPTIONS ResourceId: Ref: WorkTrackerApitimeE76CB408 RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: NONE Integration: IntegrationResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: "'Authorization'" method.response.header.Access-Control-Allow-Origin: "'*'" method.response.header.Access-Control-Allow-Methods: "'OPTIONS,GET,PUT,POST,DELETE,PATCH,HEAD'" method.response.header.Access-Control-Allow-Credentials: "'true'" StatusCode: "204" RequestTemplates: application/json: "{ statusCode: 200 }" Type: MOCK MethodResponses: - ResponseParameters: method.response.header.Access-Control-Allow-Headers: true method.response.header.Access-Control-Allow-Origin: true method.response.header.Access-Control-Allow-Methods: true method.response.header.Access-Control-Allow-Credentials: true StatusCode: "204" Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/OPTIONS/Resource WorkTrackerApitimeGETApiPermissionworktrackerWorkTrackerApi0D6A680AGETtime5350A293: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - / - Ref: WorkTrackerApiDeploymentStageprodE0E1DD82 - /GET/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/GET/ApiPermission.worktrackerWorkTrackerApi0D6A680A.GET..time WorkTrackerApitimeGETApiPermissionTestworktrackerWorkTrackerApi0D6A680AGETtime05C7716C: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - /test-invoke-stage/GET/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/GET/ApiPermission.Test.worktrackerWorkTrackerApi0D6A680A.GET..time WorkTrackerApitimeGETAC751B15: Type: AWS::ApiGateway::Method Properties: HttpMethod: GET ResourceId: Ref: WorkTrackerApitimeE76CB408 RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: COGNITO_USER_POOLS AuthorizerId: Ref: CognitoAuthorizer Integration: IntegrationHttpMethod: POST Type: AWS_PROXY Uri: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":apigateway:" - Ref: AWS::Region - :lambda:path/2015-03-31/functions/ - Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn - /invocations Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/GET/Resource WorkTrackerApitimePOSTApiPermissionworktrackerWorkTrackerApi0D6A680APOSTtime9B216B47: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - / - Ref: WorkTrackerApiDeploymentStageprodE0E1DD82 - /POST/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/POST/ApiPermission.worktrackerWorkTrackerApi0D6A680A.POST..time WorkTrackerApitimePOSTApiPermissionTestworktrackerWorkTrackerApi0D6A680APOSTtimeA2323B78: Type: AWS::Lambda::Permission Properties: Action: lambda:InvokeFunction FunctionName: Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn Principal: apigateway.amazonaws.com SourceArn: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":execute-api:" - Ref: AWS::Region - ":" - Ref: AWS::AccountId - ":" - Ref: WorkTrackerApiA3820895 - /test-invoke-stage/POST/time Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/POST/ApiPermission.Test.worktrackerWorkTrackerApi0D6A680A.POST..time WorkTrackerApitimePOSTB776371B: Type: AWS::ApiGateway::Method Properties: HttpMethod: POST ResourceId: Ref: WorkTrackerApitimeE76CB408 RestApiId: Ref: WorkTrackerApiA3820895 AuthorizationType: COGNITO_USER_POOLS AuthorizerId: Ref: CognitoAuthorizer Integration: IntegrationHttpMethod: POST Type: AWS_PROXY Uri: Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - ":apigateway:" - Ref: AWS::Region - :lambda:path/2015-03-31/functions/ - Fn::GetAtt: - getTimesOfDayC45A06A5 - Arn - /invocations Metadata: aws:cdk:path: work-tracker/WorkTrackerApi/Default/time/POST/Resource getTimesOfDayServiceRole7C5D69EC: Type: AWS::IAM::Role Properties: AssumeRolePolicyDocument: Statement: - Action: sts:AssumeRole Effect: Allow Principal: Service: lambda.amazonaws.com Version: "2012-10-17" ManagedPolicyArns: - Fn::Join: - "" - - "arn:" - Ref: AWS::Partition - :iam::aws:policy/service-role/AWSLambdaBasicExecutionRole Metadata: aws:cdk:path: work-tracker/getTimesOfDay/ServiceRole/Resource getTimesOfDayServiceRoleDefaultPolicy371E6EE6: Type: AWS::IAM::Policy Properties: PolicyDocument: Statement: - Action: - dynamodb:BatchGetItem - dynamodb:GetRecords - dynamodb:GetShardIterator - dynamodb:Query - dynamodb:GetItem - dynamodb:Scan Effect: Allow Resource: - Fn::GetAtt: - WorkTrackerBC4B47B0 - Arn - Ref: AWS::NoValue Version: "2012-10-17" PolicyName: getTimesOfDayServiceRoleDefaultPolicy371E6EE6 Roles: - Ref: getTimesOfDayServiceRole7C5D69EC Metadata: aws:cdk:path: work-tracker/getTimesOfDay/ServiceRole/DefaultPolicy/Resource getTimesOfDayC45A06A5: Type: AWS::Lambda::Function Properties: Code: S3Bucket: Ref: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3BucketB50D3E5E S3Key: Fn::Join: - "" - - Fn::Select: - 0 - Fn::Split: - "||" - Ref: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3VersionKey9E270B2B - Fn::Select: - 1 - Fn::Split: - "||" - Ref: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3VersionKey9E270B2B Handler: daily-time.handler Role: Fn::GetAtt: - getTimesOfDayServiceRole7C5D69EC - Arn Runtime: nodejs12.x DependsOn: - getTimesOfDayServiceRoleDefaultPolicy371E6EE6 - getTimesOfDayServiceRole7C5D69EC Metadata: aws:cdk:path: work-tracker/getTimesOfDay/Resource aws:asset:path: /home/sakkaku/Documents/workspaces/myin-work/dist/libs/daily-time aws:asset:property: Code CognitoAuthorizer: Type: AWS::ApiGateway::Authorizer Properties: RestApiId: Ref: WorkTrackerApiA3820895 Type: COGNITO_USER_POOLS IdentitySource: method.request.header.Authorization Name: MainPoolAuthorizer ProviderARNs: [] Metadata: aws:cdk:path: work-tracker/CognitoAuthorizer CDKMetadata: Type: AWS::CDK::Metadata Properties: Modules: aws-cdk=1.65.0,@aws-cdk/assets=1.65.0,@aws-cdk/aws-apigateway=1.65.0,@aws-cdk/aws-applicationautoscaling=1.65.0,@aws-cdk/aws-autoscaling-common=1.65.0,@aws-cdk/aws-cloudwatch=1.65.0,@aws-cdk/aws-codeguruprofiler=1.65.0,@aws-cdk/aws-dynamodb=1.65.0,@aws-cdk/aws-ec2=1.65.0,@aws-cdk/aws-events=1.65.0,@aws-cdk/aws-iam=1.65.0,@aws-cdk/aws-kms=1.65.0,@aws-cdk/aws-lambda=1.65.0,@aws-cdk/aws-logs=1.65.0,@aws-cdk/aws-s3=1.65.0,@aws-cdk/aws-s3-assets=1.65.0,@aws-cdk/aws-sqs=1.65.0,@aws-cdk/aws-ssm=1.65.0,@aws-cdk/cloud-assembly-schema=1.65.0,@aws-cdk/core=1.65.0,@aws-cdk/custom-resources=1.65.0,@aws-cdk/cx-api=1.65.0,@aws-cdk/region-info=1.65.0,jsii-runtime=node.js/v14.10.1 Metadata: aws:cdk:path: work-tracker/CDKMetadata/Default Condition: CDKMetadataAvailable Outputs: WorkTrackerApiEndpoint24C3B084: Value: Fn::Join: - "" - - https:// - Ref: WorkTrackerApiA3820895 - .execute-api. - Ref: AWS::Region - "." - Ref: AWS::URLSuffix - / - Ref: WorkTrackerApiDeploymentStageprodE0E1DD82 - / Parameters: AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3BucketB50D3E5E: Type: String Description: S3 bucket for asset "f4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781ba" AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baS3VersionKey9E270B2B: Type: String Description: S3 key for asset version "f4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781ba" AssetParametersf4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781baArtifactHashD8875F9F: Type: String Description: Artifact hash for asset "f4569dbf4276b2d1eec276690859fd4376007e4aaf42b4f7330451c1822781ba" Conditions: CDKMetadataAvailable: Fn::Or: - Fn::Or: - Fn::Equals: - Ref: AWS::Region - ap-east-1 - Fn::Equals: - Ref: AWS::Region - ap-northeast-1 - Fn::Equals: - Ref: AWS::Region - ap-northeast-2 - Fn::Equals: - Ref: AWS::Region - ap-south-1 - Fn::Equals: - Ref: AWS::Region - ap-southeast-1 - Fn::Equals: - Ref: AWS::Region - ap-southeast-2 - Fn::Equals: - Ref: AWS::Region - ca-central-1 - Fn::Equals: - Ref: AWS::Region - cn-north-1 - Fn::Equals: - Ref: AWS::Region - cn-northwest-1 - Fn::Equals: - Ref: AWS::Region - eu-central-1 - Fn::Or: - Fn::Equals: - Ref: AWS::Region - eu-north-1 - Fn::Equals: - Ref: AWS::Region - eu-west-1 - Fn::Equals: - Ref: AWS::Region - eu-west-2 - Fn::Equals: - Ref: AWS::Region - eu-west-3 - Fn::Equals: - Ref: AWS::Region - me-south-1 - Fn::Equals: - Ref: AWS::Region - sa-east-1 - Fn::Equals: - Ref: AWS::Region - us-east-1 - Fn::Equals: - Ref: AWS::Region - us-east-2 - Fn::Equals: - Ref: AWS::Region - us-west-1 - Fn::Equals: - Ref: AWS::Region - us-west-2 ```
main
aws cdk sam local error on generated template when i try to run sam local start api with the generated template from aws cdk i get the following error when i make an options request to the time endpoint error exception on time traceback most recent call last file usr lib site packages flask app py line in wsgi app response self full dispatch request file usr lib site packages flask app py line in full dispatch request rv self handle user exception e file usr lib site packages flask app py line in handle user exception reraise exc type exc value tb file usr lib site packages flask compat py line in reraise raise value file usr lib site packages flask app py line in full dispatch request rv self dispatch request file usr lib site packages flask app py line in dispatch request return self view functions req view args file usr lib site packages samcli local apigw local apigw service py line in request handler self lambda runner invoke route function name event stdout stdout stream writer stderr self stderr file usr lib site packages samcli commands local lib local lambda py line in invoke function self provider get function name file usr lib site packages samcli lib providers sam function provider py line in get raise valueerror function name is required valueerror function name is required the template file resources type aws dynamodb table properties keyschema attributename user keytype hash attributename timestamp keytype range attributedefinitions attributename user attributetype s attributename timestamp attributetype s provisionedthroughput readcapacityunits writecapacityunits tablename worktracker updatereplacepolicy retain deletionpolicy retain metadata aws cdk path work tracker worktracker resource type aws apigateway restapi properties name worktrackerapi metadata aws cdk path work tracker worktrackerapi resource type aws iam role properties assumerolepolicydocument statement action sts assumerole effect allow principal service apigateway amazonaws com version managedpolicyarns fn join arn ref aws partition iam aws policy service role amazonapigatewaypushtocloudwatchlogs metadata aws cdk path work tracker worktrackerapi cloudwatchrole resource type aws apigateway account properties cloudwatchrolearn fn getatt arn dependson metadata aws cdk path work tracker worktrackerapi account type aws apigateway deployment properties restapiid ref description automatically created by the restapi construct dependson metadata aws cdk path work tracker worktrackerapi deployment resource type aws apigateway stage properties restapiid ref deploymentid ref methodsettings httpmethod resourcepath throttlingburstlimit throttlingratelimit stagename prod metadata aws cdk path work tracker worktrackerapi deploymentstage prod resource type aws apigateway method properties httpmethod options resourceid fn getatt rootresourceid restapiid ref authorizationtype none integration integrationresponses responseparameters method response header access control allow headers authorization method response header access control allow origin method response header access control allow methods options get put post delete patch head method response header access control allow credentials true statuscode requesttemplates application json statuscode type mock methodresponses responseparameters method response header access control allow headers true method response header access control allow origin true method response header access control allow methods true method response header access control allow credentials true statuscode metadata aws cdk path work tracker worktrackerapi default options resource type aws apigateway resource properties parentid fn getatt rootresourceid pathpart time restapiid ref metadata aws cdk path work tracker worktrackerapi default time resource type aws apigateway method properties httpmethod options resourceid ref restapiid ref authorizationtype none integration integrationresponses responseparameters method response header access control allow headers authorization method response header access control allow origin method response header access control allow methods options get put post delete patch head method response header access control allow credentials true statuscode requesttemplates application json statuscode type mock methodresponses responseparameters method response header access control allow headers true method response header access control allow origin true method response header access control allow methods true method response header access control allow credentials true statuscode metadata aws cdk path work tracker worktrackerapi default time options resource type aws lambda permission properties action lambda invokefunction functionname fn getatt arn principal apigateway amazonaws com sourcearn fn join arn ref aws partition execute api ref aws region ref aws accountid ref ref get time metadata aws cdk path work tracker worktrackerapi default time get apipermission get time type aws lambda permission properties action lambda invokefunction functionname fn getatt arn principal apigateway amazonaws com sourcearn fn join arn ref aws partition execute api ref aws region ref aws accountid ref test invoke stage get time metadata aws cdk path work tracker worktrackerapi default time get apipermission test get time type aws apigateway method properties httpmethod get resourceid ref restapiid ref authorizationtype cognito user pools authorizerid ref cognitoauthorizer integration integrationhttpmethod post type aws proxy uri fn join arn ref aws partition apigateway ref aws region lambda path functions fn getatt arn invocations metadata aws cdk path work tracker worktrackerapi default time get resource type aws lambda permission properties action lambda invokefunction functionname fn getatt arn principal apigateway amazonaws com sourcearn fn join arn ref aws partition execute api ref aws region ref aws accountid ref ref post time metadata aws cdk path work tracker worktrackerapi default time post apipermission post time type aws lambda permission properties action lambda invokefunction functionname fn getatt arn principal apigateway amazonaws com sourcearn fn join arn ref aws partition execute api ref aws region ref aws accountid ref test invoke stage post time metadata aws cdk path work tracker worktrackerapi default time post apipermission test post time type aws apigateway method properties httpmethod post resourceid ref restapiid ref authorizationtype cognito user pools authorizerid ref cognitoauthorizer integration integrationhttpmethod post type aws proxy uri fn join arn ref aws partition apigateway ref aws region lambda path functions fn getatt arn invocations metadata aws cdk path work tracker worktrackerapi default time post resource type aws iam role properties assumerolepolicydocument statement action sts assumerole effect allow principal service lambda amazonaws com version managedpolicyarns fn join arn ref aws partition iam aws policy service role awslambdabasicexecutionrole metadata aws cdk path work tracker gettimesofday servicerole resource type aws iam policy properties policydocument statement action dynamodb batchgetitem dynamodb getrecords dynamodb getsharditerator dynamodb query dynamodb getitem dynamodb scan effect allow resource fn getatt arn ref aws novalue version policyname roles ref metadata aws cdk path work tracker gettimesofday servicerole defaultpolicy resource type aws lambda function properties code ref fn join fn select fn split ref fn select fn split ref handler daily time handler role fn getatt arn runtime x dependson metadata aws cdk path work tracker gettimesofday resource aws asset path home sakkaku documents workspaces myin work dist libs daily time aws asset property code cognitoauthorizer type aws apigateway authorizer properties restapiid ref type cognito user pools identitysource method request header authorization name mainpoolauthorizer providerarns metadata aws cdk path work tracker cognitoauthorizer cdkmetadata type aws cdk metadata properties modules aws cdk aws cdk assets aws cdk aws apigateway aws cdk aws applicationautoscaling aws cdk aws autoscaling common aws cdk aws cloudwatch aws cdk aws codeguruprofiler aws cdk aws dynamodb aws cdk aws aws cdk aws events aws cdk aws iam aws cdk aws kms aws cdk aws lambda aws cdk aws logs aws cdk aws aws cdk aws assets aws cdk aws sqs aws cdk aws ssm aws cdk cloud assembly schema aws cdk core aws cdk custom resources aws cdk cx api aws cdk region info jsii runtime node js metadata aws cdk path work tracker cdkmetadata default condition cdkmetadataavailable outputs value fn join https ref execute api ref aws region ref aws urlsuffix ref parameters type string description bucket for asset type string description key for asset version type string description artifact hash for asset conditions cdkmetadataavailable fn or fn or fn equals ref aws region ap east fn equals ref aws region ap northeast fn equals ref aws region ap northeast fn equals ref aws region ap south fn equals ref aws region ap southeast fn equals ref aws region ap southeast fn equals ref aws region ca central fn equals ref aws region cn north fn equals ref aws region cn northwest fn equals ref aws region eu central fn or fn equals ref aws region eu north fn equals ref aws region eu west fn equals ref aws region eu west fn equals ref aws region eu west fn equals ref aws region me south fn equals ref aws region sa east fn equals ref aws region us east fn equals ref aws region us east fn equals ref aws region us west fn equals ref aws region us west
1
365,785
10,797,729,419
IssuesEvent
2019-11-06 08:34:01
oceanprotocol/commons
https://api.github.com/repos/oceanprotocol/commons
opened
Can't parse description field in asset registration
bug help wanted priority:high
## Current Behavior The following description will cause errors and break the registration flow; ``` Line 1 Line 2 ``` Error trace (but just see for yourself) ``` Response message: [{"message":"'Line 1\\n\\nLine 2' does not match '^(.*)$'","path":"base/description"}] Error fetching querying metadata: Response {type: "cors", url: "https://aquarius.commons.oceanprotocol.com/api/v1/aquarius/assets/ddo", redirected: false, status: 400, ok: false, …} Logger.js:101 DDO stored Logger.js:101 error: Cannot read property 'id' of null ``` This occurs in BMW, Commons, Nile, ...! ### Possible Solution Plecos? Try the /validate endpoint in Aquarius service?
1.0
Can't parse description field in asset registration - ## Current Behavior The following description will cause errors and break the registration flow; ``` Line 1 Line 2 ``` Error trace (but just see for yourself) ``` Response message: [{"message":"'Line 1\\n\\nLine 2' does not match '^(.*)$'","path":"base/description"}] Error fetching querying metadata: Response {type: "cors", url: "https://aquarius.commons.oceanprotocol.com/api/v1/aquarius/assets/ddo", redirected: false, status: 400, ok: false, …} Logger.js:101 DDO stored Logger.js:101 error: Cannot read property 'id' of null ``` This occurs in BMW, Commons, Nile, ...! ### Possible Solution Plecos? Try the /validate endpoint in Aquarius service?
non_main
can t parse description field in asset registration current behavior the following description will cause errors and break the registration flow line line error trace but just see for yourself response message error fetching querying metadata response type cors url redirected false status ok false … logger js ddo stored logger js error cannot read property id of null this occurs in bmw commons nile possible solution plecos try the validate endpoint in aquarius service
0
141,609
21,573,150,592
IssuesEvent
2022-05-02 10:44:20
SiLeBAT/zoonotify-client
https://api.github.com/repos/SiLeBAT/zoonotify-client
closed
Change text on HomePage
ui-design/content 🎨
ZooNotify offers the possibility to query a simplified database of the pathogen detection in the food chain collected according to RL 99/2003/EU (Zoonoses Directive) and their properties, which are transmitted to the Federal Institute for Risk Assessment (BfR). With individual queries or by editing predefined sample queries, tables and graphics can be generated as required. Ändern in ZooNotify offers the possibility to query a simplified dataset of the pathogen detections in the food chain collected according to RL 99/2003/EU (Zoonoses Directive) and their properties submitted to the Federal Institute for Risk Assessment (BfR). With individual queries or by editing predefined sample queries, tables and graphs can be generated according to your own needs.
1.0
Change text on HomePage - ZooNotify offers the possibility to query a simplified database of the pathogen detection in the food chain collected according to RL 99/2003/EU (Zoonoses Directive) and their properties, which are transmitted to the Federal Institute for Risk Assessment (BfR). With individual queries or by editing predefined sample queries, tables and graphics can be generated as required. Ändern in ZooNotify offers the possibility to query a simplified dataset of the pathogen detections in the food chain collected according to RL 99/2003/EU (Zoonoses Directive) and their properties submitted to the Federal Institute for Risk Assessment (BfR). With individual queries or by editing predefined sample queries, tables and graphs can be generated according to your own needs.
non_main
change text on homepage zoonotify offers the possibility to query a simplified database of the pathogen detection in the food chain collected according to rl eu zoonoses directive and their properties which are transmitted to the federal institute for risk assessment bfr with individual queries or by editing predefined sample queries tables and graphics can be generated as required ändern in zoonotify offers the possibility to query a simplified dataset of the pathogen detections in the food chain collected according to rl eu zoonoses directive and their properties submitted to the federal institute for risk assessment bfr with individual queries or by editing predefined sample queries tables and graphs can be generated according to your own needs
0
542,614
15,863,487,633
IssuesEvent
2021-04-08 12:50:44
crcn/paperclip
https://api.github.com/repos/crcn/paperclip
closed
fix performance issue with history
area: playground effort: 3 priority: 4
For history, probably best to store patches instead of entire documents
1.0
fix performance issue with history - For history, probably best to store patches instead of entire documents
non_main
fix performance issue with history for history probably best to store patches instead of entire documents
0
363,743
25,464,423,429
IssuesEvent
2022-11-25 01:32:18
HardstuckGuild/HsBuildCodes
https://api.github.com/repos/HardstuckGuild/HsBuildCodes
opened
Add readme into the nuget package
documentation good first issue low priority
The readme is even already there, its just not included properly.
1.0
Add readme into the nuget package - The readme is even already there, its just not included properly.
non_main
add readme into the nuget package the readme is even already there its just not included properly
0
1,620
6,572,644,548
IssuesEvent
2017-09-11 04:01:43
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
s3_bucket error: Your previous request to create the named bucket succeeded and you already own it
affects_2.1 aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME s3_bucket ##### ANSIBLE VERSION ansible 2.1.2.0 ansible 2.2.0.0 ##### OS / ENVIRONMENT Linux Debian ##### SUMMARY Trying to run s3_bucket task when the bucket already exists results in an "Your previous request to create the named bucket succeeded and you already own it" error. ##### STEPS TO REPRODUCE Run more than once the following task: ``` - s3_bucket: name: my-test-bucket-12 state: present region: eu-central-1 versioning: yes delegate_to: management-server run_once: true ``` ##### EXPECTED RESULTS No error should be returned if the bucket already exists and is owned by the AWS account trying to create it. The "BucketAlreadyOwnedByYou" error returned by Boto can probably be ignored. However, the "versioning" setting for the bucket should be updated if necessary. It's possible that this error is returned only for buckets in regions other that US Standard: "BucketAlreadyOwnedByYou errors will only be returned outside of the US Standard region. Inside the US Standard region (i.e. when you don't specify a location constraint), attempting to recreate a bucket you already own will succeed." http://stackoverflow.com/a/16561451/226640 We use the `eu-central-1` region. ##### ACTUAL RESULTS ``` fatal: [dev -> management-server]: FAILED! => {"changed": false, "failed": true, "msg": "Your previous request to create the named bucket succeeded and you already own it."} ```
True
s3_bucket error: Your previous request to create the named bucket succeeded and you already own it - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME s3_bucket ##### ANSIBLE VERSION ansible 2.1.2.0 ansible 2.2.0.0 ##### OS / ENVIRONMENT Linux Debian ##### SUMMARY Trying to run s3_bucket task when the bucket already exists results in an "Your previous request to create the named bucket succeeded and you already own it" error. ##### STEPS TO REPRODUCE Run more than once the following task: ``` - s3_bucket: name: my-test-bucket-12 state: present region: eu-central-1 versioning: yes delegate_to: management-server run_once: true ``` ##### EXPECTED RESULTS No error should be returned if the bucket already exists and is owned by the AWS account trying to create it. The "BucketAlreadyOwnedByYou" error returned by Boto can probably be ignored. However, the "versioning" setting for the bucket should be updated if necessary. It's possible that this error is returned only for buckets in regions other that US Standard: "BucketAlreadyOwnedByYou errors will only be returned outside of the US Standard region. Inside the US Standard region (i.e. when you don't specify a location constraint), attempting to recreate a bucket you already own will succeed." http://stackoverflow.com/a/16561451/226640 We use the `eu-central-1` region. ##### ACTUAL RESULTS ``` fatal: [dev -> management-server]: FAILED! => {"changed": false, "failed": true, "msg": "Your previous request to create the named bucket succeeded and you already own it."} ```
main
bucket error your previous request to create the named bucket succeeded and you already own it issue type bug report component name bucket ansible version ansible ansible os environment linux debian summary trying to run bucket task when the bucket already exists results in an your previous request to create the named bucket succeeded and you already own it error steps to reproduce run more than once the following task bucket name my test bucket state present region eu central versioning yes delegate to management server run once true expected results no error should be returned if the bucket already exists and is owned by the aws account trying to create it the bucketalreadyownedbyyou error returned by boto can probably be ignored however the versioning setting for the bucket should be updated if necessary it s possible that this error is returned only for buckets in regions other that us standard bucketalreadyownedbyyou errors will only be returned outside of the us standard region inside the us standard region i e when you don t specify a location constraint attempting to recreate a bucket you already own will succeed we use the eu central region actual results fatal failed changed false failed true msg your previous request to create the named bucket succeeded and you already own it
1
8,285
2,611,486,159
IssuesEvent
2015-02-27 05:27:25
chrsmith/switchlist
https://api.github.com/repos/chrsmith/switchlist
closed
Changing list of stops doesn't change text version in Train panel.
auto-migrated Priority-Medium Type-Defect
``` Open your favorite layout, select the Trains tab, and select one train. Press "Change stops" to modify the list of stops. Add a new first or second station, and close the dialog. Notice the main Trains tab didn't update the list of stops in its text form. ``` Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 21 Mar 2012 at 4:43
1.0
Changing list of stops doesn't change text version in Train panel. - ``` Open your favorite layout, select the Trains tab, and select one train. Press "Change stops" to modify the list of stops. Add a new first or second station, and close the dialog. Notice the main Trains tab didn't update the list of stops in its text form. ``` Original issue reported on code.google.com by `rwbowdi...@gmail.com` on 21 Mar 2012 at 4:43
non_main
changing list of stops doesn t change text version in train panel open your favorite layout select the trains tab and select one train press change stops to modify the list of stops add a new first or second station and close the dialog notice the main trains tab didn t update the list of stops in its text form original issue reported on code google com by rwbowdi gmail com on mar at
0
104,407
8,972,308,133
IssuesEvent
2019-01-29 17:56:27
nasa-gibs/worldview
https://api.github.com/repos/nasa-gibs/worldview
closed
When you remove layers that have been added by an event, and load event again, the layers do not load in the layer list
bug testing
**Describe the bug** When you remove layers that have been added by an event, and load event again, the layers do not load in the layer list (occurs on production and develop). **To Reproduce** Steps to reproduce the behavior: 1. Go to the Events tab 2. Click on Tropical Cyclone Penny 3. Go to Layer List and see IMERG Rain Rate and Nighttime imagery loaded on list 4. Remove IMERG Rain rate and Nighttime imagery from Layer List 5. Go to Events tab and click twice on Tropical Cyclone Penny to reactivate it 6. See IMERG Rain rate load in map, but it does not load in the Layer List 7. Remove "Place Labels" from Layer List and see IMERG Rain Rate and Nighttime imagery appear in the Layer List. **Expected behavior** The layers should re-load in the layer list after re-clicking on an event even if you have previously removed them from the layer list. **Desktop (please complete the following information):** - OS: Mac OSX - Browser: Chrome and Firefox - Version 71.0.3578.98 (Official Build) (64-bit); 60.3.0esr (64-bit)
1.0
When you remove layers that have been added by an event, and load event again, the layers do not load in the layer list - **Describe the bug** When you remove layers that have been added by an event, and load event again, the layers do not load in the layer list (occurs on production and develop). **To Reproduce** Steps to reproduce the behavior: 1. Go to the Events tab 2. Click on Tropical Cyclone Penny 3. Go to Layer List and see IMERG Rain Rate and Nighttime imagery loaded on list 4. Remove IMERG Rain rate and Nighttime imagery from Layer List 5. Go to Events tab and click twice on Tropical Cyclone Penny to reactivate it 6. See IMERG Rain rate load in map, but it does not load in the Layer List 7. Remove "Place Labels" from Layer List and see IMERG Rain Rate and Nighttime imagery appear in the Layer List. **Expected behavior** The layers should re-load in the layer list after re-clicking on an event even if you have previously removed them from the layer list. **Desktop (please complete the following information):** - OS: Mac OSX - Browser: Chrome and Firefox - Version 71.0.3578.98 (Official Build) (64-bit); 60.3.0esr (64-bit)
non_main
when you remove layers that have been added by an event and load event again the layers do not load in the layer list describe the bug when you remove layers that have been added by an event and load event again the layers do not load in the layer list occurs on production and develop to reproduce steps to reproduce the behavior go to the events tab click on tropical cyclone penny go to layer list and see imerg rain rate and nighttime imagery loaded on list remove imerg rain rate and nighttime imagery from layer list go to events tab and click twice on tropical cyclone penny to reactivate it see imerg rain rate load in map but it does not load in the layer list remove place labels from layer list and see imerg rain rate and nighttime imagery appear in the layer list expected behavior the layers should re load in the layer list after re clicking on an event even if you have previously removed them from the layer list desktop please complete the following information os mac osx browser chrome and firefox version official build bit bit
0
3,914
17,481,969,725
IssuesEvent
2021-08-09 04:58:52
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
opened
Application to join: Allsite
Maintainer application
Hello and welcome to the contrib application process! We're happy to have you :) ## Please note these 3 requirements for new contrib projects: - [ ] Include a README.md file containing license and maintainer information. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md - [ ] Include a LICENSE.txt file. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt. - [ ] If porting a Drupal 7 project, Maintain the Git history from Drupal. ## Please provide the following information: **The name of your module, theme, or layout** <!-- example: Forum Access --> content_limiter dark_seven **(Optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project** <!-- example: https://www.drupal.org/project/forum_access/issues/3070491 --> **Post a link to your new Backdrop project under your own GitHub account (option #1)** <!-- example: https://github.com/jenlampton/forum_access --> https://github.com/allsite/content_limiter **OR (option #2) If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits** **OR (option #3) If you do not intend to contribute code, but would like to update documentation, manage issue queues, etc, please tag an existing contrib group member so they can post their recommendation** <!-- example: @jenlampton --> **If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)** YES/no YES **If you have chosen option #3 above, do you agree to undergo this same maintainer application process again, should you decide to contribute code in the future?** YES/no <!-- Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. --> <!-- Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
True
Application to join: Allsite - Hello and welcome to the contrib application process! We're happy to have you :) ## Please note these 3 requirements for new contrib projects: - [ ] Include a README.md file containing license and maintainer information. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/README.md - [ ] Include a LICENSE.txt file. You can use this example: https://raw.githubusercontent.com/backdrop-ops/contrib/master/examples/LICENSE.txt. - [ ] If porting a Drupal 7 project, Maintain the Git history from Drupal. ## Please provide the following information: **The name of your module, theme, or layout** <!-- example: Forum Access --> content_limiter dark_seven **(Optional) Post a link here to an issue in the drupal.org queue notifying the Drupal 7 maintainers that you are working on a Backdrop port of their project** <!-- example: https://www.drupal.org/project/forum_access/issues/3070491 --> **Post a link to your new Backdrop project under your own GitHub account (option #1)** <!-- example: https://github.com/jenlampton/forum_access --> https://github.com/allsite/content_limiter **OR (option #2) If you have already contributed code to Backdrop core or contrib projects, please provide 1-3 links to pull requests or commits** **OR (option #3) If you do not intend to contribute code, but would like to update documentation, manage issue queues, etc, please tag an existing contrib group member so they can post their recommendation** <!-- example: @jenlampton --> **If you have chosen option #2 or #1 above, do you agree to the [Backdrop Contributed Project Agreement](https://github.com/backdrop-ops/contrib#backdrop-contributed-project-agreement)** YES/no YES **If you have chosen option #3 above, do you agree to undergo this same maintainer application process again, should you decide to contribute code in the future?** YES/no <!-- Once we have a chance to review your project, we will check for the 3 requirements at the top of this issue. If those requirements are met, you will be invited to the @backdrop-contrib group. At that point you will be able to transfer the project. --> <!-- Please note that we may also include additional feedback in the code review, but anything else is only intended to be helpful, and is NOT a requirement for joining the contrib group. -->
main
application to join allsite hello and welcome to the contrib application process we re happy to have you please note these requirements for new contrib projects include a readme md file containing license and maintainer information you can use this example include a license txt file you can use this example if porting a drupal project maintain the git history from drupal please provide the following information the name of your module theme or layout content limiter dark seven optional post a link here to an issue in the drupal org queue notifying the drupal maintainers that you are working on a backdrop port of their project post a link to your new backdrop project under your own github account option or option if you have already contributed code to backdrop core or contrib projects please provide links to pull requests or commits or option if you do not intend to contribute code but would like to update documentation manage issue queues etc please tag an existing contrib group member so they can post their recommendation if you have chosen option or above do you agree to the yes no yes if you have chosen option above do you agree to undergo this same maintainer application process again should you decide to contribute code in the future yes no
1
1,709
6,574,437,930
IssuesEvent
2017-09-11 12:54:00
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Add support for upgrading packages to package module
affects_2.1 feature_idea waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME package ##### ANSIBLE VERSION 2.1.2.0 ##### OS / ENVIRONMENT N/A ##### SUMMARY As you know, yum can be used as follows to upgrade all installed packages: > yum: name=* state=latest That doesn't work for apt for example. That module requires using an upgrade parameter. If we try to apply the same logic to package module as in yum, we would write something like: > package: name=* state=latest However this will try to install all packages available in the repository. What I suggest is to add a parameter to package module that translates into an 'upgrade' for any of the packaging systems it supports.
True
Add support for upgrading packages to package module - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME package ##### ANSIBLE VERSION 2.1.2.0 ##### OS / ENVIRONMENT N/A ##### SUMMARY As you know, yum can be used as follows to upgrade all installed packages: > yum: name=* state=latest That doesn't work for apt for example. That module requires using an upgrade parameter. If we try to apply the same logic to package module as in yum, we would write something like: > package: name=* state=latest However this will try to install all packages available in the repository. What I suggest is to add a parameter to package module that translates into an 'upgrade' for any of the packaging systems it supports.
main
add support for upgrading packages to package module issue type feature idea component name package ansible version os environment n a summary as you know yum can be used as follows to upgrade all installed packages yum name state latest that doesn t work for apt for example that module requires using an upgrade parameter if we try to apply the same logic to package module as in yum we would write something like package name state latest however this will try to install all packages available in the repository what i suggest is to add a parameter to package module that translates into an upgrade for any of the packaging systems it supports
1
344,172
10,340,479,676
IssuesEvent
2019-09-03 22:06:15
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
opened
SearchBar shouldn't have internal state
DataTable Docs: not needed Effort: small Priority: High Refactor
## The suggestion `SearchBar` component from `react-native-ui-components` uses internal state. This state cannot be effected outside of the component. **Why?** You can't, for example, programmatically reset the search text. ## Describe alternatives you've considered Not resetting the search text? ## Additional context I think we should just create a `SearchBar` component in our own repo. It doesn't seem the project is used by anyone?
1.0
SearchBar shouldn't have internal state - ## The suggestion `SearchBar` component from `react-native-ui-components` uses internal state. This state cannot be effected outside of the component. **Why?** You can't, for example, programmatically reset the search text. ## Describe alternatives you've considered Not resetting the search text? ## Additional context I think we should just create a `SearchBar` component in our own repo. It doesn't seem the project is used by anyone?
non_main
searchbar shouldn t have internal state the suggestion searchbar component from react native ui components uses internal state this state cannot be effected outside of the component why you can t for example programmatically reset the search text describe alternatives you ve considered not resetting the search text additional context i think we should just create a searchbar component in our own repo it doesn t seem the project is used by anyone
0
5,769
30,577,524,872
IssuesEvent
2023-07-21 07:04:14
danswer-ai/danswer
https://api.github.com/repos/danswer-ai/danswer
closed
Notion Connector
good first issue maintainer approved
Notion gained tons of popularity as a team wiki tool. Would be great to have a Notion connecter a la [Notion Loader](https://llama-hub-ui.vercel.app/l/notion)
True
Notion Connector - Notion gained tons of popularity as a team wiki tool. Would be great to have a Notion connecter a la [Notion Loader](https://llama-hub-ui.vercel.app/l/notion)
main
notion connector notion gained tons of popularity as a team wiki tool would be great to have a notion connecter a la
1
1,016
4,803,535,415
IssuesEvent
2016-11-02 10:26:42
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
apache2_module fails for libapache2-mod-proxy-uwsgi
affects_2.2 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME apache2_module ##### ANSIBLE VERSION ``` $ ./ansible --version ansible 2.2.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION None ##### OS / ENVIRONMENT Ubuntu 14.04 ##### SUMMARY A playbook that was working with ansible 2.1.2.0 started failing when 2.2.0.0 was released. The playbook: 1) apt installs a bunch of packages, including apache2, libapache2-mod-proxy-html, and libapache2-mod-proxy-uwsgi, uwsgi, etc. 2) Uses apache2_module to enable proxy_http #### Analysis The apache2_module module was changed to use `apache2ctl -M` to list modules in 2.2.0.0, which fails with the playbook due to an oddity of the Ubuntu packages: - libapache2-mod-proxy-html : installs the modules disabled. - libapache2-mod-proxy-uwsgi : installs the module enabled. But, mod_proxy_uwsgi requires mod_proxy_http to be enabled, and since mod_proxy_uwsgi is enabled but mod_proxy_http is disabled, `apache2ctl -M` fails with a confusing error (see output below) See https://github.com/ansible/ansible-modules-core/pull/2417 for the change that added use of `apache2ctl -M`. A potential fix is to provide special handling for mod_proxy_uwsgi so that it doesn't use `apache2ctl -M` or something. ##### STEPS TO REPRODUCE ``` - name: Install packages become: true apt: name={{ item }} with_items: - apache2 - git - libapache2-mod-proxy-html - libapache2-mod-proxy-uwsgi - python-dev - python-pip - python-virtualenv - uwsgi-emperor - uwsgi-plugin-python - name: Enable modules for mod_proxy_uwsgi become: true apache2_module: name={{ item }} with_items: - proxy_http notify: - Restart httpd ``` ##### EXPECTED RESULTS Expected the old playbook to work. ##### ACTUAL RESULTS ``` failed: [default] (item=proxy_http) => { "failed": true, "invocation": { "module_args": { "force": false, "name": "proxy_http", "state": "present" }, "module_name": "apache2_module" }, "item": "proxy_http", "msg": "Error executing /usr/sbin/apache2ctl: apache2: Syntax error on line 140 of /etc/apache2/apache2.conf: Syntax error on line 1 of /etc/apache2/mods-enabled/proxy_uwsgi.load: Cannot load /usr/lib/apache2/modules/mod_proxy_uwsgi.so into server: /usr/lib/apache2/modules/mod_proxy_uwsgi.so: undefined symbol: ap_proxy_backend_broke\n" } ``` As you can see the error isn't very useful as far as figuring out the problem. ##### Workaround The following steps worked around the issue: ``` - name: Install packages become: true apt: name={{ item }} with_items: - apache2 - git - libapache2-mod-proxy-html - python-dev - python-pip - python-virtualenv - uwsgi-emperor - uwsgi-plugin-python - name: Enable modules for mod_proxy_uwsgi become: true apache2_module: name={{ item }} with_items: - proxy_http notify: - Restart httpd - name: Install mod-proxy-uwsgi become: true apt: name={{ item }} with_items: - libapache2-mod-proxy-uwsgi notify: - Restart httpd ```
True
apache2_module fails for libapache2-mod-proxy-uwsgi - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME apache2_module ##### ANSIBLE VERSION ``` $ ./ansible --version ansible 2.2.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION None ##### OS / ENVIRONMENT Ubuntu 14.04 ##### SUMMARY A playbook that was working with ansible 2.1.2.0 started failing when 2.2.0.0 was released. The playbook: 1) apt installs a bunch of packages, including apache2, libapache2-mod-proxy-html, and libapache2-mod-proxy-uwsgi, uwsgi, etc. 2) Uses apache2_module to enable proxy_http #### Analysis The apache2_module module was changed to use `apache2ctl -M` to list modules in 2.2.0.0, which fails with the playbook due to an oddity of the Ubuntu packages: - libapache2-mod-proxy-html : installs the modules disabled. - libapache2-mod-proxy-uwsgi : installs the module enabled. But, mod_proxy_uwsgi requires mod_proxy_http to be enabled, and since mod_proxy_uwsgi is enabled but mod_proxy_http is disabled, `apache2ctl -M` fails with a confusing error (see output below) See https://github.com/ansible/ansible-modules-core/pull/2417 for the change that added use of `apache2ctl -M`. A potential fix is to provide special handling for mod_proxy_uwsgi so that it doesn't use `apache2ctl -M` or something. ##### STEPS TO REPRODUCE ``` - name: Install packages become: true apt: name={{ item }} with_items: - apache2 - git - libapache2-mod-proxy-html - libapache2-mod-proxy-uwsgi - python-dev - python-pip - python-virtualenv - uwsgi-emperor - uwsgi-plugin-python - name: Enable modules for mod_proxy_uwsgi become: true apache2_module: name={{ item }} with_items: - proxy_http notify: - Restart httpd ``` ##### EXPECTED RESULTS Expected the old playbook to work. ##### ACTUAL RESULTS ``` failed: [default] (item=proxy_http) => { "failed": true, "invocation": { "module_args": { "force": false, "name": "proxy_http", "state": "present" }, "module_name": "apache2_module" }, "item": "proxy_http", "msg": "Error executing /usr/sbin/apache2ctl: apache2: Syntax error on line 140 of /etc/apache2/apache2.conf: Syntax error on line 1 of /etc/apache2/mods-enabled/proxy_uwsgi.load: Cannot load /usr/lib/apache2/modules/mod_proxy_uwsgi.so into server: /usr/lib/apache2/modules/mod_proxy_uwsgi.so: undefined symbol: ap_proxy_backend_broke\n" } ``` As you can see the error isn't very useful as far as figuring out the problem. ##### Workaround The following steps worked around the issue: ``` - name: Install packages become: true apt: name={{ item }} with_items: - apache2 - git - libapache2-mod-proxy-html - python-dev - python-pip - python-virtualenv - uwsgi-emperor - uwsgi-plugin-python - name: Enable modules for mod_proxy_uwsgi become: true apache2_module: name={{ item }} with_items: - proxy_http notify: - Restart httpd - name: Install mod-proxy-uwsgi become: true apt: name={{ item }} with_items: - libapache2-mod-proxy-uwsgi notify: - Restart httpd ```
main
module fails for mod proxy uwsgi issue type bug report component name module ansible version ansible version ansible config file configured module search path default w o overrides configuration none os environment ubuntu summary a playbook that was working with ansible started failing when was released the playbook apt installs a bunch of packages including mod proxy html and mod proxy uwsgi uwsgi etc uses module to enable proxy http analysis the module module was changed to use m to list modules in which fails with the playbook due to an oddity of the ubuntu packages mod proxy html installs the modules disabled mod proxy uwsgi installs the module enabled but mod proxy uwsgi requires mod proxy http to be enabled and since mod proxy uwsgi is enabled but mod proxy http is disabled m fails with a confusing error see output below see for the change that added use of m a potential fix is to provide special handling for mod proxy uwsgi so that it doesn t use m or something steps to reproduce name install packages become true apt name item with items git mod proxy html mod proxy uwsgi python dev python pip python virtualenv uwsgi emperor uwsgi plugin python name enable modules for mod proxy uwsgi become true module name item with items proxy http notify restart httpd expected results expected the old playbook to work actual results failed item proxy http failed true invocation module args force false name proxy http state present module name module item proxy http msg error executing usr sbin syntax error on line of etc conf syntax error on line of etc mods enabled proxy uwsgi load cannot load usr lib modules mod proxy uwsgi so into server usr lib modules mod proxy uwsgi so undefined symbol ap proxy backend broke n as you can see the error isn t very useful as far as figuring out the problem workaround the following steps worked around the issue name install packages become true apt name item with items git mod proxy html python dev python pip python virtualenv uwsgi emperor uwsgi plugin python name enable modules for mod proxy uwsgi become true module name item with items proxy http notify restart httpd name install mod proxy uwsgi become true apt name item with items mod proxy uwsgi notify restart httpd
1
934
4,644,120,046
IssuesEvent
2016-09-30 15:26:39
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2 module hangs if spot request fails
affects_2.1 aws bug_report cloud P2 waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2 module ##### ANSIBLE VERSION ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### CONFIGURATION ANSIBLE_HOSTS=/etc/ansible/ec2.py ##### OS / ENVIRONMENT Ubuntu on Windows 10 ##### SUMMARY When an ec2 spot bid fails, the ansible console hangs (making it seem like it's just taking a while to providion). ##### STEPS TO REPRODUCE This task failed due to a bad volume name (/dev/sda): ```yml - name: Provision a set of instances ec2: spot_price: 0.65 spot_wait_timeout: 600 key_name: ... region: us-east-1 group_id: ... instance_type: g2.2xlarge image: ami-d05e75b8 wait: true exact_count: 1 count_tag: Name: InstanceTag instance_tags: Name: InstanceTag vpc_subnet_id: ... assign_public_ip: yes zone: us-east-1d volumes: - device_name: /dev/sda volume_type: gp2 volume_size: 20 ``` ##### EXPECTED RESULTS I expected ansible to show me an error when the request failed ##### ACTUAL RESULTS Ansible hangs
True
ec2 module hangs if spot request fails - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2 module ##### ANSIBLE VERSION ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### CONFIGURATION ANSIBLE_HOSTS=/etc/ansible/ec2.py ##### OS / ENVIRONMENT Ubuntu on Windows 10 ##### SUMMARY When an ec2 spot bid fails, the ansible console hangs (making it seem like it's just taking a while to providion). ##### STEPS TO REPRODUCE This task failed due to a bad volume name (/dev/sda): ```yml - name: Provision a set of instances ec2: spot_price: 0.65 spot_wait_timeout: 600 key_name: ... region: us-east-1 group_id: ... instance_type: g2.2xlarge image: ami-d05e75b8 wait: true exact_count: 1 count_tag: Name: InstanceTag instance_tags: Name: InstanceTag vpc_subnet_id: ... assign_public_ip: yes zone: us-east-1d volumes: - device_name: /dev/sda volume_type: gp2 volume_size: 20 ``` ##### EXPECTED RESULTS I expected ansible to show me an error when the request failed ##### ACTUAL RESULTS Ansible hangs
main
module hangs if spot request fails issue type bug report component name module ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration ansible hosts etc ansible py os environment ubuntu on windows summary when an spot bid fails the ansible console hangs making it seem like it s just taking a while to providion steps to reproduce this task failed due to a bad volume name dev sda yml name provision a set of instances spot price spot wait timeout key name region us east group id instance type image ami wait true exact count count tag name instancetag instance tags name instancetag vpc subnet id assign public ip yes zone us east volumes device name dev sda volume type volume size expected results i expected ansible to show me an error when the request failed actual results ansible hangs
1
4,616
23,912,587,254
IssuesEvent
2022-09-09 09:35:44
OSSPhilippines/pinoy-made
https://api.github.com/repos/OSSPhilippines/pinoy-made
closed
Minimum of 20 stars is not being followed
in review maintainers only
As 5th in the acceptance criteria: 5. To maintain high quality projects in the list, your project must have at least 20 stars. I noticed this is not being followed on latest PRs that were merged. https://github.com/OSSPhilippines/pinoy-made/blob/23189193d4535ff0309aad8ace26b3737e18b751/CONTRIBUTING.md?plain=1#L21
True
Minimum of 20 stars is not being followed - As 5th in the acceptance criteria: 5. To maintain high quality projects in the list, your project must have at least 20 stars. I noticed this is not being followed on latest PRs that were merged. https://github.com/OSSPhilippines/pinoy-made/blob/23189193d4535ff0309aad8ace26b3737e18b751/CONTRIBUTING.md?plain=1#L21
main
minimum of stars is not being followed as in the acceptance criteria to maintain high quality projects in the list your project must have at least stars i noticed this is not being followed on latest prs that were merged
1
99,028
4,045,240,996
IssuesEvent
2016-05-21 21:33:11
minj/foxtrick
https://api.github.com/repos/minj/foxtrick
closed
Foxtrick shows wrong sublevel name in MatchSimulator
bug MatchOrder Priority-Medium started
This makes HatStats not match the sublevel name. --- **From:** Schumi- **PostID:** [16751289.262](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16751289%26n%3D262%26v%3D0) **To:** LA-MJ **Re:** [16751289.1](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16751289%26n%3D1%26v%3D0) **Datetime:** 2016-05-13 19:29 **Message:** If for the NT, I use the Predictor, the numbers for HatStats don't add up to what HT says. The HatStats number is 3 too high for every area. So Attack is 3 too high, Midfield is 3 too higher and Defense is 3 too high..?
1.0
Foxtrick shows wrong sublevel name in MatchSimulator - This makes HatStats not match the sublevel name. --- **From:** Schumi- **PostID:** [16751289.262](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16751289%26n%3D262%26v%3D0) **To:** LA-MJ **Re:** [16751289.1](https://www.hattrick.org/goto.ashx?path=%2FForum%2FRead.aspx%3Ft%3D16751289%26n%3D1%26v%3D0) **Datetime:** 2016-05-13 19:29 **Message:** If for the NT, I use the Predictor, the numbers for HatStats don't add up to what HT says. The HatStats number is 3 too high for every area. So Attack is 3 too high, Midfield is 3 too higher and Defense is 3 too high..?
non_main
foxtrick shows wrong sublevel name in matchsimulator this makes hatstats not match the sublevel name from schumi postid to la mj re datetime message if for the nt i use the predictor the numbers for hatstats don t add up to what ht says the hatstats number is too high for every area so attack is too high midfield is too higher and defense is too high
0
252,745
21,628,887,251
IssuesEvent
2022-05-05 07:34:02
dusk-network/dusk-blockchain
https://api.github.com/repos/dusk-network/dusk-blockchain
opened
Mempool should discard any transaction with repeated nullifier
mark:testnet
**Describe what you want implemented** On transaction retrieval, Mempool should check if any of transaction nullifiers is already used by another transaction in mempool state. The check should be executed after a successful rusk.Preverify call. **Describe "Why" this is needed** This change will allow a block generator for passing transactions without repeated nullifiers. Additionally, mempool will not need anymore to run `discardUsedNullifiers` at any block acceptance.
1.0
Mempool should discard any transaction with repeated nullifier - **Describe what you want implemented** On transaction retrieval, Mempool should check if any of transaction nullifiers is already used by another transaction in mempool state. The check should be executed after a successful rusk.Preverify call. **Describe "Why" this is needed** This change will allow a block generator for passing transactions without repeated nullifiers. Additionally, mempool will not need anymore to run `discardUsedNullifiers` at any block acceptance.
non_main
mempool should discard any transaction with repeated nullifier describe what you want implemented on transaction retrieval mempool should check if any of transaction nullifiers is already used by another transaction in mempool state the check should be executed after a successful rusk preverify call describe why this is needed this change will allow a block generator for passing transactions without repeated nullifiers additionally mempool will not need anymore to run discardusednullifiers at any block acceptance
0
38
2,585,564,678
IssuesEvent
2015-02-17 00:59:30
spyder-ide/spyder
https://api.github.com/repos/spyder-ide/spyder
closed
Optional feature handling
1 star enhancement imported Logic Maintainability Usability
_From [techtonik@gmail.com](https://code.google.com/u/techtonik@gmail.com/) on 2012-03-07T02:32:19Z_ What steps will reproduce the problem? 1.Spyder has a lot of features 2.Some features are disabled in preferences, some are missing requirements 3.I want to see a full list of features, and if they are disabled - know the reason why What is the expected output? What do you see instead? Expected is a dynamically compiled list of distinctive optional features and their status. Please use labels and text to provide additional information. This will require: 1. Features registry 2. Hardcoded core features registration for clarity and static analysis (so that core won't depend on dynamic behavior bugs) 3. Dynamic feature registration for plugins, so that they can register their features too 4. Status list widget _Original issue: http://code.google.com/p/spyderlib/issues/detail?id=964_
True
Optional feature handling - _From [techtonik@gmail.com](https://code.google.com/u/techtonik@gmail.com/) on 2012-03-07T02:32:19Z_ What steps will reproduce the problem? 1.Spyder has a lot of features 2.Some features are disabled in preferences, some are missing requirements 3.I want to see a full list of features, and if they are disabled - know the reason why What is the expected output? What do you see instead? Expected is a dynamically compiled list of distinctive optional features and their status. Please use labels and text to provide additional information. This will require: 1. Features registry 2. Hardcoded core features registration for clarity and static analysis (so that core won't depend on dynamic behavior bugs) 3. Dynamic feature registration for plugins, so that they can register their features too 4. Status list widget _Original issue: http://code.google.com/p/spyderlib/issues/detail?id=964_
main
optional feature handling from on what steps will reproduce the problem spyder has a lot of features some features are disabled in preferences some are missing requirements i want to see a full list of features and if they are disabled know the reason why what is the expected output what do you see instead expected is a dynamically compiled list of distinctive optional features and their status please use labels and text to provide additional information this will require features registry hardcoded core features registration for clarity and static analysis so that core won t depend on dynamic behavior bugs dynamic feature registration for plugins so that they can register their features too status list widget original issue
1
30,281
2,723,359,492
IssuesEvent
2015-04-14 12:03:28
OCHA-DAP/hdx-ckan
https://api.github.com/repos/OCHA-DAP/hdx-ckan
closed
Create/Edit Org Page: 3W map settings
Custom org page Priority-High
Currently, the user would need to know cryptic values for X, Y and zoom level to get the map to display. We need to understand how these setting work (from @SimonbJohnson ) and figure out a way to make it easy for the data team to configure the map.
1.0
Create/Edit Org Page: 3W map settings - Currently, the user would need to know cryptic values for X, Y and zoom level to get the map to display. We need to understand how these setting work (from @SimonbJohnson ) and figure out a way to make it easy for the data team to configure the map.
non_main
create edit org page map settings currently the user would need to know cryptic values for x y and zoom level to get the map to display we need to understand how these setting work from simonbjohnson and figure out a way to make it easy for the data team to configure the map
0
3,158
12,218,025,853
IssuesEvent
2020-05-01 18:26:56
ipfs/docs
https://api.github.com/repos/ipfs/docs
opened
IPFS and encryption
P2 dif/hard effort/days help wanted kind/enhancement need/maintainers-input status/ready topic/docs
IPFS deals with encryption in interesting ways. We need a general overview of how IPFS manages encryption, including: - Where data is and is not encrypted. - The difference between transport encryption and content encryption. - Why IPFS doesn't include encryption _as rest_. - Examples of projects that are using IPFS with encrypted data. This overview should be similar in size and layout to the [IPFS Gateway overview](https://docs-beta.ipfs.io/concepts/ipfs-gateway/). This bounty will likely lead to further investigation and research bounties regarding IPFS encryption.
True
IPFS and encryption - IPFS deals with encryption in interesting ways. We need a general overview of how IPFS manages encryption, including: - Where data is and is not encrypted. - The difference between transport encryption and content encryption. - Why IPFS doesn't include encryption _as rest_. - Examples of projects that are using IPFS with encrypted data. This overview should be similar in size and layout to the [IPFS Gateway overview](https://docs-beta.ipfs.io/concepts/ipfs-gateway/). This bounty will likely lead to further investigation and research bounties regarding IPFS encryption.
main
ipfs and encryption ipfs deals with encryption in interesting ways we need a general overview of how ipfs manages encryption including where data is and is not encrypted the difference between transport encryption and content encryption why ipfs doesn t include encryption as rest examples of projects that are using ipfs with encrypted data this overview should be similar in size and layout to the this bounty will likely lead to further investigation and research bounties regarding ipfs encryption
1
156,712
12,335,789,969
IssuesEvent
2020-05-14 12:35:28
openethereum/openethereum
https://api.github.com/repos/openethereum/openethereum
closed
[json-tests] stRevertTest blockchain test failures
F4-tests 💻 P2-asap 🌊
After https://github.com/paritytech/parity-ethereum/pull/11054 the following JSON tests fail in [`stRevertTest`](https://github.com/ethereum/tests/tree/b64fd32dd897285a3884f68b40cee358a4b3c54f/LegacyTests/Constantinople/BlockchainTests/GeneralStateTests/stRevertTest) and are added to the list of skipped tests in `ethcore/res/ethereum/tests-issues/currents.json`. ``` "RevertPrecompiledTouch_d0g0v0_Byzantium", "RevertPrecompiledTouch_d0g0v0_Constantinople", "RevertPrecompiledTouch_d0g0v0_ConstantinopleFix", "RevertPrecompiledTouch_d0g0v0_EIP158", "RevertPrecompiledTouch_d3g0v0_ConstantinopleFix", "RevertPrecompiledTouchCC_d0g0v0_Byzantium", "RevertPrecompiledTouchCC_d0g0v0_Constantinople", "RevertPrecompiledTouchCC_d0g0v0_EIP158", "RevertPrecompiledTouchDC_d0g0v0_Byzantium", "RevertPrecompiledTouchDC_d0g0v0_Constantinople", "RevertPrecompiledTouchDC_d0g0v0_EIP158", "RevertPrecompiledTouchExactOOG_d7g1v0_ConstantinopleFix", "RevertPrecompiledTouchExactOOG_d31g1v0_ConstantinopleFix", "RevertPrecompiledTouch_storage_d3g0v0_ConstantinopleFix", "RevertPrecompiledTouch_storage_d0g0v0_ConstantinopleFix" ``` Fix the tests and "un-skip" them. NOTE: of the above all are fixed except two by applying @debris's fix from #10923 to `ethcore/machine/src/executive.rs`: ``` + let ripemd = Address::from_low_u64_be(3); + if un_substate.touched.contains(&ripemd) { + substate.touched.insert(ripemd); + } ``` Failing blockchain tests with the above: ``` "RevertPrecompiledTouchExactOOG_d7g1v0_ConstantinopleFix", "RevertPrecompiledTouchExactOOG_d31g1v0_ConstantinopleFix", ```
1.0
[json-tests] stRevertTest blockchain test failures - After https://github.com/paritytech/parity-ethereum/pull/11054 the following JSON tests fail in [`stRevertTest`](https://github.com/ethereum/tests/tree/b64fd32dd897285a3884f68b40cee358a4b3c54f/LegacyTests/Constantinople/BlockchainTests/GeneralStateTests/stRevertTest) and are added to the list of skipped tests in `ethcore/res/ethereum/tests-issues/currents.json`. ``` "RevertPrecompiledTouch_d0g0v0_Byzantium", "RevertPrecompiledTouch_d0g0v0_Constantinople", "RevertPrecompiledTouch_d0g0v0_ConstantinopleFix", "RevertPrecompiledTouch_d0g0v0_EIP158", "RevertPrecompiledTouch_d3g0v0_ConstantinopleFix", "RevertPrecompiledTouchCC_d0g0v0_Byzantium", "RevertPrecompiledTouchCC_d0g0v0_Constantinople", "RevertPrecompiledTouchCC_d0g0v0_EIP158", "RevertPrecompiledTouchDC_d0g0v0_Byzantium", "RevertPrecompiledTouchDC_d0g0v0_Constantinople", "RevertPrecompiledTouchDC_d0g0v0_EIP158", "RevertPrecompiledTouchExactOOG_d7g1v0_ConstantinopleFix", "RevertPrecompiledTouchExactOOG_d31g1v0_ConstantinopleFix", "RevertPrecompiledTouch_storage_d3g0v0_ConstantinopleFix", "RevertPrecompiledTouch_storage_d0g0v0_ConstantinopleFix" ``` Fix the tests and "un-skip" them. NOTE: of the above all are fixed except two by applying @debris's fix from #10923 to `ethcore/machine/src/executive.rs`: ``` + let ripemd = Address::from_low_u64_be(3); + if un_substate.touched.contains(&ripemd) { + substate.touched.insert(ripemd); + } ``` Failing blockchain tests with the above: ``` "RevertPrecompiledTouchExactOOG_d7g1v0_ConstantinopleFix", "RevertPrecompiledTouchExactOOG_d31g1v0_ConstantinopleFix", ```
non_main
streverttest blockchain test failures after the following json tests fail in and are added to the list of skipped tests in ethcore res ethereum tests issues currents json revertprecompiledtouch byzantium revertprecompiledtouch constantinople revertprecompiledtouch constantinoplefix revertprecompiledtouch revertprecompiledtouch constantinoplefix revertprecompiledtouchcc byzantium revertprecompiledtouchcc constantinople revertprecompiledtouchcc revertprecompiledtouchdc byzantium revertprecompiledtouchdc constantinople revertprecompiledtouchdc revertprecompiledtouchexactoog constantinoplefix revertprecompiledtouchexactoog constantinoplefix revertprecompiledtouch storage constantinoplefix revertprecompiledtouch storage constantinoplefix fix the tests and un skip them note of the above all are fixed except two by applying debris s fix from to ethcore machine src executive rs let ripemd address from low be if un substate touched contains ripemd substate touched insert ripemd failing blockchain tests with the above revertprecompiledtouchexactoog constantinoplefix revertprecompiledtouchexactoog constantinoplefix
0
802,577
28,967,391,497
IssuesEvent
2023-05-10 08:51:18
rpm-software-management/dnf5
https://api.github.com/repos/rpm-software-management/dnf5
closed
Add support allow_downgrade configuration from config and commandline
Priority: LOW
This is about modifying solver SOLVER_FLAG_ALLOW_DOWNGRADE option. Currently it is implemented in GoalPrivate, but there is no way how to modify it using our API or configuration. The default value is True. This default provide a better UX, but some user prefer to use False. I am proposing - [ ] to add `allow_downgrade` configuration option with default value true - [ ] add tho command line options `--allow-downgrades`, and and `--no-downgrades` NOTE - names of options are not fixed
1.0
Add support allow_downgrade configuration from config and commandline - This is about modifying solver SOLVER_FLAG_ALLOW_DOWNGRADE option. Currently it is implemented in GoalPrivate, but there is no way how to modify it using our API or configuration. The default value is True. This default provide a better UX, but some user prefer to use False. I am proposing - [ ] to add `allow_downgrade` configuration option with default value true - [ ] add tho command line options `--allow-downgrades`, and and `--no-downgrades` NOTE - names of options are not fixed
non_main
add support allow downgrade configuration from config and commandline this is about modifying solver solver flag allow downgrade option currently it is implemented in goalprivate but there is no way how to modify it using our api or configuration the default value is true this default provide a better ux but some user prefer to use false i am proposing to add allow downgrade configuration option with default value true add tho command line options allow downgrades and and no downgrades note names of options are not fixed
0
87,588
8,101,511,686
IssuesEvent
2018-08-12 14:37:11
hrydgard/ppsspp
https://api.github.com/repos/hrydgard/ppsspp
closed
Got an error in texture upload: 00000501 in silent hill games and METAL SLUG XX
Needs hardware testing
Seperated from https://github.com/hrydgard/ppsspp/issues/8173 tracestack ![01](https://cloud.githubusercontent.com/assets/3481559/12064175/4487ceae-aff7-11e5-9040-1b7f0494ade6.png) ![2](https://cloud.githubusercontent.com/assets/3481559/12069215/21e4d564-b05f-11e5-98b4-1c05a3e3bc8d.png)
1.0
Got an error in texture upload: 00000501 in silent hill games and METAL SLUG XX - Seperated from https://github.com/hrydgard/ppsspp/issues/8173 tracestack ![01](https://cloud.githubusercontent.com/assets/3481559/12064175/4487ceae-aff7-11e5-9040-1b7f0494ade6.png) ![2](https://cloud.githubusercontent.com/assets/3481559/12069215/21e4d564-b05f-11e5-98b4-1c05a3e3bc8d.png)
non_main
got an error in texture upload in silent hill games and metal slug xx seperated from tracestack
0
686,697
23,501,471,763
IssuesEvent
2022-08-18 08:48:26
CarmenMariaMP/Clap
https://api.github.com/repos/CarmenMariaMP/Clap
closed
H16 - Subscription to registered user
Epic high-priority
Allow registered users to subscribe to a content creator account or to a company account.
1.0
H16 - Subscription to registered user - Allow registered users to subscribe to a content creator account or to a company account.
non_main
subscription to registered user allow registered users to subscribe to a content creator account or to a company account
0
1,769
6,575,036,833
IssuesEvent
2017-09-11 14:50:51
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
gce_pd: libcloud with GCE support (0.17.0+) is required for this module
affects_2.3 bug_report cloud gce waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME gce_pd ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.3.0 (devel 9265016b95) last updated 2016/10/07 13:59:01 (GMT -500) lib/ansible/modules/core: (detached HEAD b4f6a25195) last updated 2016/10/06 13:07:51 (GMT -500) lib/ansible/modules/extras: (detached HEAD 118fe8283e) last updated 2016/10/06 13:07:51 (GMT -500) config file = /Users/tanner/projects/ansible.git/playbooks.git/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Control host OS X 10.11.6 ##### SUMMARY Attempting to use the gce_pd module and getting an error message that my installed libcloud is too old. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: add disk space gce_pd: name: default project_id: "{{ project_id }}" credentials_file: "{{ credentials_file }}" service_account_email: "{{ service_account_email }}" size_gb: "{{ size_gb }}" ``` ``` $ pip freeze | grep -i cloud apache-libcloud==1.2.1 gcloud==0.18.3 ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> A "{{ size_gb }}" size disk would be added to my GCE virtual machine. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` FAILED! => {"changed": false, "failed": true, "msg": "libcloud with GCE support (0.17.0+) is required for this module"} ```
True
gce_pd: libcloud with GCE support (0.17.0+) is required for this module - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME gce_pd ##### ANSIBLE VERSION ``` $ ansible --version ansible 2.3.0 (devel 9265016b95) last updated 2016/10/07 13:59:01 (GMT -500) lib/ansible/modules/core: (detached HEAD b4f6a25195) last updated 2016/10/06 13:07:51 (GMT -500) lib/ansible/modules/extras: (detached HEAD 118fe8283e) last updated 2016/10/06 13:07:51 (GMT -500) config file = /Users/tanner/projects/ansible.git/playbooks.git/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Control host OS X 10.11.6 ##### SUMMARY Attempting to use the gce_pd module and getting an error message that my installed libcloud is too old. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: add disk space gce_pd: name: default project_id: "{{ project_id }}" credentials_file: "{{ credentials_file }}" service_account_email: "{{ service_account_email }}" size_gb: "{{ size_gb }}" ``` ``` $ pip freeze | grep -i cloud apache-libcloud==1.2.1 gcloud==0.18.3 ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> A "{{ size_gb }}" size disk would be added to my GCE virtual machine. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` FAILED! => {"changed": false, "failed": true, "msg": "libcloud with GCE support (0.17.0+) is required for this module"} ```
main
gce pd libcloud with gce support is required for this module issue type bug report component name gce pd ansible version ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file users tanner projects ansible git playbooks git ansible cfg configured module search path default w o overrides configuration os environment control host os x summary attempting to use the gce pd module and getting an error message that my installed libcloud is too old steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name add disk space gce pd name default project id project id credentials file credentials file service account email service account email size gb size gb pip freeze grep i cloud apache libcloud gcloud expected results a size gb size disk would be added to my gce virtual machine actual results failed changed false failed true msg libcloud with gce support is required for this module
1
200,455
22,775,707,082
IssuesEvent
2022-07-08 14:16:39
OpenLiberty/open-liberty
https://api.github.com/repos/OpenLiberty/open-liberty
closed
IBM WebSphere Application Server Liberty is vulnerable to Identity Spoofing (CVE-2022-22475 CVSS 5.0)
in:Security team:Core Security release bug
**Describe the bug** Confidential for [CVE-2022-22475 CVSS 5.0](https://www.ibm.com/support/pages/node/6602015) . APAR PH46073 has been assigned for this issue.
True
IBM WebSphere Application Server Liberty is vulnerable to Identity Spoofing (CVE-2022-22475 CVSS 5.0) - **Describe the bug** Confidential for [CVE-2022-22475 CVSS 5.0](https://www.ibm.com/support/pages/node/6602015) . APAR PH46073 has been assigned for this issue.
non_main
ibm websphere application server liberty is vulnerable to identity spoofing cve cvss describe the bug confidential for apar has been assigned for this issue
0
5,747
30,441,485,314
IssuesEvent
2023-07-15 05:28:19
danswer-ai/danswer
https://api.github.com/repos/danswer-ai/danswer
closed
Developer Guide for contributing connectors
maintainer approved
Hi, Perhaps you could put together a Developer Guide for contributing connectors and add it to https://docs.danswer.dev/connectors/overview? I just opened a couple of connector issues myself and I see there were a couple of other ones before me. If it's not too difficult this is where the community and open source could really shine. I think this approach has worked really well for projects like [Ruff](https://github.com/astral-sh/ruff) for example. All the best!
True
Developer Guide for contributing connectors - Hi, Perhaps you could put together a Developer Guide for contributing connectors and add it to https://docs.danswer.dev/connectors/overview? I just opened a couple of connector issues myself and I see there were a couple of other ones before me. If it's not too difficult this is where the community and open source could really shine. I think this approach has worked really well for projects like [Ruff](https://github.com/astral-sh/ruff) for example. All the best!
main
developer guide for contributing connectors hi perhaps you could put together a developer guide for contributing connectors and add it to i just opened a couple of connector issues myself and i see there were a couple of other ones before me if it s not too difficult this is where the community and open source could really shine i think this approach has worked really well for projects like for example all the best
1
3,181
12,226,852,092
IssuesEvent
2020-05-03 12:51:05
gfleetwood/asteres
https://api.github.com/repos/gfleetwood/asteres
opened
elbersb/tidylog (160428551)
R maintain
https://github.com/elbersb/tidylog Tidylog provides feedback about dplyr and tidyr operations. It provides wrapper functions for the most common functions, such as filter, mutate, select, and group_by, and provides detailed output for joins.
True
elbersb/tidylog (160428551) - https://github.com/elbersb/tidylog Tidylog provides feedback about dplyr and tidyr operations. It provides wrapper functions for the most common functions, such as filter, mutate, select, and group_by, and provides detailed output for joins.
main
elbersb tidylog tidylog provides feedback about dplyr and tidyr operations it provides wrapper functions for the most common functions such as filter mutate select and group by and provides detailed output for joins
1
5,460
27,313,220,940
IssuesEvent
2023-02-24 13:52:24
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Change the name of `User` role
type: enhancement affects: ux work: frontend status: ready restricted: maintainers
## Problem The naming of the User role is a bit confusing. We are adding a new user and the name of the role is also User. We should come up with a different terminology for the non-admin user role. ![image](https://user-images.githubusercontent.com/11032856/220952071-1f5a8dbd-e7fe-43e6-be89-474aea568b49.png) ## Solution We should change the "Role" label to "User Type" and the "User" value should be changed to "Standard".
True
Change the name of `User` role - ## Problem The naming of the User role is a bit confusing. We are adding a new user and the name of the role is also User. We should come up with a different terminology for the non-admin user role. ![image](https://user-images.githubusercontent.com/11032856/220952071-1f5a8dbd-e7fe-43e6-be89-474aea568b49.png) ## Solution We should change the "Role" label to "User Type" and the "User" value should be changed to "Standard".
main
change the name of user role problem the naming of the user role is a bit confusing we are adding a new user and the name of the role is also user we should come up with a different terminology for the non admin user role solution we should change the role label to user type and the user value should be changed to standard
1
5,527
27,633,099,916
IssuesEvent
2023-03-10 12:27:47
conbench/conbench
https://api.github.com/repos/conbench/conbench
opened
UI: work towards few/no w3c validator warnings/errors
UI/UX maintainability
When I built websites for businesses one of my most important tools was the W3C validator. Using it for `https://conbench.ursa.dev/login/` (I picked it because it has little content) currently shows warnings and errors, for example: - `The align attribute on the div element is obsolete` - `Bad value for attribute action on element form: Must be non-empty` (`https://validator.w3.org/nu/?doc=https%3A%2F%2Fconbench.ursa.dev%2Flogin%2F`) It's good to keep these at a minimum; so that the validator can be used efficiently during dev (high signal to noise ratio).
True
UI: work towards few/no w3c validator warnings/errors - When I built websites for businesses one of my most important tools was the W3C validator. Using it for `https://conbench.ursa.dev/login/` (I picked it because it has little content) currently shows warnings and errors, for example: - `The align attribute on the div element is obsolete` - `Bad value for attribute action on element form: Must be non-empty` (`https://validator.w3.org/nu/?doc=https%3A%2F%2Fconbench.ursa.dev%2Flogin%2F`) It's good to keep these at a minimum; so that the validator can be used efficiently during dev (high signal to noise ratio).
main
ui work towards few no validator warnings errors when i built websites for businesses one of my most important tools was the validator using it for i picked it because it has little content currently shows warnings and errors for example the align attribute on the div element is obsolete bad value for attribute action on element form must be non empty it s good to keep these at a minimum so that the validator can be used efficiently during dev high signal to noise ratio
1
1,688
6,574,166,850
IssuesEvent
2017-09-11 11:47:30
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Exception with ios_command: AttributeError: 'list' object has no attribute 'splitlines'
affects_2.3 bug_report networking waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ios_command ##### ANSIBLE VERSION ``` ansible --version 2.3.0 (commit 20160706.246e32a) config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION inventory = ./hosts gathering = explicit roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/git-Ansible/Roles/roles private_role_vars = yes log_path = /var/log/ansible.log fact_caching = redis fact_caching_timeout = 86400 retry_files_enabled = False ##### OS / ENVIRONMENT - **Local host**: Ubuntu 16.10 4.8 - **Target nodes**: IOSv 15.6(2)T IOSv_L2 15.2(4.0.55)E ##### SUMMARY This exception happens with different types of targets (IOSv & IOSv_L2), although there is no issue running ios_facts with the same targets. ##### STEPS TO REPRODUCE With config=running-config or config=startup-config ``` - name: Fetching config from the remote node ios_command: provider: "{{ connections.ssh }}" commands: - "show {{ config }}" register: configuration ``` ##### EXPECTED RESULTS Successful "show running-config" or "show startup-config" ##### ACTUAL RESULTS ``` TASK [ios_pull_config : Fetching config from the remote node] *************************************************************************************************************** task path: /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles/ios_pull_config/tasks/main.yml:76 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/network/ios/ios_command.py Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/network/ios/ios_command.py <172.21.100.210> ESTABLISH LOCAL CONNECTION FOR USER: root <172.21.100.210> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534 `" && echo ansible-tmp-1479307602.12-88135957770534="` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534 `" ) && sleep 0' <172.21.100.220> ESTABLISH LOCAL CONNECTION FOR USER: root <172.21.100.220> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858 `" && echo ansible-tmp-1479307602.13-213866714018858="` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858 `" ) && sleep 0' <172.21.100.220> PUT /tmp/tmpT5ggUh TO /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ios_command.py <172.21.100.210> PUT /tmp/tmpqxKZuu TO /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ios_command.py <172.21.100.220> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ios_command.py && sleep 0' <172.21.100.210> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ios_command.py && sleep 0' <172.21.100.220> EXEC /bin/sh -c '/usr/bin/python /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ios_command.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/" > /dev/null 2>&1 && sleep 0' <172.21.100.210> EXEC /bin/sh -c '/usr/bin/python /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ios_command.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 126, in run res = self._execute() File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 502, in _execute result = self._handler.run(task_vars=variables) File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/normal.py", line 33, in run results = merge_hash(results, self._execute_module(tmp=tmp, task_vars=task_vars)) File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/__init__.py", line 662, in _execute_module data['stdout_lines'] = data.get('stdout', u'').splitlines() AttributeError: 'list' object has no attribute 'splitlines' fatal: [IOSv_L2_10]: FAILED! => { "failed": true, "msg": "Unexpected failure during module execution.", "stdout": "" } ``` Despite being able to retrieve its facts with ios_facts: - name: Fetching facts from the remote node ios_facts: gather_subset: all provider: "{{ connections.ssh }}" register: facts ``` TASK [ios_pull_facts : Fetching facts from the remote node] ***************************************************************************************************************** ok: [IOSv_L2_10] => {"ansible_facts": {"ansible_net_all_ipv4_addresses": ["172.21.100.210"], "ansible_net_all_ipv6_addresses": [], "ansible_net_config": "Building configuration...\n\nCurrent configuration : 7126 bytes\n!\n! Last configuration change at 14:42:21 UTC Wed Nov 16 2016 by admin\n!\nversion 15.2\nservice timestamps debug datetime msec\nservice timestamps log datetime msec\nservice password-encryption\nservice compress-config\n!\nhostname IOSv_L2_10\n!\nboot-start-marker\nboot-end-marker\n!\n!\nenable se ...nged": false, "failed_commands": []} ```
True
Exception with ios_command: AttributeError: 'list' object has no attribute 'splitlines' - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ios_command ##### ANSIBLE VERSION ``` ansible --version 2.3.0 (commit 20160706.246e32a) config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION inventory = ./hosts gathering = explicit roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/git-Ansible/Roles/roles private_role_vars = yes log_path = /var/log/ansible.log fact_caching = redis fact_caching_timeout = 86400 retry_files_enabled = False ##### OS / ENVIRONMENT - **Local host**: Ubuntu 16.10 4.8 - **Target nodes**: IOSv 15.6(2)T IOSv_L2 15.2(4.0.55)E ##### SUMMARY This exception happens with different types of targets (IOSv & IOSv_L2), although there is no issue running ios_facts with the same targets. ##### STEPS TO REPRODUCE With config=running-config or config=startup-config ``` - name: Fetching config from the remote node ios_command: provider: "{{ connections.ssh }}" commands: - "show {{ config }}" register: configuration ``` ##### EXPECTED RESULTS Successful "show running-config" or "show startup-config" ##### ACTUAL RESULTS ``` TASK [ios_pull_config : Fetching config from the remote node] *************************************************************************************************************** task path: /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles/ios_pull_config/tasks/main.yml:76 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/network/ios/ios_command.py Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/network/ios/ios_command.py <172.21.100.210> ESTABLISH LOCAL CONNECTION FOR USER: root <172.21.100.210> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534 `" && echo ansible-tmp-1479307602.12-88135957770534="` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534 `" ) && sleep 0' <172.21.100.220> ESTABLISH LOCAL CONNECTION FOR USER: root <172.21.100.220> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858 `" && echo ansible-tmp-1479307602.13-213866714018858="` echo $HOME/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858 `" ) && sleep 0' <172.21.100.220> PUT /tmp/tmpT5ggUh TO /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ios_command.py <172.21.100.210> PUT /tmp/tmpqxKZuu TO /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ios_command.py <172.21.100.220> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ios_command.py && sleep 0' <172.21.100.210> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ios_command.py && sleep 0' <172.21.100.220> EXEC /bin/sh -c '/usr/bin/python /root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/ios_command.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1479307602.13-213866714018858/" > /dev/null 2>&1 && sleep 0' <172.21.100.210> EXEC /bin/sh -c '/usr/bin/python /root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/ios_command.py; rm -rf "/root/.ansible/tmp/ansible-tmp-1479307602.12-88135957770534/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 126, in run res = self._execute() File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 502, in _execute result = self._handler.run(task_vars=variables) File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/normal.py", line 33, in run results = merge_hash(results, self._execute_module(tmp=tmp, task_vars=task_vars)) File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/__init__.py", line 662, in _execute_module data['stdout_lines'] = data.get('stdout', u'').splitlines() AttributeError: 'list' object has no attribute 'splitlines' fatal: [IOSv_L2_10]: FAILED! => { "failed": true, "msg": "Unexpected failure during module execution.", "stdout": "" } ``` Despite being able to retrieve its facts with ios_facts: - name: Fetching facts from the remote node ios_facts: gather_subset: all provider: "{{ connections.ssh }}" register: facts ``` TASK [ios_pull_facts : Fetching facts from the remote node] ***************************************************************************************************************** ok: [IOSv_L2_10] => {"ansible_facts": {"ansible_net_all_ipv4_addresses": ["172.21.100.210"], "ansible_net_all_ipv6_addresses": [], "ansible_net_config": "Building configuration...\n\nCurrent configuration : 7126 bytes\n!\n! Last configuration change at 14:42:21 UTC Wed Nov 16 2016 by admin\n!\nversion 15.2\nservice timestamps debug datetime msec\nservice timestamps log datetime msec\nservice password-encryption\nservice compress-config\n!\nhostname IOSv_L2_10\n!\nboot-start-marker\nboot-end-marker\n!\n!\nenable se ...nged": false, "failed_commands": []} ```
main
exception with ios command attributeerror list object has no attribute splitlines issue type bug report component name ios command ansible version ansible version commit config file etc ansible ansible cfg configured module search path default w o overrides configuration inventory hosts gathering explicit roles path home actionmystique program files ubuntu ansible git ansible roles roles private role vars yes log path var log ansible log fact caching redis fact caching timeout retry files enabled false os environment local host ubuntu target nodes iosv t iosv e summary this exception happens with different types of targets iosv iosv although there is no issue running ios facts with the same targets steps to reproduce with config running config or config startup config name fetching config from the remote node ios command provider connections ssh commands show config register configuration expected results successful show running config or show startup config actual results task task path home actionmystique program files ubuntu ansible roles roles ios pull config tasks main yml using module file usr lib dist packages ansible modules core network ios ios command py using module file usr lib dist packages ansible modules core network ios ios command py establish local connection for user root exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep establish local connection for user root exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to root ansible tmp ansible tmp ios command py put tmp tmpqxkzuu to root ansible tmp ansible tmp ios command py exec bin sh c chmod u x root ansible tmp ansible tmp root ansible tmp ansible tmp ios command py sleep exec bin sh c chmod u x root ansible tmp ansible tmp root ansible tmp ansible tmp ios command py sleep exec bin sh c usr bin python root ansible tmp ansible tmp ios command py rm rf root ansible tmp ansible tmp dev null sleep exec bin sh c usr bin python root ansible tmp ansible tmp ios command py rm rf root ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file usr lib dist packages ansible executor task executor py line in run res self execute file usr lib dist packages ansible executor task executor py line in execute result self handler run task vars variables file usr lib dist packages ansible plugins action normal py line in run results merge hash results self execute module tmp tmp task vars task vars file usr lib dist packages ansible plugins action init py line in execute module data data get stdout u splitlines attributeerror list object has no attribute splitlines fatal failed failed true msg unexpected failure during module execution stdout despite being able to retrieve its facts with ios facts name fetching facts from the remote node ios facts gather subset all provider connections ssh register facts task ok ansible facts ansible net all addresses ansible net all addresses ansible net config building configuration n ncurrent configuration bytes n n last configuration change at utc wed nov by admin n nversion nservice timestamps debug datetime msec nservice timestamps log datetime msec nservice password encryption nservice compress config n nhostname iosv n nboot start marker nboot end marker n n nenable se nged false failed commands
1
2,916
5,914,189,587
IssuesEvent
2017-05-22 01:10:40
AffiliateWP/AffiliateWP
https://api.github.com/repos/AffiliateWP/AffiliateWP
closed
CSV import for affiliates and referrals
batch-processing enhancement Has PR needs testing
We should provide a way to import affiliates and referrals via a CSV file. Related #337 PR: #1891
1.0
CSV import for affiliates and referrals - We should provide a way to import affiliates and referrals via a CSV file. Related #337 PR: #1891
non_main
csv import for affiliates and referrals we should provide a way to import affiliates and referrals via a csv file related pr
0
883
4,543,514,348
IssuesEvent
2016-09-10 05:42:14
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Can not use state: started when specify log_driver in docker_container module
affects_2.1 bug_report cloud docker in progress waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### OS / ENVIRONMENT N/A ##### SUMMARY When I use state: started to start container with container that created with log_driver option. It always remove container and show error Error creating container: 400 Client Error: Bad Request ("No command specified") ##### STEPS TO REPRODUCE ``` --- - name: Test create and start container hosts: localhost connection: local gather_facts: no tasks: - docker_container: name: test image: nginx log_driver: gelf log_options: gelf-address: udp://graylog.example.com:12201 state: present - docker_container: name: test state: started ``` ##### EXPECTED RESULTS The container should start correctly ##### ACTUAL RESULTS ``` Using /home/username/git/ansible/ansible.cfg as config file Loaded callback default of type stdout, v2.0 PLAYBOOK: test.yml ************************************************************* 1 plays in test.yml PLAY [Test Docker container] *************************************************** TASK [docker_container] ******************************************************** task path: /home/username/git/ansible/test.yml:7 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: username <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808 `" && echo ansible-tmp-1472553339.39-107102396799808="` echo $HOME/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808 `" ) && sleep 0' <localhost> PUT /tmp/tmpHnZwHc TO /home/username/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808/docker_container <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/username/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808/docker_container; rm -rf "/home/username/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => {"ansible_facts": {"ansible_docker_container": {"AppArmorProfile": "", "Args": ["-g", "daemon off;"], "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["nginx", "-g", "daemon off;"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "NGINX_VERSION=1.11.3-1~jessie"], "ExposedPorts": {"443/tcp": {}, "80/tcp": {}}, "Hostname": "4dba25aa3867", "Image": "nginx", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-08-30T10:35:49.623067185Z", "Driver": "overlay2", "ExecIDs": null, "GraphDriver": {"Data": {"LowerDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7-init/diff:/var/lib/docker/overlay2/66ce9000f8d2ada07c7de4162b95a7fc120937c3f577d5ce992092129bf109ce/diff:/var/lib/docker/overlay2/8765c294e83460225a9626e86bb0ab00a33d6ff32ac4886c35526a00e2d80abf/diff:/var/lib/docker/overlay2/661c6e859f6ba5dc7c36934ec20f67aafaa8baa21af6e3e666539634ba7d96a7/diff", "MergedDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7/merged", "UpperDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7/diff", "WorkDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7/work"}, "Name": "overlay2"}, "HostConfig": {"AutoRemove": false, "Binds": [], "BlkioDeviceReadBps": null, "BlkioDeviceReadIOps": null, "BlkioDeviceWriteBps": null, "BlkioDeviceWriteIOps": null, "BlkioWeight": 0, "BlkioWeightDevice": null, "CapAdd": null, "CapDrop": null, "Cgroup": "", "CgroupParent": "", "ConsoleSize": [0, 0], "ContainerIDFile": "", "CpuCount": 0, "CpuPercent": 0, "CpuPeriod": 0, "CpuQuota": 0, "CpuShares": 0, "CpusetCpus": "", "CpusetMems": "", "Devices": null, "DiskQuota": 0, "Dns": null, "DnsOptions": null, "DnsSearch": null, "ExtraHosts": null, "GroupAdd": null, "IOMaximumBandwidth": 0, "IOMaximumIOps": 0, "IpcMode": "", "Isolation": "", "KernelMemory": 0, "Links": null, "LogConfig": {"Config": {"gelf-address": "udp://graylog.example.com:12201"}, "Type": "gelf"}, "Memory": 0, "MemoryReservation": 0, "MemorySwap": 0, "MemorySwappiness": -1, "NetworkMode": "default", "OomKillDisable": false, "OomScoreAdj": 0, "PidMode": "", "PidsLimit": 0, "PortBindings": null, "Privileged": false, "PublishAllPorts": false, "ReadonlyRootfs": false, "RestartPolicy": {"MaximumRetryCount": 0, "Name": ""}, "Runtime": "runc", "SecurityOpt": null, "ShmSize": 67108864, "UTSMode": "", "Ulimits": null, "UsernsMode": "", "VolumeDriver": "", "VolumesFrom": null}, "HostnamePath": "", "HostsPath": "", "Id": "4dba25aa3867c34d9fbefeb5ac350619ec3125ffd9ef6b3623916f060631c969", "Image": "sha256:4efb2fcdb1ab05fb03c9435234343c1cc65289eeb016be86193e88d3a5d84f6b", "LogPath": "", "MountLabel": "", "Mounts": [], "Name": "/test", "NetworkSettings": {"Bridge": "", "EndpointID": "", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "HairpinMode": false, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "LinkLocalIPv6Address": "", "LinkLocalIPv6PrefixLen": 0, "MacAddress": "", "Networks": {"bridge": {"Aliases": null, "EndpointID": "", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "IPAMConfig": null, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "Links": null, "MacAddress": "", "NetworkID": ""}}, "Ports": null, "SandboxID": "", "SandboxKey": "", "SecondaryIPAddresses": null, "SecondaryIPv6Addresses": null}, "Path": "nginx", "ProcessLabel": "", "ResolvConfPath": "", "RestartCount": 0, "State": {"Dead": false, "Error": "", "ExitCode": 0, "FinishedAt": "0001-01-01T00:00:00Z", "OOMKilled": false, "Paused": false, "Pid": 0, "Restarting": false, "Running": false, "StartedAt": "0001-01-01T00:00:00Z", "Status": "created"}}}, "changed": true, "invocation": {"module_args": {"api_version": null, "blkio_weight": null, "cacert_path": null, "capabilities": null, "cert_path": null, "cleanup": false, "command": null, "cpu_period": null, "cpu_quota": null, "cpu_shares": null, "cpuset_cpus": null, "cpuset_mems": null, "debug": false, "detach": true, "devices": null, "dns_opts": null, "dns_search_domains": null, "dns_servers": null, "docker_host": null, "entrypoint": null, "env": null, "env_file": null, "etc_hosts": null, "exposed_ports": null, "filter_logger": false, "force_kill": false, "groups": null, "hostname": null, "ignore_image": false, "image": "nginx", "interactive": false, "ipc_mode": null, "keep_volumes": true, "kernel_memory": null, "key_path": null, "kill_signal": null, "labels": null, "links": null, "log_driver": "gelf", "log_options": {"gelf-address": "udp://graylog.example.com:12201"}, "mac_address": null, "memory": "0", "memory_reservation": null, "memory_swap": null, "memory_swappiness": null, "name": "test", "network_mode": null, "networks": null, "oom_killer": null, "paused": false, "pid_mode": null, "privileged": false, "published_ports": null, "pull": false, "purge_networks": null, "read_only": false, "recreate": false, "restart": false, "restart_policy": null, "restart_retries": 0, "security_opts": null, "shm_size": null, "ssl_version": null, "state": "present", "stop_signal": null, "stop_timeout": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null, "trust_image_content": false, "tty": false, "ulimits": null, "user": null, "uts": null, "volume_driver": null, "volumes": null, "volumes_from": null}, "module_name": "docker_container"}} TASK [docker_container] ******************************************************** task path: /home/username/git/ansible/test.yml:15 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: username <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115 `" && echo ansible-tmp-1472553349.83-249876067720115="` echo $HOME/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115 `" ) && sleep 0' <localhost> PUT /tmp/tmpeeGOlf TO /home/username/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115/docker_container <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/username/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115/docker_container; rm -rf "/home/username/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115/" > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"api_version": null, "blkio_weight": null, "cacert_path": null, "capabilities": null, "cert_path": null, "cleanup": false, "command": null, "cpu_period": null, "cpu_quota": null, "cpu_shares": null, "cpuset_cpus": null, "cpuset_mems": null, "debug": false, "detach": true, "devices": null, "dns_opts": null, "dns_search_domains": null, "dns_servers": null, "docker_host": null, "entrypoint": null, "env": null, "env_file": null, "etc_hosts": null, "exposed_ports": null, "filter_logger": false, "force_kill": false, "groups": null, "hostname": null, "ignore_image": false, "image": null, "interactive": false, "ipc_mode": null, "keep_volumes": true, "kernel_memory": null, "key_path": null, "kill_signal": null, "labels": null, "links": null, "log_driver": "json-file", "log_options": null, "mac_address": null, "memory": "0", "memory_reservation": null, "memory_swap": null, "memory_swappiness": null, "name": "test", "network_mode": null, "networks": null, "oom_killer": null, "paused": false, "pid_mode": null, "privileged": false, "published_ports": null, "pull": false, "purge_networks": null, "read_only": false, "recreate": false, "restart": false, "restart_policy": null, "restart_retries": 0, "security_opts": null, "shm_size": null, "ssl_version": null, "state": "started", "stop_signal": null, "stop_timeout": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null, "trust_image_content": false, "tty": false, "ulimits": null, "user": null, "uts": null, "volume_driver": null, "volumes": null, "volumes_from": null}, "module_name": "docker_container"}, "msg": "Error creating container: 400 Client Error: Bad Request (\"No command specified\")"} NO MORE HOSTS LEFT ************************************************************* PLAY RECAP ********************************************************************* localhost : ok=1 changed=1 unreachable=0 failed=1 ```
True
Can not use state: started when specify log_driver in docker_container module - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### OS / ENVIRONMENT N/A ##### SUMMARY When I use state: started to start container with container that created with log_driver option. It always remove container and show error Error creating container: 400 Client Error: Bad Request ("No command specified") ##### STEPS TO REPRODUCE ``` --- - name: Test create and start container hosts: localhost connection: local gather_facts: no tasks: - docker_container: name: test image: nginx log_driver: gelf log_options: gelf-address: udp://graylog.example.com:12201 state: present - docker_container: name: test state: started ``` ##### EXPECTED RESULTS The container should start correctly ##### ACTUAL RESULTS ``` Using /home/username/git/ansible/ansible.cfg as config file Loaded callback default of type stdout, v2.0 PLAYBOOK: test.yml ************************************************************* 1 plays in test.yml PLAY [Test Docker container] *************************************************** TASK [docker_container] ******************************************************** task path: /home/username/git/ansible/test.yml:7 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: username <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808 `" && echo ansible-tmp-1472553339.39-107102396799808="` echo $HOME/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808 `" ) && sleep 0' <localhost> PUT /tmp/tmpHnZwHc TO /home/username/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808/docker_container <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/username/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808/docker_container; rm -rf "/home/username/.ansible/tmp/ansible-tmp-1472553339.39-107102396799808/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => {"ansible_facts": {"ansible_docker_container": {"AppArmorProfile": "", "Args": ["-g", "daemon off;"], "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["nginx", "-g", "daemon off;"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin", "NGINX_VERSION=1.11.3-1~jessie"], "ExposedPorts": {"443/tcp": {}, "80/tcp": {}}, "Hostname": "4dba25aa3867", "Image": "nginx", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-08-30T10:35:49.623067185Z", "Driver": "overlay2", "ExecIDs": null, "GraphDriver": {"Data": {"LowerDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7-init/diff:/var/lib/docker/overlay2/66ce9000f8d2ada07c7de4162b95a7fc120937c3f577d5ce992092129bf109ce/diff:/var/lib/docker/overlay2/8765c294e83460225a9626e86bb0ab00a33d6ff32ac4886c35526a00e2d80abf/diff:/var/lib/docker/overlay2/661c6e859f6ba5dc7c36934ec20f67aafaa8baa21af6e3e666539634ba7d96a7/diff", "MergedDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7/merged", "UpperDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7/diff", "WorkDir": "/var/lib/docker/overlay2/ab723dd251634a8c92d1a1fc933900861a9bddf52589e1303ebaaf0a899c5af7/work"}, "Name": "overlay2"}, "HostConfig": {"AutoRemove": false, "Binds": [], "BlkioDeviceReadBps": null, "BlkioDeviceReadIOps": null, "BlkioDeviceWriteBps": null, "BlkioDeviceWriteIOps": null, "BlkioWeight": 0, "BlkioWeightDevice": null, "CapAdd": null, "CapDrop": null, "Cgroup": "", "CgroupParent": "", "ConsoleSize": [0, 0], "ContainerIDFile": "", "CpuCount": 0, "CpuPercent": 0, "CpuPeriod": 0, "CpuQuota": 0, "CpuShares": 0, "CpusetCpus": "", "CpusetMems": "", "Devices": null, "DiskQuota": 0, "Dns": null, "DnsOptions": null, "DnsSearch": null, "ExtraHosts": null, "GroupAdd": null, "IOMaximumBandwidth": 0, "IOMaximumIOps": 0, "IpcMode": "", "Isolation": "", "KernelMemory": 0, "Links": null, "LogConfig": {"Config": {"gelf-address": "udp://graylog.example.com:12201"}, "Type": "gelf"}, "Memory": 0, "MemoryReservation": 0, "MemorySwap": 0, "MemorySwappiness": -1, "NetworkMode": "default", "OomKillDisable": false, "OomScoreAdj": 0, "PidMode": "", "PidsLimit": 0, "PortBindings": null, "Privileged": false, "PublishAllPorts": false, "ReadonlyRootfs": false, "RestartPolicy": {"MaximumRetryCount": 0, "Name": ""}, "Runtime": "runc", "SecurityOpt": null, "ShmSize": 67108864, "UTSMode": "", "Ulimits": null, "UsernsMode": "", "VolumeDriver": "", "VolumesFrom": null}, "HostnamePath": "", "HostsPath": "", "Id": "4dba25aa3867c34d9fbefeb5ac350619ec3125ffd9ef6b3623916f060631c969", "Image": "sha256:4efb2fcdb1ab05fb03c9435234343c1cc65289eeb016be86193e88d3a5d84f6b", "LogPath": "", "MountLabel": "", "Mounts": [], "Name": "/test", "NetworkSettings": {"Bridge": "", "EndpointID": "", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "HairpinMode": false, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "LinkLocalIPv6Address": "", "LinkLocalIPv6PrefixLen": 0, "MacAddress": "", "Networks": {"bridge": {"Aliases": null, "EndpointID": "", "Gateway": "", "GlobalIPv6Address": "", "GlobalIPv6PrefixLen": 0, "IPAMConfig": null, "IPAddress": "", "IPPrefixLen": 0, "IPv6Gateway": "", "Links": null, "MacAddress": "", "NetworkID": ""}}, "Ports": null, "SandboxID": "", "SandboxKey": "", "SecondaryIPAddresses": null, "SecondaryIPv6Addresses": null}, "Path": "nginx", "ProcessLabel": "", "ResolvConfPath": "", "RestartCount": 0, "State": {"Dead": false, "Error": "", "ExitCode": 0, "FinishedAt": "0001-01-01T00:00:00Z", "OOMKilled": false, "Paused": false, "Pid": 0, "Restarting": false, "Running": false, "StartedAt": "0001-01-01T00:00:00Z", "Status": "created"}}}, "changed": true, "invocation": {"module_args": {"api_version": null, "blkio_weight": null, "cacert_path": null, "capabilities": null, "cert_path": null, "cleanup": false, "command": null, "cpu_period": null, "cpu_quota": null, "cpu_shares": null, "cpuset_cpus": null, "cpuset_mems": null, "debug": false, "detach": true, "devices": null, "dns_opts": null, "dns_search_domains": null, "dns_servers": null, "docker_host": null, "entrypoint": null, "env": null, "env_file": null, "etc_hosts": null, "exposed_ports": null, "filter_logger": false, "force_kill": false, "groups": null, "hostname": null, "ignore_image": false, "image": "nginx", "interactive": false, "ipc_mode": null, "keep_volumes": true, "kernel_memory": null, "key_path": null, "kill_signal": null, "labels": null, "links": null, "log_driver": "gelf", "log_options": {"gelf-address": "udp://graylog.example.com:12201"}, "mac_address": null, "memory": "0", "memory_reservation": null, "memory_swap": null, "memory_swappiness": null, "name": "test", "network_mode": null, "networks": null, "oom_killer": null, "paused": false, "pid_mode": null, "privileged": false, "published_ports": null, "pull": false, "purge_networks": null, "read_only": false, "recreate": false, "restart": false, "restart_policy": null, "restart_retries": 0, "security_opts": null, "shm_size": null, "ssl_version": null, "state": "present", "stop_signal": null, "stop_timeout": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null, "trust_image_content": false, "tty": false, "ulimits": null, "user": null, "uts": null, "volume_driver": null, "volumes": null, "volumes_from": null}, "module_name": "docker_container"}} TASK [docker_container] ******************************************************** task path: /home/username/git/ansible/test.yml:15 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: username <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115 `" && echo ansible-tmp-1472553349.83-249876067720115="` echo $HOME/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115 `" ) && sleep 0' <localhost> PUT /tmp/tmpeeGOlf TO /home/username/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115/docker_container <localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/username/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115/docker_container; rm -rf "/home/username/.ansible/tmp/ansible-tmp-1472553349.83-249876067720115/" > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"api_version": null, "blkio_weight": null, "cacert_path": null, "capabilities": null, "cert_path": null, "cleanup": false, "command": null, "cpu_period": null, "cpu_quota": null, "cpu_shares": null, "cpuset_cpus": null, "cpuset_mems": null, "debug": false, "detach": true, "devices": null, "dns_opts": null, "dns_search_domains": null, "dns_servers": null, "docker_host": null, "entrypoint": null, "env": null, "env_file": null, "etc_hosts": null, "exposed_ports": null, "filter_logger": false, "force_kill": false, "groups": null, "hostname": null, "ignore_image": false, "image": null, "interactive": false, "ipc_mode": null, "keep_volumes": true, "kernel_memory": null, "key_path": null, "kill_signal": null, "labels": null, "links": null, "log_driver": "json-file", "log_options": null, "mac_address": null, "memory": "0", "memory_reservation": null, "memory_swap": null, "memory_swappiness": null, "name": "test", "network_mode": null, "networks": null, "oom_killer": null, "paused": false, "pid_mode": null, "privileged": false, "published_ports": null, "pull": false, "purge_networks": null, "read_only": false, "recreate": false, "restart": false, "restart_policy": null, "restart_retries": 0, "security_opts": null, "shm_size": null, "ssl_version": null, "state": "started", "stop_signal": null, "stop_timeout": null, "timeout": null, "tls": null, "tls_hostname": null, "tls_verify": null, "trust_image_content": false, "tty": false, "ulimits": null, "user": null, "uts": null, "volume_driver": null, "volumes": null, "volumes_from": null}, "module_name": "docker_container"}, "msg": "Error creating container: 400 Client Error: Bad Request (\"No command specified\")"} NO MORE HOSTS LEFT ************************************************************* PLAY RECAP ********************************************************************* localhost : ok=1 changed=1 unreachable=0 failed=1 ```
main
can not use state started when specify log driver in docker container module issue type bug report component name docker container ansible version ansible os environment n a summary when i use state started to start container with container that created with log driver option it always remove container and show error error creating container client error bad request no command specified steps to reproduce name test create and start container hosts localhost connection local gather facts no tasks docker container name test image nginx log driver gelf log options gelf address udp graylog example com state present docker container name test state started expected results the container should start correctly actual results using home username git ansible ansible cfg as config file loaded callback default of type stdout playbook test yml plays in test yml play task task path home username git ansible test yml establish local connection for user username exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmphnzwhc to home username ansible tmp ansible tmp docker container exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python home username ansible tmp ansible tmp docker container rm rf home username ansible tmp ansible tmp dev null sleep changed ansible facts ansible docker container apparmorprofile args config attachstderr false attachstdin false attachstdout false cmd domainname entrypoint null env exposedports tcp tcp hostname image nginx labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir created driver execids null graphdriver data lowerdir var lib docker init diff var lib docker diff var lib docker diff var lib docker diff mergeddir var lib docker merged upperdir var lib docker diff workdir var lib docker work name hostconfig autoremove false binds blkiodevicereadbps null blkiodevicereadiops null blkiodevicewritebps null blkiodevicewriteiops null blkioweight blkioweightdevice null capadd null capdrop null cgroup cgroupparent consolesize containeridfile cpucount cpupercent cpuperiod cpuquota cpushares cpusetcpus cpusetmems devices null diskquota dns null dnsoptions null dnssearch null extrahosts null groupadd null iomaximumbandwidth iomaximumiops ipcmode isolation kernelmemory links null logconfig config gelf address udp graylog example com type gelf memory memoryreservation memoryswap memoryswappiness networkmode default oomkilldisable false oomscoreadj pidmode pidslimit portbindings null privileged false publishallports false readonlyrootfs false restartpolicy maximumretrycount name runtime runc securityopt null shmsize utsmode ulimits null usernsmode volumedriver volumesfrom null hostnamepath hostspath id image logpath mountlabel mounts name test networksettings bridge endpointid gateway hairpinmode false ipaddress ipprefixlen macaddress networks bridge aliases null endpointid gateway ipamconfig null ipaddress ipprefixlen links null macaddress networkid ports null sandboxid sandboxkey secondaryipaddresses null null path nginx processlabel resolvconfpath restartcount state dead false error exitcode finishedat oomkilled false paused false pid restarting false running false startedat status created changed true invocation module args api version null blkio weight null cacert path null capabilities null cert path null cleanup false command null cpu period null cpu quota null cpu shares null cpuset cpus null cpuset mems null debug false detach true devices null dns opts null dns search domains null dns servers null docker host null entrypoint null env null env file null etc hosts null exposed ports null filter logger false force kill false groups null hostname null ignore image false image nginx interactive false ipc mode null keep volumes true kernel memory null key path null kill signal null labels null links null log driver gelf log options gelf address udp graylog example com mac address null memory memory reservation null memory swap null memory swappiness null name test network mode null networks null oom killer null paused false pid mode null privileged false published ports null pull false purge networks null read only false recreate false restart false restart policy null restart retries security opts null shm size null ssl version null state present stop signal null stop timeout null timeout null tls null tls hostname null tls verify null trust image content false tty false ulimits null user null uts null volume driver null volumes null volumes from null module name docker container task task path home username git ansible test yml establish local connection for user username exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpeegolf to home username ansible tmp ansible tmp docker container exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python home username ansible tmp ansible tmp docker container rm rf home username ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module args api version null blkio weight null cacert path null capabilities null cert path null cleanup false command null cpu period null cpu quota null cpu shares null cpuset cpus null cpuset mems null debug false detach true devices null dns opts null dns search domains null dns servers null docker host null entrypoint null env null env file null etc hosts null exposed ports null filter logger false force kill false groups null hostname null ignore image false image null interactive false ipc mode null keep volumes true kernel memory null key path null kill signal null labels null links null log driver json file log options null mac address null memory memory reservation null memory swap null memory swappiness null name test network mode null networks null oom killer null paused false pid mode null privileged false published ports null pull false purge networks null read only false recreate false restart false restart policy null restart retries security opts null shm size null ssl version null state started stop signal null stop timeout null timeout null tls null tls hostname null tls verify null trust image content false tty false ulimits null user null uts null volume driver null volumes null volumes from null module name docker container msg error creating container client error bad request no command specified no more hosts left play recap localhost ok changed unreachable failed
1
288,832
31,930,930,661
IssuesEvent
2023-09-19 07:22:15
Trinadh465/linux-4.1.15_CVE-2023-4128
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-4128
opened
CVE-2019-15219 (Medium) detected in linuxlinux-4.6
Mend: dependency security vulnerability
## CVE-2019-15219 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/sisusbvga/sisusb.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/sisusbvga/sisusb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.1.8. There is a NULL pointer dereference caused by a malicious USB device in the drivers/usb/misc/sisusbvga/sisusb.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15219>CVE-2019-15219</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15219">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15219</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v5.2-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-15219 (Medium) detected in linuxlinux-4.6 - ## CVE-2019-15219 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/sisusbvga/sisusb.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/sisusbvga/sisusb.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.1.8. There is a NULL pointer dereference caused by a malicious USB device in the drivers/usb/misc/sisusbvga/sisusb.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15219>CVE-2019-15219</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Physical - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15219">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15219</a></p> <p>Release Date: 2019-08-19</p> <p>Fix Resolution: v5.2-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch main vulnerable source files drivers usb misc sisusbvga sisusb c drivers usb misc sisusbvga sisusb c vulnerability details an issue was discovered in the linux kernel before there is a null pointer dereference caused by a malicious usb device in the drivers usb misc sisusbvga sisusb c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector physical attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
813
4,435,313,837
IssuesEvent
2016-08-18 08:06:18
Particular/ServicePulse
https://api.github.com/repos/Particular/ServicePulse
closed
Incorrect version shown when using ServicePulse in IIS
Size: S Tag: Maintainer Prio Tag: Triaged Type: Bug
I just followed the instructions to run ServicePulse in IIS http://docs.particular.net/servicepulse/install-servicepulse-in-iis Once I navigated to the website, I immediately noticed the following message: > ServicePulse v1.2.0 ( update available) This is incorrect as I installed Service Pulse using the 1.6.5 installer. I looked into the installation folder as well as into the extracted ServicePulse under inetpub. The app.constants.js files have different version values. C:\Program Files (x86)\Particular Software\ServicePulse\app\js\app.constants.js ```javascript ;(function (window, angular, undefined) { 'use strict'; angular.module('sc') .constant('version', '1.6.5') .constant('scConfig', { service_control_url: 'http://localhost:33333/api/' }); }(window, window.angular)); ``` C:\inetpub\websites\ServicePulse\js\app.constants.js: ```javascript ;(function (window, angular, undefined) { 'use strict'; angular.module('sc') .constant('version', '1.2.0') .constant('scConfig', { service_control_url: '/api' }); }(window, window.angular)); ``` If the only duplicated and outdated file is the app.constants.js then it's not much of an issue but I don't know if the rest of the application is up to date. Here's the Powershell script I wrote to extract ServicePulse to inetpub: ```powershell Function Install-ServicePulse-IIS { param( [Parameter(Mandatory=$false)] [Int] $ServicePulsePort , [Parameter(Mandatory=$false)] [Int] $ServiceControlPort , [Parameter(Mandatory=$false)] [System.IO.FileInfo] $ServicePulseExe ) # Get Defaults If ($ServicePulsePort -eq $null) { $ServicePulsePort = 9090 } If ($ServiceControlPort -eq $null) { $ServiceControlPort = 33333 } If ($ServicePulseExe -eq $null) { $ServicePulseExe = Get-Item 'C:\Program Files (x86)\Particular Software\ServicePulse\ServicePulse.Host.exe' } # Disable default installation Set-Service -Name Particular.ServicePulse -StartupType Disabled -Status Stopped netsh http delete urlacl http://+:$ServicePulsePort/ $path = "C:\inetpub\websites\ServicePulse" # Extract to inetpub & $ServicePulseExe --extract --outPath="$path" # Create a new subdirectory called api New-Item "$path/api" -Type Directory -Force # Edit app.constants.js and change the serviceControlUrl value from http://localhost:33333/api to /api (Get-Content "$path/js/app.constants.js").Replace("http://localhost:$ServiceControlPort/api/", "/api") | Set-Content "$path/js/app.constants.js" } ```
True
Incorrect version shown when using ServicePulse in IIS - I just followed the instructions to run ServicePulse in IIS http://docs.particular.net/servicepulse/install-servicepulse-in-iis Once I navigated to the website, I immediately noticed the following message: > ServicePulse v1.2.0 ( update available) This is incorrect as I installed Service Pulse using the 1.6.5 installer. I looked into the installation folder as well as into the extracted ServicePulse under inetpub. The app.constants.js files have different version values. C:\Program Files (x86)\Particular Software\ServicePulse\app\js\app.constants.js ```javascript ;(function (window, angular, undefined) { 'use strict'; angular.module('sc') .constant('version', '1.6.5') .constant('scConfig', { service_control_url: 'http://localhost:33333/api/' }); }(window, window.angular)); ``` C:\inetpub\websites\ServicePulse\js\app.constants.js: ```javascript ;(function (window, angular, undefined) { 'use strict'; angular.module('sc') .constant('version', '1.2.0') .constant('scConfig', { service_control_url: '/api' }); }(window, window.angular)); ``` If the only duplicated and outdated file is the app.constants.js then it's not much of an issue but I don't know if the rest of the application is up to date. Here's the Powershell script I wrote to extract ServicePulse to inetpub: ```powershell Function Install-ServicePulse-IIS { param( [Parameter(Mandatory=$false)] [Int] $ServicePulsePort , [Parameter(Mandatory=$false)] [Int] $ServiceControlPort , [Parameter(Mandatory=$false)] [System.IO.FileInfo] $ServicePulseExe ) # Get Defaults If ($ServicePulsePort -eq $null) { $ServicePulsePort = 9090 } If ($ServiceControlPort -eq $null) { $ServiceControlPort = 33333 } If ($ServicePulseExe -eq $null) { $ServicePulseExe = Get-Item 'C:\Program Files (x86)\Particular Software\ServicePulse\ServicePulse.Host.exe' } # Disable default installation Set-Service -Name Particular.ServicePulse -StartupType Disabled -Status Stopped netsh http delete urlacl http://+:$ServicePulsePort/ $path = "C:\inetpub\websites\ServicePulse" # Extract to inetpub & $ServicePulseExe --extract --outPath="$path" # Create a new subdirectory called api New-Item "$path/api" -Type Directory -Force # Edit app.constants.js and change the serviceControlUrl value from http://localhost:33333/api to /api (Get-Content "$path/js/app.constants.js").Replace("http://localhost:$ServiceControlPort/api/", "/api") | Set-Content "$path/js/app.constants.js" } ```
main
incorrect version shown when using servicepulse in iis i just followed the instructions to run servicepulse in iis once i navigated to the website i immediately noticed the following message servicepulse update available this is incorrect as i installed service pulse using the installer i looked into the installation folder as well as into the extracted servicepulse under inetpub the app constants js files have different version values c program files particular software servicepulse app js app constants js javascript function window angular undefined use strict angular module sc constant version constant scconfig service control url window window angular c inetpub websites servicepulse js app constants js javascript function window angular undefined use strict angular module sc constant version constant scconfig service control url api window window angular if the only duplicated and outdated file is the app constants js then it s not much of an issue but i don t know if the rest of the application is up to date here s the powershell script i wrote to extract servicepulse to inetpub powershell function install servicepulse iis param servicepulseport servicecontrolport servicepulseexe get defaults if servicepulseport eq null servicepulseport if servicecontrolport eq null servicecontrolport if servicepulseexe eq null servicepulseexe get item c program files particular software servicepulse servicepulse host exe disable default installation set service name particular servicepulse startuptype disabled status stopped netsh http delete urlacl path c inetpub websites servicepulse extract to inetpub servicepulseexe extract outpath path create a new subdirectory called api new item path api type directory force edit app constants js and change the servicecontrolurl value from to api get content path js app constants js replace api set content path js app constants js
1
34,920
4,957,020,478
IssuesEvent
2016-12-02 02:04:25
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
kubernetes-e2e-gke-gci-1.3-container_vm-1.5-upgrade-cluster: broken test run
kind/flake priority/P2 team/test-infra
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-gci-1.3-container_vm-1.5-upgrade-cluster/339/ Multiple broken tests: Failed: [k8s.io] Deployment deployment should support rollover {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:70 Expected error: <*errors.errorString | 0xc821399680>: { s: "error waiting for deployment test-rollover-deployment status to match expectation: timed out waiting for the condition", } error waiting for deployment test-rollover-deployment status to match expectation: timed out waiting for the condition not to have occurred ``` Issues about this test specifically: #26509 #26834 #29780 #35355 Failed: [k8s.io] V1Job should fail a job {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/batch_v1_jobs.go:202 Expected error: <*errors.errorString | 0xc820019190>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred ``` Issues about this test specifically: #27704 #30127 #30602 #31070 #34383 Failed: [k8s.io] Job should fail a job {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/job.go:198 Expected error: <*errors.errorString | 0xc820019190>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred ``` Issues about this test specifically: #28773 #29506 #30699 #32734 #34585 #37177 Failed: [k8s.io] Kubectl client [k8s.io] Kubectl apply should reuse nodePort when apply to an existing SVC {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:471 Expected error: <*errors.errorString | 0xc82056f080>: { s: "Error running &{/workspace/kubernetes_skew/cluster/kubectl.sh [/workspace/kubernetes_skew/cluster/kubectl.sh --server=https://104.154.214.26 --kubeconfig=/workspace/.kube/config get service redis-master --namespace=e2e-tests-kubectl-v2cx3 -o jsonpath={.spec.ports[0].nodePort}] [] <nil> Error executing template: nodePort is not found. Printing more information for debugging the template:\n\ttemplate was:\n\t\t{.spec.ports[0].nodePort}\n\tobject given to jsonpath engine was:\n\t\tmap[string]interface {}{\"apiVersion\":\"v1\", \"metadata\":map[string]interface {}{\"selfLink\":\"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master\", \"uid\":\"9a90fe30-b7fa-11e6-bb1b-42010af00026\", \"resourceVersion\":\"1592\", \"creationTimestamp\":\"2016-12-01T19:16:04Z\", \"labels\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"name\":\"redis-master\", \"namespace\":\"e2e-tests-kubectl-v2cx3\"}, \"spec\":map[string]interface {}{\"selector\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"clusterIP\":\"10.127.249.84\", \"type\":\"ClusterIP\", \"sessionAffinity\":\"None\", \"ports\":[]interface {}{map[string]interface {}{\"protocol\":\"TCP\", \"port\":6379, \"targetPort\":\"redis-server\"}}}, \"status\":map[string]interface {}{\"loadBalancer\":map[string]interface {}{}}, \"kind\":\"Service\"}\n\n error: error executing jsonpath \"{.spec.ports[0].nodePort}\": nodePort is not found\n [] <nil> 0xc820759760 exit status 1 <nil> true [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043c8 0xc820404458] [0xa975d0 0xa975d0] 0xc820d42480}:\nCommand stdout:\nError executing template: nodePort is not found. Printing more information for debugging the template:\n\ttemplate was:\n\t\t{.spec.ports[0].nodePort}\n\tobject given to jsonpath engine was:\n\t\tmap[string]interface {}{\"apiVersion\":\"v1\", \"metadata\":map[string]interface {}{\"selfLink\":\"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master\", \"uid\":\"9a90fe30-b7fa-11e6-bb1b-42010af00026\", \"resourceVersion\":\"1592\", \"creationTimestamp\":\"2016-12-01T19:16:04Z\", \"labels\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"name\":\"redis-master\", \"namespace\":\"e2e-tests-kubectl-v2cx3\"}, \"spec\":map[string]interface {}{\"selector\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"clusterIP\":\"10.127.249.84\", \"type\":\"ClusterIP\", \"sessionAffinity\":\"None\", \"ports\":[]interface {}{map[string]interface {}{\"protocol\":\"TCP\", \"port\":6379, \"targetPort\":\"redis-server\"}}}, \"status\":map[string]interface {}{\"loadBalancer\":map[string]interface {}{}}, \"kind\":\"Service\"}\n\n\nstderr:\nerror: error executing jsonpath \"{.spec.ports[0].nodePort}\": nodePort is not found\n\nerror:\nexit status 1\n", } Error running &{/workspace/kubernetes_skew/cluster/kubectl.sh [/workspace/kubernetes_skew/cluster/kubectl.sh --server=https://104.154.214.26 --kubeconfig=/workspace/.kube/config get service redis-master --namespace=e2e-tests-kubectl-v2cx3 -o jsonpath={.spec.ports[0].nodePort}] [] <nil> Error executing template: nodePort is not found. Printing more information for debugging the template: template was: {.spec.ports[0].nodePort} object given to jsonpath engine was: map[string]interface {}{"apiVersion":"v1", "metadata":map[string]interface {}{"selfLink":"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master", "uid":"9a90fe30-b7fa-11e6-bb1b-42010af00026", "resourceVersion":"1592", "creationTimestamp":"2016-12-01T19:16:04Z", "labels":map[string]interface {}{"app":"redis", "role":"master"}, "name":"redis-master", "namespace":"e2e-tests-kubectl-v2cx3"}, "spec":map[string]interface {}{"selector":map[string]interface {}{"app":"redis", "role":"master"}, "clusterIP":"10.127.249.84", "type":"ClusterIP", "sessionAffinity":"None", "ports":[]interface {}{map[string]interface {}{"protocol":"TCP", "port":6379, "targetPort":"redis-server"}}}, "status":map[string]interface {}{"loadBalancer":map[string]interface {}{}}, "kind":"Service"} error: error executing jsonpath "{.spec.ports[0].nodePort}": nodePort is not found [] <nil> 0xc820759760 exit status 1 <nil> true [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043c8 0xc820404458] [0xa975d0 0xa975d0] 0xc820d42480}: Command stdout: Error executing template: nodePort is not found. Printing more information for debugging the template: template was: {.spec.ports[0].nodePort} object given to jsonpath engine was: map[string]interface {}{"apiVersion":"v1", "metadata":map[string]interface {}{"selfLink":"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master", "uid":"9a90fe30-b7fa-11e6-bb1b-42010af00026", "resourceVersion":"1592", "creationTimestamp":"2016-12-01T19:16:04Z", "labels":map[string]interface {}{"app":"redis", "role":"master"}, "name":"redis-master", "namespace":"e2e-tests-kubectl-v2cx3"}, "spec":map[string]interface {}{"selector":map[string]interface {}{"app":"redis", "role":"master"}, "clusterIP":"10.127.249.84", "type":"ClusterIP", "sessionAffinity":"None", "ports":[]interface {}{map[string]interface {}{"protocol":"TCP", "port":6379, "targetPort":"redis-server"}}}, "status":map[string]interface {}{"loadBalancer":map[string]interface {}{}}, "kind":"Service"} stderr: error: error executing jsonpath "{.spec.ports[0].nodePort}": nodePort is not found error: exit status 1 not to have occurred ``` Issues about this test specifically: #28523 #35741 #37820 Previous issues for this suite: #37747
1.0
kubernetes-e2e-gke-gci-1.3-container_vm-1.5-upgrade-cluster: broken test run - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gke-gci-1.3-container_vm-1.5-upgrade-cluster/339/ Multiple broken tests: Failed: [k8s.io] Deployment deployment should support rollover {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/deployment.go:70 Expected error: <*errors.errorString | 0xc821399680>: { s: "error waiting for deployment test-rollover-deployment status to match expectation: timed out waiting for the condition", } error waiting for deployment test-rollover-deployment status to match expectation: timed out waiting for the condition not to have occurred ``` Issues about this test specifically: #26509 #26834 #29780 #35355 Failed: [k8s.io] V1Job should fail a job {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/batch_v1_jobs.go:202 Expected error: <*errors.errorString | 0xc820019190>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred ``` Issues about this test specifically: #27704 #30127 #30602 #31070 #34383 Failed: [k8s.io] Job should fail a job {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/job.go:198 Expected error: <*errors.errorString | 0xc820019190>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred ``` Issues about this test specifically: #28773 #29506 #30699 #32734 #34585 #37177 Failed: [k8s.io] Kubectl client [k8s.io] Kubectl apply should reuse nodePort when apply to an existing SVC {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/kubectl.go:471 Expected error: <*errors.errorString | 0xc82056f080>: { s: "Error running &{/workspace/kubernetes_skew/cluster/kubectl.sh [/workspace/kubernetes_skew/cluster/kubectl.sh --server=https://104.154.214.26 --kubeconfig=/workspace/.kube/config get service redis-master --namespace=e2e-tests-kubectl-v2cx3 -o jsonpath={.spec.ports[0].nodePort}] [] <nil> Error executing template: nodePort is not found. Printing more information for debugging the template:\n\ttemplate was:\n\t\t{.spec.ports[0].nodePort}\n\tobject given to jsonpath engine was:\n\t\tmap[string]interface {}{\"apiVersion\":\"v1\", \"metadata\":map[string]interface {}{\"selfLink\":\"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master\", \"uid\":\"9a90fe30-b7fa-11e6-bb1b-42010af00026\", \"resourceVersion\":\"1592\", \"creationTimestamp\":\"2016-12-01T19:16:04Z\", \"labels\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"name\":\"redis-master\", \"namespace\":\"e2e-tests-kubectl-v2cx3\"}, \"spec\":map[string]interface {}{\"selector\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"clusterIP\":\"10.127.249.84\", \"type\":\"ClusterIP\", \"sessionAffinity\":\"None\", \"ports\":[]interface {}{map[string]interface {}{\"protocol\":\"TCP\", \"port\":6379, \"targetPort\":\"redis-server\"}}}, \"status\":map[string]interface {}{\"loadBalancer\":map[string]interface {}{}}, \"kind\":\"Service\"}\n\n error: error executing jsonpath \"{.spec.ports[0].nodePort}\": nodePort is not found\n [] <nil> 0xc820759760 exit status 1 <nil> true [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043c8 0xc820404458] [0xa975d0 0xa975d0] 0xc820d42480}:\nCommand stdout:\nError executing template: nodePort is not found. Printing more information for debugging the template:\n\ttemplate was:\n\t\t{.spec.ports[0].nodePort}\n\tobject given to jsonpath engine was:\n\t\tmap[string]interface {}{\"apiVersion\":\"v1\", \"metadata\":map[string]interface {}{\"selfLink\":\"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master\", \"uid\":\"9a90fe30-b7fa-11e6-bb1b-42010af00026\", \"resourceVersion\":\"1592\", \"creationTimestamp\":\"2016-12-01T19:16:04Z\", \"labels\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"name\":\"redis-master\", \"namespace\":\"e2e-tests-kubectl-v2cx3\"}, \"spec\":map[string]interface {}{\"selector\":map[string]interface {}{\"app\":\"redis\", \"role\":\"master\"}, \"clusterIP\":\"10.127.249.84\", \"type\":\"ClusterIP\", \"sessionAffinity\":\"None\", \"ports\":[]interface {}{map[string]interface {}{\"protocol\":\"TCP\", \"port\":6379, \"targetPort\":\"redis-server\"}}}, \"status\":map[string]interface {}{\"loadBalancer\":map[string]interface {}{}}, \"kind\":\"Service\"}\n\n\nstderr:\nerror: error executing jsonpath \"{.spec.ports[0].nodePort}\": nodePort is not found\n\nerror:\nexit status 1\n", } Error running &{/workspace/kubernetes_skew/cluster/kubectl.sh [/workspace/kubernetes_skew/cluster/kubectl.sh --server=https://104.154.214.26 --kubeconfig=/workspace/.kube/config get service redis-master --namespace=e2e-tests-kubectl-v2cx3 -o jsonpath={.spec.ports[0].nodePort}] [] <nil> Error executing template: nodePort is not found. Printing more information for debugging the template: template was: {.spec.ports[0].nodePort} object given to jsonpath engine was: map[string]interface {}{"apiVersion":"v1", "metadata":map[string]interface {}{"selfLink":"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master", "uid":"9a90fe30-b7fa-11e6-bb1b-42010af00026", "resourceVersion":"1592", "creationTimestamp":"2016-12-01T19:16:04Z", "labels":map[string]interface {}{"app":"redis", "role":"master"}, "name":"redis-master", "namespace":"e2e-tests-kubectl-v2cx3"}, "spec":map[string]interface {}{"selector":map[string]interface {}{"app":"redis", "role":"master"}, "clusterIP":"10.127.249.84", "type":"ClusterIP", "sessionAffinity":"None", "ports":[]interface {}{map[string]interface {}{"protocol":"TCP", "port":6379, "targetPort":"redis-server"}}}, "status":map[string]interface {}{"loadBalancer":map[string]interface {}{}}, "kind":"Service"} error: error executing jsonpath "{.spec.ports[0].nodePort}": nodePort is not found [] <nil> 0xc820759760 exit status 1 <nil> true [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043a0 0xc8204043e8 0xc820404470] [0xc8204043c8 0xc820404458] [0xa975d0 0xa975d0] 0xc820d42480}: Command stdout: Error executing template: nodePort is not found. Printing more information for debugging the template: template was: {.spec.ports[0].nodePort} object given to jsonpath engine was: map[string]interface {}{"apiVersion":"v1", "metadata":map[string]interface {}{"selfLink":"/api/v1/namespaces/e2e-tests-kubectl-v2cx3/services/redis-master", "uid":"9a90fe30-b7fa-11e6-bb1b-42010af00026", "resourceVersion":"1592", "creationTimestamp":"2016-12-01T19:16:04Z", "labels":map[string]interface {}{"app":"redis", "role":"master"}, "name":"redis-master", "namespace":"e2e-tests-kubectl-v2cx3"}, "spec":map[string]interface {}{"selector":map[string]interface {}{"app":"redis", "role":"master"}, "clusterIP":"10.127.249.84", "type":"ClusterIP", "sessionAffinity":"None", "ports":[]interface {}{map[string]interface {}{"protocol":"TCP", "port":6379, "targetPort":"redis-server"}}}, "status":map[string]interface {}{"loadBalancer":map[string]interface {}{}}, "kind":"Service"} stderr: error: error executing jsonpath "{.spec.ports[0].nodePort}": nodePort is not found error: exit status 1 not to have occurred ``` Issues about this test specifically: #28523 #35741 #37820 Previous issues for this suite: #37747
non_main
kubernetes gke gci container vm upgrade cluster broken test run multiple broken tests failed deployment deployment should support rollover kubernetes suite go src io kubernetes output dockerized go src io kubernetes test deployment go expected error s error waiting for deployment test rollover deployment status to match expectation timed out waiting for the condition error waiting for deployment test rollover deployment status to match expectation timed out waiting for the condition not to have occurred issues about this test specifically failed should fail a job kubernetes suite go src io kubernetes output dockerized go src io kubernetes test batch jobs go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred issues about this test specifically failed job should fail a job kubernetes suite go src io kubernetes output dockerized go src io kubernetes test job go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred issues about this test specifically failed kubectl client kubectl apply should reuse nodeport when apply to an existing svc kubernetes suite go src io kubernetes output dockerized go src io kubernetes test kubectl go expected error s error running workspace kubernetes skew cluster kubectl sh nodeport error executing template nodeport is not found printing more information for debugging the template n ttemplate was n t t spec ports nodeport n tobject given to jsonpath engine was n t tmap interface apiversion metadata map interface selflink api namespaces tests kubectl services redis master uid resourceversion creationtimestamp labels map interface app redis role master name redis master namespace tests kubectl spec map interface selector map interface app redis role master clusterip type clusterip sessionaffinity none ports interface map interface protocol tcp port targetport redis server status map interface loadbalancer map interface kind service n n error error executing jsonpath spec ports nodeport nodeport is not found n exit status true ncommand stdout nerror executing template nodeport is not found printing more information for debugging the template n ttemplate was n t t spec ports nodeport n tobject given to jsonpath engine was n t tmap interface apiversion metadata map interface selflink api namespaces tests kubectl services redis master uid resourceversion creationtimestamp labels map interface app redis role master name redis master namespace tests kubectl spec map interface selector map interface app redis role master clusterip type clusterip sessionaffinity none ports interface map interface protocol tcp port targetport redis server status map interface loadbalancer map interface kind service n n nstderr nerror error executing jsonpath spec ports nodeport nodeport is not found n nerror nexit status n error running workspace kubernetes skew cluster kubectl sh nodeport error executing template nodeport is not found printing more information for debugging the template template was spec ports nodeport object given to jsonpath engine was map interface apiversion metadata map interface selflink api namespaces tests kubectl services redis master uid resourceversion creationtimestamp labels map interface app redis role master name redis master namespace tests kubectl spec map interface selector map interface app redis role master clusterip type clusterip sessionaffinity none ports interface map interface protocol tcp port targetport redis server status map interface loadbalancer map interface kind service error error executing jsonpath spec ports nodeport nodeport is not found exit status true command stdout error executing template nodeport is not found printing more information for debugging the template template was spec ports nodeport object given to jsonpath engine was map interface apiversion metadata map interface selflink api namespaces tests kubectl services redis master uid resourceversion creationtimestamp labels map interface app redis role master name redis master namespace tests kubectl spec map interface selector map interface app redis role master clusterip type clusterip sessionaffinity none ports interface map interface protocol tcp port targetport redis server status map interface loadbalancer map interface kind service stderr error error executing jsonpath spec ports nodeport nodeport is not found error exit status not to have occurred issues about this test specifically previous issues for this suite
0
5,510
27,513,520,665
IssuesEvent
2023-03-06 10:24:55
toolbx-images/images
https://api.github.com/repos/toolbx-images/images
closed
Add distribution: Alpine Edge
new-image-request maintainers-wanted
### Distribution name and versions requested I'd like to start figuring out the next Alpine release and figured it wouldn't hurt to add the `:edge` container. ### Where are the official container images from the distribution published? https://github.com/alpinelinux/docker-alpine/tree/edge ### Will you be interested in maintaining this image? Sure!
True
Add distribution: Alpine Edge - ### Distribution name and versions requested I'd like to start figuring out the next Alpine release and figured it wouldn't hurt to add the `:edge` container. ### Where are the official container images from the distribution published? https://github.com/alpinelinux/docker-alpine/tree/edge ### Will you be interested in maintaining this image? Sure!
main
add distribution alpine edge distribution name and versions requested i d like to start figuring out the next alpine release and figured it wouldn t hurt to add the edge container where are the official container images from the distribution published will you be interested in maintaining this image sure
1
167,729
26,541,341,777
IssuesEvent
2023-01-19 19:32:14
rupali-codes/LinksHub
https://api.github.com/repos/rupali-codes/LinksHub
opened
[DESIGN] design a landing page for LinksHub in Figma
enhancement good first issue frontend up for grabs design
### What would you like to share or ask? As our project is open source and we don't have a landing page yet, it would be great to have a place where users can learn about the project, its features, and how to get started. This issue is to track the design of a landing page for LinksHub in Figma. ### The landing page should include: - A brief overview of the project and its purpose - A list of key features - Screenshots and/or demo videos to showcase the project - Links to github and documentation to help users get started - A section for contributors to learn how they can contribute to the project - A way for users to contact the development team for support or to report issues > Design and layout should be modern and user-friendly, and should be accessible to users of all abilities. Please provide a Figma link to the design and share it with the development team. Also, let me know if you have any questions or need clarification on any of the above points. > **It's also important to note that the design should be accessible, responsive and mobile-friendly as well.** ### Additional information _No response_
1.0
[DESIGN] design a landing page for LinksHub in Figma - ### What would you like to share or ask? As our project is open source and we don't have a landing page yet, it would be great to have a place where users can learn about the project, its features, and how to get started. This issue is to track the design of a landing page for LinksHub in Figma. ### The landing page should include: - A brief overview of the project and its purpose - A list of key features - Screenshots and/or demo videos to showcase the project - Links to github and documentation to help users get started - A section for contributors to learn how they can contribute to the project - A way for users to contact the development team for support or to report issues > Design and layout should be modern and user-friendly, and should be accessible to users of all abilities. Please provide a Figma link to the design and share it with the development team. Also, let me know if you have any questions or need clarification on any of the above points. > **It's also important to note that the design should be accessible, responsive and mobile-friendly as well.** ### Additional information _No response_
non_main
design a landing page for linkshub in figma what would you like to share or ask as our project is open source and we don t have a landing page yet it would be great to have a place where users can learn about the project its features and how to get started this issue is to track the design of a landing page for linkshub in figma the landing page should include a brief overview of the project and its purpose a list of key features screenshots and or demo videos to showcase the project links to github and documentation to help users get started a section for contributors to learn how they can contribute to the project a way for users to contact the development team for support or to report issues design and layout should be modern and user friendly and should be accessible to users of all abilities please provide a figma link to the design and share it with the development team also let me know if you have any questions or need clarification on any of the above points it s also important to note that the design should be accessible responsive and mobile friendly as well additional information no response
0
248,488
21,035,820,186
IssuesEvent
2022-03-31 07:44:35
SUNET/eduid-front
https://api.github.com/repos/SUNET/eduid-front
closed
VETTING VIA POST: Inform user of letter delivery
standard styling testing copy layout :bulb: coordination
#### Description of issue: Update vetting interface to inform user of letter process (following update of data received from the backend): - when letter was requested (no. of days since code was requested in the GET /proofing state result) - that letter is on it way - if letter has expired - opportunity to make it clear where to input confirmation code once letter has arrived ###### This issue is part of [issue 368](https://github.com/SUNET/eduid-webapp/issues/368) in eduid-webapp and requires coordination with member of the backend team.
1.0
VETTING VIA POST: Inform user of letter delivery - #### Description of issue: Update vetting interface to inform user of letter process (following update of data received from the backend): - when letter was requested (no. of days since code was requested in the GET /proofing state result) - that letter is on it way - if letter has expired - opportunity to make it clear where to input confirmation code once letter has arrived ###### This issue is part of [issue 368](https://github.com/SUNET/eduid-webapp/issues/368) in eduid-webapp and requires coordination with member of the backend team.
non_main
vetting via post inform user of letter delivery description of issue update vetting interface to inform user of letter process following update of data received from the backend when letter was requested no of days since code was requested in the get proofing state result that letter is on it way if letter has expired opportunity to make it clear where to input confirmation code once letter has arrived this issue is part of in eduid webapp and requires coordination with member of the backend team
0
165,924
6,288,282,638
IssuesEvent
2017-07-19 16:37:42
motorcyclets/motorcycle
https://api.github.com/repos/motorcyclets/motorcycle
closed
feat: switchCombine
Priority: 3 Scope: stream Size: 2 Type: Feature
```typescript export function switchCombine<A>(stream: Stream<Array<Stream<A>>>): Stream<Array<A>> ``` - must accept `ArrayLike` and `ReadonlyArray`
1.0
feat: switchCombine - ```typescript export function switchCombine<A>(stream: Stream<Array<Stream<A>>>): Stream<Array<A>> ``` - must accept `ArrayLike` and `ReadonlyArray`
non_main
feat switchcombine typescript export function switchcombine stream stream stream must accept arraylike and readonlyarray
0
738,936
25,574,587,429
IssuesEvent
2022-11-30 20:49:48
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
asset.snippets.quickstart_getfeed_test: test_get_feed failed
priority: p1 type: bug flakybot: issue flakybot: flaky
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 07e8b8145670fc1f9f9ed99e348d8fe28ca2ca7e buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e3c01052-a4f3-49b3-b93e-6123c271a1b0), [Sponge](http://sponge2/e3c01052-a4f3-49b3-b93e-6123c271a1b0) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.197.95:443 {created_time:"2022-11-30T11:14:46.152776564+00:00", grpc_status:16, grpc_message:"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project."}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/asset/snippets/conftest.py", line 72, in test_feed feed = create_feed() File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/backoff/_sync.py", line 105, in retry ret = target(*args, **kwargs) File "/workspace/asset/snippets/conftest.py", line 69, in create_feed PROJECT, feed_id, [asset_name], test_topic.name, asset_v1.ContentType.RESOURCE File "/workspace/asset/snippets/quickstart_createfeed.py", line 38, in create_feed request={"parent": parent, "feed_id": feed_id, "feed": feed} File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/asset_v1/services/asset_service/client.py", line 961, in create_feed metadata=metadata, File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.</pre></details>
1.0
asset.snippets.quickstart_getfeed_test: test_get_feed failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: 07e8b8145670fc1f9f9ed99e348d8fe28ca2ca7e buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e3c01052-a4f3-49b3-b93e-6123c271a1b0), [Sponge](http://sponge2/e3c01052-a4f3-49b3-b93e-6123c271a1b0) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNAUTHENTICATED details = "Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project." debug_error_string = "UNKNOWN:Error received from peer ipv4:74.125.197.95:443 {created_time:"2022-11-30T11:14:46.152776564+00:00", grpc_status:16, grpc_message:"Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project."}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/asset/snippets/conftest.py", line 72, in test_feed feed = create_feed() File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/backoff/_sync.py", line 105, in retry ret = target(*args, **kwargs) File "/workspace/asset/snippets/conftest.py", line 69, in create_feed PROJECT, feed_id, [asset_name], test_topic.name, asset_v1.ContentType.RESOURCE File "/workspace/asset/snippets/quickstart_createfeed.py", line 38, in create_feed request={"parent": parent, "feed_id": feed_id, "feed": feed} File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/cloud/asset_v1/services/asset_service/client.py", line 961, in create_feed metadata=metadata, File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/asset/snippets/.nox/py-3-7/lib/python3.7/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.Unauthenticated: 401 Request had invalid authentication credentials. Expected OAuth 2 access token, login cookie or other valid authentication credential. See https://developers.google.com/identity/sign-in/web/devconsole-project.</pre></details>
non_main
asset snippets quickstart getfeed test test get feed failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace asset snippets nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace asset snippets nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace asset snippets nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unauthenticated details request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see debug error string unknown error received from peer created time grpc status grpc message request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see the above exception was the direct cause of the following exception traceback most recent call last file workspace asset snippets conftest py line in test feed feed create feed file workspace asset snippets nox py lib site packages backoff sync py line in retry ret target args kwargs file workspace asset snippets conftest py line in create feed project feed id test topic name asset contenttype resource file workspace asset snippets quickstart createfeed py line in create feed request parent parent feed id feed id feed feed file workspace asset snippets nox py lib site packages google cloud asset services asset service client py line in create feed metadata metadata file workspace asset snippets nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace asset snippets nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions unauthenticated request had invalid authentication credentials expected oauth access token login cookie or other valid authentication credential see
0
305,836
9,377,613,270
IssuesEvent
2019-04-04 10:45:48
kookmin-sw/2019-cap1-2019_5
https://api.github.com/repos/kookmin-sw/2019-cap1-2019_5
opened
Add location candidates template json file
Priority-Medium enhancement
This will contain: - add location candidate template json file - upload original location candidate json file on Google Drive for sharing
1.0
Add location candidates template json file - This will contain: - add location candidate template json file - upload original location candidate json file on Google Drive for sharing
non_main
add location candidates template json file this will contain add location candidate template json file upload original location candidate json file on google drive for sharing
0
5,731
30,292,610,321
IssuesEvent
2023-07-09 13:14:58
svengreb/wand
https://api.github.com/repos/svengreb/wand
closed
`go run` support for versioned modules (Go 1.17+)
context-api context-pkg scope-compatibility scope-dx scope-maintainability scope-stability type-feature
[As of Go 1.17 the `go run` command can finally run in module-aware mode][1] while not “polluting“ the current module in the working directory, if there is one (`go.mod` file present) 🎉 This finally allows to [run commands _on-the-fly_][10] of Go `main` module packages without installing them or without changing dependencies of the current module! To support this feature with _wand_ a new [`task.GoModule`][2] will be implemented in a new [`golang/run`][3] package. It can be run using a [command runner][4] that handles tasks of kind [`KindGoModule`][5] so mainly [`gotool.Runner`][6]. The new [`golang/run.Task`][3] will be customizable through the following functions: - `WithArgs(...string) run.Option` — sets additional arguments to pass to the command. - `WithEnv(map[string]string) run.Option` — sets the task specific environment. - `WithModulePath(string) run.Option` — sets the module import path. - `WithModuleVersion(*semver.Version) run.Option` — sets the module version. Next to the new task the [`gotool.Runner`][6] will be extended with a new [`WithCache(bool)`][9] runner option to toggle the usage of the local cache directory in the root directory of the module. The runner will be made “smart“ in the way that it either… - installing the executable through a [`golang.Runner`][8], which runs `go install pkg@version` to [leverage Go 1.16‘s feature][12], and execute it afterwards. This is the current default behavior of this runner which will be used when [`WithCache(true)`][9] is used. - pass the task to a [`golang.Runner`][8], using the new [`golang/run`][3] package task, so that it can run `go run pkg@version <args>` instead. This is the new “smart“ behavior of the runner which will be used when [`WithCache(false)`][9] (default) is used. The **new default behavior will be to not use a local cache** so that caching will be a opt-in. This decision was made because native support for running commands _on-the-fly_ should always be preferred to custom logic which is what the local cache directory and [`gotool.Runner`][6] purpose is. > **Warning** > Note that the minimum Go version for task runners, the new [`golang/run` task][3] and [the _Elder_ wand][7] will be increased to `1.17.0` since this version initially [introduced `go run` support in module-aware mode][1]! > This will be enforced through a [build constraint][11] (`go:build go1.17`). The [`Elder`][7] reference implementation will also adapt to this new feature by… 1. **deprecating the `*elder.Elder.Bootstrap(...string) []error` method**! As of _wand_ version `0.9.0` it will be a no-op and will be removed in version `0.10.0`. To install executables anyway the new `*elder.Elder.CacheExecutables error` method should be used instead. To ensure that the wand is properly initialized and operational the `*elder.Elder.Validate(..task.Runner) []error` method is the way to go. A warning message will be printed when the method is called to ensure that users adapt accordionally. 2. providing a new `*elder.Elder.CacheExecutables(...string) error` method which allows to pass paths of Go modules that should be explicitly installed to the local cache directory. This method is a kind of workaround for the, now deprecated, `*elder.Elder.Bootstrap(...string) []error` method to allows users to still cache command executables locally. 3. changing the signature of the `*elder.Elder.Validate() error` method to `*elder.Elder.Validate(...task.Runner) []error` method which allows users to ensure that the _wand_ is properly initialized and operational. Optionally [command runner][4] can be passed that will be validated while passing nothing will validate all currently supported runners. [1]: https://go.dev/doc/go1.17#go%20run [2]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#GoModule [3]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run [4]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#Runner [5]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#KindGoModule [6]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/gotool#Runner [7]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/elder [8]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang#Runner [9]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run#WithCache [10]: https://pkg.go.dev/cmd/go#hdr-Compile_and_run_Go_program [11]: https://pkg.go.dev/cmd/go#hdr-Build_constraints [12]: https://github.com/svengreb/wand/issues/89
True
`go run` support for versioned modules (Go 1.17+) - [As of Go 1.17 the `go run` command can finally run in module-aware mode][1] while not “polluting“ the current module in the working directory, if there is one (`go.mod` file present) 🎉 This finally allows to [run commands _on-the-fly_][10] of Go `main` module packages without installing them or without changing dependencies of the current module! To support this feature with _wand_ a new [`task.GoModule`][2] will be implemented in a new [`golang/run`][3] package. It can be run using a [command runner][4] that handles tasks of kind [`KindGoModule`][5] so mainly [`gotool.Runner`][6]. The new [`golang/run.Task`][3] will be customizable through the following functions: - `WithArgs(...string) run.Option` — sets additional arguments to pass to the command. - `WithEnv(map[string]string) run.Option` — sets the task specific environment. - `WithModulePath(string) run.Option` — sets the module import path. - `WithModuleVersion(*semver.Version) run.Option` — sets the module version. Next to the new task the [`gotool.Runner`][6] will be extended with a new [`WithCache(bool)`][9] runner option to toggle the usage of the local cache directory in the root directory of the module. The runner will be made “smart“ in the way that it either… - installing the executable through a [`golang.Runner`][8], which runs `go install pkg@version` to [leverage Go 1.16‘s feature][12], and execute it afterwards. This is the current default behavior of this runner which will be used when [`WithCache(true)`][9] is used. - pass the task to a [`golang.Runner`][8], using the new [`golang/run`][3] package task, so that it can run `go run pkg@version <args>` instead. This is the new “smart“ behavior of the runner which will be used when [`WithCache(false)`][9] (default) is used. The **new default behavior will be to not use a local cache** so that caching will be a opt-in. This decision was made because native support for running commands _on-the-fly_ should always be preferred to custom logic which is what the local cache directory and [`gotool.Runner`][6] purpose is. > **Warning** > Note that the minimum Go version for task runners, the new [`golang/run` task][3] and [the _Elder_ wand][7] will be increased to `1.17.0` since this version initially [introduced `go run` support in module-aware mode][1]! > This will be enforced through a [build constraint][11] (`go:build go1.17`). The [`Elder`][7] reference implementation will also adapt to this new feature by… 1. **deprecating the `*elder.Elder.Bootstrap(...string) []error` method**! As of _wand_ version `0.9.0` it will be a no-op and will be removed in version `0.10.0`. To install executables anyway the new `*elder.Elder.CacheExecutables error` method should be used instead. To ensure that the wand is properly initialized and operational the `*elder.Elder.Validate(..task.Runner) []error` method is the way to go. A warning message will be printed when the method is called to ensure that users adapt accordionally. 2. providing a new `*elder.Elder.CacheExecutables(...string) error` method which allows to pass paths of Go modules that should be explicitly installed to the local cache directory. This method is a kind of workaround for the, now deprecated, `*elder.Elder.Bootstrap(...string) []error` method to allows users to still cache command executables locally. 3. changing the signature of the `*elder.Elder.Validate() error` method to `*elder.Elder.Validate(...task.Runner) []error` method which allows users to ensure that the _wand_ is properly initialized and operational. Optionally [command runner][4] can be passed that will be validated while passing nothing will validate all currently supported runners. [1]: https://go.dev/doc/go1.17#go%20run [2]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#GoModule [3]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run [4]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#Runner [5]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task#KindGoModule [6]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/gotool#Runner [7]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/elder [8]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang#Runner [9]: https://pkg.go.dev/github.com/svengreb/wand@v0.9.0/pkg/task/golang/run#WithCache [10]: https://pkg.go.dev/cmd/go#hdr-Compile_and_run_Go_program [11]: https://pkg.go.dev/cmd/go#hdr-Build_constraints [12]: https://github.com/svengreb/wand/issues/89
main
go run support for versioned modules go while not “polluting“ the current module in the working directory if there is one go mod file present 🎉 this finally allows to of go main module packages without installing them or without changing dependencies of the current module to support this feature with wand a new will be implemented in a new package it can be run using a that handles tasks of kind so mainly the new will be customizable through the following functions withargs string run option — sets additional arguments to pass to the command withenv map string run option — sets the task specific environment withmodulepath string run option — sets the module import path withmoduleversion semver version run option — sets the module version next to the new task the will be extended with a new runner option to toggle the usage of the local cache directory in the root directory of the module the runner will be made “smart“ in the way that it either… installing the executable through a which runs go install pkg version to and execute it afterwards this is the current default behavior of this runner which will be used when is used pass the task to a using the new package task so that it can run go run pkg version instead this is the new “smart“ behavior of the runner which will be used when default is used the new default behavior will be to not use a local cache so that caching will be a opt in this decision was made because native support for running commands on the fly should always be preferred to custom logic which is what the local cache directory and purpose is warning note that the minimum go version for task runners the new and will be increased to since this version initially this will be enforced through a go build the reference implementation will also adapt to this new feature by… deprecating the elder elder bootstrap string error method as of wand version it will be a no op and will be removed in version to install executables anyway the new elder elder cacheexecutables error method should be used instead to ensure that the wand is properly initialized and operational the elder elder validate task runner error method is the way to go a warning message will be printed when the method is called to ensure that users adapt accordionally providing a new elder elder cacheexecutables string error method which allows to pass paths of go modules that should be explicitly installed to the local cache directory this method is a kind of workaround for the now deprecated elder elder bootstrap string error method to allows users to still cache command executables locally changing the signature of the elder elder validate error method to elder elder validate task runner error method which allows users to ensure that the wand is properly initialized and operational optionally can be passed that will be validated while passing nothing will validate all currently supported runners
1
68,295
21,610,472,394
IssuesEvent
2022-05-04 09:34:29
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Romanian phone numbers not supported
T-Defect S-Minor A-User-Settings I18n Z-Backend Z-Infra O-Uncommon
### Steps to reproduce 1. Go to your account (matrix.org/vector.im) settings / general / phone numbers 2. Select Romania from the list 3. Insert a valid number e.g. `741 000 000` 4. Receive `country not supported error` ### Outcome #### What did you expect? For the phone number to be added and to receive a confirmation SMS as per https://www.openmarket.com/docs/Content/globalcoverage/country/romania.htm I also tried this with my own personal phone number (which can receive SMSs without problems) with the same results. #### What happened instead? `Errror` - `Phone numbers in this country are not currently supported (Status 400)` ![Screenshot 2022-01-13 at 11 06 47](https://user-images.githubusercontent.com/637564/149300171-40f97ebd-8c99-4ee8-82a9-70a128f1dbe3.png) ### Operating system macOS ### Browser information Chrome Version 97.0.4692.71 ### URL for webapp app.element.io ### Application version Element version: 1.9.8 Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? No
1.0
Romanian phone numbers not supported - ### Steps to reproduce 1. Go to your account (matrix.org/vector.im) settings / general / phone numbers 2. Select Romania from the list 3. Insert a valid number e.g. `741 000 000` 4. Receive `country not supported error` ### Outcome #### What did you expect? For the phone number to be added and to receive a confirmation SMS as per https://www.openmarket.com/docs/Content/globalcoverage/country/romania.htm I also tried this with my own personal phone number (which can receive SMSs without problems) with the same results. #### What happened instead? `Errror` - `Phone numbers in this country are not currently supported (Status 400)` ![Screenshot 2022-01-13 at 11 06 47](https://user-images.githubusercontent.com/637564/149300171-40f97ebd-8c99-4ee8-82a9-70a128f1dbe3.png) ### Operating system macOS ### Browser information Chrome Version 97.0.4692.71 ### URL for webapp app.element.io ### Application version Element version: 1.9.8 Olm version: 3.2.8 ### Homeserver matrix.org ### Will you send logs? No
non_main
romanian phone numbers not supported steps to reproduce go to your account matrix org vector im settings general phone numbers select romania from the list insert a valid number e g receive country not supported error outcome what did you expect for the phone number to be added and to receive a confirmation sms as per i also tried this with my own personal phone number which can receive smss without problems with the same results what happened instead errror phone numbers in this country are not currently supported status operating system macos browser information chrome version url for webapp app element io application version element version olm version homeserver matrix org will you send logs no
0
86,375
8,033,705,122
IssuesEvent
2018-07-29 09:50:08
meateam/seal
https://api.github.com/repos/meateam/seal
opened
Mongo indexing during testing
Testing bug?
Ensure that the indexing works after `mongoose.connection.db.dropDatabase`. Following the problems in `rabiran/Kartoffel`.
1.0
Mongo indexing during testing - Ensure that the indexing works after `mongoose.connection.db.dropDatabase`. Following the problems in `rabiran/Kartoffel`.
non_main
mongo indexing during testing ensure that the indexing works after mongoose connection db dropdatabase following the problems in rabiran kartoffel
0
662,482
22,140,812,786
IssuesEvent
2022-06-03 06:38:46
sapa/performing-arts-ch-templates
https://api.github.com/repos/sapa/performing-arts-ch-templates
opened
Form:[all] Research by URI | no 2
New Feature Priority B
See also : #17 #88 Es muss möglich sein in alle Form durch URI suchen UND durch Text. ## form:Record Set - Field:Location->rdf:type spao:ArchivalPlace - Field:Place-> URI /p/ ## form:Record - Field:Location->rdf:type spao:ArchivalPlace - Field:Place-> URI /p/ ### Record parts ## form:InstantationPhysical - Field:Location->rdf:type spao:ArchivalPlace ## form:InstantationVirtual - Field:Location->rdf:type spao:ArchivalPlace ### File ## form:Archival Location - Parent ->rdf:type spao:ArchivalPlace ## form:Place - Parent ->URI /p/ =! rdf:type spao:ArchivalPlace ## form:Venue - Location ->URI /p/ =! rdf:type spao:ArchivalPlace - Contains ->URI /p/ =! rdf:type spao:ArchivalPlace ## form:Group - Current or former residence->URI /p/ =! rdf:type spao:ArchivalPlace ## form:Person - Lieu de naissance->URI /p/ =! rdf:type spao:ArchivalPlace - Lieu de décès->URI /p/ =! rdf:type spao:ArchivalPlace
1.0
Form:[all] Research by URI | no 2 - See also : #17 #88 Es muss möglich sein in alle Form durch URI suchen UND durch Text. ## form:Record Set - Field:Location->rdf:type spao:ArchivalPlace - Field:Place-> URI /p/ ## form:Record - Field:Location->rdf:type spao:ArchivalPlace - Field:Place-> URI /p/ ### Record parts ## form:InstantationPhysical - Field:Location->rdf:type spao:ArchivalPlace ## form:InstantationVirtual - Field:Location->rdf:type spao:ArchivalPlace ### File ## form:Archival Location - Parent ->rdf:type spao:ArchivalPlace ## form:Place - Parent ->URI /p/ =! rdf:type spao:ArchivalPlace ## form:Venue - Location ->URI /p/ =! rdf:type spao:ArchivalPlace - Contains ->URI /p/ =! rdf:type spao:ArchivalPlace ## form:Group - Current or former residence->URI /p/ =! rdf:type spao:ArchivalPlace ## form:Person - Lieu de naissance->URI /p/ =! rdf:type spao:ArchivalPlace - Lieu de décès->URI /p/ =! rdf:type spao:ArchivalPlace
non_main
form research by uri no see also es muss möglich sein in alle form durch uri suchen und durch text form record set field location rdf type spao archivalplace field place uri p form record field location rdf type spao archivalplace field place uri p record parts form instantationphysical field location rdf type spao archivalplace form instantationvirtual field location rdf type spao archivalplace file form archival location parent rdf type spao archivalplace form place parent uri p rdf type spao archivalplace form venue location uri p rdf type spao archivalplace contains uri p rdf type spao archivalplace form group current or former residence uri p rdf type spao archivalplace form person lieu de naissance uri p rdf type spao archivalplace lieu de décès uri p rdf type spao archivalplace
0
3,120
11,935,420,070
IssuesEvent
2020-04-02 08:33:09
OpenLightingProject/ola
https://api.github.com/repos/OpenLightingProject/ola
closed
autoreconf error AC_CONFIG_MACRO_DIRS
Maintainability OpSys-Linux question
I just tried building from source at the current master, and hit a hurdle at the very first step: $ autoreconf -i libtoolize: putting auxiliary files in AC_CONFIG_AUX_DIR, 'config'. libtoolize: copying file 'config/ltmain.sh' .ibtoolize: error: AC_CONFIG_MACRO_DIRS([config]) conflicts with ACLOCAL_AMFLAGS=-I config autoreconf: libtoolize failed with exit status: 1 After commenting out line 13 in `Makefile.am`, `autoreconf` and `./configure` ran through from start to finish, down to "Now type 'make [<target>]' ".
True
autoreconf error AC_CONFIG_MACRO_DIRS - I just tried building from source at the current master, and hit a hurdle at the very first step: $ autoreconf -i libtoolize: putting auxiliary files in AC_CONFIG_AUX_DIR, 'config'. libtoolize: copying file 'config/ltmain.sh' .ibtoolize: error: AC_CONFIG_MACRO_DIRS([config]) conflicts with ACLOCAL_AMFLAGS=-I config autoreconf: libtoolize failed with exit status: 1 After commenting out line 13 in `Makefile.am`, `autoreconf` and `./configure` ran through from start to finish, down to "Now type 'make [<target>]' ".
main
autoreconf error ac config macro dirs i just tried building from source at the current master and hit a hurdle at the very first step autoreconf i libtoolize putting auxiliary files in ac config aux dir config libtoolize copying file config ltmain sh ibtoolize error ac config macro dirs conflicts with aclocal amflags i config autoreconf libtoolize failed with exit status after commenting out line in makefile am autoreconf and configure ran through from start to finish down to now type make
1
14,912
26,035,909,035
IssuesEvent
2022-12-22 04:56:33
seleniumbase/SeleniumBase
https://api.github.com/repos/seleniumbase/SeleniumBase
closed
Remove "options.headless" usage before SeleniumHQ deprecates it
requirements SeleniumBase 4
### Remove ``options.headless`` usage before SeleniumHQ deprecates it. This is regarding: https://github.com/SeleniumHQ/selenium/issues/11467 Instead of setting ``options.headless``, use ``options.add_argument("--headless")`` or ``options.add_argument("--headless=chrome")``. (For Chrome and Edge) This is mostly done. There were a few cases where both ways were being used together to set headless mode.
1.0
Remove "options.headless" usage before SeleniumHQ deprecates it - ### Remove ``options.headless`` usage before SeleniumHQ deprecates it. This is regarding: https://github.com/SeleniumHQ/selenium/issues/11467 Instead of setting ``options.headless``, use ``options.add_argument("--headless")`` or ``options.add_argument("--headless=chrome")``. (For Chrome and Edge) This is mostly done. There were a few cases where both ways were being used together to set headless mode.
non_main
remove options headless usage before seleniumhq deprecates it remove options headless usage before seleniumhq deprecates it this is regarding instead of setting options headless use options add argument headless or options add argument headless chrome for chrome and edge this is mostly done there were a few cases where both ways were being used together to set headless mode
0
5,650
28,502,857,058
IssuesEvent
2023-04-18 18:47:08
ansible-collections/community.general
https://api.github.com/repos/ansible-collections/community.general
closed
redhat_subscription: status or print state needed
feature module has_pr plugins os packaging needs_maintainer
### Summary The module would benefit from a way to print the registration status/state (Im using identity) for when conditionals to prevent "Re-Registration" attempts. Currently, I am using 'subscription-manager identity' shell command to get the status for my when condition but It would be nice to have all this view functionality inside the module itself. ### Issue Type Feature Idea ### Component Name redhat_subscription ### Additional Information <!--- Paste example playbooks or commands between quotes below --> ``` `- name: "[RHSM] Check RH subscription status" shell: subscription-manager identity register: subscription_check ignore_errors: true - include: register.yml when: - ansible_distribution == "RedHat" - "'not yet registered' in subscription_check.stderr" ` ``` ### Code of Conduct - [X] I agree to follow the Ansible Code of Conduct
True
redhat_subscription: status or print state needed - ### Summary The module would benefit from a way to print the registration status/state (Im using identity) for when conditionals to prevent "Re-Registration" attempts. Currently, I am using 'subscription-manager identity' shell command to get the status for my when condition but It would be nice to have all this view functionality inside the module itself. ### Issue Type Feature Idea ### Component Name redhat_subscription ### Additional Information <!--- Paste example playbooks or commands between quotes below --> ``` `- name: "[RHSM] Check RH subscription status" shell: subscription-manager identity register: subscription_check ignore_errors: true - include: register.yml when: - ansible_distribution == "RedHat" - "'not yet registered' in subscription_check.stderr" ` ``` ### Code of Conduct - [X] I agree to follow the Ansible Code of Conduct
main
redhat subscription status or print state needed summary the module would benefit from a way to print the registration status state im using identity for when conditionals to prevent re registration attempts currently i am using subscription manager identity shell command to get the status for my when condition but it would be nice to have all this view functionality inside the module itself issue type feature idea component name redhat subscription additional information name check rh subscription status shell subscription manager identity register subscription check ignore errors true include register yml when ansible distribution redhat not yet registered in subscription check stderr code of conduct i agree to follow the ansible code of conduct
1
5,150
26,249,201,643
IssuesEvent
2023-01-05 17:41:41
cosmos/ibc-rs
https://api.github.com/repos/cosmos/ibc-rs
opened
Make `Msg*` structs `pub(crate)`
A: good first issue O: maintainability
This will allow us to make changes to our message domain types with on breaking change.
True
Make `Msg*` structs `pub(crate)` - This will allow us to make changes to our message domain types with on breaking change.
main
make msg structs pub crate this will allow us to make changes to our message domain types with on breaking change
1
687
4,233,101,914
IssuesEvent
2016-07-05 05:55:41
Microsoft/DirectXTK
https://api.github.com/repos/Microsoft/DirectXTK
reopened
Retire Windows 8.1 Store and Windows phone 8.1 projects
maintainence
At some point we should remove support for these older versions in favor of UWP apps ``DirectXTK_Windows81.vcxproj`` ``DirectXTK_WindowsPhone81.vcxproj`` ``DirectXTK_XAMLSilverlight_WindowsPhone81.vcxproj`` Please put any requests for continued support for one or more of these here.
True
Retire Windows 8.1 Store and Windows phone 8.1 projects - At some point we should remove support for these older versions in favor of UWP apps ``DirectXTK_Windows81.vcxproj`` ``DirectXTK_WindowsPhone81.vcxproj`` ``DirectXTK_XAMLSilverlight_WindowsPhone81.vcxproj`` Please put any requests for continued support for one or more of these here.
main
retire windows store and windows phone projects at some point we should remove support for these older versions in favor of uwp apps directxtk vcxproj directxtk vcxproj directxtk xamlsilverlight vcxproj please put any requests for continued support for one or more of these here
1
1,501
6,496,530,120
IssuesEvent
2017-08-22 10:24:43
sapcc/maia
https://api.github.com/repos/sapcc/maia
closed
Metrics
maintainability
Provide some essential metrics for monitoring. WeaveWorks has some nice tools for instrumenting Go web apps: https://www.weave.works/blog/of-metrics-and-middleware/
True
Metrics - Provide some essential metrics for monitoring. WeaveWorks has some nice tools for instrumenting Go web apps: https://www.weave.works/blog/of-metrics-and-middleware/
main
metrics provide some essential metrics for monitoring weaveworks has some nice tools for instrumenting go web apps
1
4,486
23,375,624,130
IssuesEvent
2022-08-11 02:28:06
restqa/restqa
https://api.github.com/repos/restqa/restqa
closed
Ignoring the report files generated by restqa
enhancement wontfix pair with maintainer
Hello 👋, ### 👀 Background Currently, after the RestQA initialization (command: `restqa init`), the user use the default configuration that includes the html reporter and the json file reporter. Example: ``` version: 0.0.1 metadata: code: JSON-PLACEHOLDER name: jsonplaceholder from typecode description: An example application to demo RestQA Features environments: - name: local default: true plugins: - name: "@restqa/restqapi" config: url: https://jsonplaceholder.typicode.com outputs: - type: html enabled: true - type: file enabled: true config: path: restqa-result.json ``` It mean RestQA will generate 2 kind of report: html + json ### ✌️ What is the actual behavior? When the user is using RestQA for the first time, 2 file related to the report will be generated: 1. the folder `report` 2. the file `restqa-result.json` However those files are just reporting information. **They should be ignore from the version control tool (GIT)** ### 🕵️‍♀️ How to reproduce the current behavior? 1. Install RestQA 2. Run the command `restqa init` to initialize a brand new project 3. Run the command `restqa run` to create the first report 4. Run the command `git status` You will see the `report ` folder and the file `restqa-result.json` appear on the list of the git status. ### 🤞 What is the expected behavior? When the user will would like to commit his changes to the version control, the files `report` and `restqa-result.json` should be ignored. ### 😎 Proposed solution. During the initialization of restqa thought the command `restqa init` we should check if the file `./.gitignore` already exists or create it if it doesn't exists. Then add the following lines into it. ``` ./report ./restqa-result.json ``` Cheers.
True
Ignoring the report files generated by restqa - Hello 👋, ### 👀 Background Currently, after the RestQA initialization (command: `restqa init`), the user use the default configuration that includes the html reporter and the json file reporter. Example: ``` version: 0.0.1 metadata: code: JSON-PLACEHOLDER name: jsonplaceholder from typecode description: An example application to demo RestQA Features environments: - name: local default: true plugins: - name: "@restqa/restqapi" config: url: https://jsonplaceholder.typicode.com outputs: - type: html enabled: true - type: file enabled: true config: path: restqa-result.json ``` It mean RestQA will generate 2 kind of report: html + json ### ✌️ What is the actual behavior? When the user is using RestQA for the first time, 2 file related to the report will be generated: 1. the folder `report` 2. the file `restqa-result.json` However those files are just reporting information. **They should be ignore from the version control tool (GIT)** ### 🕵️‍♀️ How to reproduce the current behavior? 1. Install RestQA 2. Run the command `restqa init` to initialize a brand new project 3. Run the command `restqa run` to create the first report 4. Run the command `git status` You will see the `report ` folder and the file `restqa-result.json` appear on the list of the git status. ### 🤞 What is the expected behavior? When the user will would like to commit his changes to the version control, the files `report` and `restqa-result.json` should be ignored. ### 😎 Proposed solution. During the initialization of restqa thought the command `restqa init` we should check if the file `./.gitignore` already exists or create it if it doesn't exists. Then add the following lines into it. ``` ./report ./restqa-result.json ``` Cheers.
main
ignoring the report files generated by restqa hello 👋 👀 background currently after the restqa initialization command restqa init the user use the default configuration that includes the html reporter and the json file reporter example version metadata code json placeholder name jsonplaceholder from typecode description an example application to demo restqa features environments name local default true plugins name restqa restqapi config url outputs type html enabled true type file enabled true config path restqa result json it mean restqa will generate kind of report html json ✌️ what is the actual behavior when the user is using restqa for the first time file related to the report will be generated the folder report the file restqa result json however those files are just reporting information they should be ignore from the version control tool git 🕵️‍♀️ how to reproduce the current behavior install restqa run the command restqa init to initialize a brand new project run the command restqa run to create the first report run the command git status you will see the report folder and the file restqa result json appear on the list of the git status 🤞 what is the expected behavior when the user will would like to commit his changes to the version control the files report and restqa result json should be ignored 😎 proposed solution during the initialization of restqa thought the command restqa init we should check if the file gitignore already exists or create it if it doesn t exists then add the following lines into it report restqa result json cheers
1
13,153
4,643,485,984
IssuesEvent
2016-09-30 13:39:24
rubberduck-vba/Rubberduck
https://api.github.com/repos/rubberduck-vba/Rubberduck
opened
Inspection for redundant type conversions
code-inspections enhancement
If variables and assignments are strongly typed, then *some* explicit conversions will be unnecessary and redundant and unnecessary... Rubberduck should be able to find these redundant Type conversions. ```vb Dim aa As String Dim bb As String Dim cc As String aa = CStr("abc") '"abc" bb = CStr(aa) 'aa cc = CStr(Right$(CStr(bb), 2)) 'Right$(bb, 2) ```
1.0
Inspection for redundant type conversions - If variables and assignments are strongly typed, then *some* explicit conversions will be unnecessary and redundant and unnecessary... Rubberduck should be able to find these redundant Type conversions. ```vb Dim aa As String Dim bb As String Dim cc As String aa = CStr("abc") '"abc" bb = CStr(aa) 'aa cc = CStr(Right$(CStr(bb), 2)) 'Right$(bb, 2) ```
non_main
inspection for redundant type conversions if variables and assignments are strongly typed then some explicit conversions will be unnecessary and redundant and unnecessary rubberduck should be able to find these redundant type conversions vb dim aa as string dim bb as string dim cc as string aa cstr abc abc bb cstr aa aa cc cstr right cstr bb right bb
0
309,478
23,296,933,918
IssuesEvent
2022-08-06 18:25:08
openscientia/terraform-provider-atlassian
https://api.github.com/repos/openscientia/terraform-provider-atlassian
closed
Update YAML frontmatter in `atlassian_jira_issue_type_scheme` markdown template
documentation enhancement jira/issuetypeschemes
### Terraform CLI and Provider Versions v0.1.0 ### New or Affected Resource(s) - atlassian_jira_issue_type_scheme ### Use Cases or Problem Statement The [markdown template](https://github.com/openscientia/terraform-provider-atlassian/blob/main/templates/resources/jira_issue_type_scheme.md.tmpl) used to generate documentation for `atlassian_jira_issue_type_scheme` resources contains an incorrect `description` in the YAML frontmatter. ```yaml description: |- Manages {{ .Type }}. --- ``` ### Proposal The `description` in the YAML frontmatter should be as follows: ```yaml description: |- Manages {{ .Name }}. --- ``` ### How much impact is this issue causing? Low ### Additional Information _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
Update YAML frontmatter in `atlassian_jira_issue_type_scheme` markdown template - ### Terraform CLI and Provider Versions v0.1.0 ### New or Affected Resource(s) - atlassian_jira_issue_type_scheme ### Use Cases or Problem Statement The [markdown template](https://github.com/openscientia/terraform-provider-atlassian/blob/main/templates/resources/jira_issue_type_scheme.md.tmpl) used to generate documentation for `atlassian_jira_issue_type_scheme` resources contains an incorrect `description` in the YAML frontmatter. ```yaml description: |- Manages {{ .Type }}. --- ``` ### Proposal The `description` in the YAML frontmatter should be as follows: ```yaml description: |- Manages {{ .Name }}. --- ``` ### How much impact is this issue causing? Low ### Additional Information _No response_ ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
non_main
update yaml frontmatter in atlassian jira issue type scheme markdown template terraform cli and provider versions new or affected resource s atlassian jira issue type scheme use cases or problem statement the used to generate documentation for atlassian jira issue type scheme resources contains an incorrect description in the yaml frontmatter yaml description manages type proposal the description in the yaml frontmatter should be as follows yaml description manages name how much impact is this issue causing low additional information no response code of conduct i agree to follow this project s code of conduct
0
153,544
19,708,252,257
IssuesEvent
2022-01-13 01:16:09
amccool/AngularASPNETCore2WebApiAuth
https://api.github.com/repos/amccool/AngularASPNETCore2WebApiAuth
opened
CVE-2020-28469 (High) detected in glob-parent-2.0.0.tgz
security vulnerability
## CVE-2020-28469 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p> <p>Path to dependency file: /src/package.json</p> <p>Path to vulnerable library: /src/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - compiler-cli-5.1.3.tgz (Root Library) - chokidar-1.7.0.tgz - :x: **glob-parent-2.0.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-28469 (High) detected in glob-parent-2.0.0.tgz - ## CVE-2020-28469 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-2.0.0.tgz</b></p></summary> <p>Strips glob magic from a string to provide the parent path</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-2.0.0.tgz</a></p> <p>Path to dependency file: /src/package.json</p> <p>Path to vulnerable library: /src/node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - compiler-cli-5.1.3.tgz (Root Library) - chokidar-1.7.0.tgz - :x: **glob-parent-2.0.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package glob-parent before 5.1.2. The enclosure regex used to check for strings ending in enclosure containing path separator. <p>Publish Date: 2021-06-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28469>CVE-2020-28469</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28469</a></p> <p>Release Date: 2021-06-03</p> <p>Fix Resolution: glob-parent - 5.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in glob parent tgz cve high severity vulnerability vulnerable library glob parent tgz strips glob magic from a string to provide the parent path library home page a href path to dependency file src package json path to vulnerable library src node modules glob parent package json dependency hierarchy compiler cli tgz root library chokidar tgz x glob parent tgz vulnerable library vulnerability details this affects the package glob parent before the enclosure regex used to check for strings ending in enclosure containing path separator publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with whitesource
0
4,134
19,603,761,585
IssuesEvent
2022-01-06 06:22:26
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Sam sync removes node_modules in lambda layer
type/bug blocked/more-info-needed area/sync maintainer/need-followup
### Description: I have a sam application that contains a nodejs14.x Lambda function that references a Lambda layer with a library installed in `node_modules`. The first time I run `sam sync --watch`, all works fine. When I stop the watcher and restart it, it deletes the `node_modules` folder in `.aws-sam/auto-dependency-layer/SharedLayer/nodejs/node_modules` (where `SharedLayer` is my Lambda Layer). Consequently my Lambda function is broken. My workaround is to modify the `manifest_hash` value of the Lambda layer in `.aws-sam/build.toml`, and make a dummy change in the code of the Lambda layer. This triggers the watcher, so it correctly builds and syncs the layer + dependent functions. ### Steps to reproduce: 1) Clone this repo: https://github.com/cshenrik/accelerate-bug. 1) Run `sam sync --watch --stack-name <STACK_NAME>`. 1) Wait until the initial sync is done. 1) Exit the watcher by hitting ctrl + c. 1) Run `sam sync --watch --stack-name <STACK_NAME>` again. Then `.aws-sam/auto-dependency-layer/SharedLayer/nodejs/node_modules` has been deleted, and the Lambda function is broken. ### Observed result: This is the output from running the second `sam sync` in step 5 (though with --debug): ```2021-11-24 20:34:31,079 | No Parameters detected in the template 2021-11-24 20:34:31,099 | 3 stacks found in the template 2021-11-24 20:34:31,099 | No Parameters detected in the template 2021-11-24 20:34:31,113 | 3 resources found in the stack 2021-11-24 20:34:31,113 | No Parameters detected in the template 2021-11-24 20:34:31,126 | No Parameters detected in the template 2021-11-24 20:34:31,140 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,140 | Found Serverless function with name='SayHelloFunction' and CodeUri='say-hello/' 2021-11-24 20:34:31,140 | --base-dir is not presented, adjusting uri say-hello/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,140 | No Parameters detected in the template 2021-11-24 20:34:31,153 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,173 | Queued infra sync. Wating for in progress code syncs to complete... 2021-11-24 20:34:31,174 | Starting infra sync. 2021-11-24 20:34:31,184 | No Parameters detected in the template 2021-11-24 20:34:31,201 | 3 stacks found in the template 2021-11-24 20:34:31,201 | No Parameters detected in the template 2021-11-24 20:34:31,215 | 3 resources found in the stack 2021-11-24 20:34:31,215 | No Parameters detected in the template 2021-11-24 20:34:31,228 | No Parameters detected in the template 2021-11-24 20:34:31,243 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,243 | Found Serverless function with name='SayHelloFunction' and CodeUri='say-hello/' 2021-11-24 20:34:31,243 | --base-dir is not presented, adjusting uri say-hello/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,243 | No Parameters detected in the template 2021-11-24 20:34:31,255 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,256 | Instantiating build definitions 2021-11-24 20:34:31,260 | Same function build definition found, adding function (Previous: BuildDefinition(nodejs14.x, /home/henrik/code/accelerate-bug/say-hello, Zip, , d4fbbd94-b684-4533-9386-944b156f0d9c, {}, {}, x86_64, []), Current: BuildDefinition(nodejs14.x, /home/henrik/code/accelerate-bug/say-hello, Zip, , d59cc2b5-5805-4e84-bb60-3d19648d581f, {}, {}, x86_64, []), Function: Function(name='SayHelloFunction', functionname='SayHelloFunction', runtime='nodejs14.x', memory=1024, timeout=3, handler='app.lambdaHandler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/home/henrik/code/accelerate-bug/say-hello', environment=None, rolearn=None, layers=[<samcli.lib.providers.provider.LayerVersion object at 0x7f744d989fd0>], events={'HttpGet': {'Type': 'HttpApi', 'Properties': {'Path': '/say-hello', 'Method': 'get', 'ApiId': 'ServerlessHttpApi'}}}, metadata=None, inlinecode=None, codesign_config_arn=None, architectures=None, stack_path='')) 2021-11-24 20:34:31,260 | Same Layer build definition found, adding layer (Previous: LayerBuildDefinition(SharedLayer, /home/henrik/code/accelerate-bug/shared, , 3d1b6ddd-2760-449c-9d26-346390ef5a4f, nodejs14.x, ['nodejs14.x'], x86_64, {}), Current: LayerBuildDefinition(SharedLayer, /home/henrik/code/accelerate-bug/shared, , 489aa9fd-dd20-4134-bede-7fa22b03dc46, nodejs14.x, ['nodejs14.x'], x86_64, {}), Layer: <samcli.lib.providers.provider.LayerVersion object at 0x7f744d9890d0>) 2021-11-24 20:34:31,262 | Async execution started 2021-11-24 20:34:31,262 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_function_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x7f744d88b910>>, <samcli.lib.build.build_graph.FunctionBuildDefinition object at 0x7f744d8861d0>) 2021-11-24 20:34:31,262 | Running incremental build for runtime nodejs14.x for build definition d4fbbd94-b684-4533-9386-944b156f0d9c 2021-11-24 20:34:31,262 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_layer_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x7f744d88b910>>, <samcli.lib.build.build_graph.LayerBuildDefinition object at 0x7f744d886550>) 2021-11-24 20:34:31,263 | Running incremental build for runtime nodejs14.x for build definition 3d1b6ddd-2760-449c-9d26-346390ef5a4f 2021-11-24 20:34:31,263 | Waiting for async results 2021-11-24 20:34:31,264 | Manifest is not changed for 3d1b6ddd-2760-449c-9d26-346390ef5a4f, running incremental build 2021-11-24 20:34:31,264 | Building layer 'SharedLayer' 2021-11-24 20:34:31,264 | Loading workflow module 'aws_lambda_builders.workflows' 2021-11-24 20:34:31,264 | Manifest is not changed for d4fbbd94-b684-4533-9386-944b156f0d9c, running incremental build 2021-11-24 20:34:31,265 | Building codeuri: /home/henrik/code/accelerate-bug/say-hello runtime: nodejs14.x metadata: {} architecture: x86_64 functions: ['SayHelloFunction'] 2021-11-24 20:34:31,265 | Building to following folder /home/henrik/code/accelerate-bug/.aws-sam/auto-dependency-layer/SayHelloFunction 2021-11-24 20:34:31,265 | Loading workflow module 'aws_lambda_builders.workflows' 2021-11-24 20:34:31,267 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)' 2021-11-24 20:34:31,268 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-11-24 20:34:31,269 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)' 2021-11-24 20:34:31,269 | Registering workflow 'GoDepBuilder' with capability 'Capability(language='go', dependency_manager='dep', application_framework=None)' 2021-11-24 20:34:31,271 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)' 2021-11-24 20:34:31,273 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)' 2021-11-24 20:34:31,274 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)' 2021-11-24 20:34:31,277 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)' 2021-11-24 20:34:31,279 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)' 2021-11-24 20:34:31,280 | Found workflow 'NodejsNpmBuilder' to support capabilities 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-11-24 20:34:31,280 | Found workflow 'NodejsNpmBuilder' to support capabilities 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-11-24 20:34:31,280 | download_dependencies is False and dependencies_dir is None. Copying the source files into the artifacts directory. 2021-11-24 20:34:31,280 | download_dependencies is False and dependencies_dir is None. Copying the source files into the artifacts directory. 2021-11-24 20:34:31,281 | Running workflow 'NodejsNpmBuilder' 2021-11-24 20:34:31,282 | Running NodejsNpmBuilder:NpmPack 2021-11-24 20:34:31,282 | NODEJS packaging file:/home/henrik/code/accelerate-bug/shared to /tmp/tmpipo4gchb 2021-11-24 20:34:31,282 | executing NPM: ['npm', 'pack', '-q', 'file:/home/henrik/code/accelerate-bug/shared'] 2021-11-24 20:34:31,282 | Running workflow 'NodejsNpmBuilder' 2021-11-24 20:34:31,285 | Running NodejsNpmBuilder:NpmPack 2021-11-24 20:34:31,285 | NODEJS packaging file:/home/henrik/code/accelerate-bug/say-hello to /tmp/tmp_7gtln35 2021-11-24 20:34:31,286 | executing NPM: ['npm', 'pack', '-q', 'file:/home/henrik/code/accelerate-bug/say-hello'] 2021-11-24 20:34:31,555 | NODEJS packed to say-hello-1.0.0.tgz 2021-11-24 20:34:31,555 | NODEJS extracting to /tmp/tmp_7gtln35/unpacked 2021-11-24 20:34:31,556 | NODEJS packed to shared-1.0.0.tgz 2021-11-24 20:34:31,556 | NODEJS extracting to /tmp/tmpipo4gchb/unpacked 2021-11-24 20:34:31,557 | NodejsNpmBuilder:NpmPack succeeded 2021-11-24 20:34:31,557 | Running NodejsNpmBuilder:CopyNpmrc 2021-11-24 20:34:31,557 | NodejsNpmBuilder:CopyNpmrc succeeded 2021-11-24 20:34:31,558 | Running NodejsNpmBuilder:CopySource 2021-11-24 20:34:31,558 | NodejsNpmBuilder:NpmPack succeeded 2021-11-24 20:34:31,558 | Running NodejsNpmBuilder:CopyNpmrc 2021-11-24 20:34:31,558 | NodejsNpmBuilder:CopyNpmrc succeeded 2021-11-24 20:34:31,558 | Running NodejsNpmBuilder:CopySource 2021-11-24 20:34:31,559 | NodejsNpmBuilder:CopySource succeeded 2021-11-24 20:34:31,559 | Running NodejsNpmBuilder:CleanUpNpmrc 2021-11-24 20:34:31,559 | NodejsNpmBuilder:CleanUpNpmrc succeeded 2021-11-24 20:34:31,560 | NodejsNpmBuilder:CopySource succeeded 2021-11-24 20:34:31,560 | Running NodejsNpmBuilder:CleanUpNpmrc 2021-11-24 20:34:31,560 | NodejsNpmBuilder:CleanUpNpmrc succeeded 2021-11-24 20:34:31,562 | Async execution completed 2021-11-24 20:34:31,565 | Auto creating dependency layer for each function resource into a nested stack 2021-11-24 20:34:31,565 | No Parameters detected in the template 2021-11-24 20:34:31,580 | 3 resources found in the stack accelerate-bug 2021-11-24 20:34:31,580 | No Parameters detected in the template 2021-11-24 20:34:31,595 | No Parameters detected in the template 2021-11-24 20:34:31,610 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/SharedLayer relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,610 | Found Serverless function with name='SayHelloFunction' and CodeUri='.aws-sam/auto-dependency-layer/SayHelloFunction' 2021-11-24 20:34:31,610 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/SayHelloFunction relative to /home/henrik/code/accelerate-bug/template.yaml Build Succeeded Built Artifacts : .aws-sam/auto-dependency-layer Built Template : .aws-sam/auto-dependency-layer/template.yaml ``` The output that follows the above shows the cloudformation update. ### Expected result: I would expect it to not delete `node_modules` such that the Lambda function still works. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Fedora 33 2. `sam --version`: `SAM CLI, version 1.36.0` 3. AWS region: eu-west-1
True
Sam sync removes node_modules in lambda layer - ### Description: I have a sam application that contains a nodejs14.x Lambda function that references a Lambda layer with a library installed in `node_modules`. The first time I run `sam sync --watch`, all works fine. When I stop the watcher and restart it, it deletes the `node_modules` folder in `.aws-sam/auto-dependency-layer/SharedLayer/nodejs/node_modules` (where `SharedLayer` is my Lambda Layer). Consequently my Lambda function is broken. My workaround is to modify the `manifest_hash` value of the Lambda layer in `.aws-sam/build.toml`, and make a dummy change in the code of the Lambda layer. This triggers the watcher, so it correctly builds and syncs the layer + dependent functions. ### Steps to reproduce: 1) Clone this repo: https://github.com/cshenrik/accelerate-bug. 1) Run `sam sync --watch --stack-name <STACK_NAME>`. 1) Wait until the initial sync is done. 1) Exit the watcher by hitting ctrl + c. 1) Run `sam sync --watch --stack-name <STACK_NAME>` again. Then `.aws-sam/auto-dependency-layer/SharedLayer/nodejs/node_modules` has been deleted, and the Lambda function is broken. ### Observed result: This is the output from running the second `sam sync` in step 5 (though with --debug): ```2021-11-24 20:34:31,079 | No Parameters detected in the template 2021-11-24 20:34:31,099 | 3 stacks found in the template 2021-11-24 20:34:31,099 | No Parameters detected in the template 2021-11-24 20:34:31,113 | 3 resources found in the stack 2021-11-24 20:34:31,113 | No Parameters detected in the template 2021-11-24 20:34:31,126 | No Parameters detected in the template 2021-11-24 20:34:31,140 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,140 | Found Serverless function with name='SayHelloFunction' and CodeUri='say-hello/' 2021-11-24 20:34:31,140 | --base-dir is not presented, adjusting uri say-hello/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,140 | No Parameters detected in the template 2021-11-24 20:34:31,153 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,173 | Queued infra sync. Wating for in progress code syncs to complete... 2021-11-24 20:34:31,174 | Starting infra sync. 2021-11-24 20:34:31,184 | No Parameters detected in the template 2021-11-24 20:34:31,201 | 3 stacks found in the template 2021-11-24 20:34:31,201 | No Parameters detected in the template 2021-11-24 20:34:31,215 | 3 resources found in the stack 2021-11-24 20:34:31,215 | No Parameters detected in the template 2021-11-24 20:34:31,228 | No Parameters detected in the template 2021-11-24 20:34:31,243 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,243 | Found Serverless function with name='SayHelloFunction' and CodeUri='say-hello/' 2021-11-24 20:34:31,243 | --base-dir is not presented, adjusting uri say-hello/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,243 | No Parameters detected in the template 2021-11-24 20:34:31,255 | --base-dir is not presented, adjusting uri shared/ relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,256 | Instantiating build definitions 2021-11-24 20:34:31,260 | Same function build definition found, adding function (Previous: BuildDefinition(nodejs14.x, /home/henrik/code/accelerate-bug/say-hello, Zip, , d4fbbd94-b684-4533-9386-944b156f0d9c, {}, {}, x86_64, []), Current: BuildDefinition(nodejs14.x, /home/henrik/code/accelerate-bug/say-hello, Zip, , d59cc2b5-5805-4e84-bb60-3d19648d581f, {}, {}, x86_64, []), Function: Function(name='SayHelloFunction', functionname='SayHelloFunction', runtime='nodejs14.x', memory=1024, timeout=3, handler='app.lambdaHandler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/home/henrik/code/accelerate-bug/say-hello', environment=None, rolearn=None, layers=[<samcli.lib.providers.provider.LayerVersion object at 0x7f744d989fd0>], events={'HttpGet': {'Type': 'HttpApi', 'Properties': {'Path': '/say-hello', 'Method': 'get', 'ApiId': 'ServerlessHttpApi'}}}, metadata=None, inlinecode=None, codesign_config_arn=None, architectures=None, stack_path='')) 2021-11-24 20:34:31,260 | Same Layer build definition found, adding layer (Previous: LayerBuildDefinition(SharedLayer, /home/henrik/code/accelerate-bug/shared, , 3d1b6ddd-2760-449c-9d26-346390ef5a4f, nodejs14.x, ['nodejs14.x'], x86_64, {}), Current: LayerBuildDefinition(SharedLayer, /home/henrik/code/accelerate-bug/shared, , 489aa9fd-dd20-4134-bede-7fa22b03dc46, nodejs14.x, ['nodejs14.x'], x86_64, {}), Layer: <samcli.lib.providers.provider.LayerVersion object at 0x7f744d9890d0>) 2021-11-24 20:34:31,262 | Async execution started 2021-11-24 20:34:31,262 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_function_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x7f744d88b910>>, <samcli.lib.build.build_graph.FunctionBuildDefinition object at 0x7f744d8861d0>) 2021-11-24 20:34:31,262 | Running incremental build for runtime nodejs14.x for build definition d4fbbd94-b684-4533-9386-944b156f0d9c 2021-11-24 20:34:31,262 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_layer_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x7f744d88b910>>, <samcli.lib.build.build_graph.LayerBuildDefinition object at 0x7f744d886550>) 2021-11-24 20:34:31,263 | Running incremental build for runtime nodejs14.x for build definition 3d1b6ddd-2760-449c-9d26-346390ef5a4f 2021-11-24 20:34:31,263 | Waiting for async results 2021-11-24 20:34:31,264 | Manifest is not changed for 3d1b6ddd-2760-449c-9d26-346390ef5a4f, running incremental build 2021-11-24 20:34:31,264 | Building layer 'SharedLayer' 2021-11-24 20:34:31,264 | Loading workflow module 'aws_lambda_builders.workflows' 2021-11-24 20:34:31,264 | Manifest is not changed for d4fbbd94-b684-4533-9386-944b156f0d9c, running incremental build 2021-11-24 20:34:31,265 | Building codeuri: /home/henrik/code/accelerate-bug/say-hello runtime: nodejs14.x metadata: {} architecture: x86_64 functions: ['SayHelloFunction'] 2021-11-24 20:34:31,265 | Building to following folder /home/henrik/code/accelerate-bug/.aws-sam/auto-dependency-layer/SayHelloFunction 2021-11-24 20:34:31,265 | Loading workflow module 'aws_lambda_builders.workflows' 2021-11-24 20:34:31,267 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)' 2021-11-24 20:34:31,268 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-11-24 20:34:31,269 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)' 2021-11-24 20:34:31,269 | Registering workflow 'GoDepBuilder' with capability 'Capability(language='go', dependency_manager='dep', application_framework=None)' 2021-11-24 20:34:31,271 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)' 2021-11-24 20:34:31,273 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)' 2021-11-24 20:34:31,274 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)' 2021-11-24 20:34:31,277 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)' 2021-11-24 20:34:31,279 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)' 2021-11-24 20:34:31,280 | Found workflow 'NodejsNpmBuilder' to support capabilities 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-11-24 20:34:31,280 | Found workflow 'NodejsNpmBuilder' to support capabilities 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-11-24 20:34:31,280 | download_dependencies is False and dependencies_dir is None. Copying the source files into the artifacts directory. 2021-11-24 20:34:31,280 | download_dependencies is False and dependencies_dir is None. Copying the source files into the artifacts directory. 2021-11-24 20:34:31,281 | Running workflow 'NodejsNpmBuilder' 2021-11-24 20:34:31,282 | Running NodejsNpmBuilder:NpmPack 2021-11-24 20:34:31,282 | NODEJS packaging file:/home/henrik/code/accelerate-bug/shared to /tmp/tmpipo4gchb 2021-11-24 20:34:31,282 | executing NPM: ['npm', 'pack', '-q', 'file:/home/henrik/code/accelerate-bug/shared'] 2021-11-24 20:34:31,282 | Running workflow 'NodejsNpmBuilder' 2021-11-24 20:34:31,285 | Running NodejsNpmBuilder:NpmPack 2021-11-24 20:34:31,285 | NODEJS packaging file:/home/henrik/code/accelerate-bug/say-hello to /tmp/tmp_7gtln35 2021-11-24 20:34:31,286 | executing NPM: ['npm', 'pack', '-q', 'file:/home/henrik/code/accelerate-bug/say-hello'] 2021-11-24 20:34:31,555 | NODEJS packed to say-hello-1.0.0.tgz 2021-11-24 20:34:31,555 | NODEJS extracting to /tmp/tmp_7gtln35/unpacked 2021-11-24 20:34:31,556 | NODEJS packed to shared-1.0.0.tgz 2021-11-24 20:34:31,556 | NODEJS extracting to /tmp/tmpipo4gchb/unpacked 2021-11-24 20:34:31,557 | NodejsNpmBuilder:NpmPack succeeded 2021-11-24 20:34:31,557 | Running NodejsNpmBuilder:CopyNpmrc 2021-11-24 20:34:31,557 | NodejsNpmBuilder:CopyNpmrc succeeded 2021-11-24 20:34:31,558 | Running NodejsNpmBuilder:CopySource 2021-11-24 20:34:31,558 | NodejsNpmBuilder:NpmPack succeeded 2021-11-24 20:34:31,558 | Running NodejsNpmBuilder:CopyNpmrc 2021-11-24 20:34:31,558 | NodejsNpmBuilder:CopyNpmrc succeeded 2021-11-24 20:34:31,558 | Running NodejsNpmBuilder:CopySource 2021-11-24 20:34:31,559 | NodejsNpmBuilder:CopySource succeeded 2021-11-24 20:34:31,559 | Running NodejsNpmBuilder:CleanUpNpmrc 2021-11-24 20:34:31,559 | NodejsNpmBuilder:CleanUpNpmrc succeeded 2021-11-24 20:34:31,560 | NodejsNpmBuilder:CopySource succeeded 2021-11-24 20:34:31,560 | Running NodejsNpmBuilder:CleanUpNpmrc 2021-11-24 20:34:31,560 | NodejsNpmBuilder:CleanUpNpmrc succeeded 2021-11-24 20:34:31,562 | Async execution completed 2021-11-24 20:34:31,565 | Auto creating dependency layer for each function resource into a nested stack 2021-11-24 20:34:31,565 | No Parameters detected in the template 2021-11-24 20:34:31,580 | 3 resources found in the stack accelerate-bug 2021-11-24 20:34:31,580 | No Parameters detected in the template 2021-11-24 20:34:31,595 | No Parameters detected in the template 2021-11-24 20:34:31,610 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/SharedLayer relative to /home/henrik/code/accelerate-bug/template.yaml 2021-11-24 20:34:31,610 | Found Serverless function with name='SayHelloFunction' and CodeUri='.aws-sam/auto-dependency-layer/SayHelloFunction' 2021-11-24 20:34:31,610 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/SayHelloFunction relative to /home/henrik/code/accelerate-bug/template.yaml Build Succeeded Built Artifacts : .aws-sam/auto-dependency-layer Built Template : .aws-sam/auto-dependency-layer/template.yaml ``` The output that follows the above shows the cloudformation update. ### Expected result: I would expect it to not delete `node_modules` such that the Lambda function still works. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Fedora 33 2. `sam --version`: `SAM CLI, version 1.36.0` 3. AWS region: eu-west-1
main
sam sync removes node modules in lambda layer description i have a sam application that contains a x lambda function that references a lambda layer with a library installed in node modules the first time i run sam sync watch all works fine when i stop the watcher and restart it it deletes the node modules folder in aws sam auto dependency layer sharedlayer nodejs node modules where sharedlayer is my lambda layer consequently my lambda function is broken my workaround is to modify the manifest hash value of the lambda layer in aws sam build toml and make a dummy change in the code of the lambda layer this triggers the watcher so it correctly builds and syncs the layer dependent functions steps to reproduce clone this repo run sam sync watch stack name wait until the initial sync is done exit the watcher by hitting ctrl c run sam sync watch stack name again then aws sam auto dependency layer sharedlayer nodejs node modules has been deleted and the lambda function is broken observed result this is the output from running the second sam sync in step though with debug no parameters detected in the template stacks found in the template no parameters detected in the template resources found in the stack no parameters detected in the template no parameters detected in the template base dir is not presented adjusting uri shared relative to home henrik code accelerate bug template yaml found serverless function with name sayhellofunction and codeuri say hello base dir is not presented adjusting uri say hello relative to home henrik code accelerate bug template yaml no parameters detected in the template base dir is not presented adjusting uri shared relative to home henrik code accelerate bug template yaml queued infra sync wating for in progress code syncs to complete starting infra sync no parameters detected in the template stacks found in the template no parameters detected in the template resources found in the stack no parameters detected in the template no parameters detected in the template base dir is not presented adjusting uri shared relative to home henrik code accelerate bug template yaml found serverless function with name sayhellofunction and codeuri say hello base dir is not presented adjusting uri say hello relative to home henrik code accelerate bug template yaml no parameters detected in the template base dir is not presented adjusting uri shared relative to home henrik code accelerate bug template yaml instantiating build definitions same function build definition found adding function previous builddefinition x home henrik code accelerate bug say hello zip current builddefinition x home henrik code accelerate bug say hello zip function function name sayhellofunction functionname sayhellofunction runtime x memory timeout handler app lambdahandler imageuri none packagetype zip imageconfig none codeuri home henrik code accelerate bug say hello environment none rolearn none layers events httpget type httpapi properties path say hello method get apiid serverlesshttpapi metadata none inlinecode none codesign config arn none architectures none stack path same layer build definition found adding layer previous layerbuilddefinition sharedlayer home henrik code accelerate bug shared x current layerbuilddefinition sharedlayer home henrik code accelerate bug shared bede x layer async execution started invoking function functools partial running incremental build for runtime x for build definition invoking function functools partial running incremental build for runtime x for build definition waiting for async results manifest is not changed for running incremental build building layer sharedlayer loading workflow module aws lambda builders workflows manifest is not changed for running incremental build building codeuri home henrik code accelerate bug say hello runtime x metadata architecture functions building to following folder home henrik code accelerate bug aws sam auto dependency layer sayhellofunction loading workflow module aws lambda builders workflows registering workflow pythonpipbuilder with capability capability language python dependency manager pip application framework none registering workflow nodejsnpmbuilder with capability capability language nodejs dependency manager npm application framework none registering workflow rubybundlerbuilder with capability capability language ruby dependency manager bundler application framework none registering workflow godepbuilder with capability capability language go dependency manager dep application framework none registering workflow gomodulesbuilder with capability capability language go dependency manager modules application framework none registering workflow javagradleworkflow with capability capability language java dependency manager gradle application framework none registering workflow javamavenworkflow with capability capability language java dependency manager maven application framework none registering workflow dotnetclipackagebuilder with capability capability language dotnet dependency manager cli package application framework none registering workflow custommakebuilder with capability capability language provided dependency manager none application framework none found workflow nodejsnpmbuilder to support capabilities capability language nodejs dependency manager npm application framework none found workflow nodejsnpmbuilder to support capabilities capability language nodejs dependency manager npm application framework none download dependencies is false and dependencies dir is none copying the source files into the artifacts directory download dependencies is false and dependencies dir is none copying the source files into the artifacts directory running workflow nodejsnpmbuilder running nodejsnpmbuilder npmpack nodejs packaging file home henrik code accelerate bug shared to tmp executing npm running workflow nodejsnpmbuilder running nodejsnpmbuilder npmpack nodejs packaging file home henrik code accelerate bug say hello to tmp tmp executing npm nodejs packed to say hello tgz nodejs extracting to tmp tmp unpacked nodejs packed to shared tgz nodejs extracting to tmp unpacked nodejsnpmbuilder npmpack succeeded running nodejsnpmbuilder copynpmrc nodejsnpmbuilder copynpmrc succeeded running nodejsnpmbuilder copysource nodejsnpmbuilder npmpack succeeded running nodejsnpmbuilder copynpmrc nodejsnpmbuilder copynpmrc succeeded running nodejsnpmbuilder copysource nodejsnpmbuilder copysource succeeded running nodejsnpmbuilder cleanupnpmrc nodejsnpmbuilder cleanupnpmrc succeeded nodejsnpmbuilder copysource succeeded running nodejsnpmbuilder cleanupnpmrc nodejsnpmbuilder cleanupnpmrc succeeded async execution completed auto creating dependency layer for each function resource into a nested stack no parameters detected in the template resources found in the stack accelerate bug no parameters detected in the template no parameters detected in the template base dir is not presented adjusting uri aws sam auto dependency layer sharedlayer relative to home henrik code accelerate bug template yaml found serverless function with name sayhellofunction and codeuri aws sam auto dependency layer sayhellofunction base dir is not presented adjusting uri aws sam auto dependency layer sayhellofunction relative to home henrik code accelerate bug template yaml build succeeded built artifacts aws sam auto dependency layer built template aws sam auto dependency layer template yaml the output that follows the above shows the cloudformation update expected result i would expect it to not delete node modules such that the lambda function still works additional environment details ex windows mac amazon linux etc os fedora sam version sam cli version aws region eu west
1
167,301
26,483,519,369
IssuesEvent
2023-01-17 16:16:19
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
[Components and pattern standards] Design components and patterns in use are outdated. (04.08.2)
content design 508/Accessibility ia collab-cycle-feedback Staging CCIssue04.08 CC-Dashboard my-education-benefits
### General Information #### VFS team name DGIB - My Education Benefits #### VFS product name Enrollment verification #### VFS feature name #### Point of Contact/Reviewers Brian DeConinck - @briandeconinck - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue Design components and patterns in use are outdated. ### Issue Details **Accessibility note:** Since this is coded as a link and navigates to a new page, it's important that it be visually styled as a link. This kind of material honesty benefits voice command users who choose the command they speak to their computer based on what they're seeing on the page. ### Link, screenshot or steps to recreate ### VA.gov Experience Standard [Category Number 04, Issue Number 08](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 3.2.4 AA ### Platform Recommendation --- ### VFS Guidance - Close the ticket when the issue has been resolved or validated by your Product Owner - If your team has additional questions or needs Platform help validating the issue, please comment on the ticket - Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue. - If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.
1.0
[Components and pattern standards] Design components and patterns in use are outdated. (04.08.2) - ### General Information #### VFS team name DGIB - My Education Benefits #### VFS product name Enrollment verification #### VFS feature name #### Point of Contact/Reviewers Brian DeConinck - @briandeconinck - Accessibility *For more information on how to interpret this ticket, please refer to the [Anatomy of a Staging Review issue ticket](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/Anatomy-of-a-Staging-Review-Issue-ticket.2060320997.html) guidance on Platform Website. --- ### Platform Issue Design components and patterns in use are outdated. ### Issue Details **Accessibility note:** Since this is coded as a link and navigates to a new page, it's important that it be visually styled as a link. This kind of material honesty benefits voice command users who choose the command they speak to their computer based on what they're seeing on the page. ### Link, screenshot or steps to recreate ### VA.gov Experience Standard [Category Number 04, Issue Number 08](https://depo-platform-documentation.scrollhelp.site/collaboration-cycle/VA.gov-experience-standards.1683980311.html) ### Other References WCAG SC 3.2.4 AA ### Platform Recommendation --- ### VFS Guidance - Close the ticket when the issue has been resolved or validated by your Product Owner - If your team has additional questions or needs Platform help validating the issue, please comment on the ticket - Some feedback provided may be out of scope for your iteration of the product, however, Platform's OCTO leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue. - If you do not believe that this Staging Review issue ticket is the responsibility of your team, comment below providing an explanation and who you believe is responsible. Please tag the Point of Contact/Reviewers. Governance team will research and will follow up.
non_main
design components and patterns in use are outdated general information vfs team name dgib my education benefits vfs product name enrollment verification vfs feature name point of contact reviewers brian deconinck briandeconinck accessibility for more information on how to interpret this ticket please refer to the guidance on platform website platform issue design components and patterns in use are outdated issue details accessibility note since this is coded as a link and navigates to a new page it s important that it be visually styled as a link this kind of material honesty benefits voice command users who choose the command they speak to their computer based on what they re seeing on the page link screenshot or steps to recreate va gov experience standard other references wcag sc aa platform recommendation vfs guidance close the ticket when the issue has been resolved or validated by your product owner if your team has additional questions or needs platform help validating the issue please comment on the ticket some feedback provided may be out of scope for your iteration of the product however platform s octo leadership has stated that all identified issues need to be documented and it is still your responsibility to resolve the issue if you do not believe that this staging review issue ticket is the responsibility of your team comment below providing an explanation and who you believe is responsible please tag the point of contact reviewers governance team will research and will follow up
0
2,496
8,655,457,992
IssuesEvent
2018-11-27 16:00:18
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
[Issue] Vita cannot reconnect after recovering from Standby
unmaintained
My Vita is connected to my PC (Win8.1 Pro) via WiFi Everything works - Until my Vita enters Standby... - After recovering from Standby, the Vita cannot connect to QCMA - The error on the Vita says "Could not connect to network" In order to fix this, I need to completely shutdown QCMA and restart it before my Vita can connect - is this how it is supposed to work?
True
[Issue] Vita cannot reconnect after recovering from Standby - My Vita is connected to my PC (Win8.1 Pro) via WiFi Everything works - Until my Vita enters Standby... - After recovering from Standby, the Vita cannot connect to QCMA - The error on the Vita says "Could not connect to network" In order to fix this, I need to completely shutdown QCMA and restart it before my Vita can connect - is this how it is supposed to work?
main
vita cannot reconnect after recovering from standby my vita is connected to my pc pro via wifi everything works until my vita enters standby after recovering from standby the vita cannot connect to qcma the error on the vita says could not connect to network in order to fix this i need to completely shutdown qcma and restart it before my vita can connect is this how it is supposed to work
1
346
3,228,434,834
IssuesEvent
2015-10-12 02:03:10
spyder-ide/spyder
https://api.github.com/repos/spyder-ide/spyder
closed
fallback_plugin tests are failing
bug Editor Maintainability
As I was working on PR #2752, i discovered that those tests are failing. I skipped them for now in `modules_test.sh`, but they need to be fixed. @blink1073, could you take care of that (given that you wrote them)? They should be easy to fix :-)
True
fallback_plugin tests are failing - As I was working on PR #2752, i discovered that those tests are failing. I skipped them for now in `modules_test.sh`, but they need to be fixed. @blink1073, could you take care of that (given that you wrote them)? They should be easy to fix :-)
main
fallback plugin tests are failing as i was working on pr i discovered that those tests are failing i skipped them for now in modules test sh but they need to be fixed could you take care of that given that you wrote them they should be easy to fix
1
368,895
10,885,977,337
IssuesEvent
2019-11-18 11:31:59
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
myanimelist.net - design is broken
browser-firefox-mobile engine-gecko priority-important
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://myanimelist.net/forum/?topicid=1810168&show=50#post59 **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 6.0 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: the right edge of website won't show in desktop mode **Steps to Reproduce**: In this website and also in some other phpbb forums, when we request desktop mode of the page the page won't show completely and right side or left side of the website won't load. [![Screenshot Description](https://webcompat.com/uploads/2019/11/74244d22-ea99-4bb5-8c29-657f889ef9b8-thumb.jpeg)](https://webcompat.com/uploads/2019/11/74244d22-ea99-4bb5-8c29-657f889ef9b8.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190920125400</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> <p>Console Messages:</p> <pre> ['[JavaScript Warning: "Loading failed for the <script> with source https://tags-cdn.deployads.com/a/myanimelist.net.js." {file: "https://myanimelist.net/forum/?topicid=1810168&show=50#post59" line: 33}]', '[JavaScript Warning: "Loading failed for the <script> with source https://apis.google.com/js/platform.js." {file: "https://myanimelist.net/forum/?topicid=1810168&show=50#post59" line: 108}]', '[JavaScript Error: "TypeError: e is null" {file: "https://cdn.myanimelist.net/static/assets/js/pc/all-81831e9acf.js" line: 46}]\n@https://cdn.myanimelist.net/static/assets/js/pc/all-81831e9acf.js:46:414550\nc@https://cdn.myanimelist.net/static/assets/js/pc/header-4681929590.js:24:25054\ni/</l<@https://cdn.myanimelist.net/static/assets/js/pc/header-4681929590.js:24:25357\n', '[JavaScript Warning: "This site appears to use a scroll-linked positioning effect. This may not work well with asynchronous panning; see https://developer.mozilla.org/docs/Mozilla/Performance/ScrollLinkedEffects for further details and to join the discussion on related tools and features!" {file: "https://myanimelist.net/forum/?topicid=1810168&show=50#post59" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
myanimelist.net - design is broken - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 6.0; Mobile; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://myanimelist.net/forum/?topicid=1810168&show=50#post59 **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 6.0 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: the right edge of website won't show in desktop mode **Steps to Reproduce**: In this website and also in some other phpbb forums, when we request desktop mode of the page the page won't show completely and right side or left side of the website won't load. [![Screenshot Description](https://webcompat.com/uploads/2019/11/74244d22-ea99-4bb5-8c29-657f889ef9b8-thumb.jpeg)](https://webcompat.com/uploads/2019/11/74244d22-ea99-4bb5-8c29-657f889ef9b8.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20190920125400</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> <p>Console Messages:</p> <pre> ['[JavaScript Warning: "Loading failed for the <script> with source https://tags-cdn.deployads.com/a/myanimelist.net.js." {file: "https://myanimelist.net/forum/?topicid=1810168&show=50#post59" line: 33}]', '[JavaScript Warning: "Loading failed for the <script> with source https://apis.google.com/js/platform.js." {file: "https://myanimelist.net/forum/?topicid=1810168&show=50#post59" line: 108}]', '[JavaScript Error: "TypeError: e is null" {file: "https://cdn.myanimelist.net/static/assets/js/pc/all-81831e9acf.js" line: 46}]\n@https://cdn.myanimelist.net/static/assets/js/pc/all-81831e9acf.js:46:414550\nc@https://cdn.myanimelist.net/static/assets/js/pc/header-4681929590.js:24:25054\ni/</l<@https://cdn.myanimelist.net/static/assets/js/pc/header-4681929590.js:24:25357\n', '[JavaScript Warning: "This site appears to use a scroll-linked positioning effect. This may not work well with asynchronous panning; see https://developer.mozilla.org/docs/Mozilla/Performance/ScrollLinkedEffects for further details and to join the discussion on related tools and features!" {file: "https://myanimelist.net/forum/?topicid=1810168&show=50#post59" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
myanimelist net design is broken url browser version firefox mobile operating system android tested another browser yes problem type design is broken description the right edge of website won t show in desktop mode steps to reproduce in this website and also in some other phpbb forums when we request desktop mode of the page the page won t show completely and right side or left side of the website won t load browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false console messages n from with ❤️
0
3,499
13,654,540,022
IssuesEvent
2020-09-27 17:54:35
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
opened
Migrate to travis-ci.com
CI/CD maintainability to be reviewed
As [announced in 2018](https://blog.travis-ci.com/2018-05-02-open-source-projects-on-travis-ci-com-with-github-apps) travis-ci.com is to be used for both open source and commercial projects. In December 2020, travis-ci.org will be [going away](https://mailchi.mp/3d439eeb1098/travis-ciorg-is-moving-to-travis-cicom), so we need to move before then. Migration instructions are here: https://docs.travis-ci.com/user/migrate/open-source-repository-migration The travis-ci.org builds will become read-only after the migration, so they can't be restarted, etc, but activity seems pretty low right now, so I don't know if we need to do any upfront coordination of the migration. Opinions?
True
Migrate to travis-ci.com - As [announced in 2018](https://blog.travis-ci.com/2018-05-02-open-source-projects-on-travis-ci-com-with-github-apps) travis-ci.com is to be used for both open source and commercial projects. In December 2020, travis-ci.org will be [going away](https://mailchi.mp/3d439eeb1098/travis-ciorg-is-moving-to-travis-cicom), so we need to move before then. Migration instructions are here: https://docs.travis-ci.com/user/migrate/open-source-repository-migration The travis-ci.org builds will become read-only after the migration, so they can't be restarted, etc, but activity seems pretty low right now, so I don't know if we need to do any upfront coordination of the migration. Opinions?
main
migrate to travis ci com as travis ci com is to be used for both open source and commercial projects in december travis ci org will be so we need to move before then migration instructions are here the travis ci org builds will become read only after the migration so they can t be restarted etc but activity seems pretty low right now so i don t know if we need to do any upfront coordination of the migration opinions
1
120,867
17,644,324,705
IssuesEvent
2021-08-20 02:12:47
Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting
https://api.github.com/repos/Baneeishaque/Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting
opened
CVE-2021-37643 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl
security vulnerability
## CVE-2021-37643 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p>Path to vulnerable library: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. If a user does not provide a valid padding value to `tf.raw_ops.MatrixDiagPartOp`, then the code triggers a null pointer dereference (if input is empty) or produces invalid behavior, ignoring all values after the first. The [implementation](https://github.com/tensorflow/tensorflow/blob/8d72537c6abf5a44103b57b9c2e22c14f5f49698/tensorflow/core/kernels/linalg/matrix_diag_op.cc#L89) reads the first value from a tensor buffer without first checking that the tensor has values to read from. We have patched the issue in GitHub commit 482da92095c4d48f8784b1f00dda4f81c28d2988. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range. <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37643>CVE-2021-37643</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-fcwc-p4fc-c5cc">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-fcwc-p4fc-c5cc</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37643 (High) detected in tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl - ## CVE-2021-37643 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/92/2b/e3af15221da9ff323521565fa3324b0d7c7c5b1d7a8ca66984c8d59cb0ce/tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p>Path to vulnerable library: Raindrop-Removal-With-Light-Field-Image-Using-Image-Inpainting/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp37-cp37m-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. If a user does not provide a valid padding value to `tf.raw_ops.MatrixDiagPartOp`, then the code triggers a null pointer dereference (if input is empty) or produces invalid behavior, ignoring all values after the first. The [implementation](https://github.com/tensorflow/tensorflow/blob/8d72537c6abf5a44103b57b9c2e22c14f5f49698/tensorflow/core/kernels/linalg/matrix_diag_op.cc#L89) reads the first value from a tensor buffer without first checking that the tensor has values to read from. We have patched the issue in GitHub commit 482da92095c4d48f8784b1f00dda4f81c28d2988. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range. <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37643>CVE-2021-37643</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-fcwc-p4fc-c5cc">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-fcwc-p4fc-c5cc</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file raindrop removal with light field image using image inpainting requirements txt path to vulnerable library raindrop removal with light field image using image inpainting requirements txt dependency hierarchy x tensorflow whl vulnerable library found in base branch main vulnerability details tensorflow is an end to end open source platform for machine learning if a user does not provide a valid padding value to tf raw ops matrixdiagpartop then the code triggers a null pointer dereference if input is empty or produces invalid behavior ignoring all values after the first the reads the first value from a tensor buffer without first checking that the tensor has values to read from we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
1,881
6,577,510,856
IssuesEvent
2017-09-12 01:25:19
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
setting iam_role state to active or inactive fails with python UnboundLocal error
affects_2.0 aws bug_report cloud P2 waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME iam module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT OS X 10.11.3 ##### SUMMARY It seems to be impossible to deactivate or revoke AWS access keys via Ansible. ##### STEPS TO REPRODUCE - Attempt to revoke or make inactive an IAM Access Key via the following playbook: ``` --- - hosts: all tasks: - name: Create TSE users with approrpiate group and fetch IAM keys iam: iam_type: user name: "joe.user" state: present access_key_state: "inactive" access_key_ids: - "AKIABRACADABRA" groups: "DEVUSER" profile: dev register: newusers - debug: var=newusers ``` Run with: `ansible-playbook -i "localhost," -c local ./tsecreds.yaml -vvvv` ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269919.04-56226139783789/iam", line 2920, in <module> main() File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269919.04-56226139783789/iam", line 554, in main if any([n in key_state for n in ['active', 'inactive']]) and not key_ids: UnboundLocalError: local variable 'key_ids' referenced before assignment ``` This happens for key states of "Active,Inactive". Using "Remove" will provide this error: ``` fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"access_key_ids": ["AKIABRACADABRA"], "access_key_state": "remove", "aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "groups": ["TSE"], "iam_type": "user", "key_count": 1, "name": "joe.user", "new_name": null, "new_path": null, "password": null, "path": "/", "profile": "dev", "region": null, "security_token": null, "state": "present", "update_password": "always", "validate_certs": true}, "module_name": "iam"}, "msg": "BotoServerError: 400 Bad Request\n<ErrorResponse xmlns=\"https://iam.amazonaws.com/doc/2010-05-08/\">\n <Error>\n <Type>Sender</Type>\n <Code>ValidationError</Code>\n <Message>1 validation error detected: Value 'Remove' at 'status' failed to satisfy constraint: Member must satisfy enum value set: [Active, Inactive]</Message>\n </Error>\n <RequestId>079855bd-f5ce-11e5-88c6-23b4e6cc6158</RequestId>\n</ErrorResponse>\n"} ``` ##### EXPECTED RESULTS I expected 'remove' to remove the keys entirely without first having to run a 'inactive' operation on them (is that instead a feature request?) and I expected the 'inactive' and 'active' operations to work as expected instead of throwing a Python error. Additionally, when running 'create' against a user without any keys, I expected the keys to be available in the debug information after I registered the variable. ##### ACTUAL RESULTS See above inline in 'steps to reproduce' Full output below: ### access_key_state: inactive ``` C02KN05HFFT3:oneoff-tsecreds karl.katzke$ ansible-playbook -i "localhost," -c local ./tsecreds.yaml -vvv No config file found; using defaults 1 plays in ./tsecreds.yaml PLAY *************************************************************************** TASK [setup] ******************************************************************* ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248 `" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmpfBc49B TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248/setup localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248/setup; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248/" > /dev/null 2>&1' ok: [localhost] TASK [Create TSE users with approrpiate group and fetch IAM keys] ************** task path: /Users/karl.katzke/Work/ansible/oneoff-tsecreds/tsecreds.yaml:4 ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406 `" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmpnY3azX TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/" > /dev/null 2>&1' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam", line 2920, in <module> main() File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam", line 554, in main if any([n in key_state for n in ['active', 'inactive']]) and not key_ids: UnboundLocalError: local variable 'key_ids' referenced before assignment fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "iam"}, "parsed": false} NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @./tsecreds.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ``` ### access_key_state: remove C02KN05HFFT3:oneoff-tsecreds karl.katzke$ ansible-playbook -i "localhost," -c local ./tsecreds.yaml -vvv No config file found; using defaults 1 plays in ./tsecreds.yaml PLAY *************************************************************************** TASK [setup] ******************************************************************* ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "`echo $HOME/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273`" && echo "`echo $HOME/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273`" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmprZfcCw TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273/setup localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273/setup; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273/" > /dev/null 2>&1' ok: [localhost] TASK [Create TSE users with approrpiate group and fetch IAM keys] ************** task path: /Users/karl.katzke/Work/ansible/oneoff-tsecreds/tsecreds.yaml:4 ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "`echo $HOME/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934`" && echo "`echo $HOME/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934`" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmpHq1d7t TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934/iam localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934/iam; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934/" > /dev/null 2>&1' fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"access_key_ids": ["AKIAI3BJJ6AEJCLR5MSA"], "access_key_state": "remove", "aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "groups": ["TSE"], "iam_type": "user", "key_count": 1, "name": "brian.outlaw", "new_name": null, "new_path": null, "password": null, "path": "/", "profile": "dev", "region": null, "security_token": null, "state": "present", "update_password": "always", "validate_certs": true}, "module_name": "iam"}, "msg": "BotoServerError: 400 Bad Request\n<ErrorResponse xmlns=\"https://iam.amazonaws.com/doc/2010-05-08/\">\n <Error>\n <Type>Sender</Type>\n <Code>ValidationError</Code>\n <Message>1 validation error detected: Value 'Remove' at 'status' failed to satisfy constraint: Member must satisfy enum value set: [Active, Inactive]</Message>\n </Error>\n <RequestId>079855bd-f5ce-11e5-88c6-23b4e6cc6158</RequestId>\n</ErrorResponse>\n"} NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @./tsecreds.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1
True
setting iam_role state to active or inactive fails with python UnboundLocal error - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME iam module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT OS X 10.11.3 ##### SUMMARY It seems to be impossible to deactivate or revoke AWS access keys via Ansible. ##### STEPS TO REPRODUCE - Attempt to revoke or make inactive an IAM Access Key via the following playbook: ``` --- - hosts: all tasks: - name: Create TSE users with approrpiate group and fetch IAM keys iam: iam_type: user name: "joe.user" state: present access_key_state: "inactive" access_key_ids: - "AKIABRACADABRA" groups: "DEVUSER" profile: dev register: newusers - debug: var=newusers ``` Run with: `ansible-playbook -i "localhost," -c local ./tsecreds.yaml -vvvv` ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269919.04-56226139783789/iam", line 2920, in <module> main() File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269919.04-56226139783789/iam", line 554, in main if any([n in key_state for n in ['active', 'inactive']]) and not key_ids: UnboundLocalError: local variable 'key_ids' referenced before assignment ``` This happens for key states of "Active,Inactive". Using "Remove" will provide this error: ``` fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"access_key_ids": ["AKIABRACADABRA"], "access_key_state": "remove", "aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "groups": ["TSE"], "iam_type": "user", "key_count": 1, "name": "joe.user", "new_name": null, "new_path": null, "password": null, "path": "/", "profile": "dev", "region": null, "security_token": null, "state": "present", "update_password": "always", "validate_certs": true}, "module_name": "iam"}, "msg": "BotoServerError: 400 Bad Request\n<ErrorResponse xmlns=\"https://iam.amazonaws.com/doc/2010-05-08/\">\n <Error>\n <Type>Sender</Type>\n <Code>ValidationError</Code>\n <Message>1 validation error detected: Value 'Remove' at 'status' failed to satisfy constraint: Member must satisfy enum value set: [Active, Inactive]</Message>\n </Error>\n <RequestId>079855bd-f5ce-11e5-88c6-23b4e6cc6158</RequestId>\n</ErrorResponse>\n"} ``` ##### EXPECTED RESULTS I expected 'remove' to remove the keys entirely without first having to run a 'inactive' operation on them (is that instead a feature request?) and I expected the 'inactive' and 'active' operations to work as expected instead of throwing a Python error. Additionally, when running 'create' against a user without any keys, I expected the keys to be available in the debug information after I registered the variable. ##### ACTUAL RESULTS See above inline in 'steps to reproduce' Full output below: ### access_key_state: inactive ``` C02KN05HFFT3:oneoff-tsecreds karl.katzke$ ansible-playbook -i "localhost," -c local ./tsecreds.yaml -vvv No config file found; using defaults 1 plays in ./tsecreds.yaml PLAY *************************************************************************** TASK [setup] ******************************************************************* ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248 `" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmpfBc49B TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248/setup localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248/setup; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269887.0-254944181699248/" > /dev/null 2>&1' ok: [localhost] TASK [Create TSE users with approrpiate group and fetch IAM keys] ************** task path: /Users/karl.katzke/Work/ansible/oneoff-tsecreds/tsecreds.yaml:4 ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406 `" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmpnY3azX TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/" > /dev/null 2>&1' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam", line 2920, in <module> main() File "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459269890.51-92058580067406/iam", line 554, in main if any([n in key_state for n in ['active', 'inactive']]) and not key_ids: UnboundLocalError: local variable 'key_ids' referenced before assignment fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "iam"}, "parsed": false} NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @./tsecreds.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ``` ### access_key_state: remove C02KN05HFFT3:oneoff-tsecreds karl.katzke$ ansible-playbook -i "localhost," -c local ./tsecreds.yaml -vvv No config file found; using defaults 1 plays in ./tsecreds.yaml PLAY *************************************************************************** TASK [setup] ******************************************************************* ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "`echo $HOME/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273`" && echo "`echo $HOME/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273`" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmprZfcCw TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273/setup localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273/setup; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270090.04-124925881060273/" > /dev/null 2>&1' ok: [localhost] TASK [Create TSE users with approrpiate group and fetch IAM keys] ************** task path: /Users/karl.katzke/Work/ansible/oneoff-tsecreds/tsecreds.yaml:4 ESTABLISH LOCAL CONNECTION FOR USER: karl.katzke localhost EXEC /bin/sh -c '( umask 22 && mkdir -p "`echo $HOME/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934`" && echo "`echo $HOME/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934`" )' localhost PUT /var/folders/tp/mry8ntpn6kj551z091w9bl0d_tl3s2/T/tmpHq1d7t TO /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934/iam localhost EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934/iam; rm -rf "/Users/karl.katzke/.ansible/tmp/ansible-tmp-1459270093.31-143517777805934/" > /dev/null 2>&1' fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"access_key_ids": ["AKIAI3BJJ6AEJCLR5MSA"], "access_key_state": "remove", "aws_access_key": null, "aws_secret_key": null, "ec2_url": null, "groups": ["TSE"], "iam_type": "user", "key_count": 1, "name": "brian.outlaw", "new_name": null, "new_path": null, "password": null, "path": "/", "profile": "dev", "region": null, "security_token": null, "state": "present", "update_password": "always", "validate_certs": true}, "module_name": "iam"}, "msg": "BotoServerError: 400 Bad Request\n<ErrorResponse xmlns=\"https://iam.amazonaws.com/doc/2010-05-08/\">\n <Error>\n <Type>Sender</Type>\n <Code>ValidationError</Code>\n <Message>1 validation error detected: Value 'Remove' at 'status' failed to satisfy constraint: Member must satisfy enum value set: [Active, Inactive]</Message>\n </Error>\n <RequestId>079855bd-f5ce-11e5-88c6-23b4e6cc6158</RequestId>\n</ErrorResponse>\n"} NO MORE HOSTS LEFT ************************************************************* to retry, use: --limit @./tsecreds.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1
main
setting iam role state to active or inactive fails with python unboundlocal error issue type bug report component name iam module ansible version ansible config file configured module search path default w o overrides os environment os x summary it seems to be impossible to deactivate or revoke aws access keys via ansible steps to reproduce attempt to revoke or make inactive an iam access key via the following playbook hosts all tasks name create tse users with approrpiate group and fetch iam keys iam iam type user name joe user state present access key state inactive access key ids akiabracadabra groups devuser profile dev register newusers debug var newusers run with ansible playbook i localhost c local tsecreds yaml vvvv an exception occurred during task execution the full traceback is traceback most recent call last file users karl katzke ansible tmp ansible tmp iam line in main file users karl katzke ansible tmp ansible tmp iam line in main if any and not key ids unboundlocalerror local variable key ids referenced before assignment this happens for key states of active inactive using remove will provide this error fatal failed changed false failed true invocation module args access key ids access key state remove aws access key null aws secret key null url null groups iam type user key count name joe user new name null new path null password null path profile dev region null security token null state present update password always validate certs true module name iam msg botoservererror bad request n n sender n validationerror n validation error detected value remove at status failed to satisfy constraint member must satisfy enum value set n n n n expected results i expected remove to remove the keys entirely without first having to run a inactive operation on them is that instead a feature request and i expected the inactive and active operations to work as expected instead of throwing a python error additionally when running create against a user without any keys i expected the keys to be available in the debug information after i registered the variable actual results see above inline in steps to reproduce full output below access key state inactive oneoff tsecreds karl katzke ansible playbook i localhost c local tsecreds yaml vvv no config file found using defaults plays in tsecreds yaml play task establish local connection for user karl katzke localhost exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp localhost put var folders tp t to users karl katzke ansible tmp ansible tmp setup localhost exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users karl katzke ansible tmp ansible tmp setup rm rf users karl katzke ansible tmp ansible tmp dev null ok task task path users karl katzke work ansible oneoff tsecreds tsecreds yaml establish local connection for user karl katzke localhost exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp localhost put var folders tp t to users karl katzke ansible tmp ansible tmp iam localhost exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users karl katzke ansible tmp ansible tmp iam rm rf users karl katzke ansible tmp ansible tmp dev null an exception occurred during task execution the full traceback is traceback most recent call last file users karl katzke ansible tmp ansible tmp iam line in main file users karl katzke ansible tmp ansible tmp iam line in main if any and not key ids unboundlocalerror local variable key ids referenced before assignment fatal failed changed false failed true invocation module name iam parsed false no more hosts left to retry use limit tsecreds retry play recap localhost ok changed unreachable failed access key state remove oneoff tsecreds karl katzke ansible playbook i localhost c local tsecreds yaml vvv no config file found using defaults plays in tsecreds yaml play task establish local connection for user karl katzke localhost exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp localhost put var folders tp t tmprzfccw to users karl katzke ansible tmp ansible tmp setup localhost exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users karl katzke ansible tmp ansible tmp setup rm rf users karl katzke ansible tmp ansible tmp dev null ok task task path users karl katzke work ansible oneoff tsecreds tsecreds yaml establish local connection for user karl katzke localhost exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp localhost put var folders tp t to users karl katzke ansible tmp ansible tmp iam localhost exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python users karl katzke ansible tmp ansible tmp iam rm rf users karl katzke ansible tmp ansible tmp dev null fatal failed changed false failed true invocation module args access key ids access key state remove aws access key null aws secret key null url null groups iam type user key count name brian outlaw new name null new path null password null path profile dev region null security token null state present update password always validate certs true module name iam msg botoservererror bad request n n sender n validationerror n validation error detected value remove at status failed to satisfy constraint member must satisfy enum value set n n n n no more hosts left to retry use limit tsecreds retry play recap localhost ok changed unreachable failed
1
920
4,622,139,259
IssuesEvent
2016-09-27 06:04:18
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_service: timeout option not respected
affects_2.1 bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_service ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### SUMMARY The `timeout` option specified in a docker_service task seems to be not actually used. ##### STEPS TO REPRODUCE SEE https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker_service.py#L837-L862
True
docker_service: timeout option not respected - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_service ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### SUMMARY The `timeout` option specified in a docker_service task seems to be not actually used. ##### STEPS TO REPRODUCE SEE https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker_service.py#L837-L862
main
docker service timeout option not respected issue type bug report component name docker service ansible version ansible summary the timeout option specified in a docker service task seems to be not actually used steps to reproduce see
1
2,958
10,616,585,270
IssuesEvent
2019-10-12 12:54:52
arcticicestudio/snowsaw
https://api.github.com/repos/arcticicestudio/snowsaw
opened
Update to Go 1.13 and latest dependency versions
context-workflow scope-compatibility scope-maintainability scope-performance scope-quality scope-security scope-stability type-task
[Go 1.13 has been released][blog] over a month ago that comes with some great features and a lot stability, performance and security improvements and bug fixes. The [new `os.UserConfigDir()` function][os] is a great addition for the handling for snowsaw's configuration files that will be implemented late on. See the [Go 1.13 official release notes][rln] for more details. Since there are no breaking changes snowsaw will now require Go 1.13 as minimum version. With the update to Go 1.13.x all outdated dependencies should be updated to the latest version too to prevent possible module incompatibilities as well as including the latest improvements and bug fixes. [blog]: https://blog.golang.org/go1.13 [os]: https://golang.org/pkg/os/#UserConfigDir [rln]: https://golang.org/doc/go1.13
True
Update to Go 1.13 and latest dependency versions - [Go 1.13 has been released][blog] over a month ago that comes with some great features and a lot stability, performance and security improvements and bug fixes. The [new `os.UserConfigDir()` function][os] is a great addition for the handling for snowsaw's configuration files that will be implemented late on. See the [Go 1.13 official release notes][rln] for more details. Since there are no breaking changes snowsaw will now require Go 1.13 as minimum version. With the update to Go 1.13.x all outdated dependencies should be updated to the latest version too to prevent possible module incompatibilities as well as including the latest improvements and bug fixes. [blog]: https://blog.golang.org/go1.13 [os]: https://golang.org/pkg/os/#UserConfigDir [rln]: https://golang.org/doc/go1.13
main
update to go and latest dependency versions over a month ago that comes with some great features and a lot stability performance and security improvements and bug fixes the is a great addition for the handling for snowsaw s configuration files that will be implemented late on see the for more details since there are no breaking changes snowsaw will now require go as minimum version with the update to go x all outdated dependencies should be updated to the latest version too to prevent possible module incompatibilities as well as including the latest improvements and bug fixes
1
4,373
22,194,725,363
IssuesEvent
2022-06-07 05:22:40
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
opened
[MAINTAIN] `libarmadillo` should be named `armadillo`
maintain
<!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** In general, library in archlinux should **NOT** contain the `lib` prefix, see, for example, `boost`, there is also a similar aur package, see https://aur.archlinux.org/packages/armadillo Also, `depends` and `repo_depends` in `phyx` should be changed accordingly. **Packages (please complete the following information):** - Involved packages: `libarmadillo`, `phyx` **Description** Add any other context about the problem here.
True
[MAINTAIN] `libarmadillo` should be named `armadillo` - <!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** In general, library in archlinux should **NOT** contain the `lib` prefix, see, for example, `boost`, there is also a similar aur package, see https://aur.archlinux.org/packages/armadillo Also, `depends` and `repo_depends` in `phyx` should be changed accordingly. **Packages (please complete the following information):** - Involved packages: `libarmadillo`, `phyx` **Description** Add any other context about the problem here.
main
libarmadillo should be named armadillo please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug in general library in archlinux should not contain the lib prefix see for example boost there is also a similar aur package see also depends and repo depends in phyx should be changed accordingly packages please complete the following information involved packages libarmadillo phyx description add any other context about the problem here
1
316,864
27,190,120,006
IssuesEvent
2023-02-19 17:49:09
RotherOSS/otobo
https://api.github.com/repos/RotherOSS/otobo
closed
Make creation of test driver object optional in test scripts
enhancement unittests
Using the package `Kernel::System::UnitTest::Driver` adds little value for test scripts that are using mostly `Test2::V0`. Yet an instances of that class is created when `Kernel::System::UnitTest::RegisterDriver;` is loaded. It would be nice to not having to call `use Kernel::System::UnitTest::RegisterDriver;` just for creating the `$Kernel::OM` object. Implementation could be done by moving the import sub `Kernel::System::UnitTest::RegisterDriver::import();` to a dedicated module `Kernel::System::UnitTest::RegisterOM;`. Care must be taken that existing test scripts do not have to be updated. This can be achievied by `Kernel::System::UnitTest::RegisterDriver` depending on `Kernel::System::UnitTest::RegisterOM`.
1.0
Make creation of test driver object optional in test scripts - Using the package `Kernel::System::UnitTest::Driver` adds little value for test scripts that are using mostly `Test2::V0`. Yet an instances of that class is created when `Kernel::System::UnitTest::RegisterDriver;` is loaded. It would be nice to not having to call `use Kernel::System::UnitTest::RegisterDriver;` just for creating the `$Kernel::OM` object. Implementation could be done by moving the import sub `Kernel::System::UnitTest::RegisterDriver::import();` to a dedicated module `Kernel::System::UnitTest::RegisterOM;`. Care must be taken that existing test scripts do not have to be updated. This can be achievied by `Kernel::System::UnitTest::RegisterDriver` depending on `Kernel::System::UnitTest::RegisterOM`.
non_main
make creation of test driver object optional in test scripts using the package kernel system unittest driver adds little value for test scripts that are using mostly yet an instances of that class is created when kernel system unittest registerdriver is loaded it would be nice to not having to call use kernel system unittest registerdriver just for creating the kernel om object implementation could be done by moving the import sub kernel system unittest registerdriver import to a dedicated module kernel system unittest registerom care must be taken that existing test scripts do not have to be updated this can be achievied by kernel system unittest registerdriver depending on kernel system unittest registerom
0
189,017
14,482,792,178
IssuesEvent
2020-12-10 14:24:55
eth-cscs/reframe
https://api.github.com/repos/eth-cscs/reframe
closed
Add openacc host_data test
prio: low regression test
## Crayport Case #164153 - `acc host_data` directive fails if array size is larger than 224x224 you may want to use that code as a new openacc test, if not, just close the issue.
1.0
Add openacc host_data test - ## Crayport Case #164153 - `acc host_data` directive fails if array size is larger than 224x224 you may want to use that code as a new openacc test, if not, just close the issue.
non_main
add openacc host data test crayport case acc host data directive fails if array size is larger than you may want to use that code as a new openacc test if not just close the issue
0
167,708
20,726,285,773
IssuesEvent
2022-03-14 02:33:02
Baneeishaque/Ethiopian_Construction_Directory
https://api.github.com/repos/Baneeishaque/Ethiopian_Construction_Directory
opened
CVE-2021-37712 (High) detected in tar-2.2.1.tgz
security vulnerability
## CVE-2021-37712 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/tar/package.json</p> <p> Dependency Hierarchy: - gulp-sass-4.0.1.tgz (Root Library) - node-sass-4.8.3.tgz - node-gyp-3.6.2.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37712>CVE-2021-37712</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (tar): 4.4.18</p> <p>Direct dependency fix Resolution (gulp-sass): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37712 (High) detected in tar-2.2.1.tgz - ## CVE-2021-37712 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/tar/package.json</p> <p> Dependency Hierarchy: - gulp-sass-4.0.1.tgz (Root Library) - node-sass-4.8.3.tgz - node-gyp-3.6.2.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.18, 5.0.10, and 6.1.9 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value. Additionally, on Windows systems, long path portions would resolve to the same file system entities as their 8.3 "short path" counterparts. A specially crafted tar archive could thus include a directory with one form of the path, followed by a symbolic link with a different string that resolves to the same file system entity, followed by a file using the first form. By first creating a directory, and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. These issues were addressed in releases 4.4.18, 5.0.10 and 6.1.9. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-qq89-hq3f-393p. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37712>CVE-2021-37712</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p">https://github.com/npm/node-tar/security/advisories/GHSA-qq89-hq3f-393p</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution (tar): 4.4.18</p> <p>Direct dependency fix Resolution (gulp-sass): 5.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in tar tgz cve high severity vulnerability vulnerable library tar tgz tar for node library home page a href path to dependency file package json path to vulnerable library node modules tar package json dependency hierarchy gulp sass tgz root library node sass tgz node gyp tgz x tar tgz vulnerable library vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with names containing unicode values that normalized to the same value additionally on windows systems long path portions would resolve to the same file system entities as their short path counterparts a specially crafted tar archive could thus include a directory with one form of the path followed by a symbolic link with a different string that resolves to the same file system entity followed by a file using the first form by first creating a directory and then replacing that directory with a symlink that had a different apparent name that resolved to the same entry in the filesystem it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar direct dependency fix resolution gulp sass step up your open source security game with whitesource
0
980
4,745,802,108
IssuesEvent
2016-10-21 08:48:34
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Issue with aws rds module - returned endpoint address has a bunch of asterisks.
affects_2.0 aws bug_report cloud waiting_on_maintainer
## Issue Type Bug Report ## Component Name rds ## Ansible Version 2.0.0.2 ## Environment Ansible 2.0.0.2 Ubuntu 14.03 AWS ## Summary I am trying to create an AWS RDS service and then write out the endpoint address to a file, but for some reason once the RDS service is available the endpoint address retrieved from the registered variable has a bunch of asterisks in the name? ### playbook ```` tasks: - name: Launch RDS instances rds: region: "{{ ec2_region }}" ec2_access_key: "{{ ec2_access_key }}" ec2_secret_key: "{{ ec2_secret_key }}" command: create instance_name: oraclelinkstardb db_engine: oracle-se instance_type: db.t2.micro username: linkstar password: xxxxx size: 10 wait: yes wait_timeout: 2000 vpc_security_groups: sg-7ab71302 tags: Environment: dev Application: linkstar with_items: ec2_instances register: rds_result - name: write rds host info to file local_action: module: copy content: "{{ item.instance.endpoint }}" dest: ./files/rdsendpoint.txt with_items: "{{ rds_result.results|default([]) }}" ```` ### PlayBook Output ```` TASK [write rds host info to file] ********************************************* task path: /home/arlindo/projects/aws/provision.yml:37 ESTABLISH LOCAL CONNECTION FOR USER: arlindo localhost EXEC rc=flag; [ -r ./files/rdsendpoint.txt ] || rc=2; [ -f ./files/rdsendpoint.txt ] || rc=1; [ -d ./files/rdsendpoint.txt ] && rc=3; python -V 2>/dev/null || rc=4; [ x"$rc" != "xflag" ] && echo "${rc} "./files/rdsendpoint.txt && exit 0; (python -c 'import hashlib; BLOCKSIZE = 65536; hasher = hashlib.sha1(); afile = open("'./files/rdsendpoint.txt'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (python -c 'import sha; BLOCKSIZE = 65536; hasher = sha.sha(); afile = open("'./files/rdsendpoint.txt'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (echo '0 './files/rdsendpoint.txt) localhost EXEC ( umask 22 && mkdir -p "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852 )" && echo "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852 )" ) localhost PUT /tmp/tmpBne1eN TO /home/arlindo/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852/source localhost EXEC ( umask 22 && mkdir -p "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681 )" && echo "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681 )" ) localhost PUT /tmp/tmpEMzg0F TO /home/arlindo/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681/copy localhost EXEC LANG=en_CA.UTF-8 LC_ALL=en_CA.UTF-8 LC_MESSAGES=en_CA.UTF-8 /usr/bin/python /home/arlindo/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681/copy; rm -rf "/home/arlindo/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681/" > /dev/null 2>&1 changed: [localhost -> localhost] => (item={'invocation': {'module_name': u'rds', u'module_args': {u'profile': None, u'db_engine': u'oracle-se', u'iops': None, u'publicly_accessible': None, u'ec2_url': None, u'backup_retention': None, u'port': None, u'security_groups': None, u'size': 10, u'aws_secret_key': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'subnet': None, u'vpc_security_groups': [u'sg-7ab71302'], u'upgrade': False, u'zone': None, u'source_instance': None, u'parameter_group': None, u'command': u'create', u'multi_zone': False, u'new_instance_name': None, u'username': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'tags': {u'Environment': u'dev', u'Application': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER'}, u'character_set_name': None, u'db_name': None, u'license_model': None, u'ec2_access_key': u'', u'ec2_secret_key': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'apply_immediately': False, u'wait': True, u'aws_access_key': u'', u'security_token': None, u'force_failover': False, u'maint_window': None, u'region': u'us-east-1', u'option_group': None, u'engine_version': None, u'instance_name': u'oracle********db', u'instance_type': u'db.t2.micro', u'password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'wait_timeout': 2000, u'snapshot': None, u'backup_window': None, u'validate_certs': True}}, u'instance': {u'status': u'available', u'username': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'vpc_security_groups': u'sg-7ab71302', u'endpoint': u'oracle********db.cgxmjxhypelr.us-east-1.rds.amazonaws.com', u'availability_zone': u'us-east-1e', u'port': 1521, u'replication_source': None, u'instance_type': u'db.t2.micro', u'iops': None, u'create_time': 1456801030.478, u'backup_retention': 1, u'multi_zone': False, u'id': u'oracle********db', u'maintenance_window': u'mon:09:02-mon:09:32'}, u'changed': True, '_ansible_no_log': False, 'item': {u'group': [u'web', u'winrdp', u'default'], u'count_tag': {u'Name': u'webwin'}, u'exact_count': 1, u'instance_type': u't2.micro', u'keypair': u'ansible', u'instance_tags': {u'Name': u'webwin'}, u'image': u'ami-42596f28'}}) => {"changed": true, "checksum": "92761a1042a66cf97c67626d7a307b6e4342f8c4", "dest": "./files/rdsendpoint.txt", "gid": 1000, "group": "arlindo", "invocation": {"module_args": {"backup": false, "content": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "delimiter": null, "dest": "./files/rdsendpoint.txt", "directory_mode": null, "follow": false, "force": true, "group": null, "mode": null, "original_basename": "tmpBne1eN", "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": "/home/arlindo/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852/source", "validate": null}}, "item": {"_ansible_no_log": false, "changed": true, "instance": {"availability_zone": "us-east-1e", "backup_retention": 1, "create_time": 1456801030.478, "endpoint": "oracle********db.cgxmjxhypelr.us-east-1.rds.amazonaws.com", "id": "oracle********db", "instance_type": "db.t2.micro", "iops": null, "maintenance_window": "mon:09:02-mon:09:32", "multi_zone": false, "port": 1521, "replication_source": null, "status": "available", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "vpc_security_groups": "sg-7ab71302"}, "invocation": {"module_args": {"apply_immediately": false, "aws_access_key": "", "aws_secret_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "backup_retention": null, "backup_window": null, "character_set_name": null, "command": "create", "db_engine": "oracle-se", "db_name": null, "ec2_access_key": "", "ec2_secret_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "ec2_url": null, "engine_version": null, "force_failover": false, "instance_name": "oracle********db", "instance_type": "db.t2.micro", "iops": null, "license_model": null, "maint_window": null, "multi_zone": false, "new_instance_name": null, "option_group": null, "parameter_group": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "profile": null, "publicly_accessible": null, "region": "us-east-1", "security_groups": null, "security_token": null, "size": 10, "snapshot": null, "source_instance": null, "subnet": null, "tags": {"Application": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "Environment": "dev"}, "upgrade": false, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true, "vpc_security_groups": ["sg-7ab71302"], "wait": true, "wait_timeout": 2000, "zone": null}, "module_name": "rds"}, "item": {"count_tag": {"Name": "webwin"}, "exact_count": 1, "group": ["web", "winrdp", "default"], "image": "ami-42596f28", "instance_tags": {"Name": "webwin"}, "instance_type": "t2.micro", "keypair": "ansible"}}, "md5sum": "aa41800bfab755d8f1787ea90e7f0233", "mode": "0664", "owner": "arlindo", "size": 57, "src": "/home/arlindo/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852/source", "state": "file", "uid": 1000} ````
True
Issue with aws rds module - returned endpoint address has a bunch of asterisks. - ## Issue Type Bug Report ## Component Name rds ## Ansible Version 2.0.0.2 ## Environment Ansible 2.0.0.2 Ubuntu 14.03 AWS ## Summary I am trying to create an AWS RDS service and then write out the endpoint address to a file, but for some reason once the RDS service is available the endpoint address retrieved from the registered variable has a bunch of asterisks in the name? ### playbook ```` tasks: - name: Launch RDS instances rds: region: "{{ ec2_region }}" ec2_access_key: "{{ ec2_access_key }}" ec2_secret_key: "{{ ec2_secret_key }}" command: create instance_name: oraclelinkstardb db_engine: oracle-se instance_type: db.t2.micro username: linkstar password: xxxxx size: 10 wait: yes wait_timeout: 2000 vpc_security_groups: sg-7ab71302 tags: Environment: dev Application: linkstar with_items: ec2_instances register: rds_result - name: write rds host info to file local_action: module: copy content: "{{ item.instance.endpoint }}" dest: ./files/rdsendpoint.txt with_items: "{{ rds_result.results|default([]) }}" ```` ### PlayBook Output ```` TASK [write rds host info to file] ********************************************* task path: /home/arlindo/projects/aws/provision.yml:37 ESTABLISH LOCAL CONNECTION FOR USER: arlindo localhost EXEC rc=flag; [ -r ./files/rdsendpoint.txt ] || rc=2; [ -f ./files/rdsendpoint.txt ] || rc=1; [ -d ./files/rdsendpoint.txt ] && rc=3; python -V 2>/dev/null || rc=4; [ x"$rc" != "xflag" ] && echo "${rc} "./files/rdsendpoint.txt && exit 0; (python -c 'import hashlib; BLOCKSIZE = 65536; hasher = hashlib.sha1(); afile = open("'./files/rdsendpoint.txt'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (python -c 'import sha; BLOCKSIZE = 65536; hasher = sha.sha(); afile = open("'./files/rdsendpoint.txt'", "rb") buf = afile.read(BLOCKSIZE) while len(buf) > 0: hasher.update(buf) buf = afile.read(BLOCKSIZE) afile.close() print(hasher.hexdigest())' 2>/dev/null) || (echo '0 './files/rdsendpoint.txt) localhost EXEC ( umask 22 && mkdir -p "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852 )" && echo "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852 )" ) localhost PUT /tmp/tmpBne1eN TO /home/arlindo/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852/source localhost EXEC ( umask 22 && mkdir -p "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681 )" && echo "$( echo $HOME/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681 )" ) localhost PUT /tmp/tmpEMzg0F TO /home/arlindo/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681/copy localhost EXEC LANG=en_CA.UTF-8 LC_ALL=en_CA.UTF-8 LC_MESSAGES=en_CA.UTF-8 /usr/bin/python /home/arlindo/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681/copy; rm -rf "/home/arlindo/.ansible/tmp/ansible-tmp-1456801607.24-41001705013681/" > /dev/null 2>&1 changed: [localhost -> localhost] => (item={'invocation': {'module_name': u'rds', u'module_args': {u'profile': None, u'db_engine': u'oracle-se', u'iops': None, u'publicly_accessible': None, u'ec2_url': None, u'backup_retention': None, u'port': None, u'security_groups': None, u'size': 10, u'aws_secret_key': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'subnet': None, u'vpc_security_groups': [u'sg-7ab71302'], u'upgrade': False, u'zone': None, u'source_instance': None, u'parameter_group': None, u'command': u'create', u'multi_zone': False, u'new_instance_name': None, u'username': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'tags': {u'Environment': u'dev', u'Application': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER'}, u'character_set_name': None, u'db_name': None, u'license_model': None, u'ec2_access_key': u'', u'ec2_secret_key': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'apply_immediately': False, u'wait': True, u'aws_access_key': u'', u'security_token': None, u'force_failover': False, u'maint_window': None, u'region': u'us-east-1', u'option_group': None, u'engine_version': None, u'instance_name': u'oracle********db', u'instance_type': u'db.t2.micro', u'password': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'wait_timeout': 2000, u'snapshot': None, u'backup_window': None, u'validate_certs': True}}, u'instance': {u'status': u'available', u'username': u'VALUE_SPECIFIED_IN_NO_LOG_PARAMETER', u'vpc_security_groups': u'sg-7ab71302', u'endpoint': u'oracle********db.cgxmjxhypelr.us-east-1.rds.amazonaws.com', u'availability_zone': u'us-east-1e', u'port': 1521, u'replication_source': None, u'instance_type': u'db.t2.micro', u'iops': None, u'create_time': 1456801030.478, u'backup_retention': 1, u'multi_zone': False, u'id': u'oracle********db', u'maintenance_window': u'mon:09:02-mon:09:32'}, u'changed': True, '_ansible_no_log': False, 'item': {u'group': [u'web', u'winrdp', u'default'], u'count_tag': {u'Name': u'webwin'}, u'exact_count': 1, u'instance_type': u't2.micro', u'keypair': u'ansible', u'instance_tags': {u'Name': u'webwin'}, u'image': u'ami-42596f28'}}) => {"changed": true, "checksum": "92761a1042a66cf97c67626d7a307b6e4342f8c4", "dest": "./files/rdsendpoint.txt", "gid": 1000, "group": "arlindo", "invocation": {"module_args": {"backup": false, "content": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "delimiter": null, "dest": "./files/rdsendpoint.txt", "directory_mode": null, "follow": false, "force": true, "group": null, "mode": null, "original_basename": "tmpBne1eN", "owner": null, "regexp": null, "remote_src": null, "selevel": null, "serole": null, "setype": null, "seuser": null, "src": "/home/arlindo/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852/source", "validate": null}}, "item": {"_ansible_no_log": false, "changed": true, "instance": {"availability_zone": "us-east-1e", "backup_retention": 1, "create_time": 1456801030.478, "endpoint": "oracle********db.cgxmjxhypelr.us-east-1.rds.amazonaws.com", "id": "oracle********db", "instance_type": "db.t2.micro", "iops": null, "maintenance_window": "mon:09:02-mon:09:32", "multi_zone": false, "port": 1521, "replication_source": null, "status": "available", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "vpc_security_groups": "sg-7ab71302"}, "invocation": {"module_args": {"apply_immediately": false, "aws_access_key": "", "aws_secret_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "backup_retention": null, "backup_window": null, "character_set_name": null, "command": "create", "db_engine": "oracle-se", "db_name": null, "ec2_access_key": "", "ec2_secret_key": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "ec2_url": null, "engine_version": null, "force_failover": false, "instance_name": "oracle********db", "instance_type": "db.t2.micro", "iops": null, "license_model": null, "maint_window": null, "multi_zone": false, "new_instance_name": null, "option_group": null, "parameter_group": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "profile": null, "publicly_accessible": null, "region": "us-east-1", "security_groups": null, "security_token": null, "size": 10, "snapshot": null, "source_instance": null, "subnet": null, "tags": {"Application": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "Environment": "dev"}, "upgrade": false, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true, "vpc_security_groups": ["sg-7ab71302"], "wait": true, "wait_timeout": 2000, "zone": null}, "module_name": "rds"}, "item": {"count_tag": {"Name": "webwin"}, "exact_count": 1, "group": ["web", "winrdp", "default"], "image": "ami-42596f28", "instance_tags": {"Name": "webwin"}, "instance_type": "t2.micro", "keypair": "ansible"}}, "md5sum": "aa41800bfab755d8f1787ea90e7f0233", "mode": "0664", "owner": "arlindo", "size": 57, "src": "/home/arlindo/.ansible/tmp/ansible-tmp-1456801607.23-107237530777852/source", "state": "file", "uid": 1000} ````
main
issue with aws rds module returned endpoint address has a bunch of asterisks issue type bug report component name rds ansible version environment ansible ubuntu aws summary i am trying to create an aws rds service and then write out the endpoint address to a file but for some reason once the rds service is available the endpoint address retrieved from the registered variable has a bunch of asterisks in the name playbook tasks name launch rds instances rds region region access key access key secret key secret key command create instance name oraclelinkstardb db engine oracle se instance type db micro username linkstar password xxxxx size wait yes wait timeout vpc security groups sg tags environment dev application linkstar with items instances register rds result name write rds host info to file local action module copy content item instance endpoint dest files rdsendpoint txt with items rds result results default playbook output task task path home arlindo projects aws provision yml establish local connection for user arlindo localhost exec rc flag rc rc rc python v dev null rc echo rc files rdsendpoint txt exit python c import hashlib blocksize hasher hashlib afile open files rdsendpoint txt rb buf afile read blocksize while len buf hasher update buf buf afile read blocksize afile close print hasher hexdigest dev null python c import sha blocksize hasher sha sha afile open files rdsendpoint txt rb buf afile read blocksize while len buf hasher update buf buf afile read blocksize afile close print hasher hexdigest dev null echo files rdsendpoint txt localhost exec umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp localhost put tmp to home arlindo ansible tmp ansible tmp source localhost exec umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp localhost put tmp to home arlindo ansible tmp ansible tmp copy localhost exec lang en ca utf lc all en ca utf lc messages en ca utf usr bin python home arlindo ansible tmp ansible tmp copy rm rf home arlindo ansible tmp ansible tmp dev null changed item invocation module name u rds u module args u profile none u db engine u oracle se u iops none u publicly accessible none u url none u backup retention none u port none u security groups none u size u aws secret key u value specified in no log parameter u subnet none u vpc security groups u upgrade false u zone none u source instance none u parameter group none u command u create u multi zone false u new instance name none u username u value specified in no log parameter u tags u environment u dev u application u value specified in no log parameter u character set name none u db name none u license model none u access key u u secret key u value specified in no log parameter u apply immediately false u wait true u aws access key u u security token none u force failover false u maint window none u region u us east u option group none u engine version none u instance name u oracle db u instance type u db micro u password u value specified in no log parameter u wait timeout u snapshot none u backup window none u validate certs true u instance u status u available u username u value specified in no log parameter u vpc security groups u sg u endpoint u oracle db cgxmjxhypelr us east rds amazonaws com u availability zone u us east u port u replication source none u instance type u db micro u iops none u create time u backup retention u multi zone false u id u oracle db u maintenance window u mon mon u changed true ansible no log false item u group u count tag u name u webwin u exact count u instance type u micro u keypair u ansible u instance tags u name u webwin u image u ami changed true checksum dest files rdsendpoint txt gid group arlindo invocation module args backup false content value specified in no log parameter delimiter null dest files rdsendpoint txt directory mode null follow false force true group null mode null original basename owner null regexp null remote src null selevel null serole null setype null seuser null src home arlindo ansible tmp ansible tmp source validate null item ansible no log false changed true instance availability zone us east backup retention create time endpoint oracle db cgxmjxhypelr us east rds amazonaws com id oracle db instance type db micro iops null maintenance window mon mon multi zone false port replication source null status available username value specified in no log parameter vpc security groups sg invocation module args apply immediately false aws access key aws secret key value specified in no log parameter backup retention null backup window null character set name null command create db engine oracle se db name null access key secret key value specified in no log parameter url null engine version null force failover false instance name oracle db instance type db micro iops null license model null maint window null multi zone false new instance name null option group null parameter group null password value specified in no log parameter port null profile null publicly accessible null region us east security groups null security token null size snapshot null source instance null subnet null tags application value specified in no log parameter environment dev upgrade false username value specified in no log parameter validate certs true vpc security groups wait true wait timeout zone null module name rds item count tag name webwin exact count group image ami instance tags name webwin instance type micro keypair ansible mode owner arlindo size src home arlindo ansible tmp ansible tmp source state file uid
1
3,032
11,215,932,668
IssuesEvent
2020-01-07 04:13:57
SethMMorton/natsort
https://api.github.com/repos/SethMMorton/natsort
closed
Use "on_fail" option of fastnumbers instead of key
maintainability wontfix
When https://github.com/SethMMorton/fastnumbers/issues/32 is merged and released as version 3.0.0, `natsort` should change the (optionally) required `fastnumbers` version to 3.0.0 and change its use of `key` to `on_fail`.
True
Use "on_fail" option of fastnumbers instead of key - When https://github.com/SethMMorton/fastnumbers/issues/32 is merged and released as version 3.0.0, `natsort` should change the (optionally) required `fastnumbers` version to 3.0.0 and change its use of `key` to `on_fail`.
main
use on fail option of fastnumbers instead of key when is merged and released as version natsort should change the optionally required fastnumbers version to and change its use of key to on fail
1
635,245
20,382,485,663
IssuesEvent
2022-02-22 00:37:36
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Priority inversion in CAN
bug priority: medium area: CAN has-pr
The drivers in the CAN subsystem suffer from priority inversion. This causes urgent frames to be delayed arbitrarily, and in the case of the FlexCAN drivers, to cause the CPU to spin for an arbitrary time. I explain CAN priority inversion in a blog post here: https://kentindell.github.io/2020/06/29/can-priority-inversion/ The solution to this is to rewrite the CAN drivers. Some software may rely on FIFO processing (e.g. for segmented messages) and a separate API for sending frames according to FIFO will probably be necessary. The blog post discusses this in more detail.
1.0
Priority inversion in CAN - The drivers in the CAN subsystem suffer from priority inversion. This causes urgent frames to be delayed arbitrarily, and in the case of the FlexCAN drivers, to cause the CPU to spin for an arbitrary time. I explain CAN priority inversion in a blog post here: https://kentindell.github.io/2020/06/29/can-priority-inversion/ The solution to this is to rewrite the CAN drivers. Some software may rely on FIFO processing (e.g. for segmented messages) and a separate API for sending frames according to FIFO will probably be necessary. The blog post discusses this in more detail.
non_main
priority inversion in can the drivers in the can subsystem suffer from priority inversion this causes urgent frames to be delayed arbitrarily and in the case of the flexcan drivers to cause the cpu to spin for an arbitrary time i explain can priority inversion in a blog post here the solution to this is to rewrite the can drivers some software may rely on fifo processing e g for segmented messages and a separate api for sending frames according to fifo will probably be necessary the blog post discusses this in more detail
0
318,936
27,333,446,183
IssuesEvent
2023-02-25 22:59:51
fedimint/fedimint
https://api.github.com/repos/fedimint/fedimint
opened
Test that canceling HTLCs actually cancels HTLCs
lightning testing
In our lightning integrations, we send messages to the CLN / LND to cancel, or ignore HTLCs. We don't have test coverage checking that these behave as we expect.
1.0
Test that canceling HTLCs actually cancels HTLCs - In our lightning integrations, we send messages to the CLN / LND to cancel, or ignore HTLCs. We don't have test coverage checking that these behave as we expect.
non_main
test that canceling htlcs actually cancels htlcs in our lightning integrations we send messages to the cln lnd to cancel or ignore htlcs we don t have test coverage checking that these behave as we expect
0
340,315
10,270,608,956
IssuesEvent
2019-08-23 12:07:06
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
open.spotify.com - video or audio doesn't play
browser-firefox engine-gecko priority-important type-tracking-protection-basic
<!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN **Browser / Version**: Firefox 68.0 **Operating System**: Windows 10 **Tested Another Browser**: Unknown **Problem type**: Video or audio doesn't play **Description**: Sound stoped, controls dont work **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/8/9c9099bc-f2bd-48db-8db9-0e079bdf8f33-thumb.jpeg)](https://webcompat.com/uploads/2019/8/9c9099bc-f2bd-48db-8db9-0e079bdf8f33.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190819184224</li><li>tracking content blocked: true (basic)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Se carg script desde https://www.fastly-insights.com/insights.js?k=040e3997-282c-4275-ba9b-a406ce78b133&dnt=1 aunque su tipo MIME (text/html) no es un tipo MIME vlido de JavaScript." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://www.google-analytics.com/analytics.js porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 2ms)" {file: "https://open.scdn.co/static/vendor.f4dcb69a.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 0ms)" {file: "https://open.scdn.co/static/vendor.f4dcb69a.js" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://sb.scorecardresearch.com/beacon.js porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignorando \'unsafe-inline\' dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando https: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando http: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Se bloque el recurso en https://tapestry.tapad.com/tapestry/1?ta_partner_id=950&ta_redirect=https%3A%2F%2Ft.myvisualiq.net%2Fsync%3Fprid%3D1001%26ao%3D0%26pruuid%3DTAPAD_%24%7BIDS%3Akey%7D porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignorando x-frame-options por la directiva frame-ancestors."]', u'[JavaScript Warning: "Content Security Policy: Ignorando \'unsafe-inline\' dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando https: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando http: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando \'unsafe-inline\' dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando https: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando http: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Se bloque el recurso en https://idsync.rlcdn.com/420356.gif?partner_uid=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://dpm.demdex.net/ibs:dpid=125310&dpuuid=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6&redir=https%3A%2F%2Ft.myvisualiq.net%2Fsync%3Fprid%3DAOEPNR1%26ao%3D0%26pruuid%3D%24%7BDD_UUID%7D porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://loadus.exelator.com/load/?p=204&g=1260&buid=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://www.facebook.com/tr?id=1901136573502950&ev=PageView&cd[order_id]=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://tags.bluekai.com/site/21398?id=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Blocked https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN from extracting canvas data because no user input was detected." {file: "https://open.scdn.co/static/web-player.a3c930d1.js" line: 1}]', u'[JavaScript Warning: "Content Security Policy: Ignorando x-frame-options por la directiva frame-ancestors."]', u'[JavaScript Warning: "Content Security Policy: Ignorando x-frame-options por la directiva frame-ancestors."]', u'[JavaScript Error: "uncaught exception: Object"]', u'[JavaScript Warning: "onmozfullscreenchange est en desuso." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "onmozfullscreenchange est en desuso." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
open.spotify.com - video or audio doesn't play - <!-- @browser: Firefox 68.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; rv:68.0) Gecko/20100101 Firefox/68.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-tracking-protection-basic --> **URL**: https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN **Browser / Version**: Firefox 68.0 **Operating System**: Windows 10 **Tested Another Browser**: Unknown **Problem type**: Video or audio doesn't play **Description**: Sound stoped, controls dont work **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/8/9c9099bc-f2bd-48db-8db9-0e079bdf8f33-thumb.jpeg)](https://webcompat.com/uploads/2019/8/9c9099bc-f2bd-48db-8db9-0e079bdf8f33.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190819184224</li><li>tracking content blocked: true (basic)</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: beta</li> </ul> <p>Console Messages:</p> <pre> [u'[JavaScript Warning: "Se carg script desde https://www.fastly-insights.com/insights.js?k=040e3997-282c-4275-ba9b-a406ce78b133&dnt=1 aunque su tipo MIME (text/html) no es un tipo MIME vlido de JavaScript." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://www.google-analytics.com/analytics.js porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 2ms)" {file: "https://open.scdn.co/static/vendor.f4dcb69a.js" line: 0}]', u'[JavaScript Warning: "Successfully compiled asm.js code (total compilation time 0ms)" {file: "https://open.scdn.co/static/vendor.f4dcb69a.js" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://sb.scorecardresearch.com/beacon.js porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignorando \'unsafe-inline\' dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando https: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando http: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Se bloque el recurso en https://tapestry.tapad.com/tapestry/1?ta_partner_id=950&ta_redirect=https%3A%2F%2Ft.myvisualiq.net%2Fsync%3Fprid%3D1001%26ao%3D0%26pruuid%3DTAPAD_%24%7BIDS%3Akey%7D porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Content Security Policy: Ignorando x-frame-options por la directiva frame-ancestors."]', u'[JavaScript Warning: "Content Security Policy: Ignorando \'unsafe-inline\' dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando https: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando http: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando \'unsafe-inline\' dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando https: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Content Security Policy: Ignorando http: dentro de script-src: strict-dynamic especificado"]', u'[JavaScript Warning: "Se bloque el recurso en https://idsync.rlcdn.com/420356.gif?partner_uid=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://dpm.demdex.net/ibs:dpid=125310&dpuuid=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6&redir=https%3A%2F%2Ft.myvisualiq.net%2Fsync%3Fprid%3DAOEPNR1%26ao%3D0%26pruuid%3D%24%7BDD_UUID%7D porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://loadus.exelator.com/load/?p=204&g=1260&buid=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://www.facebook.com/tr?id=1901136573502950&ev=PageView&cd[order_id]=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Se bloque el recurso en https://tags.bluekai.com/site/21398?id=4dbbb790-5a6a-4f7c-ac13-6990c7e176a6 porque la proteccin de rastreo est habilitada." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "Blocked https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN from extracting canvas data because no user input was detected." {file: "https://open.scdn.co/static/web-player.a3c930d1.js" line: 1}]', u'[JavaScript Warning: "Content Security Policy: Ignorando x-frame-options por la directiva frame-ancestors."]', u'[JavaScript Warning: "Content Security Policy: Ignorando x-frame-options por la directiva frame-ancestors."]', u'[JavaScript Error: "uncaught exception: Object"]', u'[JavaScript Warning: "onmozfullscreenchange est en desuso." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]', u'[JavaScript Warning: "onmozfullscreenchange est en desuso." {file: "https://open.spotify.com/album/02GpJacLYR0D41zssOxvlN" line: 0}]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
open spotify com video or audio doesn t play url browser version firefox operating system windows tested another browser unknown problem type video or audio doesn t play description sound stoped controls dont work steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked true basic gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel beta console messages u u u u u u u u u u u u u u u u u u u porque la proteccin de rastreo est habilitada file line u u u u u u u from with ❤️
0