Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
260,220
8,205,125,964
IssuesEvent
2018-09-03 09:08:12
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
closed
ONE_MANAGED should be renamed to VCENTER_IMPORTED
Category: vCenter Priority: Low Status: Accepted Type: Bug
**Description** A clear and concise description of what the bug is. OPENNEBULA_MANAGED is a attribute for disks and nics with the purpose of mark these vm resources with a flag that tell us that are managed by vCenter. Raw resources like Images or networks should have the VCENTER_IMPORTED attribute. Used as protection (dont remove imported resources). ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
ONE_MANAGED should be renamed to VCENTER_IMPORTED - **Description** A clear and concise description of what the bug is. OPENNEBULA_MANAGED is a attribute for disks and nics with the purpose of mark these vm resources with a flag that tell us that are managed by vCenter. Raw resources like Images or networks should have the VCENTER_IMPORTED attribute. Used as protection (dont remove imported resources). ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
non_main
one managed should be renamed to vcenter imported description a clear and concise description of what the bug is opennebula managed is a attribute for disks and nics with the purpose of mark these vm resources with a flag that tell us that are managed by vcenter raw resources like images or networks should have the vcenter imported attribute used as protection dont remove imported resources progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
0
4,152
19,760,947,002
IssuesEvent
2022-01-16 12:04:44
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
closed
Ctor should set default values for a property that returns an Enum
feature Area: analyzer Area: maintainability
To make it clear to the developer and to prevent accidentally changed behavior, properties that return `Enum` values should be set via the `.ctor`. Following code **SHOULD** report an issue: ```c# public class TestMe { public TestMe() { } public StringComparison Comparison { get; } } ``` Reason: The `.ctor` does not set the value directly, and there is also no value set via any property initializer. In contrast to the example above, following code **should NOT** report an issue: ```c# public class TestMe { public TestMe() { } public StringComparison Comparison { get; } = StringComparison.OrdinalIgnoreCase; } ``` Reason: The `.ctor` does not set the value directly, but it is set via property initializer.
True
Ctor should set default values for a property that returns an Enum - To make it clear to the developer and to prevent accidentally changed behavior, properties that return `Enum` values should be set via the `.ctor`. Following code **SHOULD** report an issue: ```c# public class TestMe { public TestMe() { } public StringComparison Comparison { get; } } ``` Reason: The `.ctor` does not set the value directly, and there is also no value set via any property initializer. In contrast to the example above, following code **should NOT** report an issue: ```c# public class TestMe { public TestMe() { } public StringComparison Comparison { get; } = StringComparison.OrdinalIgnoreCase; } ``` Reason: The `.ctor` does not set the value directly, but it is set via property initializer.
main
ctor should set default values for a property that returns an enum to make it clear to the developer and to prevent accidentally changed behavior properties that return enum values should be set via the ctor following code should report an issue c public class testme public testme public stringcomparison comparison get reason the ctor does not set the value directly and there is also no value set via any property initializer in contrast to the example above following code should not report an issue c public class testme public testme public stringcomparison comparison get stringcomparison ordinalignorecase reason the ctor does not set the value directly but it is set via property initializer
1
2,644
9,022,892,595
IssuesEvent
2019-02-07 04:10:56
reactiveops/reckoner
https://api.github.com/repos/reactiveops/reckoner
reopened
certain subcommands don't work
Maintainability Usability good first issue
Might suggest removing them as they do not work today. `reckoner generate` Stacktrace -- see below: ``` reckoner generate 2019-01-04 14:23:03 xps13-9370 root[16862] INFO Generating example course as course.yml Traceback (most recent call last): File "/home/user/.pentagon/venv/bin/reckoner", line 11, in <module> sys.exit(cli()) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 697, in main rv = self.invoke(ctx) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 1066, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/reckoner/cli.py", line 62, in generate src = pkg_resources.resource_string("reckoner", "example-course.yml") File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/pkg_resources/__init__.py", line 1152, in resource_string self, resource_name File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/pkg_resources/__init__.py", line 1396, in get_resource_string return self._get(self._fn(self.module_path, resource_name)) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/pkg_resources/__init__.py", line 1519, in _get with open(path, 'rb') as stream: IOError: [Errno 2] No such file or directory: '/home/user/.pentagon/venv/local/lib/python2.7/site-packages/reckoner/example-course.yml' ``` `reckoner version` ``` Traceback (most recent call last): File "/home/nhuanca/.pentagon/venv/bin/reckoner", line 11, in <module> sys.exit(cli()) File "/path/to/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/path/to/click/core.py", line 697, in main rv = self.invoke(ctx) File "/path/to/click/core.py", line 1066, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/path/to/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/path/to/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/path/to/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/path/to/reckoner/cli.py", line 72, in version print(reckoner.__version__) NameError: global name 'reckoner' is not defined ```
True
certain subcommands don't work - Might suggest removing them as they do not work today. `reckoner generate` Stacktrace -- see below: ``` reckoner generate 2019-01-04 14:23:03 xps13-9370 root[16862] INFO Generating example course as course.yml Traceback (most recent call last): File "/home/user/.pentagon/venv/bin/reckoner", line 11, in <module> sys.exit(cli()) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 697, in main rv = self.invoke(ctx) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 1066, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/reckoner/cli.py", line 62, in generate src = pkg_resources.resource_string("reckoner", "example-course.yml") File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/pkg_resources/__init__.py", line 1152, in resource_string self, resource_name File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/pkg_resources/__init__.py", line 1396, in get_resource_string return self._get(self._fn(self.module_path, resource_name)) File "/home/user/.pentagon/venv/local/lib/python2.7/site-packages/pkg_resources/__init__.py", line 1519, in _get with open(path, 'rb') as stream: IOError: [Errno 2] No such file or directory: '/home/user/.pentagon/venv/local/lib/python2.7/site-packages/reckoner/example-course.yml' ``` `reckoner version` ``` Traceback (most recent call last): File "/home/nhuanca/.pentagon/venv/bin/reckoner", line 11, in <module> sys.exit(cli()) File "/path/to/click/core.py", line 722, in __call__ return self.main(*args, **kwargs) File "/path/to/click/core.py", line 697, in main rv = self.invoke(ctx) File "/path/to/click/core.py", line 1066, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/path/to/click/core.py", line 895, in invoke return ctx.invoke(self.callback, **ctx.params) File "/path/to/click/core.py", line 535, in invoke return callback(*args, **kwargs) File "/path/to/click/decorators.py", line 17, in new_func return f(get_current_context(), *args, **kwargs) File "/path/to/reckoner/cli.py", line 72, in version print(reckoner.__version__) NameError: global name 'reckoner' is not defined ```
main
certain subcommands don t work might suggest removing them as they do not work today reckoner generate stacktrace see below reckoner generate root info generating example course as course yml traceback most recent call last file home user pentagon venv bin reckoner line in sys exit cli file home user pentagon venv local lib site packages click core py line in call return self main args kwargs file home user pentagon venv local lib site packages click core py line in main rv self invoke ctx file home user pentagon venv local lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file home user pentagon venv local lib site packages click core py line in invoke return ctx invoke self callback ctx params file home user pentagon venv local lib site packages click core py line in invoke return callback args kwargs file home user pentagon venv local lib site packages click decorators py line in new func return f get current context args kwargs file home user pentagon venv local lib site packages reckoner cli py line in generate src pkg resources resource string reckoner example course yml file home user pentagon venv local lib site packages pkg resources init py line in resource string self resource name file home user pentagon venv local lib site packages pkg resources init py line in get resource string return self get self fn self module path resource name file home user pentagon venv local lib site packages pkg resources init py line in get with open path rb as stream ioerror no such file or directory home user pentagon venv local lib site packages reckoner example course yml reckoner version traceback most recent call last file home nhuanca pentagon venv bin reckoner line in sys exit cli file path to click core py line in call return self main args kwargs file path to click core py line in main rv self invoke ctx file path to click core py line in invoke return process result sub ctx command invoke sub ctx file path to click core py line in invoke return ctx invoke self callback ctx params file path to click core py line in invoke return callback args kwargs file path to click decorators py line in new func return f get current context args kwargs file path to reckoner cli py line in version print reckoner version nameerror global name reckoner is not defined
1
75,224
20,732,500,213
IssuesEvent
2022-03-14 10:43:06
denoland/deno
https://api.github.com/repos/denoland/deno
opened
Disable snapshots when cross compiling
build suggestion
When cross-compiling, snapshots are built for the host architecture instead of the target. There is no easy workaround for this. Unblocks our CI to produce the following targets: - `aarch64-android-linux` - `aarch64-unknown-linux-gnu`
1.0
Disable snapshots when cross compiling - When cross-compiling, snapshots are built for the host architecture instead of the target. There is no easy workaround for this. Unblocks our CI to produce the following targets: - `aarch64-android-linux` - `aarch64-unknown-linux-gnu`
non_main
disable snapshots when cross compiling when cross compiling snapshots are built for the host architecture instead of the target there is no easy workaround for this unblocks our ci to produce the following targets android linux unknown linux gnu
0
39,929
2,860,895,353
IssuesEvent
2015-06-03 17:59:43
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
rewrite CondorPlugIn for Condor 8 or upper version
Medium Priority New Feature ToDo WMAgent
take advantage of python bindings provided by Condor 8. Possibly improve the performance and reduce the memory usage.
1.0
rewrite CondorPlugIn for Condor 8 or upper version - take advantage of python bindings provided by Condor 8. Possibly improve the performance and reduce the memory usage.
non_main
rewrite condorplugin for condor or upper version take advantage of python bindings provided by condor possibly improve the performance and reduce the memory usage
0
4,896
25,139,034,818
IssuesEvent
2022-11-09 21:16:49
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Cell selection logic prevents user from modifying column types options
type: bug work: frontend status: ready restricted: maintainers
## Reproduce 1. Navigate to the Publications Table Page. 1. Select a cell within the Year column. 1. Open the Column tab within the Table Inspector. 1. Within Data Type, click Number to edit the type options. 1. Open the Digit Grouping select element. 1. Choose a new value. 1. Expect the select element to close but remain visible. 1. Instead, observe that the cell selection has been reset to empty, removing all the UI Column tab of the Table Inspector because the column is no longer selected. CC @rajatvijay
True
Cell selection logic prevents user from modifying column types options - ## Reproduce 1. Navigate to the Publications Table Page. 1. Select a cell within the Year column. 1. Open the Column tab within the Table Inspector. 1. Within Data Type, click Number to edit the type options. 1. Open the Digit Grouping select element. 1. Choose a new value. 1. Expect the select element to close but remain visible. 1. Instead, observe that the cell selection has been reset to empty, removing all the UI Column tab of the Table Inspector because the column is no longer selected. CC @rajatvijay
main
cell selection logic prevents user from modifying column types options reproduce navigate to the publications table page select a cell within the year column open the column tab within the table inspector within data type click number to edit the type options open the digit grouping select element choose a new value expect the select element to close but remain visible instead observe that the cell selection has been reset to empty removing all the ui column tab of the table inspector because the column is no longer selected cc rajatvijay
1
5,645
28,369,956,689
IssuesEvent
2023-04-12 16:13:59
deislabs/spiderlightning
https://api.github.com/repos/deislabs/spiderlightning
opened
can't install "sudo apt-add-repository ppa:mosquitto-dev/mosquitto-ppa" per `make install-deps"
🐛 bug 🚧 maintainer issue
**Description of the bug** on Ubuntu 20.04 WSL2, doing `make install-deps` to build (because installation fails due to https://github.com/deislabs/spiderlightning/issues/377 ) results in: ╭── ~/work/squillace/spiderlightning  main ✘✘✘ ✭ ╰────▶ sudo apt-add-repository ppa:mosquitto-dev/mosquitto-ppa More info: https://launchpad.net/~mosquitto-dev/+archive/ubuntu/mosquitto-ppa Press [ENTER] to continue or Ctrl-c to cancel adding it. Hit:1 https://packages.microsoft.com/repos/azure-cli focal InRelease Hit:2 https://packages.microsoft.com/ubuntu/20.04/prod focal InRelease Hit:3 http://archive.ubuntu.com/ubuntu focal InRelease Hit:4 http://archive.ubuntu.com/ubuntu focal-updates InRelease Hit:5 https://download.docker.com/linux/ubuntu focal InRelease Hit:6 http://archive.ubuntu.com/ubuntu focal-backports InRelease Hit:7 http://security.ubuntu.com/ubuntu focal-security InRelease Hit:8 https://download.mono-project.com/repo/ubuntu stable-focal InRelease Hit:9 http://ppa.launchpad.net/mosquitto-dev/mosquitto-ppa/ubuntu focal InRelease Get:10 https://packages.cloud.google.com/apt cloud-sdk InRelease [6361 B] Err:10 https://packages.cloud.google.com/apt cloud-sdk InRelease The following signatures couldn't be verified because the public key is not available: NO_PUBKEY B53DC80D13EDEF05 Hit:11 http://prerelease.keybase.io/deb stable InRelease Reading package lists... Done W: An error occurred during the signature verification. The repository is not updated and the previous index files will be used. GPG error: https://packages.cloud.google.com/apt cloud-sdk InRelease: The following signatures couldn't be verified because the public key is not available: NO_PUBKEY B53DC80D13EDEF05 W: Failed to fetch https://packages.cloud.google.com/apt/dists/cloud-sdk/InRelease The following signatures couldn't be verified because the public key is not available: NO_PUBKEY B53DC80D13EDEF05 W: Some index files failed to download. They have been ignored, or old ones used instead. **To Reproduce** sudo apt-add-repository ppa:mosquitto-dev/mosquitto-ppa on ubuntu 20.04 should do it **Additional context**
True
can't install "sudo apt-add-repository ppa:mosquitto-dev/mosquitto-ppa" per `make install-deps" - **Description of the bug** on Ubuntu 20.04 WSL2, doing `make install-deps` to build (because installation fails due to https://github.com/deislabs/spiderlightning/issues/377 ) results in: ╭── ~/work/squillace/spiderlightning  main ✘✘✘ ✭ ╰────▶ sudo apt-add-repository ppa:mosquitto-dev/mosquitto-ppa More info: https://launchpad.net/~mosquitto-dev/+archive/ubuntu/mosquitto-ppa Press [ENTER] to continue or Ctrl-c to cancel adding it. Hit:1 https://packages.microsoft.com/repos/azure-cli focal InRelease Hit:2 https://packages.microsoft.com/ubuntu/20.04/prod focal InRelease Hit:3 http://archive.ubuntu.com/ubuntu focal InRelease Hit:4 http://archive.ubuntu.com/ubuntu focal-updates InRelease Hit:5 https://download.docker.com/linux/ubuntu focal InRelease Hit:6 http://archive.ubuntu.com/ubuntu focal-backports InRelease Hit:7 http://security.ubuntu.com/ubuntu focal-security InRelease Hit:8 https://download.mono-project.com/repo/ubuntu stable-focal InRelease Hit:9 http://ppa.launchpad.net/mosquitto-dev/mosquitto-ppa/ubuntu focal InRelease Get:10 https://packages.cloud.google.com/apt cloud-sdk InRelease [6361 B] Err:10 https://packages.cloud.google.com/apt cloud-sdk InRelease The following signatures couldn't be verified because the public key is not available: NO_PUBKEY B53DC80D13EDEF05 Hit:11 http://prerelease.keybase.io/deb stable InRelease Reading package lists... Done W: An error occurred during the signature verification. The repository is not updated and the previous index files will be used. GPG error: https://packages.cloud.google.com/apt cloud-sdk InRelease: The following signatures couldn't be verified because the public key is not available: NO_PUBKEY B53DC80D13EDEF05 W: Failed to fetch https://packages.cloud.google.com/apt/dists/cloud-sdk/InRelease The following signatures couldn't be verified because the public key is not available: NO_PUBKEY B53DC80D13EDEF05 W: Some index files failed to download. They have been ignored, or old ones used instead. **To Reproduce** sudo apt-add-repository ppa:mosquitto-dev/mosquitto-ppa on ubuntu 20.04 should do it **Additional context**
main
can t install sudo apt add repository ppa mosquitto dev mosquitto ppa per make install deps description of the bug on ubuntu doing make install deps to build because installation fails due to results in ╭── work squillace spiderlightning  main ✘✘✘ ✭ ╰────▶ sudo apt add repository ppa mosquitto dev mosquitto ppa more info press to continue or ctrl c to cancel adding it hit focal inrelease hit focal inrelease hit focal inrelease hit focal updates inrelease hit focal inrelease hit focal backports inrelease hit focal security inrelease hit stable focal inrelease hit focal inrelease get cloud sdk inrelease err cloud sdk inrelease the following signatures couldn t be verified because the public key is not available no pubkey hit stable inrelease reading package lists done w an error occurred during the signature verification the repository is not updated and the previous index files will be used gpg error cloud sdk inrelease the following signatures couldn t be verified because the public key is not available no pubkey w failed to fetch the following signatures couldn t be verified because the public key is not available no pubkey w some index files failed to download they have been ignored or old ones used instead to reproduce sudo apt add repository ppa mosquitto dev mosquitto ppa on ubuntu should do it additional context
1
2,122
7,245,243,524
IssuesEvent
2018-02-14 17:26:08
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
opened
Unused parameters should be removed
Area: analyzer Area: maintainability feature
Methods that have parameters which are not used should be reworked and the unused parameters should be removed.
True
Unused parameters should be removed - Methods that have parameters which are not used should be reworked and the unused parameters should be removed.
main
unused parameters should be removed methods that have parameters which are not used should be reworked and the unused parameters should be removed
1
813,888
30,477,995,239
IssuesEvent
2023-07-17 18:00:00
npm/cli
https://api.github.com/repos/npm/cli
closed
[BUG] npx 'canceled' error reason lacks clarity
Enhancement Priority 2 Release 9.x
### Is there an existing issue for this? - [X] I have searched the existing issues ### This issue exists in the latest npm version - [X] I am using the latest npm ### Current Behavior Using npm version 9.5.0 running the command: ```bash npx --no-install esbuild --bundle "~/typescript-graphql-service/service/src/lambda.ts" --target=node18 --platform=node --outfile="~/typescript-graphql-service/AWS/cdk.out/bundling-temp-d2395ea5f5b16624b9255daf70671b29ee231644c99ea4f7718828f55c1a0f13/index.js" --external="@aws-sdk/*" --loader:.graphql=file ``` I get ```bash npm ERR! canceled npm ERR! A complete log of this run can be found in: npm ERR! /Users/lancegliser/.npm/_logs/2023-03-01T18_11_17_043Z-debug-0.log ``` ### Expected Behavior A more useful explanation of why it was cancelled. It was unclear what module was throwing an error or canceling at first. ```bash npm ERR! npx canceled due to missing packages and no YES option: ["esbuild@0.17.10"] npm ERR! A complete log of this run can be found in: npm ERR! /Users/lancegliser/.npm/_logs/2023-03-01T18_11_17_043Z-debug-0.log ``` ### Steps To Reproduce 1. In a general termal 2. Without `esbuild` in path 3. Run `npx --no-install esbuild --bundle` 4. See error... ### Environment - npm: 9.5.1 - Node.js: 18.13.0 - OS Name: MacOS 12.6.3 - System Model Name: Mackbook Pro - npm config: ```ini ; "user" config from /Users/lancegliser/.npmrc init-author-name = "Lance Gliser" ; node bin location = /usr/local/bin/node ; node version = v18.13.0 ; npm local prefix = /Users/lancegliser ; npm version = 9.5.1 ; cwd = /Users/lancegliser ; HOME = /Users/lancegliser ; Run `npm config ls -l` to show all defaults.```
1.0
[BUG] npx 'canceled' error reason lacks clarity - ### Is there an existing issue for this? - [X] I have searched the existing issues ### This issue exists in the latest npm version - [X] I am using the latest npm ### Current Behavior Using npm version 9.5.0 running the command: ```bash npx --no-install esbuild --bundle "~/typescript-graphql-service/service/src/lambda.ts" --target=node18 --platform=node --outfile="~/typescript-graphql-service/AWS/cdk.out/bundling-temp-d2395ea5f5b16624b9255daf70671b29ee231644c99ea4f7718828f55c1a0f13/index.js" --external="@aws-sdk/*" --loader:.graphql=file ``` I get ```bash npm ERR! canceled npm ERR! A complete log of this run can be found in: npm ERR! /Users/lancegliser/.npm/_logs/2023-03-01T18_11_17_043Z-debug-0.log ``` ### Expected Behavior A more useful explanation of why it was cancelled. It was unclear what module was throwing an error or canceling at first. ```bash npm ERR! npx canceled due to missing packages and no YES option: ["esbuild@0.17.10"] npm ERR! A complete log of this run can be found in: npm ERR! /Users/lancegliser/.npm/_logs/2023-03-01T18_11_17_043Z-debug-0.log ``` ### Steps To Reproduce 1. In a general termal 2. Without `esbuild` in path 3. Run `npx --no-install esbuild --bundle` 4. See error... ### Environment - npm: 9.5.1 - Node.js: 18.13.0 - OS Name: MacOS 12.6.3 - System Model Name: Mackbook Pro - npm config: ```ini ; "user" config from /Users/lancegliser/.npmrc init-author-name = "Lance Gliser" ; node bin location = /usr/local/bin/node ; node version = v18.13.0 ; npm local prefix = /Users/lancegliser ; npm version = 9.5.1 ; cwd = /Users/lancegliser ; HOME = /Users/lancegliser ; Run `npm config ls -l` to show all defaults.```
non_main
npx canceled error reason lacks clarity is there an existing issue for this i have searched the existing issues this issue exists in the latest npm version i am using the latest npm current behavior using npm version running the command bash npx no install esbuild bundle typescript graphql service service src lambda ts target platform node outfile typescript graphql service aws cdk out bundling temp index js external aws sdk loader graphql file i get bash npm err canceled npm err a complete log of this run can be found in npm err users lancegliser npm logs debug log expected behavior a more useful explanation of why it was cancelled it was unclear what module was throwing an error or canceling at first bash npm err npx canceled due to missing packages and no yes option npm err a complete log of this run can be found in npm err users lancegliser npm logs debug log steps to reproduce in a general termal without esbuild in path run npx no install esbuild bundle see error environment npm node js os name macos system model name mackbook pro npm config ini user config from users lancegliser npmrc init author name lance gliser node bin location usr local bin node node version npm local prefix users lancegliser npm version cwd users lancegliser home users lancegliser run npm config ls l to show all defaults
0
1,859
6,577,407,957
IssuesEvent
2017-09-12 00:42:13
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
lineinfile: regexp="...\b..." no longer matches in v2
affects_2.0 bug_report waiting_on_maintainer
##### ISSUE TYPE Bug Report ##### COMPONENT NAME lineinfile module ##### ANSIBLE VERSION 2.0.0.2 ##### SUMMARY This is a re-submit of bug #14449 by user nichovo which was filed against ansible-core, but should be filed here. Given this line in /etc/ssh/sshd_config: # AddressFamily any this task in Ansible v1.9.4: lineinfile: dest=/etc/ssh/sshd_config regexp='^[ \t#]_AddressFamily\b._' line='AddressFamily inet' would replace the line as expected with: AddressFamily inet In v2.0.0.2, the regex doesn't match and the line is instead added to the bottom of the file. However, escaping the \b and using \b in the regex works! Backslashes for other character classes like \w don't need to be escaped - this regex matches: '^[ \t#]*AddressFamily \w+' Is it a "bug" or a "feature" that only \b now needs to be \b in Ansible v2?
True
lineinfile: regexp="...\b..." no longer matches in v2 - ##### ISSUE TYPE Bug Report ##### COMPONENT NAME lineinfile module ##### ANSIBLE VERSION 2.0.0.2 ##### SUMMARY This is a re-submit of bug #14449 by user nichovo which was filed against ansible-core, but should be filed here. Given this line in /etc/ssh/sshd_config: # AddressFamily any this task in Ansible v1.9.4: lineinfile: dest=/etc/ssh/sshd_config regexp='^[ \t#]_AddressFamily\b._' line='AddressFamily inet' would replace the line as expected with: AddressFamily inet In v2.0.0.2, the regex doesn't match and the line is instead added to the bottom of the file. However, escaping the \b and using \b in the regex works! Backslashes for other character classes like \w don't need to be escaped - this regex matches: '^[ \t#]*AddressFamily \w+' Is it a "bug" or a "feature" that only \b now needs to be \b in Ansible v2?
main
lineinfile regexp b no longer matches in issue type bug report component name lineinfile module ansible version summary this is a re submit of bug by user nichovo which was filed against ansible core but should be filed here given this line in etc ssh sshd config addressfamily any this task in ansible lineinfile dest etc ssh sshd config regexp addressfamily b line addressfamily inet would replace the line as expected with addressfamily inet in the regex doesn t match and the line is instead added to the bottom of the file however escaping the b and using b in the regex works backslashes for other character classes like w don t need to be escaped this regex matches addressfamily w is it a bug or a feature that only b now needs to be b in ansible
1
5,817
30,792,519,056
IssuesEvent
2023-07-31 17:14:34
jupyter-naas/awesome-notebooks
https://api.github.com/repos/jupyter-naas/awesome-notebooks
closed
JSON - Pretty Print JSON Data
templates maintainer
This notebook will show how to pretty print JSON data and how it is useful for organizations. It will explain how to parse a JSON string and display it in a readable format.
True
JSON - Pretty Print JSON Data - This notebook will show how to pretty print JSON data and how it is useful for organizations. It will explain how to parse a JSON string and display it in a readable format.
main
json pretty print json data this notebook will show how to pretty print json data and how it is useful for organizations it will explain how to parse a json string and display it in a readable format
1
26,107
11,268,134,049
IssuesEvent
2020-01-14 05:01:59
istio/istio
https://api.github.com/repos/istio/istio
closed
Path based basic auth
area/security kind/enhancement lifecycle/needs-triage lifecycle/stale
**Describe the feature request** I would expect that istio supports basic authentication for routing. E.g. like this: ```yaml apiVersion: "authentication.istio.io/v1alpha1" kind: "Policy" metadata: name: "jwt-example" spec: targets: - name: httpbin origins: - basic: secret: "k8s-basic-auth-secret" trigger_rules: - included_paths: - exact: /documentation principalBinding: USE_ORIGIN ``` or ```yaml apiVersion: "authentication.istio.io/v1alpha1" kind: "Policy" metadata: name: "jwt-example" spec: targets: - name: httpbin origins: - basic: users: - "myuser:mypassword" trigger_rules: - included_paths: - exact: /documentation principalBinding: USE_ORIGIN ``` **Describe alternatives you've considered** I thought about adding another NGINX proxy in front of the the service, but I do not really want to decrease the performance of each HTTP request by adding another layer. **Affected product area (please put an X in all that apply)** [ ] Configuration Infrastructure [ ] Docs [ ] Installation [ ] Networking [ ] Performance and Scalability [ ] Policies and Telemetry [x] Security [ ] Test and Release [ ] User Experience [ ] Developer Infrastructure **Additional context**
True
Path based basic auth - **Describe the feature request** I would expect that istio supports basic authentication for routing. E.g. like this: ```yaml apiVersion: "authentication.istio.io/v1alpha1" kind: "Policy" metadata: name: "jwt-example" spec: targets: - name: httpbin origins: - basic: secret: "k8s-basic-auth-secret" trigger_rules: - included_paths: - exact: /documentation principalBinding: USE_ORIGIN ``` or ```yaml apiVersion: "authentication.istio.io/v1alpha1" kind: "Policy" metadata: name: "jwt-example" spec: targets: - name: httpbin origins: - basic: users: - "myuser:mypassword" trigger_rules: - included_paths: - exact: /documentation principalBinding: USE_ORIGIN ``` **Describe alternatives you've considered** I thought about adding another NGINX proxy in front of the the service, but I do not really want to decrease the performance of each HTTP request by adding another layer. **Affected product area (please put an X in all that apply)** [ ] Configuration Infrastructure [ ] Docs [ ] Installation [ ] Networking [ ] Performance and Scalability [ ] Policies and Telemetry [x] Security [ ] Test and Release [ ] User Experience [ ] Developer Infrastructure **Additional context**
non_main
path based basic auth describe the feature request i would expect that istio supports basic authentication for routing e g like this yaml apiversion authentication istio io kind policy metadata name jwt example spec targets name httpbin origins basic secret basic auth secret trigger rules included paths exact documentation principalbinding use origin or yaml apiversion authentication istio io kind policy metadata name jwt example spec targets name httpbin origins basic users myuser mypassword trigger rules included paths exact documentation principalbinding use origin describe alternatives you ve considered i thought about adding another nginx proxy in front of the the service but i do not really want to decrease the performance of each http request by adding another layer affected product area please put an x in all that apply configuration infrastructure docs installation networking performance and scalability policies and telemetry security test and release user experience developer infrastructure additional context
0
146,790
23,121,391,635
IssuesEvent
2022-07-27 21:58:32
clap-rs/clap
https://api.github.com/repos/clap-rs/clap
closed
Add an option to show aliases to `--help`
C-enhancement A-help S-waiting-on-design
### Please complete the following tasks - [X] I have searched the [discussions](https://github.com/clap-rs/clap/discussions) - [X] I have searched the [open](https://github.com/clap-rs/clap/issues) and [rejected](https://github.com/clap-rs/clap/issues?q=is%3Aissue+label%3AS-wont-fix+is%3Aclosed) issues ### Clap Version 3.0.10 ### Describe your use case Currently the developer either specifies `alias` or `visible_alias` so that aliases for commands are either always shown or always hidden in the help. On one hand, the developers want users to be able to discover shortcuts for the commands from help message, but on the other hand having aliases for each command clutters the screen and makes it less readable. Example with aliases: ``` to-dec Convert hex value into a decimal number. [aliases: td, 2d] to-fix Convert an integer into a fixed point number. [aliases: tf, 2f] to-hex Convert an integer to hex. [aliases: th, 2h] to-hexdata Normalize the input to lowercase, 0x-prefixed hex. [aliases: thd, 2hd] to-int256 Convert a number to a hex-encoded int256. [aliases: ti, 2i] to-rlp RLP encodes hex data, or an array of hex data to-uint256 Convert a number to a hex-encoded uint256. [aliases: tu, 2u] ``` Without: ``` to-dec Convert hex value into a decimal number. to-fix Convert an integer into a fixed point number. to-hex Convert an integer to hex. to-hexdata Normalize the input to lowercase, 0x-prefixed hex. to-int256 Convert a number to a hex-encoded int256. to-rlp RLP encodes hex data, or an array of hex data to-uint256 Convert a number to a hex-encoded uint256. ``` ### Describe the solution you'd like Add an additional argument to `--help` that prints help as if all `alias` items were `visible_alias` ones. This way aliases would stay hidden by default to avoid screen clutter, but there would still be a way to discover them. Not sure how to better implement it ### Alternatives, if applicable _No response_ ### Additional Context _No response_
1.0
Add an option to show aliases to `--help` - ### Please complete the following tasks - [X] I have searched the [discussions](https://github.com/clap-rs/clap/discussions) - [X] I have searched the [open](https://github.com/clap-rs/clap/issues) and [rejected](https://github.com/clap-rs/clap/issues?q=is%3Aissue+label%3AS-wont-fix+is%3Aclosed) issues ### Clap Version 3.0.10 ### Describe your use case Currently the developer either specifies `alias` or `visible_alias` so that aliases for commands are either always shown or always hidden in the help. On one hand, the developers want users to be able to discover shortcuts for the commands from help message, but on the other hand having aliases for each command clutters the screen and makes it less readable. Example with aliases: ``` to-dec Convert hex value into a decimal number. [aliases: td, 2d] to-fix Convert an integer into a fixed point number. [aliases: tf, 2f] to-hex Convert an integer to hex. [aliases: th, 2h] to-hexdata Normalize the input to lowercase, 0x-prefixed hex. [aliases: thd, 2hd] to-int256 Convert a number to a hex-encoded int256. [aliases: ti, 2i] to-rlp RLP encodes hex data, or an array of hex data to-uint256 Convert a number to a hex-encoded uint256. [aliases: tu, 2u] ``` Without: ``` to-dec Convert hex value into a decimal number. to-fix Convert an integer into a fixed point number. to-hex Convert an integer to hex. to-hexdata Normalize the input to lowercase, 0x-prefixed hex. to-int256 Convert a number to a hex-encoded int256. to-rlp RLP encodes hex data, or an array of hex data to-uint256 Convert a number to a hex-encoded uint256. ``` ### Describe the solution you'd like Add an additional argument to `--help` that prints help as if all `alias` items were `visible_alias` ones. This way aliases would stay hidden by default to avoid screen clutter, but there would still be a way to discover them. Not sure how to better implement it ### Alternatives, if applicable _No response_ ### Additional Context _No response_
non_main
add an option to show aliases to help please complete the following tasks i have searched the i have searched the and issues clap version describe your use case currently the developer either specifies alias or visible alias so that aliases for commands are either always shown or always hidden in the help on one hand the developers want users to be able to discover shortcuts for the commands from help message but on the other hand having aliases for each command clutters the screen and makes it less readable example with aliases to dec convert hex value into a decimal number to fix convert an integer into a fixed point number to hex convert an integer to hex to hexdata normalize the input to lowercase prefixed hex to convert a number to a hex encoded to rlp rlp encodes hex data or an array of hex data to convert a number to a hex encoded without to dec convert hex value into a decimal number to fix convert an integer into a fixed point number to hex convert an integer to hex to hexdata normalize the input to lowercase prefixed hex to convert a number to a hex encoded to rlp rlp encodes hex data or an array of hex data to convert a number to a hex encoded describe the solution you d like add an additional argument to help that prints help as if all alias items were visible alias ones this way aliases would stay hidden by default to avoid screen clutter but there would still be a way to discover them not sure how to better implement it alternatives if applicable no response additional context no response
0
322,728
23,920,639,975
IssuesEvent
2022-09-09 16:30:43
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Crash when opening docs for a second time on Fit Script Generator
Bug Documentation ISIS Team: Spectroscopy
**Describe the bug** Opening the documentation page from the `Fit Script Generator` for a second time, after closing and re-opening the interface, causes a crash. **To Reproduce** 1. Open `General`->`Fit Script Generator` 2. Click `?` and the documentation should open 3. Close the interface 4. Open the same interface and click `?` again. Crash! **Expected behavior** The documentation should open fine **Platform/Version (please complete the following information):** - Windows, Redhat, and possibly others
1.0
Crash when opening docs for a second time on Fit Script Generator - **Describe the bug** Opening the documentation page from the `Fit Script Generator` for a second time, after closing and re-opening the interface, causes a crash. **To Reproduce** 1. Open `General`->`Fit Script Generator` 2. Click `?` and the documentation should open 3. Close the interface 4. Open the same interface and click `?` again. Crash! **Expected behavior** The documentation should open fine **Platform/Version (please complete the following information):** - Windows, Redhat, and possibly others
non_main
crash when opening docs for a second time on fit script generator describe the bug opening the documentation page from the fit script generator for a second time after closing and re opening the interface causes a crash to reproduce open general fit script generator click and the documentation should open close the interface open the same interface and click again crash expected behavior the documentation should open fine platform version please complete the following information windows redhat and possibly others
0
800
4,417,453,085
IssuesEvent
2016-08-15 05:26:00
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
cron module erases other cronjobs of user root and puts cronjobs of other users in /var/cron/tabs/root
bug_report waiting_on_maintainer
This issue is not found yet in bug tracker ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME cron module ##### ANSIBLE VERSION ansible 2.0.1.0 config file = /usr/local/etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### CONFIGURATION forks = 200 ssh_args = -o ControlMaster=auto -o ControlPersist=60s pipelining = True ##### OS / ENVIRONMENT FreeBSD 10.x ##### SUMMARY cron module erases other cronjobs of definite user root and puts cronjobs of other users in /var/cron/tabs/root. ##### STEPS TO REPRODUCE This problem meets at 2.0.1.0 only. Example: ``` $ less unixadm/roles/host-monitoring/tasks/main.yml ... - cron: name="raidstat" minute="*/15" user=root job="/usr/local/sbin/raidstat -w >/dev/null 2>&1" tags: host-monitoring become: yes - cron: name="check snmpd" minute="*/5" user=root job="/usr/local/sbin/check_snmpd >/dev/null 2>&1" tags: host-monitoring become: yes - cron: name="tcp_states" minute="*/5" user=snmp job="/usr/local/sbin/tcp_states > /tmp/tcp_states" tags: host-monitoring become: yes ... playbook: $ ansible-playbook -i hosts_store unixadm/store.yml --tags='common,host-monitoring' -l 'store6*' ``` ##### EXPECTED RESULTS ``` # crontab -l #Ansible: freebsd-update 0 14 * * * /usr/sbin/freebsd-update -f /usr/local/etc/freebsd-update.conf -t /dev/null cron #Ansible: raidstat */15 * * * * /usr/local/sbin/raidstat -w >/dev/null 2>&1 #Ansible: check snmpd */5 * * * * /usr/local/sbin/check_snmpd >/dev/null 2>&1 # crontab -l -u snmp #Ansible: tcp_states */5 * * * * /usr/local/sbin/tcp_states > /tmp/tcp_states ``` ##### ACTUAL RESULTS ``` # crontab -l -u root #Ansible: tcp_states */5 * * * * /usr/local/sbin/tcp_states > /tmp/tcp_states ``` We analized code of cron.py, making diffs with 2.0.0.2 version: ``` --- ansible-2.0.0.2/lib/ansible/modules/core/system/cron.py 2016-01-15 01:33:30.000000000 +0300 +++ ansible-2.0.1.0/lib/ansible/modules/core/system/cron.py 2016-02-25 06:00:58.000000000 +0300 @@ -45,7 +45,7 @@ options: name: description: - - Description of a crontab entry. + - Description of a crontab entry. Required if state=absent default: null required: false user: @@ -383,7 +383,7 @@ return "chown %s %s ; su '%s' -c '%s %s'" % (pipes.quote(self.user), pipes.quote(path), pipes.quote(self.user), CRONCMD, pipes.quote(path)) else: user = '-u %s' % pipes.quote(self.user) - return "%s %s %s" % (CRONCMD , user, pipes.quote(path)) + return "%s %s %s" % (CRONCMD , pipes.quote(path), user) ``` Then, we changed one line in new cron.py, and now it works! Please fix this ASAP.
True
cron module erases other cronjobs of user root and puts cronjobs of other users in /var/cron/tabs/root - This issue is not found yet in bug tracker ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME cron module ##### ANSIBLE VERSION ansible 2.0.1.0 config file = /usr/local/etc/ansible/ansible.cfg configured module search path = Default w/o overrides ##### CONFIGURATION forks = 200 ssh_args = -o ControlMaster=auto -o ControlPersist=60s pipelining = True ##### OS / ENVIRONMENT FreeBSD 10.x ##### SUMMARY cron module erases other cronjobs of definite user root and puts cronjobs of other users in /var/cron/tabs/root. ##### STEPS TO REPRODUCE This problem meets at 2.0.1.0 only. Example: ``` $ less unixadm/roles/host-monitoring/tasks/main.yml ... - cron: name="raidstat" minute="*/15" user=root job="/usr/local/sbin/raidstat -w >/dev/null 2>&1" tags: host-monitoring become: yes - cron: name="check snmpd" minute="*/5" user=root job="/usr/local/sbin/check_snmpd >/dev/null 2>&1" tags: host-monitoring become: yes - cron: name="tcp_states" minute="*/5" user=snmp job="/usr/local/sbin/tcp_states > /tmp/tcp_states" tags: host-monitoring become: yes ... playbook: $ ansible-playbook -i hosts_store unixadm/store.yml --tags='common,host-monitoring' -l 'store6*' ``` ##### EXPECTED RESULTS ``` # crontab -l #Ansible: freebsd-update 0 14 * * * /usr/sbin/freebsd-update -f /usr/local/etc/freebsd-update.conf -t /dev/null cron #Ansible: raidstat */15 * * * * /usr/local/sbin/raidstat -w >/dev/null 2>&1 #Ansible: check snmpd */5 * * * * /usr/local/sbin/check_snmpd >/dev/null 2>&1 # crontab -l -u snmp #Ansible: tcp_states */5 * * * * /usr/local/sbin/tcp_states > /tmp/tcp_states ``` ##### ACTUAL RESULTS ``` # crontab -l -u root #Ansible: tcp_states */5 * * * * /usr/local/sbin/tcp_states > /tmp/tcp_states ``` We analized code of cron.py, making diffs with 2.0.0.2 version: ``` --- ansible-2.0.0.2/lib/ansible/modules/core/system/cron.py 2016-01-15 01:33:30.000000000 +0300 +++ ansible-2.0.1.0/lib/ansible/modules/core/system/cron.py 2016-02-25 06:00:58.000000000 +0300 @@ -45,7 +45,7 @@ options: name: description: - - Description of a crontab entry. + - Description of a crontab entry. Required if state=absent default: null required: false user: @@ -383,7 +383,7 @@ return "chown %s %s ; su '%s' -c '%s %s'" % (pipes.quote(self.user), pipes.quote(path), pipes.quote(self.user), CRONCMD, pipes.quote(path)) else: user = '-u %s' % pipes.quote(self.user) - return "%s %s %s" % (CRONCMD , user, pipes.quote(path)) + return "%s %s %s" % (CRONCMD , pipes.quote(path), user) ``` Then, we changed one line in new cron.py, and now it works! Please fix this ASAP.
main
cron module erases other cronjobs of user root and puts cronjobs of other users in var cron tabs root this issue is not found yet in bug tracker issue type bug report component name cron module ansible version ansible config file usr local etc ansible ansible cfg configured module search path default w o overrides configuration forks ssh args o controlmaster auto o controlpersist pipelining true os environment freebsd x summary cron module erases other cronjobs of definite user root and puts cronjobs of other users in var cron tabs root steps to reproduce this problem meets at only example less unixadm roles host monitoring tasks main yml cron name raidstat minute user root job usr local sbin raidstat w dev null tags host monitoring become yes cron name check snmpd minute user root job usr local sbin check snmpd dev null tags host monitoring become yes cron name tcp states minute user snmp job usr local sbin tcp states tmp tcp states tags host monitoring become yes playbook ansible playbook i hosts store unixadm store yml tags common host monitoring l expected results crontab l ansible freebsd update usr sbin freebsd update f usr local etc freebsd update conf t dev null cron ansible raidstat usr local sbin raidstat w dev null ansible check snmpd usr local sbin check snmpd dev null crontab l u snmp ansible tcp states usr local sbin tcp states tmp tcp states actual results crontab l u root ansible tcp states usr local sbin tcp states tmp tcp states we analized code of cron py making diffs with version ansible lib ansible modules core system cron py ansible lib ansible modules core system cron py options name description description of a crontab entry description of a crontab entry required if state absent default null required false user return chown s s su s c s s pipes quote self user pipes quote path pipes quote self user croncmd pipes quote path else user u s pipes quote self user return s s s croncmd user pipes quote path return s s s croncmd pipes quote path user then we changed one line in new cron py and now it works please fix this asap
1
698,508
23,982,997,562
IssuesEvent
2022-09-13 16:29:57
bcgov/entity
https://api.github.com/repos/bcgov/entity
closed
Backend/Filer: business founding date mismatch between firms and other entity types
bug Priority1 ENTITY
#### New description Create UI is saving Start Date as `yyyy-mm-dd`, as expected. The Filer should take this and add the filing time, so that the eventual Founding Date is `yyyy-mm-ddThh:mm:ss` (in UTC). (Same thing with Dissolution Date -- please check if this is incorrect as well.) (In the db, for some firms, Founding Date is `yyyy-mm-dd 00:00:00+00`, which is incorrect. If possible, please fix these in Dev db; they should include the UTC offset so that the date remains correct after Pacific time conversion.) #### Old description For a benefit company, the Founding Date property in the business response is an actual UTC date-time, eg: ![image.png](https://images.zenhubusercontent.com/5d0a7edda4644173e93bf808/8bda3295-5f41-46a3-a16c-822cbd89b5ed) However, for a firm, the Founding Date is actually just a date in Pacific timezone (ignore the zero time), eg: ![image.png](https://images.zenhubusercontent.com/5d0a7edda4644173e93bf808/12490245-a5d1-4331-86d3-f25d3e321f91) **Proposed to do:** - [ ] change `foundingDate` to a date-only (in Pacific timezone) for the entity types that this applies to - [ ] rename `foundingDate` to `foundingDateTime` (or Founding Timestamp or something else that indicates a time is present, and which will be interpreted as a UTC datetime) for the entity types that this applies to - [ ] or some other design such that the UI does not have to handle the same property in different ways depending on the entity type Note that some of the changes above will impact Filings UI and possibly Create UI, since they expect to see the foundingDate property. PS - Also look at this property for other entity types (eg, Coop).
1.0
Backend/Filer: business founding date mismatch between firms and other entity types - #### New description Create UI is saving Start Date as `yyyy-mm-dd`, as expected. The Filer should take this and add the filing time, so that the eventual Founding Date is `yyyy-mm-ddThh:mm:ss` (in UTC). (Same thing with Dissolution Date -- please check if this is incorrect as well.) (In the db, for some firms, Founding Date is `yyyy-mm-dd 00:00:00+00`, which is incorrect. If possible, please fix these in Dev db; they should include the UTC offset so that the date remains correct after Pacific time conversion.) #### Old description For a benefit company, the Founding Date property in the business response is an actual UTC date-time, eg: ![image.png](https://images.zenhubusercontent.com/5d0a7edda4644173e93bf808/8bda3295-5f41-46a3-a16c-822cbd89b5ed) However, for a firm, the Founding Date is actually just a date in Pacific timezone (ignore the zero time), eg: ![image.png](https://images.zenhubusercontent.com/5d0a7edda4644173e93bf808/12490245-a5d1-4331-86d3-f25d3e321f91) **Proposed to do:** - [ ] change `foundingDate` to a date-only (in Pacific timezone) for the entity types that this applies to - [ ] rename `foundingDate` to `foundingDateTime` (or Founding Timestamp or something else that indicates a time is present, and which will be interpreted as a UTC datetime) for the entity types that this applies to - [ ] or some other design such that the UI does not have to handle the same property in different ways depending on the entity type Note that some of the changes above will impact Filings UI and possibly Create UI, since they expect to see the foundingDate property. PS - Also look at this property for other entity types (eg, Coop).
non_main
backend filer business founding date mismatch between firms and other entity types new description create ui is saving start date as yyyy mm dd as expected the filer should take this and add the filing time so that the eventual founding date is yyyy mm ddthh mm ss in utc same thing with dissolution date please check if this is incorrect as well in the db for some firms founding date is yyyy mm dd which is incorrect if possible please fix these in dev db they should include the utc offset so that the date remains correct after pacific time conversion old description for a benefit company the founding date property in the business response is an actual utc date time eg however for a firm the founding date is actually just a date in pacific timezone ignore the zero time eg proposed to do change foundingdate to a date only in pacific timezone for the entity types that this applies to rename foundingdate to foundingdatetime or founding timestamp or something else that indicates a time is present and which will be interpreted as a utc datetime for the entity types that this applies to or some other design such that the ui does not have to handle the same property in different ways depending on the entity type note that some of the changes above will impact filings ui and possibly create ui since they expect to see the foundingdate property ps also look at this property for other entity types eg coop
0
783
4,387,451,399
IssuesEvent
2016-08-08 15:49:44
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
zfs module doesn't support management of OpenZFS pools
bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME zfs ##### ANSIBLE VERSION ansible 2.1.0.0 ##### CONFIGURATION Nothing ##### OS / ENVIRONMENT Freshly installed latest version of OpenZFS on OpenIndiana. This also affects every illumos-based operating systems (SmartOS, OmniOS etc) ##### SUMMARY The problem is with zpool version. OpenZFS uses feature flags and sets zpool version to "-", which breaks compatibility with this module. It fails when it tries to compare OpenZFS zpool version with zpool version 34 on illumos-based system (https://github.com/ansible/ansible-modules-extras/blob/2a0c5e2a8fd7ed3ce6d6eedd08e85e01e1617113/system/zfs.py#L99) This module should check for Solaris presence not by os.uname, but by ansible facts if possible. ##### STEPS TO REPRODUCE ansible all -m zfs -a 'name=rpool/data state=present' -vvv ##### EXPECTED RESULTS ZFS dataset rpool/data should be created. ##### ACTUAL RESULTS rpool/data wasn't created. <!--- Paste verbatim command output between quotes below --> ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 245, in <module> main() File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 227, in main zfs = Zfs(module, name, properties) File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 90, in __init__ self.enhanced_sharing = self.check_enhanced_sharing() File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 99, in check_enhanced_sharing if int(version) >= 34: ValueError: invalid literal for int() with base 10: '-' ```
True
zfs module doesn't support management of OpenZFS pools - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME zfs ##### ANSIBLE VERSION ansible 2.1.0.0 ##### CONFIGURATION Nothing ##### OS / ENVIRONMENT Freshly installed latest version of OpenZFS on OpenIndiana. This also affects every illumos-based operating systems (SmartOS, OmniOS etc) ##### SUMMARY The problem is with zpool version. OpenZFS uses feature flags and sets zpool version to "-", which breaks compatibility with this module. It fails when it tries to compare OpenZFS zpool version with zpool version 34 on illumos-based system (https://github.com/ansible/ansible-modules-extras/blob/2a0c5e2a8fd7ed3ce6d6eedd08e85e01e1617113/system/zfs.py#L99) This module should check for Solaris presence not by os.uname, but by ansible facts if possible. ##### STEPS TO REPRODUCE ansible all -m zfs -a 'name=rpool/data state=present' -vvv ##### EXPECTED RESULTS ZFS dataset rpool/data should be created. ##### ACTUAL RESULTS rpool/data wasn't created. <!--- Paste verbatim command output between quotes below --> ``` An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 245, in <module> main() File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 227, in main zfs = Zfs(module, name, properties) File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 90, in __init__ self.enhanced_sharing = self.check_enhanced_sharing() File "/tmp/ansible_Y3bKol/ansible_module_zfs.py", line 99, in check_enhanced_sharing if int(version) >= 34: ValueError: invalid literal for int() with base 10: '-' ```
main
zfs module doesn t support management of openzfs pools issue type bug report component name zfs ansible version ansible configuration nothing os environment freshly installed latest version of openzfs on openindiana this also affects every illumos based operating systems smartos omnios etc summary the problem is with zpool version openzfs uses feature flags and sets zpool version to which breaks compatibility with this module it fails when it tries to compare openzfs zpool version with zpool version on illumos based system this module should check for solaris presence not by os uname but by ansible facts if possible steps to reproduce ansible all m zfs a name rpool data state present vvv expected results zfs dataset rpool data should be created actual results rpool data wasn t created an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible ansible module zfs py line in main file tmp ansible ansible module zfs py line in main zfs zfs module name properties file tmp ansible ansible module zfs py line in init self enhanced sharing self check enhanced sharing file tmp ansible ansible module zfs py line in check enhanced sharing if int version valueerror invalid literal for int with base
1
169,079
26,744,998,773
IssuesEvent
2023-01-30 15:26:05
linagora/Twake-Mobile
https://api.github.com/repos/linagora/Twake-Mobile
closed
[Story] Workspace limitation
story to design mobile
## User story summary [Definition](#Definition) [Business case](#Business-case) [Criteria](#criteria) [UI/UX Design](#UI-UX-design) [Notes](#notes) [Related](#related) ## Definition ```gherkin Given I am a user in a free company, When I want to create a second workspace in my company Then I can see a popup saying that I have to upgrade to create more than 1 workspace ``` ## Criteria - [ ] We can't create more than 1 workspace in a free company - [ ] Link of "learn more" redirects to the pricing page if I am not the company owner - [ ] Link of "learn more" redirects to the plan page of the console if I am the company owner ## UI UX Design ![workspaceL.png](https://images.zenhubusercontent.com/5f06c5a0e15425885ef10c4a/351d9922-198a-46e9-b1b4-cd181459d0d9) ## Notes ## Related Epic : #872 Web PR related : https://github.com/linagora/Twake/pull/1339
1.0
[Story] Workspace limitation - ## User story summary [Definition](#Definition) [Business case](#Business-case) [Criteria](#criteria) [UI/UX Design](#UI-UX-design) [Notes](#notes) [Related](#related) ## Definition ```gherkin Given I am a user in a free company, When I want to create a second workspace in my company Then I can see a popup saying that I have to upgrade to create more than 1 workspace ``` ## Criteria - [ ] We can't create more than 1 workspace in a free company - [ ] Link of "learn more" redirects to the pricing page if I am not the company owner - [ ] Link of "learn more" redirects to the plan page of the console if I am the company owner ## UI UX Design ![workspaceL.png](https://images.zenhubusercontent.com/5f06c5a0e15425885ef10c4a/351d9922-198a-46e9-b1b4-cd181459d0d9) ## Notes ## Related Epic : #872 Web PR related : https://github.com/linagora/Twake/pull/1339
non_main
workspace limitation user story summary definition business case criteria ui ux design notes related definition gherkin given i am a user in a free company when i want to create a second workspace in my company then i can see a popup saying that i have to upgrade to create more than workspace criteria we can t create more than workspace in a free company link of learn more redirects to the pricing page if i am not the company owner link of learn more redirects to the plan page of the console if i am the company owner ui ux design notes related epic web pr related
0
56,121
14,938,204,405
IssuesEvent
2021-01-25 15:30:35
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
508-defect-2 [COGNITION, SCREEN READER]: Buttons SHOULD have unique and accessible labels on the Military Service History page
508-defect-2 508-issue-cognition 508-issue-screenreader 508/Accessibility BDD vsa-benefits
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. <hr/> ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Josh_ ## User Story or Problem Statement As a screen reader user, I want to hear an accessible label like "Edit issues eligible for review" instead of "Edit" when the button has focus or virtual cursor. ## Details There are several issues with this page which may ultimately be unsolvable from a remediation standpoint. I would recommend revisiting this page from a design lens. Issues are as follows: - Edit buttons do not have unique and accessible labels. As such, they all currently announce "Edit" which doesn't tell the user _what_ they are editing. - Inconsistency exists in legends when adding new military service as only the _last_ item in the stack has the legend of "New Service Period" despite the possibility of other _new_ service periods being added previously in the same session. - The legend "New Service Period" does not change to "Edit Service Period" upon editing again _after_ saving which may confuse the user. This will require a larger conversation or meeting to discuss solutions, so I'll only be documenting issues as opposed to recommendations. For more details, view the screenshots attached on this ticket. ## Acceptance Criteria - [ ] TBD ## Solution TBD ## WCAG or Vendor Guidance (optional) * [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html) ## Screenshots or Trace Logs <img width="1792" alt="Screen Shot 2021-01-25 at 10 00 31 AM" src="https://user-images.githubusercontent.com/14154792/105726517-29231f00-5ef8-11eb-8878-f42e2336db1d.png"> <img width="1792" alt="Screen Shot 2021-01-25 at 10 09 47 AM" src="https://user-images.githubusercontent.com/14154792/105726526-2aece280-5ef8-11eb-9d17-c3996b923d58.png"> <img width="1337" alt="Screen Shot 2021-01-25 at 10 14 12 AM" src="https://user-images.githubusercontent.com/14154792/105726532-2c1e0f80-5ef8-11eb-86bb-272bb7782613.png">
1.0
508-defect-2 [COGNITION, SCREEN READER]: Buttons SHOULD have unique and accessible labels on the Military Service History page - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. <hr/> ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Josh_ ## User Story or Problem Statement As a screen reader user, I want to hear an accessible label like "Edit issues eligible for review" instead of "Edit" when the button has focus or virtual cursor. ## Details There are several issues with this page which may ultimately be unsolvable from a remediation standpoint. I would recommend revisiting this page from a design lens. Issues are as follows: - Edit buttons do not have unique and accessible labels. As such, they all currently announce "Edit" which doesn't tell the user _what_ they are editing. - Inconsistency exists in legends when adding new military service as only the _last_ item in the stack has the legend of "New Service Period" despite the possibility of other _new_ service periods being added previously in the same session. - The legend "New Service Period" does not change to "Edit Service Period" upon editing again _after_ saving which may confuse the user. This will require a larger conversation or meeting to discuss solutions, so I'll only be documenting issues as opposed to recommendations. For more details, view the screenshots attached on this ticket. ## Acceptance Criteria - [ ] TBD ## Solution TBD ## WCAG or Vendor Guidance (optional) * [Info and Relationships: Understanding SC 1.3.1](https://www.w3.org/TR/UNDERSTANDING-WCAG20/content-structure-separation-programmatic.html) ## Screenshots or Trace Logs <img width="1792" alt="Screen Shot 2021-01-25 at 10 00 31 AM" src="https://user-images.githubusercontent.com/14154792/105726517-29231f00-5ef8-11eb-8878-f42e2336db1d.png"> <img width="1792" alt="Screen Shot 2021-01-25 at 10 09 47 AM" src="https://user-images.githubusercontent.com/14154792/105726526-2aece280-5ef8-11eb-9d17-c3996b923d58.png"> <img width="1337" alt="Screen Shot 2021-01-25 at 10 14 12 AM" src="https://user-images.githubusercontent.com/14154792/105726532-2c1e0f80-5ef8-11eb-86bb-272bb7782613.png">
non_main
defect buttons should have unique and accessible labels on the military service history page enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact josh user story or problem statement as a screen reader user i want to hear an accessible label like edit issues eligible for review instead of edit when the button has focus or virtual cursor details there are several issues with this page which may ultimately be unsolvable from a remediation standpoint i would recommend revisiting this page from a design lens issues are as follows edit buttons do not have unique and accessible labels as such they all currently announce edit which doesn t tell the user what they are editing inconsistency exists in legends when adding new military service as only the last item in the stack has the legend of new service period despite the possibility of other new service periods being added previously in the same session the legend new service period does not change to edit service period upon editing again after saving which may confuse the user this will require a larger conversation or meeting to discuss solutions so i ll only be documenting issues as opposed to recommendations for more details view the screenshots attached on this ticket acceptance criteria tbd solution tbd wcag or vendor guidance optional screenshots or trace logs img width alt screen shot at am src img width alt screen shot at am src img width alt screen shot at am src
0
823,980
31,075,953,353
IssuesEvent
2023-08-12 13:34:49
fossasia/open-event-frontend
https://api.github.com/repos/fossasia/open-event-frontend
opened
Mobile Schedule filter dropdowns not following defined order and dropdown too short
bug Priority: High
The mobile schedule filter dropdowns are too short and do not follow the defined order of the wizard. 1. Follow in wizard defined order 2. Increase dropdown height to show at least 10 items. ![Screenshot from 2023-08-12 21-27-49](https://github.com/fossasia/open-event-frontend/assets/1583873/4bae5eda-61eb-47ff-824c-9c8f3d412e9d)
1.0
Mobile Schedule filter dropdowns not following defined order and dropdown too short - The mobile schedule filter dropdowns are too short and do not follow the defined order of the wizard. 1. Follow in wizard defined order 2. Increase dropdown height to show at least 10 items. ![Screenshot from 2023-08-12 21-27-49](https://github.com/fossasia/open-event-frontend/assets/1583873/4bae5eda-61eb-47ff-824c-9c8f3d412e9d)
non_main
mobile schedule filter dropdowns not following defined order and dropdown too short the mobile schedule filter dropdowns are too short and do not follow the defined order of the wizard follow in wizard defined order increase dropdown height to show at least items
0
5,225
26,507,019,092
IssuesEvent
2023-01-18 14:32:26
precice/precice
https://api.github.com/repos/precice/precice
closed
Clarify mesh API
enhancement usability maintainability breaking change
**Please describe the problem you are trying to solve.** The API for setting mesh primitives is confusing and tedious. * Triangles have to be set using edges * `setMeshTriangleWithEdges` sounds like it would take edges, but it actually takes vertices. #1057 * There are no bulk functions for setting edges and triangles. #465 * Adding `setMeshTetrahedron` requiring triangles would be a huge pain for users. #1314 * Exposing handles to connectivity (EdgeID) prevents us from optimizing meshes #1313 **Describe the solution you propose.** 1. Change the API to a vertex-only style. **:warning: breaking** 2. Remove `XWithEdges` 3. Add bulk functions Function | Inputs + MeshID | Outputs | Comment --- | --- | --- | --- setMeshVertex | Coords | VertexID | _unchanged_ setMeshVertices | Count, Coords | VertexIDs | _unchanged_ setMeshEdge | VertexIDs | | _no return_ setMeshEdges | Count, VertexIDs | | _no return_ setMeshTriangle | VertexIDs | | _changed input_ setMeshTriangles | Count, VertexIDs | | _new_ setMeshQuad | VertexIDs | | _changed input_ setMeshQuads | Count, VertexIDs | | _new_ setMeshTetrahedron | VertexIDs | | _unchanged_ setMeshTetrahedra | Count, VertexIDs | | _new_ **Describe alternatives you've considered** Leave it as it and end up with an increasingly confusing API.
True
Clarify mesh API - **Please describe the problem you are trying to solve.** The API for setting mesh primitives is confusing and tedious. * Triangles have to be set using edges * `setMeshTriangleWithEdges` sounds like it would take edges, but it actually takes vertices. #1057 * There are no bulk functions for setting edges and triangles. #465 * Adding `setMeshTetrahedron` requiring triangles would be a huge pain for users. #1314 * Exposing handles to connectivity (EdgeID) prevents us from optimizing meshes #1313 **Describe the solution you propose.** 1. Change the API to a vertex-only style. **:warning: breaking** 2. Remove `XWithEdges` 3. Add bulk functions Function | Inputs + MeshID | Outputs | Comment --- | --- | --- | --- setMeshVertex | Coords | VertexID | _unchanged_ setMeshVertices | Count, Coords | VertexIDs | _unchanged_ setMeshEdge | VertexIDs | | _no return_ setMeshEdges | Count, VertexIDs | | _no return_ setMeshTriangle | VertexIDs | | _changed input_ setMeshTriangles | Count, VertexIDs | | _new_ setMeshQuad | VertexIDs | | _changed input_ setMeshQuads | Count, VertexIDs | | _new_ setMeshTetrahedron | VertexIDs | | _unchanged_ setMeshTetrahedra | Count, VertexIDs | | _new_ **Describe alternatives you've considered** Leave it as it and end up with an increasingly confusing API.
main
clarify mesh api please describe the problem you are trying to solve the api for setting mesh primitives is confusing and tedious triangles have to be set using edges setmeshtrianglewithedges sounds like it would take edges but it actually takes vertices there are no bulk functions for setting edges and triangles adding setmeshtetrahedron requiring triangles would be a huge pain for users exposing handles to connectivity edgeid prevents us from optimizing meshes describe the solution you propose change the api to a vertex only style warning breaking remove xwithedges add bulk functions function inputs meshid outputs comment setmeshvertex coords vertexid unchanged setmeshvertices count coords vertexids unchanged setmeshedge vertexids no return setmeshedges count vertexids no return setmeshtriangle vertexids changed input setmeshtriangles count vertexids new setmeshquad vertexids changed input setmeshquads count vertexids new setmeshtetrahedron vertexids unchanged setmeshtetrahedra count vertexids new describe alternatives you ve considered leave it as it and end up with an increasingly confusing api
1
5,193
26,405,381,143
IssuesEvent
2023-01-13 07:25:06
seleniumbase/SeleniumBase
https://api.github.com/repos/seleniumbase/SeleniumBase
closed
Maintainer on vacation from January 6 to January 12, 2023
maintainer on vacation
Hey, this is the maintainer of SeleniumBase. I'll be on vacation January 6 to January 12. Will be in Florida again, like my previous vacation a year ago (https://github.com/seleniumbase/SeleniumBase/issues/1224). Responses will be very delayed. I might be able to respond to simple questions via smartphone, but I plan on keeping the laptop lid closed during the trip, so I probably won't answer any questions until next Thursday when I'm back.
True
Maintainer on vacation from January 6 to January 12, 2023 - Hey, this is the maintainer of SeleniumBase. I'll be on vacation January 6 to January 12. Will be in Florida again, like my previous vacation a year ago (https://github.com/seleniumbase/SeleniumBase/issues/1224). Responses will be very delayed. I might be able to respond to simple questions via smartphone, but I plan on keeping the laptop lid closed during the trip, so I probably won't answer any questions until next Thursday when I'm back.
main
maintainer on vacation from january to january hey this is the maintainer of seleniumbase i ll be on vacation january to january will be in florida again like my previous vacation a year ago responses will be very delayed i might be able to respond to simple questions via smartphone but i plan on keeping the laptop lid closed during the trip so i probably won t answer any questions until next thursday when i m back
1
167,232
20,725,929,149
IssuesEvent
2022-03-14 01:52:02
jinuem/Cordova-Examples
https://api.github.com/repos/jinuem/Cordova-Examples
opened
CVE-2021-37701 (High) detected in tar-2.2.1.tgz, nodev11.15.0
security vulnerability
## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-2.2.1.tgz</b>, <b>nodev11.15.0</b></p></summary> <p> <details><summary><b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /Cordova-Examples/ionicWeatherV2/package.json</p> <p>Path to vulnerable library: /fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - ionic-gulp-sass-build-1.0.0.tgz (Root Library) - gulp-sass-2.3.2.tgz - node-sass-3.13.1.tgz - node-gyp-3.8.0.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.16,5.0.8,6.1.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37701 (High) detected in tar-2.2.1.tgz, nodev11.15.0 - ## CVE-2021-37701 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tar-2.2.1.tgz</b>, <b>nodev11.15.0</b></p></summary> <p> <details><summary><b>tar-2.2.1.tgz</b></p></summary> <p>tar for node</p> <p>Library home page: <a href="https://registry.npmjs.org/tar/-/tar-2.2.1.tgz">https://registry.npmjs.org/tar/-/tar-2.2.1.tgz</a></p> <p>Path to dependency file: /Cordova-Examples/ionicWeatherV2/package.json</p> <p>Path to vulnerable library: /fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json,/fixcontacts/node_modules/tar/package.json</p> <p> Dependency Hierarchy: - ionic-gulp-sass-build-1.0.0.tgz (Root Library) - gulp-sass-2.3.2.tgz - node-sass-3.13.1.tgz - node-gyp-3.8.0.tgz - :x: **tar-2.2.1.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The npm package "tar" (aka node-tar) before versions 4.4.16, 5.0.8, and 6.1.7 has an arbitrary file creation/overwrite and arbitrary code execution vulnerability. node-tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted. This is, in part, achieved by ensuring that extracted directories are not symlinks. Additionally, in order to prevent unnecessary stat calls to determine whether a given path is a directory, paths are cached when directories are created. This logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory, where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems. The cache checking logic used both `\` and `/` characters as path separators, however `\` is a valid filename character on posix systems. By first creating a directory, and then replacing that directory with a symlink, it was thus possible to bypass node-tar symlink checks on directories, essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location, thus allowing arbitrary file creation and overwrite. Additionally, a similar confusion could arise on case-insensitive filesystems. If a tar archive contained a directory at `FOO`, followed by a symbolic link named `foo`, then on case-insensitive file systems, the creation of the symbolic link would remove the directory from the filesystem, but _not_ from the internal directory cache, as it would not be treated as a cache hit. A subsequent file entry within the `FOO` directory would then be placed in the target of the symbolic link, thinking that the directory had already been created. These issues were addressed in releases 4.4.16, 5.0.8 and 6.1.7. The v3 branch of node-tar has been deprecated and did not receive patches for these issues. If you are still using a v3 release we recommend you update to a more recent version of node-tar. If this is not possible, a workaround is available in the referenced GHSA-9r2w-394v-53qc. <p>Publish Date: 2021-08-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37701>CVE-2021-37701</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc">https://github.com/npm/node-tar/security/advisories/GHSA-9r2w-394v-53qc</a></p> <p>Release Date: 2021-08-31</p> <p>Fix Resolution: tar - 4.4.16,5.0.8,6.1.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in tar tgz cve high severity vulnerability vulnerable libraries tar tgz tar tgz tar for node library home page a href path to dependency file cordova examples package json path to vulnerable library fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json fixcontacts node modules tar package json dependency hierarchy ionic gulp sass build tgz root library gulp sass tgz node sass tgz node gyp tgz x tar tgz vulnerable library vulnerability details the npm package tar aka node tar before versions and has an arbitrary file creation overwrite and arbitrary code execution vulnerability node tar aims to guarantee that any file whose location would be modified by a symbolic link is not extracted this is in part achieved by ensuring that extracted directories are not symlinks additionally in order to prevent unnecessary stat calls to determine whether a given path is a directory paths are cached when directories are created this logic was insufficient when extracting tar files that contained both a directory and a symlink with the same name as the directory where the symlink and directory names in the archive entry used backslashes as a path separator on posix systems the cache checking logic used both and characters as path separators however is a valid filename character on posix systems by first creating a directory and then replacing that directory with a symlink it was thus possible to bypass node tar symlink checks on directories essentially allowing an untrusted tar file to symlink into an arbitrary location and subsequently extracting arbitrary files into that location thus allowing arbitrary file creation and overwrite additionally a similar confusion could arise on case insensitive filesystems if a tar archive contained a directory at foo followed by a symbolic link named foo then on case insensitive file systems the creation of the symbolic link would remove the directory from the filesystem but not from the internal directory cache as it would not be treated as a cache hit a subsequent file entry within the foo directory would then be placed in the target of the symbolic link thinking that the directory had already been created these issues were addressed in releases and the branch of node tar has been deprecated and did not receive patches for these issues if you are still using a release we recommend you update to a more recent version of node tar if this is not possible a workaround is available in the referenced ghsa publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tar step up your open source security game with whitesource
0
143,974
11,589,563,405
IssuesEvent
2020-02-24 02:50:55
qmk/qmk_firmware
https://api.github.com/repos/qmk/qmk_firmware
closed
Bad Lag sometimes happens with I2C on split boards
bug core discussion help wanted needs testing
On some of my builds, if USE_I2C is used, that half and the whole board lags a bunch in registering keypresses. ## Describe the Bug I’ve tried to trace things down a bit, like erasing EEPROM, compiling with and without EE_HANDS and RGB, but no luck so far. Some of my halves work fine. I’ve heard someone else report a similar issue on Discord as well. ## System Information - Keyboard: Levinson Rev 2 & 3 - Revision (if applicable): - Operating system: Mac - AVR GCC version: 8.3 <!-- Run `avr-gcc --version` to find this out. --> - ARM GCC version: <!-- Run `arm-none-eabi-gcc --version` to find this out. --> - QMK Firmware version: Latest <!-- Run `git describe --abbrev=0 --tags` to find this out. --> - Any keyboard related software installed? - [ ] AutoHotKey - [ ] Karabiner - [ ] Other: ## Additional Context <!-- Add any other relevant information about the problem here. -->
1.0
Bad Lag sometimes happens with I2C on split boards - On some of my builds, if USE_I2C is used, that half and the whole board lags a bunch in registering keypresses. ## Describe the Bug I’ve tried to trace things down a bit, like erasing EEPROM, compiling with and without EE_HANDS and RGB, but no luck so far. Some of my halves work fine. I’ve heard someone else report a similar issue on Discord as well. ## System Information - Keyboard: Levinson Rev 2 & 3 - Revision (if applicable): - Operating system: Mac - AVR GCC version: 8.3 <!-- Run `avr-gcc --version` to find this out. --> - ARM GCC version: <!-- Run `arm-none-eabi-gcc --version` to find this out. --> - QMK Firmware version: Latest <!-- Run `git describe --abbrev=0 --tags` to find this out. --> - Any keyboard related software installed? - [ ] AutoHotKey - [ ] Karabiner - [ ] Other: ## Additional Context <!-- Add any other relevant information about the problem here. -->
non_main
bad lag sometimes happens with on split boards on some of my builds if use is used that half and the whole board lags a bunch in registering keypresses describe the bug i’ve tried to trace things down a bit like erasing eeprom compiling with and without ee hands and rgb but no luck so far some of my halves work fine i’ve heard someone else report a similar issue on discord as well system information keyboard levinson rev revision if applicable operating system mac avr gcc version arm gcc version qmk firmware version latest any keyboard related software installed autohotkey karabiner other additional context
0
714,219
24,554,820,073
IssuesEvent
2022-10-12 15:06:17
episphere/dashboard
https://api.github.com/repos/episphere/dashboard
closed
Participant Status=Error for suspended contact on all pts pg stage
High Priority Pending PROD Deployment
Adding github issue so we can track. The stage pt with a suspended contact has their participation status=ERROR on the all participants pg of the dshbrd. But their participation status looks correct on the pt details pg. Their status should = no refusal. For pt 3249429176.
1.0
Participant Status=Error for suspended contact on all pts pg stage - Adding github issue so we can track. The stage pt with a suspended contact has their participation status=ERROR on the all participants pg of the dshbrd. But their participation status looks correct on the pt details pg. Their status should = no refusal. For pt 3249429176.
non_main
participant status error for suspended contact on all pts pg stage adding github issue so we can track the stage pt with a suspended contact has their participation status error on the all participants pg of the dshbrd but their participation status looks correct on the pt details pg their status should no refusal for pt
0
527
3,925,502,773
IssuesEvent
2016-04-22 19:13:20
heiglandreas/authLdap
https://api.github.com/repos/heiglandreas/authLdap
closed
bug password reset should be disabled for authLdap users
feature maintainer reply expected
This adds the code to disable password resets for LDAP users. Apologies for not providing diffs. I added this function below authLdap_show_password_fields: ``` /** * This function disables the password reset for a user. * * It does not make sense to authenticate via LDAP and then allow the user to * reset the password only in the wordpress database. And changing the password * LDAP-wide can not be the scope of Wordpress! * * Whether the user is an LDAP-User or not is determined using the authLDAP-Flag * of the users meta-informations * * @return false, if the user is an LDAP-User, true if he isn't * @conf boolean authLDAP */ function authLdap_allow_password_reset($return, $userid) { if (!(isset($userid))) { return true; } if (get_user_meta($userid, 'authLDAP')) { return false; } return $return; } ``` This code goes at the bottom of authLdap.php: ``` add_filter('allow_password_reset', 'authLdap_allow_password_reset', 10, 2); ``` I've done some limited testing and it seems to work. I also have the plugin network activated.
True
bug password reset should be disabled for authLdap users - This adds the code to disable password resets for LDAP users. Apologies for not providing diffs. I added this function below authLdap_show_password_fields: ``` /** * This function disables the password reset for a user. * * It does not make sense to authenticate via LDAP and then allow the user to * reset the password only in the wordpress database. And changing the password * LDAP-wide can not be the scope of Wordpress! * * Whether the user is an LDAP-User or not is determined using the authLDAP-Flag * of the users meta-informations * * @return false, if the user is an LDAP-User, true if he isn't * @conf boolean authLDAP */ function authLdap_allow_password_reset($return, $userid) { if (!(isset($userid))) { return true; } if (get_user_meta($userid, 'authLDAP')) { return false; } return $return; } ``` This code goes at the bottom of authLdap.php: ``` add_filter('allow_password_reset', 'authLdap_allow_password_reset', 10, 2); ``` I've done some limited testing and it seems to work. I also have the plugin network activated.
main
bug password reset should be disabled for authldap users this adds the code to disable password resets for ldap users apologies for not providing diffs i added this function below authldap show password fields this function disables the password reset for a user it does not make sense to authenticate via ldap and then allow the user to reset the password only in the wordpress database and changing the password ldap wide can not be the scope of wordpress whether the user is an ldap user or not is determined using the authldap flag of the users meta informations return false if the user is an ldap user true if he isn t conf boolean authldap function authldap allow password reset return userid if isset userid return true if get user meta userid authldap return false return return this code goes at the bottom of authldap php add filter allow password reset authldap allow password reset i ve done some limited testing and it seems to work i also have the plugin network activated
1
3,378
13,082,397,771
IssuesEvent
2020-08-01 14:22:15
OpenLightingProject/ola
https://api.github.com/repos/OpenLightingProject/ola
closed
"readdir_r" deprecated in newer glibc versions
Difficulty-Easy Language-C++ Maintainability OpSys-Linux
In fedora's rawhide version of glibc `readdir_r` is deprecated ([build.log](https://kojipkgs.fedoraproject.org//work/tasks/9706/13179706/build.log)), the [patch](https://www.sourceware.org/ml/libc-alpha/2016-02/msg00093.html) or [bug](https://sourceware.org/bugzilla/show_bug.cgi?id=19056) don't really explain why or what to use else.
True
"readdir_r" deprecated in newer glibc versions - In fedora's rawhide version of glibc `readdir_r` is deprecated ([build.log](https://kojipkgs.fedoraproject.org//work/tasks/9706/13179706/build.log)), the [patch](https://www.sourceware.org/ml/libc-alpha/2016-02/msg00093.html) or [bug](https://sourceware.org/bugzilla/show_bug.cgi?id=19056) don't really explain why or what to use else.
main
readdir r deprecated in newer glibc versions in fedora s rawhide version of glibc readdir r is deprecated the or don t really explain why or what to use else
1
65,402
27,092,076,425
IssuesEvent
2023-02-14 22:03:22
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Not mentioning filter types
service-bus-messaging/svc triaged assigned-to-author doc-enhancement Pri2
Microsoft makes a categorization on the filter types: sql, boolean, and correlation. But that is not reflected in this examples. I would advise Microsoft to not make these categorizations that looks simple but generates more questions. Or specify the difference in these types with clear examples. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 1bce37cf-a134-9a2c-7c1e-81329ab5af72 * Version Independent ID: f97acc7d-9936-ea76-0bdb-1838333abe57 * Content: [Set subscriptions filters in Azure Service Bus - Azure Service Bus](https://learn.microsoft.com/en-us/azure/service-bus-messaging/service-bus-filter-examples) * Content Source: [articles/service-bus-messaging/service-bus-filter-examples.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/service-bus-messaging/service-bus-filter-examples.md) * Service: **service-bus-messaging** * GitHub Login: @spelluru * Microsoft Alias: **spelluru**
1.0
Not mentioning filter types - Microsoft makes a categorization on the filter types: sql, boolean, and correlation. But that is not reflected in this examples. I would advise Microsoft to not make these categorizations that looks simple but generates more questions. Or specify the difference in these types with clear examples. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 1bce37cf-a134-9a2c-7c1e-81329ab5af72 * Version Independent ID: f97acc7d-9936-ea76-0bdb-1838333abe57 * Content: [Set subscriptions filters in Azure Service Bus - Azure Service Bus](https://learn.microsoft.com/en-us/azure/service-bus-messaging/service-bus-filter-examples) * Content Source: [articles/service-bus-messaging/service-bus-filter-examples.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/service-bus-messaging/service-bus-filter-examples.md) * Service: **service-bus-messaging** * GitHub Login: @spelluru * Microsoft Alias: **spelluru**
non_main
not mentioning filter types microsoft makes a categorization on the filter types sql boolean and correlation but that is not reflected in this examples i would advise microsoft to not make these categorizations that looks simple but generates more questions or specify the difference in these types with clear examples document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service service bus messaging github login spelluru microsoft alias spelluru
0
187,803
22,045,920,308
IssuesEvent
2022-05-30 01:40:22
ngocdai94/Hacker-Scripts
https://api.github.com/repos/ngocdai94/Hacker-Scripts
opened
CVE-2020-36242 (High) detected in cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl
security vulnerability
## CVE-2020-36242 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/c3/c1/cf8665c955c9393e9ff0872ba6cd3dc6f46ef915e94afcf6e0410508ca69/cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/c3/c1/cf8665c955c9393e9ff0872ba6cd3dc6f46ef915e94afcf6e0410508ca69/cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /Hacker-Scripts/python3/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_65f23061-8160-4967-8e57-7f3ead083caf/20190530200043_59909/20190530200026_depth_0/cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64/cryptography</p> <p> Dependency Hierarchy: - :x: **cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the cryptography package before 3.3.2 for Python, certain sequences of update calls to symmetrically encrypt multi-GB values could result in an integer overflow and buffer overflow, as demonstrated by the Fernet class. <p>Publish Date: 2021-02-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36242>CVE-2020-36242</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/pyca/cryptography/blob/master/CHANGELOG.rst">https://github.com/pyca/cryptography/blob/master/CHANGELOG.rst</a></p> <p>Release Date: 2021-02-07</p> <p>Fix Resolution: cryptography - 3.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36242 (High) detected in cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2020-36242 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/c3/c1/cf8665c955c9393e9ff0872ba6cd3dc6f46ef915e94afcf6e0410508ca69/cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/c3/c1/cf8665c955c9393e9ff0872ba6cd3dc6f46ef915e94afcf6e0410508ca69/cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: /Hacker-Scripts/python3/requirements.txt</p> <p>Path to vulnerable library: /teSource-ArchiveExtractor_65f23061-8160-4967-8e57-7f3ead083caf/20190530200043_59909/20190530200026_depth_0/cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64/cryptography</p> <p> Dependency Hierarchy: - :x: **cryptography-2.6.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In the cryptography package before 3.3.2 for Python, certain sequences of update calls to symmetrically encrypt multi-GB values could result in an integer overflow and buffer overflow, as demonstrated by the Fernet class. <p>Publish Date: 2021-02-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36242>CVE-2020-36242</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/pyca/cryptography/blob/master/CHANGELOG.rst">https://github.com/pyca/cryptography/blob/master/CHANGELOG.rst</a></p> <p>Release Date: 2021-02-07</p> <p>Fix Resolution: cryptography - 3.3.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in cryptography whl cve high severity vulnerability vulnerable library cryptography whl cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file hacker scripts requirements txt path to vulnerable library tesource archiveextractor depth cryptography cryptography dependency hierarchy x cryptography whl vulnerable library vulnerability details in the cryptography package before for python certain sequences of update calls to symmetrically encrypt multi gb values could result in an integer overflow and buffer overflow as demonstrated by the fernet class publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cryptography step up your open source security game with mend
0
342,310
10,315,036,815
IssuesEvent
2019-08-30 06:16:19
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.ncbi.nlm.nih.gov - The page is displayed with overlapping elements
browser-fenix engine-gecko priority-important severity-important
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.ncbi.nlm.nih.gov/pubmed/advanced **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: improper site layout **Steps to Reproduce**: Just typed in the web address [![Screenshot Description](https://webcompat.com/uploads/2019/7/726a746b-8413-4fec-8099-92b9ce8a6d96-thumb.jpg)](https://webcompat.com/uploads/2019/7/726a746b-8413-4fec-8099-92b9ce8a6d96.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.ncbi.nlm.nih.gov - The page is displayed with overlapping elements - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.ncbi.nlm.nih.gov/pubmed/advanced **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: improper site layout **Steps to Reproduce**: Just typed in the web address [![Screenshot Description](https://webcompat.com/uploads/2019/7/726a746b-8413-4fec-8099-92b9ce8a6d96-thumb.jpg)](https://webcompat.com/uploads/2019/7/726a746b-8413-4fec-8099-92b9ce8a6d96.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
the page is displayed with overlapping elements url browser version firefox mobile operating system android tested another browser yes problem type site is not usable description improper site layout steps to reproduce just typed in the web address browser configuration none from with ❤️
0
5,437
27,245,989,737
IssuesEvent
2023-02-22 02:04:31
VA-Explorer/va_explorer
https://api.github.com/repos/VA-Explorer/va_explorer
opened
Change dashboard trends to only display last 3 years worth of data
Type: Maintainance good first issue Domain: Frontend
**What is the expected state?** When I view the dashboard I want to see data that is no older than Jan 1 2020 in the trends panel. **What is the actual state?** When I view the dashboard I see data from the 1960s (possible corrupt VAs?) **Relevant context** - Consider tackling #263 first as it will affect this issue. - Question to consider: Should another issue be created to allow not keeping data older than X years in the first place? (ex. don't pull data older than 3 years from ODK and delete data in VA explorer if it's older than 3 years)
True
Change dashboard trends to only display last 3 years worth of data - **What is the expected state?** When I view the dashboard I want to see data that is no older than Jan 1 2020 in the trends panel. **What is the actual state?** When I view the dashboard I see data from the 1960s (possible corrupt VAs?) **Relevant context** - Consider tackling #263 first as it will affect this issue. - Question to consider: Should another issue be created to allow not keeping data older than X years in the first place? (ex. don't pull data older than 3 years from ODK and delete data in VA explorer if it's older than 3 years)
main
change dashboard trends to only display last years worth of data what is the expected state when i view the dashboard i want to see data that is no older than jan in the trends panel what is the actual state when i view the dashboard i see data from the possible corrupt vas relevant context consider tackling first as it will affect this issue question to consider should another issue be created to allow not keeping data older than x years in the first place ex don t pull data older than years from odk and delete data in va explorer if it s older than years
1
1,515
6,548,401,121
IssuesEvent
2017-09-04 21:32:29
commercialhaskell/path
https://api.github.com/repos/commercialhaskell/path
opened
Sharing Travis capacity with the rest of the commercialhaskell organization
discuss help wanted maintainability
It's annoying that my Travis job is stuck waiting (apparently) for a stack job. I'm not sure if this can be fixed in any other way than moving this repo to a different (new) organization or account.
True
Sharing Travis capacity with the rest of the commercialhaskell organization - It's annoying that my Travis job is stuck waiting (apparently) for a stack job. I'm not sure if this can be fixed in any other way than moving this repo to a different (new) organization or account.
main
sharing travis capacity with the rest of the commercialhaskell organization it s annoying that my travis job is stuck waiting apparently for a stack job i m not sure if this can be fixed in any other way than moving this repo to a different new organization or account
1
711,325
24,458,567,271
IssuesEvent
2022-10-07 09:03:28
root-project/root
https://api.github.com/repos/root-project/root
closed
TExecutorCRTP and TSequentialExecutor errors when building application with Kokkos 3.7
priority:default
I'm seeing an error when building an application with root 6.26.06, gcc 11.2 and Kokkos 3.7. I didn't see this when building the same source against root 6.24.00, Kokkos 3.4 with gcc 9.3. Nothing in the source is explicitly calling TExecutorCRTP. Before I start playing with all permutations of root / gcc / kokkos to see where the issue starts, I'm wondering if anyone has seen this before or has some advice. This is the error I see: ``` /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:191:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 191 | auto TExecutorCRTP<SubC>::Map(F func, ROOT::TSeq<INTEGER> args) -> std::vector<typename std::result_of<F(INTEGER)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:203:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 203 | auto TExecutorCRTP<SubC>::Map(F func, std::initializer_list<T> args) -> std::vector<typename std::result_of<F(T)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:217:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 217 | auto TExecutorCRTP<SubC>::Map(F func, std::vector<T> &args) -> std::vector<typename std::result_of<F(T)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:230:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 230 | auto TExecutorCRTP<SubC>::Map(F func, const std::vector<T> &args) -> std::vector<typename std::result_of<F(T)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:257:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 257 | auto TExecutorCRTP<SubC>::MapReduce(F func, ROOT::TSeq<INTEGER> args, R redfunc) -> typename std::result_of<F(INTEGER)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 134 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:132:93: note: ‘template<class SubC> template<class F, class INTEGER, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 132 | template<class F, class INTEGER, class R, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:130:69: note: ‘template<class SubC> template<class F, class R, class Cond> typename std::result_of<F()>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, unsigned int, R)’ 130 | template<class F, class R, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:270:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 270 | auto TExecutorCRTP<SubC>::MapReduce(F func, std::initializer_list<T> args, R redfunc) -> typename std::result_of<F(T)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 134 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:132:93: note: ‘template<class SubC> template<class F, class INTEGER, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 132 | template<class F, class INTEGER, class R, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:130:69: note: ‘template<class SubC> template<class F, class R, class Cond> typename std::result_of<F()>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, unsigned int, R)’ 130 | template<class F, class R, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:284:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 284 | auto TExecutorCRTP<SubC>::MapReduce(F func, std::vector<T> &args, R redfunc) -> typename std::result_of<F(T)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 134 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:132:93: note: ‘template<class SubC> template<class F, class INTEGER, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 132 | template<class F, class INTEGER, class R, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:130:69: note: ‘template<class SubC> template<class F, class R, class Cond> typename std::result_of<F()>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, unsigned int, R)’ 130 | template<class F, class R, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:296:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 296 | auto TExecutorCRTP<SubC>::MapReduce(F func, const std::vector<T> &args, R redfunc) -> typename std::result_of<F(T)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ ``` similarly for `TSequentialExecutor`
1.0
TExecutorCRTP and TSequentialExecutor errors when building application with Kokkos 3.7 - I'm seeing an error when building an application with root 6.26.06, gcc 11.2 and Kokkos 3.7. I didn't see this when building the same source against root 6.24.00, Kokkos 3.4 with gcc 9.3. Nothing in the source is explicitly calling TExecutorCRTP. Before I start playing with all permutations of root / gcc / kokkos to see where the issue starts, I'm wondering if anyone has seen this before or has some advice. This is the error I see: ``` /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:191:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 191 | auto TExecutorCRTP<SubC>::Map(F func, ROOT::TSeq<INTEGER> args) -> std::vector<typename std::result_of<F(INTEGER)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:203:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 203 | auto TExecutorCRTP<SubC>::Map(F func, std::initializer_list<T> args) -> std::vector<typename std::result_of<F(T)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:217:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 217 | auto TExecutorCRTP<SubC>::Map(F func, std::vector<T> &args) -> std::vector<typename std::result_of<F(T)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:230:1: error: no declaration matches ‘std::vector<typename std::result_of<F(INTEGER)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 230 | auto TExecutorCRTP<SubC>::Map(F func, const std::vector<T> &args) -> std::vector<typename std::result_of<F(T)>::type> | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:124:72: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, const std::vector<T>&)’ 124 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:122:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::vector<T>&)’ 122 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:120:72: note: ‘template<class SubC> template<class F, class T, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, std::initializer_list<T>)’ 120 | template<class F, class T, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:118:84: note: ‘template<class SubC> template<class F, class INTEGER, class Cond> std::vector<typename std::result_of<F(INTEGER, ...)>::type> ROOT::TExecutorCRTP<SubC>::Map(F, ROOT::TSeq<INTEGER>)’ 118 | template<class F, class INTEGER, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:116:60: note: ‘template<class SubC> template<class F, class Cond> std::vector<typename std::result_of<F()>::type> ROOT::TExecutorCRTP<SubC>::Map(F, unsigned int)’ 116 | template<class F, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:257:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 257 | auto TExecutorCRTP<SubC>::MapReduce(F func, ROOT::TSeq<INTEGER> args, R redfunc) -> typename std::result_of<F(INTEGER)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 134 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:132:93: note: ‘template<class SubC> template<class F, class INTEGER, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 132 | template<class F, class INTEGER, class R, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:130:69: note: ‘template<class SubC> template<class F, class R, class Cond> typename std::result_of<F()>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, unsigned int, R)’ 130 | template<class F, class R, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:270:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 270 | auto TExecutorCRTP<SubC>::MapReduce(F func, std::initializer_list<T> args, R redfunc) -> typename std::result_of<F(T)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 134 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:132:93: note: ‘template<class SubC> template<class F, class INTEGER, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 132 | template<class F, class INTEGER, class R, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:130:69: note: ‘template<class SubC> template<class F, class R, class Cond> typename std::result_of<F()>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, unsigned int, R)’ 130 | template<class F, class R, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:284:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 284 | auto TExecutorCRTP<SubC>::MapReduce(F func, std::vector<T> &args, R redfunc) -> typename std::result_of<F(T)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ 134 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:132:93: note: ‘template<class SubC> template<class F, class INTEGER, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, ROOT::TSeq<INTEGER>, R)’ 132 | template<class F, class INTEGER, class R, class Cond = noReferenceCond<F, INTEGER>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:130:69: note: ‘template<class SubC> template<class F, class R, class Cond> typename std::result_of<F()>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, unsigned int, R)’ 130 | template<class F, class R, class Cond = noReferenceCond<F>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:102:7: note: ‘class ROOT::TExecutorCRTP<SubC>’ defined here 102 | class TExecutorCRTP { | ^~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:296:1: error: no declaration matches ‘typename std::result_of<F(INTEGER)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 296 | auto TExecutorCRTP<SubC>::MapReduce(F func, const std::vector<T> &args, R redfunc) -> typename std::result_of<F(T)>::type | ^~~~~~~~~~~~~~~~~~~~ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:142:70: note: candidates are: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T*>&)’ 142 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:140:70: note: ‘template<class SubC> template<class F, class T, class Cond> T* ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T*>&)’ 140 | template<class F, class T,class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:138:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::vector<T>&, R)’ 138 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:136:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, const std::vector<T>&, R)’ 136 | template<class F, class T, class R, class Cond = noReferenceCond<F, T>> | ^ /opt/root/v6-26-06_gcc112_c17/include/ROOT/TExecutorCRTP.hxx:134:81: note: ‘template<class SubC> template<class F, class T, class R, class Cond> typename std::result_of<F(INTEGER, ...)>::type ROOT::TExecutorCRTP<SubC>::MapReduce(F, std::initializer_list<T>, R)’ ``` similarly for `TSequentialExecutor`
non_main
texecutorcrtp and tsequentialexecutor errors when building application with kokkos i m seeing an error when building an application with root gcc and kokkos i didn t see this when building the same source against root kokkos with gcc nothing in the source is explicitly calling texecutorcrtp before i start playing with all permutations of root gcc kokkos to see where the issue starts i m wondering if anyone has seen this before or has some advice this is the error i see opt root include root texecutorcrtp hxx error no declaration matches ‘std vector type root texecutorcrtp map f root tseq ’ auto texecutorcrtp map f func root tseq args std vector type opt root include root texecutorcrtp hxx note candidates are ‘template template std vector type root texecutorcrtp map f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std initializer list ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f root tseq ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f unsigned int ’ template opt root include root texecutorcrtp hxx note ‘class root texecutorcrtp ’ defined here class texecutorcrtp opt root include root texecutorcrtp hxx error no declaration matches ‘std vector type root texecutorcrtp map f std initializer list ’ auto texecutorcrtp map f func std initializer list args std vector type opt root include root texecutorcrtp hxx note candidates are ‘template template std vector type root texecutorcrtp map f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std initializer list ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f root tseq ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f unsigned int ’ template opt root include root texecutorcrtp hxx note ‘class root texecutorcrtp ’ defined here class texecutorcrtp opt root include root texecutorcrtp hxx error no declaration matches ‘std vector type root texecutorcrtp map f std vector ’ auto texecutorcrtp map f func std vector args std vector type opt root include root texecutorcrtp hxx note candidates are ‘template template std vector type root texecutorcrtp map f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std initializer list ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f root tseq ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f unsigned int ’ template opt root include root texecutorcrtp hxx note ‘class root texecutorcrtp ’ defined here class texecutorcrtp opt root include root texecutorcrtp hxx error no declaration matches ‘std vector type root texecutorcrtp map f const std vector ’ auto texecutorcrtp map f func const std vector args std vector type opt root include root texecutorcrtp hxx note candidates are ‘template template std vector type root texecutorcrtp map f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f std initializer list ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f root tseq ’ template opt root include root texecutorcrtp hxx note ‘template template std vector type root texecutorcrtp map f unsigned int ’ template opt root include root texecutorcrtp hxx note ‘class root texecutorcrtp ’ defined here class texecutorcrtp opt root include root texecutorcrtp hxx error no declaration matches ‘typename std result of type root texecutorcrtp mapreduce f root tseq r ’ auto texecutorcrtp mapreduce f func root tseq args r redfunc typename std result of type opt root include root texecutorcrtp hxx note candidates are ‘template template t root texecutorcrtp mapreduce f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template t root texecutorcrtp mapreduce f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f const std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std initializer list r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f root tseq r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f unsigned int r ’ template opt root include root texecutorcrtp hxx note ‘class root texecutorcrtp ’ defined here class texecutorcrtp opt root include root texecutorcrtp hxx error no declaration matches ‘typename std result of type root texecutorcrtp mapreduce f std initializer list r ’ auto texecutorcrtp mapreduce f func std initializer list args r redfunc typename std result of type opt root include root texecutorcrtp hxx note candidates are ‘template template t root texecutorcrtp mapreduce f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template t root texecutorcrtp mapreduce f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f const std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std initializer list r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f root tseq r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f unsigned int r ’ template opt root include root texecutorcrtp hxx note ‘class root texecutorcrtp ’ defined here class texecutorcrtp opt root include root texecutorcrtp hxx error no declaration matches ‘typename std result of type root texecutorcrtp mapreduce f std vector r ’ auto texecutorcrtp mapreduce f func std vector args r redfunc typename std result of type opt root include root texecutorcrtp hxx note candidates are ‘template template t root texecutorcrtp mapreduce f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template t root texecutorcrtp mapreduce f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f const std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std initializer list r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f root tseq r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f unsigned int r ’ template opt root include root texecutorcrtp hxx note ‘class root texecutorcrtp ’ defined here class texecutorcrtp opt root include root texecutorcrtp hxx error no declaration matches ‘typename std result of type root texecutorcrtp mapreduce f const std vector r ’ auto texecutorcrtp mapreduce f func const std vector args r redfunc typename std result of type opt root include root texecutorcrtp hxx note candidates are ‘template template t root texecutorcrtp mapreduce f const std vector ’ template opt root include root texecutorcrtp hxx note ‘template template t root texecutorcrtp mapreduce f std vector ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f const std vector r ’ template opt root include root texecutorcrtp hxx note ‘template template typename std result of type root texecutorcrtp mapreduce f std initializer list r ’ similarly for tsequentialexecutor
0
5,531
27,646,120,352
IssuesEvent
2023-03-10 23:19:52
microsoft/DirectXTex
https://api.github.com/repos/microsoft/DirectXTex
closed
BC6 BC7 CPU /analyze warning
maintainence
Looking at this /analyze warning, it looks like there's a bug in the code. The helpers **OptimizeRGB** and **OptimizeRGBA** are never updated ``fError`` in the loop. ``` BC6HBC7.cpp(1477) : warning C6240: (<expression> && <non-zero constant>) always evaluates to the result of <expression>. Did you intend to use the bitwise-and operator? BC6HBC7.cpp(1477) : warning C6326: Potential comparison of a constant with another constant. ```
True
BC6 BC7 CPU /analyze warning - Looking at this /analyze warning, it looks like there's a bug in the code. The helpers **OptimizeRGB** and **OptimizeRGBA** are never updated ``fError`` in the loop. ``` BC6HBC7.cpp(1477) : warning C6240: (<expression> && <non-zero constant>) always evaluates to the result of <expression>. Did you intend to use the bitwise-and operator? BC6HBC7.cpp(1477) : warning C6326: Potential comparison of a constant with another constant. ```
main
cpu analyze warning looking at this analyze warning it looks like there s a bug in the code the helpers optimizergb and optimizergba are never updated ferror in the loop cpp warning always evaluates to the result of did you intend to use the bitwise and operator cpp warning potential comparison of a constant with another constant
1
433,567
12,506,855,622
IssuesEvent
2020-06-02 13:15:57
grpc/grpc
https://api.github.com/repos/grpc/grpc
opened
How can I use boringssl instead of openssl ?
kind/question priority/P3
Hi, I want to use grpc as a static library. That is, linking libgrpc++.a and libgrpc.a into my libmy.so. But I found that, if openssl has installed on my machine, grpc will use openssl instead of boringssl, which makes libmy.so depends on libopenssl.so. That is not what I want. Can I find some options to disable openssl ? My version is 1.26.x. Thanks a lot !
1.0
How can I use boringssl instead of openssl ? - Hi, I want to use grpc as a static library. That is, linking libgrpc++.a and libgrpc.a into my libmy.so. But I found that, if openssl has installed on my machine, grpc will use openssl instead of boringssl, which makes libmy.so depends on libopenssl.so. That is not what I want. Can I find some options to disable openssl ? My version is 1.26.x. Thanks a lot !
non_main
how can i use boringssl instead of openssl hi i want to use grpc as a static library that is linking libgrpc a and libgrpc a into my libmy so but i found that if openssl has installed on my machine grpc will use openssl instead of boringssl which makes libmy so depends on libopenssl so that is not what i want can i find some options to disable openssl my version is x thanks a lot
0
150,239
13,327,082,618
IssuesEvent
2020-08-27 12:40:13
EntCollNet/BugFlow
https://api.github.com/repos/EntCollNet/BugFlow
closed
Modules 4A-J (Imaging) Need Created
documentation
4A - Pinned 2D photo-stacked specimen imaging 4B - Pinned 3D specimen imaging 4C - Enveloped/flat (dragonfly?) Imaging 4D - Frozen/tissue imaging (subcase of "wet" museum specimens?) 4E - Slide photo-stacked imaging 4F - Slide label/low-res imaging 4G - Wet photo-stacked imaging 4H - Dry Bulk low-res Imaging 4I - Wet Soup low-res Imaging 4J - Image Processing
1.0
Modules 4A-J (Imaging) Need Created - 4A - Pinned 2D photo-stacked specimen imaging 4B - Pinned 3D specimen imaging 4C - Enveloped/flat (dragonfly?) Imaging 4D - Frozen/tissue imaging (subcase of "wet" museum specimens?) 4E - Slide photo-stacked imaging 4F - Slide label/low-res imaging 4G - Wet photo-stacked imaging 4H - Dry Bulk low-res Imaging 4I - Wet Soup low-res Imaging 4J - Image Processing
non_main
modules j imaging need created pinned photo stacked specimen imaging pinned specimen imaging enveloped flat dragonfly imaging frozen tissue imaging subcase of wet museum specimens slide photo stacked imaging slide label low res imaging wet photo stacked imaging dry bulk low res imaging wet soup low res imaging image processing
0
356,964
10,599,839,360
IssuesEvent
2019-10-10 08:50:20
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.google.com - design is broken
ML Correct ML ON browser-fenix engine-gecko priority-critical
<!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.google.com/search?q=test **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Searches look really ugly compared to other browser. The layout of Google searches is rounder in other browsers. **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.google.com - design is broken - <!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.google.com/search?q=test **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: Searches look really ugly compared to other browser. The layout of Google searches is rounder in other browsers. **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
design is broken url browser version firefox mobile operating system android tested another browser yes problem type design is broken description searches look really ugly compared to other browser the layout of google searches is rounder in other browsers steps to reproduce browser configuration none from with ❤️
0
2,652
9,083,443,576
IssuesEvent
2019-02-17 20:31:23
lrozenblyum/chess
https://api.github.com/repos/lrozenblyum/chess
closed
Update Maven plug-ins
maintainability
Related to #245. - [x] maven-compiler-plugin - [x] maven-clean-plugin - already up-to-date - [x] maven-jar-plugin - [x] maven-dependency-plugin - [x] maven-assembly-plugin - [x] github-release-plugin - already up-to-date - [x] pitest-maven - split to #262 - [x] implicit plugins - [x] https://stackoverflow.com/questions/21128372/how-does-maven-resolve-plugin-versions - [x] maven-install-plugin - [x] maven-site-plugin - [x] maven-surefire-plugin - [x] maven-deploy-plugin - [x] maven-resources-plugin - [x] Interesting tool: http://www.mojohaus.org/versions-maven-plugin/display-plugin-updates-mojo.html In this task we may review and drop support of local sonar-related configurations if they're not topical anymore (or better split to another one)
True
Update Maven plug-ins - Related to #245. - [x] maven-compiler-plugin - [x] maven-clean-plugin - already up-to-date - [x] maven-jar-plugin - [x] maven-dependency-plugin - [x] maven-assembly-plugin - [x] github-release-plugin - already up-to-date - [x] pitest-maven - split to #262 - [x] implicit plugins - [x] https://stackoverflow.com/questions/21128372/how-does-maven-resolve-plugin-versions - [x] maven-install-plugin - [x] maven-site-plugin - [x] maven-surefire-plugin - [x] maven-deploy-plugin - [x] maven-resources-plugin - [x] Interesting tool: http://www.mojohaus.org/versions-maven-plugin/display-plugin-updates-mojo.html In this task we may review and drop support of local sonar-related configurations if they're not topical anymore (or better split to another one)
main
update maven plug ins related to maven compiler plugin maven clean plugin already up to date maven jar plugin maven dependency plugin maven assembly plugin github release plugin already up to date pitest maven split to implicit plugins maven install plugin maven site plugin maven surefire plugin maven deploy plugin maven resources plugin interesting tool in this task we may review and drop support of local sonar related configurations if they re not topical anymore or better split to another one
1
310,811
26,746,196,929
IssuesEvent
2023-01-30 16:09:03
r-three/git-theta
https://api.github.com/repos/r-three/git-theta
opened
Fast test target
testing
Some of our unittests do probablistic/randomized testing and run things many times. We should have 2 ways to run the unittests a fast on that run more normal SWE tests and is really quick and a more through one the runs the computational tests too.
1.0
Fast test target - Some of our unittests do probablistic/randomized testing and run things many times. We should have 2 ways to run the unittests a fast on that run more normal SWE tests and is really quick and a more through one the runs the computational tests too.
non_main
fast test target some of our unittests do probablistic randomized testing and run things many times we should have ways to run the unittests a fast on that run more normal swe tests and is really quick and a more through one the runs the computational tests too
0
2,849
10,219,571,462
IssuesEvent
2019-08-15 18:56:09
arcticicestudio/styleguide-javascript
https://api.github.com/repos/arcticicestudio/styleguide-javascript
closed
Husky
context-workflow scope-dx scope-maintainability scope-quality type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/48658801-30ad2a80-ea48-11e8-9323-16bb0b25002b.png" width="20%" /></p> > Epic: #8 > Must be resolved **after** #12 Integrate [Husky][gh-husky], the tool that make Git hooks easy and can prevent bad Git commits, pushes and more :dog: _woof_! ### Configuration The configuration file `.huskyrc.js` will be placed in the project root and includes the command to run for any [supported Git hook][gh-husky-docs-hooks]. It will at least contain configs for the following hooks: - `pre-commit` - Run lint-staged (#12) before each commit (via `lint-staged` command) to ensure all staged files are compliant to all style guides. ## Tasks - [x] Install [husky][npm-husky] package. - [x] Implement `.huskyrc.js` configuration file. [gh-husky-docs-hooks]: https://github.com/typicode/husky/blob/master/DOCS.md#supported-hooks [gh-husky]: https://github.com/typicode/husky [npm-husky]: https://www.npmjs.com/package/husky
True
Husky - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48658801-30ad2a80-ea48-11e8-9323-16bb0b25002b.png" width="20%" /></p> > Epic: #8 > Must be resolved **after** #12 Integrate [Husky][gh-husky], the tool that make Git hooks easy and can prevent bad Git commits, pushes and more :dog: _woof_! ### Configuration The configuration file `.huskyrc.js` will be placed in the project root and includes the command to run for any [supported Git hook][gh-husky-docs-hooks]. It will at least contain configs for the following hooks: - `pre-commit` - Run lint-staged (#12) before each commit (via `lint-staged` command) to ensure all staged files are compliant to all style guides. ## Tasks - [x] Install [husky][npm-husky] package. - [x] Implement `.huskyrc.js` configuration file. [gh-husky-docs-hooks]: https://github.com/typicode/husky/blob/master/DOCS.md#supported-hooks [gh-husky]: https://github.com/typicode/husky [npm-husky]: https://www.npmjs.com/package/husky
main
husky epic must be resolved after integrate the tool that make git hooks easy and can prevent bad git commits pushes and more dog woof configuration the configuration file huskyrc js will be placed in the project root and includes the command to run for any it will at least contain configs for the following hooks pre commit run lint staged before each commit via lint staged command to ensure all staged files are compliant to all style guides tasks install package implement huskyrc js configuration file
1
1,981
6,694,193,362
IssuesEvent
2017-10-10 00:13:33
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Editor: Needs UX improvement
Maintainer Input Requested
User feedback: "The idea is great but... how can I run the python code? This is not intuitive at all." --- IA Page: http://duck.co/ia/view/editor [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jmg
True
Editor: Needs UX improvement - User feedback: "The idea is great but... how can I run the python code? This is not intuitive at all." --- IA Page: http://duck.co/ia/view/editor [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jmg
main
editor needs ux improvement user feedback the idea is great but how can i run the python code this is not intuitive at all ia page jmg
1
1,576
6,572,341,584
IssuesEvent
2017-09-11 01:32:47
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
lxc_container: logging is not applied to any action except container creation
affects_2.1 bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> lxc_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> None ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> N/A ##### SUMMARY <!--- Explain the problem briefly --> The `container_log` and `container_log_level` parameters are only ever used for the container creation. It is never applied to the changes of state, to the command execution or to any other actions taken against the container. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> The details steps to demonstrate the bug are in https://gist.github.com/odyssey4me/dc94e44d96d94e36bf11c0b8faa0c7ee <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expect that the `container_log` and `container_log_level` parameters are applied to every state change and every command execution. Any state changes introduced by container_config changes should also be logged if the parameters are provided. The detailed actual results and expected results are demonstrated in https://gist.github.com/odyssey4me/dc94e44d96d94e36bf11c0b8faa0c7ee ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> The detailed actual results and expected results are demonstrated in https://gist.github.com/odyssey4me/dc94e44d96d94e36bf11c0b8faa0c7ee
True
lxc_container: logging is not applied to any action except container creation - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> lxc_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> None ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> N/A ##### SUMMARY <!--- Explain the problem briefly --> The `container_log` and `container_log_level` parameters are only ever used for the container creation. It is never applied to the changes of state, to the command execution or to any other actions taken against the container. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> The details steps to demonstrate the bug are in https://gist.github.com/odyssey4me/dc94e44d96d94e36bf11c0b8faa0c7ee <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expect that the `container_log` and `container_log_level` parameters are applied to every state change and every command execution. Any state changes introduced by container_config changes should also be logged if the parameters are provided. The detailed actual results and expected results are demonstrated in https://gist.github.com/odyssey4me/dc94e44d96d94e36bf11c0b8faa0c7ee ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> The detailed actual results and expected results are demonstrated in https://gist.github.com/odyssey4me/dc94e44d96d94e36bf11c0b8faa0c7ee
main
lxc container logging is not applied to any action except container creation issue type bug report component name lxc container ansible version ansible config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables none os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific n a summary the container log and container log level parameters are only ever used for the container creation it is never applied to the changes of state to the command execution or to any other actions taken against the container steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used the details steps to demonstrate the bug are in expected results i expect that the container log and container log level parameters are applied to every state change and every command execution any state changes introduced by container config changes should also be logged if the parameters are provided the detailed actual results and expected results are demonstrated in actual results the detailed actual results and expected results are demonstrated in
1
3,717
15,353,415,800
IssuesEvent
2021-03-01 08:32:45
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
opened
Exceptional Log statements should end with colon
Area: analyzer Area: maintainability feature
Log statements that log exceptions should end with a colon, to indicate that there is an exception.
True
Exceptional Log statements should end with colon - Log statements that log exceptions should end with a colon, to indicate that there is an exception.
main
exceptional log statements should end with colon log statements that log exceptions should end with a colon to indicate that there is an exception
1
5,003
25,732,477,115
IssuesEvent
2022-12-07 21:27:26
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
ComboBox doesn't support entry of Text not found in list items
type: bug 🐛 component: combobox status: waiting for maintainer response 💬
## ComboBox has onInputChange but no way to know when Enter is pressed ## Though documentation claims the ComboBox allows people to enter text not found in the list items, there is no notification that this has occurred. There should be an onSubmit method to allow capture of Text that was in the input field when enter was pressed. If text matches a list item, that becomes selected and onChange is fired. ## What package(s) are you using? - [x] carbon-components-react-7.18.0.tgz - [x] carbon-components-10.18.0.tgz ## Detailed description > I can not determine if the user has pressed enter in the ComboBox input to get its value for the ID I'd used to identify the ComboBox if what the user typed is not in the list. > Is this issue related to a specific component? Yes, ComboBox > What did you expect to happen? What happened instead? What would you like to > see changed? I would like an onSubmit or onChange to report an event with the text containing the value of the ComboBox when the user has pressed enter for anything they typed, not just for matching items in the list. > What browser are you working in? Fails in Chrome and Firefox > What version of the Carbon Design System are you using? 10.18.0 for carbon-components
True
ComboBox doesn't support entry of Text not found in list items - ## ComboBox has onInputChange but no way to know when Enter is pressed ## Though documentation claims the ComboBox allows people to enter text not found in the list items, there is no notification that this has occurred. There should be an onSubmit method to allow capture of Text that was in the input field when enter was pressed. If text matches a list item, that becomes selected and onChange is fired. ## What package(s) are you using? - [x] carbon-components-react-7.18.0.tgz - [x] carbon-components-10.18.0.tgz ## Detailed description > I can not determine if the user has pressed enter in the ComboBox input to get its value for the ID I'd used to identify the ComboBox if what the user typed is not in the list. > Is this issue related to a specific component? Yes, ComboBox > What did you expect to happen? What happened instead? What would you like to > see changed? I would like an onSubmit or onChange to report an event with the text containing the value of the ComboBox when the user has pressed enter for anything they typed, not just for matching items in the list. > What browser are you working in? Fails in Chrome and Firefox > What version of the Carbon Design System are you using? 10.18.0 for carbon-components
main
combobox doesn t support entry of text not found in list items combobox has oninputchange but no way to know when enter is pressed though documentation claims the combobox allows people to enter text not found in the list items there is no notification that this has occurred there should be an onsubmit method to allow capture of text that was in the input field when enter was pressed if text matches a list item that becomes selected and onchange is fired what package s are you using carbon components react tgz carbon components tgz detailed description i can not determine if the user has pressed enter in the combobox input to get its value for the id i d used to identify the combobox if what the user typed is not in the list is this issue related to a specific component yes combobox what did you expect to happen what happened instead what would you like to see changed i would like an onsubmit or onchange to report an event with the text containing the value of the combobox when the user has pressed enter for anything they typed not just for matching items in the list what browser are you working in fails in chrome and firefox what version of the carbon design system are you using for carbon components
1
1,051
4,863,245,681
IssuesEvent
2016-11-14 14:58:33
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Unarchive module thinks existing local archive is nonexistent
affects_2.1 bug_report waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> Issue close to this one: #932 ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> - unarchive module ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.2.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> Nothing changed. ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> from Fedora 24 workstation, to RHEL 7.2 workstation ##### SUMMARY <!--- Explain the problem briefly --> Unarchive module is unable to unzip file located on remote machine (I am using remote_src=yes), failing with: > FAILED! => {"failed": true, "msg": "file or module does not exist: /home/cdk/cdk.zip"} When using default remote_src=no, module is able to unzip /home/agajdosi/cdk.zip on local machine and put it in /home/cdk/ on remote machine. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> 1. Put .zip file in remote machine. 2. Write unarchive task into playbook. 3. Run playbook. <!--- Paste example playbooks or commands between quotes below --> ``` --- - hosts: [rhel] remote_user: cdk tasks: - name: Unzip CDK zip unarchive: remote_src: yes src: /home/cdk/cdk.zip dest: /home/cdk/ ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Should unzip local cdk.zip file and extract it into home directory of user cdk. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Fails. <!--- Paste verbatim command output between quotes below --> ``` fatal: [foo]: FAILED! => {"failed": true, "msg": "file or module does not exist: /home/cdk/cdk.zip"} ```
True
Unarchive module thinks existing local archive is nonexistent - <!--- Verify first that your issue/request is not already reported in GitHub --> Issue close to this one: #932 ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> - unarchive module ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.2.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> Nothing changed. ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> from Fedora 24 workstation, to RHEL 7.2 workstation ##### SUMMARY <!--- Explain the problem briefly --> Unarchive module is unable to unzip file located on remote machine (I am using remote_src=yes), failing with: > FAILED! => {"failed": true, "msg": "file or module does not exist: /home/cdk/cdk.zip"} When using default remote_src=no, module is able to unzip /home/agajdosi/cdk.zip on local machine and put it in /home/cdk/ on remote machine. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> 1. Put .zip file in remote machine. 2. Write unarchive task into playbook. 3. Run playbook. <!--- Paste example playbooks or commands between quotes below --> ``` --- - hosts: [rhel] remote_user: cdk tasks: - name: Unzip CDK zip unarchive: remote_src: yes src: /home/cdk/cdk.zip dest: /home/cdk/ ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> Should unzip local cdk.zip file and extract it into home directory of user cdk. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Fails. <!--- Paste verbatim command output between quotes below --> ``` fatal: [foo]: FAILED! => {"failed": true, "msg": "file or module does not exist: /home/cdk/cdk.zip"} ```
main
unarchive module thinks existing local archive is nonexistent issue close to this one issue type bug report component name unarchive module ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables nothing changed os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific from fedora workstation to rhel workstation summary unarchive module is unable to unzip file located on remote machine i am using remote src yes failing with failed failed true msg file or module does not exist home cdk cdk zip when using default remote src no module is able to unzip home agajdosi cdk zip on local machine and put it in home cdk on remote machine steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used put zip file in remote machine write unarchive task into playbook run playbook hosts remote user cdk tasks name unzip cdk zip unarchive remote src yes src home cdk cdk zip dest home cdk expected results should unzip local cdk zip file and extract it into home directory of user cdk actual results fails fatal failed failed true msg file or module does not exist home cdk cdk zip
1
255,159
21,903,264,400
IssuesEvent
2022-05-20 15:19:38
ORNL-Fusion/xolotl
https://api.github.com/repos/ORNL-Fusion/xolotl
closed
Add tests for bursting and other processes
testing
Set the RNG seed to a given value to be able to reproduce the results.
1.0
Add tests for bursting and other processes - Set the RNG seed to a given value to be able to reproduce the results.
non_main
add tests for bursting and other processes set the rng seed to a given value to be able to reproduce the results
0
3,688
15,057,685,839
IssuesEvent
2021-02-03 22:05:24
IITIDIDX597/sp_2021_team2
https://api.github.com/repos/IITIDIDX597/sp_2021_team2
opened
Getting quickest supply of items
Maintaining Inventory Story
As a shared space manager I want to understand which supplier will be able to get an item quickest prior to placing an order So that kitchen users will have access to items as soon as possible
True
Getting quickest supply of items - As a shared space manager I want to understand which supplier will be able to get an item quickest prior to placing an order So that kitchen users will have access to items as soon as possible
main
getting quickest supply of items as a shared space manager i want to understand which supplier will be able to get an item quickest prior to placing an order so that kitchen users will have access to items as soon as possible
1
201,394
15,192,841,673
IssuesEvent
2021-02-15 23:01:50
istio/istio.io
https://api.github.com/repos/istio/istio.io
closed
Use of Pilot, Mixer, Galley
community/testing days
Is there a doc somewhere that says how the various components should be used in the current doc. For example, I see - Each Istio component (Pilot, Galley, Mixer) - Mixer, Pilot, and Galley all implement
1.0
Use of Pilot, Mixer, Galley - Is there a doc somewhere that says how the various components should be used in the current doc. For example, I see - Each Istio component (Pilot, Galley, Mixer) - Mixer, Pilot, and Galley all implement
non_main
use of pilot mixer galley is there a doc somewhere that says how the various components should be used in the current doc for example i see each istio component pilot galley mixer mixer pilot and galley all implement
0
2,399
8,518,507,200
IssuesEvent
2018-11-01 11:52:44
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Proper typing of beacons that dispense items
Maintainability/Hinders improvements
`/obj/item/musicbeacon` `/obj/item/herobeacon` `/obj/item/holybeacon` These all do literally the exact same thing in nearly the same way, certainly enough to require they be typed properly. `obj/item/beacon` is already taken so may need to be typed to `/obj/item/choice_beacon` with each of these three a subtype Putting this as an issue to remind myself and others should they want to do it for easy GBP.
True
Proper typing of beacons that dispense items - `/obj/item/musicbeacon` `/obj/item/herobeacon` `/obj/item/holybeacon` These all do literally the exact same thing in nearly the same way, certainly enough to require they be typed properly. `obj/item/beacon` is already taken so may need to be typed to `/obj/item/choice_beacon` with each of these three a subtype Putting this as an issue to remind myself and others should they want to do it for easy GBP.
main
proper typing of beacons that dispense items obj item musicbeacon obj item herobeacon obj item holybeacon these all do literally the exact same thing in nearly the same way certainly enough to require they be typed properly obj item beacon is already taken so may need to be typed to obj item choice beacon with each of these three a subtype putting this as an issue to remind myself and others should they want to do it for easy gbp
1
2,262
7,940,756,390
IssuesEvent
2018-07-10 00:15:08
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
Requesting access to the backdrop-contrib group (New Relic RPM module)
Maintainer application Port status
I would like access to the backdrop-contrib repository. I am a maintainer for the New Relic RPM module and i have ported it to backdrop. [Drupal module](https://www.drupal.org/project/new_relic_rpm) [Backdrop module](https://github.com/hosef/new_relic) I agree to all items in the Backdrop Contributed Project Agreement.
True
Requesting access to the backdrop-contrib group (New Relic RPM module) - I would like access to the backdrop-contrib repository. I am a maintainer for the New Relic RPM module and i have ported it to backdrop. [Drupal module](https://www.drupal.org/project/new_relic_rpm) [Backdrop module](https://github.com/hosef/new_relic) I agree to all items in the Backdrop Contributed Project Agreement.
main
requesting access to the backdrop contrib group new relic rpm module i would like access to the backdrop contrib repository i am a maintainer for the new relic rpm module and i have ported it to backdrop i agree to all items in the backdrop contributed project agreement
1
12,807
8,717,407,534
IssuesEvent
2018-12-07 17:02:51
istio/istio
https://api.github.com/repos/istio/istio
closed
Tool to run 'deep validation' for authN policy and related destination rule
area/security/aaa kind/enhancement stale
We may want to provide a tool for customers to verify authN policy setup (mainly focus on mTLS, due to the complexity of having authN policy and destination rule that control server and clients independently). The tool could be part of `istioctl`, or as a stand-alone script/binary. 1. Check for authentication policies *conflicts* (i.e if there are more than one policies with the same target scope). 1. Similar, check for destination rule *conflicts*. 1. Check if mTLS is enabled for a service, but destination rule for that service doesn't use TLS (and vice-versa). (Note: the first 2 items could be generic for all types of CRD-based config/policy)
True
Tool to run 'deep validation' for authN policy and related destination rule - We may want to provide a tool for customers to verify authN policy setup (mainly focus on mTLS, due to the complexity of having authN policy and destination rule that control server and clients independently). The tool could be part of `istioctl`, or as a stand-alone script/binary. 1. Check for authentication policies *conflicts* (i.e if there are more than one policies with the same target scope). 1. Similar, check for destination rule *conflicts*. 1. Check if mTLS is enabled for a service, but destination rule for that service doesn't use TLS (and vice-versa). (Note: the first 2 items could be generic for all types of CRD-based config/policy)
non_main
tool to run deep validation for authn policy and related destination rule we may want to provide a tool for customers to verify authn policy setup mainly focus on mtls due to the complexity of having authn policy and destination rule that control server and clients independently the tool could be part of istioctl or as a stand alone script binary check for authentication policies conflicts i e if there are more than one policies with the same target scope similar check for destination rule conflicts check if mtls is enabled for a service but destination rule for that service doesn t use tls and vice versa note the first items could be generic for all types of crd based config policy
0
69,677
15,025,333,168
IssuesEvent
2021-02-01 20:56:07
AlexRogalskiy/gradle-kotlin-sample
https://api.github.com/repos/AlexRogalskiy/gradle-kotlin-sample
opened
CVE-2015-7501 (High) detected in commons-collections-3.2.1.jar
security vulnerability
## CVE-2015-7501 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: gradle-kotlin-sample/buildSrc/build.gradle.kts</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-collections/commons-collections/3.2.1/761ea405b9b37ced573d2df0d1e3a4e0f9edc668/commons-collections-3.2.1.jar</p> <p> Dependency Hierarchy: - gradle-nexus-staging-plugin-0.22.0.jar (Root Library) - http-builder-0.7.1.jar - json-lib-2.3.pom - :x: **commons-collections-3.2.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/gradle-kotlin-sample/commit/e3765366fde896c6160e145c7bff772fafd08cfd">e3765366fde896c6160e145c7bff772fafd08cfd</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Red Hat JBoss A-MQ 6.x; BPM Suite (BPMS) 6.x; BRMS 6.x and 5.x; Data Grid (JDG) 6.x; Data Virtualization (JDV) 6.x and 5.x; Enterprise Application Platform 6.x, 5.x, and 4.3.x; Fuse 6.x; Fuse Service Works (FSW) 6.x; Operations Network (JBoss ON) 3.x; Portal 6.x; SOA Platform (SOA-P) 5.x; Web Server (JWS) 3.x; Red Hat OpenShift/xPAAS 3.x; and Red Hat Subscription Asset Manager 1.3 allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2017-11-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-7501>CVE-2015-7501</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1279330">https://bugzilla.redhat.com/show_bug.cgi?id=1279330</a></p> <p>Release Date: 2017-11-09</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2;org.apache.commons:commons-collections4:4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-7501 (High) detected in commons-collections-3.2.1.jar - ## CVE-2015-7501 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-collections-3.2.1.jar</b></p></summary> <p>Types that extend and augment the Java Collections Framework.</p> <p>Path to dependency file: gradle-kotlin-sample/buildSrc/build.gradle.kts</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/commons-collections/commons-collections/3.2.1/761ea405b9b37ced573d2df0d1e3a4e0f9edc668/commons-collections-3.2.1.jar</p> <p> Dependency Hierarchy: - gradle-nexus-staging-plugin-0.22.0.jar (Root Library) - http-builder-0.7.1.jar - json-lib-2.3.pom - :x: **commons-collections-3.2.1.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/gradle-kotlin-sample/commit/e3765366fde896c6160e145c7bff772fafd08cfd">e3765366fde896c6160e145c7bff772fafd08cfd</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Red Hat JBoss A-MQ 6.x; BPM Suite (BPMS) 6.x; BRMS 6.x and 5.x; Data Grid (JDG) 6.x; Data Virtualization (JDV) 6.x and 5.x; Enterprise Application Platform 6.x, 5.x, and 4.3.x; Fuse 6.x; Fuse Service Works (FSW) 6.x; Operations Network (JBoss ON) 3.x; Portal 6.x; SOA Platform (SOA-P) 5.x; Web Server (JWS) 3.x; Red Hat OpenShift/xPAAS 3.x; and Red Hat Subscription Asset Manager 1.3 allow remote attackers to execute arbitrary commands via a crafted serialized Java object, related to the Apache Commons Collections (ACC) library. <p>Publish Date: 2017-11-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-7501>CVE-2015-7501</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=1279330">https://bugzilla.redhat.com/show_bug.cgi?id=1279330</a></p> <p>Release Date: 2017-11-09</p> <p>Fix Resolution: commons-collections:commons-collections:3.2.2;org.apache.commons:commons-collections4:4.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in commons collections jar cve high severity vulnerability vulnerable library commons collections jar types that extend and augment the java collections framework path to dependency file gradle kotlin sample buildsrc build gradle kts path to vulnerable library home wss scanner gradle caches modules files commons collections commons collections commons collections jar dependency hierarchy gradle nexus staging plugin jar root library http builder jar json lib pom x commons collections jar vulnerable library found in head commit a href found in base branch master vulnerability details red hat jboss a mq x bpm suite bpms x brms x and x data grid jdg x data virtualization jdv x and x enterprise application platform x x and x fuse x fuse service works fsw x operations network jboss on x portal x soa platform soa p x web server jws x red hat openshift xpaas x and red hat subscription asset manager allow remote attackers to execute arbitrary commands via a crafted serialized java object related to the apache commons collections acc library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution commons collections commons collections org apache commons commons step up your open source security game with whitesource
0
264,723
20,030,245,205
IssuesEvent
2022-02-02 04:19:23
endojs/endo
https://api.github.com/repos/endojs/endo
closed
SES / endo security policy documentation out of sync with github, hackerone
documentation endo
https://hackerone.com/agoric?type=team refers to https://github.com/Agoric/ses-shim ; that redirects to https://github.com/endojs/endo , which is good, but SECURITY.md is under packages/ses , so github doesn't treat it as the security policy for the repository. Also, SECURITY.md doesn't mention hackerone. cc @jessysaurusrex
1.0
SES / endo security policy documentation out of sync with github, hackerone - https://hackerone.com/agoric?type=team refers to https://github.com/Agoric/ses-shim ; that redirects to https://github.com/endojs/endo , which is good, but SECURITY.md is under packages/ses , so github doesn't treat it as the security policy for the repository. Also, SECURITY.md doesn't mention hackerone. cc @jessysaurusrex
non_main
ses endo security policy documentation out of sync with github hackerone refers to that redirects to which is good but security md is under packages ses so github doesn t treat it as the security policy for the repository also security md doesn t mention hackerone cc jessysaurusrex
0
1,979
6,694,191,273
IssuesEvent
2017-10-10 00:12:44
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Instant Answers on Beta: Display scrolls back to the first tile under the time frame
Maintainer Input Requested
If we click on an IA tile, the details are of that IA are shown and the display scrolls back to the first tile for that time frame. For example, if I click 'Political Polls' which is the last IA under 'Last 30 Days' currently, the display scrolls back to 'Ace Editor Cheat Sheet'. Was this intended? People to notify: @jdorweiler --- IA Page: http://duck.co/ia/view/experimentalia [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @vikashvverma
True
Instant Answers on Beta: Display scrolls back to the first tile under the time frame - If we click on an IA tile, the details are of that IA are shown and the display scrolls back to the first tile for that time frame. For example, if I click 'Political Polls' which is the last IA under 'Last 30 Days' currently, the display scrolls back to 'Ace Editor Cheat Sheet'. Was this intended? People to notify: @jdorweiler --- IA Page: http://duck.co/ia/view/experimentalia [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @vikashvverma
main
instant answers on beta display scrolls back to the first tile under the time frame if we click on an ia tile the details are of that ia are shown and the display scrolls back to the first tile for that time frame for example if i click political polls which is the last ia under last days currently the display scrolls back to ace editor cheat sheet was this intended people to notify jdorweiler ia page vikashvverma
1
37,629
6,625,171,918
IssuesEvent
2017-09-22 14:31:07
FlinkML/flink-jpmml
https://api.github.com/repos/FlinkML/flink-jpmml
opened
Add maven libraries version badges
Documentation enhancement
Add maven libraries version badges to show on README the last published versions for both scala 2.10 and 2.11
1.0
Add maven libraries version badges - Add maven libraries version badges to show on README the last published versions for both scala 2.10 and 2.11
non_main
add maven libraries version badges add maven libraries version badges to show on readme the last published versions for both scala and
0
76,216
14,582,295,762
IssuesEvent
2020-12-18 12:08:18
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
Crash, likely bot speech capitalization error.
Bug Code Crash
*Sent by WordPress (issue@barotraumagame.com). Created by [fire](https://fire.fundersclub.com/).* --- From: Crazy Ivan **Description** >From the log, I think maybe a bot tried to say something. I was playing with 3 other bots. **Steps To Reproduce** **Version** 0.11.0.9 **Operating System** Mac -- This issue was sent from a contact form on https://barotraumagame.com/bug --- *Attachments:* 1. [servercrashreport-3.log](https://firebot-prod-media.s3.amazonaws.com:443/email-attachments/2bec0b53-f861-4fbf-b635-802a1601b1e8/servercrashreport-3.log)
1.0
Crash, likely bot speech capitalization error. - *Sent by WordPress (issue@barotraumagame.com). Created by [fire](https://fire.fundersclub.com/).* --- From: Crazy Ivan **Description** >From the log, I think maybe a bot tried to say something. I was playing with 3 other bots. **Steps To Reproduce** **Version** 0.11.0.9 **Operating System** Mac -- This issue was sent from a contact form on https://barotraumagame.com/bug --- *Attachments:* 1. [servercrashreport-3.log](https://firebot-prod-media.s3.amazonaws.com:443/email-attachments/2bec0b53-f861-4fbf-b635-802a1601b1e8/servercrashreport-3.log)
non_main
crash likely bot speech capitalization error sent by wordpress issue barotraumagame com created by from crazy ivan description from the log i think maybe a bot tried to say something i was playing with other bots steps to reproduce version operating system mac this issue was sent from a contact form on attachments
0
2,965
10,618,705,144
IssuesEvent
2019-10-13 07:11:30
arcticicestudio/nord-docs
https://api.github.com/repos/arcticicestudio/nord-docs
opened
Dependency Update April-October 2019
scope-compatibility scope-maintainability scope-performance scope-quality scope-security scope-stability type-task
This is the regular batch update for outdated dependencies. The following testing related packages have also been deprecated in favor of new (scoped) packages: - `eslint-config-arcticicestudio` ➔ `@arcticicestudio/eslint-config` - `gatsby-mdx` ➔ `gatsby-plugin-mdx` - `jest-dom` ➔ `@testing-library/jest-dom` - `react-hooks-testing-library` ➔ `@testing-library/react-hooks` - `react-testing-library` ➔ `@testing-library/react` ###### Production Dependencies - @mdx-js/react `1.0.6` ➔ `1.5.1` - axios `0.18.0` ➔ `0.19.0` - body-scroll-lock `2.6.1` ➔ `2.6.4` - date-fns `2.0.0-alpha.27` ➔ `2.4.1` - gatsby `2.3.29` ➔ `2.15.36` - gatsby-image `2.0.40` ➔ `2.2.27` - ~~gatsby-mdx `0.6.2`~~ ➔ gatsby-plugin-mdx `1.0.51` - gatsby-plugin-canonical-urls `2.0.12` ➔ `2.1.11` - gatsby-plugin-catch-links `2.0.13` ➔ `2.1.13` - gatsby-plugin-google-gtag `1.0.16` ➔ `1.1.12` - gatsby-plugin-lodash `3.0.5` ➔ `3.1.11` - gatsby-plugin-manifest `2.0.29` ➔ `2.2.21` - gatsby-plugin-netlify `2.0.15` ➔ 2.1.19`` - gatsby-plugin-no-sourcemaps `2.0.2` ➔ `2.1.1` - gatsby-plugin-offline `2.0.25` ➔ `3.0.14` - gatsby-plugin-react-helmet `3.0.12` ➔ `3.1.11` - gatsby-plugin-remove-trailing-slashes `2.0.11` ➔ `2.1.10` - gatsby-plugin-robots-txt `1.4.0` ➔ `1.5.0` - gatsby-plugin-sharp `2.0.35` ➔ `2.2.29` - gatsby-plugin-sitemap `2.0.12` ➔ `2.2.17` - gatsby-plugin-styled-components `3.0.7` ➔ `3.1.9` - gatsby-plugin-webpack-size `0.0.3` ➔ `1.0.0` - gatsby-source-filesystem `2.0.32` ➔ `2.1.31` - gatsby-source-graphql `2.0.18` ➔ `2.1.18` - gatsby-transformer-sharp `2.1.18` ➔ `2.2.21` - gatsby-transformer-yaml `2.1.12` ➔ `2.2.13` - inter-ui `3.5.0` ➔ `3.10.0` - lodash `4.17.11` ➔ `4.17.15` - polished `3.2.0` ➔ `3.4.1` - react `16.8.6` ➔ `16.10.2` - react-dom `16.8.6` ➔ `16.10.2` - react-helmet `5.2.0` ➔ `5.2.1` - react-pose `4.0.8` ➔ `4.0.9` - react-spring `8.0.7` ➔ `8.0.27` - semver `6.0.0` ➔ `6.3.0` - styled-components `4.2.0` ➔ `4.4.0` ###### Development Dependencies - @babel/core `7.4.3` ➔ `7.6.4` - @babel/plugin-proposal-class-properties `7.4.0` ➔ `7.5.5` - @babel/plugin-proposal-export-default-from `7.2.0` ➔ `7.5.2` - @babel/plugin-proposal-nullish-coalescing-operator `7.4.3` ➔ `7.4.4` - @babel/plugin-proposal-optional-chaining `7.2.0` ➔ `7.6.0` - @mdx-js/mdx `1.0.14` ➔ `1.5.1` - @svgr/webpack `4.2.0` ➔ `4.3.3` - babel-eslint `10.0.1` ➔ `10.0.3` - babel-jest `24.7.1` ➔ `24.9.0` - babel-plugin-styled-components `1.10.0` ➔ `1.10.6` - babel-preset-gatsby `0.1.11` ➔ `0.2.18` - del-cli `1.1.0` ➔ `3.0.0` - eslint `5.16.0` ➔ `6.5.1` - ~~eslint-config-arcticicestudio `>=0.4.0 <1.0.0`~~ ➔ @arcticicestudio/eslint-config `>=0.8.0 <1.0.0` - eslint-import-resolver-jest `2.1.1` ➔ `3.0.0` - eslint-plugin-import `2.17.2` ➔ `2.18.2` - eslint-plugin-jsx-a11y `6.2.1` ➔ `6.2.3` - eslint-plugin-prettier `3.0.1` ➔ `3.1.1` - eslint-plugin-react `7.12.4` ➔ `7.16.0` - eslint-plugin-react-hooks `1.6.0` ➔ `2.1.2` - git-revision-webpack-plugin `3.0.3` ➔ `3.0.4` - glob `7.1.3` ➔ `7.1.4` - husky `2.1.0` ➔ `3.0.9` - jest `24.7.1` ➔ `24.9.0` - ~~jest-dom `3.1.3`~~ ➔ @testing-library/jest-dom `4.1.2` - jest-junit `6.3.0` ➔ `8.0.0` - jest-styled-components `6.3.1` ➔ `6.3.3` - lint-staged `8.1.5` ➔ `9.4.2` - prettier `1.17.0` ➔ `1.18.2` - rehype-slug `2.0.2` ➔ `2.0.3` - ~~react-hooks-testing-library `0.5.0`~~ ➔ @testing-library/react `9.3.0` - ~~react-testing-library `6.1.2`~~ ➔ @testing-library/react-hooks `2.0.3` - remark-breaks `1.0.2` ➔ `1.0.3` - remark-cli `6.0.1` ➔ `7.0.0` - remark-github `7.0.6` ➔ `8.0.0` - remark-preset-lint-arcticicestudio `>=0.2.0 <1.0.0` ➔ `>=0.3.0 <1.0.0` - webpack-bundle-analyzer `3.3.2` ➔ `3.5.2`
True
Dependency Update April-October 2019 - This is the regular batch update for outdated dependencies. The following testing related packages have also been deprecated in favor of new (scoped) packages: - `eslint-config-arcticicestudio` ➔ `@arcticicestudio/eslint-config` - `gatsby-mdx` ➔ `gatsby-plugin-mdx` - `jest-dom` ➔ `@testing-library/jest-dom` - `react-hooks-testing-library` ➔ `@testing-library/react-hooks` - `react-testing-library` ➔ `@testing-library/react` ###### Production Dependencies - @mdx-js/react `1.0.6` ➔ `1.5.1` - axios `0.18.0` ➔ `0.19.0` - body-scroll-lock `2.6.1` ➔ `2.6.4` - date-fns `2.0.0-alpha.27` ➔ `2.4.1` - gatsby `2.3.29` ➔ `2.15.36` - gatsby-image `2.0.40` ➔ `2.2.27` - ~~gatsby-mdx `0.6.2`~~ ➔ gatsby-plugin-mdx `1.0.51` - gatsby-plugin-canonical-urls `2.0.12` ➔ `2.1.11` - gatsby-plugin-catch-links `2.0.13` ➔ `2.1.13` - gatsby-plugin-google-gtag `1.0.16` ➔ `1.1.12` - gatsby-plugin-lodash `3.0.5` ➔ `3.1.11` - gatsby-plugin-manifest `2.0.29` ➔ `2.2.21` - gatsby-plugin-netlify `2.0.15` ➔ 2.1.19`` - gatsby-plugin-no-sourcemaps `2.0.2` ➔ `2.1.1` - gatsby-plugin-offline `2.0.25` ➔ `3.0.14` - gatsby-plugin-react-helmet `3.0.12` ➔ `3.1.11` - gatsby-plugin-remove-trailing-slashes `2.0.11` ➔ `2.1.10` - gatsby-plugin-robots-txt `1.4.0` ➔ `1.5.0` - gatsby-plugin-sharp `2.0.35` ➔ `2.2.29` - gatsby-plugin-sitemap `2.0.12` ➔ `2.2.17` - gatsby-plugin-styled-components `3.0.7` ➔ `3.1.9` - gatsby-plugin-webpack-size `0.0.3` ➔ `1.0.0` - gatsby-source-filesystem `2.0.32` ➔ `2.1.31` - gatsby-source-graphql `2.0.18` ➔ `2.1.18` - gatsby-transformer-sharp `2.1.18` ➔ `2.2.21` - gatsby-transformer-yaml `2.1.12` ➔ `2.2.13` - inter-ui `3.5.0` ➔ `3.10.0` - lodash `4.17.11` ➔ `4.17.15` - polished `3.2.0` ➔ `3.4.1` - react `16.8.6` ➔ `16.10.2` - react-dom `16.8.6` ➔ `16.10.2` - react-helmet `5.2.0` ➔ `5.2.1` - react-pose `4.0.8` ➔ `4.0.9` - react-spring `8.0.7` ➔ `8.0.27` - semver `6.0.0` ➔ `6.3.0` - styled-components `4.2.0` ➔ `4.4.0` ###### Development Dependencies - @babel/core `7.4.3` ➔ `7.6.4` - @babel/plugin-proposal-class-properties `7.4.0` ➔ `7.5.5` - @babel/plugin-proposal-export-default-from `7.2.0` ➔ `7.5.2` - @babel/plugin-proposal-nullish-coalescing-operator `7.4.3` ➔ `7.4.4` - @babel/plugin-proposal-optional-chaining `7.2.0` ➔ `7.6.0` - @mdx-js/mdx `1.0.14` ➔ `1.5.1` - @svgr/webpack `4.2.0` ➔ `4.3.3` - babel-eslint `10.0.1` ➔ `10.0.3` - babel-jest `24.7.1` ➔ `24.9.0` - babel-plugin-styled-components `1.10.0` ➔ `1.10.6` - babel-preset-gatsby `0.1.11` ➔ `0.2.18` - del-cli `1.1.0` ➔ `3.0.0` - eslint `5.16.0` ➔ `6.5.1` - ~~eslint-config-arcticicestudio `>=0.4.0 <1.0.0`~~ ➔ @arcticicestudio/eslint-config `>=0.8.0 <1.0.0` - eslint-import-resolver-jest `2.1.1` ➔ `3.0.0` - eslint-plugin-import `2.17.2` ➔ `2.18.2` - eslint-plugin-jsx-a11y `6.2.1` ➔ `6.2.3` - eslint-plugin-prettier `3.0.1` ➔ `3.1.1` - eslint-plugin-react `7.12.4` ➔ `7.16.0` - eslint-plugin-react-hooks `1.6.0` ➔ `2.1.2` - git-revision-webpack-plugin `3.0.3` ➔ `3.0.4` - glob `7.1.3` ➔ `7.1.4` - husky `2.1.0` ➔ `3.0.9` - jest `24.7.1` ➔ `24.9.0` - ~~jest-dom `3.1.3`~~ ➔ @testing-library/jest-dom `4.1.2` - jest-junit `6.3.0` ➔ `8.0.0` - jest-styled-components `6.3.1` ➔ `6.3.3` - lint-staged `8.1.5` ➔ `9.4.2` - prettier `1.17.0` ➔ `1.18.2` - rehype-slug `2.0.2` ➔ `2.0.3` - ~~react-hooks-testing-library `0.5.0`~~ ➔ @testing-library/react `9.3.0` - ~~react-testing-library `6.1.2`~~ ➔ @testing-library/react-hooks `2.0.3` - remark-breaks `1.0.2` ➔ `1.0.3` - remark-cli `6.0.1` ➔ `7.0.0` - remark-github `7.0.6` ➔ `8.0.0` - remark-preset-lint-arcticicestudio `>=0.2.0 <1.0.0` ➔ `>=0.3.0 <1.0.0` - webpack-bundle-analyzer `3.3.2` ➔ `3.5.2`
main
dependency update april october this is the regular batch update for outdated dependencies the following testing related packages have also been deprecated in favor of new scoped packages eslint config arcticicestudio ➔ arcticicestudio eslint config gatsby mdx ➔ gatsby plugin mdx jest dom ➔ testing library jest dom react hooks testing library ➔ testing library react hooks react testing library ➔ testing library react production dependencies mdx js react ➔ axios ➔ body scroll lock ➔ date fns alpha ➔ gatsby ➔ gatsby image ➔ gatsby mdx ➔ gatsby plugin mdx gatsby plugin canonical urls ➔ gatsby plugin catch links ➔ gatsby plugin google gtag ➔ gatsby plugin lodash ➔ gatsby plugin manifest ➔ gatsby plugin netlify ➔ gatsby plugin no sourcemaps ➔ gatsby plugin offline ➔ gatsby plugin react helmet ➔ gatsby plugin remove trailing slashes ➔ gatsby plugin robots txt ➔ gatsby plugin sharp ➔ gatsby plugin sitemap ➔ gatsby plugin styled components ➔ gatsby plugin webpack size ➔ gatsby source filesystem ➔ gatsby source graphql ➔ gatsby transformer sharp ➔ gatsby transformer yaml ➔ inter ui ➔ lodash ➔ polished ➔ react ➔ react dom ➔ react helmet ➔ react pose ➔ react spring ➔ semver ➔ styled components ➔ development dependencies babel core ➔ babel plugin proposal class properties ➔ babel plugin proposal export default from ➔ babel plugin proposal nullish coalescing operator ➔ babel plugin proposal optional chaining ➔ mdx js mdx ➔ svgr webpack ➔ babel eslint ➔ babel jest ➔ babel plugin styled components ➔ babel preset gatsby ➔ del cli ➔ eslint ➔ eslint config arcticicestudio eslint import resolver jest ➔ eslint plugin import ➔ eslint plugin jsx ➔ eslint plugin prettier ➔ eslint plugin react ➔ eslint plugin react hooks ➔ git revision webpack plugin ➔ glob ➔ husky ➔ jest ➔ jest dom ➔ testing library jest dom jest junit ➔ jest styled components ➔ lint staged ➔ prettier ➔ rehype slug ➔ react hooks testing library ➔ testing library react react testing library ➔ testing library react hooks remark breaks ➔ remark cli ➔ remark github ➔ remark preset lint arcticicestudio webpack bundle analyzer ➔
1
5,281
26,683,320,785
IssuesEvent
2023-01-26 19:35:39
ipfs/ipfs-docs
https://api.github.com/repos/ipfs/ipfs-docs
opened
Add a "report abuse" resource to the docs somewhere
kind/enhancement need/maintainers-input
As discussed in this thread: https://filecoinproject.slack.com/archives/C04ENJL5S84/p1674748260526859 Currently, every now and then we'll get an issue opened in this repo reporting some kind of abuse of our CoC or ToS, and asking for instruction on what to do about it. It would be useful to have a very visible resource regarding how to report abuse, and also link to the community code of conduct, this may also reduce the number of these issues that get opened in this repo. We may also want to provide a link to these: https://discuss.ipfs.tech/tos and https://protocol.ai/legal/#terms-of-service
True
Add a "report abuse" resource to the docs somewhere - As discussed in this thread: https://filecoinproject.slack.com/archives/C04ENJL5S84/p1674748260526859 Currently, every now and then we'll get an issue opened in this repo reporting some kind of abuse of our CoC or ToS, and asking for instruction on what to do about it. It would be useful to have a very visible resource regarding how to report abuse, and also link to the community code of conduct, this may also reduce the number of these issues that get opened in this repo. We may also want to provide a link to these: https://discuss.ipfs.tech/tos and https://protocol.ai/legal/#terms-of-service
main
add a report abuse resource to the docs somewhere as discussed in this thread currently every now and then we ll get an issue opened in this repo reporting some kind of abuse of our coc or tos and asking for instruction on what to do about it it would be useful to have a very visible resource regarding how to report abuse and also link to the community code of conduct this may also reduce the number of these issues that get opened in this repo we may also want to provide a link to these and
1
64,819
6,923,913,817
IssuesEvent
2017-11-30 10:43:47
ocadotechnology/aimmo
https://api.github.com/repos/ocadotechnology/aimmo
closed
Integration Testing
Epic testing
We need some integration testing. My idea was to create a mock django server that would capture all request and responses and potentially deploy the server to an actual cluster. The idea seems to work locally with threads and minikube(not finished yet) -- you can see the work on branch "tests".
1.0
Integration Testing - We need some integration testing. My idea was to create a mock django server that would capture all request and responses and potentially deploy the server to an actual cluster. The idea seems to work locally with threads and minikube(not finished yet) -- you can see the work on branch "tests".
non_main
integration testing we need some integration testing my idea was to create a mock django server that would capture all request and responses and potentially deploy the server to an actual cluster the idea seems to work locally with threads and minikube not finished yet you can see the work on branch tests
0
786,562
27,658,844,152
IssuesEvent
2023-03-12 09:33:49
nabu-catalog/nabu
https://api.github.com/repos/nabu-catalog/nabu
closed
Enable gzip for static resources in nginx configuration
Priority Major Urgent (#1) 2019 Priorities
Hi, As per title, please enable gzip'ing of static resources in the nginx configuration to speed up content serving. Following is a browser console screenshot showing that the response is not zipped. ![screen shot 2018-04-11 at 9 46 06 am](https://user-images.githubusercontent.com/2639995/38589267-598337c0-3d6d-11e8-964f-102d3fb1a848.png) An example is @ https://gist.github.com/rcanepa/0b9beaac137f7d45754b58822bfd0092 Or if the nginx version is new enough it looks like there's a module that can be compiled / loaded to figure it all out automatically (I've not used it so YMMV): https://nginx.org/en/docs/http/ngx_http_gzip_static_module.html Thanks!
1.0
Enable gzip for static resources in nginx configuration - Hi, As per title, please enable gzip'ing of static resources in the nginx configuration to speed up content serving. Following is a browser console screenshot showing that the response is not zipped. ![screen shot 2018-04-11 at 9 46 06 am](https://user-images.githubusercontent.com/2639995/38589267-598337c0-3d6d-11e8-964f-102d3fb1a848.png) An example is @ https://gist.github.com/rcanepa/0b9beaac137f7d45754b58822bfd0092 Or if the nginx version is new enough it looks like there's a module that can be compiled / loaded to figure it all out automatically (I've not used it so YMMV): https://nginx.org/en/docs/http/ngx_http_gzip_static_module.html Thanks!
non_main
enable gzip for static resources in nginx configuration hi as per title please enable gzip ing of static resources in the nginx configuration to speed up content serving following is a browser console screenshot showing that the response is not zipped an example is or if the nginx version is new enough it looks like there s a module that can be compiled loaded to figure it all out automatically i ve not used it so ymmv thanks
0
58,120
16,342,528,793
IssuesEvent
2021-05-13 00:31:03
darshan-hpc/darshan
https://api.github.com/repos/darshan-hpc/darshan
closed
MPIIO_F_WRITE_START_TIMESTAMP counter incorrect
defect
In GitLab by @carns on Oct 5, 2016, 09:08 Reported by William Yoo. This counter doesn't match the POSIX level write start timestamp or system-level instrumentation for a VPIC benchmark that does collective writes via HDF5.
1.0
MPIIO_F_WRITE_START_TIMESTAMP counter incorrect - In GitLab by @carns on Oct 5, 2016, 09:08 Reported by William Yoo. This counter doesn't match the POSIX level write start timestamp or system-level instrumentation for a VPIC benchmark that does collective writes via HDF5.
non_main
mpiio f write start timestamp counter incorrect in gitlab by carns on oct reported by william yoo this counter doesn t match the posix level write start timestamp or system level instrumentation for a vpic benchmark that does collective writes via
0
1,107
4,981,804,283
IssuesEvent
2016-12-07 09:18:52
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Floor tile system limits ability to add new colors and patterns
Maintainability - Hinders improvements - Not a bug Sprites
As-is, if you want to add a new floor tile color with all the directions and such, you need 13 sprites, four for each half-colored tile, four for the three-quarter colored tiles, four for the corner colored tiles, and one for the full-colored tile. And if you want the color to have a checkerboard pattern, that's two more sprites. And that's just for one base tile color, so if you want that tile color on white or black tiles, you've got to make that many more sprites just to add one color. This significantly limits the ability to add more colors and patterns, as well as bloating the floor.dmi excessively due to how many sprites it adds to make a single additional color.
True
Floor tile system limits ability to add new colors and patterns - As-is, if you want to add a new floor tile color with all the directions and such, you need 13 sprites, four for each half-colored tile, four for the three-quarter colored tiles, four for the corner colored tiles, and one for the full-colored tile. And if you want the color to have a checkerboard pattern, that's two more sprites. And that's just for one base tile color, so if you want that tile color on white or black tiles, you've got to make that many more sprites just to add one color. This significantly limits the ability to add more colors and patterns, as well as bloating the floor.dmi excessively due to how many sprites it adds to make a single additional color.
main
floor tile system limits ability to add new colors and patterns as is if you want to add a new floor tile color with all the directions and such you need sprites four for each half colored tile four for the three quarter colored tiles four for the corner colored tiles and one for the full colored tile and if you want the color to have a checkerboard pattern that s two more sprites and that s just for one base tile color so if you want that tile color on white or black tiles you ve got to make that many more sprites just to add one color this significantly limits the ability to add more colors and patterns as well as bloating the floor dmi excessively due to how many sprites it adds to make a single additional color
1
2,498
8,655,458,078
IssuesEvent
2018-11-27 16:00:19
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
Cannot connect to PSTV
unmaintained
As titled. Rather than necro bumping another thread I choose to open this new one. I'm running Windows 10 x64 and PSTV (FW 3.57) just can't find qcma. I checked Windows firewall and qcma is not blocked by it. PSTV is connected to LAN by Ethernet, and PSTV can go online like PSN or web browser.
True
Cannot connect to PSTV - As titled. Rather than necro bumping another thread I choose to open this new one. I'm running Windows 10 x64 and PSTV (FW 3.57) just can't find qcma. I checked Windows firewall and qcma is not blocked by it. PSTV is connected to LAN by Ethernet, and PSTV can go online like PSN or web browser.
main
cannot connect to pstv as titled rather than necro bumping another thread i choose to open this new one i m running windows and pstv fw just can t find qcma i checked windows firewall and qcma is not blocked by it pstv is connected to lan by ethernet and pstv can go online like psn or web browser
1
5,315
26,821,766,949
IssuesEvent
2023-02-02 09:59:21
toolbx-images/images
https://api.github.com/repos/toolbx-images/images
closed
Request for Amazon Linux based images
new-image-request maintainers-wanted
### Distribution name and versions requested Amazon Linux 2 & Amazon Linux 2022 https://github.com/thimslugga/amazonlinux-toolbx ### Where are the official container images from the distribution published? FROM public.ecr.aws/amazonlinux/amazonlinux:2 FROM public.ecr.aws/amazonlinux/amazonlinux:2022 ### Will you be interested in maintaining this image? Yes, I’ll make a pull request with work I’ve already done
True
Request for Amazon Linux based images - ### Distribution name and versions requested Amazon Linux 2 & Amazon Linux 2022 https://github.com/thimslugga/amazonlinux-toolbx ### Where are the official container images from the distribution published? FROM public.ecr.aws/amazonlinux/amazonlinux:2 FROM public.ecr.aws/amazonlinux/amazonlinux:2022 ### Will you be interested in maintaining this image? Yes, I’ll make a pull request with work I’ve already done
main
request for amazon linux based images distribution name and versions requested amazon linux amazon linux where are the official container images from the distribution published from public ecr aws amazonlinux amazonlinux from public ecr aws amazonlinux amazonlinux will you be interested in maintaining this image yes i’ll make a pull request with work i’ve already done
1
5,184
26,394,234,889
IssuesEvent
2023-01-12 17:54:56
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFP - kite
Status: Available For Maintainer(s)
## Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://www.kite.com Direct download URL for the software / installer : https://alpha.kite.com/release/dls/windows/current Software summary / short description: AI-Powered code completion <!-- ## Package Expectations Install without e-mail for all editors. -->
True
RFP - kite - ## Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://www.kite.com Direct download URL for the software / installer : https://alpha.kite.com/release/dls/windows/current Software summary / short description: AI-Powered code completion <!-- ## Package Expectations Install without e-mail for all editors. -->
main
rfp kite checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts with rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description ai powered code completion package expectations install without e mail for all editors
1
151,031
13,388,362,292
IssuesEvent
2020-09-02 17:15:38
kubesphere/kubesphere
https://api.github.com/repos/kubesphere/kubesphere
closed
2.0.2 升级2.1 出现coredns 配置丢失
area/installation kind/documentation stale
**问题描述** 从2.0.2 离线版升级 2.1 离线版 后,出现coredns 配置丢失。 **安装环境的硬件配置** easystack 云主机 **Installer版本** 2.0.2 >> 2.1
1.0
2.0.2 升级2.1 出现coredns 配置丢失 - **问题描述** 从2.0.2 离线版升级 2.1 离线版 后,出现coredns 配置丢失。 **安装环境的硬件配置** easystack 云主机 **Installer版本** 2.0.2 >> 2.1
non_main
出现coredns 配置丢失 问题描述 离线版升级 离线版 后,出现coredns 配置丢失。 安装环境的硬件配置 easystack 云主机 installer版本
0
5,095
26,007,472,462
IssuesEvent
2022-12-20 20:56:11
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
[Feature Request] Cognito Offline Testing
type/feature area/local/start-api stage/pm-review maintainer/need-followup
Just curious whether SAM will support cognito offline testing? Pasting in some links to the same feature request in various other forums: - https://github.com/localstack/localstack/issues/26 - https://stackoverflow.com/questions/41633800/use-cognito-locally-offline - https://stackoverflow.com/questions/51719765/invoke-lambda-function-without-authorization-on-local-serverless-offline-environ
True
[Feature Request] Cognito Offline Testing - Just curious whether SAM will support cognito offline testing? Pasting in some links to the same feature request in various other forums: - https://github.com/localstack/localstack/issues/26 - https://stackoverflow.com/questions/41633800/use-cognito-locally-offline - https://stackoverflow.com/questions/51719765/invoke-lambda-function-without-authorization-on-local-serverless-offline-environ
main
cognito offline testing just curious whether sam will support cognito offline testing pasting in some links to the same feature request in various other forums
1
35,742
5,005,664,868
IssuesEvent
2016-12-12 11:28:49
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
[Upgrade test] "ScheduledJob should replace jobs when ReplaceConcurrent" in ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster
kind/upgrade-test-failure non-release-blocker
Test has been failing [100% of the time](https://k8s-testgrid.appspot.com/google-1.4-1.5-upgrade#gce-cvm-1.4-cvm-1.5-upgrade-cluster&width=20&sort-by-failures=&include-filter-by-regex=ScheduledJob%20should%20replace%20jobs%20when%20ReplaceConcurrent) in [ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster](https://k8s-gubernator.appspot.com/builds/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/). Sample failure: [link](https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/4) Based on spreadsheet tracking 1.5 upgrade test failures created by @krousey.
1.0
[Upgrade test] "ScheduledJob should replace jobs when ReplaceConcurrent" in ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster - Test has been failing [100% of the time](https://k8s-testgrid.appspot.com/google-1.4-1.5-upgrade#gce-cvm-1.4-cvm-1.5-upgrade-cluster&width=20&sort-by-failures=&include-filter-by-regex=ScheduledJob%20should%20replace%20jobs%20when%20ReplaceConcurrent) in [ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster](https://k8s-gubernator.appspot.com/builds/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/). Sample failure: [link](https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gce-1.4-1.5-upgrade-cluster/4) Based on spreadsheet tracking 1.5 upgrade test failures created by @krousey.
non_main
scheduledjob should replace jobs when replaceconcurrent in ci kubernetes gce upgrade cluster test has been failing in sample failure based on spreadsheet tracking upgrade test failures created by krousey
0
344,521
10,345,888,173
IssuesEvent
2019-09-04 14:18:26
GarkGarcia/icon_baker
https://api.github.com/repos/GarkGarcia/icon_baker
closed
IconBaker's Svg renderer
priority
Follow up to https://github.com/GarkGarcia/icon-pie/issues/15. As you can see in [`resvg`](https://github.com/GarkGarcia/icon_baker/tree/resvg), I've ported the `svg` functionalities from `nsvg` to `resvg`. Everything seems to be working fine. However, I'm (unfortunately) using a Windows machine and I was not able to install `cairo`. This makes it hard for me to debug the library. For now, I'm using `wsl` to debug, but help on installing `cairo` would be appreciated (since the [official guides](https://www.cairographics.org/download/) where not able to help me much).
1.0
IconBaker's Svg renderer - Follow up to https://github.com/GarkGarcia/icon-pie/issues/15. As you can see in [`resvg`](https://github.com/GarkGarcia/icon_baker/tree/resvg), I've ported the `svg` functionalities from `nsvg` to `resvg`. Everything seems to be working fine. However, I'm (unfortunately) using a Windows machine and I was not able to install `cairo`. This makes it hard for me to debug the library. For now, I'm using `wsl` to debug, but help on installing `cairo` would be appreciated (since the [official guides](https://www.cairographics.org/download/) where not able to help me much).
non_main
iconbaker s svg renderer follow up to as you can see in i ve ported the svg functionalities from nsvg to resvg everything seems to be working fine however i m unfortunately using a windows machine and i was not able to install cairo this makes it hard for me to debug the library for now i m using wsl to debug but help on installing cairo would be appreciated since the where not able to help me much
0
5,548
27,776,201,954
IssuesEvent
2023-03-16 17:23:26
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
The database extension overrides CSS styles from core
bug UI maintainability extension
The database extensions loads Yahoo's prue.css library which styles elements regardless of classes and as a result one must always ensure that core got styles which take priority to avoid side effects. ### To Reproduce Inspect any element on either the index or project page and check for styles coming from prue.css.
True
The database extension overrides CSS styles from core - The database extensions loads Yahoo's prue.css library which styles elements regardless of classes and as a result one must always ensure that core got styles which take priority to avoid side effects. ### To Reproduce Inspect any element on either the index or project page and check for styles coming from prue.css.
main
the database extension overrides css styles from core the database extensions loads yahoo s prue css library which styles elements regardless of classes and as a result one must always ensure that core got styles which take priority to avoid side effects to reproduce inspect any element on either the index or project page and check for styles coming from prue css
1
159,995
20,092,432,575
IssuesEvent
2022-02-06 01:01:58
directoryxx/Django-Polls-Tutorial
https://api.github.com/repos/directoryxx/Django-Polls-Tutorial
opened
CVE-2022-22818 (Medium) detected in django2.2.2
security vulnerability
## CVE-2022-22818 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>django2.2.2</b></p></summary> <p> <p>The Web framework for perfectionists with deadlines.</p> <p>Library home page: <a href=https://github.com/django/django.git>https://github.com/django/django.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/Django-Polls-Tutorial/venv/lib/python3.7/site-packages/django/template/defaulttags.py</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The {% debug %} template tag in Django 2.2 before 2.2.27, 3.2 before 3.2.12, and 4.0 before 4.0.2 does not properly encode the current context. This may lead to XSS. <p>Publish Date: 2022-02-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22818>CVE-2022-22818</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2022/feb/01/security-releases/">https://www.djangoproject.com/weblog/2022/feb/01/security-releases/</a></p> <p>Release Date: 2022-02-03</p> <p>Fix Resolution: Django - 2.2.27,3.2.12,4.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-22818 (Medium) detected in django2.2.2 - ## CVE-2022-22818 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>django2.2.2</b></p></summary> <p> <p>The Web framework for perfectionists with deadlines.</p> <p>Library home page: <a href=https://github.com/django/django.git>https://github.com/django/django.git</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/Django-Polls-Tutorial/venv/lib/python3.7/site-packages/django/template/defaulttags.py</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The {% debug %} template tag in Django 2.2 before 2.2.27, 3.2 before 3.2.12, and 4.0 before 4.0.2 does not properly encode the current context. This may lead to XSS. <p>Publish Date: 2022-02-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22818>CVE-2022-22818</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.djangoproject.com/weblog/2022/feb/01/security-releases/">https://www.djangoproject.com/weblog/2022/feb/01/security-releases/</a></p> <p>Release Date: 2022-02-03</p> <p>Fix Resolution: Django - 2.2.27,3.2.12,4.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in cve medium severity vulnerability vulnerable library the web framework for perfectionists with deadlines library home page a href vulnerable source files django polls tutorial venv lib site packages django template defaulttags py vulnerability details the debug template tag in django before before and before does not properly encode the current context this may lead to xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution django step up your open source security game with whitesource
0
4,203
20,604,219,346
IssuesEvent
2022-03-06 18:36:01
bromite/bromite
https://api.github.com/repos/bromite/bromite
closed
Enable encrypted Backups
enhancement enhancement-without-maintainer
### Is your feature request related to privacy? Yes, with [Seedvault](https://github.com/seedvault-app/seedvault) it is possible to do privacy respecting backups - IF the app allows it... ### Is there a patch available for this feature somewhere? Yes, see: https://github.com/grote/Transportr/commit/4dc38f429f75909a088d8bd8a5b3b5ddd8030f71 from [seedvault FAQ#why-do-some-apps-not-allow-to-get-backed-up](https://github.com/seedvault-app/seedvault/wiki/FAQ#why-do-some-apps-not-allow-to-get-backed-up) ### Describe the solution you would like Bromite should allow backups which are encrypted. ### Describe alternatives you have considered None.
True
Enable encrypted Backups - ### Is your feature request related to privacy? Yes, with [Seedvault](https://github.com/seedvault-app/seedvault) it is possible to do privacy respecting backups - IF the app allows it... ### Is there a patch available for this feature somewhere? Yes, see: https://github.com/grote/Transportr/commit/4dc38f429f75909a088d8bd8a5b3b5ddd8030f71 from [seedvault FAQ#why-do-some-apps-not-allow-to-get-backed-up](https://github.com/seedvault-app/seedvault/wiki/FAQ#why-do-some-apps-not-allow-to-get-backed-up) ### Describe the solution you would like Bromite should allow backups which are encrypted. ### Describe alternatives you have considered None.
main
enable encrypted backups is your feature request related to privacy yes with it is possible to do privacy respecting backups if the app allows it is there a patch available for this feature somewhere yes see from describe the solution you would like bromite should allow backups which are encrypted describe alternatives you have considered none
1
96
2,518,727,816
IssuesEvent
2015-01-17 01:29:15
tgstation/-tg-station
https://api.github.com/repos/tgstation/-tg-station
closed
chemistry heater is shit
Bug Usability
a. it can't eject the beakers you put in until it's done heating b. it doesn't have a gui c. it heats INCREDIBLY slowly. 2 per tick may not seem too slow, but there is no chemical that is BELOW 200k to heat, so it takes incredibly long most of the time d. there's no way to tell what the temperature is at the given moment
True
chemistry heater is shit - a. it can't eject the beakers you put in until it's done heating b. it doesn't have a gui c. it heats INCREDIBLY slowly. 2 per tick may not seem too slow, but there is no chemical that is BELOW 200k to heat, so it takes incredibly long most of the time d. there's no way to tell what the temperature is at the given moment
non_main
chemistry heater is shit a it can t eject the beakers you put in until it s done heating b it doesn t have a gui c it heats incredibly slowly per tick may not seem too slow but there is no chemical that is below to heat so it takes incredibly long most of the time d there s no way to tell what the temperature is at the given moment
0
958
4,702,357,466
IssuesEvent
2016-10-13 01:43:18
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ios_config save: true - "unable to load backup configuration"
affects_2.2 bug_report networking P1 waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ios_config iosxr_config ##### ANSIBLE VERSION ``` ansible --version ansible 2.2.0 (devel 1c7890bf86) last updated 2016/09/26 10:42:34 (GMT +100) lib/ansible/modules/core: (devel cf243860ff) last updated 2016/09/26 10:42:39 (GMT +100) lib/ansible/modules/extras: (devel 7aab9cd93b) last updated 2016/09/26 10:42:41 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY From the code in `ios_config` looks like we've got an exception in the exception handler ``` try: run(module, result) # Triggers exception except NetworkError: load_backup(module) exc = get_exception() module.fail_json(msg=str(exc)) ... def load_backup(module): try: module.cli(['exit', 'config replace flash:/ansible-rollback force']) # triggers exception except NetworkError: module.fail_json(msg='unable to load backup configuration') ``` 1) Underlying issue needs fixing 2) Would it be worth Should we capture the exception earlier on in `main()` 3) In `load_backup` can be give the user more feedback about why the error occured? ##### STEPS TO REPRODUCE ```yaml - name: setup ios_config: commands: - no description - no shutdown parents: - interface Loopback999 match: none provider: "{{ cli }}" - name: save config ios_config: save: true provider: "{{ cli }}" register: result ``` ##### EXPECTED RESULTS ##### ACTUAL RESULTS ```yaml TASK [test_ios_config : save config] ******************************************* task path: /home/johnb/git/ansible-inc/testing-ios/roles/test_ios_config/tests/cli/save.yaml:15 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/ios/ios_config.py <ios01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <ios01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396 `" && echo ansible-tmp-1474897504.34-175123254234396="` echo $HOME/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396 `" ) && sleep 0' <ios01> PUT /tmp/tmp5Z2AeZ TO /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ios_config.py <ios01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ios_config.py && sleep 0' <ios01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ios_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/" > /dev/null 2>&1 && sleep 0' fatal: [ios01]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "after": null, "auth_pass": null, "authorize": false, "backup": false, "before": null, "config": null, "defaults": false, "force": false, "host": "ios01", "lines": null, "match": "line", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "ios01", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "transport": "cli", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER" }, "replace": "line", "save": true, "src": null, "ssh_keyfile": null, "timeout": 10, "transport": "cli", "use_ssl": true, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true } }, "msg": "unable to load backup configuration" ```
True
ios_config save: true - "unable to load backup configuration" - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ios_config iosxr_config ##### ANSIBLE VERSION ``` ansible --version ansible 2.2.0 (devel 1c7890bf86) last updated 2016/09/26 10:42:34 (GMT +100) lib/ansible/modules/core: (devel cf243860ff) last updated 2016/09/26 10:42:39 (GMT +100) lib/ansible/modules/extras: (devel 7aab9cd93b) last updated 2016/09/26 10:42:41 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY From the code in `ios_config` looks like we've got an exception in the exception handler ``` try: run(module, result) # Triggers exception except NetworkError: load_backup(module) exc = get_exception() module.fail_json(msg=str(exc)) ... def load_backup(module): try: module.cli(['exit', 'config replace flash:/ansible-rollback force']) # triggers exception except NetworkError: module.fail_json(msg='unable to load backup configuration') ``` 1) Underlying issue needs fixing 2) Would it be worth Should we capture the exception earlier on in `main()` 3) In `load_backup` can be give the user more feedback about why the error occured? ##### STEPS TO REPRODUCE ```yaml - name: setup ios_config: commands: - no description - no shutdown parents: - interface Loopback999 match: none provider: "{{ cli }}" - name: save config ios_config: save: true provider: "{{ cli }}" register: result ``` ##### EXPECTED RESULTS ##### ACTUAL RESULTS ```yaml TASK [test_ios_config : save config] ******************************************* task path: /home/johnb/git/ansible-inc/testing-ios/roles/test_ios_config/tests/cli/save.yaml:15 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/ios/ios_config.py <ios01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <ios01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396 `" && echo ansible-tmp-1474897504.34-175123254234396="` echo $HOME/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396 `" ) && sleep 0' <ios01> PUT /tmp/tmp5Z2AeZ TO /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ios_config.py <ios01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ios_config.py && sleep 0' <ios01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/ios_config.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1474897504.34-175123254234396/" > /dev/null 2>&1 && sleep 0' fatal: [ios01]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "after": null, "auth_pass": null, "authorize": false, "backup": false, "before": null, "config": null, "defaults": false, "force": false, "host": "ios01", "lines": null, "match": "line", "parents": null, "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "port": null, "provider": { "host": "ios01", "password": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "transport": "cli", "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER" }, "replace": "line", "save": true, "src": null, "ssh_keyfile": null, "timeout": 10, "transport": "cli", "use_ssl": true, "username": "VALUE_SPECIFIED_IN_NO_LOG_PARAMETER", "validate_certs": true } }, "msg": "unable to load backup configuration" ```
main
ios config save true unable to load backup configuration issue type bug report component name ios config iosxr config ansible version ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary from the code in ios config looks like we ve got an exception in the exception handler try run module result triggers exception except networkerror load backup module exc get exception module fail json msg str exc def load backup module try module cli triggers exception except networkerror module fail json msg unable to load backup configuration underlying issue needs fixing would it be worth should we capture the exception earlier on in main in load backup can be give the user more feedback about why the error occured steps to reproduce yaml name setup ios config commands no description no shutdown parents interface match none provider cli name save config ios config save true provider cli register result expected results actual results yaml task task path home johnb git ansible inc testing ios roles test ios config tests cli save yaml using module file home johnb git ansible inc ansible lib ansible modules core network ios ios config py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home johnb ansible tmp ansible tmp ios config py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp ios config py sleep exec bin sh c python home johnb ansible tmp ansible tmp ios config py rm rf home johnb ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module args after null auth pass null authorize false backup false before null config null defaults false force false host lines null match line parents null password value specified in no log parameter port null provider host password value specified in no log parameter transport cli username value specified in no log parameter replace line save true src null ssh keyfile null timeout transport cli use ssl true username value specified in no log parameter validate certs true msg unable to load backup configuration
1
118,965
15,383,449,376
IssuesEvent
2021-03-03 02:45:16
PostHog/posthog.com
https://api.github.com/repos/PostHog/posthog.com
opened
Features pages
design
First stab at a layout for feature pages. This has evolved a bit from my original thinking. [**Invision prototype of Product section tabs**](https://balsamiq.cloud/sd0i9zq/pxvojo4/r60A1?f=N4IgUiBcAMA0IDkpxAYSgRngGU%2FAQsvALJEgBaZA0lANoC6AvkA%3D) (Overview | Product analytics) ![image](https://user-images.githubusercontent.com/154479/109743921-7a7d9880-7b9f-11eb-9797-7e2e14477b13.png) ### New idea for product subnavigation (seen above) - Overview (pictured above) - _use cases that PostHog handles, showing the breadth of the platform_ - Product analytics (click to this in the prototype) - Insights & experimentation - Open source platform ### Backstory: Existing plan My current plan for the Product dropdown menu had been to split the menu items into three groups: 1. Product suite 2. Top features 3. Product overview/Compare features (a little clunky since there's more to the product) **BUT,** in the back of my mind, there have been some problems with this format I've been thinking about: - We'll have at least a couple comparison pages (vs competitor, compare editions), and there's no obvious place for these - There should be a clear product overview landing page - This lists top features, but there are a whole lot more features that deserve mentions outside of a feature comparison table - Once you click into one of these top-level options, how do you maintain a clear understanding of where you are _inside_ the Product section? ### New structure A subnav that sticks with you throughout the features section **Overview** - Use cases **Product analytics** - Trends - Funnels - Cohorts - Retention - Paths - User sessions - _Revenue tracking_ **Product learning & experimentation** - Session recording - Heatmaps - Feature flags - _A/B testing_ - _User feedback_ **Platform** Some sort of interesting, masonry-style layout (or newspaper), highlighting a bunch of important things with a sidebar or two. (I have some ideas...) - Compare features - Compare editions - Top features - Event autocapture - Self-hosted option - Open source - API - Data privacy/ad blocker buster - SQL access - Plugins - Plugin list
1.0
Features pages - First stab at a layout for feature pages. This has evolved a bit from my original thinking. [**Invision prototype of Product section tabs**](https://balsamiq.cloud/sd0i9zq/pxvojo4/r60A1?f=N4IgUiBcAMA0IDkpxAYSgRngGU%2FAQsvALJEgBaZA0lANoC6AvkA%3D) (Overview | Product analytics) ![image](https://user-images.githubusercontent.com/154479/109743921-7a7d9880-7b9f-11eb-9797-7e2e14477b13.png) ### New idea for product subnavigation (seen above) - Overview (pictured above) - _use cases that PostHog handles, showing the breadth of the platform_ - Product analytics (click to this in the prototype) - Insights & experimentation - Open source platform ### Backstory: Existing plan My current plan for the Product dropdown menu had been to split the menu items into three groups: 1. Product suite 2. Top features 3. Product overview/Compare features (a little clunky since there's more to the product) **BUT,** in the back of my mind, there have been some problems with this format I've been thinking about: - We'll have at least a couple comparison pages (vs competitor, compare editions), and there's no obvious place for these - There should be a clear product overview landing page - This lists top features, but there are a whole lot more features that deserve mentions outside of a feature comparison table - Once you click into one of these top-level options, how do you maintain a clear understanding of where you are _inside_ the Product section? ### New structure A subnav that sticks with you throughout the features section **Overview** - Use cases **Product analytics** - Trends - Funnels - Cohorts - Retention - Paths - User sessions - _Revenue tracking_ **Product learning & experimentation** - Session recording - Heatmaps - Feature flags - _A/B testing_ - _User feedback_ **Platform** Some sort of interesting, masonry-style layout (or newspaper), highlighting a bunch of important things with a sidebar or two. (I have some ideas...) - Compare features - Compare editions - Top features - Event autocapture - Self-hosted option - Open source - API - Data privacy/ad blocker buster - SQL access - Plugins - Plugin list
non_main
features pages first stab at a layout for feature pages this has evolved a bit from my original thinking overview product analytics new idea for product subnavigation seen above overview pictured above use cases that posthog handles showing the breadth of the platform product analytics click to this in the prototype insights experimentation open source platform backstory existing plan my current plan for the product dropdown menu had been to split the menu items into three groups product suite top features product overview compare features a little clunky since there s more to the product but in the back of my mind there have been some problems with this format i ve been thinking about we ll have at least a couple comparison pages vs competitor compare editions and there s no obvious place for these there should be a clear product overview landing page this lists top features but there are a whole lot more features that deserve mentions outside of a feature comparison table once you click into one of these top level options how do you maintain a clear understanding of where you are inside the product section new structure a subnav that sticks with you throughout the features section overview use cases product analytics trends funnels cohorts retention paths user sessions revenue tracking product learning experimentation session recording heatmaps feature flags a b testing user feedback platform some sort of interesting masonry style layout or newspaper highlighting a bunch of important things with a sidebar or two i have some ideas compare features compare editions top features event autocapture self hosted option open source api data privacy ad blocker buster sql access plugins plugin list
0
182,658
14,145,411,599
IssuesEvent
2020-11-10 17:42:25
istio/istio.io
https://api.github.com/repos/istio/istio.io
closed
[istio.io testing] Need test for Ingress Control
area/test and release kind/docs kind/enhancement lifecycle/needs-triage
**Describe the feature request** We need to develop a test for the [Ingress Control](https://preliminary.istio.io/docs/tasks/traffic-management/ingress/ingress-control/) task in order to provide automated testing of the Istio.io website for future releases. Example tests and the documentation for the framework can be found here: https://github.com/istio/istio/tree/master/tests/integration/istioio A YouTube video describing the usage of the framework can be found here: https://www.youtube.com/watch?v=3y-z8NaVwr0 Please reach out to the Test and Release channel on the [Istio Slack](https://discuss.istio.io/t/istio-slack-channel/1527) or [Test and Release Discuss](https://discuss.istio.io/c/test-and-release) with any questions.
1.0
[istio.io testing] Need test for Ingress Control - **Describe the feature request** We need to develop a test for the [Ingress Control](https://preliminary.istio.io/docs/tasks/traffic-management/ingress/ingress-control/) task in order to provide automated testing of the Istio.io website for future releases. Example tests and the documentation for the framework can be found here: https://github.com/istio/istio/tree/master/tests/integration/istioio A YouTube video describing the usage of the framework can be found here: https://www.youtube.com/watch?v=3y-z8NaVwr0 Please reach out to the Test and Release channel on the [Istio Slack](https://discuss.istio.io/t/istio-slack-channel/1527) or [Test and Release Discuss](https://discuss.istio.io/c/test-and-release) with any questions.
non_main
need test for ingress control describe the feature request we need to develop a test for the task in order to provide automated testing of the istio io website for future releases example tests and the documentation for the framework can be found here a youtube video describing the usage of the framework can be found here please reach out to the test and release channel on the or with any questions
0
155,832
12,279,356,651
IssuesEvent
2020-05-08 12:01:38
awslabs/s2n
https://api.github.com/repos/awslabs/s2n
closed
Dynamic record size tests fail sometimes
good first issue type/flaky_test type/integration_test
## **Problem:** From time to time we run into failures with dynamic record size tests. https://travis-ci.org/awslabs/s2n/jobs/650075690?utm_medium=notification&utm_source=github_status ```sh Running s2n dynamic record size tests Cipher: AES128-SHA Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES128-SHA Vers: TLSv1.2 ... PASSED Cipher: AES256-SHA Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES256-SHA Vers: TLSv1.2 ... PASSED Cipher: AES128-SHA256 Vers: TLSv1.2 ... PASSED Cipher: AES256-SHA256 Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES128-SHA256 Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES256-SHA256 Vers: TLSv1.2 ... FAILED Makefile:52: recipe for target 'dynamic_record' failed make[2]: *** [dynamic_record] Error 1 make[2]: Leaving directory '/home/travis/build/awslabs/s2n/tests/integration' Makefile:39: recipe for target 'integration' failed make[1]: *** [integration] Error 2 make[1]: Leaving directory '/home/travis/build/awslabs/s2n/tests' Makefile:73: recipe for target 'integration' failed make: *** [integration] Error 2 The command ".travis/s2n_travis_build.sh" exited with 2. 0.00s$ .travis/s2n_after_travis_build.sh ``` https://travis-ci.org/awslabs/s2n/jobs/649153883?utm_medium=notification&utm_source=github_status ```sh Traceback (most recent call last): File "s2n_dynamic_record_size_test.py", line 293, in <module> sys.exit(main()) File "s2n_dynamic_record_size_test.py", line 282, in main failed += test(host, port, test_ciphers, int(file_size / 2)) File "s2n_dynamic_record_size_test.py", line 196, in test result = run_test(host, port, ssl_version, cipher, threshold) File "s2n_dynamic_record_size_test.py", line 179, in run_test failed += analyze_throughput_dump(out_array) File "s2n_dynamic_record_size_test.py", line 233, in analyze_throughput_dump output = array[i] IndexError: list index out of range Makefile:52: recipe for target 'dynamic_record' failed ``` ## **Proposed Solution:** Investigate, fix. [//]: # (NOTE: If you believe this might be a security issue, please email aws-security@amazon.com instead of creating a GitHub issue. For more details, see the AWS Vulnerability Reporting Guide: https://aws.amazon.com/security/vulnerability-reporting/ )
2.0
Dynamic record size tests fail sometimes - ## **Problem:** From time to time we run into failures with dynamic record size tests. https://travis-ci.org/awslabs/s2n/jobs/650075690?utm_medium=notification&utm_source=github_status ```sh Running s2n dynamic record size tests Cipher: AES128-SHA Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES128-SHA Vers: TLSv1.2 ... PASSED Cipher: AES256-SHA Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES256-SHA Vers: TLSv1.2 ... PASSED Cipher: AES128-SHA256 Vers: TLSv1.2 ... PASSED Cipher: AES256-SHA256 Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES128-SHA256 Vers: TLSv1.2 ... PASSED Cipher: DHE-RSA-AES256-SHA256 Vers: TLSv1.2 ... FAILED Makefile:52: recipe for target 'dynamic_record' failed make[2]: *** [dynamic_record] Error 1 make[2]: Leaving directory '/home/travis/build/awslabs/s2n/tests/integration' Makefile:39: recipe for target 'integration' failed make[1]: *** [integration] Error 2 make[1]: Leaving directory '/home/travis/build/awslabs/s2n/tests' Makefile:73: recipe for target 'integration' failed make: *** [integration] Error 2 The command ".travis/s2n_travis_build.sh" exited with 2. 0.00s$ .travis/s2n_after_travis_build.sh ``` https://travis-ci.org/awslabs/s2n/jobs/649153883?utm_medium=notification&utm_source=github_status ```sh Traceback (most recent call last): File "s2n_dynamic_record_size_test.py", line 293, in <module> sys.exit(main()) File "s2n_dynamic_record_size_test.py", line 282, in main failed += test(host, port, test_ciphers, int(file_size / 2)) File "s2n_dynamic_record_size_test.py", line 196, in test result = run_test(host, port, ssl_version, cipher, threshold) File "s2n_dynamic_record_size_test.py", line 179, in run_test failed += analyze_throughput_dump(out_array) File "s2n_dynamic_record_size_test.py", line 233, in analyze_throughput_dump output = array[i] IndexError: list index out of range Makefile:52: recipe for target 'dynamic_record' failed ``` ## **Proposed Solution:** Investigate, fix. [//]: # (NOTE: If you believe this might be a security issue, please email aws-security@amazon.com instead of creating a GitHub issue. For more details, see the AWS Vulnerability Reporting Guide: https://aws.amazon.com/security/vulnerability-reporting/ )
non_main
dynamic record size tests fail sometimes problem from time to time we run into failures with dynamic record size tests sh running dynamic record size tests cipher sha vers passed cipher dhe rsa sha vers passed cipher sha vers passed cipher dhe rsa sha vers passed cipher vers passed cipher vers passed cipher dhe rsa vers passed cipher dhe rsa vers failed makefile recipe for target dynamic record failed make error make leaving directory home travis build awslabs tests integration makefile recipe for target integration failed make error make leaving directory home travis build awslabs tests makefile recipe for target integration failed make error the command travis travis build sh exited with travis after travis build sh sh traceback most recent call last file dynamic record size test py line in sys exit main file dynamic record size test py line in main failed test host port test ciphers int file size file dynamic record size test py line in test result run test host port ssl version cipher threshold file dynamic record size test py line in run test failed analyze throughput dump out array file dynamic record size test py line in analyze throughput dump output array indexerror list index out of range makefile recipe for target dynamic record failed proposed solution investigate fix note if you believe this might be a security issue please email aws security amazon com instead of creating a github issue for more details see the aws vulnerability reporting guide
0
1,039
4,836,057,250
IssuesEvent
2016-11-08 18:30:39
daemonraco/toobasic
https://api.github.com/repos/daemonraco/toobasic
closed
JSON Validator for Database Specs
Core Logic Database Database Structure Maintainer JSONValidator
## What to do Implement validations for database specifications using _JSON Validator_, both version 1 and 2.
True
JSON Validator for Database Specs - ## What to do Implement validations for database specifications using _JSON Validator_, both version 1 and 2.
main
json validator for database specs what to do implement validations for database specifications using json validator both version and
1
54,939
11,352,958,206
IssuesEvent
2020-01-24 14:41:28
mozilla/addons-server
https://api.github.com/repos/mozilla/addons-server
closed
Investigate pygit2 performance by iterating differently through the diffs
component: code manager priority: p3
This is just temporary for now till I played more with it but I found https://github.com/libgit2/pygit2/issues/736 and it seems that there might be some more performant API usage that we could use to speed up the diff api. I'll investigate and update the issue accordingly.
1.0
Investigate pygit2 performance by iterating differently through the diffs - This is just temporary for now till I played more with it but I found https://github.com/libgit2/pygit2/issues/736 and it seems that there might be some more performant API usage that we could use to speed up the diff api. I'll investigate and update the issue accordingly.
non_main
investigate performance by iterating differently through the diffs this is just temporary for now till i played more with it but i found and it seems that there might be some more performant api usage that we could use to speed up the diff api i ll investigate and update the issue accordingly
0
60,818
7,390,218,221
IssuesEvent
2018-03-16 11:32:11
otavanopisto/muikku
https://api.github.com/repos/otavanopisto/muikku
closed
Issues for implementation of the new UI design
REDESIGN2017 enhancement review ui
It would be a great idea to atomize rest of the new UI design for smaller issues and place them to milestones. Lankinen is about to leave for his vacation, so this should be done next week.
1.0
Issues for implementation of the new UI design - It would be a great idea to atomize rest of the new UI design for smaller issues and place them to milestones. Lankinen is about to leave for his vacation, so this should be done next week.
non_main
issues for implementation of the new ui design it would be a great idea to atomize rest of the new ui design for smaller issues and place them to milestones lankinen is about to leave for his vacation so this should be done next week
0
3,744
15,733,398,958
IssuesEvent
2021-03-29 19:30:55
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
opened
Add runtime option to disable start/stop API and annotations and then remove the ifdefs
Maintainability Type-Feature
For running untrusted applications, it might be nice to disable all cases where DR listens to the app for what to do: the start/stop API and annotations. Today we have ifdefs for both of those: but we'd like to remove the DR_APP_EXPORTS define (xref #2971), and the ANNOTATIONS define for that matter once it's ported to AArchXX. Plus it is harder to maintain a separate build. This option proposes adding a runtime option to disable these features and adding a test to the CI for this option. I think the overhead of checking the option in the inner interp loop will be worth the code cleanliness and maintenance gain?
True
Add runtime option to disable start/stop API and annotations and then remove the ifdefs - For running untrusted applications, it might be nice to disable all cases where DR listens to the app for what to do: the start/stop API and annotations. Today we have ifdefs for both of those: but we'd like to remove the DR_APP_EXPORTS define (xref #2971), and the ANNOTATIONS define for that matter once it's ported to AArchXX. Plus it is harder to maintain a separate build. This option proposes adding a runtime option to disable these features and adding a test to the CI for this option. I think the overhead of checking the option in the inner interp loop will be worth the code cleanliness and maintenance gain?
main
add runtime option to disable start stop api and annotations and then remove the ifdefs for running untrusted applications it might be nice to disable all cases where dr listens to the app for what to do the start stop api and annotations today we have ifdefs for both of those but we d like to remove the dr app exports define xref and the annotations define for that matter once it s ported to aarchxx plus it is harder to maintain a separate build this option proposes adding a runtime option to disable these features and adding a test to the ci for this option i think the overhead of checking the option in the inner interp loop will be worth the code cleanliness and maintenance gain
1
146,257
13,174,993,485
IssuesEvent
2020-08-12 00:09:37
nicholaschiasson/ngx_upstream_jdomain
https://api.github.com/repos/nicholaschiasson/ngx_upstream_jdomain
closed
Improve README.md
documentation good first issue help wanted
Document the project better. - [ ] Features: all supported directive attributes and what they do. - [ ] Development tools to use for local development: docker, [act](https://github.com/nektos/act), etc. - [ ] General instructions for local development: building, testing, running github actions locally, etc.
1.0
Improve README.md - Document the project better. - [ ] Features: all supported directive attributes and what they do. - [ ] Development tools to use for local development: docker, [act](https://github.com/nektos/act), etc. - [ ] General instructions for local development: building, testing, running github actions locally, etc.
non_main
improve readme md document the project better features all supported directive attributes and what they do development tools to use for local development docker etc general instructions for local development building testing running github actions locally etc
0
472,005
13,614,525,834
IssuesEvent
2020-09-23 13:22:03
magento/adobe-stock-integration
https://api.github.com/repos/magento/adobe-stock-integration
closed
Error when saving image from stock that does not have Category
Priority: P3 Progress: PR created Severity: S3
### Steps to reproduce (*) 1. Open Adobe Stock Panel from Media Gallery 2. Search for "Distant deer in a misty forest" 4. In the search results find an image with category "None" 5. Save Preview ![image](https://user-images.githubusercontent.com/2028541/93813917-7cd8df80-fc5c-11ea-91ec-b4c84945e141.png) ### Expected result (*) Image is saved successfully, no errors appear ### Actual result (*) Image is saved but an error appears "An error occurred during adobe stock asset save." Error in the exception log: ``` report.CRITICAL: SQLSTATE[23000]: Integrity constraint violation: 1452 Cannot add or update a child row: a foreign key constraint fails (`main`.`adobe_stock_asset`, CONSTRAINT `ADOBE_STOCK_ASSET_CATEGORY_ID_ADOBE_STOCK_CATEGORY_ID` FOREIGN KEY (`category_id`) REFERENCES `adobe_stock_category` (`id`) ON DELETE SET NULL), query was: INSERT INTO `adobe_stock_asset` (`id`,`media_gallery_id`,`category_id`,`creator_id`,`is_licensed`,`creation_date`) VALUES (?, ?, ?, ?, ?, ?) ON DUPLICATE KEY UPDATE `id` = VALUES(`id`), `media_gallery_id` = VALUES(`media_gallery_id`), `category_id` = VALUES(`category_id`), `creator_id` = VALUES(`creator_id`), `is_licensed` = VALUES(`is_licensed`), `creation_date` = VALUES(`creation_date`) {"exception":"[object] (Zend_Db_Statement_Exception(code: 23000): SQLSTATE[23000]: Integrity constraint violation: 1452 Cannot add or update a child row: a foreign key constraint fails (`main`.`adobe_stock_asset`, CONSTRAINT `ADOBE_STOCK_ASSET_CATEGORY_ID_ADOBE_STOCK_CATEGORY_ID` FOREIGN KEY (`category_id`) REFERENCES `adobe_stock_category` (`id`) ON DELETE SET NULL), query was: INSERT INTO `adobe_stock_asset` (`id`,`media_gallery_id`,`category_id`,`creator_id`,`is_licensed`,`creation_date`) VALUES (?, ?, ?, ?, ?, ?) ON DUPLICATE KEY UPDATE `id` = VALUES(`id`), `media_gallery_id` = VALUES(`media_gallery_id`), `category_id` = VALUES(`category_id`), `creator_id` = VALUES(`creator_id`), `is_licensed` = VALUES(`is_licensed`), `creation_date` = VALUES(`creation_date`) at /app/vendor/magento/framework/DB/Statement/Pdo/Mysql.php:110, PDOException(code: 23000): SQLSTATE[23000]: Integrity constraint violation: 1452 Cannot add or update a child row: a foreign key constraint fails (`main`.`adobe_stock_asset`, CONSTRAINT `ADOBE_STOCK_ASSET_CATEGORY_ID_ADOBE_STOCK_CATEGORY_ID` FOREIGN KEY (`category_id`) REFERENCES `adobe_stock_category` (`id`) ON DELETE SET NULL) at /app/vendor/magento/framework/DB/Statement/Pdo/Mysql.php:91)"} [] ``` ### Proposed solution (backward compatible) In case Category is empty - we should load/create "None" category and use it for the asset
1.0
Error when saving image from stock that does not have Category - ### Steps to reproduce (*) 1. Open Adobe Stock Panel from Media Gallery 2. Search for "Distant deer in a misty forest" 4. In the search results find an image with category "None" 5. Save Preview ![image](https://user-images.githubusercontent.com/2028541/93813917-7cd8df80-fc5c-11ea-91ec-b4c84945e141.png) ### Expected result (*) Image is saved successfully, no errors appear ### Actual result (*) Image is saved but an error appears "An error occurred during adobe stock asset save." Error in the exception log: ``` report.CRITICAL: SQLSTATE[23000]: Integrity constraint violation: 1452 Cannot add or update a child row: a foreign key constraint fails (`main`.`adobe_stock_asset`, CONSTRAINT `ADOBE_STOCK_ASSET_CATEGORY_ID_ADOBE_STOCK_CATEGORY_ID` FOREIGN KEY (`category_id`) REFERENCES `adobe_stock_category` (`id`) ON DELETE SET NULL), query was: INSERT INTO `adobe_stock_asset` (`id`,`media_gallery_id`,`category_id`,`creator_id`,`is_licensed`,`creation_date`) VALUES (?, ?, ?, ?, ?, ?) ON DUPLICATE KEY UPDATE `id` = VALUES(`id`), `media_gallery_id` = VALUES(`media_gallery_id`), `category_id` = VALUES(`category_id`), `creator_id` = VALUES(`creator_id`), `is_licensed` = VALUES(`is_licensed`), `creation_date` = VALUES(`creation_date`) {"exception":"[object] (Zend_Db_Statement_Exception(code: 23000): SQLSTATE[23000]: Integrity constraint violation: 1452 Cannot add or update a child row: a foreign key constraint fails (`main`.`adobe_stock_asset`, CONSTRAINT `ADOBE_STOCK_ASSET_CATEGORY_ID_ADOBE_STOCK_CATEGORY_ID` FOREIGN KEY (`category_id`) REFERENCES `adobe_stock_category` (`id`) ON DELETE SET NULL), query was: INSERT INTO `adobe_stock_asset` (`id`,`media_gallery_id`,`category_id`,`creator_id`,`is_licensed`,`creation_date`) VALUES (?, ?, ?, ?, ?, ?) ON DUPLICATE KEY UPDATE `id` = VALUES(`id`), `media_gallery_id` = VALUES(`media_gallery_id`), `category_id` = VALUES(`category_id`), `creator_id` = VALUES(`creator_id`), `is_licensed` = VALUES(`is_licensed`), `creation_date` = VALUES(`creation_date`) at /app/vendor/magento/framework/DB/Statement/Pdo/Mysql.php:110, PDOException(code: 23000): SQLSTATE[23000]: Integrity constraint violation: 1452 Cannot add or update a child row: a foreign key constraint fails (`main`.`adobe_stock_asset`, CONSTRAINT `ADOBE_STOCK_ASSET_CATEGORY_ID_ADOBE_STOCK_CATEGORY_ID` FOREIGN KEY (`category_id`) REFERENCES `adobe_stock_category` (`id`) ON DELETE SET NULL) at /app/vendor/magento/framework/DB/Statement/Pdo/Mysql.php:91)"} [] ``` ### Proposed solution (backward compatible) In case Category is empty - we should load/create "None" category and use it for the asset
non_main
error when saving image from stock that does not have category steps to reproduce open adobe stock panel from media gallery search for distant deer in a misty forest in the search results find an image with category none save preview expected result image is saved successfully no errors appear actual result image is saved but an error appears an error occurred during adobe stock asset save error in the exception log report critical sqlstate integrity constraint violation cannot add or update a child row a foreign key constraint fails main adobe stock asset constraint adobe stock asset category id adobe stock category id foreign key category id references adobe stock category id on delete set null query was insert into adobe stock asset id media gallery id category id creator id is licensed creation date values on duplicate key update id values id media gallery id values media gallery id category id values category id creator id values creator id is licensed values is licensed creation date values creation date exception zend db statement exception code sqlstate integrity constraint violation cannot add or update a child row a foreign key constraint fails main adobe stock asset constraint adobe stock asset category id adobe stock category id foreign key category id references adobe stock category id on delete set null query was insert into adobe stock asset id media gallery id category id creator id is licensed creation date values on duplicate key update id values id media gallery id values media gallery id category id values category id creator id values creator id is licensed values is licensed creation date values creation date at app vendor magento framework db statement pdo mysql php pdoexception code sqlstate integrity constraint violation cannot add or update a child row a foreign key constraint fails main adobe stock asset constraint adobe stock asset category id adobe stock category id foreign key category id references adobe stock category id on delete set null at app vendor magento framework db statement pdo mysql php proposed solution backward compatible in case category is empty we should load create none category and use it for the asset
0
2,641
8,960,709,765
IssuesEvent
2019-01-28 07:12:09
pound-python/infobob
https://api.github.com/repos/pound-python/infobob
closed
Set up CI
maintainability
Travis is probably fine. Need to investigate if Travis has hooks we can use for CD. Requires #22
True
Set up CI - Travis is probably fine. Need to investigate if Travis has hooks we can use for CD. Requires #22
main
set up ci travis is probably fine need to investigate if travis has hooks we can use for cd requires
1
146,085
22,954,172,908
IssuesEvent
2022-07-19 10:01:30
dotnet/aspnetcore
https://api.github.com/repos/dotnet/aspnetcore
closed
QuickGrid: QuickGridItemsProvider naming suggestion
area-blazor :heavy_check_mark: Resolution: Answered Status: Resolved design-proposal
I was having a look at the QuickGrid implementation and noticed that the `GridItemsProviderResult` and `GridItemsProviderRequest` structs as well as the `GridItemsProvider` delegate could be generally useful in custom list components. In this case, I was wondering if it would make more sense for the naming to not assume that your component is going to use a grid layout.
1.0
QuickGrid: QuickGridItemsProvider naming suggestion - I was having a look at the QuickGrid implementation and noticed that the `GridItemsProviderResult` and `GridItemsProviderRequest` structs as well as the `GridItemsProvider` delegate could be generally useful in custom list components. In this case, I was wondering if it would make more sense for the naming to not assume that your component is going to use a grid layout.
non_main
quickgrid quickgriditemsprovider naming suggestion i was having a look at the quickgrid implementation and noticed that the griditemsproviderresult and griditemsproviderrequest structs as well as the griditemsprovider delegate could be generally useful in custom list components in this case i was wondering if it would make more sense for the naming to not assume that your component is going to use a grid layout
0
5,245
26,565,988,528
IssuesEvent
2023-01-20 20:16:30
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[QUESTION] Suddenly showing "It looks like something's gone wrong"
🤷‍♂️ Question 👤 Awaiting Maintainer Response
<h3 dir="auto">Question</h3> <p dir="auto">Hello :)</p> <p dir="auto">I've been using dashy for a year or more now, and kept it updated using watchtower (it runs once every 5 minutes so I am pretty sure that I use the latest version). Since today, I always get this error message:<br> <a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/1122571/213356668-54c693d4-1992-4612-89a5-63af424df6cb.png"><img src="https://user-images.githubusercontent.com/1122571/213356668-54c693d4-1992-4612-89a5-63af424df6cb.png" alt="image" style="max-width: 100%;"></a></p> <p dir="auto">However, after waiting for like 20 seconds, it shows the actual panel without errors, so I am a bit confused lol. I am not really sure on what to do or how to debug it :/</p> <p dir="auto">Any help would be appreciated!</p> <p dir="auto">My dashy url is <a href="https://hub.jeff-media.com" rel="nofollow">https://hub.jeff-media.com</a></p> <h3 dir="auto">Category</h3> <p dir="auto">Setup and Deployment</p> <h3 dir="auto">Please tick the boxes</h3> <ul class="contains-task-list"> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You are using a <a href="https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions">supported</a> version of Dashy (check the first two digits of the version number)</li> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked that this <a href="https://github.com/Lissy93/dashy/issues?q=is%3Aissue">question hasn't already been raised</a></li> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked the <a href="https://github.com/Lissy93/dashy/tree/master/docs#readme">docs</a> and <a href="https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting">troubleshooting</a> guide</li> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You agree to the <a href="https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct">code of conduct</a></li> </ul> <!-- unfurl begin --> <blockquote> <div> <strong> <a href="https://hub.jeff-media.com/">Dashy</a> </strong> </div> </blockquote> <!-- unfurl end -->
True
[QUESTION] Suddenly showing "It looks like something's gone wrong" - <h3 dir="auto">Question</h3> <p dir="auto">Hello :)</p> <p dir="auto">I've been using dashy for a year or more now, and kept it updated using watchtower (it runs once every 5 minutes so I am pretty sure that I use the latest version). Since today, I always get this error message:<br> <a target="_blank" rel="noopener noreferrer nofollow" href="https://user-images.githubusercontent.com/1122571/213356668-54c693d4-1992-4612-89a5-63af424df6cb.png"><img src="https://user-images.githubusercontent.com/1122571/213356668-54c693d4-1992-4612-89a5-63af424df6cb.png" alt="image" style="max-width: 100%;"></a></p> <p dir="auto">However, after waiting for like 20 seconds, it shows the actual panel without errors, so I am a bit confused lol. I am not really sure on what to do or how to debug it :/</p> <p dir="auto">Any help would be appreciated!</p> <p dir="auto">My dashy url is <a href="https://hub.jeff-media.com" rel="nofollow">https://hub.jeff-media.com</a></p> <h3 dir="auto">Category</h3> <p dir="auto">Setup and Deployment</p> <h3 dir="auto">Please tick the boxes</h3> <ul class="contains-task-list"> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You are using a <a href="https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions">supported</a> version of Dashy (check the first two digits of the version number)</li> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked that this <a href="https://github.com/Lissy93/dashy/issues?q=is%3Aissue">question hasn't already been raised</a></li> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You've checked the <a href="https://github.com/Lissy93/dashy/tree/master/docs#readme">docs</a> and <a href="https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting">troubleshooting</a> guide</li> <li class="task-list-item"><input type="checkbox" id="" disabled="" class="task-list-item-checkbox" checked=""> You agree to the <a href="https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct">code of conduct</a></li> </ul> <!-- unfurl begin --> <blockquote> <div> <strong> <a href="https://hub.jeff-media.com/">Dashy</a> </strong> </div> </blockquote> <!-- unfurl end -->
main
suddenly showing it looks like something s gone wrong question hello i ve been using dashy for a year or more now and kept it updated using watchtower it runs once every minutes so i am pretty sure that i use the latest version since today i always get this error message however after waiting for like seconds it shows the actual panel without errors so i am a bit confused lol i am not really sure on what to do or how to debug it any help would be appreciated my dashy url is category setup and deployment please tick the boxes you are using a you ve checked that this you ve checked the you agree to the
1
686
4,231,993,004
IssuesEvent
2016-07-04 19:17:31
Microsoft/UVAtlas
https://api.github.com/repos/Microsoft/UVAtlas
opened
Retire Windows 8.1 Store projects
maintainence
At some point we should remove support for this older version in favor of UWP apps ``UVatlas_Windows81.vcxproj`` Please put any requests for continued support here.
True
Retire Windows 8.1 Store projects - At some point we should remove support for this older version in favor of UWP apps ``UVatlas_Windows81.vcxproj`` Please put any requests for continued support here.
main
retire windows store projects at some point we should remove support for this older version in favor of uwp apps uvatlas vcxproj please put any requests for continued support here
1
22,427
3,645,494,735
IssuesEvent
2016-02-15 14:53:55
contao/core
https://api.github.com/repos/contao/core
closed
ModuleRegistration does not throw an error when username and password are identical
defect
The following comparison fails: `// Check whether the password matches the username` `if ($objWidget instanceof \FormPassword && $varValue == \Input::post('username'))` `{` ` $objWidget->addError($GLOBALS['TL_LANG']['ERR']['passwordName']);` `} Maybe related to #8144
1.0
ModuleRegistration does not throw an error when username and password are identical - The following comparison fails: `// Check whether the password matches the username` `if ($objWidget instanceof \FormPassword && $varValue == \Input::post('username'))` `{` ` $objWidget->addError($GLOBALS['TL_LANG']['ERR']['passwordName']);` `} Maybe related to #8144
non_main
moduleregistration does not throw an error when username and password are identical the following comparison fails check whether the password matches the username if objwidget instanceof formpassword varvalue input post username objwidget adderror globals maybe related to
0
3,166
12,226,516,301
IssuesEvent
2020-05-03 11:17:59
gfleetwood/asteres
https://api.github.com/repos/gfleetwood/asteres
opened
nocomplexity/SecurityPrivacyReferenceArchitecture (44663811)
Python maintain
https://github.com/nocomplexity/SecurityPrivacyReferenceArchitecture Open Repository for the Open Security and Privacy Reference Architecture
True
nocomplexity/SecurityPrivacyReferenceArchitecture (44663811) - https://github.com/nocomplexity/SecurityPrivacyReferenceArchitecture Open Repository for the Open Security and Privacy Reference Architecture
main
nocomplexity securityprivacyreferencearchitecture open repository for the open security and privacy reference architecture
1
4,272
21,465,353,929
IssuesEvent
2022-04-26 02:43:45
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Suit accessories with a storage component (ie pocket protectors) can no longer be accessed without first removing them from your uniform
Maintainability/Hinders improvements Bug Cleanup Flagged
Master Repo: Put pocket protector on your uniform Try to click it end up taking off your uniform alt: Put pocket protector on your uniform Try to drag it to yourself nothing happens I have a couple features in the works that depends on this working, so tagging it as hinders improvements
True
Suit accessories with a storage component (ie pocket protectors) can no longer be accessed without first removing them from your uniform - Master Repo: Put pocket protector on your uniform Try to click it end up taking off your uniform alt: Put pocket protector on your uniform Try to drag it to yourself nothing happens I have a couple features in the works that depends on this working, so tagging it as hinders improvements
main
suit accessories with a storage component ie pocket protectors can no longer be accessed without first removing them from your uniform master repo put pocket protector on your uniform try to click it end up taking off your uniform alt put pocket protector on your uniform try to drag it to yourself nothing happens i have a couple features in the works that depends on this working so tagging it as hinders improvements
1
5,503
27,485,366,823
IssuesEvent
2023-03-04 02:36:04
Windham-High-School/CubeServer
https://api.github.com/repos/Windham-High-School/CubeServer
closed
Pin exact versions of dependencies
maintainability :scroll: Convention Established :scroll:
Dependency versions will be pinned to prevent dependency updates from screwing with functionality and/or causing bugs. # :scroll: - Convention Establishment - :scroll: **Henceforth**, the following _shall_ be performed _prior to_ **every** major release (`X.0.0`): - New dependency versions should be sought after, considered, and tested. - Dependencies should be pinned to the _exact_ version in the `/Docker/*/requirements.txt` files, as well as looser requirements specified in the `/src/*/setup.py` files. **Exact Dependency Pinning** - `/Docker/CubeServer-api/requirements.txt` - `/Docker/CubeServer-app/requirements.txt` - `/Docker/CubeServer-beaconserver/requirements.txt` **Approximate Dependency Compatibility** - `/src/CubeServer-api/setup.py` - `/src/CubeServer-app/setup.py` - `/src/CubeServer-beaconserver/setup.py` - `/src/CubeServer-common/setup.py` - Docker container dependencies should be pinned with an exact minor version (`X.Y`) as well in every `/Docker/*/Dockerfile` file. **Affected Files** - `/Docker/CubeServer-api/Dockerfile` - `/Docker/CubeServer-app/Dockerfile` - `/Docker/CubeServer-beaconserver/Dockerfile` - `/Docker/access-point/Dockerfile` - **If issues arise**, patches and minor releases _may_ modify the required versions, so long as the exact versions are specified so as to ensure that the dependencies of the development instance are _identical_ to those in the production environment _for as long as_ the exact version of CubeServer remains constant - **Other things should also be considered** for replacement with newer software. Consider: - Python version - Javascript dependencies - Other Tools _If any problem is to be found with keeping this convention, this issue should be reopened for discussion._
True
Pin exact versions of dependencies - Dependency versions will be pinned to prevent dependency updates from screwing with functionality and/or causing bugs. # :scroll: - Convention Establishment - :scroll: **Henceforth**, the following _shall_ be performed _prior to_ **every** major release (`X.0.0`): - New dependency versions should be sought after, considered, and tested. - Dependencies should be pinned to the _exact_ version in the `/Docker/*/requirements.txt` files, as well as looser requirements specified in the `/src/*/setup.py` files. **Exact Dependency Pinning** - `/Docker/CubeServer-api/requirements.txt` - `/Docker/CubeServer-app/requirements.txt` - `/Docker/CubeServer-beaconserver/requirements.txt` **Approximate Dependency Compatibility** - `/src/CubeServer-api/setup.py` - `/src/CubeServer-app/setup.py` - `/src/CubeServer-beaconserver/setup.py` - `/src/CubeServer-common/setup.py` - Docker container dependencies should be pinned with an exact minor version (`X.Y`) as well in every `/Docker/*/Dockerfile` file. **Affected Files** - `/Docker/CubeServer-api/Dockerfile` - `/Docker/CubeServer-app/Dockerfile` - `/Docker/CubeServer-beaconserver/Dockerfile` - `/Docker/access-point/Dockerfile` - **If issues arise**, patches and minor releases _may_ modify the required versions, so long as the exact versions are specified so as to ensure that the dependencies of the development instance are _identical_ to those in the production environment _for as long as_ the exact version of CubeServer remains constant - **Other things should also be considered** for replacement with newer software. Consider: - Python version - Javascript dependencies - Other Tools _If any problem is to be found with keeping this convention, this issue should be reopened for discussion._
main
pin exact versions of dependencies dependency versions will be pinned to prevent dependency updates from screwing with functionality and or causing bugs scroll convention establishment scroll henceforth the following shall be performed prior to every major release x new dependency versions should be sought after considered and tested dependencies should be pinned to the exact version in the docker requirements txt files as well as looser requirements specified in the src setup py files exact dependency pinning docker cubeserver api requirements txt docker cubeserver app requirements txt docker cubeserver beaconserver requirements txt approximate dependency compatibility src cubeserver api setup py src cubeserver app setup py src cubeserver beaconserver setup py src cubeserver common setup py docker container dependencies should be pinned with an exact minor version x y as well in every docker dockerfile file affected files docker cubeserver api dockerfile docker cubeserver app dockerfile docker cubeserver beaconserver dockerfile docker access point dockerfile if issues arise patches and minor releases may modify the required versions so long as the exact versions are specified so as to ensure that the dependencies of the development instance are identical to those in the production environment for as long as the exact version of cubeserver remains constant other things should also be considered for replacement with newer software consider python version javascript dependencies other tools if any problem is to be found with keeping this convention this issue should be reopened for discussion
1
5,262
26,621,056,519
IssuesEvent
2023-01-24 11:15:08
NaluKit/nalu
https://api.github.com/repos/NaluKit/nalu
closed
use INSTANCE instead .get() to get a factory instance
maintainance
In some cases factories provide a `get()`-method to provide a facotry instnace other a static `INSTANCE`-variable. Change nalu in this way, that all factories use a a static `INSTANCE`-variable.
True
use INSTANCE instead .get() to get a factory instance - In some cases factories provide a `get()`-method to provide a facotry instnace other a static `INSTANCE`-variable. Change nalu in this way, that all factories use a a static `INSTANCE`-variable.
main
use instance instead get to get a factory instance in some cases factories provide a get method to provide a facotry instnace other a static instance variable change nalu in this way that all factories use a a static instance variable
1
3,043
11,297,456,986
IssuesEvent
2020-01-17 06:05:16
microsoft/DirectXTK
https://api.github.com/repos/microsoft/DirectXTK
closed
directxtk build failure for x86-uwp #9486
maintainence
#9486 "C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln" (Rebuild target) (1) -> (ValidateSolutionConfiguration target) -> C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln.metaproj : error MSB4126: The specified solution configuration "Release|Win32" is invalid. Please specify a valid solution configuration using the Configuration and Platform properties (e.g. MSBuild.exe Solution.sln /p:Configuration=Debug /p:Platform="Any CPU") or leave those properties blank to use the default solution configuration. [C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln] 0 Warning(s) 1 Error(s) Build FAILED. "C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln" (Rebuild target) (1) -> (ValidateSolutionConfiguration target) -> C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln.metaproj : error MSB4126: The specified solution configuration "Release|Win32" is invalid. Please specify a valid solution configuration using the Configuration and Platform properties (e.g. MSBuild.exe Solution.sln /p:Configuration=Debug /p:Platform="Any CPU") or leave those properties blank to use the default solution configuration. [C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln] 0 Warning(s) 1 Error(s)
True
directxtk build failure for x86-uwp #9486 - #9486 "C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln" (Rebuild target) (1) -> (ValidateSolutionConfiguration target) -> C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln.metaproj : error MSB4126: The specified solution configuration "Release|Win32" is invalid. Please specify a valid solution configuration using the Configuration and Platform properties (e.g. MSBuild.exe Solution.sln /p:Configuration=Debug /p:Platform="Any CPU") or leave those properties blank to use the default solution configuration. [C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln] 0 Warning(s) 1 Error(s) Build FAILED. "C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln" (Rebuild target) (1) -> (ValidateSolutionConfiguration target) -> C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln.metaproj : error MSB4126: The specified solution configuration "Release|Win32" is invalid. Please specify a valid solution configuration using the Configuration and Platform properties (e.g. MSBuild.exe Solution.sln /p:Configuration=Debug /p:Platform="Any CPU") or leave those properties blank to use the default solution configuration. [C:\MyDevProjects\MyProjectsThirdPartyLibs\vcpkg2\vcpkg\buildtrees\directxtk\src\dec2019-b08f6255ba\DirectXTK_Windows10_2019.sln] 0 Warning(s) 1 Error(s)
main
directxtk build failure for uwp c mydevprojects myprojectsthirdpartylibs vcpkg buildtrees directxtk src directxtk sln rebuild target validatesolutionconfiguration target c mydevprojects myprojectsthirdpartylibs vcpkg buildtrees directxtk src directxtk sln metaproj error the specified solution configuration release is invalid please specify a valid solution configuration using the configuration and platform properties e g msbuild exe solution sln p configuration debug p platform any cpu or leave those properties blank to use the default solution configuration warning s error s build failed c mydevprojects myprojectsthirdpartylibs vcpkg buildtrees directxtk src directxtk sln rebuild target validatesolutionconfiguration target c mydevprojects myprojectsthirdpartylibs vcpkg buildtrees directxtk src directxtk sln metaproj error the specified solution configuration release is invalid please specify a valid solution configuration using the configuration and platform properties e g msbuild exe solution sln p configuration debug p platform any cpu or leave those properties blank to use the default solution configuration warning s error s
1
331,635
10,075,749,943
IssuesEvent
2019-07-24 14:51:16
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
login.live.com - site is not usable
browser-firefox-mobile engine-gecko priority-critical
<!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> **URL**: https://login.live.com/login.srf?wa=wsignin1.0 **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: slow and unresponsive **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
login.live.com - site is not usable - <!-- @browser: Firefox Mobile 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: --> **URL**: https://login.live.com/login.srf?wa=wsignin1.0 **Browser / Version**: Firefox Mobile 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: slow and unresponsive **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
login live com site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description slow and unresponsive steps to reproduce browser configuration none from with ❤️
0
330,915
28,495,527,170
IssuesEvent
2023-04-18 13:58:20
NationalSecurityAgency/skills-service
https://api.github.com/repos/NationalSecurityAgency/skills-service
closed
Admin users are not able to view quiz status or run a quiz on behalf of another user
bug test
- navigate to user's Skills Display preview (``/administrator/projects/<project-id>/subjects/<subject-id>/skills/<skill-id>/users/<user-id>/`` - drill down to a skill that has a quiz or a survey assigned to it - attempt to take the survey/quiz
1.0
Admin users are not able to view quiz status or run a quiz on behalf of another user - - navigate to user's Skills Display preview (``/administrator/projects/<project-id>/subjects/<subject-id>/skills/<skill-id>/users/<user-id>/`` - drill down to a skill that has a quiz or a survey assigned to it - attempt to take the survey/quiz
non_main
admin users are not able to view quiz status or run a quiz on behalf of another user navigate to user s skills display preview administrator projects subjects skills users drill down to a skill that has a quiz or a survey assigned to it attempt to take the survey quiz
0
32,608
13,890,852,656
IssuesEvent
2020-10-19 09:50:19
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
Support for composite indexes in `azurerm_cosmosdb_sql_container`
enhancement service/cosmosdb
Please add support for composite indexes in `azurerm_cosmosdb_sql_container`. I'll probably be working on this myself, but opening up an issue in case someone feels compelled as well :)
1.0
Support for composite indexes in `azurerm_cosmosdb_sql_container` - Please add support for composite indexes in `azurerm_cosmosdb_sql_container`. I'll probably be working on this myself, but opening up an issue in case someone feels compelled as well :)
non_main
support for composite indexes in azurerm cosmosdb sql container please add support for composite indexes in azurerm cosmosdb sql container i ll probably be working on this myself but opening up an issue in case someone feels compelled as well
0
35,301
7,929,391,818
IssuesEvent
2018-07-06 14:56:01
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
to change the level at a menu - starange sort. in the menu
No Code Attached Yet
### Steps to reproduce the issue 1 open a menu 2 change the level (i do from 2 to 1) ### Expected result Error: BIGINT UNSIGNED value is out of range in '(`maniacsql30`.`#__menu`.`level` + -(1))' ### Actual result strange sort in the menu. different to the sort in the table xx_menu ### System information (as much as possible) 3.8.10 - linux - PHP erstellt für | Linux cobath-5723-1 3.16.0-4-amd64 #1 SMP Debian 3.16.51-2 (2017-12-03) x86_64 -- | -- Database Type | mysql Datenbankversion | 5.5.60-0+deb8u1 Datenbankzeichensatz | latin1_swedish_ci Datenbankverbindungszeichensatz | utf8mb4_general_ci PHP-Version | 7.0.29 Webserver | Apache/2.4.10 (Debian) PHP-Interface für den Webserver | cgi-fcgi Joomla!-Version | Joomla! 3.8.10 Stable [ Amani ] 26-June-2018 15:45 GMT Joomla!-Plattform-Version | Joomla Platform 13.1.0 Stable [ Curiosity ] 24-Apr-2013 00:00 GMT ### Additional comments
1.0
to change the level at a menu - starange sort. in the menu - ### Steps to reproduce the issue 1 open a menu 2 change the level (i do from 2 to 1) ### Expected result Error: BIGINT UNSIGNED value is out of range in '(`maniacsql30`.`#__menu`.`level` + -(1))' ### Actual result strange sort in the menu. different to the sort in the table xx_menu ### System information (as much as possible) 3.8.10 - linux - PHP erstellt für | Linux cobath-5723-1 3.16.0-4-amd64 #1 SMP Debian 3.16.51-2 (2017-12-03) x86_64 -- | -- Database Type | mysql Datenbankversion | 5.5.60-0+deb8u1 Datenbankzeichensatz | latin1_swedish_ci Datenbankverbindungszeichensatz | utf8mb4_general_ci PHP-Version | 7.0.29 Webserver | Apache/2.4.10 (Debian) PHP-Interface für den Webserver | cgi-fcgi Joomla!-Version | Joomla! 3.8.10 Stable [ Amani ] 26-June-2018 15:45 GMT Joomla!-Plattform-Version | Joomla Platform 13.1.0 Stable [ Curiosity ] 24-Apr-2013 00:00 GMT ### Additional comments
non_main
to change the level at a menu starange sort in the menu steps to reproduce the issue open a menu change the level i do from to expected result error bigint unsigned value is out of range in menu level actual result strange sort in the menu different to the sort in the table xx menu system information as much as possible linux php erstellt für linux cobath smp debian database type mysql datenbankversion datenbankzeichensatz swedish ci datenbankverbindungszeichensatz general ci php version webserver apache debian php interface für den webserver cgi fcgi joomla version joomla stable june gmt joomla plattform version joomla platform stable apr gmt additional comments
0
120,881
25,886,812,881
IssuesEvent
2022-12-14 15:05:50
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
[Factions] Hammerhead Matriarch Genes are experiencing incredibly odd behaviour
Bug Need more info Code Severe
### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? ### The following is a recollection of my experiences shared in the dedicated channel. I noticed that my Hammerhead Matriarch genes were not working at all. I had combined Tiger Thresher (100%) genes with Hammerhead Matriarch (100%, and will be referred to as simply "Matriarch") genes with a tainted result (Decrepify). After suffering blunt force trauma and lacerations from my crewmember, the wounds did not heal after extended periods of time had passed. The first images show this. ![image](https://user-images.githubusercontent.com/105495756/201224174-4356b6fb-1ac6-4bc4-995b-cfcd7b5ffc7f.png) _Figure 1. Evidence of Matriarch genes_ ![image](https://user-images.githubusercontent.com/105495756/201224285-23234865-2486-494b-a541-3f9b438cdd43.png) _Figure 2. Evidence of injuries. Taken at 21:53_ After taking further damage and waiting for roughly three minutes, the wounds did not heal. ![image](https://user-images.githubusercontent.com/105495756/201224612-35db44b8-b9a5-461b-958a-68b4434f0591.png) _Figure 3. Evidence of further injuries. Taken at 21:56_ My crewmate also used Matriarch genes. His were working as intended. I checked my Tiger Thresher genes, they were also working as intended. The issue was solely with my Matriarch genes. The following is an exert taken directly from my comments in the discord channel: _"I've used opium on myself to see if that resets it further tests show that his healing works rapidly whilst mine is non existent perhaps there's an issue with gene placement? the only thing I did differently was when giving him his genes, I put the genes in the genetic splicer and then gave it to him whereas with my genes, I put the genes into the genetic splicer within my head that's the only thing I did differently. Would that break the genes' functionality? The OTHER thing that's super weird is that the Tiger Thresher genes which I'm using in combination work perfectly fine so only half of the genetic content is active"_ Upon starting a new mission, however, the ill-effects completely subsided and acted normal. When we docked at the next outpost, there were further issues with genes. Notably, the healing rate of the genes was significantly lower in my character than previously observed. In the case of my crewmember, he reported "rubber-banding healing" of his health. Upon further inspection, it was revealed that his health was healing from 68% bite wounds down to 65% bite wounds, only to reset to 68% for half a frame and jump down to 67%. ![image](https://user-images.githubusercontent.com/105495756/201225521-37e3bb85-e4b8-463d-b22b-2b22d31769e4.png) _Figure 4. Initial health using health HUD_ ![image](https://user-images.githubusercontent.com/105495756/201225401-5705b0cf-8c2a-433e-a0f8-b8a8935fffad.png) _Figure 5. Evidence of wounds healing_ ![image](https://user-images.githubusercontent.com/105495756/201225442-57616e6b-4213-4cad-8729-a909086accd3.png) _Figure 6. Evidence of wounds resetting_ In complete honesty, I have absolutely no idea what caused this or how to reproduce it. I do not know the scale of which it works upon, whether it is exclusive to Matriarch genes or if it has wider reaching influences. I am hoping someone can shed light on this issue or a dev more knowledgeable in these areas can work out what is going on. ### Reproduction steps _No response_ ### Bug prevalence Happens regularly ### Version Faction test branch ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
1.0
[Factions] Hammerhead Matriarch Genes are experiencing incredibly odd behaviour - ### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? ### The following is a recollection of my experiences shared in the dedicated channel. I noticed that my Hammerhead Matriarch genes were not working at all. I had combined Tiger Thresher (100%) genes with Hammerhead Matriarch (100%, and will be referred to as simply "Matriarch") genes with a tainted result (Decrepify). After suffering blunt force trauma and lacerations from my crewmember, the wounds did not heal after extended periods of time had passed. The first images show this. ![image](https://user-images.githubusercontent.com/105495756/201224174-4356b6fb-1ac6-4bc4-995b-cfcd7b5ffc7f.png) _Figure 1. Evidence of Matriarch genes_ ![image](https://user-images.githubusercontent.com/105495756/201224285-23234865-2486-494b-a541-3f9b438cdd43.png) _Figure 2. Evidence of injuries. Taken at 21:53_ After taking further damage and waiting for roughly three minutes, the wounds did not heal. ![image](https://user-images.githubusercontent.com/105495756/201224612-35db44b8-b9a5-461b-958a-68b4434f0591.png) _Figure 3. Evidence of further injuries. Taken at 21:56_ My crewmate also used Matriarch genes. His were working as intended. I checked my Tiger Thresher genes, they were also working as intended. The issue was solely with my Matriarch genes. The following is an exert taken directly from my comments in the discord channel: _"I've used opium on myself to see if that resets it further tests show that his healing works rapidly whilst mine is non existent perhaps there's an issue with gene placement? the only thing I did differently was when giving him his genes, I put the genes in the genetic splicer and then gave it to him whereas with my genes, I put the genes into the genetic splicer within my head that's the only thing I did differently. Would that break the genes' functionality? The OTHER thing that's super weird is that the Tiger Thresher genes which I'm using in combination work perfectly fine so only half of the genetic content is active"_ Upon starting a new mission, however, the ill-effects completely subsided and acted normal. When we docked at the next outpost, there were further issues with genes. Notably, the healing rate of the genes was significantly lower in my character than previously observed. In the case of my crewmember, he reported "rubber-banding healing" of his health. Upon further inspection, it was revealed that his health was healing from 68% bite wounds down to 65% bite wounds, only to reset to 68% for half a frame and jump down to 67%. ![image](https://user-images.githubusercontent.com/105495756/201225521-37e3bb85-e4b8-463d-b22b-2b22d31769e4.png) _Figure 4. Initial health using health HUD_ ![image](https://user-images.githubusercontent.com/105495756/201225401-5705b0cf-8c2a-433e-a0f8-b8a8935fffad.png) _Figure 5. Evidence of wounds healing_ ![image](https://user-images.githubusercontent.com/105495756/201225442-57616e6b-4213-4cad-8729-a909086accd3.png) _Figure 6. Evidence of wounds resetting_ In complete honesty, I have absolutely no idea what caused this or how to reproduce it. I do not know the scale of which it works upon, whether it is exclusive to Matriarch genes or if it has wider reaching influences. I am hoping someone can shed light on this issue or a dev more knowledgeable in these areas can work out what is going on. ### Reproduction steps _No response_ ### Bug prevalence Happens regularly ### Version Faction test branch ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
non_main
hammerhead matriarch genes are experiencing incredibly odd behaviour disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened the following is a recollection of my experiences shared in the dedicated channel i noticed that my hammerhead matriarch genes were not working at all i had combined tiger thresher genes with hammerhead matriarch and will be referred to as simply matriarch genes with a tainted result decrepify after suffering blunt force trauma and lacerations from my crewmember the wounds did not heal after extended periods of time had passed the first images show this figure evidence of matriarch genes figure evidence of injuries taken at after taking further damage and waiting for roughly three minutes the wounds did not heal figure evidence of further injuries taken at my crewmate also used matriarch genes his were working as intended i checked my tiger thresher genes they were also working as intended the issue was solely with my matriarch genes the following is an exert taken directly from my comments in the discord channel i ve used opium on myself to see if that resets it further tests show that his healing works rapidly whilst mine is non existent perhaps there s an issue with gene placement the only thing i did differently was when giving him his genes i put the genes in the genetic splicer and then gave it to him whereas with my genes i put the genes into the genetic splicer within my head that s the only thing i did differently would that break the genes functionality the other thing that s super weird is that the tiger thresher genes which i m using in combination work perfectly fine so only half of the genetic content is active upon starting a new mission however the ill effects completely subsided and acted normal when we docked at the next outpost there were further issues with genes notably the healing rate of the genes was significantly lower in my character than previously observed in the case of my crewmember he reported rubber banding healing of his health upon further inspection it was revealed that his health was healing from bite wounds down to bite wounds only to reset to for half a frame and jump down to figure initial health using health hud figure evidence of wounds healing figure evidence of wounds resetting in complete honesty i have absolutely no idea what caused this or how to reproduce it i do not know the scale of which it works upon whether it is exclusive to matriarch genes or if it has wider reaching influences i am hoping someone can shed light on this issue or a dev more knowledgeable in these areas can work out what is going on reproduction steps no response bug prevalence happens regularly version faction test branch no response which operating system did you encounter this bug on windows relevant error messages and crash reports no response
0
50,858
13,608,189,155
IssuesEvent
2020-09-23 01:38:56
peilin823/peilin823-security_test_demo1
https://api.github.com/repos/peilin823/peilin823-security_test_demo1
opened
CVE-2017-16113 (High) detected in parsejson-0.0.3.tgz
security vulnerability
## CVE-2017-16113 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parsejson-0.0.3.tgz</b></p></summary> <p>Method that parses a JSON string and returns a JSON object</p> <p>Library home page: <a href="https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz">https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz</a></p> <p>Path to dependency file: peilin823-security_test_demo1/package.json</p> <p>Path to vulnerable library: peilin823-security_test_demo1/node_modules/parsejson/package.json</p> <p> Dependency Hierarchy: - reveal.js-3.4.1.tgz (Root Library) - socket.io-1.7.4.tgz - socket.io-client-1.7.4.tgz - engine.io-client-1.8.5.tgz - :x: **parsejson-0.0.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/peilin823/peilin823-security_test_demo1/commit/9cbc07e9a44117e1b35984fd315cbd4aec36fd41">9cbc07e9a44117e1b35984fd315cbd4aec36fd41</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16113>CVE-2017-16113</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-16113 (High) detected in parsejson-0.0.3.tgz - ## CVE-2017-16113 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parsejson-0.0.3.tgz</b></p></summary> <p>Method that parses a JSON string and returns a JSON object</p> <p>Library home page: <a href="https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz">https://registry.npmjs.org/parsejson/-/parsejson-0.0.3.tgz</a></p> <p>Path to dependency file: peilin823-security_test_demo1/package.json</p> <p>Path to vulnerable library: peilin823-security_test_demo1/node_modules/parsejson/package.json</p> <p> Dependency Hierarchy: - reveal.js-3.4.1.tgz (Root Library) - socket.io-1.7.4.tgz - socket.io-client-1.7.4.tgz - engine.io-client-1.8.5.tgz - :x: **parsejson-0.0.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/peilin823/peilin823-security_test_demo1/commit/9cbc07e9a44117e1b35984fd315cbd4aec36fd41">9cbc07e9a44117e1b35984fd315cbd4aec36fd41</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16113>CVE-2017-16113</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in parsejson tgz cve high severity vulnerability vulnerable library parsejson tgz method that parses a json string and returns a json object library home page a href path to dependency file security test package json path to vulnerable library security test node modules parsejson package json dependency hierarchy reveal js tgz root library socket io tgz socket io client tgz engine io client tgz x parsejson tgz vulnerable library found in head commit a href found in base branch master vulnerability details the parsejson module is vulnerable to regular expression denial of service when untrusted user input is passed into it to be parsed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource
0
1,294
5,477,711,492
IssuesEvent
2017-03-12 11:22:56
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
lvol is trying to shrink logical volume with fixed size when executed twice
affects_2.2 bug_report waiting_on_maintainer
ISSUE TYPE: Bug Report COMPONENT NAME: lvol ANSIBLE VERSION: ansible 2.2.0.0 CONFIGURATION: no changes OS / ENVIRONMENT: Ubuntu xenial 16.04 (Linux 4.4.0-47-generic #68-Ubuntu SMP x86_64) - ##### SUMMARY lvol is creating the logical volume correctly at the first run, but after running the same playbook again without modifications, an error message is displayed: "Sorry, no shrinking of lv_default without force=yes." ##### STEPS TO REPRODUCE 1) Log on an ec2 instance with ansible 2.2 installed and ec2 dynamic inventory enabled 2) Create lvm.yml file with the following content: ``` - hosts: localhost tasks: - name: Get ec2 facts local_action: ec2_facts - name: Create and attach EBS volume ec2_vol: volume_size: 1 region: us-east-1 instance: "{{ ansible_ec2_instance_id }}" device_name: /dev/xvdz - name: Create LVM volume group become: yes lvg: vg: vg_default pvs: /dev/xvdz - name: Create LVM logical volumes inside the volume group become: yes lvol: vg=vg_default lv=lv_default size="1013m" ``` 3) Execute `ansible-playbook lvm-bug.yml` - the volume will be created 3) Execute the same playbook again - an error message will be displayed: "Sorry, no shrinking of lv_default without force=yes." ##### EXPECTED RESULTS Volume should be created at the first run, and subsequent runs should have no effect and no error messages should be displayed ##### ACTUAL RESULTS An error message is displayed: ``` talves@ip-10-69-11-150:~$ ansible-playbook lvm-bug.yml -vvvvv Using /etc/ansible/ansible.cfg as config file [WARNING]: provided hosts list is empty, only localhost is available Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/__init__.pyc PLAYBOOK: lvm-bug.yml ********************************************************** 1 plays in lvm-bug.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/system/setup.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139 `" && echo ansible-tmp-1479749874.91-104034913293139="` echo $HOME/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpv78jbF TO /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/setup.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/ /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/setup.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/setup.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Get ec2 facts] *********************************************************** task path: /home/talves/lvm-bug.yml:3 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/cloud/amazon/ec2_facts.py <localhost> ESTABLISH LOCAL CONNECTION FOR USER: talves <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672 `" && echo ansible-tmp-1479749875.36-75305735143672="` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672 `" ) && sleep 0' <localhost> PUT /tmp/tmp4zkGha TO /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ec2_facts.py <localhost> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ec2_facts.py && sleep 0' <localhost> EXEC /bin/sh -c '/usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ec2_facts.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/" > /dev/null 2>&1 && sleep 0' ok: [localhost -> localhost] => { "ansible_facts": { "ansible_ec2_ami_id": "ami-78b4976f", "ansible_ec2_ami_launch_index": "0", "ansible_ec2_ami_manifest_path": "(unknown)", "ansible_ec2_block_device_mapping_ami": "/dev/sda1", "ansible_ec2_block_device_mapping_root": "/dev/sda1", "ansible_ec2_hostname": "ip-10-69-11-150.devfactory.local", "ansible_ec2_instance_action": "none", "ansible_ec2_instance_id": "i-0c7d71e6ac0595b17", "ansible_ec2_instance_type": "t2.micro", "ansible_ec2_local_hostname": "ip-10-69-11-150.devfactory.local", "ansible_ec2_local_ipv4": "10.69.11.150", "ansible_ec2_mac": "12:27:d0:65:16:aa", "ansible_ec2_metrics_vhostmd": "<?xml version=\"1.0\" encoding=\"UTF-8\"?>", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_device_number": "0", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_interface_id": "eni-8f846c70", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_ipv4_associations_34.192.201.42": "10.69.11.150", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_local_hostname": "ip-10-69-11-150.devfactory.local", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_local_ipv4s": "10.69.11.150", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_mac": "12:27:d0:65:16:aa", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_owner_id": "646253092271", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_public_hostname": "ec2-34-192-201-42.compute-1.amazonaws.com", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_public_ipv4s": "34.192.201.42", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_security_group_ids": "sg-ed3d1296", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_security_groups": "ecs-private", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_subnet_id": "subnet-d64c60fc", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_subnet_ipv4_cidr_block": "10.69.11.0/24", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_vpc_id": "vpc-a79317c0", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_vpc_ipv4_cidr_block": "10.69.0.0/20", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_vpc_ipv4_cidr_blocks": "10.69.0.0/20", "ansible_ec2_placement_availability_zone": "us-east-1c", "ansible_ec2_placement_region": "us-east-1", "ansible_ec2_profile": "default-hvm", "ansible_ec2_public_hostname": "ec2-34-192-201-42.compute-1.amazonaws.com", "ansible_ec2_public_ipv4": "34.192.201.42", "ansible_ec2_public_key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCihoNkm/3LelqzXoqJtZ1ORi5EGc97+03Q5+pBeWxJ7Lf14UemxYJEWWXlQo+RbebvsSBAha0o8sfPWg7qJEOyMRWITQG0w9K+eX362Bl2IY0EY2cFzj+gXdZdMlUTZTGR/C3g2YmzmZxE35YwEfJ/BcAEhyKf6L4x9bxvOTVPPa6a1+wsqLlIIy+1Nh4aTLGlEwnCfpA5d+TbTB+H505Pu1m82x9CO5u5aOPG5ZMPX8xwA3zzPepLbQsH/HSK5w2WPGB/Fe2yA2j0vDwiOhtcx9SG+SZMZtzs5rvn7QbHSnTjPp5oNUjj1JTXbKX6JatdZT2uUCmCilO7K2yZO7Tt central-docker-team\n", "ansible_ec2_reservation_id": "r-0b3aab7de7254c9bd", "ansible_ec2_security_groups": "ecs-private", "ansible_ec2_services_domain": "amazonaws.com", "ansible_ec2_services_partition": "aws", "ansible_ec2_user_data": null }, "changed": false, "invocation": { "module_args": { "force": false, "force_basic_auth": false, "http_agent": "ansible-httpget", "url": null, "url_password": null, "url_username": null, "use_proxy": true, "validate_certs": true }, "module_name": "ec2_facts" } } TASK [Create and attach EBS volume] ******************************************** task path: /home/talves/lvm-bug.yml:6 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/cloud/amazon/ec2_vol.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352 `" && echo ansible-tmp-1479749875.76-182263987502352="` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmp3YBDj6 TO /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ec2_vol.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ec2_vol.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ec2_vol.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => { "changed": false, "device": "/dev/xvdz", "invocation": { "module_args": { "aws_access_key": null, "aws_secret_key": null, "delete_on_termination": false, "device_name": "/dev/xvdz", "ec2_url": null, "encrypted": false, "id": null, "instance": "i-0c7d71e6ac0595b17", "iops": null, "name": null, "profile": null, "region": "us-east-1", "security_token": null, "snapshot": null, "state": "present", "validate_certs": true, "volume_size": "1", "volume_type": "standard", "zone": null }, "module_name": "ec2_vol" }, "msg": "Volume mapping for /dev/xvdz already exists on instance i-0c7d71e6ac0595b17", "volume_id": "vol-606c80f3" } TASK [Create LVM volume group] ************************************************* task path: /home/talves/lvm-bug.yml:13 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/extras/system/lvg.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141 `" && echo ansible-tmp-1479749876.26-228386796421141="` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpHpRJZq TO /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/lvg.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/ /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/lvg.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-ptheuhxyltmmgljgmwhjwolcddmexicc; /usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/lvg.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/" > /dev/null 2>&1'"'"' && sleep 0' ok: [localhost] => { "changed": false, "invocation": { "module_args": { "force": false, "pesize": 4, "pvs": [ "/dev/xvdz" ], "state": "present", "vg": "vg_default", "vg_options": "" }, "module_name": "lvg" } } TASK [Create LVM logical volumes inside the volume group] ********************** task path: /home/talves/lvm-bug.yml:19 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/extras/system/lvol.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585 `" && echo ansible-tmp-1479749876.63-200489056793585="` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpfKsPsA TO /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/lvol.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/ /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/lvol.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-dvukcdtkvvbvnxuopibtdcxeczhjasyt; /usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/lvol.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/" > /dev/null 2>&1'"'"' && sleep 0' fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "active": true, "force": false, "lv": "lv_default", "opts": null, "pvs": null, "shrink": true, "size": "1013m", "snapshot": null, "state": "present", "vg": "vg_default" }, "module_name": "lvol" }, "msg": "Sorry, no shrinking of lv_default without force=yes." } to retry, use: --limit @/home/talves/lvm-bug.retry PLAY RECAP ********************************************************************* localhost : ok=4 changed=0 unreachable=0 failed=1 ```
True
lvol is trying to shrink logical volume with fixed size when executed twice - ISSUE TYPE: Bug Report COMPONENT NAME: lvol ANSIBLE VERSION: ansible 2.2.0.0 CONFIGURATION: no changes OS / ENVIRONMENT: Ubuntu xenial 16.04 (Linux 4.4.0-47-generic #68-Ubuntu SMP x86_64) - ##### SUMMARY lvol is creating the logical volume correctly at the first run, but after running the same playbook again without modifications, an error message is displayed: "Sorry, no shrinking of lv_default without force=yes." ##### STEPS TO REPRODUCE 1) Log on an ec2 instance with ansible 2.2 installed and ec2 dynamic inventory enabled 2) Create lvm.yml file with the following content: ``` - hosts: localhost tasks: - name: Get ec2 facts local_action: ec2_facts - name: Create and attach EBS volume ec2_vol: volume_size: 1 region: us-east-1 instance: "{{ ansible_ec2_instance_id }}" device_name: /dev/xvdz - name: Create LVM volume group become: yes lvg: vg: vg_default pvs: /dev/xvdz - name: Create LVM logical volumes inside the volume group become: yes lvol: vg=vg_default lv=lv_default size="1013m" ``` 3) Execute `ansible-playbook lvm-bug.yml` - the volume will be created 3) Execute the same playbook again - an error message will be displayed: "Sorry, no shrinking of lv_default without force=yes." ##### EXPECTED RESULTS Volume should be created at the first run, and subsequent runs should have no effect and no error messages should be displayed ##### ACTUAL RESULTS An error message is displayed: ``` talves@ip-10-69-11-150:~$ ansible-playbook lvm-bug.yml -vvvvv Using /etc/ansible/ansible.cfg as config file [WARNING]: provided hosts list is empty, only localhost is available Loading callback plugin default of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/__init__.pyc PLAYBOOK: lvm-bug.yml ********************************************************** 1 plays in lvm-bug.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/system/setup.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139 `" && echo ansible-tmp-1479749874.91-104034913293139="` echo $HOME/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpv78jbF TO /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/setup.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/ /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/setup.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/setup.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749874.91-104034913293139/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [Get ec2 facts] *********************************************************** task path: /home/talves/lvm-bug.yml:3 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/cloud/amazon/ec2_facts.py <localhost> ESTABLISH LOCAL CONNECTION FOR USER: talves <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672 `" && echo ansible-tmp-1479749875.36-75305735143672="` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672 `" ) && sleep 0' <localhost> PUT /tmp/tmp4zkGha TO /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ec2_facts.py <localhost> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ec2_facts.py && sleep 0' <localhost> EXEC /bin/sh -c '/usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/ec2_facts.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749875.36-75305735143672/" > /dev/null 2>&1 && sleep 0' ok: [localhost -> localhost] => { "ansible_facts": { "ansible_ec2_ami_id": "ami-78b4976f", "ansible_ec2_ami_launch_index": "0", "ansible_ec2_ami_manifest_path": "(unknown)", "ansible_ec2_block_device_mapping_ami": "/dev/sda1", "ansible_ec2_block_device_mapping_root": "/dev/sda1", "ansible_ec2_hostname": "ip-10-69-11-150.devfactory.local", "ansible_ec2_instance_action": "none", "ansible_ec2_instance_id": "i-0c7d71e6ac0595b17", "ansible_ec2_instance_type": "t2.micro", "ansible_ec2_local_hostname": "ip-10-69-11-150.devfactory.local", "ansible_ec2_local_ipv4": "10.69.11.150", "ansible_ec2_mac": "12:27:d0:65:16:aa", "ansible_ec2_metrics_vhostmd": "<?xml version=\"1.0\" encoding=\"UTF-8\"?>", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_device_number": "0", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_interface_id": "eni-8f846c70", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_ipv4_associations_34.192.201.42": "10.69.11.150", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_local_hostname": "ip-10-69-11-150.devfactory.local", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_local_ipv4s": "10.69.11.150", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_mac": "12:27:d0:65:16:aa", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_owner_id": "646253092271", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_public_hostname": "ec2-34-192-201-42.compute-1.amazonaws.com", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_public_ipv4s": "34.192.201.42", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_security_group_ids": "sg-ed3d1296", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_security_groups": "ecs-private", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_subnet_id": "subnet-d64c60fc", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_subnet_ipv4_cidr_block": "10.69.11.0/24", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_vpc_id": "vpc-a79317c0", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_vpc_ipv4_cidr_block": "10.69.0.0/20", "ansible_ec2_network_interfaces_macs_12_27_d0_65_16_aa_vpc_ipv4_cidr_blocks": "10.69.0.0/20", "ansible_ec2_placement_availability_zone": "us-east-1c", "ansible_ec2_placement_region": "us-east-1", "ansible_ec2_profile": "default-hvm", "ansible_ec2_public_hostname": "ec2-34-192-201-42.compute-1.amazonaws.com", "ansible_ec2_public_ipv4": "34.192.201.42", "ansible_ec2_public_key": "ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAABAQCihoNkm/3LelqzXoqJtZ1ORi5EGc97+03Q5+pBeWxJ7Lf14UemxYJEWWXlQo+RbebvsSBAha0o8sfPWg7qJEOyMRWITQG0w9K+eX362Bl2IY0EY2cFzj+gXdZdMlUTZTGR/C3g2YmzmZxE35YwEfJ/BcAEhyKf6L4x9bxvOTVPPa6a1+wsqLlIIy+1Nh4aTLGlEwnCfpA5d+TbTB+H505Pu1m82x9CO5u5aOPG5ZMPX8xwA3zzPepLbQsH/HSK5w2WPGB/Fe2yA2j0vDwiOhtcx9SG+SZMZtzs5rvn7QbHSnTjPp5oNUjj1JTXbKX6JatdZT2uUCmCilO7K2yZO7Tt central-docker-team\n", "ansible_ec2_reservation_id": "r-0b3aab7de7254c9bd", "ansible_ec2_security_groups": "ecs-private", "ansible_ec2_services_domain": "amazonaws.com", "ansible_ec2_services_partition": "aws", "ansible_ec2_user_data": null }, "changed": false, "invocation": { "module_args": { "force": false, "force_basic_auth": false, "http_agent": "ansible-httpget", "url": null, "url_password": null, "url_username": null, "use_proxy": true, "validate_certs": true }, "module_name": "ec2_facts" } } TASK [Create and attach EBS volume] ******************************************** task path: /home/talves/lvm-bug.yml:6 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/cloud/amazon/ec2_vol.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352 `" && echo ansible-tmp-1479749875.76-182263987502352="` echo $HOME/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmp3YBDj6 TO /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ec2_vol.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ec2_vol.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/ec2_vol.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749875.76-182263987502352/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => { "changed": false, "device": "/dev/xvdz", "invocation": { "module_args": { "aws_access_key": null, "aws_secret_key": null, "delete_on_termination": false, "device_name": "/dev/xvdz", "ec2_url": null, "encrypted": false, "id": null, "instance": "i-0c7d71e6ac0595b17", "iops": null, "name": null, "profile": null, "region": "us-east-1", "security_token": null, "snapshot": null, "state": "present", "validate_certs": true, "volume_size": "1", "volume_type": "standard", "zone": null }, "module_name": "ec2_vol" }, "msg": "Volume mapping for /dev/xvdz already exists on instance i-0c7d71e6ac0595b17", "volume_id": "vol-606c80f3" } TASK [Create LVM volume group] ************************************************* task path: /home/talves/lvm-bug.yml:13 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/extras/system/lvg.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141 `" && echo ansible-tmp-1479749876.26-228386796421141="` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpHpRJZq TO /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/lvg.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/ /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/lvg.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-ptheuhxyltmmgljgmwhjwolcddmexicc; /usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/lvg.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749876.26-228386796421141/" > /dev/null 2>&1'"'"' && sleep 0' ok: [localhost] => { "changed": false, "invocation": { "module_args": { "force": false, "pesize": 4, "pvs": [ "/dev/xvdz" ], "state": "present", "vg": "vg_default", "vg_options": "" }, "module_name": "lvg" } } TASK [Create LVM logical volumes inside the volume group] ********************** task path: /home/talves/lvm-bug.yml:19 Using module file /usr/lib/python2.7/dist-packages/ansible/modules/extras/system/lvol.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: talves <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585 `" && echo ansible-tmp-1479749876.63-200489056793585="` echo $HOME/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpfKsPsA TO /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/lvol.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/ /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/lvol.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c 'sudo -H -S -n -u root /bin/sh -c '"'"'echo BECOME-SUCCESS-dvukcdtkvvbvnxuopibtdcxeczhjasyt; /usr/bin/python /home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/lvol.py; rm -rf "/home/talves/.ansible/tmp/ansible-tmp-1479749876.63-200489056793585/" > /dev/null 2>&1'"'"' && sleep 0' fatal: [localhost]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_args": { "active": true, "force": false, "lv": "lv_default", "opts": null, "pvs": null, "shrink": true, "size": "1013m", "snapshot": null, "state": "present", "vg": "vg_default" }, "module_name": "lvol" }, "msg": "Sorry, no shrinking of lv_default without force=yes." } to retry, use: --limit @/home/talves/lvm-bug.retry PLAY RECAP ********************************************************************* localhost : ok=4 changed=0 unreachable=0 failed=1 ```
main
lvol is trying to shrink logical volume with fixed size when executed twice issue type bug report component name lvol ansible version ansible configuration no changes os environment ubuntu xenial linux generic ubuntu smp summary lvol is creating the logical volume correctly at the first run but after running the same playbook again without modifications an error message is displayed sorry no shrinking of lv default without force yes steps to reproduce log on an instance with ansible installed and dynamic inventory enabled create lvm yml file with the following content hosts localhost tasks name get facts local action facts name create and attach ebs volume vol volume size region us east instance ansible instance id device name dev xvdz name create lvm volume group become yes lvg vg vg default pvs dev xvdz name create lvm logical volumes inside the volume group become yes lvol vg vg default lv lv default size execute ansible playbook lvm bug yml the volume will be created execute the same playbook again an error message will be displayed sorry no shrinking of lv default without force yes expected results volume should be created at the first run and subsequent runs should have no effect and no error messages should be displayed actual results an error message is displayed talves ip ansible playbook lvm bug yml vvvvv using etc ansible ansible cfg as config file provided hosts list is empty only localhost is available loading callback plugin default of type stdout from usr lib dist packages ansible plugins callback init pyc playbook lvm bug yml plays in lvm bug yml play task using module file usr lib dist packages ansible modules core system setup py establish local connection for user talves exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home talves ansible tmp ansible tmp setup py exec bin sh c chmod u x home talves ansible tmp ansible tmp home talves ansible tmp ansible tmp setup py sleep exec bin sh c usr bin python home talves ansible tmp ansible tmp setup py rm rf home talves ansible tmp ansible tmp dev null sleep ok task task path home talves lvm bug yml using module file usr lib dist packages ansible modules core cloud amazon facts py establish local connection for user talves exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home talves ansible tmp ansible tmp facts py exec bin sh c chmod u x home talves ansible tmp ansible tmp home talves ansible tmp ansible tmp facts py sleep exec bin sh c usr bin python home talves ansible tmp ansible tmp facts py rm rf home talves ansible tmp ansible tmp dev null sleep ok ansible facts ansible ami id ami ansible ami launch index ansible ami manifest path unknown ansible block device mapping ami dev ansible block device mapping root dev ansible hostname ip devfactory local ansible instance action none ansible instance id i ansible instance type micro ansible local hostname ip devfactory local ansible local ansible mac aa ansible metrics vhostmd ansible network interfaces macs aa device number ansible network interfaces macs aa interface id eni ansible network interfaces macs aa associations ansible network interfaces macs aa local hostname ip devfactory local ansible network interfaces macs aa local ansible network interfaces macs aa mac aa ansible network interfaces macs aa owner id ansible network interfaces macs aa public hostname compute amazonaws com ansible network interfaces macs aa public ansible network interfaces macs aa security group ids sg ansible network interfaces macs aa security groups ecs private ansible network interfaces macs aa subnet id subnet ansible network interfaces macs aa subnet cidr block ansible network interfaces macs aa vpc id vpc ansible network interfaces macs aa vpc cidr block ansible network interfaces macs aa vpc cidr blocks ansible placement availability zone us east ansible placement region us east ansible profile default hvm ansible public hostname compute amazonaws com ansible public ansible public key ssh rsa gxdzdmlutztgr wsqlliiy tbtb central docker team n ansible reservation id r ansible security groups ecs private ansible services domain amazonaws com ansible services partition aws ansible user data null changed false invocation module args force false force basic auth false http agent ansible httpget url null url password null url username null use proxy true validate certs true module name facts task task path home talves lvm bug yml using module file usr lib dist packages ansible modules core cloud amazon vol py establish local connection for user talves exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home talves ansible tmp ansible tmp vol py exec bin sh c chmod u x home talves ansible tmp ansible tmp home talves ansible tmp ansible tmp vol py sleep exec bin sh c usr bin python home talves ansible tmp ansible tmp vol py rm rf home talves ansible tmp ansible tmp dev null sleep ok changed false device dev xvdz invocation module args aws access key null aws secret key null delete on termination false device name dev xvdz url null encrypted false id null instance i iops null name null profile null region us east security token null snapshot null state present validate certs true volume size volume type standard zone null module name vol msg volume mapping for dev xvdz already exists on instance i volume id vol task task path home talves lvm bug yml using module file usr lib dist packages ansible modules extras system lvg py establish local connection for user talves exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmphprjzq to home talves ansible tmp ansible tmp lvg py exec bin sh c chmod u x home talves ansible tmp ansible tmp home talves ansible tmp ansible tmp lvg py sleep exec bin sh c sudo h s n u root bin sh c echo become success ptheuhxyltmmgljgmwhjwolcddmexicc usr bin python home talves ansible tmp ansible tmp lvg py rm rf home talves ansible tmp ansible tmp dev null sleep ok changed false invocation module args force false pesize pvs dev xvdz state present vg vg default vg options module name lvg task task path home talves lvm bug yml using module file usr lib dist packages ansible modules extras system lvol py establish local connection for user talves exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpfkspsa to home talves ansible tmp ansible tmp lvol py exec bin sh c chmod u x home talves ansible tmp ansible tmp home talves ansible tmp ansible tmp lvol py sleep exec bin sh c sudo h s n u root bin sh c echo become success dvukcdtkvvbvnxuopibtdcxeczhjasyt usr bin python home talves ansible tmp ansible tmp lvol py rm rf home talves ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module args active true force false lv lv default opts null pvs null shrink true size snapshot null state present vg vg default module name lvol msg sorry no shrinking of lv default without force yes to retry use limit home talves lvm bug retry play recap localhost ok changed unreachable failed
1