Unnamed: 0
int64 1
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 3
438
| labels
stringlengths 4
308
| body
stringlengths 7
254k
| index
stringclasses 7
values | text_combine
stringlengths 96
254k
| label
stringclasses 2
values | text
stringlengths 96
246k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,983
| 6,694,204,027
|
IssuesEvent
|
2017-10-10 00:18:02
|
duckduckgo/zeroclickinfo-spice
|
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
|
closed
|
Dictionary Definition: raw food
|
Maintainer Input Requested
|
Food that was not heated above 42 Celsius or 107.60000 Fahrenheit
Examples: fresh fruit, vegetables, sprouts (also sprouted grains), grasses
Further examples of raw "gourmet" food: nuts and seeds; soaked grains
---
IA Page: http://duck.co/ia/view/dictionary_definition
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jagtalon
|
True
|
Dictionary Definition: raw food - Food that was not heated above 42 Celsius or 107.60000 Fahrenheit
Examples: fresh fruit, vegetables, sprouts (also sprouted grains), grasses
Further examples of raw "gourmet" food: nuts and seeds; soaked grains
---
IA Page: http://duck.co/ia/view/dictionary_definition
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jagtalon
|
main
|
dictionary definition raw food food that was not heated above celsius or fahrenheit examples fresh fruit vegetables sprouts also sprouted grains grasses further examples of raw gourmet food nuts and seeds soaked grains ia page jagtalon
| 1
|
1,068
| 4,889,235,855
|
IssuesEvent
|
2016-11-18 09:31:52
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
include_role does not support tags
|
affects_2.2 bug_report waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/task/feature -->
include_role
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
Linux, Red Hat Enterprise 7.2
##### SUMMARY
Originally filed this issue here because I was told on IRC this was part of ansible because it was a action plugin. The bot told me i was wrong after a few days, so now it's here.
https://github.com/ansible/ansible/issues/17761
Forgive my terminology. I will do my best to describe that issue I am seeing. I am using a pre-release version of Ansible 2.2 for this.
When using include_role module with tags, the include_role seems to ignore the tags and does not apply them when being called.
I am using include_role in my playbook in the task list. I apply a tag on it. When I call ansible-playbook -t tagname it does not run my include_role task.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
---
- name: test playbook
hosts: localhost
tasks:
- debug:
msg: "hello from debug"
tags:
- test
- include_role:
name: role1
tags:
- test
```
OUTPUT
```
$ ansible-playbook -t test test.yml
PLAY [test playbook] ***********************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [debug] *******************************************************************
ok: [localhost] => {
"msg": "hello from debug"
}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=0 unreachable=0 failed=0
```
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
I would expect it to run the role1, because it was called with the tags
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
role1 was never called
|
True
|
include_role does not support tags - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/task/feature -->
include_role
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
Linux, Red Hat Enterprise 7.2
##### SUMMARY
Originally filed this issue here because I was told on IRC this was part of ansible because it was a action plugin. The bot told me i was wrong after a few days, so now it's here.
https://github.com/ansible/ansible/issues/17761
Forgive my terminology. I will do my best to describe that issue I am seeing. I am using a pre-release version of Ansible 2.2 for this.
When using include_role module with tags, the include_role seems to ignore the tags and does not apply them when being called.
I am using include_role in my playbook in the task list. I apply a tag on it. When I call ansible-playbook -t tagname it does not run my include_role task.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
---
- name: test playbook
hosts: localhost
tasks:
- debug:
msg: "hello from debug"
tags:
- test
- include_role:
name: role1
tags:
- test
```
OUTPUT
```
$ ansible-playbook -t test test.yml
PLAY [test playbook] ***********************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [debug] *******************************************************************
ok: [localhost] => {
"msg": "hello from debug"
}
PLAY RECAP *********************************************************************
localhost : ok=2 changed=0 unreachable=0 failed=0
```
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
I would expect it to run the role1, because it was called with the tags
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
role1 was never called
|
main
|
include role does not support tags issue type bug report component name include role ansible version ansible configuration n a os environment linux red hat enterprise summary originally filed this issue here because i was told on irc this was part of ansible because it was a action plugin the bot told me i was wrong after a few days so now it s here forgive my terminology i will do my best to describe that issue i am seeing i am using a pre release version of ansible for this when using include role module with tags the include role seems to ignore the tags and does not apply them when being called i am using include role in my playbook in the task list i apply a tag on it when i call ansible playbook t tagname it does not run my include role task steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name test playbook hosts localhost tasks debug msg hello from debug tags test include role name tags test output ansible playbook t test test yml play task ok task ok msg hello from debug play recap localhost ok changed unreachable failed expected results i would expect it to run the because it was called with the tags actual results was never called
| 1
|
2,394
| 8,499,960,216
|
IssuesEvent
|
2018-10-29 18:30:42
|
Qo2770/Algorithms
|
https://api.github.com/repos/Qo2770/Algorithms
|
closed
|
Algorithms is searching for Maintainers!
|
Maintainers Wanted
|
*This issue was created by [Maintainers Wanted](https://maintainerswanted.com)* :nerd_face:
*Support us by leaving a star on [Github](https://github.com/flxwu/maintainerswanted.com)!* :star2:
## Algorithms is searching for new Maintainers! :man_technologist: :mailbox_with_mail:
Do you use Algorithms personally or at work and would like this project to be further developed and improved?
Or are you already a contributor and ready to take the next step to becoming a maintainer?
If you are interested, comment here below on this issue! :point_down::raised_hands:
|
True
|
Algorithms is searching for Maintainers! - *This issue was created by [Maintainers Wanted](https://maintainerswanted.com)* :nerd_face:
*Support us by leaving a star on [Github](https://github.com/flxwu/maintainerswanted.com)!* :star2:
## Algorithms is searching for new Maintainers! :man_technologist: :mailbox_with_mail:
Do you use Algorithms personally or at work and would like this project to be further developed and improved?
Or are you already a contributor and ready to take the next step to becoming a maintainer?
If you are interested, comment here below on this issue! :point_down::raised_hands:
|
main
|
algorithms is searching for maintainers this issue was created by nerd face support us by leaving a star on algorithms is searching for new maintainers man technologist mailbox with mail do you use algorithms personally or at work and would like this project to be further developed and improved or are you already a contributor and ready to take the next step to becoming a maintainer if you are interested comment here below on this issue point down raised hands
| 1
|
5,128
| 26,142,624,704
|
IssuesEvent
|
2022-12-29 21:05:03
|
backdrop-ops/contrib
|
https://api.github.com/repos/backdrop-ops/contrib
|
closed
|
Field Permissions: Become a maintainer for this project
|
Maintainer change request
|
I would like to become the maintainer for the Field Permissions project (which currently is "looking for maintainers").
Issue in that project: https://github.com/backdrop-contrib/field_permissions/issues/7
PR to update the README file: https://github.com/backdrop-contrib/field_permissions/pull/8
|
True
|
Field Permissions: Become a maintainer for this project - I would like to become the maintainer for the Field Permissions project (which currently is "looking for maintainers").
Issue in that project: https://github.com/backdrop-contrib/field_permissions/issues/7
PR to update the README file: https://github.com/backdrop-contrib/field_permissions/pull/8
|
main
|
field permissions become a maintainer for this project i would like to become the maintainer for the field permissions project which currently is looking for maintainers issue in that project pr to update the readme file
| 1
|
2,257
| 7,933,144,877
|
IssuesEvent
|
2018-07-08 00:42:12
|
Subsurface-divelog/subsurface
|
https://api.github.com/repos/Subsurface-divelog/subsurface
|
closed
|
Syncing with cloud/ Creating commit fails
|
bug cloud-storage needs-maintainer-feedback ui-desktop
|
<!-- Lines like this one are comments and will not be shown in the final output. -->
<!-- If you are a collaborator, please add labels and assign other collaborators for a review. -->
### Describe the issue:
- [x] Bug
### Issue long description:
Subsurface fails to sync with cloud storage. A red banner shows "creating commit failed".
### Operating system:
archlinux
### Subsurface version:
Self compiled binary:
Subsurface v4.7.8-289-g9611e92cf0d9,
built with libdivecomputer v0.7.0-devel-Subsurface-NG (e97a47cca55973199715df0f818b4955e60d3a31)
built with Qt Version 5.11.0, runtime from Qt Version 5.11.0
built with libgit2 0.27.0
Also happens with Appimage:
Subsurface v4.7.8,
built with libdivecomputer v0.7.0-devel-Subsurface-branch (d02f1c3cdcc6a04d085538578d872ec6e3282382)
built with Qt Version 5.9.3, runtime from Qt Version 5.9.3
built with libgit2 0.26.0
Also happens with Archlinux package:
Subsurface v,
built with libdivecomputer v0.7.0-devel-Subsurface-branch ()
built with Qt Version 5.11.0, runtime from Qt Version 5.11.0
built with libgit2 0.27.0
### Steps to reproduce:
1) open local logbook *.ssrf
2) select file-> save to cloud storage
3) red banner appears, "creating commit failed", new dives are not saved to cloud
### Current behavior:
No new commit is created in ./subsurface/cloudstorage/hash/
### Expected behavior:
Create a new commit with the changes from the local file and push it to the cloud server.
### Additional information:
Command line log:
...
cloud URL set as "https://cloud.subsurface-divelog.org//git/___@gmail.com[___@gmail.com]"
Saving cloud storage to: "https://cloud.subsurface-divelog.org//git/___@gmail.com[__@gmail.com]"
git storage: Synchronising data file
git storage: Sync with cloud storage
git storage: Successful cloud connection, fetch remote
git storage: Done syncing with cloud storage
git storage: Preparing to save data
git storage: Start saving data
git storage: Start saving dives
git storage: Done creating local cache

seems to be related to
https://github.com/Subsurface-divelog/subsurface/blob/2209054226f463a3fd0b88a34e9dcd18e15b363b/core/save-git.c#L1271
|
True
|
Syncing with cloud/ Creating commit fails - <!-- Lines like this one are comments and will not be shown in the final output. -->
<!-- If you are a collaborator, please add labels and assign other collaborators for a review. -->
### Describe the issue:
- [x] Bug
### Issue long description:
Subsurface fails to sync with cloud storage. A red banner shows "creating commit failed".
### Operating system:
archlinux
### Subsurface version:
Self compiled binary:
Subsurface v4.7.8-289-g9611e92cf0d9,
built with libdivecomputer v0.7.0-devel-Subsurface-NG (e97a47cca55973199715df0f818b4955e60d3a31)
built with Qt Version 5.11.0, runtime from Qt Version 5.11.0
built with libgit2 0.27.0
Also happens with Appimage:
Subsurface v4.7.8,
built with libdivecomputer v0.7.0-devel-Subsurface-branch (d02f1c3cdcc6a04d085538578d872ec6e3282382)
built with Qt Version 5.9.3, runtime from Qt Version 5.9.3
built with libgit2 0.26.0
Also happens with Archlinux package:
Subsurface v,
built with libdivecomputer v0.7.0-devel-Subsurface-branch ()
built with Qt Version 5.11.0, runtime from Qt Version 5.11.0
built with libgit2 0.27.0
### Steps to reproduce:
1) open local logbook *.ssrf
2) select file-> save to cloud storage
3) red banner appears, "creating commit failed", new dives are not saved to cloud
### Current behavior:
No new commit is created in ./subsurface/cloudstorage/hash/
### Expected behavior:
Create a new commit with the changes from the local file and push it to the cloud server.
### Additional information:
Command line log:
...
cloud URL set as "https://cloud.subsurface-divelog.org//git/___@gmail.com[___@gmail.com]"
Saving cloud storage to: "https://cloud.subsurface-divelog.org//git/___@gmail.com[__@gmail.com]"
git storage: Synchronising data file
git storage: Sync with cloud storage
git storage: Successful cloud connection, fetch remote
git storage: Done syncing with cloud storage
git storage: Preparing to save data
git storage: Start saving data
git storage: Start saving dives
git storage: Done creating local cache

seems to be related to
https://github.com/Subsurface-divelog/subsurface/blob/2209054226f463a3fd0b88a34e9dcd18e15b363b/core/save-git.c#L1271
|
main
|
syncing with cloud creating commit fails describe the issue bug issue long description subsurface fails to sync with cloud storage a red banner shows creating commit failed operating system archlinux subsurface version self compiled binary subsurface built with libdivecomputer devel subsurface ng built with qt version runtime from qt version built with also happens with appimage subsurface built with libdivecomputer devel subsurface branch built with qt version runtime from qt version built with also happens with archlinux package subsurface v built with libdivecomputer devel subsurface branch built with qt version runtime from qt version built with steps to reproduce open local logbook ssrf select file save to cloud storage red banner appears creating commit failed new dives are not saved to cloud current behavior no new commit is created in subsurface cloudstorage hash expected behavior create a new commit with the changes from the local file and push it to the cloud server additional information command line log cloud url set as saving cloud storage to git storage synchronising data file git storage sync with cloud storage git storage successful cloud connection fetch remote git storage done syncing with cloud storage git storage preparing to save data git storage start saving data git storage start saving dives git storage done creating local cache seems to be related to
| 1
|
3,212
| 12,288,607,697
|
IssuesEvent
|
2020-05-09 17:30:01
|
short-d/short
|
https://api.github.com/repos/short-d/short
|
opened
|
[Refactor] Add foreign keys for `user` referenced tables.
|
maintainability
|
There are a lot of DB tables that hold the `user_id` reference which leads to data integrity violation, as anytime a `user` record can be deleted or modified.
**Your solution**
After the migration process (primary key user_email -> user_id) for `user` table to be completed, it requires adding foreign key constraints in referenced tables.
|
True
|
[Refactor] Add foreign keys for `user` referenced tables. - There are a lot of DB tables that hold the `user_id` reference which leads to data integrity violation, as anytime a `user` record can be deleted or modified.
**Your solution**
After the migration process (primary key user_email -> user_id) for `user` table to be completed, it requires adding foreign key constraints in referenced tables.
|
main
|
add foreign keys for user referenced tables there are a lot of db tables that hold the user id reference which leads to data integrity violation as anytime a user record can be deleted or modified your solution after the migration process primary key user email user id for user table to be completed it requires adding foreign key constraints in referenced tables
| 1
|
4,463
| 23,253,711,311
|
IssuesEvent
|
2022-08-04 07:22:10
|
JENOT-ANT/ENIGMA
|
https://api.github.com/repos/JENOT-ANT/ENIGMA
|
opened
|
Konsolowy interfejs użytkownika
|
dev task(s) waiting-for-maintainers
|
Trzebaby dodać jakieś menu użytkownika z opcjami typu:
- szyfruj z pliku
- szyfruj w czasie rzeczywistym
- konfiguracja maszyny
- itp.
|
True
|
Konsolowy interfejs użytkownika - Trzebaby dodać jakieś menu użytkownika z opcjami typu:
- szyfruj z pliku
- szyfruj w czasie rzeczywistym
- konfiguracja maszyny
- itp.
|
main
|
konsolowy interfejs użytkownika trzebaby dodać jakieś menu użytkownika z opcjami typu szyfruj z pliku szyfruj w czasie rzeczywistym konfiguracja maszyny itp
| 1
|
896
| 4,554,432,795
|
IssuesEvent
|
2016-09-13 09:25:25
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
eos_template: TypeError: load_config() got an unexpected keyword argument 'session'
|
affects_2.2 bug_report networking waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
eos_template
##### ANSIBLE VERSION
```
ansible --version
ansible 2.2.0 (devel 9868117d1f) last updated 2016/09/12 20:18:28 (GMT +100)
lib/ansible/modules/core: (devel 432ee70da1) last updated 2016/09/12 20:19:14 (GMT +100)
lib/ansible/modules/extras: (devel 67a1bebbd3) last updated 2016/09/12 12:05:15 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
https://github.com/ansible/ansible-modules-core/commit/e464599632340f247c593f2db770be9782f7dec5
changed `module.config.load_config` to pass in `session` however that doesn't seem to be in https://github.com/ansible/ansible/blame/devel/lib/ansible/module_utils/netcfg.py#L56
@privateip When you ran
`ansible-playbook -vvv eos.yaml `
Did you have other local changes
##### STEPS TO REPRODUCE
```
ansible-playbook -vvv eos.yaml -i ../inventory-testnetwork -e "limit_to=eos_template"
```
##### EXPECTED RESULTS
Tests to pass
##### ACTUAL RESULTS
```
TASK [test_eos_template : configure device with config] ************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_eos_template/tests/cli/force.yaml:14
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/eos/eos_template.py
<veos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<veos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" && echo ansible-tmp-1473707985.23-162291198567673="` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" ) && sleep 0'
<veos01> PUT /tmp/tmpsPqOzL TO /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py
<veos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/ /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py && sleep 0'
<veos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 215, in <module>
main()
File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 205, in main
commit=True)
File "/tmp/ansible_aJfsia/ansible_modlib.zip/ansible/module_utils/netcfg.py", line 58, in load_config
TypeError: load_config() got an unexpected keyword argument 'session'
```
|
True
|
eos_template: TypeError: load_config() got an unexpected keyword argument 'session' - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
eos_template
##### ANSIBLE VERSION
```
ansible --version
ansible 2.2.0 (devel 9868117d1f) last updated 2016/09/12 20:18:28 (GMT +100)
lib/ansible/modules/core: (devel 432ee70da1) last updated 2016/09/12 20:19:14 (GMT +100)
lib/ansible/modules/extras: (devel 67a1bebbd3) last updated 2016/09/12 12:05:15 (GMT +100)
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
https://github.com/ansible/ansible-modules-core/commit/e464599632340f247c593f2db770be9782f7dec5
changed `module.config.load_config` to pass in `session` however that doesn't seem to be in https://github.com/ansible/ansible/blame/devel/lib/ansible/module_utils/netcfg.py#L56
@privateip When you ran
`ansible-playbook -vvv eos.yaml `
Did you have other local changes
##### STEPS TO REPRODUCE
```
ansible-playbook -vvv eos.yaml -i ../inventory-testnetwork -e "limit_to=eos_template"
```
##### EXPECTED RESULTS
Tests to pass
##### ACTUAL RESULTS
```
TASK [test_eos_template : configure device with config] ************************
task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_eos_template/tests/cli/force.yaml:14
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/eos/eos_template.py
<veos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<veos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" && echo ansible-tmp-1473707985.23-162291198567673="` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" ) && sleep 0'
<veos01> PUT /tmp/tmpsPqOzL TO /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py
<veos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/ /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py && sleep 0'
<veos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 215, in <module>
main()
File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 205, in main
commit=True)
File "/tmp/ansible_aJfsia/ansible_modlib.zip/ansible/module_utils/netcfg.py", line 58, in load_config
TypeError: load_config() got an unexpected keyword argument 'session'
```
|
main
|
eos template typeerror load config got an unexpected keyword argument session issue type bug report component name eos template ansible version ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary changed module config load config to pass in session however that doesn t seem to be in privateip when you ran ansible playbook vvv eos yaml did you have other local changes steps to reproduce ansible playbook vvv eos yaml i inventory testnetwork e limit to eos template expected results tests to pass actual results task task path home johnb git ansible inc test network modules roles test eos template tests cli force yaml using module file home johnb git ansible inc ansible lib ansible modules core network eos eos template py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpspqozl to home johnb ansible tmp ansible tmp eos template py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp eos template py sleep exec bin sh c python home johnb ansible tmp ansible tmp eos template py rm rf home johnb ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible ajfsia ansible module eos template py line in main file tmp ansible ajfsia ansible module eos template py line in main commit true file tmp ansible ajfsia ansible modlib zip ansible module utils netcfg py line in load config typeerror load config got an unexpected keyword argument session
| 1
|
365,823
| 25,554,076,383
|
IssuesEvent
|
2022-11-30 04:05:46
|
simonw/datasette
|
https://api.github.com/repos/simonw/datasette
|
closed
|
Interactive demo of Datasette 1.0 write APIs
|
enhancement documentation ops
|
I'm going to try to get this working on https://latest.datasette.io/ - it already has a way for people to sign in as root, but none of the databases there are writable.
So I'm going to build a plugin which adds a writable named in-memory database.
And some kind of mechanism for clearing out that database on a regular basis - maybe tables in that database get deleted automatically an hour after they are created?
(Would be neat to display their time-left-until-deleted too)
|
1.0
|
Interactive demo of Datasette 1.0 write APIs - I'm going to try to get this working on https://latest.datasette.io/ - it already has a way for people to sign in as root, but none of the databases there are writable.
So I'm going to build a plugin which adds a writable named in-memory database.
And some kind of mechanism for clearing out that database on a regular basis - maybe tables in that database get deleted automatically an hour after they are created?
(Would be neat to display their time-left-until-deleted too)
|
non_main
|
interactive demo of datasette write apis i m going to try to get this working on it already has a way for people to sign in as root but none of the databases there are writable so i m going to build a plugin which adds a writable named in memory database and some kind of mechanism for clearing out that database on a regular basis maybe tables in that database get deleted automatically an hour after they are created would be neat to display their time left until deleted too
| 0
|
1,138
| 4,998,877,174
|
IssuesEvent
|
2016-12-09 21:19:46
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
ec2_vpc: Must specify resource_tags when deleting(!) VPC
|
affects_2.1 aws bug_report cloud waiting_on_maintainer
|
<!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_vpc
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
##### SUMMARY
According to [the doc](https://docs.ansible.com/ansible/ec2_vpc_module.html) you don't have to specify the `resource_tags` parameter when deleting a VPC.
However, this module barfs when this parameter is missing.
##### STEPS TO REPRODUCE
This is my task:
<!--- Paste example playbooks or commands between quotes below -->
```
- name: "Destroy PREPROD VPC"
ec2_vpc:
region: "{{ region }}"
state: absent
vpc_id: "{{ preprod_vpc.vpcs[0].id }}"
Name: "{{ owner }}_vpc_{{ env }}"
Env: "{{ owner }}_{{ env }}"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
I expected that the task would be executed successfully.
##### ACTUAL RESULTS
`ec2_vpc` complained as follows:
<!--- Paste verbatim command output between quotes below -->
```
missing required arguments: resource_tags
```
After I added the parameter the call was successful.
|
True
|
ec2_vpc: Must specify resource_tags when deleting(!) VPC - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_vpc
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
##### SUMMARY
According to [the doc](https://docs.ansible.com/ansible/ec2_vpc_module.html) you don't have to specify the `resource_tags` parameter when deleting a VPC.
However, this module barfs when this parameter is missing.
##### STEPS TO REPRODUCE
This is my task:
<!--- Paste example playbooks or commands between quotes below -->
```
- name: "Destroy PREPROD VPC"
ec2_vpc:
region: "{{ region }}"
state: absent
vpc_id: "{{ preprod_vpc.vpcs[0].id }}"
Name: "{{ owner }}_vpc_{{ env }}"
Env: "{{ owner }}_{{ env }}"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
I expected that the task would be executed successfully.
##### ACTUAL RESULTS
`ec2_vpc` complained as follows:
<!--- Paste verbatim command output between quotes below -->
```
missing required arguments: resource_tags
```
After I added the parameter the call was successful.
|
main
|
vpc must specify resource tags when deleting vpc issue type bug report component name vpc ansible version ansible config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific summary according to you don t have to specify the resource tags parameter when deleting a vpc however this module barfs when this parameter is missing steps to reproduce this is my task name destroy preprod vpc vpc region region state absent vpc id preprod vpc vpcs id name owner vpc env env owner env expected results i expected that the task would be executed successfully actual results vpc complained as follows missing required arguments resource tags after i added the parameter the call was successful
| 1
|
35,133
| 12,311,802,245
|
IssuesEvent
|
2020-05-12 13:01:10
|
vondenstein/gitblast.com
|
https://api.github.com/repos/vondenstein/gitblast.com
|
opened
|
CVE-2019-6286 (Medium) detected in opennms-opennms-source-22.0.1-1, node-sass-4.13.1.tgz
|
security vulnerability
|
## CVE-2019-6286 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.13.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/gitblast.com/package.json</p>
<p>Path to vulnerable library: /gitblast.com/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.13.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/vondenstein/gitblast.com/commit/9ba0bd32c0c48dfe3dc35a01cf167196a1e46913">9ba0bd32c0c48dfe3dc35a01cf167196a1e46913</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::skip_over_scopes in prelexer.hpp when called from Sass::Parser::parse_import(), a similar issue to CVE-2018-11693.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6286>CVE-2019-6286</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-6286 (Medium) detected in opennms-opennms-source-22.0.1-1, node-sass-4.13.1.tgz - ## CVE-2019-6286 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.1.tgz</b></p></summary>
<p>
<details><summary><b>node-sass-4.13.1.tgz</b></p></summary>
<p>Wrapper around libsass</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/gitblast.com/package.json</p>
<p>Path to vulnerable library: /gitblast.com/node_modules/node-sass/package.json</p>
<p>
Dependency Hierarchy:
- :x: **node-sass-4.13.1.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/vondenstein/gitblast.com/commit/9ba0bd32c0c48dfe3dc35a01cf167196a1e46913">9ba0bd32c0c48dfe3dc35a01cf167196a1e46913</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::skip_over_scopes in prelexer.hpp when called from Sass::Parser::parse_import(), a similar issue to CVE-2018-11693.
<p>Publish Date: 2019-01-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6286>CVE-2019-6286</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286</a></p>
<p>Release Date: 2019-08-06</p>
<p>Fix Resolution: LibSass - 3.6.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve medium detected in opennms opennms source node sass tgz cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm gitblast com package json path to vulnerable library gitblast com node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details in libsass a heap based buffer over read exists in sass prelexer skip over scopes in prelexer hpp when called from sass parser parse import a similar issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
| 0
|
3,768
| 15,833,528,410
|
IssuesEvent
|
2021-04-06 15:44:22
|
ethersphere/bee-status
|
https://api.github.com/repos/ethersphere/bee-status
|
closed
|
Add Repository Description
|
need:maintainers-input type:issue
|
Currently the repo doesn't contain a description which may make it harder for people to discern what the project is.
|
True
|
Add Repository Description - Currently the repo doesn't contain a description which may make it harder for people to discern what the project is.
|
main
|
add repository description currently the repo doesn t contain a description which may make it harder for people to discern what the project is
| 1
|
1,557
| 6,572,253,565
|
IssuesEvent
|
2017-09-11 00:39:03
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
Jira module is not sending valid json when operation = edit
|
affects_2.1 bug_report waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
jira (module)
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
Default configuration - no changes
##### OS / ENVIRONMENT
Tested on Mac OS - however seems be global (not related with local platform)
##### SUMMARY
According documentation jira module should allow create / modify items in Jira. Creating tickets is working fine, adding comments are working fine - however when operation == 'edit', module complain with error.
##### STEPS TO REPRODUCE
This can be reproduce with `netcat` command and `jira` task. Running `ansible -vvvv` not helps in this case - not displaying outgoing json string which is sending to Jira URL.
Repeat steps:
On console 1 (this allows display json which is sending without sending this to real JIRA url)
```
# Run netcat and listen for connections
nc -l 8080
```
On another console run following playbook:
```
# Playbook:
- name: Updating ticket on Jira (testing)
hosts: localhost
gather_facts: no
vars:
- soft : 'Ansible'
- url : 'http://127.0.0.1:8080' # Server: nc -l 8080
- issue_test : 'Test'
vars_files:
# Copy 'common_jira' to 'custom_jira' and adjust with valid credentials
- [ "custom_jira.yml", "common_jira.yml" ]
tasks:
- name: Edit issue - use arguments with fields
jira:
username: "{{ jira_user.name }}"
password: "{{ jira_user.pass }}"
uri: "{{ url }}"
operation: "edit"
issue: "200"
register: issue_edit
ignore_errors: yes
args:
fields:
- labels:
- autocreated
- ansible
- yaml
# Create file with variables: custom_jira.yml
---
jira_user:
name: 'put_your_login_here'
pass: 'put_your_pass_here'
# Create basic hosts.txt (Empty - local host will be added automatically)
# Run command
ansible-playbook -i hosts.txt playbook.yml
```
##### EXPECTED RESULTS
After debugging with Jira following Json should be created during sending
```
{"fields": {"labels": ["autocreated", "ansible"]}}
```
Module should connect with JIRA via REST API and edit issue based on above fields.
##### ACTUAL RESULTS
```
TASK [Edit issue - use arguments with fields] **********************************
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "HTTP Error 400: Bad Request"}
# JSON: Difference of inverted commas.
{"fields": "{'labels': ['autocreated', 'ansible']}"}
```
Runnning ansible -vvvv not shows outgoing json - this is why netcat was used to see what 'server' is receiving.
Last info:
Result from Jira based on wrong json:
```
HTTP/1.1 400 Bad Request
```
Result from Jira based on valid json:
```
HTTP/1.1 204 No Content
# And Jira issue were updated with valid fields
```
|
True
|
Jira module is not sending valid json when operation = edit - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
jira (module)
##### ANSIBLE VERSION
```
ansible 2.1.0.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
Default configuration - no changes
##### OS / ENVIRONMENT
Tested on Mac OS - however seems be global (not related with local platform)
##### SUMMARY
According documentation jira module should allow create / modify items in Jira. Creating tickets is working fine, adding comments are working fine - however when operation == 'edit', module complain with error.
##### STEPS TO REPRODUCE
This can be reproduce with `netcat` command and `jira` task. Running `ansible -vvvv` not helps in this case - not displaying outgoing json string which is sending to Jira URL.
Repeat steps:
On console 1 (this allows display json which is sending without sending this to real JIRA url)
```
# Run netcat and listen for connections
nc -l 8080
```
On another console run following playbook:
```
# Playbook:
- name: Updating ticket on Jira (testing)
hosts: localhost
gather_facts: no
vars:
- soft : 'Ansible'
- url : 'http://127.0.0.1:8080' # Server: nc -l 8080
- issue_test : 'Test'
vars_files:
# Copy 'common_jira' to 'custom_jira' and adjust with valid credentials
- [ "custom_jira.yml", "common_jira.yml" ]
tasks:
- name: Edit issue - use arguments with fields
jira:
username: "{{ jira_user.name }}"
password: "{{ jira_user.pass }}"
uri: "{{ url }}"
operation: "edit"
issue: "200"
register: issue_edit
ignore_errors: yes
args:
fields:
- labels:
- autocreated
- ansible
- yaml
# Create file with variables: custom_jira.yml
---
jira_user:
name: 'put_your_login_here'
pass: 'put_your_pass_here'
# Create basic hosts.txt (Empty - local host will be added automatically)
# Run command
ansible-playbook -i hosts.txt playbook.yml
```
##### EXPECTED RESULTS
After debugging with Jira following Json should be created during sending
```
{"fields": {"labels": ["autocreated", "ansible"]}}
```
Module should connect with JIRA via REST API and edit issue based on above fields.
##### ACTUAL RESULTS
```
TASK [Edit issue - use arguments with fields] **********************************
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "HTTP Error 400: Bad Request"}
# JSON: Difference of inverted commas.
{"fields": "{'labels': ['autocreated', 'ansible']}"}
```
Runnning ansible -vvvv not shows outgoing json - this is why netcat was used to see what 'server' is receiving.
Last info:
Result from Jira based on wrong json:
```
HTTP/1.1 400 Bad Request
```
Result from Jira based on valid json:
```
HTTP/1.1 204 No Content
# And Jira issue were updated with valid fields
```
|
main
|
jira module is not sending valid json when operation edit issue type bug report component name jira module ansible version ansible config file configured module search path default w o overrides configuration default configuration no changes os environment tested on mac os however seems be global not related with local platform summary according documentation jira module should allow create modify items in jira creating tickets is working fine adding comments are working fine however when operation edit module complain with error steps to reproduce this can be reproduce with netcat command and jira task running ansible vvvv not helps in this case not displaying outgoing json string which is sending to jira url repeat steps on console this allows display json which is sending without sending this to real jira url run netcat and listen for connections nc l on another console run following playbook playbook name updating ticket on jira testing hosts localhost gather facts no vars soft ansible url server nc l issue test test vars files copy common jira to custom jira and adjust with valid credentials tasks name edit issue use arguments with fields jira username jira user name password jira user pass uri url operation edit issue register issue edit ignore errors yes args fields labels autocreated ansible yaml create file with variables custom jira yml jira user name put your login here pass put your pass here create basic hosts txt empty local host will be added automatically run command ansible playbook i hosts txt playbook yml expected results after debugging with jira following json should be created during sending fields labels module should connect with jira via rest api and edit issue based on above fields actual results task fatal failed changed false failed true msg http error bad request json difference of inverted commas fields labels runnning ansible vvvv not shows outgoing json this is why netcat was used to see what server is receiving last info result from jira based on wrong json http bad request result from jira based on valid json http no content and jira issue were updated with valid fields
| 1
|
815,328
| 30,546,736,723
|
IssuesEvent
|
2023-07-20 04:59:54
|
wso2/product-is
|
https://api.github.com/repos/wso2/product-is
|
closed
|
Sample web app breaks with below logs when the tomcat directory exists in a location where there is any space
|
Priority/High Severity/Major bug Affected/5.11.0 samples
|
**Describe the issue:**
Sample web app breaks with below logs when **tomcat** (parent) directory exists in a location where there is any space.
```
18-Jul-2023 19:09:05.981 INFO [main] org.wso2.sample.identity.jks.JKSLoader.contextInitialized Setting trust store path to : /Users/ratnajothy/Desktop/apache-tomcat%20%20%20%20%209.0.41/webapps/pickup-dispatch/WEB-INF/classes/wso2carbon.jks
18-Jul-2023 19:09:06.104 SEVERE [main] org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse Error while retrieving response.
java.net.SocketException: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext)
at javax.net.ssl.DefaultSSLSocketFactory.throwException(SSLSocketFactory.java:248)
at javax.net.ssl.DefaultSSLSocketFactory.createSocket(SSLSocketFactory.java:255)
at sun.net.www.protocol.https.HttpsClient.createSocket(HttpsClient.java:405)
at sun.net.NetworkClient.doConnect(NetworkClient.java:162)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:463)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:558)
at sun.net.www.protocol.https.HttpsClient.<init>(HttpsClient.java:264)
at sun.net.www.protocol.https.HttpsClient.New(HttpsClient.java:367)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.getNewHttpClient(AbstractDelegateHttpsURLConnection.java:203)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1162)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1056)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:189)
at sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1340)
at sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1315)
at sun.net.www.protocol.https.HttpsURLConnectionImpl.getOutputStream(HttpsURLConnectionImpl.java:264)
at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:261)
at org.wso2.samples.claims.manager.ClaimManagerProxy.getOIDCDialectNodes(ClaimManagerProxy.java:192)
at org.wso2.samples.claims.manager.ClaimManagerProxy.<init>(ClaimManagerProxy.java:74)
at org.wso2.sample.identity.oauth2.SampleContextEventListener.contextInitialized(SampleContextEventListener.java:50)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4768)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5230)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:728)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:700)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:696)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1024)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1911)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75)
at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:112)
at org.apache.catalina.startup.HostConfig.deployWARs(HostConfig.java:825)
at org.apache.catalina.startup.HostConfig.deployApps(HostConfig.java:475)
at org.apache.catalina.startup.HostConfig.start(HostConfig.java:1618)
at org.apache.catalina.startup.HostConfig.lifecycleEvent(HostConfig.java:319)
at org.apache.catalina.util.LifecycleBase.fireLifecycleEvent(LifecycleBase.java:123)
at org.apache.catalina.util.LifecycleBase.setStateInternal(LifecycleBase.java:423)
at org.apache.catalina.util.LifecycleBase.setState(LifecycleBase.java:366)
at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:948)
at org.apache.catalina.core.StandardHost.startInternal(StandardHost.java:835)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1398)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1388)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75)
at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:134)
at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:921)
at org.apache.catalina.core.StandardEngine.startInternal(StandardEngine.java:263)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.StandardService.startInternal(StandardService.java:437)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.StandardServer.startInternal(StandardServer.java:934)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.startup.Catalina.start(Catalina.java:772)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:345)
at org.apache.catalina.startup.Bootstrap.main(Bootstrap.java:476)
Caused by: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext)
at java.security.Provider$Service.newInstance(Provider.java:1617)
at sun.security.jca.GetInstance.getInstance(GetInstance.java:236)
at sun.security.jca.GetInstance.getInstance(GetInstance.java:164)
at javax.net.ssl.SSLContext.getInstance(SSLContext.java:156)
at javax.net.ssl.SSLContext.getDefault(SSLContext.java:96)
at javax.net.ssl.SSLSocketFactory.getDefault(SSLSocketFactory.java:122)
at javax.net.ssl.HttpsURLConnection.getDefaultSSLSocketFactory(HttpsURLConnection.java:332)
at javax.net.ssl.HttpsURLConnection.<init>(HttpsURLConnection.java:289)
at sun.net.www.protocol.https.HttpsURLConnectionImpl.<init>(HttpsURLConnectionImpl.java:99)
at sun.net.www.protocol.https.Handler.openConnection(Handler.java:62)
at sun.net.www.protocol.https.Handler.openConnection(Handler.java:57)
at java.net.URL.openConnection(URL.java:1002)
at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:251)
... 44 more
Caused by: java.security.KeyStoreException: problem accessing trust store
at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:73)
at javax.net.ssl.TrustManagerFactory.init(TrustManagerFactory.java:250)
at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.getTrustManagers(SSLContextImpl.java:1041)
at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.<clinit>(SSLContextImpl.java:1011)
at sun.security.ssl.SSLContextImpl$DefaultSSLContext.<init>(SSLContextImpl.java:1186)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at java.security.Provider$Service.newInstance(Provider.java:1595)
... 56 more
Caused by: java.io.IOException: Keystore was tampered with, or password was incorrect
at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:792)
at sun.security.provider.JavaKeyStore$JKS.engineLoad(JavaKeyStore.java:57)
at sun.security.provider.KeyStoreDelegator.engineLoad(KeyStoreDelegator.java:224)
at sun.security.provider.JavaKeyStore$DualFormatJKS.engineLoad(JavaKeyStore.java:71)
at java.security.KeyStore.load(KeyStore.java:1445)
at sun.security.ssl.TrustStoreManager$TrustAnchorManager.loadKeyStore(TrustStoreManager.java:365)
at sun.security.ssl.TrustStoreManager$TrustAnchorManager.getTrustedCerts(TrustStoreManager.java:313)
at sun.security.ssl.TrustStoreManager.getTrustedCerts(TrustStoreManager.java:55)
at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:49)
... 65 more
Caused by: java.security.UnrecoverableKeyException: Password verification failed
at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:790)
... 73 more
[Fatal Error] :-1:-1: Premature end of file.
```
**How to reproduce:**
1. Rename the tomcat directory by giving a space in its name
2. Make sure there are no other webapps except for the Pickup-manager or Pickup-dispatch
3. Pickup-manager and Pickup-dispatch fails to deploy successfully
**Expected behaviour:**
<!-- A clear and concise description of what you expected to happen. -->
Should work despite having spaces in the located path of the tomcat directory.
**Suggestion**
Fix the logic in obtaining resource path as URL when setting Trust Store Path : https://github.com/wso2/samples-is/blob/345db6b9228396d6ac74a8eb7dcb441dde7574cf/sso-samples/oidc-sso-sample/oidc-jks-loader/src/main/org/wso2/sample/identity/jks/JKSLoader.java#L58
_Fix should be for Pickup-manager and dispatch (both)_
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: ....., 5.10.0, 5.11.0
- OS: Windows, Linux, Mac
(Verified this issue with ORA JDK 8 and Tomcat 9.0.41)
|
1.0
|
Sample web app breaks with below logs when the tomcat directory exists in a location where there is any space - **Describe the issue:**
Sample web app breaks with below logs when **tomcat** (parent) directory exists in a location where there is any space.
```
18-Jul-2023 19:09:05.981 INFO [main] org.wso2.sample.identity.jks.JKSLoader.contextInitialized Setting trust store path to : /Users/ratnajothy/Desktop/apache-tomcat%20%20%20%20%209.0.41/webapps/pickup-dispatch/WEB-INF/classes/wso2carbon.jks
18-Jul-2023 19:09:06.104 SEVERE [main] org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse Error while retrieving response.
java.net.SocketException: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext)
at javax.net.ssl.DefaultSSLSocketFactory.throwException(SSLSocketFactory.java:248)
at javax.net.ssl.DefaultSSLSocketFactory.createSocket(SSLSocketFactory.java:255)
at sun.net.www.protocol.https.HttpsClient.createSocket(HttpsClient.java:405)
at sun.net.NetworkClient.doConnect(NetworkClient.java:162)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:463)
at sun.net.www.http.HttpClient.openServer(HttpClient.java:558)
at sun.net.www.protocol.https.HttpsClient.<init>(HttpsClient.java:264)
at sun.net.www.protocol.https.HttpsClient.New(HttpsClient.java:367)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.getNewHttpClient(AbstractDelegateHttpsURLConnection.java:203)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1162)
at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1056)
at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:189)
at sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1340)
at sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1315)
at sun.net.www.protocol.https.HttpsURLConnectionImpl.getOutputStream(HttpsURLConnectionImpl.java:264)
at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:261)
at org.wso2.samples.claims.manager.ClaimManagerProxy.getOIDCDialectNodes(ClaimManagerProxy.java:192)
at org.wso2.samples.claims.manager.ClaimManagerProxy.<init>(ClaimManagerProxy.java:74)
at org.wso2.sample.identity.oauth2.SampleContextEventListener.contextInitialized(SampleContextEventListener.java:50)
at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4768)
at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5230)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:728)
at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:700)
at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:696)
at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1024)
at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1911)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75)
at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:112)
at org.apache.catalina.startup.HostConfig.deployWARs(HostConfig.java:825)
at org.apache.catalina.startup.HostConfig.deployApps(HostConfig.java:475)
at org.apache.catalina.startup.HostConfig.start(HostConfig.java:1618)
at org.apache.catalina.startup.HostConfig.lifecycleEvent(HostConfig.java:319)
at org.apache.catalina.util.LifecycleBase.fireLifecycleEvent(LifecycleBase.java:123)
at org.apache.catalina.util.LifecycleBase.setStateInternal(LifecycleBase.java:423)
at org.apache.catalina.util.LifecycleBase.setState(LifecycleBase.java:366)
at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:948)
at org.apache.catalina.core.StandardHost.startInternal(StandardHost.java:835)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1398)
at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1388)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75)
at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:134)
at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:921)
at org.apache.catalina.core.StandardEngine.startInternal(StandardEngine.java:263)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.StandardService.startInternal(StandardService.java:437)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.core.StandardServer.startInternal(StandardServer.java:934)
at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183)
at org.apache.catalina.startup.Catalina.start(Catalina.java:772)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:345)
at org.apache.catalina.startup.Bootstrap.main(Bootstrap.java:476)
Caused by: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext)
at java.security.Provider$Service.newInstance(Provider.java:1617)
at sun.security.jca.GetInstance.getInstance(GetInstance.java:236)
at sun.security.jca.GetInstance.getInstance(GetInstance.java:164)
at javax.net.ssl.SSLContext.getInstance(SSLContext.java:156)
at javax.net.ssl.SSLContext.getDefault(SSLContext.java:96)
at javax.net.ssl.SSLSocketFactory.getDefault(SSLSocketFactory.java:122)
at javax.net.ssl.HttpsURLConnection.getDefaultSSLSocketFactory(HttpsURLConnection.java:332)
at javax.net.ssl.HttpsURLConnection.<init>(HttpsURLConnection.java:289)
at sun.net.www.protocol.https.HttpsURLConnectionImpl.<init>(HttpsURLConnectionImpl.java:99)
at sun.net.www.protocol.https.Handler.openConnection(Handler.java:62)
at sun.net.www.protocol.https.Handler.openConnection(Handler.java:57)
at java.net.URL.openConnection(URL.java:1002)
at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:251)
... 44 more
Caused by: java.security.KeyStoreException: problem accessing trust store
at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:73)
at javax.net.ssl.TrustManagerFactory.init(TrustManagerFactory.java:250)
at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.getTrustManagers(SSLContextImpl.java:1041)
at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.<clinit>(SSLContextImpl.java:1011)
at sun.security.ssl.SSLContextImpl$DefaultSSLContext.<init>(SSLContextImpl.java:1186)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at java.security.Provider$Service.newInstance(Provider.java:1595)
... 56 more
Caused by: java.io.IOException: Keystore was tampered with, or password was incorrect
at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:792)
at sun.security.provider.JavaKeyStore$JKS.engineLoad(JavaKeyStore.java:57)
at sun.security.provider.KeyStoreDelegator.engineLoad(KeyStoreDelegator.java:224)
at sun.security.provider.JavaKeyStore$DualFormatJKS.engineLoad(JavaKeyStore.java:71)
at java.security.KeyStore.load(KeyStore.java:1445)
at sun.security.ssl.TrustStoreManager$TrustAnchorManager.loadKeyStore(TrustStoreManager.java:365)
at sun.security.ssl.TrustStoreManager$TrustAnchorManager.getTrustedCerts(TrustStoreManager.java:313)
at sun.security.ssl.TrustStoreManager.getTrustedCerts(TrustStoreManager.java:55)
at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:49)
... 65 more
Caused by: java.security.UnrecoverableKeyException: Password verification failed
at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:790)
... 73 more
[Fatal Error] :-1:-1: Premature end of file.
```
**How to reproduce:**
1. Rename the tomcat directory by giving a space in its name
2. Make sure there are no other webapps except for the Pickup-manager or Pickup-dispatch
3. Pickup-manager and Pickup-dispatch fails to deploy successfully
**Expected behaviour:**
<!-- A clear and concise description of what you expected to happen. -->
Should work despite having spaces in the located path of the tomcat directory.
**Suggestion**
Fix the logic in obtaining resource path as URL when setting Trust Store Path : https://github.com/wso2/samples-is/blob/345db6b9228396d6ac74a8eb7dcb441dde7574cf/sso-samples/oidc-sso-sample/oidc-jks-loader/src/main/org/wso2/sample/identity/jks/JKSLoader.java#L58
_Fix should be for Pickup-manager and dispatch (both)_
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: ....., 5.10.0, 5.11.0
- OS: Windows, Linux, Mac
(Verified this issue with ORA JDK 8 and Tomcat 9.0.41)
|
non_main
|
sample web app breaks with below logs when the tomcat directory exists in a location where there is any space describe the issue sample web app breaks with below logs when tomcat parent directory exists in a location where there is any space jul info org sample identity jks jksloader contextinitialized setting trust store path to users ratnajothy desktop apache tomcat webapps pickup dispatch web inf classes jks jul severe org samples claims manager claimmanagerproxy getclaimmanagementresponse error while retrieving response java net socketexception java security nosuchalgorithmexception error constructing implementation algorithm default provider sunjsse class sun security ssl sslcontextimpl defaultsslcontext at javax net ssl defaultsslsocketfactory throwexception sslsocketfactory java at javax net ssl defaultsslsocketfactory createsocket sslsocketfactory java at sun net at sun net networkclient doconnect networkclient java at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at org samples claims manager claimmanagerproxy getclaimmanagementresponse claimmanagerproxy java at org samples claims manager claimmanagerproxy getoidcdialectnodes claimmanagerproxy java at org samples claims manager claimmanagerproxy claimmanagerproxy java at org sample identity samplecontexteventlistener contextinitialized samplecontexteventlistener java at org apache catalina core standardcontext listenerstart standardcontext java at org apache catalina core standardcontext startinternal standardcontext java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core containerbase addchildinternal containerbase java at org apache catalina core containerbase addchild containerbase java at org apache catalina core standardhost addchild standardhost java at org apache catalina startup hostconfig deploywar hostconfig java at org apache catalina startup hostconfig deploywar run hostconfig java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at org apache tomcat util threads inlineexecutorservice execute inlineexecutorservice java at java util concurrent abstractexecutorservice submit abstractexecutorservice java at org apache catalina startup hostconfig deploywars hostconfig java at org apache catalina startup hostconfig deployapps hostconfig java at org apache catalina startup hostconfig start hostconfig java at org apache catalina startup hostconfig lifecycleevent hostconfig java at org apache catalina util lifecyclebase firelifecycleevent lifecyclebase java at org apache catalina util lifecyclebase setstateinternal lifecyclebase java at org apache catalina util lifecyclebase setstate lifecyclebase java at org apache catalina core containerbase startinternal containerbase java at org apache catalina core standardhost startinternal standardhost java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core containerbase startchild call containerbase java at org apache catalina core containerbase startchild call containerbase java at java util concurrent futuretask run futuretask java at org apache tomcat util threads inlineexecutorservice execute inlineexecutorservice java at java util concurrent abstractexecutorservice submit abstractexecutorservice java at org apache catalina core containerbase startinternal containerbase java at org apache catalina core standardengine startinternal standardengine java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core standardservice startinternal standardservice java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core standardserver startinternal standardserver java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina startup catalina start catalina java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache catalina startup bootstrap start bootstrap java at org apache catalina startup bootstrap main bootstrap java caused by java security nosuchalgorithmexception error constructing implementation algorithm default provider sunjsse class sun security ssl sslcontextimpl defaultsslcontext at java security provider service newinstance provider java at sun security jca getinstance getinstance getinstance java at sun security jca getinstance getinstance getinstance java at javax net ssl sslcontext getinstance sslcontext java at javax net ssl sslcontext getdefault sslcontext java at javax net ssl sslsocketfactory getdefault sslsocketfactory java at javax net ssl httpsurlconnection getdefaultsslsocketfactory httpsurlconnection java at javax net ssl httpsurlconnection httpsurlconnection java at sun net at sun net at sun net at java net url openconnection url java at org samples claims manager claimmanagerproxy getclaimmanagementresponse claimmanagerproxy java more caused by java security keystoreexception problem accessing trust store at sun security ssl trustmanagerfactoryimpl engineinit trustmanagerfactoryimpl java at javax net ssl trustmanagerfactory init trustmanagerfactory java at sun security ssl sslcontextimpl defaultmanagersholder gettrustmanagers sslcontextimpl java at sun security ssl sslcontextimpl defaultmanagersholder sslcontextimpl java at sun security ssl sslcontextimpl defaultsslcontext sslcontextimpl java at sun reflect nativeconstructoraccessorimpl native method at sun reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at sun reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java lang reflect constructor newinstance constructor java at java security provider service newinstance provider java more caused by java io ioexception keystore was tampered with or password was incorrect at sun security provider javakeystore engineload javakeystore java at sun security provider javakeystore jks engineload javakeystore java at sun security provider keystoredelegator engineload keystoredelegator java at sun security provider javakeystore dualformatjks engineload javakeystore java at java security keystore load keystore java at sun security ssl truststoremanager trustanchormanager loadkeystore truststoremanager java at sun security ssl truststoremanager trustanchormanager gettrustedcerts truststoremanager java at sun security ssl truststoremanager gettrustedcerts truststoremanager java at sun security ssl trustmanagerfactoryimpl engineinit trustmanagerfactoryimpl java more caused by java security unrecoverablekeyexception password verification failed at sun security provider javakeystore engineload javakeystore java more premature end of file how to reproduce rename the tomcat directory by giving a space in its name make sure there are no other webapps except for the pickup manager or pickup dispatch pickup manager and pickup dispatch fails to deploy successfully expected behaviour should work despite having spaces in the located path of the tomcat directory suggestion fix the logic in obtaining resource path as url when setting trust store path fix should be for pickup manager and dispatch both environment information please complete the following information remove any unnecessary fields product version os windows linux mac verified this issue with ora jdk and tomcat
| 0
|
244,921
| 18,769,439,701
|
IssuesEvent
|
2021-11-06 15:06:07
|
AY2122S1-CS2103T-T15-2/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-T15-2/tp
|
closed
|
[PE-D] Visuals in UG
|
type.Documentation
|
Perhaps could have included more visuals for things like task and event commands that aren't covered in persons command.
<!--session: 1635494514989-f88cdb01-2d06-45de-8a15-2a40c14ffa30-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.DocumentationBug` `severity.VeryLow`
original: Shruthi0310/ped#14
|
1.0
|
[PE-D] Visuals in UG - Perhaps could have included more visuals for things like task and event commands that aren't covered in persons command.
<!--session: 1635494514989-f88cdb01-2d06-45de-8a15-2a40c14ffa30-->
<!--Version: Web v3.4.1-->
-------------
Labels: `type.DocumentationBug` `severity.VeryLow`
original: Shruthi0310/ped#14
|
non_main
|
visuals in ug perhaps could have included more visuals for things like task and event commands that aren t covered in persons command labels type documentationbug severity verylow original ped
| 0
|
1,169
| 5,087,807,764
|
IssuesEvent
|
2016-12-31 09:47:21
|
openwrt/packages
|
https://api.github.com/repos/openwrt/packages
|
closed
|
transmission has wrong maintainer
|
waiting for maintainer
|
net/transmission still has the maintainer from the old packages feed, which does not match the pull request initiator.
@mstorchak since you added the package, please change the maintainer to yourself.
|
True
|
transmission has wrong maintainer - net/transmission still has the maintainer from the old packages feed, which does not match the pull request initiator.
@mstorchak since you added the package, please change the maintainer to yourself.
|
main
|
transmission has wrong maintainer net transmission still has the maintainer from the old packages feed which does not match the pull request initiator mstorchak since you added the package please change the maintainer to yourself
| 1
|
142,567
| 19,099,352,334
|
IssuesEvent
|
2021-11-29 20:26:23
|
harrinry/stream-chat-react
|
https://api.github.com/repos/harrinry/stream-chat-react
|
opened
|
CVE-2021-32723 (Medium) detected in prismjs-1.22.0.tgz
|
security vulnerability
|
## CVE-2021-32723 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.22.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz</a></p>
<p>Path to dependency file: stream-chat-react/package.json</p>
<p>Path to vulnerable library: stream-chat-react/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- react-styleguidist-11.1.4.tgz (Root Library)
- :x: **prismjs-1.22.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/harrinry/stream-chat-react/commit/a80df0ece7581dc566b7c05d5f5ae6ecc66b730f">a80df0ece7581dc566b7c05d5f5ae6ecc66b730f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text.
<p>Publish Date: 2021-06-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723>CVE-2021-32723</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg">https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg</a></p>
<p>Release Date: 2021-06-28</p>
<p>Fix Resolution: prismjs - 1.24.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.22.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-styleguidist:11.1.4;prismjs:1.22.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"prismjs - 1.24.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32723","vulnerabilityDetails":"Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-32723 (Medium) detected in prismjs-1.22.0.tgz - ## CVE-2021-32723 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.22.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz</a></p>
<p>Path to dependency file: stream-chat-react/package.json</p>
<p>Path to vulnerable library: stream-chat-react/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- react-styleguidist-11.1.4.tgz (Root Library)
- :x: **prismjs-1.22.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/harrinry/stream-chat-react/commit/a80df0ece7581dc566b7c05d5f5ae6ecc66b730f">a80df0ece7581dc566b7c05d5f5ae6ecc66b730f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text.
<p>Publish Date: 2021-06-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723>CVE-2021-32723</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg">https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg</a></p>
<p>Release Date: 2021-06-28</p>
<p>Fix Resolution: prismjs - 1.24.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.22.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-styleguidist:11.1.4;prismjs:1.22.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"prismjs - 1.24.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32723","vulnerabilityDetails":"Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
|
non_main
|
cve medium detected in prismjs tgz cve medium severity vulnerability vulnerable library prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file stream chat react package json path to vulnerable library stream chat react node modules prismjs package json dependency hierarchy react styleguidist tgz root library x prismjs tgz vulnerable library found in head commit a href found in base branch master vulnerability details prism is a syntax highlighting library some languages before are vulnerable to regular expression denial of service redos when prism is used to highlight untrusted user given text an attacker can craft a string that will take a very very long time to highlight this problem has been fixed in prism as a workaround do not use asciidoc or erb to highlight untrusted text other languages are not affected and can be used to highlight untrusted text publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution prismjs isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree react styleguidist prismjs isminimumfixversionavailable true minimumfixversion prismjs isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails prism is a syntax highlighting library some languages before are vulnerable to regular expression denial of service redos when prism is used to highlight untrusted user given text an attacker can craft a string that will take a very very long time to highlight this problem has been fixed in prism as a workaround do not use asciidoc or erb to highlight untrusted text other languages are not affected and can be used to highlight untrusted text vulnerabilityurl
| 0
|
943
| 4,674,565,765
|
IssuesEvent
|
2016-10-07 02:00:43
|
tomchentw/react-google-maps
|
https://api.github.com/repos/tomchentw/react-google-maps
|
closed
|
InfoWindow clears on update
|
CALL_FOR_MAINTAINERS
|
Hi guys,
I'm currently using `react-google-maps` to create a public transit map and I'm using the `InfoWindow` component to display stop timetables.
When I open the `InfoWindow` component through a marker, the window opens as normal, however, I am using JQuery and Ajax to load the data from a server into the React element. When the Ajax request finishes and the data is stored in the React element state, the `InfoWindow` clears and empties.
My code is here: https://gist.github.com/uniquoooo/c80e6bda7bc901d8a53753c2d10e15a1
I'm not sure why this is happening, could someone look at my code and see if I am doing anything wrong?
Thanks
|
True
|
InfoWindow clears on update - Hi guys,
I'm currently using `react-google-maps` to create a public transit map and I'm using the `InfoWindow` component to display stop timetables.
When I open the `InfoWindow` component through a marker, the window opens as normal, however, I am using JQuery and Ajax to load the data from a server into the React element. When the Ajax request finishes and the data is stored in the React element state, the `InfoWindow` clears and empties.
My code is here: https://gist.github.com/uniquoooo/c80e6bda7bc901d8a53753c2d10e15a1
I'm not sure why this is happening, could someone look at my code and see if I am doing anything wrong?
Thanks
|
main
|
infowindow clears on update hi guys i m currently using react google maps to create a public transit map and i m using the infowindow component to display stop timetables when i open the infowindow component through a marker the window opens as normal however i am using jquery and ajax to load the data from a server into the react element when the ajax request finishes and the data is stored in the react element state the infowindow clears and empties my code is here i m not sure why this is happening could someone look at my code and see if i am doing anything wrong thanks
| 1
|
545,403
| 15,949,885,119
|
IssuesEvent
|
2021-04-15 08:00:01
|
mobigen/IRIS-BigData-Platform
|
https://api.github.com/repos/mobigen/IRIS-BigData-Platform
|
closed
|
[Studio][개발요청] 변수에 저장된 text type 값들을 "예제1", "예제2", "예제3" 처럼 따옴표와 콤마로 만드는 함수 또는 명령어
|
#DSL #Studio Priority: P0 Status: Backlog
|
## 기능 요청 ##
b-iris.mobigen.com demo 계정
- studio : EDU_예제_그룹_여집합구하기
http://b-iris.mobigen.com:80/studio/exported/89d7a85f1c784c24ac0306f366dfc80623435f48882e47ae9704e66a76b26d5e
에서
- "common_data_3"
<img width="479" alt="스크린샷 2021-03-22 오전 8 29 48" src="https://user-images.githubusercontent.com/5126497/111924633-e9884600-8ae8-11eb-87cf-d76ef35b8177.png">
- "common_data_2"
<img width="478" alt="스크린샷 2021-03-22 오전 8 31 48" src="https://user-images.githubusercontent.com/5126497/111924683-0fade600-8ae9-11eb-9020-c18d4e37921f.png">
- 원하는 구문
```
* |
sql "SELECT explode(array_except( array( 'setosa','virginica' ), array( 'versicolor','virginica' ) ) ) as my_uniqSpecies
from angora"
```
- 변수명에서는 문자열에 따옴표가 있지만, studio 에서 사용할 때는 **따옴표가 삭제** 됩니다. 그래서 nunjucks 로 for loop 구문을 써서 따옴표를 붙이는 작업을 해야 합니다.
```
* | {% set my_string = '' %}
{% for i in common_data_3.results -%}
{% set my_string = my_string ~ "'" ~ i ~ "'," %}
{%- endfor %}
{% set my_string = my_string.slice(0,-1) %}
{% set my_string2 = '' %}
{% for i in common_data_2.results -%}
{% set my_string2 = my_string2 ~ "'" ~ i ~ "'," %}
{%- endfor %}
{% set my_string2 = my_string2.slice(0,-1) %}
sql "SELECT explode(array_except( array( {{ my_string }} ), array( {{ my_string2 }} ) ) ) as my_uniqSpecies
from angora"
```
## 원하는 솔루션 설명 ##
nunjucks 없이
변수에 TYPE : text 이면 따옴표가 붙은 채로 사용할 수 있는 기능
검색어에서
- 현재
```
* | where A in ( ${common_data_3.results} ) => * | where A in ( setosa,virginica )
```
- 요청
```
* | where A in ( ${common_data_3.results} ) => * | where A in ( 'setosa', 'virginica' )
```
로 사용할 수 있는 방법이 필요합니다.
## 고려한 다른 대안 ##
{ 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. }
## 기타 ##
기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다.
|
1.0
|
[Studio][개발요청] 변수에 저장된 text type 값들을 "예제1", "예제2", "예제3" 처럼 따옴표와 콤마로 만드는 함수 또는 명령어 - ## 기능 요청 ##
b-iris.mobigen.com demo 계정
- studio : EDU_예제_그룹_여집합구하기
http://b-iris.mobigen.com:80/studio/exported/89d7a85f1c784c24ac0306f366dfc80623435f48882e47ae9704e66a76b26d5e
에서
- "common_data_3"
<img width="479" alt="스크린샷 2021-03-22 오전 8 29 48" src="https://user-images.githubusercontent.com/5126497/111924633-e9884600-8ae8-11eb-87cf-d76ef35b8177.png">
- "common_data_2"
<img width="478" alt="스크린샷 2021-03-22 오전 8 31 48" src="https://user-images.githubusercontent.com/5126497/111924683-0fade600-8ae9-11eb-9020-c18d4e37921f.png">
- 원하는 구문
```
* |
sql "SELECT explode(array_except( array( 'setosa','virginica' ), array( 'versicolor','virginica' ) ) ) as my_uniqSpecies
from angora"
```
- 변수명에서는 문자열에 따옴표가 있지만, studio 에서 사용할 때는 **따옴표가 삭제** 됩니다. 그래서 nunjucks 로 for loop 구문을 써서 따옴표를 붙이는 작업을 해야 합니다.
```
* | {% set my_string = '' %}
{% for i in common_data_3.results -%}
{% set my_string = my_string ~ "'" ~ i ~ "'," %}
{%- endfor %}
{% set my_string = my_string.slice(0,-1) %}
{% set my_string2 = '' %}
{% for i in common_data_2.results -%}
{% set my_string2 = my_string2 ~ "'" ~ i ~ "'," %}
{%- endfor %}
{% set my_string2 = my_string2.slice(0,-1) %}
sql "SELECT explode(array_except( array( {{ my_string }} ), array( {{ my_string2 }} ) ) ) as my_uniqSpecies
from angora"
```
## 원하는 솔루션 설명 ##
nunjucks 없이
변수에 TYPE : text 이면 따옴표가 붙은 채로 사용할 수 있는 기능
검색어에서
- 현재
```
* | where A in ( ${common_data_3.results} ) => * | where A in ( setosa,virginica )
```
- 요청
```
* | where A in ( ${common_data_3.results} ) => * | where A in ( 'setosa', 'virginica' )
```
로 사용할 수 있는 방법이 필요합니다.
## 고려한 다른 대안 ##
{ 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. }
## 기타 ##
기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다.
|
non_main
|
변수에 저장된 text type 값들을 처럼 따옴표와 콤마로 만드는 함수 또는 명령어 기능 요청 b iris mobigen com demo 계정 studio edu 예제 그룹 여집합구하기 에서 common data img width alt 스크린샷 오전 src common data img width alt 스크린샷 오전 src 원하는 구문 sql select explode array except array setosa virginica array versicolor virginica as my uniqspecies from angora 변수명에서는 문자열에 따옴표가 있지만 studio 에서 사용할 때는 따옴표가 삭제 됩니다 그래서 nunjucks 로 for loop 구문을 써서 따옴표를 붙이는 작업을 해야 합니다 set my string for i in common data results set my string my string i endfor set my string my string slice set my for i in common data results set my my i endfor set my my slice sql select explode array except array my string array my as my uniqspecies from angora 원하는 솔루션 설명 nunjucks 없이 변수에 type text 이면 따옴표가 붙은 채로 사용할 수 있는 기능 검색어에서 현재 where a in common data results where a in setosa virginica 요청 where a in common data results where a in setosa virginica 로 사용할 수 있는 방법이 필요합니다 고려한 다른 대안 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다 기타 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다
| 0
|
35
| 2,582,640,188
|
IssuesEvent
|
2015-02-15 13:50:23
|
0robustus1/savage
|
https://api.github.com/repos/0robustus1/savage
|
opened
|
switch test-framework
|
maintainability
|
I don't really like the standard unit-testing syntax. We should take a look at other clojure libraries. Maybe try out Midje?
|
True
|
switch test-framework - I don't really like the standard unit-testing syntax. We should take a look at other clojure libraries. Maybe try out Midje?
|
main
|
switch test framework i don t really like the standard unit testing syntax we should take a look at other clojure libraries maybe try out midje
| 1
|
270,465
| 8,460,817,487
|
IssuesEvent
|
2018-10-22 19:57:14
|
the-better-gibgab/gibgab-frontend
|
https://api.github.com/repos/the-better-gibgab/gibgab-frontend
|
opened
|
View Post
|
Priority: High Status: Available
|
### View Post
As a user
I want to view the content of a post
So that I can interact with the social network
#### Acceptance Criteria
Given : Post data
When : The user chooses a post
Then : Post and comment data is displayed to the user
#### Story Points
Points: 4
|
1.0
|
View Post - ### View Post
As a user
I want to view the content of a post
So that I can interact with the social network
#### Acceptance Criteria
Given : Post data
When : The user chooses a post
Then : Post and comment data is displayed to the user
#### Story Points
Points: 4
|
non_main
|
view post view post as a user i want to view the content of a post so that i can interact with the social network acceptance criteria given post data when the user chooses a post then post and comment data is displayed to the user story points points
| 0
|
5,219
| 26,479,630,348
|
IssuesEvent
|
2023-01-17 13:49:16
|
OpenRefine/OpenRefine
|
https://api.github.com/repos/OpenRefine/OpenRefine
|
closed
|
Simplify the CI/CD workflows to use the builtin Maven cache feature in setup-java action that now supports it
|
bug maintainability CI/CD
|
We can now remove the actions/cache steps and replace them with a config setting under `setup-java` action. It automatically uses actions/cache already under the hood and will create the fileHash for the Maven `**/pom.xml` automatically.
Documention: https://github.com/actions/setup-java/blob/main/README.md#caching-packages-dependencies
Usage such as:
```
steps:
- uses: actions/checkout@v3
- uses: actions/setup-java@v3
with:
distribution: 'temurin'
java-version: '17'
cache: 'maven'
- name: Build with Maven
run: mvn -B package --file pom.xml
```
### To Reproduce
Steps to reproduce the behavior:
1. Run a PR workflow.
### Current Results
Extra steps not need for `actions/cache` separately.
### Expected Behavior
cleaner and less verbose build/test output in logs.
|
True
|
Simplify the CI/CD workflows to use the builtin Maven cache feature in setup-java action that now supports it - We can now remove the actions/cache steps and replace them with a config setting under `setup-java` action. It automatically uses actions/cache already under the hood and will create the fileHash for the Maven `**/pom.xml` automatically.
Documention: https://github.com/actions/setup-java/blob/main/README.md#caching-packages-dependencies
Usage such as:
```
steps:
- uses: actions/checkout@v3
- uses: actions/setup-java@v3
with:
distribution: 'temurin'
java-version: '17'
cache: 'maven'
- name: Build with Maven
run: mvn -B package --file pom.xml
```
### To Reproduce
Steps to reproduce the behavior:
1. Run a PR workflow.
### Current Results
Extra steps not need for `actions/cache` separately.
### Expected Behavior
cleaner and less verbose build/test output in logs.
|
main
|
simplify the ci cd workflows to use the builtin maven cache feature in setup java action that now supports it we can now remove the actions cache steps and replace them with a config setting under setup java action it automatically uses actions cache already under the hood and will create the filehash for the maven pom xml automatically documention usage such as steps uses actions checkout uses actions setup java with distribution temurin java version cache maven name build with maven run mvn b package file pom xml to reproduce steps to reproduce the behavior run a pr workflow current results extra steps not need for actions cache separately expected behavior cleaner and less verbose build test output in logs
| 1
|
4,739
| 24,460,271,323
|
IssuesEvent
|
2022-10-07 10:28:21
|
mozilla/foundation.mozilla.org
|
https://api.github.com/repos/mozilla/foundation.mozilla.org
|
closed
|
Numbers in templates are still getting localized
|
bug 🦠 engineering Maintain
|
This is caused by `L10N=True`: even though we removed `USE_THOUSAND_SEPARATOR`, which turns this behaviour off for English/default, the `L10N` setting "overrules" that and turns on silent number reformatting with even less control. There is the `{{ value | unlocalize }}` template filter from the `l10n` tags that we can use, but it would be fantastic if we can find out how to actually turns this off for every locale, except where `intcomma` is explicitly used.
|
True
|
Numbers in templates are still getting localized - This is caused by `L10N=True`: even though we removed `USE_THOUSAND_SEPARATOR`, which turns this behaviour off for English/default, the `L10N` setting "overrules" that and turns on silent number reformatting with even less control. There is the `{{ value | unlocalize }}` template filter from the `l10n` tags that we can use, but it would be fantastic if we can find out how to actually turns this off for every locale, except where `intcomma` is explicitly used.
|
main
|
numbers in templates are still getting localized this is caused by true even though we removed use thousand separator which turns this behaviour off for english default the setting overrules that and turns on silent number reformatting with even less control there is the value unlocalize template filter from the tags that we can use but it would be fantastic if we can find out how to actually turns this off for every locale except where intcomma is explicitly used
| 1
|
157,373
| 12,371,339,342
|
IssuesEvent
|
2020-05-18 18:23:14
|
fitbenchmarking/fitbenchmarking
|
https://api.github.com/repos/fitbenchmarking/fitbenchmarking
|
opened
|
Better windows testing in test_parsers.py
|
Maintenance Priority - low Testing
|
**Which module and class/method/function does this relate to?**
This relates to test_parsers.py.
Tests on windows limited to `formats = ['nist']`
Extend this to all combinations that are supported by basic install (i.e. pip install .) in some way
Relates to #526
|
1.0
|
Better windows testing in test_parsers.py - **Which module and class/method/function does this relate to?**
This relates to test_parsers.py.
Tests on windows limited to `formats = ['nist']`
Extend this to all combinations that are supported by basic install (i.e. pip install .) in some way
Relates to #526
|
non_main
|
better windows testing in test parsers py which module and class method function does this relate to this relates to test parsers py tests on windows limited to formats extend this to all combinations that are supported by basic install i e pip install in some way relates to
| 0
|
99,949
| 12,491,085,395
|
IssuesEvent
|
2020-06-01 02:47:25
|
WestpacGEL/GEL
|
https://api.github.com/repos/WestpacGEL/GEL
|
closed
|
Component: Modal
|
KTM-Design bug modal
|
## Features
- [x] Sizing
- [x] Examples
- [ ] Jest tests
- [ ] Cypress tests
- [x] Docs
## To discuss
- Whether separate modal header and body components are necessary
- Use of GEL Button for the close icon
## Feedback from Kate's review
- **modal/default** - Doesn't seem to have all styling applied, ie Hero coloured header separator line. _JPS: This has now been fixed... In the Modal PR (https://github.com/WestpacGEL/GEL/pull/199), awaiting merge)
## Jonny review
- Currently in @jaortiz ’s PR https://github.com/WestpacGEL/GEL/pull/199. Will wait before giving to Intopia.
_JPS: now merged_
|
1.0
|
Component: Modal - ## Features
- [x] Sizing
- [x] Examples
- [ ] Jest tests
- [ ] Cypress tests
- [x] Docs
## To discuss
- Whether separate modal header and body components are necessary
- Use of GEL Button for the close icon
## Feedback from Kate's review
- **modal/default** - Doesn't seem to have all styling applied, ie Hero coloured header separator line. _JPS: This has now been fixed... In the Modal PR (https://github.com/WestpacGEL/GEL/pull/199), awaiting merge)
## Jonny review
- Currently in @jaortiz ’s PR https://github.com/WestpacGEL/GEL/pull/199. Will wait before giving to Intopia.
_JPS: now merged_
|
non_main
|
component modal features sizing examples jest tests cypress tests docs to discuss whether separate modal header and body components are necessary use of gel button for the close icon feedback from kate s review modal default doesn t seem to have all styling applied ie hero coloured header separator line jps this has now been fixed in the modal pr awaiting merge jonny review currently in jaortiz ’s pr will wait before giving to intopia jps now merged
| 0
|
3,674
| 15,036,029,509
|
IssuesEvent
|
2021-02-02 14:48:38
|
IITIDIDX597/sp_2021_team1
|
https://api.github.com/repos/IITIDIDX597/sp_2021_team1
|
opened
|
Tagging articles for better search
|
Epic: 5 Maintaining the system Story Week 3
|
**Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care.
**Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform.
**Sub-Hill Statements:**
1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments.
### **Story Details:**
As an: administrator
I want: to be able to tag the article with various labels according to the topic
So that: it's easier for people to search
|
True
|
Tagging articles for better search - **Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care.
**Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform.
**Sub-Hill Statements:**
1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments.
### **Story Details:**
As an: administrator
I want: to be able to tag the article with various labels according to the topic
So that: it's easier for people to search
|
main
|
tagging articles for better search project goal s lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way while at the same time foster deeper learning experiences in order to deliver better abilitylab patient care hill statement individual clinicians can reference relevant continuously evolving information for their patient s therapy needs to self manage their approach patient care plan development in a single platform sub hill statements the learning platform will be routinely updated with s lab s own research advancements as well as outside discoveries and best practices developed for rehabilitation treatments story details as an administrator i want to be able to tag the article with various labels according to the topic so that it s easier for people to search
| 1
|
2,887
| 10,319,603,894
|
IssuesEvent
|
2019-08-30 17:59:53
|
backdrop-ops/contrib
|
https://api.github.com/repos/backdrop-ops/contrib
|
closed
|
First contrib module: simple EU cookies
|
Maintainer application
|
I wrote my first module for backdrop. Displays a warning about the use of cookies.
It is simply an implementation of jquery module: cookiebar.
I would like other people to take advantage of this work. I would appreciate it got in the repository modules backdrop.
The link to the module: https://github.com/davidvina/simple_EU_cookies
best regards,
and excuse me for my poor English.
|
True
|
First contrib module: simple EU cookies - I wrote my first module for backdrop. Displays a warning about the use of cookies.
It is simply an implementation of jquery module: cookiebar.
I would like other people to take advantage of this work. I would appreciate it got in the repository modules backdrop.
The link to the module: https://github.com/davidvina/simple_EU_cookies
best regards,
and excuse me for my poor English.
|
main
|
first contrib module simple eu cookies i wrote my first module for backdrop displays a warning about the use of cookies it is simply an implementation of jquery module cookiebar i would like other people to take advantage of this work i would appreciate it got in the repository modules backdrop the link to the module best regards and excuse me for my poor english
| 1
|
4,444
| 23,085,904,417
|
IssuesEvent
|
2022-07-26 11:22:52
|
software-mansion/react-native-reanimated
|
https://api.github.com/repos/software-mansion/react-native-reanimated
|
opened
|
Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from` on `export * as default from './Animated';`
|
needs-review maintainer-issue
|
<!--
NOTE: please submit only bug reports here, any new questions or feature requests should be submitted in Discussions:
https://github.com/software-mansion/react-native-reanimated/discussions
-->
## Description
<!--
Tell us what's happening here.
-->
Starting from Reanimated 2.9.0, **if you forget to add Reanimated's Babel plugin**, instead of a user-friendly error message, the following error will appear:
```sh
error: node_modules/react-native-reanimated/src/index.ts: /Users/tomekzaw/RNOS/Issue3397/node_modules/react-native-reanimated/src/index.ts: Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from`.
5 | export * from './reanimated1';
6 | export * from './reanimated2';
> 7 | export * as default from './Animated';
| ^^^^^^^^^^^^
8 |
```
| Android | iOS |
|:-:|:-:|
| <img src="https://user-images.githubusercontent.com/20516055/180992073-0c43b86b-84bb-4cc0-8a0f-fae2cd89beda.png" alt="iOS error" height="700" /> | <img src="https://user-images.githubusercontent.com/20516055/180992066-da1f2227-7dc0-46e8-9e1b-55eb0bd095f0.png" alt="Android error" height="700" /> |

## Solution
1. Add Reanimated's babel plugin to your `babel.config.js`:
```diff
module.exports = {
presets: ['module:metro-react-native-babel-preset'],
+ plugins: ['react-native-reanimated/plugin'],
};
```
2. Clear application cache (depending on your workflow or favourite package manager):
* `yarn start --reset-cache`
* `npm start -- --reset-cache`
* `expo start -c`
3. Reload app
See [installation docs](https://docs.swmansion.com/react-native-reanimated/docs/fundamentals/installation#babel-plugin) for more details.
|
True
|
Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from` on `export * as default from './Animated';` - <!--
NOTE: please submit only bug reports here, any new questions or feature requests should be submitted in Discussions:
https://github.com/software-mansion/react-native-reanimated/discussions
-->
## Description
<!--
Tell us what's happening here.
-->
Starting from Reanimated 2.9.0, **if you forget to add Reanimated's Babel plugin**, instead of a user-friendly error message, the following error will appear:
```sh
error: node_modules/react-native-reanimated/src/index.ts: /Users/tomekzaw/RNOS/Issue3397/node_modules/react-native-reanimated/src/index.ts: Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from`.
5 | export * from './reanimated1';
6 | export * from './reanimated2';
> 7 | export * as default from './Animated';
| ^^^^^^^^^^^^
8 |
```
| Android | iOS |
|:-:|:-:|
| <img src="https://user-images.githubusercontent.com/20516055/180992073-0c43b86b-84bb-4cc0-8a0f-fae2cd89beda.png" alt="iOS error" height="700" /> | <img src="https://user-images.githubusercontent.com/20516055/180992066-da1f2227-7dc0-46e8-9e1b-55eb0bd095f0.png" alt="Android error" height="700" /> |

## Solution
1. Add Reanimated's babel plugin to your `babel.config.js`:
```diff
module.exports = {
presets: ['module:metro-react-native-babel-preset'],
+ plugins: ['react-native-reanimated/plugin'],
};
```
2. Clear application cache (depending on your workflow or favourite package manager):
* `yarn start --reset-cache`
* `npm start -- --reset-cache`
* `expo start -c`
3. Reload app
See [installation docs](https://docs.swmansion.com/react-native-reanimated/docs/fundamentals/installation#babel-plugin) for more details.
|
main
|
export namespace should be first transformed by babel plugin proposal export namespace from on export as default from animated note please submit only bug reports here any new questions or feature requests should be submitted in discussions description tell us what s happening here starting from reanimated if you forget to add reanimated s babel plugin instead of a user friendly error message the following error will appear sh error node modules react native reanimated src index ts users tomekzaw rnos node modules react native reanimated src index ts export namespace should be first transformed by babel plugin proposal export namespace from export from export from export as default from animated android ios solution add reanimated s babel plugin to your babel config js diff module exports presets plugins clear application cache depending on your workflow or favourite package manager yarn start reset cache npm start reset cache expo start c reload app see for more details
| 1
|
99,562
| 4,056,886,409
|
IssuesEvent
|
2016-05-24 20:11:22
|
ndoit/fenrir
|
https://api.github.com/repos/ndoit/fenrir
|
closed
|
Partial word search not working
|
Priority 2
|
When you type ahead you see results.
Try typing "Empl", you see:

When you hit enter you see:

|
1.0
|
Partial word search not working - When you type ahead you see results.
Try typing "Empl", you see:

When you hit enter you see:

|
non_main
|
partial word search not working when you type ahead you see results try typing empl you see when you hit enter you see
| 0
|
430,797
| 30,200,593,346
|
IssuesEvent
|
2023-07-05 05:06:53
|
sulton-max/backend.caching
|
https://api.github.com/repos/sulton-max/backend.caching
|
opened
|
Add In Memory Cache documentation
|
documentation
|
### Task
Add In Memory Cache documentation including
* When to use it
* What is it
* Why to use it
* How to use it
|
1.0
|
Add In Memory Cache documentation - ### Task
Add In Memory Cache documentation including
* When to use it
* What is it
* Why to use it
* How to use it
|
non_main
|
add in memory cache documentation task add in memory cache documentation including when to use it what is it why to use it how to use it
| 0
|
63,860
| 15,726,854,261
|
IssuesEvent
|
2021-03-29 11:54:31
|
rticommunity/rticonnextdds-examples
|
https://api.github.com/repos/rticommunity/rticonnextdds-examples
|
closed
|
Port flat_data_latency C++11 example to CMake
|
build documentation enhancement good first issue style
|
The example `/examples/connext_dds/flat_data_latency/c++11` is missing the new `README.md` template and the `CMakeLists.txt`.
|
1.0
|
Port flat_data_latency C++11 example to CMake - The example `/examples/connext_dds/flat_data_latency/c++11` is missing the new `README.md` template and the `CMakeLists.txt`.
|
non_main
|
port flat data latency c example to cmake the example examples connext dds flat data latency c is missing the new readme md template and the cmakelists txt
| 0
|
4,841
| 24,955,745,275
|
IssuesEvent
|
2022-11-01 11:33:58
|
Lissy93/dashy
|
https://api.github.com/repos/Lissy93/dashy
|
closed
|
[QUESTION] Can I use actual domain not second level domain to fetch favicon?
|
🤷♂️ Question 👤 Awaiting Maintainer Response
|
### Question
hi, I found dashy's favicon api always use sencod-level domain to fetch a favicon, just like google.com for google. Is there any method to use the actual domain to fetch a favicon,no matter it domain level,just like docs.google.com for google Doc
### Category
Using Icons
### Please tick the boxes
- [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number)
- [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue)
- [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide
- [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
|
True
|
[QUESTION] Can I use actual domain not second level domain to fetch favicon? - ### Question
hi, I found dashy's favicon api always use sencod-level domain to fetch a favicon, just like google.com for google. Is there any method to use the actual domain to fetch a favicon,no matter it domain level,just like docs.google.com for google Doc
### Category
Using Icons
### Please tick the boxes
- [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number)
- [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue)
- [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide
- [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
|
main
|
can i use actual domain not second level domain to fetch favicon question hi i found dashy s favicon api always use sencod level domain to fetch a favicon just like google com for google is there any method to use the actual domain to fetch a favicon,no matter it domain level,just like docs google com for google doc category using icons please tick the boxes you are using a version of dashy check the first two digits of the version number you ve checked that this you ve checked the and guide you agree to the
| 1
|
276
| 3,041,430,828
|
IssuesEvent
|
2015-08-07 21:16:07
|
angular/universal
|
https://api.github.com/repos/angular/universal
|
opened
|
Move preboot tests back to modules/preboot/test
|
module: preboot.js type: RepoMaintainerExperience
|
e2e tests are global and remain in root while module tests should remain within the module
|
True
|
Move preboot tests back to modules/preboot/test - e2e tests are global and remain in root while module tests should remain within the module
|
main
|
move preboot tests back to modules preboot test tests are global and remain in root while module tests should remain within the module
| 1
|
5,797
| 30,708,962,495
|
IssuesEvent
|
2023-07-27 08:27:31
|
precice/precice
|
https://api.github.com/repos/precice/precice
|
opened
|
Refactor B-Spline interpolation into a separate function
|
maintainability
|
**Please describe the problem you are trying to solve.**
We are currently using B-Spline interpolation in waveforms provided by Eigen.
This is implemented as part of the `time::Storage`.
The interface is clear though: `(timestamps, samples, timeToSample) -> sample`
We should extract this into a separate file, and test it.
**Additional context**
#1746
|
True
|
Refactor B-Spline interpolation into a separate function - **Please describe the problem you are trying to solve.**
We are currently using B-Spline interpolation in waveforms provided by Eigen.
This is implemented as part of the `time::Storage`.
The interface is clear though: `(timestamps, samples, timeToSample) -> sample`
We should extract this into a separate file, and test it.
**Additional context**
#1746
|
main
|
refactor b spline interpolation into a separate function please describe the problem you are trying to solve we are currently using b spline interpolation in waveforms provided by eigen this is implemented as part of the time storage the interface is clear though timestamps samples timetosample sample we should extract this into a separate file and test it additional context
| 1
|
4,971
| 25,540,480,841
|
IssuesEvent
|
2022-11-29 14:59:45
|
centerofci/mathesar
|
https://api.github.com/repos/centerofci/mathesar
|
closed
|
Provide targeted error messages when tables & columns involved in an exploration are no longer present
|
type: bug work: backend work: frontend restricted: maintainers status: started
|
* Create a new exploration with a base_table and some initial columns
* Delete some of the columns in the table directly
- Notice that the API responds with a 500 error code
* Delete the table which is the base_table of the query
- The query also gets deleted
- We need to show these in the warning message while prompting for user confirmation
* Delete a linked table whose columns we use as initial_columns in the query
- The API responds with a 500
Expected:
* The frontend needs to show the columns and highlight the ones that are no longer present
* The API response needs to provide both the column structure and targetted errors
|
True
|
Provide targeted error messages when tables & columns involved in an exploration are no longer present - * Create a new exploration with a base_table and some initial columns
* Delete some of the columns in the table directly
- Notice that the API responds with a 500 error code
* Delete the table which is the base_table of the query
- The query also gets deleted
- We need to show these in the warning message while prompting for user confirmation
* Delete a linked table whose columns we use as initial_columns in the query
- The API responds with a 500
Expected:
* The frontend needs to show the columns and highlight the ones that are no longer present
* The API response needs to provide both the column structure and targetted errors
|
main
|
provide targeted error messages when tables columns involved in an exploration are no longer present create a new exploration with a base table and some initial columns delete some of the columns in the table directly notice that the api responds with a error code delete the table which is the base table of the query the query also gets deleted we need to show these in the warning message while prompting for user confirmation delete a linked table whose columns we use as initial columns in the query the api responds with a expected the frontend needs to show the columns and highlight the ones that are no longer present the api response needs to provide both the column structure and targetted errors
| 1
|
3,993
| 18,495,026,731
|
IssuesEvent
|
2021-10-19 07:43:50
|
camunda-cloud/zeebe
|
https://api.github.com/repos/camunda-cloud/zeebe
|
closed
|
Define a way to manage the state access
|
Type: Maintenance Impact: Performance Impact: Maintainability
|
**Description**
Define a way how to
* avoid multiple reads/writes of one state entity within one BPMN step.
* check invariants
* track state changes for debugging
Related to #4227
|
True
|
Define a way to manage the state access - **Description**
Define a way how to
* avoid multiple reads/writes of one state entity within one BPMN step.
* check invariants
* track state changes for debugging
Related to #4227
|
main
|
define a way to manage the state access description define a way how to avoid multiple reads writes of one state entity within one bpmn step check invariants track state changes for debugging related to
| 1
|
597,560
| 18,166,487,258
|
IssuesEvent
|
2021-09-27 15:05:43
|
OpenNebula/one
|
https://api.github.com/repos/OpenNebula/one
|
opened
|
Expose cleanup flag in oneprovisoin
|
Type: Feature Status: Accepted Priority: Normal Category: Provision
|
**Description**
Delete a provision fails if trying to delete a provision that contains virtual machines and images. To avoid this issue you need to use --cleanup flag in oneprovision, but this is not exposed through the GUI.
**Use case**
Erase a provision from the GUI
**Interface Changes**
Add a checkbox on the delete dialog to activate the flag.
**Additional Context**
Add any other context or screenshots about the feature request here. Or any other alternative you have considered to addressed this new feature.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
|
1.0
|
Expose cleanup flag in oneprovisoin - **Description**
Delete a provision fails if trying to delete a provision that contains virtual machines and images. To avoid this issue you need to use --cleanup flag in oneprovision, but this is not exposed through the GUI.
**Use case**
Erase a provision from the GUI
**Interface Changes**
Add a checkbox on the delete dialog to activate the flag.
**Additional Context**
Add any other context or screenshots about the feature request here. Or any other alternative you have considered to addressed this new feature.
<!--////////////////////////////////////////////-->
<!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM -->
<!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS -->
<!-- PROGRESS WILL BE REFLECTED HERE -->
<!--////////////////////////////////////////////-->
## Progress Status
- [ ] Branch created
- [ ] Code committed to development branch
- [ ] Testing - QA
- [ ] Documentation
- [ ] Release notes - resolved issues, compatibility, known issues
- [ ] Code committed to upstream release/hotfix branches
- [ ] Documentation committed to upstream release/hotfix branches
|
non_main
|
expose cleanup flag in oneprovisoin description delete a provision fails if trying to delete a provision that contains virtual machines and images to avoid this issue you need to use cleanup flag in oneprovision but this is not exposed through the gui use case erase a provision from the gui interface changes add a checkbox on the delete dialog to activate the flag additional context add any other context or screenshots about the feature request here or any other alternative you have considered to addressed this new feature progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
| 0
|
91,686
| 8,316,500,842
|
IssuesEvent
|
2018-09-25 09:13:37
|
Microsoft/AzureStorageExplorer
|
https://api.github.com/repos/Microsoft/AzureStorageExplorer
|
opened
|
Display a tedious message when opening blobs/files
|
testing
|
**Storage Explorer Version**: 1.5.0
**Platform/OS Version**: Windows 10/ Linux Ubuntu 16.04/ MacOS High Sierra
**Architecture**: ia32/x64
**Build Number**: 20180924.1
**Commit**: 1a75b88f
**Regression From**: Previous release 1.4.2(20180917.1)
#### Steps to Reproduce: ####
1. Open one blob container then upload a blob to it.
2. Try to open the uploaded blob -> Check the displayed message in Activities.
#### Expected Experience: ####
The tedious message disappears after opening the blob/file successfully.
#### Actual Experience: ####
Display a tedious message in Activities.

|
1.0
|
Display a tedious message when opening blobs/files - **Storage Explorer Version**: 1.5.0
**Platform/OS Version**: Windows 10/ Linux Ubuntu 16.04/ MacOS High Sierra
**Architecture**: ia32/x64
**Build Number**: 20180924.1
**Commit**: 1a75b88f
**Regression From**: Previous release 1.4.2(20180917.1)
#### Steps to Reproduce: ####
1. Open one blob container then upload a blob to it.
2. Try to open the uploaded blob -> Check the displayed message in Activities.
#### Expected Experience: ####
The tedious message disappears after opening the blob/file successfully.
#### Actual Experience: ####
Display a tedious message in Activities.

|
non_main
|
display a tedious message when opening blobs files storage explorer version platform os version windows linux ubuntu macos high sierra architecture build number commit regression from previous release steps to reproduce open one blob container then upload a blob to it try to open the uploaded blob check the displayed message in activities expected experience the tedious message disappears after opening the blob file successfully actual experience display a tedious message in activities
| 0
|
2,036
| 6,848,588,102
|
IssuesEvent
|
2017-11-13 19:02:29
|
tgstation/tgstation-server
|
https://api.github.com/repos/tgstation/tgstation-server
|
closed
|
Console mode for the service/Seperate actual hosting from service integration
|
Feature Request Maintainability Issue Work In Progress
|
Would make debugging much easier
|
True
|
Console mode for the service/Seperate actual hosting from service integration - Would make debugging much easier
|
main
|
console mode for the service seperate actual hosting from service integration would make debugging much easier
| 1
|
176,124
| 14,564,131,811
|
IssuesEvent
|
2020-12-17 04:17:43
|
lammps/lammps
|
https://api.github.com/repos/lammps/lammps
|
opened
|
Review use of THERMO_ENERGY in `Fix::setmask()` and methods of computing of the conserved quantity in MD
|
documentation maintenance
|
There are some inconsistencies and unusual behavior due to choices made a long time ago, that need to be reviewed and possibly some changes made. For details, please see the discussion on PR #2515
Further discussion results and proposed changes should be documented here.
|
1.0
|
Review use of THERMO_ENERGY in `Fix::setmask()` and methods of computing of the conserved quantity in MD - There are some inconsistencies and unusual behavior due to choices made a long time ago, that need to be reviewed and possibly some changes made. For details, please see the discussion on PR #2515
Further discussion results and proposed changes should be documented here.
|
non_main
|
review use of thermo energy in fix setmask and methods of computing of the conserved quantity in md there are some inconsistencies and unusual behavior due to choices made a long time ago that need to be reviewed and possibly some changes made for details please see the discussion on pr further discussion results and proposed changes should be documented here
| 0
|
894
| 4,553,934,052
|
IssuesEvent
|
2016-09-13 07:35:41
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
docker: 'reloaded' state recreates container everytime if volumes-from is used
|
affects_2.0 bug_report cloud docker P3 waiting_on_maintainer
|
Hi,
according to the docs it should recreate the container if parameters changed:
> "reloaded" asserts that all matching containers are running and restarts any that have any images or configuration out of date.
But if you run this simple play multiple times, ansible will recreate the container every time even though no parameters have changed:
```
- hosts: 127.0.0.1
connection: local
tasks:
- docker: image=busybox name=test-data volumes=/data command=/bin/true state=present
- docker:
image: ubuntu
name: test
state: reloaded
command: "nc -l -k 2342"
volumes_from: test-data
```
In general state reloaded doesn't seem to work properly. See also #1129 which I just opened for a different issue with this.
|
True
|
docker: 'reloaded' state recreates container everytime if volumes-from is used - Hi,
according to the docs it should recreate the container if parameters changed:
> "reloaded" asserts that all matching containers are running and restarts any that have any images or configuration out of date.
But if you run this simple play multiple times, ansible will recreate the container every time even though no parameters have changed:
```
- hosts: 127.0.0.1
connection: local
tasks:
- docker: image=busybox name=test-data volumes=/data command=/bin/true state=present
- docker:
image: ubuntu
name: test
state: reloaded
command: "nc -l -k 2342"
volumes_from: test-data
```
In general state reloaded doesn't seem to work properly. See also #1129 which I just opened for a different issue with this.
|
main
|
docker reloaded state recreates container everytime if volumes from is used hi according to the docs it should recreate the container if parameters changed reloaded asserts that all matching containers are running and restarts any that have any images or configuration out of date but if you run this simple play multiple times ansible will recreate the container every time even though no parameters have changed hosts connection local tasks docker image busybox name test data volumes data command bin true state present docker image ubuntu name test state reloaded command nc l k volumes from test data in general state reloaded doesn t seem to work properly see also which i just opened for a different issue with this
| 1
|
3,373
| 13,057,324,638
|
IssuesEvent
|
2020-07-30 07:05:29
|
short-d/short
|
https://api.github.com/repos/short-d/short
|
closed
|
[Refactor] Replace mustParseTime with must.Time across the codebase
|
maintainability
|
**What is frustrating you?**
The existing tests use [mustParseTime](https://github.com/short-d/short/blob/master/backend/app/adapter/sqldb/main_test.go#L54) helper to convert human readable time in the test case to help test readers understand the test case. However, it does NOT inherently support returning `Time pointer`, preventing time string being directly embed into the test case. https://github.com/short-d/short/pull/965 is adding `must.Time()` to support point return value. This lead to code duplication & inconsistency.
**Your solution**
Remove or refactor `mustParseTime` across the codebase.
|
True
|
[Refactor] Replace mustParseTime with must.Time across the codebase - **What is frustrating you?**
The existing tests use [mustParseTime](https://github.com/short-d/short/blob/master/backend/app/adapter/sqldb/main_test.go#L54) helper to convert human readable time in the test case to help test readers understand the test case. However, it does NOT inherently support returning `Time pointer`, preventing time string being directly embed into the test case. https://github.com/short-d/short/pull/965 is adding `must.Time()` to support point return value. This lead to code duplication & inconsistency.
**Your solution**
Remove or refactor `mustParseTime` across the codebase.
|
main
|
replace mustparsetime with must time across the codebase what is frustrating you the existing tests use helper to convert human readable time in the test case to help test readers understand the test case however it does not inherently support returning time pointer preventing time string being directly embed into the test case is adding must time to support point return value this lead to code duplication inconsistency your solution remove or refactor mustparsetime across the codebase
| 1
|
373,283
| 26,047,476,668
|
IssuesEvent
|
2022-12-22 15:33:14
|
arcanus55/neodigm55
|
https://api.github.com/repos/arcanus55/neodigm55
|
closed
|
Enchanted CTA | Support Material Design icons (google font) via neodigm-icon element
|
documentation enhancement
|
Adding a neodigm-icon element within the button text should display an inline icon.
Create a wiki recipe.
|
1.0
|
Enchanted CTA | Support Material Design icons (google font) via neodigm-icon element - Adding a neodigm-icon element within the button text should display an inline icon.
Create a wiki recipe.
|
non_main
|
enchanted cta support material design icons google font via neodigm icon element adding a neodigm icon element within the button text should display an inline icon create a wiki recipe
| 0
|
279,372
| 30,702,524,042
|
IssuesEvent
|
2023-07-27 01:37:35
|
Satheesh575555/linux-4.1.15
|
https://api.github.com/repos/Satheesh575555/linux-4.1.15
|
closed
|
CVE-2019-19061 (High) detected in linuxlinux-4.6 - autoclosed
|
Mend: dependency security vulnerability
|
## CVE-2019-19061 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15/commit/951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1">951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/iio/imu/adis_buffer.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the adis_update_scan_mode_burst() function in drivers/iio/imu/adis_buffer.c in the Linux kernel before 5.3.9 allows attackers to cause a denial of service (memory consumption), aka CID-9c0530e898f3.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19061>CVE-2019-19061</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061</a></p>
<p>Release Date: 2020-09-25</p>
<p>Fix Resolution: v5.4-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-19061 (High) detected in linuxlinux-4.6 - autoclosed - ## CVE-2019-19061 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15/commit/951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1">951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/iio/imu/adis_buffer.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the adis_update_scan_mode_burst() function in drivers/iio/imu/adis_buffer.c in the Linux kernel before 5.3.9 allows attackers to cause a denial of service (memory consumption), aka CID-9c0530e898f3.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19061>CVE-2019-19061</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061</a></p>
<p>Release Date: 2020-09-25</p>
<p>Fix Resolution: v5.4-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_main
|
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers iio imu adis buffer c vulnerability details a memory leak in the adis update scan mode burst function in drivers iio imu adis buffer c in the linux kernel before allows attackers to cause a denial of service memory consumption aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
4,059
| 18,981,825,719
|
IssuesEvent
|
2021-11-21 02:10:09
|
aws/aws-sam-cli
|
https://api.github.com/repos/aws/aws-sam-cli
|
closed
|
Change set is throwing an error while deploying the template.
|
area/installation blocked/close-if-inactive stage/waiting-for-release maintainer/need-followup
|
<!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- Briefly describe the bug you are facing.-->
When I am deploying the template using SAM deploy. It is giving an error :
**botocore.exceptions.ParamValidationError: Parameter validation failed:
Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken**
I am not using either of **--disable-rollback | --no-disable-rollback** while deploying
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
The command I am using :
sam deploy --template-file packaged.yaml --stack-name <name of stack> --s3-bucket <name of bucket> --capabilities CAPABILITY_IAM --region $AWS_DEFAULT_REGION --parameter-overrides <list of parameters>
### Observed result:
<!-- Please provide command output with `--debug` flag set. -->
Error with --debug flag:
Changeset created successfully. ***
2021-10-27 09:29:38,186 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'fb75737e-a71f-40ff-b874-f9c6d5b53f53', 'installationId': '47f3640c-0c6d-4159-9dac-4f18e6526b4d', 'sessionId': 'a629f491-1932-4811-b3b4-7532d1cdceb3', 'executionEnvironment': 'GitLab', 'ci': True, 'pyversion': '3.8.12', 'samcliVersion': '1.34.1', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': 'eu-west-1', 'commandName': 'sam deploy', 'duration': 16611, 'exitReason': 'ParamValidationError', 'exitCode': 255}}]}
2021-10-27 09:29:38,810 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Traceback (most recent call last):
File "/usr/local/bin/sam", line 8, in <module>
sys.exit(cli())
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/cli_validation/image_repository_validation.py", line 85, in wrapped
return func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/click/decorators.py", line 73, in new_func
return ctx.invoke(f, obj, *args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 157, in wrapped
raise exception # pylint: disable=raising-bad-type
File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 122, in wrapped
return_value = func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/utils/version_checker.py", line 41, in wrapped
actual_result = func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/cli/main.py", line 87, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 163, in cli
do_cli(
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 317, in do_cli
deploy_context.run()
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 161, in run
return self.deploy(
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 265, in deploy
self.deployer.execute_changeset(result["Id"], stack_name, disable_rollback)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/deploy/deployer.py", line 333, in execute_changeset
return self._client.execute_change_set(
File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 386, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 677, in _make_api_call
request_dict = self._convert_to_request_dict(
File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 725, in _convert_to_request_dict
request_dict = self._serializer.serialize_to_request(
File "/usr/local/lib/python3.8/site-packages/botocore/validate.py", line 319, in serialize_to_request
raise ParamValidationError(report=report.generate_report())
botocore.exceptions.ParamValidationError: Parameter validation failed:
Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken
### Expected result:
<!-- Describe what you expected. -->
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: Linux
2. `sam --version`: 1.34.1
3. AWS region: eu-west-1
`Add --debug flag to command you are running`
|
True
|
Change set is throwing an error while deploying the template. - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- Briefly describe the bug you are facing.-->
When I am deploying the template using SAM deploy. It is giving an error :
**botocore.exceptions.ParamValidationError: Parameter validation failed:
Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken**
I am not using either of **--disable-rollback | --no-disable-rollback** while deploying
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
The command I am using :
sam deploy --template-file packaged.yaml --stack-name <name of stack> --s3-bucket <name of bucket> --capabilities CAPABILITY_IAM --region $AWS_DEFAULT_REGION --parameter-overrides <list of parameters>
### Observed result:
<!-- Please provide command output with `--debug` flag set. -->
Error with --debug flag:
Changeset created successfully. ***
2021-10-27 09:29:38,186 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'fb75737e-a71f-40ff-b874-f9c6d5b53f53', 'installationId': '47f3640c-0c6d-4159-9dac-4f18e6526b4d', 'sessionId': 'a629f491-1932-4811-b3b4-7532d1cdceb3', 'executionEnvironment': 'GitLab', 'ci': True, 'pyversion': '3.8.12', 'samcliVersion': '1.34.1', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': 'eu-west-1', 'commandName': 'sam deploy', 'duration': 16611, 'exitReason': 'ParamValidationError', 'exitCode': 255}}]}
2021-10-27 09:29:38,810 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Traceback (most recent call last):
File "/usr/local/bin/sam", line 8, in <module>
sys.exit(cli())
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/cli_validation/image_repository_validation.py", line 85, in wrapped
return func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/click/decorators.py", line 73, in new_func
return ctx.invoke(f, obj, *args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 157, in wrapped
raise exception # pylint: disable=raising-bad-type
File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 122, in wrapped
return_value = func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/utils/version_checker.py", line 41, in wrapped
actual_result = func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/cli/main.py", line 87, in wrapper
return func(*args, **kwargs)
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 163, in cli
do_cli(
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 317, in do_cli
deploy_context.run()
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 161, in run
return self.deploy(
File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 265, in deploy
self.deployer.execute_changeset(result["Id"], stack_name, disable_rollback)
File "/usr/local/lib/python3.8/site-packages/samcli/lib/deploy/deployer.py", line 333, in execute_changeset
return self._client.execute_change_set(
File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 386, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 677, in _make_api_call
request_dict = self._convert_to_request_dict(
File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 725, in _convert_to_request_dict
request_dict = self._serializer.serialize_to_request(
File "/usr/local/lib/python3.8/site-packages/botocore/validate.py", line 319, in serialize_to_request
raise ParamValidationError(report=report.generate_report())
botocore.exceptions.ParamValidationError: Parameter validation failed:
Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken
### Expected result:
<!-- Describe what you expected. -->
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: Linux
2. `sam --version`: 1.34.1
3. AWS region: eu-west-1
`Add --debug flag to command you are running`
|
main
|
change set is throwing an error while deploying the template make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description when i am deploying the template using sam deploy it is giving an error botocore exceptions paramvalidationerror parameter validation failed unknown parameter in input disablerollback must be one of changesetname stackname clientrequesttoken i am not using either of disable rollback no disable rollback while deploying steps to reproduce the command i am using sam deploy template file packaged yaml stack name bucket capabilities capability iam region aws default region parameter overrides observed result error with debug flag changeset created successfully sending telemetry metrics httpsconnectionpool host aws serverless tools telemetry us west amazonaws com port read timed out read timeout traceback most recent call last file usr local bin sam line in sys exit cli file usr local lib site packages click core py line in call return self main args kwargs file usr local lib site packages click core py line in main rv self invoke ctx file usr local lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file usr local lib site packages click core py line in invoke return ctx invoke self callback ctx params file usr local lib site packages click core py line in invoke return callback args kwargs file usr local lib site packages samcli lib cli validation image repository validation py line in wrapped return func args kwargs file usr local lib site packages click decorators py line in new func return ctx invoke f obj args kwargs file usr local lib site packages click core py line in invoke return callback args kwargs file usr local lib site packages samcli lib telemetry metric py line in wrapped raise exception pylint disable raising bad type file usr local lib site packages samcli lib telemetry metric py line in wrapped return value func args kwargs file usr local lib site packages samcli lib utils version checker py line in wrapped actual result func args kwargs file usr local lib site packages samcli cli main py line in wrapper return func args kwargs file usr local lib site packages samcli commands deploy command py line in cli do cli file usr local lib site packages samcli commands deploy command py line in do cli deploy context run file usr local lib site packages samcli commands deploy deploy context py line in run return self deploy file usr local lib site packages samcli commands deploy deploy context py line in deploy self deployer execute changeset result stack name disable rollback file usr local lib site packages samcli lib deploy deployer py line in execute changeset return self client execute change set file usr local lib site packages botocore client py line in api call return self make api call operation name kwargs file usr local lib site packages botocore client py line in make api call request dict self convert to request dict file usr local lib site packages botocore client py line in convert to request dict request dict self serializer serialize to request file usr local lib site packages botocore validate py line in serialize to request raise paramvalidationerror report report generate report botocore exceptions paramvalidationerror parameter validation failed unknown parameter in input disablerollback must be one of changesetname stackname clientrequesttoken expected result additional environment details ex windows mac amazon linux etc os linux sam version aws region eu west add debug flag to command you are running
| 1
|
319,213
| 23,760,456,735
|
IssuesEvent
|
2022-09-01 08:26:11
|
Danilo-Danijel-team/sms_sender
|
https://api.github.com/repos/Danilo-Danijel-team/sms_sender
|
closed
|
Create postman documentation tests for created server routes.
|
documentation test
|
- [x] Create tests based on what services are returning from their API responses. Some of the routes that will be present
are listed in #10 issue.
- [x] Create tests before @danilojezernik will start doing #10 issue, there must be postman documentation tests prepared that will serve as a testing environment for server development. Those tests will @danilojezernik use while developing server logic.
Be aware that this list of routes and requirements will change many times in the future,
and you will add new tests and correct existing ones over many iterations.
|
1.0
|
Create postman documentation tests for created server routes. - - [x] Create tests based on what services are returning from their API responses. Some of the routes that will be present
are listed in #10 issue.
- [x] Create tests before @danilojezernik will start doing #10 issue, there must be postman documentation tests prepared that will serve as a testing environment for server development. Those tests will @danilojezernik use while developing server logic.
Be aware that this list of routes and requirements will change many times in the future,
and you will add new tests and correct existing ones over many iterations.
|
non_main
|
create postman documentation tests for created server routes create tests based on what services are returning from their api responses some of the routes that will be present are listed in issue create tests before danilojezernik will start doing issue there must be postman documentation tests prepared that will serve as a testing environment for server development those tests will danilojezernik use while developing server logic be aware that this list of routes and requirements will change many times in the future and you will add new tests and correct existing ones over many iterations
| 0
|
2,724
| 9,612,818,489
|
IssuesEvent
|
2019-05-13 09:47:37
|
precice/precice
|
https://api.github.com/repos/precice/precice
|
opened
|
Refactor static ids
|
maintainability
|
preCICE currently has to manage a set of `static` IDs due to some prior design decisions regarding simpler object creation. The additional bookkeeping, however, is error-prone and already lead to issues #378.
Affected classes are:
1) Participant via `_participantsSize`
2) Mesh via `_managePropertyID`
3) Data via `_dataCount`
These static IDs should be refactored out of the affected classes into the objects responsible for the generation of IDed objects. The object lifetime of the generators should then correctly handle the bookkeeping.
These generators are most likely going to be configuration objects.
|
True
|
Refactor static ids - preCICE currently has to manage a set of `static` IDs due to some prior design decisions regarding simpler object creation. The additional bookkeeping, however, is error-prone and already lead to issues #378.
Affected classes are:
1) Participant via `_participantsSize`
2) Mesh via `_managePropertyID`
3) Data via `_dataCount`
These static IDs should be refactored out of the affected classes into the objects responsible for the generation of IDed objects. The object lifetime of the generators should then correctly handle the bookkeeping.
These generators are most likely going to be configuration objects.
|
main
|
refactor static ids precice currently has to manage a set of static ids due to some prior design decisions regarding simpler object creation the additional bookkeeping however is error prone and already lead to issues affected classes are participant via participantssize mesh via managepropertyid data via datacount these static ids should be refactored out of the affected classes into the objects responsible for the generation of ided objects the object lifetime of the generators should then correctly handle the bookkeeping these generators are most likely going to be configuration objects
| 1
|
2,603
| 3,916,938,910
|
IssuesEvent
|
2016-04-21 05:29:30
|
steveworley/drush-audit
|
https://api.github.com/repos/steveworley/drush-audit
|
closed
|
Login security is enabled and configured
|
enhancement security
|
- [Login security](https://www.drupal.org/project/login_security), recommend for projects if it is not available
- If the module exists ensure that it is enabled and that it is configured correctly
|
True
|
Login security is enabled and configured - - [Login security](https://www.drupal.org/project/login_security), recommend for projects if it is not available
- If the module exists ensure that it is enabled and that it is configured correctly
|
non_main
|
login security is enabled and configured recommend for projects if it is not available if the module exists ensure that it is enabled and that it is configured correctly
| 0
|
88
| 2,608,261,350
|
IssuesEvent
|
2015-02-26 03:56:33
|
mranney/node_pcap
|
https://api.github.com/repos/mranney/node_pcap
|
closed
|
Setup travis CI build which runs tests and hints code
|
maintainance
|
Having a travis ci build really helps in determining the state of pull requests.
- [x] Setup travis
- [x] Run hinter on travis
- [x] Run unit tests on travis
- [x] Setup coveralls
|
True
|
Setup travis CI build which runs tests and hints code - Having a travis ci build really helps in determining the state of pull requests.
- [x] Setup travis
- [x] Run hinter on travis
- [x] Run unit tests on travis
- [x] Setup coveralls
|
main
|
setup travis ci build which runs tests and hints code having a travis ci build really helps in determining the state of pull requests setup travis run hinter on travis run unit tests on travis setup coveralls
| 1
|
2,310
| 8,279,120,360
|
IssuesEvent
|
2018-09-18 01:18:27
|
spacetelescope/wfc3tools
|
https://api.github.com/repos/spacetelescope/wfc3tools
|
opened
|
TST: Add real tests and put them on Jenkins/Artifactory
|
maintainance
|
For working examples, see `hstcal`, `acstools`, `stistools`, or `calcos`.
|
True
|
TST: Add real tests and put them on Jenkins/Artifactory - For working examples, see `hstcal`, `acstools`, `stistools`, or `calcos`.
|
main
|
tst add real tests and put them on jenkins artifactory for working examples see hstcal acstools stistools or calcos
| 1
|
4,596
| 23,831,872,185
|
IssuesEvent
|
2022-09-05 22:23:23
|
MDAnalysis/mdanalysis
|
https://api.github.com/repos/MDAnalysis/mdanalysis
|
opened
|
MAINT: improve auto-labeling
|
maintainability
|
I should mention that the approach to auto-labeling that we currently use for PRs and stuff was rejected upstream and replaced with something that is supposed to work better re: not mutating/removing labels humans already placed on PRs: https://github.com/scipy/scipy/pull/16870
We may want to look into that--I'm almost certain I've seen the auto-labeling doing that kind of annoying thing of removing a human label over here as well.
|
True
|
MAINT: improve auto-labeling - I should mention that the approach to auto-labeling that we currently use for PRs and stuff was rejected upstream and replaced with something that is supposed to work better re: not mutating/removing labels humans already placed on PRs: https://github.com/scipy/scipy/pull/16870
We may want to look into that--I'm almost certain I've seen the auto-labeling doing that kind of annoying thing of removing a human label over here as well.
|
main
|
maint improve auto labeling i should mention that the approach to auto labeling that we currently use for prs and stuff was rejected upstream and replaced with something that is supposed to work better re not mutating removing labels humans already placed on prs we may want to look into that i m almost certain i ve seen the auto labeling doing that kind of annoying thing of removing a human label over here as well
| 1
|
2,580
| 8,773,417,146
|
IssuesEvent
|
2018-12-18 16:48:41
|
arcticicestudio/nord-docs
|
https://api.github.com/repos/arcticicestudio/nord-docs
|
opened
|
Google Analytics
|
context-workflow scope-maintainability scope-quality scope-stability type-feature
|
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167256-14bbd380-02e9-11e9-8aca-a31baf745cd8.png" width="20%"/></p>
> Associated epics: #86
This issue documents the implementation of [Google Analytics][ga-mark] like documented in the [“Analytics & Statistics” design concept][gh-86].
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167593-c824c800-02e9-11e9-9b70-84b6fc40c05f.png " width="20%"/></p>
The main tool to collect and analyze data will be [Google Analytics][ga-mark]. It is a stable and proven service with a lot of useful configurable features and a reliable persistence.
_Nord Docs_ will use the latest and recommended [gtag.js][gdev-ga-gtag] library that optionally allows, next to Google Analytics itself, the integration of almost all Google Marketing services like e.g. [Google Tag Manager][gdev-tm].
The library will be integrated through [gatsby-plugin-google-gtag][gh-gb-p-ga-tag].
## Tasks
- [ ] Install required packages:
- [gatsby-plugin-google-gtag][npm-gp-gtag]
- [ ] Implement required internal constants.
- [ ] Implement the plugin configuration.
[g-sup-anonip]: https://support.google.com/analytics/answer/2763052
[gh-gb-p-ga-tag]: https://github.com/gatsbyjs/gatsby/tree/master/packages/gatsby-plugin-google-gtag
[gh-86]: https://github.com/arcticicestudio/nord-docs/issues/86
[ga-mark]: https://marketingplatform.google.com/about/analytics
[gdev-ga-gtag]: https://developers.google.com/analytics/devguides/collection/gtagjs
[gdev-tm]: https://developers.google.com/tag-manager
[wiki-a]: https://en.wikipedia.org/wiki/Analytics
[wiki-s]: https://en.wikipedia.org/wiki/Statistics
[wiki-dnt]: https://en.wikipedia.org/wiki/Do_Not_Track
[npm-gp-gtag]: https://www.npmjs.com/package/gatsby-plugin-google-gtag
|
True
|
Google Analytics - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167256-14bbd380-02e9-11e9-8aca-a31baf745cd8.png" width="20%"/></p>
> Associated epics: #86
This issue documents the implementation of [Google Analytics][ga-mark] like documented in the [“Analytics & Statistics” design concept][gh-86].
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167593-c824c800-02e9-11e9-9b70-84b6fc40c05f.png " width="20%"/></p>
The main tool to collect and analyze data will be [Google Analytics][ga-mark]. It is a stable and proven service with a lot of useful configurable features and a reliable persistence.
_Nord Docs_ will use the latest and recommended [gtag.js][gdev-ga-gtag] library that optionally allows, next to Google Analytics itself, the integration of almost all Google Marketing services like e.g. [Google Tag Manager][gdev-tm].
The library will be integrated through [gatsby-plugin-google-gtag][gh-gb-p-ga-tag].
## Tasks
- [ ] Install required packages:
- [gatsby-plugin-google-gtag][npm-gp-gtag]
- [ ] Implement required internal constants.
- [ ] Implement the plugin configuration.
[g-sup-anonip]: https://support.google.com/analytics/answer/2763052
[gh-gb-p-ga-tag]: https://github.com/gatsbyjs/gatsby/tree/master/packages/gatsby-plugin-google-gtag
[gh-86]: https://github.com/arcticicestudio/nord-docs/issues/86
[ga-mark]: https://marketingplatform.google.com/about/analytics
[gdev-ga-gtag]: https://developers.google.com/analytics/devguides/collection/gtagjs
[gdev-tm]: https://developers.google.com/tag-manager
[wiki-a]: https://en.wikipedia.org/wiki/Analytics
[wiki-s]: https://en.wikipedia.org/wiki/Statistics
[wiki-dnt]: https://en.wikipedia.org/wiki/Do_Not_Track
[npm-gp-gtag]: https://www.npmjs.com/package/gatsby-plugin-google-gtag
|
main
|
google analytics associated epics this issue documents the implementation of like documented in the the main tool to collect and analyze data will be it is a stable and proven service with a lot of useful configurable features and a reliable persistence nord docs will use the latest and recommended library that optionally allows next to google analytics itself the integration of almost all google marketing services like e g the library will be integrated through tasks install required packages implement required internal constants implement the plugin configuration
| 1
|
1,971
| 6,694,170,457
|
IssuesEvent
|
2017-10-10 00:04:14
|
duckduckgo/zeroclickinfo-spice
|
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
|
closed
|
Amazon: localised amazon results
|
Maintainer Input Requested
|
Currently, the results are returned from "amazon.com" (by default). It would be **useful** if we could return **localized results**. For example, the user located in Germany should see results from "amazon.de" instead of "amazon.com".
---
IA Page: http://duck.co/ia/view/products
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @bsstoner
|
True
|
Amazon: localised amazon results - Currently, the results are returned from "amazon.com" (by default). It would be **useful** if we could return **localized results**. For example, the user located in Germany should see results from "amazon.de" instead of "amazon.com".
---
IA Page: http://duck.co/ia/view/products
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @bsstoner
|
main
|
amazon localised amazon results currently the results are returned from amazon com by default it would be useful if we could return localized results for example the user located in germany should see results from amazon de instead of amazon com ia page bsstoner
| 1
|
4,947
| 25,455,552,139
|
IssuesEvent
|
2022-11-24 13:55:25
|
pace/bricks
|
https://api.github.com/repos/pace/bricks
|
closed
|
Remove the service argument
|
T::Maintainance
|
Remove the `service` argument in the command so that we can create new microservices with `pace new pay`
|
True
|
Remove the service argument - Remove the `service` argument in the command so that we can create new microservices with `pace new pay`
|
main
|
remove the service argument remove the service argument in the command so that we can create new microservices with pace new pay
| 1
|
2,555
| 8,695,162,685
|
IssuesEvent
|
2018-12-04 14:33:09
|
arcticicestudio/nord-docs
|
https://api.github.com/repos/arcticicestudio/nord-docs
|
closed
|
Theme Global Base Styles
|
context-ui scope-maintainability scope-ux type-feature
|
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/49328407-a4832300-f570-11e8-8613-21c78267b558.png" width="20%" /></p>
> Associated epic: #51
This issue handles the implementation of the basic theme styles like font and background colors. They will be initially used in the global theme styles implemented in #53.
## Tasks
- [x] Implement `background` and `font` CSS module styles.
- [x] Integrate basic `background-color` and `color` in global CSS theme styles.
- Include a `transition` for both attributes to ensure a smooth theme mode transition.
|
True
|
Theme Global Base Styles - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/49328407-a4832300-f570-11e8-8613-21c78267b558.png" width="20%" /></p>
> Associated epic: #51
This issue handles the implementation of the basic theme styles like font and background colors. They will be initially used in the global theme styles implemented in #53.
## Tasks
- [x] Implement `background` and `font` CSS module styles.
- [x] Integrate basic `background-color` and `color` in global CSS theme styles.
- Include a `transition` for both attributes to ensure a smooth theme mode transition.
|
main
|
theme global base styles associated epic this issue handles the implementation of the basic theme styles like font and background colors they will be initially used in the global theme styles implemented in tasks implement background and font css module styles integrate basic background color and color in global css theme styles include a transition for both attributes to ensure a smooth theme mode transition
| 1
|
2,927
| 10,452,286,943
|
IssuesEvent
|
2019-09-19 14:25:49
|
substrate-developer-hub/hacktoberfest
|
https://api.github.com/repos/substrate-developer-hub/hacktoberfest
|
closed
|
Publish Substrate Runtime Library Modules v1.0 to crates.io
|
m-prep m-wontfix t-maintaince
|
Required steps:
- Raise a PR against the substrate 1.0 branch that:
- [ ] clean up dependency tree of sr- and srml-modules (see `cargo tree`)
- [ ] Updates the license, description and other missing fields for each module (see `cargo package` warnings)
- [ ] pins down versions of locally used dependencies
- Publish _(this can only be done by a Parity employee)_:
- publish all srml-, sr- and dependent crates on crates.io (see `cargo publish`)
- make sure [paritytech:core-devs](https://crates.io/teams/github:paritytech:core-devs) is one of the owners of the package on crates.io (see [`cargo owner`](https://doc.rust-lang.org/cargo/reference/publishing.html#cargo-owner))
- Raise a PR against [the templates repository](https://github.com/shawntabrizi/substrate-package), replacing all published crates with version numbers
|
True
|
Publish Substrate Runtime Library Modules v1.0 to crates.io - Required steps:
- Raise a PR against the substrate 1.0 branch that:
- [ ] clean up dependency tree of sr- and srml-modules (see `cargo tree`)
- [ ] Updates the license, description and other missing fields for each module (see `cargo package` warnings)
- [ ] pins down versions of locally used dependencies
- Publish _(this can only be done by a Parity employee)_:
- publish all srml-, sr- and dependent crates on crates.io (see `cargo publish`)
- make sure [paritytech:core-devs](https://crates.io/teams/github:paritytech:core-devs) is one of the owners of the package on crates.io (see [`cargo owner`](https://doc.rust-lang.org/cargo/reference/publishing.html#cargo-owner))
- Raise a PR against [the templates repository](https://github.com/shawntabrizi/substrate-package), replacing all published crates with version numbers
|
main
|
publish substrate runtime library modules to crates io required steps raise a pr against the substrate branch that clean up dependency tree of sr and srml modules see cargo tree updates the license description and other missing fields for each module see cargo package warnings pins down versions of locally used dependencies publish this can only be done by a parity employee publish all srml sr and dependent crates on crates io see cargo publish make sure is one of the owners of the package on crates io see raise a pr against replacing all published crates with version numbers
| 1
|
188,781
| 14,475,410,351
|
IssuesEvent
|
2020-12-10 01:33:14
|
kalexmills/github-vet-tests-dec2020
|
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
|
closed
|
radondb/radon: src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go; 13 LoC
|
fresh small test
|
Found a possible issue in [radondb/radon](https://www.github.com/radondb/radon) at [src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable s used in defer or goroutine at line 154
[Click here to see the code in its original context.](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159)
<details>
<summary>Click here to show the 13 line(s) of Go which triggered the analyzer.</summary>
```go
for _, s := range servers {
wg.Add(1)
go func() {
defer wg.Done()
var a int64
var b string
err := s.db.QueryRow("SELECT a, b FROM tbl WHERE id=1").Scan(&a, &b)
assert.Nil(t, err)
assert.Equal(t, int64(1), a)
assert.Equal(t, "hello world", b)
}()
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: e3744d066da240cb8657ac95a44373495369df46
|
1.0
|
radondb/radon: src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go; 13 LoC -
Found a possible issue in [radondb/radon](https://www.github.com/radondb/radon) at [src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159)
Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first
issue it finds, so please do not limit your consideration to the contents of the below message.
> range-loop variable s used in defer or goroutine at line 154
[Click here to see the code in its original context.](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159)
<details>
<summary>Click here to show the 13 line(s) of Go which triggered the analyzer.</summary>
```go
for _, s := range servers {
wg.Add(1)
go func() {
defer wg.Done()
var a int64
var b string
err := s.db.QueryRow("SELECT a, b FROM tbl WHERE id=1").Scan(&a, &b)
assert.Nil(t, err)
assert.Equal(t, int64(1), a)
assert.Equal(t, "hello world", b)
}()
}
```
</details>
Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket:
See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information.
commit ID: e3744d066da240cb8657ac95a44373495369df46
|
non_main
|
radondb radon src vendor github com radondb shift vendor github com siddontang go mysql server server test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable s used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for s range servers wg add go func defer wg done var a var b string err s db queryrow select a b from tbl where id scan a b assert nil t err assert equal t a assert equal t hello world b leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
| 0
|
145,418
| 13,150,421,095
|
IssuesEvent
|
2020-08-09 11:27:00
|
JuliaReach/LazySets.jl
|
https://api.github.com/repos/JuliaReach/LazySets.jl
|
opened
|
Invalid local links in documentation
|
documentation :book: fix :face_with_head_bandage:
|
```julia
┌ Warning: invalid local link: unresolved path in lib/interfaces.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "is_right_turn")
│ link.url = "utils.html#LazySets.Arrays.is_right_turn"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/LineSegment.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "plot_recipe(::LazySet{<:Real})")
│ link.url = "interfaces.html#LazySets.plot_recipe-Union{Tuple{LazySet{N}}, Tuple{N}, Tuple{LazySet{N},N}} where N<:Real"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "_isbounded_unit_dimensions")
│ link.url = "interfaces.html#LazySets._isbounded_unit_dimensions-Union{Tuple{LazySet{N}}, Tuple{N}} where N<:Real"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "_isbounded_stiemke")
│ link.url = "interfaces.html#LazySets._isbounded_stiemke-Union{Tuple{HPolyhedron{N,VN} where VN<:AbstractArray{N,1}}, Tuple{N}} where N<:Real"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "_isbounded_stiemke")
│ link.url
```
|
1.0
|
Invalid local links in documentation - ```julia
┌ Warning: invalid local link: unresolved path in lib/interfaces.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "is_right_turn")
│ link.url = "utils.html#LazySets.Arrays.is_right_turn"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/LineSegment.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "plot_recipe(::LazySet{<:Real})")
│ link.url = "interfaces.html#LazySets.plot_recipe-Union{Tuple{LazySet{N}}, Tuple{N}, Tuple{LazySet{N},N}} where N<:Real"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "_isbounded_unit_dimensions")
│ link.url = "interfaces.html#LazySets._isbounded_unit_dimensions-Union{Tuple{LazySet{N}}, Tuple{N}} where N<:Real"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "_isbounded_stiemke")
│ link.url = "interfaces.html#LazySets._isbounded_stiemke-Union{Tuple{HPolyhedron{N,VN} where VN<:AbstractArray{N,1}}, Tuple{N}} where N<:Real"
└ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725
┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md
│ link.text =
│ 1-element Array{Any,1}:
│ Markdown.Code("", "_isbounded_stiemke")
│ link.url
```
|
non_main
|
invalid local links in documentation julia ┌ warning invalid local link unresolved path in lib interfaces md │ link text │ element array any │ markdown code is right turn │ link url utils html lazysets arrays is right turn └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets linesegment md │ link text │ element array any │ markdown code plot recipe lazyset real │ link url interfaces html lazysets plot recipe union tuple lazyset n tuple n tuple lazyset n n where n real └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets hpolyhedron md │ link text │ element array any │ markdown code isbounded unit dimensions │ link url interfaces html lazysets isbounded unit dimensions union tuple lazyset n tuple n where n real └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets hpolyhedron md │ link text │ element array any │ markdown code isbounded stiemke │ link url interfaces html lazysets isbounded stiemke union tuple hpolyhedron n vn where vn abstractarray n tuple n where n real └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets hpolyhedron md │ link text │ element array any │ markdown code isbounded stiemke │ link url
| 0
|
345,122
| 30,789,683,064
|
IssuesEvent
|
2023-07-31 15:20:31
|
vedmack/yadcf
|
https://api.github.com/repos/vedmack/yadcf
|
closed
|
Android 12 / Chrome 11 / Fixed Headers
|
waiting for response / test page link
|
With fixedHeader : true and YADCF this bug happens when selecting a text input field. Once i disable fixedHeader it does not have the input bug.
**video of issue in my app**
https://drive.google.com/file/d/1DcZ2jp_WYvgcKpdqeddUO7_oX-eRKeUR/view
**demo url to see issue**
[https://www.otelconsulting.com/bug.html](https://www.otelconsulting.com/bug.html)
**demo url with latest beta 0.9.4.beta.46**
[https://www.otelconsulting.com/bug-094beta46.html](https://www.otelconsulting.com/bug-094beta46.html)
**How to replicate**
visit url and then click on the column filter input in the position column
iOS latest and Chrome do not produce the same issue.
|
1.0
|
Android 12 / Chrome 11 / Fixed Headers - With fixedHeader : true and YADCF this bug happens when selecting a text input field. Once i disable fixedHeader it does not have the input bug.
**video of issue in my app**
https://drive.google.com/file/d/1DcZ2jp_WYvgcKpdqeddUO7_oX-eRKeUR/view
**demo url to see issue**
[https://www.otelconsulting.com/bug.html](https://www.otelconsulting.com/bug.html)
**demo url with latest beta 0.9.4.beta.46**
[https://www.otelconsulting.com/bug-094beta46.html](https://www.otelconsulting.com/bug-094beta46.html)
**How to replicate**
visit url and then click on the column filter input in the position column
iOS latest and Chrome do not produce the same issue.
|
non_main
|
android chrome fixed headers with fixedheader true and yadcf this bug happens when selecting a text input field once i disable fixedheader it does not have the input bug video of issue in my app demo url to see issue demo url with latest beta beta how to replicate visit url and then click on the column filter input in the position column ios latest and chrome do not produce the same issue
| 0
|
529
| 3,925,714,914
|
IssuesEvent
|
2016-04-22 20:05:39
|
heiglandreas/authLdap
|
https://api.github.com/repos/heiglandreas/authLdap
|
closed
|
Check for already successfull authentication before "reauthenticating"
|
bug maintainer reply expected
|
This is an issue reported at https://wordpress.org/support/topic/authldap-doesnt-work-with-other-auth-plugins.
|
True
|
Check for already successfull authentication before "reauthenticating" - This is an issue reported at https://wordpress.org/support/topic/authldap-doesnt-work-with-other-auth-plugins.
|
main
|
check for already successfull authentication before reauthenticating this is an issue reported at
| 1
|
52,587
| 13,753,231,334
|
IssuesEvent
|
2020-10-06 15:23:46
|
flyingcircusio/nixpkgs
|
https://api.github.com/repos/flyingcircusio/nixpkgs
|
opened
|
Vulnerability roundup 9: bind-9.12.4-P2: 3 advisories [7.5]
|
1.severity: security
|
[search](https://search.nix.gsc.io/?q=bind&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=bind+in%3Apath&type=Code)
* [ ] [CVE-2020-8623](https://nvd.nist.gov/vuln/detail/CVE-2020-8623) CVSSv3=7.5 (nixos-19.03)
* [ ] [CVE-2020-8622](https://nvd.nist.gov/vuln/detail/CVE-2020-8622) CVSSv3=6.5 (nixos-19.03)
* [ ] [CVE-2020-8624](https://nvd.nist.gov/vuln/detail/CVE-2020-8624) CVSSv3=4.3 (nixos-19.03)
Scanned versions: nixos-19.03: 715db89e7d3.
|
True
|
Vulnerability roundup 9: bind-9.12.4-P2: 3 advisories [7.5] - [search](https://search.nix.gsc.io/?q=bind&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=bind+in%3Apath&type=Code)
* [ ] [CVE-2020-8623](https://nvd.nist.gov/vuln/detail/CVE-2020-8623) CVSSv3=7.5 (nixos-19.03)
* [ ] [CVE-2020-8622](https://nvd.nist.gov/vuln/detail/CVE-2020-8622) CVSSv3=6.5 (nixos-19.03)
* [ ] [CVE-2020-8624](https://nvd.nist.gov/vuln/detail/CVE-2020-8624) CVSSv3=4.3 (nixos-19.03)
Scanned versions: nixos-19.03: 715db89e7d3.
|
non_main
|
vulnerability roundup bind advisories nixos nixos nixos scanned versions nixos
| 0
|
264
| 3,023,990,424
|
IssuesEvent
|
2015-08-02 04:06:44
|
daemonraco/toobasic
|
https://api.github.com/repos/daemonraco/toobasic
|
closed
|
BLOB defaults error on MySQL
|
bug Database Structure Maintainer
|
## Error
What __TooBasic__ attempts to create a table using BLOB columns it failes on the default value.
|
True
|
BLOB defaults error on MySQL - ## Error
What __TooBasic__ attempts to create a table using BLOB columns it failes on the default value.
|
main
|
blob defaults error on mysql error what toobasic attempts to create a table using blob columns it failes on the default value
| 1
|
441,022
| 30,765,734,901
|
IssuesEvent
|
2023-07-30 09:19:54
|
openhab/openhab-distro
|
https://api.github.com/repos/openhab/openhab-distro
|
closed
|
Migration guide from v 2.x.x to 3.x.x Breaking changes link is broken
|
documentation
|
Official docs https://www.openhab.org/docs/configuration/migration/ guides the user to the wiki page "Breaking Changes in openHAB 3" (https://github.com/openhab/openhab-distro/wiki/Breaking-Changes-in-openHAB-3), which in turn
guides the user to another document (https://github.com/openhab/openhab-distro/releases/tag/3.0.0#breaking-changes-that-require-manual-interaction-after-the-upgrade).
It has a quick link "Breaking Changes" which refers the document that does not exist: https://github.com/openhab/openhab-distro/blob/3.0.0/breaking-changes-that-require-manual-interaction-after-the-upgrade.
As a result there is no way to see the breaking changes when migrating from v2.x.x to v3.x.x
|
1.0
|
Migration guide from v 2.x.x to 3.x.x Breaking changes link is broken - Official docs https://www.openhab.org/docs/configuration/migration/ guides the user to the wiki page "Breaking Changes in openHAB 3" (https://github.com/openhab/openhab-distro/wiki/Breaking-Changes-in-openHAB-3), which in turn
guides the user to another document (https://github.com/openhab/openhab-distro/releases/tag/3.0.0#breaking-changes-that-require-manual-interaction-after-the-upgrade).
It has a quick link "Breaking Changes" which refers the document that does not exist: https://github.com/openhab/openhab-distro/blob/3.0.0/breaking-changes-that-require-manual-interaction-after-the-upgrade.
As a result there is no way to see the breaking changes when migrating from v2.x.x to v3.x.x
|
non_main
|
migration guide from v x x to x x breaking changes link is broken official docs guides the user to the wiki page breaking changes in openhab which in turn guides the user to another document it has a quick link breaking changes which refers the document that does not exist as a result there is no way to see the breaking changes when migrating from x x to x x
| 0
|
251,865
| 18,977,314,437
|
IssuesEvent
|
2021-11-20 07:44:10
|
corona-warn-app/cwa-wishlist
|
https://api.github.com/repos/corona-warn-app/cwa-wishlist
|
closed
|
Vaccination certificate from another country
|
documentation enhancement
|
<!--
Thanks for pointing us to missing information 🙌 ❤️
Before opening a new issue, please make sure that we do not have any duplicates already open. You can ensure this by searching the issue list for this repository. If there is a duplicate, please close your issue and add a comment to the existing issue instead.
To browse existing issues by category, please see these overview issues: https://github.com/corona-warn-app/cwa-wishlist/issues/337
Specifically, please check if your suggestion has already been raised here: https://github.com/corona-warn-app/cwa-wishlist/issues/394
-->
## What is missing
<!-- Outline the information that you would like to see added. Please be rather specific (e.g., not only 'more information about', but what exactly is missing). -->
I couldn’t find if the app supports loading certificates generated in other EU countries, and if it’s possible to mix them up. For example, if one gets the first two doses in Germany and the booster in Italy or Spain.
## Why should it be included
<!-- Which aspects of the corona warn app project cannot be properly understood without this information? -->
Considering that we will probably get one dose every 6 to 12 months, more and more people will have a mix situation and would like to make sure that the app supports it.
## Where should it be included
<!-- If you think the information should be part of a specific existing document, please let us know. -->
Not sure.
|
1.0
|
Vaccination certificate from another country - <!--
Thanks for pointing us to missing information 🙌 ❤️
Before opening a new issue, please make sure that we do not have any duplicates already open. You can ensure this by searching the issue list for this repository. If there is a duplicate, please close your issue and add a comment to the existing issue instead.
To browse existing issues by category, please see these overview issues: https://github.com/corona-warn-app/cwa-wishlist/issues/337
Specifically, please check if your suggestion has already been raised here: https://github.com/corona-warn-app/cwa-wishlist/issues/394
-->
## What is missing
<!-- Outline the information that you would like to see added. Please be rather specific (e.g., not only 'more information about', but what exactly is missing). -->
I couldn’t find if the app supports loading certificates generated in other EU countries, and if it’s possible to mix them up. For example, if one gets the first two doses in Germany and the booster in Italy or Spain.
## Why should it be included
<!-- Which aspects of the corona warn app project cannot be properly understood without this information? -->
Considering that we will probably get one dose every 6 to 12 months, more and more people will have a mix situation and would like to make sure that the app supports it.
## Where should it be included
<!-- If you think the information should be part of a specific existing document, please let us know. -->
Not sure.
|
non_main
|
vaccination certificate from another country thanks for pointing us to missing information 🙌 ❤️ before opening a new issue please make sure that we do not have any duplicates already open you can ensure this by searching the issue list for this repository if there is a duplicate please close your issue and add a comment to the existing issue instead to browse existing issues by category please see these overview issues specifically please check if your suggestion has already been raised here what is missing i couldn’t find if the app supports loading certificates generated in other eu countries and if it’s possible to mix them up for example if one gets the first two doses in germany and the booster in italy or spain why should it be included considering that we will probably get one dose every to months more and more people will have a mix situation and would like to make sure that the app supports it where should it be included not sure
| 0
|
5,201
| 26,440,582,887
|
IssuesEvent
|
2023-01-15 22:49:04
|
MarcusWolschon/osmeditor4android
|
https://api.github.com/repos/MarcusWolschon/osmeditor4android
|
opened
|
Rework GPX file loading
|
Maintainability
|
A lot of the GPX file loading code has no function anymore since everything is being done in setupLayers, this should be cleaned up.
|
True
|
Rework GPX file loading - A lot of the GPX file loading code has no function anymore since everything is being done in setupLayers, this should be cleaned up.
|
main
|
rework gpx file loading a lot of the gpx file loading code has no function anymore since everything is being done in setuplayers this should be cleaned up
| 1
|
4,170
| 19,985,454,752
|
IssuesEvent
|
2022-01-30 15:41:55
|
BioArchLinux/Packages
|
https://api.github.com/repos/BioArchLinux/Packages
|
opened
|
[MAINTAIN] r-stan
|
maintain
|
<!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**Log of the bug**
<details>
```
rix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes]
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h: In instantiation of ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’:
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:300:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 1>’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:551:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 3>’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseBase.h:41:34: required from ‘class Eigen::DenseBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MatrixBase.h:48:34: required from ‘class Eigen::MatrixBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:37:34: required from ‘class Eigen::MapBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:221:34: [ skipping 20 instantiation contexts, use -ftemplate-backtrace-limit=0 to disable ]
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:251:64: required from ‘static void Eigen::internal::generic_product_impl<Lhs, Rhs, Eigen::DenseShape, Eigen::DenseShape, 6>::evalTo(Dst&, const Lhs&, const Rhs&) [with Dst = Eigen::Matrix<double, 1, 1, 0, 1, 1>; Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:124:75: required from ‘Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::product_evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes]
g++ -std=gnu++14 -shared -L/usr/lib64/R/lib -Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now -o rstan.so Module.o chains.o init.o misc.o pointer-tools.o sparse_extractors.o stan_fit_base.o stan_fit_rccp.o stanc.o stan/lang/ast_def.o stan/lang/grammars/bare_type_grammar_inst.o stan/lang/grammars/block_var_decls_grammar_inst.o stan/lang/grammars/expression07_grammar_inst.o stan/lang/grammars/expression_grammar_inst.o stan/lang/grammars/functions_grammar_inst.o stan/lang/grammars/indexes_grammar_inst.o stan/lang/grammars/local_var_decls_grammar_inst.o stan/lang/grammars/program_grammar_inst.o stan/lang/grammars/semantic_actions_def.o stan/lang/grammars/statement_2_grammar_inst.o stan/lang/grammars/statement_grammar_inst.o stan/lang/grammars/term_grammar_inst.o stan/lang/grammars/whitespace_grammar_inst.o -L/usr/lib64/R/lib -lR
ar -rs ../inst/lib//libStanServices.a stan_fit.o stan_fit_base.o
ar: creating ../inst/lib//libStanServices.a
installing to /build/r-rstan/src/00LOCK-rstan/00new/rstan/libs
** R
** inst
** byte-compile and prepare package for lazy loading
Error in as.POSIXlt.character(x, tz, ...) :
character string is not in a standard unambiguous format
Error: unable to load R code in package ‘rstan’
Execution halted
ERROR: lazy loading failed for package ‘rstan’
```
</details>
**Packages (please complete the following information):**
- Package Name: r-rstan
**Description**
https://log.bioarchlinux.org/2022-01-28T13%3A17%3A39/r-rstan.log
|
True
|
[MAINTAIN] r-stan - <!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**Log of the bug**
<details>
```
rix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes]
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h: In instantiation of ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’:
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:300:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 1>’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:551:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 3>’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseBase.h:41:34: required from ‘class Eigen::DenseBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MatrixBase.h:48:34: required from ‘class Eigen::MatrixBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:37:34: required from ‘class Eigen::MapBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:221:34: [ skipping 20 instantiation contexts, use -ftemplate-backtrace-limit=0 to disable ]
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:251:64: required from ‘static void Eigen::internal::generic_product_impl<Lhs, Rhs, Eigen::DenseShape, Eigen::DenseShape, 6>::evalTo(Dst&, const Lhs&, const Rhs&) [with Dst = Eigen::Matrix<double, 1, 1, 0, 1, 1>; Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:124:75: required from ‘Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::product_evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’
./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here
/usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes]
g++ -std=gnu++14 -shared -L/usr/lib64/R/lib -Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now -o rstan.so Module.o chains.o init.o misc.o pointer-tools.o sparse_extractors.o stan_fit_base.o stan_fit_rccp.o stanc.o stan/lang/ast_def.o stan/lang/grammars/bare_type_grammar_inst.o stan/lang/grammars/block_var_decls_grammar_inst.o stan/lang/grammars/expression07_grammar_inst.o stan/lang/grammars/expression_grammar_inst.o stan/lang/grammars/functions_grammar_inst.o stan/lang/grammars/indexes_grammar_inst.o stan/lang/grammars/local_var_decls_grammar_inst.o stan/lang/grammars/program_grammar_inst.o stan/lang/grammars/semantic_actions_def.o stan/lang/grammars/statement_2_grammar_inst.o stan/lang/grammars/statement_grammar_inst.o stan/lang/grammars/term_grammar_inst.o stan/lang/grammars/whitespace_grammar_inst.o -L/usr/lib64/R/lib -lR
ar -rs ../inst/lib//libStanServices.a stan_fit.o stan_fit_base.o
ar: creating ../inst/lib//libStanServices.a
installing to /build/r-rstan/src/00LOCK-rstan/00new/rstan/libs
** R
** inst
** byte-compile and prepare package for lazy loading
Error in as.POSIXlt.character(x, tz, ...) :
character string is not in a standard unambiguous format
Error: unable to load R code in package ‘rstan’
Execution halted
ERROR: lazy loading failed for package ‘rstan’
```
</details>
**Packages (please complete the following information):**
- Package Name: r-rstan
**Description**
https://log.bioarchlinux.org/2022-01-28T13%3A17%3A39/r-rstan.log
|
main
|
r stan please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug rix eigen matrix rhs eigen matrix int options int producttag lhsshape eigen denseshape rhsshape eigen denseshape typename eigen internal traits rhs scalar double typename eigen product rhs eigen matrix typename eigen internal traits lhs scalar double typename eigen product lhs eigen product const eigen cwisenullaryop const eigen matrix const eigen transpose eigen matrix eigen internal product evaluator producttag lhsshape rhsshape xprtype eigen product const eigen cwisenullaryop const eigen matrix const eigen transpose eigen matrix eigen matrix ’ usr lib r library rcppeigen include eigen src core productevaluators h required from ‘eigen internal evaluator evaluator const xprtype ’ usr lib r library rcppeigen include eigen src core product h required from ‘eigen internal dense product base operator const scalar const ’ stan mcmc hmc hamiltonians dense e metric hpp required from ‘double stan mcmc dense e metric t stan mcmc dense e point ’ stan mcmc hmc hamiltonians dense e metric hpp required from here usr lib r library rcppeigen include eigen src core densecoeffsbase h warning ignoring attributes on template argument ‘eigen internal packet traits type’ aka ‘ ’ usr lib r library rcppeigen include eigen src core densecoeffsbase h in instantiation of ‘class eigen densecoeffsbase true ’ usr lib r library rcppeigen include eigen src core densecoeffsbase h required from ‘class eigen densecoeffsbase true ’ usr lib r library rcppeigen include eigen src core densecoeffsbase h required from ‘class eigen densecoeffsbase true ’ usr lib r library rcppeigen include eigen src core densebase h required from ‘class eigen densebase true ’ usr lib r library rcppeigen include eigen src core matrixbase h required from ‘class eigen matrixbase true ’ usr lib r library rcppeigen include eigen src core mapbase h required from ‘class eigen mapbase true ’ usr lib r library rcppeigen include eigen src core mapbase h usr lib r library rcppeigen include eigen src core productevaluators h required from ‘static void eigen internal generic product impl evalto dst const lhs const rhs ’ usr lib r library rcppeigen include eigen src core productevaluators h required from ‘eigen internal product evaluator producttag lhsshape rhsshape product evaluator const xprtype ’ usr lib r library rcppeigen include eigen src core productevaluators h required from ‘eigen internal evaluator evaluator const xprtype ’ usr lib r library rcppeigen include eigen src core product h required from ‘eigen internal dense product base operator const scalar const ’ stan mcmc hmc hamiltonians dense e metric hpp required from ‘double stan mcmc dense e metric t stan mcmc dense e point ’ stan mcmc hmc hamiltonians dense e metric hpp required from here usr lib r library rcppeigen include eigen src core densecoeffsbase h warning ignoring attributes on template argument ‘eigen internal packet traits type’ aka ‘ ’ g std gnu shared l usr r lib wl sort common as needed z relro z now o rstan so module o chains o init o misc o pointer tools o sparse extractors o stan fit base o stan fit rccp o stanc o stan lang ast def o stan lang grammars bare type grammar inst o stan lang grammars block var decls grammar inst o stan lang grammars grammar inst o stan lang grammars expression grammar inst o stan lang grammars functions grammar inst o stan lang grammars indexes grammar inst o stan lang grammars local var decls grammar inst o stan lang grammars program grammar inst o stan lang grammars semantic actions def o stan lang grammars statement grammar inst o stan lang grammars statement grammar inst o stan lang grammars term grammar inst o stan lang grammars whitespace grammar inst o l usr r lib lr ar rs inst lib libstanservices a stan fit o stan fit base o ar creating inst lib libstanservices a installing to build r rstan src rstan rstan libs r inst byte compile and prepare package for lazy loading error in as posixlt character x tz character string is not in a standard unambiguous format error unable to load r code in package ‘rstan’ execution halted error lazy loading failed for package ‘rstan’ packages please complete the following information package name r rstan description
| 1
|
1,535
| 6,572,225,490
|
IssuesEvent
|
2017-09-11 00:17:08
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
nmcli: conn_name required
|
affects_2.1 bug_report docs_report networking waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
nmcli
##### ANSIBLE VERSION
```
# latest head of 2016-04-16
ansible 2.1.0
config file =
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
CentOS Linux release 7.2.1511 (Core)
##### SUMMARY
Conn Name Required althouth stated otherwise in documentation:
> Where conn_name will be the name used to call the connection. **when not provided a default name is generated: `<type>[-<ifname>][-<num>]`**
##### STEPS TO REPRODUCE
```
- name: Configure network adapter
nmcli: type=ethernet ip4="192.168.1.1/16" ifname="enp0s1" autoconnect=yes state=present
```
##### EXPECTED RESULTS
OK
##### ACTUAL RESULTS
```
FAILED! => {"changed": false, "failed": true, "msg": "missing required arguments: conn_name"}
```
|
True
|
nmcli: conn_name required - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
nmcli
##### ANSIBLE VERSION
```
# latest head of 2016-04-16
ansible 2.1.0
config file =
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
CentOS Linux release 7.2.1511 (Core)
##### SUMMARY
Conn Name Required althouth stated otherwise in documentation:
> Where conn_name will be the name used to call the connection. **when not provided a default name is generated: `<type>[-<ifname>][-<num>]`**
##### STEPS TO REPRODUCE
```
- name: Configure network adapter
nmcli: type=ethernet ip4="192.168.1.1/16" ifname="enp0s1" autoconnect=yes state=present
```
##### EXPECTED RESULTS
OK
##### ACTUAL RESULTS
```
FAILED! => {"changed": false, "failed": true, "msg": "missing required arguments: conn_name"}
```
|
main
|
nmcli conn name required issue type bug report component name nmcli ansible version latest head of ansible config file configured module search path default w o overrides os environment centos linux release core summary conn name required althouth stated otherwise in documentation where conn name will be the name used to call the connection when not provided a default name is generated steps to reproduce name configure network adapter nmcli type ethernet ifname autoconnect yes state present expected results ok actual results failed changed false failed true msg missing required arguments conn name
| 1
|
2,925
| 10,427,223,538
|
IssuesEvent
|
2019-09-16 19:25:38
|
NaluKit/nalu
|
https://api.github.com/repos/NaluKit/nalu
|
closed
|
Remove errorRoute implementation
|
maintainance
|
With the new upcoming error event feature, it gets clearer, that supporting both features will end up in a lot of more code. Some methods gets inconsistent and make the framework harder to use.
Due to this fact, the errorRoute implementation gets removed.
The error implementation of Nalu v1.0.0 can be easily reused. There will be an example how to do this.
|
True
|
Remove errorRoute implementation - With the new upcoming error event feature, it gets clearer, that supporting both features will end up in a lot of more code. Some methods gets inconsistent and make the framework harder to use.
Due to this fact, the errorRoute implementation gets removed.
The error implementation of Nalu v1.0.0 can be easily reused. There will be an example how to do this.
|
main
|
remove errorroute implementation with the new upcoming error event feature it gets clearer that supporting both features will end up in a lot of more code some methods gets inconsistent and make the framework harder to use due to this fact the errorroute implementation gets removed the error implementation of nalu can be easily reused there will be an example how to do this
| 1
|
94,561
| 10,831,668,028
|
IssuesEvent
|
2019-11-11 08:53:25
|
p-amonpitakpun/clock-drawing-dnn
|
https://api.github.com/repos/p-amonpitakpun/clock-drawing-dnn
|
closed
|
Metric Result Exporting.
|
documentation enhancement question
|
How can we export the result of the classifier so that we can use in the presentation and document (such as putting it in README) ?
|
1.0
|
Metric Result Exporting. - How can we export the result of the classifier so that we can use in the presentation and document (such as putting it in README) ?
|
non_main
|
metric result exporting how can we export the result of the classifier so that we can use in the presentation and document such as putting it in readme
| 0
|
550,703
| 16,130,585,677
|
IssuesEvent
|
2021-04-29 03:39:24
|
rainlab/translate-plugin
|
https://api.github.com/repos/rainlab/translate-plugin
|
closed
|
Back-end locale switcher single vs. multi field default behavior
|
Priority: Low Type: Discussion
|
There's been some question in our shop about whether from a UX perspective the back-end locale switcher should default to switching all translatable fields (the current CTRL/CMD click behavior), with the CTRL/CMD click being used to change only one field. I wanted to pose this question to see if there's agreement on that front, or if we'd need to customize if we decide that's the behavior we want to offer our clients.
|
1.0
|
Back-end locale switcher single vs. multi field default behavior - There's been some question in our shop about whether from a UX perspective the back-end locale switcher should default to switching all translatable fields (the current CTRL/CMD click behavior), with the CTRL/CMD click being used to change only one field. I wanted to pose this question to see if there's agreement on that front, or if we'd need to customize if we decide that's the behavior we want to offer our clients.
|
non_main
|
back end locale switcher single vs multi field default behavior there s been some question in our shop about whether from a ux perspective the back end locale switcher should default to switching all translatable fields the current ctrl cmd click behavior with the ctrl cmd click being used to change only one field i wanted to pose this question to see if there s agreement on that front or if we d need to customize if we decide that s the behavior we want to offer our clients
| 0
|
570
| 4,047,242,928
|
IssuesEvent
|
2016-05-23 03:35:00
|
duckduckgo/zeroclickinfo-spice
|
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
|
closed
|
Rust Cargo Packages: should accept the term "crate"
|
Maintainer Input Requested Suggestion Triggering
|
The unit of compilation in Rust is a crate, and packages can only contain a single library crate, so when searching, people are frequently going to want to search for a "rust crate" or "cargo crate". You should be able to use "cargo crate time" in the same way you would use "cargo package time".
------
IA Page: http://duck.co/ia/view/rust_cargo
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @TomBebbington
|
True
|
Rust Cargo Packages: should accept the term "crate" - The unit of compilation in Rust is a crate, and packages can only contain a single library crate, so when searching, people are frequently going to want to search for a "rust crate" or "cargo crate". You should be able to use "cargo crate time" in the same way you would use "cargo package time".
------
IA Page: http://duck.co/ia/view/rust_cargo
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @TomBebbington
|
main
|
rust cargo packages should accept the term crate the unit of compilation in rust is a crate and packages can only contain a single library crate so when searching people are frequently going to want to search for a rust crate or cargo crate you should be able to use cargo crate time in the same way you would use cargo package time ia page tombebbington
| 1
|
403
| 3,460,612,650
|
IssuesEvent
|
2015-12-19 09:33:04
|
espeak-ng/espeak-ng
|
https://api.github.com/repos/espeak-ng/espeak-ng
|
closed
|
Remove redundant `end of ...` comments
|
maintainability resolved/fixed
|
Various functions inconsistently have `{//===...` at the start and `} // end of ...` comments. These make the code harder to read and should be removed.
|
True
|
Remove redundant `end of ...` comments - Various functions inconsistently have `{//===...` at the start and `} // end of ...` comments. These make the code harder to read and should be removed.
|
main
|
remove redundant end of comments various functions inconsistently have at the start and end of comments these make the code harder to read and should be removed
| 1
|
1,370
| 5,933,073,146
|
IssuesEvent
|
2017-05-24 10:59:41
|
aroberge/reeborg
|
https://api.github.com/repos/aroberge/reeborg
|
closed
|
Remove RUR.CURRENT_WORLD almost everywhere
|
easier to maintain
|
The world state is put in a global variable named `RUR.CURRENT_WORLD`. If possible, it should be removed from almost all functions, except when it needs to be created. Elsewhere, an exported function should be used to access it.
|
True
|
Remove RUR.CURRENT_WORLD almost everywhere - The world state is put in a global variable named `RUR.CURRENT_WORLD`. If possible, it should be removed from almost all functions, except when it needs to be created. Elsewhere, an exported function should be used to access it.
|
main
|
remove rur current world almost everywhere the world state is put in a global variable named rur current world if possible it should be removed from almost all functions except when it needs to be created elsewhere an exported function should be used to access it
| 1
|
20,469
| 15,571,286,835
|
IssuesEvent
|
2021-03-17 04:34:08
|
SanderMertens/flecs
|
https://api.github.com/repos/SanderMertens/flecs
|
opened
|
Update API & query parser to new entity-relation terminology/notation
|
enhancement usability
|
**Describe the problem you are trying to solve.**
With the upcoming new query implementation (#295) the primary use case for traits will be as a way to store relationships between entities. The terminology used by the current API does not make this obvious or intuitive.
Additionally, there are a few inconsistencies between the C and C++ API that need to be addressed (the C API uses the reverse order of specifying traits as the C++ API).
**Describe the solution you'd like**
Modify the API to use terminology as described here:
https://github.com/SanderMertens/flecs/discussions/344
|
True
|
Update API & query parser to new entity-relation terminology/notation - **Describe the problem you are trying to solve.**
With the upcoming new query implementation (#295) the primary use case for traits will be as a way to store relationships between entities. The terminology used by the current API does not make this obvious or intuitive.
Additionally, there are a few inconsistencies between the C and C++ API that need to be addressed (the C API uses the reverse order of specifying traits as the C++ API).
**Describe the solution you'd like**
Modify the API to use terminology as described here:
https://github.com/SanderMertens/flecs/discussions/344
|
non_main
|
update api query parser to new entity relation terminology notation describe the problem you are trying to solve with the upcoming new query implementation the primary use case for traits will be as a way to store relationships between entities the terminology used by the current api does not make this obvious or intuitive additionally there are a few inconsistencies between the c and c api that need to be addressed the c api uses the reverse order of specifying traits as the c api describe the solution you d like modify the api to use terminology as described here
| 0
|
3,835
| 16,686,917,247
|
IssuesEvent
|
2021-06-08 08:59:00
|
keptn/community
|
https://api.github.com/repos/keptn/community
|
closed
|
REQUEST: Maintainer membership for @grabnerandi @AloisReitbauer @danielkhan
|
membership:maintainer status:approved
|
### Multi-request for Keptn maintainer team
Please note, that these people are already maintainers. This issue is created for bookkeeping purposes.
* Andreas Grabner @grabnerandi
* Alois Reitbauer @AloisReitbauer
* Daniel Khan @danielkhan
### Requirements
- [x] We have reviewed the community membership guidelines (https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md)
- [x] We have enabled 2FA on my GitHub account. See https://github.com/settings/security
- [x] We have subscribed to the [Keptn Slack channel](http://slack.keptn.sh/)
- [x] We are actively contributing to 1 or more Keptn subprojects in terms of guiding the direction of the project
- [x] We have two sponsors that meet the sponsor requirements listed in the community membership guidelines. Among other requirements, sponsors must be approvers or maintainers of at least one repository in the organization
- [x] I have spoken to my sponsors ahead of this application, and they have agreed to sponsor my application
### Sponsors
<!-- Replace (at) with the `@` sign -->
- @christian-kreuzberger-dtx
- @johannes-b
Each sponsor should reply to this issue with the comment "*I support*".
Please remember, it is an applicant's responsibility to get their sponsors' confirmation before submitting the request.
### List of contributions to the Keptn project
The aforementioned members have been actively contributing to the Keptn project for several months and are driving the direction of the project. Their contributions range from design decisions, technical guidance, roadmap, and even code contributions to the project.
|
True
|
REQUEST: Maintainer membership for @grabnerandi @AloisReitbauer @danielkhan - ### Multi-request for Keptn maintainer team
Please note, that these people are already maintainers. This issue is created for bookkeeping purposes.
* Andreas Grabner @grabnerandi
* Alois Reitbauer @AloisReitbauer
* Daniel Khan @danielkhan
### Requirements
- [x] We have reviewed the community membership guidelines (https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md)
- [x] We have enabled 2FA on my GitHub account. See https://github.com/settings/security
- [x] We have subscribed to the [Keptn Slack channel](http://slack.keptn.sh/)
- [x] We are actively contributing to 1 or more Keptn subprojects in terms of guiding the direction of the project
- [x] We have two sponsors that meet the sponsor requirements listed in the community membership guidelines. Among other requirements, sponsors must be approvers or maintainers of at least one repository in the organization
- [x] I have spoken to my sponsors ahead of this application, and they have agreed to sponsor my application
### Sponsors
<!-- Replace (at) with the `@` sign -->
- @christian-kreuzberger-dtx
- @johannes-b
Each sponsor should reply to this issue with the comment "*I support*".
Please remember, it is an applicant's responsibility to get their sponsors' confirmation before submitting the request.
### List of contributions to the Keptn project
The aforementioned members have been actively contributing to the Keptn project for several months and are driving the direction of the project. Their contributions range from design decisions, technical guidance, roadmap, and even code contributions to the project.
|
main
|
request maintainer membership for grabnerandi aloisreitbauer danielkhan multi request for keptn maintainer team please note that these people are already maintainers this issue is created for bookkeeping purposes andreas grabner grabnerandi alois reitbauer aloisreitbauer daniel khan danielkhan requirements we have reviewed the community membership guidelines we have enabled on my github account see we have subscribed to the we are actively contributing to or more keptn subprojects in terms of guiding the direction of the project we have two sponsors that meet the sponsor requirements listed in the community membership guidelines among other requirements sponsors must be approvers or maintainers of at least one repository in the organization i have spoken to my sponsors ahead of this application and they have agreed to sponsor my application sponsors christian kreuzberger dtx johannes b each sponsor should reply to this issue with the comment i support please remember it is an applicant s responsibility to get their sponsors confirmation before submitting the request list of contributions to the keptn project the aforementioned members have been actively contributing to the keptn project for several months and are driving the direction of the project their contributions range from design decisions technical guidance roadmap and even code contributions to the project
| 1
|
353,817
| 25,137,220,389
|
IssuesEvent
|
2022-11-09 19:39:34
|
Merck/gsDesign2
|
https://api.github.com/repos/Merck/gsDesign2
|
closed
|
Remove `simtrial` dependency
|
documentation
|
Could we include required functions from `simtrial` to make `gsDesign2` self-contained?
cc: @nanxstats
|
1.0
|
Remove `simtrial` dependency - Could we include required functions from `simtrial` to make `gsDesign2` self-contained?
cc: @nanxstats
|
non_main
|
remove simtrial dependency could we include required functions from simtrial to make self contained cc nanxstats
| 0
|
11,986
| 3,247,994,598
|
IssuesEvent
|
2015-10-17 00:10:07
|
realm/realm-cocoa
|
https://api.github.com/repos/realm/realm-cocoa
|
closed
|
RealmTests.testCompact fails on iOS 7 device
|
bug P2 tests
|
This happens consistently with the 64-bit iOS 7 device I just added to CI: https://ci.realm.io/job/objc_pr/2112/configuration=Release,swift_version=2.0,target=ios-device-objc/testReport/junit/(root)/RealmTests/testCompact/
Maybe the unit test assumes more than it should about the behavior of compaction?
We have 3 options:
1. fix the code/test
2. disable the test for iOS 7 devices
3. remove the iOS 7 device from CI
Option 1 is of course ideal, but if no one will be able to dig into this seriously for a bit, I propose option 2.
|
1.0
|
RealmTests.testCompact fails on iOS 7 device - This happens consistently with the 64-bit iOS 7 device I just added to CI: https://ci.realm.io/job/objc_pr/2112/configuration=Release,swift_version=2.0,target=ios-device-objc/testReport/junit/(root)/RealmTests/testCompact/
Maybe the unit test assumes more than it should about the behavior of compaction?
We have 3 options:
1. fix the code/test
2. disable the test for iOS 7 devices
3. remove the iOS 7 device from CI
Option 1 is of course ideal, but if no one will be able to dig into this seriously for a bit, I propose option 2.
|
non_main
|
realmtests testcompact fails on ios device this happens consistently with the bit ios device i just added to ci maybe the unit test assumes more than it should about the behavior of compaction we have options fix the code test disable the test for ios devices remove the ios device from ci option is of course ideal but if no one will be able to dig into this seriously for a bit i propose option
| 0
|
66,739
| 3,257,697,271
|
IssuesEvent
|
2015-10-20 18:58:29
|
openshift/origin
|
https://api.github.com/repos/openshift/origin
|
closed
|
oc deploy should allow the user to view the logs of the latest deployment
|
component/deployments kind/enhancement priority/P2
|
`oc deploy --logs` should show the logs of the current deployment (if available). It maybe should support looking at older logs.
|
1.0
|
oc deploy should allow the user to view the logs of the latest deployment - `oc deploy --logs` should show the logs of the current deployment (if available). It maybe should support looking at older logs.
|
non_main
|
oc deploy should allow the user to view the logs of the latest deployment oc deploy logs should show the logs of the current deployment if available it maybe should support looking at older logs
| 0
|
672,456
| 22,826,905,249
|
IssuesEvent
|
2022-07-12 09:23:02
|
bedita/manager
|
https://api.github.com/repos/bedita/manager
|
closed
|
Data Modeling Object Type modify
|
feature Priority - Normal UI / UX
|
Provide the following enhancements:
- [x] Single custom property Hide/Show button should populate `Hidden` field properly (https://github.com/bedita/manager/pull/810)
- [x] Single custom property Delete button should make an ajax call to api delete proper endpoint (https://github.com/bedita/manager/pull/812)
- [x] Parent name should be an abstract objects only select combo (https://github.com/bedita/manager/pull/815)
- [x] Table should be a select combo (static list for now, result of an api call in the future) (https://github.com/bedita/manager/pull/815)
- [x] Associations should be a list of checkboxes (static list for now: DateRanges, Streams, Categories, Tags; there could be an extra value from api: show it as readonly checked) (https://github.com/bedita/manager/pull/816)
- [x] Single custom property Type should be a property_types select combo (https://github.com/bedita/manager/pull/816)
|
1.0
|
Data Modeling Object Type modify - Provide the following enhancements:
- [x] Single custom property Hide/Show button should populate `Hidden` field properly (https://github.com/bedita/manager/pull/810)
- [x] Single custom property Delete button should make an ajax call to api delete proper endpoint (https://github.com/bedita/manager/pull/812)
- [x] Parent name should be an abstract objects only select combo (https://github.com/bedita/manager/pull/815)
- [x] Table should be a select combo (static list for now, result of an api call in the future) (https://github.com/bedita/manager/pull/815)
- [x] Associations should be a list of checkboxes (static list for now: DateRanges, Streams, Categories, Tags; there could be an extra value from api: show it as readonly checked) (https://github.com/bedita/manager/pull/816)
- [x] Single custom property Type should be a property_types select combo (https://github.com/bedita/manager/pull/816)
|
non_main
|
data modeling object type modify provide the following enhancements single custom property hide show button should populate hidden field properly single custom property delete button should make an ajax call to api delete proper endpoint parent name should be an abstract objects only select combo table should be a select combo static list for now result of an api call in the future associations should be a list of checkboxes static list for now dateranges streams categories tags there could be an extra value from api show it as readonly checked single custom property type should be a property types select combo
| 0
|
714,562
| 24,566,453,789
|
IssuesEvent
|
2022-10-13 03:49:17
|
AY2223S1-CS2113-T17-1/tp
|
https://api.github.com/repos/AY2223S1-CS2113-T17-1/tp
|
closed
|
[List] As an AOM, I can view the details of a passenger
|
type.Story priority.High
|
so that I am able to have an overview of the passenger list in terminal 1.
Vignesh: Class creation (With accompanying methods)
Ivan: implement in main class
Due Date: 11th Oct 2022 (Tuesday)
|
1.0
|
[List] As an AOM, I can view the details of a passenger - so that I am able to have an overview of the passenger list in terminal 1.
Vignesh: Class creation (With accompanying methods)
Ivan: implement in main class
Due Date: 11th Oct 2022 (Tuesday)
|
non_main
|
as an aom i can view the details of a passenger so that i am able to have an overview of the passenger list in terminal vignesh class creation with accompanying methods ivan implement in main class due date oct tuesday
| 0
|
149,010
| 13,253,535,284
|
IssuesEvent
|
2020-08-20 07:46:43
|
nolar/kopf
|
https://api.github.com/repos/nolar/kopf
|
opened
|
Fork Kopf
|
documentation
|
Kopf has a new GitHub home:
* https://github.com/nolar/kopf.
Announcement (a link to go around paywall):
* https://medium.com/@nolar/kopf-is-forked-cdca40026ea7?source=friends_link&sk=d476cc32ec728382d664506d2cf08b69
|
1.0
|
Fork Kopf - Kopf has a new GitHub home:
* https://github.com/nolar/kopf.
Announcement (a link to go around paywall):
* https://medium.com/@nolar/kopf-is-forked-cdca40026ea7?source=friends_link&sk=d476cc32ec728382d664506d2cf08b69
|
non_main
|
fork kopf kopf has a new github home announcement a link to go around paywall
| 0
|
2,418
| 8,580,906,017
|
IssuesEvent
|
2018-11-13 13:22:34
|
ansible/ansible
|
https://api.github.com/repos/ansible/ansible
|
closed
|
ec2_ami_copy: When using tag_equality with wait, error occurs
|
affects_2.8 aws bug cloud has_pr module needs_info needs_maintainer support:community traceback
|
<!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
When using the ec2_ami_copy module with the tag_equality and wait parameters set to true, if the AMI exists, an error occurs. Unhandled exception. (local variable 'image_id' referenced before assignment)
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below -->
ec2_ami_copy.py
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
ansible 2.7.0
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/etc/ansible/custom_modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.5 (default, May 31 2018, 09:41:32) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)]
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
When tag_equality and wait parameters are set and AMI exists, the problem presents itself.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
ec2_ami_copy_pdx:
source_region: "us-east-1"
source_image_id: "ami-0a65asfd4"
name: TestAMI
tags:
Name: "TestAMI"
tag_equality: yes
wait: yes
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
The ec2_ami_copy module should not create a new AMI and should recognize that one exists.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
An error message occurs when running the module.
<!--- Paste verbatim command output between quotes -->
```paste below
The full traceback is:
WARNING: The below traceback may *not* be related to the actual failure.
File "/tmp/ansible_ec2_ami_copy_payload_bJvVsP/__main__.py", line 193, in copy_image
ImageIds=[image_id],
fatal: [localhost]: FAILED! => changed=false
invocation:
module_args:
aws_access_key: null
aws_secret_key: null
description: null
ec2_url: null
encrypted: false
name: TestAMI
profile: null
region: null
security_token: null
source_image_id: ami-0a65asfd4
source_region: us-east-1
tag_equality: true
tags:
Name: TestAMI
validate_certs: true
wait: true
wait_timeout: 600
msg: Unhandled exception. (local variable 'image_id' referenced before assignment)
```
|
True
|
ec2_ami_copy: When using tag_equality with wait, error occurs - <!--- Verify first that your issue is not already reported on GitHub -->
<!--- Also test if the latest release and devel branch are affected too -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Explain the problem briefly below -->
When using the ec2_ami_copy module with the tag_equality and wait parameters set to true, if the AMI exists, an error occurs. Unhandled exception. (local variable 'image_id' referenced before assignment)
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below -->
ec2_ami_copy.py
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
ansible 2.7.0
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/etc/ansible/custom_modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible
python version = 2.7.5 (default, May 31 2018, 09:41:32) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)]
```
##### OS / ENVIRONMENT
<!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. -->
##### STEPS TO REPRODUCE
<!--- Describe exactly how to reproduce the problem, using a minimal test-case -->
When tag_equality and wait parameters are set and AMI exists, the problem presents itself.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
ec2_ami_copy_pdx:
source_region: "us-east-1"
source_image_id: "ami-0a65asfd4"
name: TestAMI
tags:
Name: "TestAMI"
tag_equality: yes
wait: yes
```
<!--- HINT: You can paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- Describe what you expected to happen when running the steps above -->
The ec2_ami_copy module should not create a new AMI and should recognize that one exists.
##### ACTUAL RESULTS
<!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) -->
An error message occurs when running the module.
<!--- Paste verbatim command output between quotes -->
```paste below
The full traceback is:
WARNING: The below traceback may *not* be related to the actual failure.
File "/tmp/ansible_ec2_ami_copy_payload_bJvVsP/__main__.py", line 193, in copy_image
ImageIds=[image_id],
fatal: [localhost]: FAILED! => changed=false
invocation:
module_args:
aws_access_key: null
aws_secret_key: null
description: null
ec2_url: null
encrypted: false
name: TestAMI
profile: null
region: null
security_token: null
source_image_id: ami-0a65asfd4
source_region: us-east-1
tag_equality: true
tags:
Name: TestAMI
validate_certs: true
wait: true
wait_timeout: 600
msg: Unhandled exception. (local variable 'image_id' referenced before assignment)
```
|
main
|
ami copy when using tag equality with wait error occurs summary when using the ami copy module with the tag equality and wait parameters set to true if the ami exists an error occurs unhandled exception local variable image id referenced before assignment issue type bug report component name ami copy py ansible version paste below configuration paste below ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible python version default may os environment steps to reproduce when tag equality and wait parameters are set and ami exists the problem presents itself yaml ami copy pdx source region us east source image id ami name testami tags name testami tag equality yes wait yes expected results the ami copy module should not create a new ami and should recognize that one exists actual results an error message occurs when running the module paste below the full traceback is warning the below traceback may not be related to the actual failure file tmp ansible ami copy payload bjvvsp main py line in copy image imageids fatal failed changed false invocation module args aws access key null aws secret key null description null url null encrypted false name testami profile null region null security token null source image id ami source region us east tag equality true tags name testami validate certs true wait true wait timeout msg unhandled exception local variable image id referenced before assignment
| 1
|
56,111
| 13,757,517,391
|
IssuesEvent
|
2020-10-06 21:48:37
|
sitmun/sitmun-backend-core
|
https://api.github.com/repos/sitmun/sitmun-backend-core
|
closed
|
Fix generation of API documentation
|
bug build
|
Generation of documentation does not work as expected. Swagger.json file is generated at the root of the project rather than at the build folder. In addition tests are run twice in Travis-CI because `convertSwagger2markup` always depends on tests.
|
1.0
|
Fix generation of API documentation - Generation of documentation does not work as expected. Swagger.json file is generated at the root of the project rather than at the build folder. In addition tests are run twice in Travis-CI because `convertSwagger2markup` always depends on tests.
|
non_main
|
fix generation of api documentation generation of documentation does not work as expected swagger json file is generated at the root of the project rather than at the build folder in addition tests are run twice in travis ci because always depends on tests
| 0
|
924
| 4,627,883,043
|
IssuesEvent
|
2016-09-28 00:59:40
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
junos_command: ValueError: need more than 1 value to unpack
|
affects_2.2 bug_report in progress networking P1 waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
junos_command
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel eb33ed4219) last updated 2016/09/27 09:18:44 (GMT +100)
lib/ansible/modules/core: (devel c03697c81e) last updated 2016/09/27 09:18:49 (GMT +100)
lib/ansible/modules/extras: (devel 119bc466be) last updated 2016/09/27 09:18:51 (GMT +100)
```
plus the following patch
```
diff --git a/lib/ansible/module_utils/junos.py b/lib/ansible/module_utils/junos.py
index 9cf36da..d3e3cb2 100644
--- a/lib/ansible/module_utils/junos.py
+++ b/lib/ansible/module_utils/junos.py
@@ -171,7 +171,7 @@ class Netconf(object):
return ele
def load_config(self, config, commit=False, replace=False, confirm=None,
- comment=None, config_format='text'):
+ comment=None, config_format='text', **kwargs):
if replace:
merge = False
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
Issue may need fixing in `module_utils/netcli.py`, though I'm triggering it from `junos_command`, so raising it in ansible-modules-core.
##### STEPS TO REPRODUCE
```yaml
- name: test gt operator
junos_command:
commands:
- show version
- show interfaces fxp0
format: xml
wait_for:
- "stdout[1]['interface-information']['physical-interface'].mtu[0]"
provider: "{{ netconf }}"
register: result
```
##### EXPECTED RESULTS
##### ACTUAL RESULTS
```yaml
TASK [test_junos_command : test gt operator] ***********************************
task path: /home/johnb/git/ansible-inc/test-junos/roles/test_junos_command/tests/netconf_xml/greaterthan.yaml:4
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/junos/junos_command.py
<vsrx01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<vsrx01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" && echo ansible-tmp-1475001874.51-177586194093154="` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" ) && sleep 0'
<vsrx01> PUT /tmp/tmpbDvPHR TO /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py
<vsrx01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/ /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py && sleep 0'
<vsrx01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 304, in <module>
main()
File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 269, in main
runner.add_conditional(item)
File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 153, in add_conditional
File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 195, in __init__
ValueError: need more than 1 value to unpack
fatal: [vsrx01]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "junos_command"
},
"module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 304, in <module>\n main()\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 269, in main\n runner.add_conditional(item)\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 153, in add_conditional\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 195, in __init__\nValueError: need more than 1 value to unpack\n",
"module_stdout": "",
"msg": "MODULE FAILURE"
}
```
|
True
|
junos_command: ValueError: need more than 1 value to unpack - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
junos_command
##### ANSIBLE VERSION
```
ansible 2.2.0 (devel eb33ed4219) last updated 2016/09/27 09:18:44 (GMT +100)
lib/ansible/modules/core: (devel c03697c81e) last updated 2016/09/27 09:18:49 (GMT +100)
lib/ansible/modules/extras: (devel 119bc466be) last updated 2016/09/27 09:18:51 (GMT +100)
```
plus the following patch
```
diff --git a/lib/ansible/module_utils/junos.py b/lib/ansible/module_utils/junos.py
index 9cf36da..d3e3cb2 100644
--- a/lib/ansible/module_utils/junos.py
+++ b/lib/ansible/module_utils/junos.py
@@ -171,7 +171,7 @@ class Netconf(object):
return ele
def load_config(self, config, commit=False, replace=False, confirm=None,
- comment=None, config_format='text'):
+ comment=None, config_format='text', **kwargs):
if replace:
merge = False
```
##### CONFIGURATION
##### OS / ENVIRONMENT
##### SUMMARY
Issue may need fixing in `module_utils/netcli.py`, though I'm triggering it from `junos_command`, so raising it in ansible-modules-core.
##### STEPS TO REPRODUCE
```yaml
- name: test gt operator
junos_command:
commands:
- show version
- show interfaces fxp0
format: xml
wait_for:
- "stdout[1]['interface-information']['physical-interface'].mtu[0]"
provider: "{{ netconf }}"
register: result
```
##### EXPECTED RESULTS
##### ACTUAL RESULTS
```yaml
TASK [test_junos_command : test gt operator] ***********************************
task path: /home/johnb/git/ansible-inc/test-junos/roles/test_junos_command/tests/netconf_xml/greaterthan.yaml:4
Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/junos/junos_command.py
<vsrx01> ESTABLISH LOCAL CONNECTION FOR USER: johnb
<vsrx01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" && echo ansible-tmp-1475001874.51-177586194093154="` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" ) && sleep 0'
<vsrx01> PUT /tmp/tmpbDvPHR TO /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py
<vsrx01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/ /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py && sleep 0'
<vsrx01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 304, in <module>
main()
File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 269, in main
runner.add_conditional(item)
File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 153, in add_conditional
File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 195, in __init__
ValueError: need more than 1 value to unpack
fatal: [vsrx01]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "junos_command"
},
"module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 304, in <module>\n main()\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 269, in main\n runner.add_conditional(item)\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 153, in add_conditional\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 195, in __init__\nValueError: need more than 1 value to unpack\n",
"module_stdout": "",
"msg": "MODULE FAILURE"
}
```
|
main
|
junos command valueerror need more than value to unpack issue type bug report component name junos command ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt plus the following patch diff git a lib ansible module utils junos py b lib ansible module utils junos py index a lib ansible module utils junos py b lib ansible module utils junos py class netconf object return ele def load config self config commit false replace false confirm none comment none config format text comment none config format text kwargs if replace merge false configuration os environment summary issue may need fixing in module utils netcli py though i m triggering it from junos command so raising it in ansible modules core steps to reproduce yaml name test gt operator junos command commands show version show interfaces format xml wait for stdout mtu provider netconf register result expected results actual results yaml task task path home johnb git ansible inc test junos roles test junos command tests netconf xml greaterthan yaml using module file home johnb git ansible inc ansible lib ansible modules core network junos junos command py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpbdvphr to home johnb ansible tmp ansible tmp junos command py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp junos command py sleep exec bin sh c python home johnb ansible tmp ansible tmp junos command py rm rf home johnb ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible ansible module junos command py line in main file tmp ansible ansible module junos command py line in main runner add conditional item file tmp ansible ansible modlib zip ansible module utils netcli py line in add conditional file tmp ansible ansible modlib zip ansible module utils netcli py line in init valueerror need more than value to unpack fatal failed changed false failed true invocation module name junos command module stderr traceback most recent call last n file tmp ansible ansible module junos command py line in n main n file tmp ansible ansible module junos command py line in main n runner add conditional item n file tmp ansible ansible modlib zip ansible module utils netcli py line in add conditional n file tmp ansible ansible modlib zip ansible module utils netcli py line in init nvalueerror need more than value to unpack n module stdout msg module failure
| 1
|
3,342
| 12,962,193,174
|
IssuesEvent
|
2020-07-20 16:46:08
|
short-d/short
|
https://api.github.com/repos/short-d/short
|
closed
|
[Refactor] Update user short link relation in Usecase
|
maintainability
|
**What is frustrating you?**
When updating the alias of a short link, the relation of user short link is updated only because they are joint in postgres database but it is not updated in updater usecase
https://github.com/short-d/short/blob/b945aac5192b0f4bb674fd6dfdbdb7d7b1f05fd3/backend/app/usecase/shortlink/updater.go#L80-L81
**Your solution**
Update user short link relation in Usecase
|
True
|
[Refactor] Update user short link relation in Usecase - **What is frustrating you?**
When updating the alias of a short link, the relation of user short link is updated only because they are joint in postgres database but it is not updated in updater usecase
https://github.com/short-d/short/blob/b945aac5192b0f4bb674fd6dfdbdb7d7b1f05fd3/backend/app/usecase/shortlink/updater.go#L80-L81
**Your solution**
Update user short link relation in Usecase
|
main
|
update user short link relation in usecase what is frustrating you when updating the alias of a short link the relation of user short link is updated only because they are joint in postgres database but it is not updated in updater usecase your solution update user short link relation in usecase
| 1
|
4,096
| 19,323,198,690
|
IssuesEvent
|
2021-12-14 08:37:45
|
WarenGonzaga/daisy.js
|
https://api.github.com/repos/WarenGonzaga/daisy.js
|
opened
|
move site to docs folder
|
chore maintainers only tweak
|
It is more reasonable to move the demo site to the docs folder instead. 👌
|
True
|
move site to docs folder - It is more reasonable to move the demo site to the docs folder instead. 👌
|
main
|
move site to docs folder it is more reasonable to move the demo site to the docs folder instead 👌
| 1
|
5,613
| 28,095,593,858
|
IssuesEvent
|
2023-03-30 15:34:57
|
MozillaFoundation/foundation.mozilla.org
|
https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org
|
opened
|
Implement security relevant upgrades
|
engineering maintain needs grooming security
|
## Description
See this list of security relevant upgraded: https://github.com/MozillaFoundation/foundation.mozilla.org/security/dependabot
## Acceptance criteria
- [ ] All security relevant upgrades have been implemented.
|
True
|
Implement security relevant upgrades - ## Description
See this list of security relevant upgraded: https://github.com/MozillaFoundation/foundation.mozilla.org/security/dependabot
## Acceptance criteria
- [ ] All security relevant upgrades have been implemented.
|
main
|
implement security relevant upgrades description see this list of security relevant upgraded acceptance criteria all security relevant upgrades have been implemented
| 1
|
1,477
| 6,404,174,472
|
IssuesEvent
|
2017-08-07 01:23:43
|
caskroom/homebrew-cask
|
https://api.github.com/repos/caskroom/homebrew-cask
|
closed
|
microsoft-office uninstall does not remove .app files from /Applications
|
awaiting maintainer feedback
|
#### General troubleshooting steps
- [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue.
- [X] None of the templates was appropriate for my issue, or I’m not sure.
- [X] I ran `brew update-reset && brew update` and retried my command.
- [X] I ran `brew doctor`, fixed as many issues as possible and retried my command.
- [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md).
#### Description of issue
`brew cask zap microsoft-office` should get rid of the apps, but the apps (microsoft-office is a suite) remain on /Applications/.
#### Output of your command with `--verbose --debug`
```
[I] gtklocker@schwarz ~> brew cask zap microsoft-office --verbose --debug
==> Zapping Cask microsoft-office
==> Implied "brew cask uninstall microsoft-office"
==> Un-installing artifacts
==> Determining which artifacts are present in Cask microsoft-office
==> 3 artifact/s defined
#<Hbc::Artifact::Uninstall:0x007fe6ad109b90>
#<Hbc::Artifact::Pkg:0x007fe6ad109758>
#<Hbc::Artifact::Zap:0x007fe6ad109460>
==> Un-installing artifact of class Hbc::Artifact::Uninstall
==> Running uninstall process for microsoft-office; your password may be necessary
==> Removing launchctl service com.microsoft.autoupdate.helpertool
==> Executing: ["/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
Password:
==> Removing launchctl service com.microsoft.office.licensing.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Removing launchctl service com.microsoft.office.licensingV2.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Uninstalling packages:
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.package.*"]
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.pkg.licensing"]
==> Dispatching zap stanza
==> Running zap process for microsoft-office; your password may be necessary
==> Removing files:
~/Library/Application Scripts/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.excel.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.powerpoint.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.word.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate.fba.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate2.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.Office
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.OfficeOsfWebHost
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.ms
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Excel.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Powerpoint.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Word.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate.fba.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate2.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.autoupdate2.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.office.setupassistant.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
==> Removing directories if empty:
~/Library/Caches/Microsoft/uls
~/Library/Caches/Microsoft
==> Removing all staged versions of Cask 'microsoft-office'
==> Purging all staged versions of Cask microsoft-office
[I] gtklocker@schwarz ~> ls -d1 /Applications/Microsoft\ *
/Applications/Microsoft Excel.app
/Applications/Microsoft OneNote.app
/Applications/Microsoft Outlook.app
/Applications/Microsoft PowerPoint.app
/Applications/Microsoft Word.app
```
#### Output of `brew cask doctor`
```
[I] gtklocker@schwarz ~> brew cask doctor
==> Homebrew-Cask Version
Homebrew-Cask 1.3.0-39-gf57a172
caskroom/homebrew-cask (git revision ebc89; last commit 2017-08-07)
==> Homebrew-Cask Install Location
<NONE>
==> Homebrew-Cask Staging Location
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads
~/Library/Caches/Homebrew/Cask (47 files, 3.5GB)
==> Homebrew-Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3676 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1107 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (160 casks)
==> Contents of $LOAD_PATH
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Environment Variables
LANG="en_GB.UTF-8"
PATH="/usr/local/bin:/usr/local/opt/fzf/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/Homebrew/Library/Homebrew/shims/scm"
SHELL="/usr/local/bin/fish"
```
|
True
|
microsoft-office uninstall does not remove .app files from /Applications - #### General troubleshooting steps
- [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue.
- [X] None of the templates was appropriate for my issue, or I’m not sure.
- [X] I ran `brew update-reset && brew update` and retried my command.
- [X] I ran `brew doctor`, fixed as many issues as possible and retried my command.
- [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md).
#### Description of issue
`brew cask zap microsoft-office` should get rid of the apps, but the apps (microsoft-office is a suite) remain on /Applications/.
#### Output of your command with `--verbose --debug`
```
[I] gtklocker@schwarz ~> brew cask zap microsoft-office --verbose --debug
==> Zapping Cask microsoft-office
==> Implied "brew cask uninstall microsoft-office"
==> Un-installing artifacts
==> Determining which artifacts are present in Cask microsoft-office
==> 3 artifact/s defined
#<Hbc::Artifact::Uninstall:0x007fe6ad109b90>
#<Hbc::Artifact::Pkg:0x007fe6ad109758>
#<Hbc::Artifact::Zap:0x007fe6ad109460>
==> Un-installing artifact of class Hbc::Artifact::Uninstall
==> Running uninstall process for microsoft-office; your password may be necessary
==> Removing launchctl service com.microsoft.autoupdate.helpertool
==> Executing: ["/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"]
Password:
==> Removing launchctl service com.microsoft.office.licensing.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensing.helper"]
==> Removing launchctl service com.microsoft.office.licensingV2.helper
==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"]
==> Uninstalling packages:
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.package.*"]
==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.pkg.licensing"]
==> Dispatching zap stanza
==> Running zap process for microsoft-office; your password may be necessary
==> Removing files:
~/Library/Application Scripts/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Scripts/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.excel.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.powerpoint.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.word.sfl
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate.fba
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Caches/com.microsoft.autoupdate2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Excel
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Office365ServiceV2
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Outlook
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Powerpoint
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.Word
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.errorreporting
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Containers/com.microsoft.onenote.mac
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate.fba.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Cookies/com.microsoft.autoupdate2.binarycookies
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.Office
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.OfficeOsfWebHost
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Group Containers/UBF8T346G9.ms
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Excel.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Powerpoint.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.Word.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate.fba.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Preferences/com.microsoft.autoupdate2.plist
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.autoupdate2.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
~/Library/Saved Application State/com.microsoft.office.setupassistant.savedState
==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"]
==> Removing directories if empty:
~/Library/Caches/Microsoft/uls
~/Library/Caches/Microsoft
==> Removing all staged versions of Cask 'microsoft-office'
==> Purging all staged versions of Cask microsoft-office
[I] gtklocker@schwarz ~> ls -d1 /Applications/Microsoft\ *
/Applications/Microsoft Excel.app
/Applications/Microsoft OneNote.app
/Applications/Microsoft Outlook.app
/Applications/Microsoft PowerPoint.app
/Applications/Microsoft Word.app
```
#### Output of `brew cask doctor`
```
[I] gtklocker@schwarz ~> brew cask doctor
==> Homebrew-Cask Version
Homebrew-Cask 1.3.0-39-gf57a172
caskroom/homebrew-cask (git revision ebc89; last commit 2017-08-07)
==> Homebrew-Cask Install Location
<NONE>
==> Homebrew-Cask Staging Location
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads
~/Library/Caches/Homebrew/Cask (47 files, 3.5GB)
==> Homebrew-Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3676 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1107 casks)
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (160 casks)
==> Contents of $LOAD_PATH
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Environment Variables
LANG="en_GB.UTF-8"
PATH="/usr/local/bin:/usr/local/opt/fzf/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/Homebrew/Library/Homebrew/shims/scm"
SHELL="/usr/local/bin/fish"
```
|
main
|
microsoft office uninstall does not remove app files from applications general troubleshooting steps i have checked the instructions for or before opening the issue none of the templates was appropriate for my issue or i’m not sure i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i understand that description of issue brew cask zap microsoft office should get rid of the apps but the apps microsoft office is a suite remain on applications output of your command with verbose debug gtklocker schwarz brew cask zap microsoft office verbose debug zapping cask microsoft office implied brew cask uninstall microsoft office un installing artifacts determining which artifacts are present in cask microsoft office artifact s defined un installing artifact of class hbc artifact uninstall running uninstall process for microsoft office your password may be necessary removing launchctl service com microsoft autoupdate helpertool executing executing password removing launchctl service com microsoft office licensing helper executing executing removing launchctl service com microsoft office helper executing executing uninstalling packages executing executing dispatching zap stanza running zap process for microsoft office your password may be necessary removing files library application scripts com microsoft excel executing library application scripts com microsoft executing library application scripts com microsoft outlook executing library application scripts com microsoft powerpoint executing library application scripts com microsoft word executing library application scripts com microsoft errorreporting executing library application scripts com microsoft onenote mac executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft excel sfl executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft powerpoint sfl executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft word sfl executing library caches microsoft uls com microsoft autoupdate fba executing library caches microsoft uls com microsoft executing library caches com microsoft autoupdate fba executing library caches com microsoft executing library containers com microsoft excel executing library containers com microsoft executing library containers com microsoft outlook executing library containers com microsoft powerpoint executing library containers com microsoft word executing library containers com microsoft errorreporting executing library containers com microsoft onenote mac executing library cookies com microsoft autoupdate fba binarycookies executing library cookies com microsoft binarycookies executing library group containers office executing library group containers officeosfwebhost executing library group containers ms executing library preferences com microsoft excel plist executing library preferences com microsoft powerpoint plist executing library preferences com microsoft word plist executing library preferences com microsoft autoupdate fba plist executing library preferences com microsoft plist executing library saved application state com microsoft savedstate executing library saved application state com microsoft office setupassistant savedstate executing removing directories if empty library caches microsoft uls library caches microsoft removing all staged versions of cask microsoft office purging all staged versions of cask microsoft office gtklocker schwarz ls applications microsoft applications microsoft excel app applications microsoft onenote app applications microsoft outlook app applications microsoft powerpoint app applications microsoft word app output of brew cask doctor gtklocker schwarz brew cask doctor homebrew cask version homebrew cask caskroom homebrew cask git revision last commit homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask cached downloads library caches homebrew cask files homebrew cask taps usr local homebrew library taps caskroom homebrew cask casks usr local homebrew library taps caskroom homebrew fonts casks usr local homebrew library taps caskroom homebrew versions casks contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal environment variables lang en gb utf path usr local bin usr local opt fzf bin usr bin bin usr sbin sbin opt bin usr local homebrew library homebrew shims scm shell usr local bin fish
| 1
|
506,425
| 14,664,989,563
|
IssuesEvent
|
2020-12-29 13:17:23
|
teamforus/forus
|
https://api.github.com/repos/teamforus/forus
|
closed
|
Time based criteria, prevalidations adjustments
|
Epic Priority: Must have project-107 project-108
|
## Main asssignee: @
## Context/goal:
- [ ] Fund configuration for how long a record_type is trusted
- All record_types from criteria should have the trusted period (if you want to activate voucher directly)
- Describe all edge cases if trusted period is different:
- One record is valid, other is not
- All records are invalid
- [ ] create a prevalidation -> date of creation of prevalidation should added to record/record_validation
- [ ] Requester should be able to get a new voucher when they don’t have an active voucher.
- [ ] Use the expiration date of the voucher itself, ignore the expiration date of the fund.
- Math.min(fund.expiration_date, voucher.expiration_date);
- [ ] Replace or update ‘creation’/’trusted’ date of a prevalidation
- [ ] Soft deleting a single prevalidation from table
|
1.0
|
Time based criteria, prevalidations adjustments - ## Main asssignee: @
## Context/goal:
- [ ] Fund configuration for how long a record_type is trusted
- All record_types from criteria should have the trusted period (if you want to activate voucher directly)
- Describe all edge cases if trusted period is different:
- One record is valid, other is not
- All records are invalid
- [ ] create a prevalidation -> date of creation of prevalidation should added to record/record_validation
- [ ] Requester should be able to get a new voucher when they don’t have an active voucher.
- [ ] Use the expiration date of the voucher itself, ignore the expiration date of the fund.
- Math.min(fund.expiration_date, voucher.expiration_date);
- [ ] Replace or update ‘creation’/’trusted’ date of a prevalidation
- [ ] Soft deleting a single prevalidation from table
|
non_main
|
time based criteria prevalidations adjustments main asssignee context goal fund configuration for how long a record type is trusted all record types from criteria should have the trusted period if you want to activate voucher directly describe all edge cases if trusted period is different one record is valid other is not all records are invalid create a prevalidation date of creation of prevalidation should added to record record validation requester should be able to get a new voucher when they don’t have an active voucher use the expiration date of the voucher itself ignore the expiration date of the fund math min fund expiration date voucher expiration date replace or update ‘creation’ ’trusted’ date of a prevalidation soft deleting a single prevalidation from table
| 0
|
2,308
| 8,271,873,700
|
IssuesEvent
|
2018-09-16 14:14:45
|
react-navigation/react-navigation
|
https://api.github.com/repos/react-navigation/react-navigation
|
reopened
|
Params are not passed from Stack navigator to Stack navigator if they are children of SwitchNavigator
|
needs action from maintainer
|
## The issue tracker is reserved for bug reports only.
this is the bug only
previous same bugs that seem are not fixed yet
https://github.com/react-navigation/react-navigation/issues/143
https://github.com/react-navigation/react-navigation/issues/3787
### Current Behavior
Params are not passed from one screen to another. Screens are initial screens of different Stack navigators.
**Config**
```
SwitchNavigator
LoadingScreen: Component
App: StackNavigator
ProfileScreen
DashboardScreen
SignInFlow: StackNavigator
LoginScreen // from this screen (and vice versa)
LikeScreen
SignUpFlow: StackNavigator
SignUpScreen // to this screen (and vice versa)
TCScreen
```
**Flow**
From `LoginScreen` I try to navigate to `SignUpScreen` AND pass params.
As you can see above, these screens belong to different stacks.
Those stacks are children of `SwitchNavigator`.
**Issue**
Since `react-navigation v1` renders screen twice if I navigate from stack to stack (it is infamous bug of v1), I have to navigate to stack itself, not its particular screen.
```javascript
// params are passed OK, but RN renders screen twice. So it is not an option
this.props.navigation.navigate('LoginScreen', { ...this.state.data }) }
// RN renders screen once as expected, but params are not being passed
this.props.navigation.navigate('SignInFlow', { ...this.state.data }) }
```
### Expected Behavior
I get params in initial screen as if I was navigating to it, not to it's parent `Stacknavigator`
### How to reproduce
Same way
### Your Environment
| software | version
| ---------------- | -------
| react-navigation | 1.5.8
| react-native |0.54.2
| node |8.11.3
| npm or yarn |5.6.0
|
True
|
Params are not passed from Stack navigator to Stack navigator if they are children of SwitchNavigator - ## The issue tracker is reserved for bug reports only.
this is the bug only
previous same bugs that seem are not fixed yet
https://github.com/react-navigation/react-navigation/issues/143
https://github.com/react-navigation/react-navigation/issues/3787
### Current Behavior
Params are not passed from one screen to another. Screens are initial screens of different Stack navigators.
**Config**
```
SwitchNavigator
LoadingScreen: Component
App: StackNavigator
ProfileScreen
DashboardScreen
SignInFlow: StackNavigator
LoginScreen // from this screen (and vice versa)
LikeScreen
SignUpFlow: StackNavigator
SignUpScreen // to this screen (and vice versa)
TCScreen
```
**Flow**
From `LoginScreen` I try to navigate to `SignUpScreen` AND pass params.
As you can see above, these screens belong to different stacks.
Those stacks are children of `SwitchNavigator`.
**Issue**
Since `react-navigation v1` renders screen twice if I navigate from stack to stack (it is infamous bug of v1), I have to navigate to stack itself, not its particular screen.
```javascript
// params are passed OK, but RN renders screen twice. So it is not an option
this.props.navigation.navigate('LoginScreen', { ...this.state.data }) }
// RN renders screen once as expected, but params are not being passed
this.props.navigation.navigate('SignInFlow', { ...this.state.data }) }
```
### Expected Behavior
I get params in initial screen as if I was navigating to it, not to it's parent `Stacknavigator`
### How to reproduce
Same way
### Your Environment
| software | version
| ---------------- | -------
| react-navigation | 1.5.8
| react-native |0.54.2
| node |8.11.3
| npm or yarn |5.6.0
|
main
|
params are not passed from stack navigator to stack navigator if they are children of switchnavigator the issue tracker is reserved for bug reports only this is the bug only previous same bugs that seem are not fixed yet current behavior params are not passed from one screen to another screens are initial screens of different stack navigators config switchnavigator loadingscreen component app stacknavigator profilescreen dashboardscreen signinflow stacknavigator loginscreen from this screen and vice versa likescreen signupflow stacknavigator signupscreen to this screen and vice versa tcscreen flow from loginscreen i try to navigate to signupscreen and pass params as you can see above these screens belong to different stacks those stacks are children of switchnavigator issue since react navigation renders screen twice if i navigate from stack to stack it is infamous bug of i have to navigate to stack itself not its particular screen javascript params are passed ok but rn renders screen twice so it is not an option this props navigation navigate loginscreen this state data rn renders screen once as expected but params are not being passed this props navigation navigate signinflow this state data expected behavior i get params in initial screen as if i was navigating to it not to it s parent stacknavigator how to reproduce same way your environment software version react navigation react native node npm or yarn
| 1
|
1,508
| 6,523,561,302
|
IssuesEvent
|
2017-08-29 09:09:50
|
Kristinita/Erics-Green-Room
|
https://api.github.com/repos/Kristinita/Erics-Green-Room
|
opened
|
[Feature request] Улучшения дизайна комнаты
|
need-maintainer
|
### 1. Моё программное обеспечение
На котором сделаны скриншоты.
+ Windows 10 Enterprise LTSB 64-bit EN,
+ Firefox 55.0.3 (64-bit).
### 2. Кнопка «Выйти»

Полагаю, не очень удачно расположена, перекрывая скроллбар, верхнее меню и список пользователей.
### 3. Полоса под полем ответа

Не понимаю предназначение голубой полосы под полем ответа. Если она действительно не несёт никаких функций, можно её убрать, как занимающую лишнее место.
### 4. Настройка цвета в комнатах
#### 1. Желательно
Предоставить пользователям возможность выбора предпочтительного для них цвета бэкграунда.
#### 2. Альтернатива
Добавить в бэкграунд зелёный оттенок, поскольку комната называется «Зелёной комнатой Эрика». Сейчас фон при заходе через браузер бело-голубой.
Спасибо.
|
True
|
[Feature request] Улучшения дизайна комнаты - ### 1. Моё программное обеспечение
На котором сделаны скриншоты.
+ Windows 10 Enterprise LTSB 64-bit EN,
+ Firefox 55.0.3 (64-bit).
### 2. Кнопка «Выйти»

Полагаю, не очень удачно расположена, перекрывая скроллбар, верхнее меню и список пользователей.
### 3. Полоса под полем ответа

Не понимаю предназначение голубой полосы под полем ответа. Если она действительно не несёт никаких функций, можно её убрать, как занимающую лишнее место.
### 4. Настройка цвета в комнатах
#### 1. Желательно
Предоставить пользователям возможность выбора предпочтительного для них цвета бэкграунда.
#### 2. Альтернатива
Добавить в бэкграунд зелёный оттенок, поскольку комната называется «Зелёной комнатой Эрика». Сейчас фон при заходе через браузер бело-голубой.
Спасибо.
|
main
|
улучшения дизайна комнаты моё программное обеспечение на котором сделаны скриншоты windows enterprise ltsb bit en firefox bit кнопка «выйти» полагаю не очень удачно расположена перекрывая скроллбар верхнее меню и список пользователей полоса под полем ответа не понимаю предназначение голубой полосы под полем ответа если она действительно не несёт никаких функций можно её убрать как занимающую лишнее место настройка цвета в комнатах желательно предоставить пользователям возможность выбора предпочтительного для них цвета бэкграунда альтернатива добавить в бэкграунд зелёный оттенок поскольку комната называется «зелёной комнатой эрика» сейчас фон при заходе через браузер бело голубой спасибо
| 1
|
5,690
| 29,952,334,921
|
IssuesEvent
|
2023-06-23 03:04:18
|
spicetify/spicetify-themes
|
https://api.github.com/repos/spicetify/spicetify-themes
|
closed
|
[Dreary] Playlists not able to scroll
|
☠️ unmaintained
|
**Describe the bug**
There is no scroll bar on the playlists section nor is there an outline on them even though I enabled `spicetify config sidebar_config 1`
**To Reproduce**
Steps to reproduce the behavior:
Not sure but I can send my installed things if needed
**Expected behavior**
Be able to scroll through my playlists in the sidebar
**Screenshots**

**Logs**
```console
Failed to load resource: the server responded with a status of 404 ()
extension.js:1 Initializing Spicetify Marketplace v0.8.5
extension.js:1 Loaded Marketplace extension
hidePodcasts.js?time=1681738153179:1 HidePodcasts: Loaded
xpui.js:1 cosmos-connector: fetchAdForSlot endpoint=%o sp://ads/v1/ads/hpto
hidePodcasts.js?time=1681738153179:5 / <section class="main-shelf-shelf Shelf" aria-label="Jump back in">…</section> flex
hidePodcasts.js?time=1681738153179:5 Tagging carousel: Your shows
vendor~xpui.js:29 NotFoundError: GET request to sp://core-playlist/v1/playlist/spotify%3Aplaylist%3A37i9dQZF1EYkqdzj48dyYq/?responseFormat=protobufJson&start=0&length=0 failed with status code 404
at l.onSuccess (xpui.js:1:2001835)
at onSuccess (xpui.js:1:2001649)
onError @ vendor~xpui.js:29
raw.githubusercontent.com/3raxton/spicetify-custom-apps-and-extensions/main/manifest.json:1 Failed to load resource: the server responded with a status of 404 ()
raw.githubusercontent.com/TheRandomLabs/Scoop-Spotify/master/manifest.json:1 Failed to load resource: the server responded with a status of 404 ()
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/vendor~xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/9692.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/8934.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-hpto.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui-desktop-modals.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-v2.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/7542.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
extension.js:1 GET https://raw.githubusercontent.com/catppuccin/spicetify/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/itsmeow/Spicetify-Canvas/master/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/getomni/spicetify/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/schnensch0/ziro/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404
```
**Specifics (please complete the following information):**
- OS: [Windows]
- Spotify version [Spotify for Windows
1.2.9.743.g85d9593d]
- Spicetify version [Spicetify v2.17.2]
- Dreary

|
True
|
[Dreary] Playlists not able to scroll - **Describe the bug**
There is no scroll bar on the playlists section nor is there an outline on them even though I enabled `spicetify config sidebar_config 1`
**To Reproduce**
Steps to reproduce the behavior:
Not sure but I can send my installed things if needed
**Expected behavior**
Be able to scroll through my playlists in the sidebar
**Screenshots**

**Logs**
```console
Failed to load resource: the server responded with a status of 404 ()
extension.js:1 Initializing Spicetify Marketplace v0.8.5
extension.js:1 Loaded Marketplace extension
hidePodcasts.js?time=1681738153179:1 HidePodcasts: Loaded
xpui.js:1 cosmos-connector: fetchAdForSlot endpoint=%o sp://ads/v1/ads/hpto
hidePodcasts.js?time=1681738153179:5 / <section class="main-shelf-shelf Shelf" aria-label="Jump back in">…</section> flex
hidePodcasts.js?time=1681738153179:5 Tagging carousel: Your shows
vendor~xpui.js:29 NotFoundError: GET request to sp://core-playlist/v1/playlist/spotify%3Aplaylist%3A37i9dQZF1EYkqdzj48dyYq/?responseFormat=protobufJson&start=0&length=0 failed with status code 404
at l.onSuccess (xpui.js:1:2001835)
at onSuccess (xpui.js:1:2001649)
onError @ vendor~xpui.js:29
raw.githubusercontent.com/3raxton/spicetify-custom-apps-and-extensions/main/manifest.json:1 Failed to load resource: the server responded with a status of 404 ()
raw.githubusercontent.com/TheRandomLabs/Scoop-Spotify/master/manifest.json:1 Failed to load resource: the server responded with a status of 404 ()
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/vendor~xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/9692.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/8934.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-hpto.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui-desktop-modals.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-v2.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/7542.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED
extension.js:1 GET https://raw.githubusercontent.com/catppuccin/spicetify/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/itsmeow/Spicetify-Canvas/master/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/getomni/spicetify/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/schnensch0/ziro/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404
R @ extension.js:1
(anonymous) @ extension.js:1
(anonymous) @ extension.js:1
extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404
```
**Specifics (please complete the following information):**
- OS: [Windows]
- Spotify version [Spotify for Windows
1.2.9.743.g85d9593d]
- Spicetify version [Spicetify v2.17.2]
- Dreary

|
main
|
playlists not able to scroll describe the bug there is no scroll bar on the playlists section nor is there an outline on them even though i enabled spicetify config sidebar config to reproduce steps to reproduce the behavior not sure but i can send my installed things if needed expected behavior be able to scroll through my playlists in the sidebar screenshots logs console failed to load resource the server responded with a status of extension js initializing spicetify marketplace extension js loaded marketplace extension hidepodcasts js time hidepodcasts loaded xpui js cosmos connector fetchadforslot endpoint o sp ads ads hpto hidepodcasts js time … flex hidepodcasts js time tagging carousel your shows vendor xpui js notfounderror get request to sp core playlist playlist spotify responseformat protobufjson start length failed with status code at l onsuccess xpui js at onsuccess xpui js onerror vendor xpui js raw githubusercontent com spicetify custom apps and extensions main manifest json failed to load resource the server responded with a status of raw githubusercontent com therandomlabs scoop spotify master manifest json failed to load resource the server responded with a status of devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get specifics please complete the following information os spotify version spotify for windows spicetify version dreary
| 1
|
119,227
| 17,604,737,773
|
IssuesEvent
|
2021-08-17 15:40:37
|
Dima2021/NodeGoat
|
https://api.github.com/repos/Dima2021/NodeGoat
|
closed
|
WS-2020-0344 (High) detected in is-my-json-valid-2.19.0.tgz, is-my-json-valid-2.15.0.tgz - autoclosed
|
security vulnerability
|
## WS-2020-0344 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>is-my-json-valid-2.19.0.tgz</b>, <b>is-my-json-valid-2.15.0.tgz</b></p></summary>
<p>
<details><summary><b>is-my-json-valid-2.19.0.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- grunt-retire-0.3.12.tgz (Root Library)
- request-2.67.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.19.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>is-my-json-valid-2.15.0.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Dima2021/NodeGoat/commit/3b25a27fed9a418e332bfca71ab656c04a404f1e">3b25a27fed9a418e332bfca71ab656c04a404f1e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.
<p>Publish Date: 2020-06-09
<p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p>
<p>Release Date: 2020-06-09</p>
<p>Fix Resolution: is-my-json-valid - 2.20.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.19.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;har-validator:2.0.6;is-my-json-valid:2.19.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"},{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0344","vulnerabilityDetails":"Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
WS-2020-0344 (High) detected in is-my-json-valid-2.19.0.tgz, is-my-json-valid-2.15.0.tgz - autoclosed - ## WS-2020-0344 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>is-my-json-valid-2.19.0.tgz</b>, <b>is-my-json-valid-2.15.0.tgz</b></p></summary>
<p>
<details><summary><b>is-my-json-valid-2.19.0.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- grunt-retire-0.3.12.tgz (Root Library)
- request-2.67.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.19.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>is-my-json-valid-2.15.0.tgz</b></p></summary>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p>
<p>Path to dependency file: NodeGoat/package.json</p>
<p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- grunt-npm-install-0.3.1.tgz (Root Library)
- npm-3.10.10.tgz
- request-2.75.0.tgz
- har-validator-2.0.6.tgz
- :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/Dima2021/NodeGoat/commit/3b25a27fed9a418e332bfca71ab656c04a404f1e">3b25a27fed9a418e332bfca71ab656c04a404f1e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.
<p>Publish Date: 2020-06-09
<p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p>
<p>Release Date: 2020-06-09</p>
<p>Fix Resolution: is-my-json-valid - 2.20.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.19.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;har-validator:2.0.6;is-my-json-valid:2.19.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"},{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0344","vulnerabilityDetails":"Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_main
|
ws high detected in is my json valid tgz is my json valid tgz autoclosed ws high severity vulnerability vulnerable libraries is my json valid tgz is my json valid tgz is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules is my json valid package json dependency hierarchy grunt retire tgz root library request tgz har validator tgz x is my json valid tgz vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules npm node modules request node modules har validator node modules is my json valid package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch master vulnerability details arbitrary code execution vulnerability was found in is my json valid before via the fromatname function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt retire request har validator is my json valid isminimumfixversionavailable true minimumfixversion is my json valid packagetype javascript node js packagename is my json valid packageversion packagefilepaths istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid isminimumfixversionavailable true minimumfixversion is my json valid basebranches vulnerabilityidentifier ws vulnerabilitydetails arbitrary code execution vulnerability was found in is my json valid before via the fromatname function vulnerabilityurl
| 0
|
234,643
| 18,011,518,911
|
IssuesEvent
|
2021-09-16 09:08:58
|
gardener/documentation
|
https://api.github.com/repos/gardener/documentation
|
closed
|
Local preview doesn't work correctly
|
kind/bug component/documentation priority/2
|
**What happened**:
With the shift to `docforge` some folders were removed from documentation because they are fetched by `docforge`. Consequently when doing `make serve` for local preview they are missing form the website because the step with docforge build has not been executed. The only way round that is to use docforge, before doing local preview, similar to what we do in the automated central build. That unfortunately, makes it difficult to support live changes, but is still a way to preview before submitting.
**What you expected to happen**:
Supply the documentation manifest for which we want to see preview and have it built and shown for us.
/priority critical
|
1.0
|
Local preview doesn't work correctly - **What happened**:
With the shift to `docforge` some folders were removed from documentation because they are fetched by `docforge`. Consequently when doing `make serve` for local preview they are missing form the website because the step with docforge build has not been executed. The only way round that is to use docforge, before doing local preview, similar to what we do in the automated central build. That unfortunately, makes it difficult to support live changes, but is still a way to preview before submitting.
**What you expected to happen**:
Supply the documentation manifest for which we want to see preview and have it built and shown for us.
/priority critical
|
non_main
|
local preview doesn t work correctly what happened with the shift to docforge some folders were removed from documentation because they are fetched by docforge consequently when doing make serve for local preview they are missing form the website because the step with docforge build has not been executed the only way round that is to use docforge before doing local preview similar to what we do in the automated central build that unfortunately makes it difficult to support live changes but is still a way to preview before submitting what you expected to happen supply the documentation manifest for which we want to see preview and have it built and shown for us priority critical
| 0
|
5,318
| 26,839,240,214
|
IssuesEvent
|
2023-02-02 22:23:05
|
aws/aws-sam-cli
|
https://api.github.com/repos/aws/aws-sam-cli
|
closed
|
add deploy --outputs-file option, like AWS CDK
|
type/feature area/deploy stage/pm-review maintainer/need-followup
|
### Describe your idea/feature/enhancement
I wish SAM CLI would have an `--outputs-file` optional CLI argument for `sam deploy`, like the one `cdk deploy` has, see [the CDK docs](https://docs.aws.amazon.com/cdk/latest/guide/cli.html#w109aac23b7c33c13). Right now, it only prints the outputs in an ASCII table, mixed in with all the other stuff that goes to stdout, which of course is not machine readable, forcing a user to have to write a separate program to query CloudFormation to get the outputs.
|
True
|
add deploy --outputs-file option, like AWS CDK - ### Describe your idea/feature/enhancement
I wish SAM CLI would have an `--outputs-file` optional CLI argument for `sam deploy`, like the one `cdk deploy` has, see [the CDK docs](https://docs.aws.amazon.com/cdk/latest/guide/cli.html#w109aac23b7c33c13). Right now, it only prints the outputs in an ASCII table, mixed in with all the other stuff that goes to stdout, which of course is not machine readable, forcing a user to have to write a separate program to query CloudFormation to get the outputs.
|
main
|
add deploy outputs file option like aws cdk describe your idea feature enhancement i wish sam cli would have an outputs file optional cli argument for sam deploy like the one cdk deploy has see right now it only prints the outputs in an ascii table mixed in with all the other stuff that goes to stdout which of course is not machine readable forcing a user to have to write a separate program to query cloudformation to get the outputs
| 1
|
809,644
| 30,203,303,886
|
IssuesEvent
|
2023-07-05 07:43:59
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Filter box in table has horizontal scroll
|
Type:Bug Priority:P3 Querying/Parameters & Variables .Frontend Visualization/Tables .Team/42 :milky_way:
|
### Describe the bug
When filtering a field that has long strings, the filter box creates an unnecessary horizontal scroll.
### To Reproduce
1. Sample database -> People
2. Password (long strings field): filter by this column
3. Is: "8bee" or any known value (dont copy and paste)
4. Horizontal scroll
<img width="544" alt="Screenshot 2023-06-07 at 5 43 39 PM" src="https://github.com/metabase/metabase/assets/132273646/6ef8a49a-b8b6-45f8-bdfa-14b700d5a90d">
### Expected behavior
No horizontal scroll
### Logs
_No response_
### Information about your Metabase installation
```JSON
v1.46.4
```
### Severity
Minor
### Additional context
_No response_
|
1.0
|
Filter box in table has horizontal scroll - ### Describe the bug
When filtering a field that has long strings, the filter box creates an unnecessary horizontal scroll.
### To Reproduce
1. Sample database -> People
2. Password (long strings field): filter by this column
3. Is: "8bee" or any known value (dont copy and paste)
4. Horizontal scroll
<img width="544" alt="Screenshot 2023-06-07 at 5 43 39 PM" src="https://github.com/metabase/metabase/assets/132273646/6ef8a49a-b8b6-45f8-bdfa-14b700d5a90d">
### Expected behavior
No horizontal scroll
### Logs
_No response_
### Information about your Metabase installation
```JSON
v1.46.4
```
### Severity
Minor
### Additional context
_No response_
|
non_main
|
filter box in table has horizontal scroll describe the bug when filtering a field that has long strings the filter box creates an unnecessary horizontal scroll to reproduce sample database people password long strings field filter by this column is or any known value dont copy and paste horizontal scroll img width alt screenshot at pm src expected behavior no horizontal scroll logs no response information about your metabase installation json severity minor additional context no response
| 0
|
99,815
| 4,065,670,387
|
IssuesEvent
|
2016-05-26 12:17:27
|
Taeir/ContextProject-MIGI2
|
https://api.github.com/repos/Taeir/ContextProject-MIGI2
|
closed
|
Refactoring - Refactoring Packages
|
Priority A
|
The packages needs to be clear and correct to improve the structure of the code.
This is a subtask of #207 - Refactoring
|
1.0
|
Refactoring - Refactoring Packages - The packages needs to be clear and correct to improve the structure of the code.
This is a subtask of #207 - Refactoring
|
non_main
|
refactoring refactoring packages the packages needs to be clear and correct to improve the structure of the code this is a subtask of refactoring
| 0
|
23
| 2,524,070,209
|
IssuesEvent
|
2015-01-20 15:33:41
|
simplesamlphp/simplesamlphp
|
https://api.github.com/repos/simplesamlphp/simplesamlphp
|
closed
|
Remove backwards-compatibility code in SimpleSAML_SessionHandlerStore
|
enhancement maintainability started
|
The `loadSession()` method has some backwards-compatible code that must go away.
|
True
|
Remove backwards-compatibility code in SimpleSAML_SessionHandlerStore - The `loadSession()` method has some backwards-compatible code that must go away.
|
main
|
remove backwards compatibility code in simplesaml sessionhandlerstore the loadsession method has some backwards compatible code that must go away
| 1
|
1,554
| 6,572,249,517
|
IssuesEvent
|
2017-09-11 00:35:58
|
ansible/ansible-modules-extras
|
https://api.github.com/repos/ansible/ansible-modules-extras
|
closed
|
New monit release breaks monit module
|
affects_2.1 bug_report waiting_on_maintainer
|
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
monitoring/monit
##### ANSIBLE VERSION
```
ansible 2.1.0.0
```
##### SUMMARY
Since Monit changed the output of the `summary` command in version 5.18, the monit module is not able to parse it.
##### STEPS TO REPRODUCE
Running the following task:
```
- name: Start server
monit: name=server state=monitored
```
with the corresponding monit service definition in place, results in
```
fatal: [xhost]: FAILED! => {
"changed": false,
"failed": true,
"msg": "server process not presently configured with monit",
"name": "server",
"state": "monitored"
}
```
This happens due to the change of the output of the `monit summary` command:
before:
```
The Monit daemon 5.17.1 uptime: 0m
System 'xhost' Running
Process 'server' Running
```
after upgrade to Monit 5.18:
```
Monit uptime: 56m
Service Name Status Type
xhost Running System
server Running Process
```
Notice the additional header and the re-ordered columns.
|
True
|
New monit release breaks monit module - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
monitoring/monit
##### ANSIBLE VERSION
```
ansible 2.1.0.0
```
##### SUMMARY
Since Monit changed the output of the `summary` command in version 5.18, the monit module is not able to parse it.
##### STEPS TO REPRODUCE
Running the following task:
```
- name: Start server
monit: name=server state=monitored
```
with the corresponding monit service definition in place, results in
```
fatal: [xhost]: FAILED! => {
"changed": false,
"failed": true,
"msg": "server process not presently configured with monit",
"name": "server",
"state": "monitored"
}
```
This happens due to the change of the output of the `monit summary` command:
before:
```
The Monit daemon 5.17.1 uptime: 0m
System 'xhost' Running
Process 'server' Running
```
after upgrade to Monit 5.18:
```
Monit uptime: 56m
Service Name Status Type
xhost Running System
server Running Process
```
Notice the additional header and the re-ordered columns.
|
main
|
new monit release breaks monit module issue type bug report component name monitoring monit ansible version ansible summary since monit changed the output of the summary command in version the monit module is not able to parse it steps to reproduce running the following task name start server monit name server state monitored with the corresponding monit service definition in place results in fatal failed changed false failed true msg server process not presently configured with monit name server state monitored this happens due to the change of the output of the monit summary command before the monit daemon uptime system xhost running process server running after upgrade to monit monit uptime service name status type xhost running system server running process notice the additional header and the re ordered columns
| 1
|
1,896
| 6,577,539,092
|
IssuesEvent
|
2017-09-12 01:37:11
|
ansible/ansible-modules-core
|
https://api.github.com/repos/ansible/ansible-modules-core
|
closed
|
ansible-modules-core/cloud/amazon/ec2_vpc.py tag issue - feature request
|
affects_2.0 aws cloud feature_idea waiting_on_maintainer
|
##### Issue Type:
- Feature Idea
##### Plugin Name:
- ec2_vpc.py
##### Ansible Version:
ansible 2.0.0.2
config file = /home/naslanidis/code/ag-vpc-management/ansible.cfg
configured module search path = /usr/share/ansible/modules
##### Ansible Configuration:
N/A
##### Environment:
Fedora 22 x64
##### Summary:
When creating VPC's using this module, if any tag is changed or added a completely new VPC is created rather than the tags being added or updated. Looking at the documentation, all tags are used along with the cidr range to uniquely identify an existing VPC. In my opinion this should just be a NAME tag and other tags should be able to be added, changed and removed for an existing VPC
##### Steps To Reproduce:
1. Create a VPC using this module with a number of tags in place
2. Add a new tag (e.g. cost code), and rerun the script
3. A new duplicate VPC is created rather than adding the cost code tag to the existing VPC
```
- name: test out vpc module 1st run though
ec2_vpc:
state: present
region: ap-southeast-2
profile: non_prod
cidr_block: 10.33.0.0/16
resource_tags:
Name: AAG-VPC-33
Owner: ABC
register: created_vpc
```
```
- name: test out vpc module 2nd run though
ec2_vpc:
state: present
region: ap-southeast-2
profile: non_prod
cidr_block: 10.33.0.0/16
resource_tags:
Name: AAG-VPC-33
Owner: ABC
Cost_code: 3
register: created_vpc
```
##### Expected Results:
The existing VPC should be updated with the new Cost_code tag.
##### Actual Results:
A new VPC is created that duplicates the first VPC.
|
True
|
ansible-modules-core/cloud/amazon/ec2_vpc.py tag issue - feature request - ##### Issue Type:
- Feature Idea
##### Plugin Name:
- ec2_vpc.py
##### Ansible Version:
ansible 2.0.0.2
config file = /home/naslanidis/code/ag-vpc-management/ansible.cfg
configured module search path = /usr/share/ansible/modules
##### Ansible Configuration:
N/A
##### Environment:
Fedora 22 x64
##### Summary:
When creating VPC's using this module, if any tag is changed or added a completely new VPC is created rather than the tags being added or updated. Looking at the documentation, all tags are used along with the cidr range to uniquely identify an existing VPC. In my opinion this should just be a NAME tag and other tags should be able to be added, changed and removed for an existing VPC
##### Steps To Reproduce:
1. Create a VPC using this module with a number of tags in place
2. Add a new tag (e.g. cost code), and rerun the script
3. A new duplicate VPC is created rather than adding the cost code tag to the existing VPC
```
- name: test out vpc module 1st run though
ec2_vpc:
state: present
region: ap-southeast-2
profile: non_prod
cidr_block: 10.33.0.0/16
resource_tags:
Name: AAG-VPC-33
Owner: ABC
register: created_vpc
```
```
- name: test out vpc module 2nd run though
ec2_vpc:
state: present
region: ap-southeast-2
profile: non_prod
cidr_block: 10.33.0.0/16
resource_tags:
Name: AAG-VPC-33
Owner: ABC
Cost_code: 3
register: created_vpc
```
##### Expected Results:
The existing VPC should be updated with the new Cost_code tag.
##### Actual Results:
A new VPC is created that duplicates the first VPC.
|
main
|
ansible modules core cloud amazon vpc py tag issue feature request issue type feature idea plugin name vpc py ansible version ansible config file home naslanidis code ag vpc management ansible cfg configured module search path usr share ansible modules ansible configuration n a environment fedora summary when creating vpc s using this module if any tag is changed or added a completely new vpc is created rather than the tags being added or updated looking at the documentation all tags are used along with the cidr range to uniquely identify an existing vpc in my opinion this should just be a name tag and other tags should be able to be added changed and removed for an existing vpc steps to reproduce create a vpc using this module with a number of tags in place add a new tag e g cost code and rerun the script a new duplicate vpc is created rather than adding the cost code tag to the existing vpc name test out vpc module run though vpc state present region ap southeast profile non prod cidr block resource tags name aag vpc owner abc register created vpc name test out vpc module run though vpc state present region ap southeast profile non prod cidr block resource tags name aag vpc owner abc cost code register created vpc expected results the existing vpc should be updated with the new cost code tag actual results a new vpc is created that duplicates the first vpc
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.