Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
1,983
6,694,204,027
IssuesEvent
2017-10-10 00:18:02
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Dictionary Definition: raw food
Maintainer Input Requested
Food that was not heated above 42 Celsius or 107.60000 Fahrenheit Examples: fresh fruit, vegetables, sprouts (also sprouted grains), grasses Further examples of raw "gourmet" food: nuts and seeds; soaked grains --- IA Page: http://duck.co/ia/view/dictionary_definition [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jagtalon
True
Dictionary Definition: raw food - Food that was not heated above 42 Celsius or 107.60000 Fahrenheit Examples: fresh fruit, vegetables, sprouts (also sprouted grains), grasses Further examples of raw "gourmet" food: nuts and seeds; soaked grains --- IA Page: http://duck.co/ia/view/dictionary_definition [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jagtalon
main
dictionary definition raw food food that was not heated above celsius or fahrenheit examples fresh fruit vegetables sprouts also sprouted grains grasses further examples of raw gourmet food nuts and seeds soaked grains ia page jagtalon
1
1,068
4,889,235,855
IssuesEvent
2016-11-18 09:31:52
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
include_role does not support tags
affects_2.2 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/task/feature --> include_role ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Linux, Red Hat Enterprise 7.2 ##### SUMMARY Originally filed this issue here because I was told on IRC this was part of ansible because it was a action plugin. The bot told me i was wrong after a few days, so now it's here. https://github.com/ansible/ansible/issues/17761 Forgive my terminology. I will do my best to describe that issue I am seeing. I am using a pre-release version of Ansible 2.2 for this. When using include_role module with tags, the include_role seems to ignore the tags and does not apply them when being called. I am using include_role in my playbook in the task list. I apply a tag on it. When I call ansible-playbook -t tagname it does not run my include_role task. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: test playbook hosts: localhost tasks: - debug: msg: "hello from debug" tags: - test - include_role: name: role1 tags: - test ``` OUTPUT ``` $ ansible-playbook -t test test.yml PLAY [test playbook] *********************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [debug] ******************************************************************* ok: [localhost] => { "msg": "hello from debug" } PLAY RECAP ********************************************************************* localhost : ok=2 changed=0 unreachable=0 failed=0 ``` ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I would expect it to run the role1, because it was called with the tags ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> role1 was never called
True
include_role does not support tags - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/task/feature --> include_role ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT Linux, Red Hat Enterprise 7.2 ##### SUMMARY Originally filed this issue here because I was told on IRC this was part of ansible because it was a action plugin. The bot told me i was wrong after a few days, so now it's here. https://github.com/ansible/ansible/issues/17761 Forgive my terminology. I will do my best to describe that issue I am seeing. I am using a pre-release version of Ansible 2.2 for this. When using include_role module with tags, the include_role seems to ignore the tags and does not apply them when being called. I am using include_role in my playbook in the task list. I apply a tag on it. When I call ansible-playbook -t tagname it does not run my include_role task. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` --- - name: test playbook hosts: localhost tasks: - debug: msg: "hello from debug" tags: - test - include_role: name: role1 tags: - test ``` OUTPUT ``` $ ansible-playbook -t test test.yml PLAY [test playbook] *********************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [debug] ******************************************************************* ok: [localhost] => { "msg": "hello from debug" } PLAY RECAP ********************************************************************* localhost : ok=2 changed=0 unreachable=0 failed=0 ``` ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I would expect it to run the role1, because it was called with the tags ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> role1 was never called
main
include role does not support tags issue type bug report component name include role ansible version ansible configuration n a os environment linux red hat enterprise summary originally filed this issue here because i was told on irc this was part of ansible because it was a action plugin the bot told me i was wrong after a few days so now it s here forgive my terminology i will do my best to describe that issue i am seeing i am using a pre release version of ansible for this when using include role module with tags the include role seems to ignore the tags and does not apply them when being called i am using include role in my playbook in the task list i apply a tag on it when i call ansible playbook t tagname it does not run my include role task steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name test playbook hosts localhost tasks debug msg hello from debug tags test include role name tags test output ansible playbook t test test yml play task ok task ok msg hello from debug play recap localhost ok changed unreachable failed expected results i would expect it to run the because it was called with the tags actual results was never called
1
2,394
8,499,960,216
IssuesEvent
2018-10-29 18:30:42
Qo2770/Algorithms
https://api.github.com/repos/Qo2770/Algorithms
closed
Algorithms is searching for Maintainers!
Maintainers Wanted
*This issue was created by [Maintainers Wanted](https://maintainerswanted.com)* :nerd_face: *Support us by leaving a star on [Github](https://github.com/flxwu/maintainerswanted.com)!* :star2: ## Algorithms is searching for new Maintainers! :man_technologist: :mailbox_with_mail: Do you use Algorithms personally or at work and would like this project to be further developed and improved? Or are you already a contributor and ready to take the next step to becoming a maintainer? If you are interested, comment here below on this issue! :point_down::raised_hands:
True
Algorithms is searching for Maintainers! - *This issue was created by [Maintainers Wanted](https://maintainerswanted.com)* :nerd_face: *Support us by leaving a star on [Github](https://github.com/flxwu/maintainerswanted.com)!* :star2: ## Algorithms is searching for new Maintainers! :man_technologist: :mailbox_with_mail: Do you use Algorithms personally or at work and would like this project to be further developed and improved? Or are you already a contributor and ready to take the next step to becoming a maintainer? If you are interested, comment here below on this issue! :point_down::raised_hands:
main
algorithms is searching for maintainers this issue was created by nerd face support us by leaving a star on algorithms is searching for new maintainers man technologist mailbox with mail do you use algorithms personally or at work and would like this project to be further developed and improved or are you already a contributor and ready to take the next step to becoming a maintainer if you are interested comment here below on this issue point down raised hands
1
5,128
26,142,624,704
IssuesEvent
2022-12-29 21:05:03
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
Field Permissions: Become a maintainer for this project
Maintainer change request
I would like to become the maintainer for the Field Permissions project (which currently is "looking for maintainers"). Issue in that project: https://github.com/backdrop-contrib/field_permissions/issues/7 PR to update the README file: https://github.com/backdrop-contrib/field_permissions/pull/8
True
Field Permissions: Become a maintainer for this project - I would like to become the maintainer for the Field Permissions project (which currently is "looking for maintainers"). Issue in that project: https://github.com/backdrop-contrib/field_permissions/issues/7 PR to update the README file: https://github.com/backdrop-contrib/field_permissions/pull/8
main
field permissions become a maintainer for this project i would like to become the maintainer for the field permissions project which currently is looking for maintainers issue in that project pr to update the readme file
1
2,257
7,933,144,877
IssuesEvent
2018-07-08 00:42:12
Subsurface-divelog/subsurface
https://api.github.com/repos/Subsurface-divelog/subsurface
closed
Syncing with cloud/ Creating commit fails
bug cloud-storage needs-maintainer-feedback ui-desktop
<!-- Lines like this one are comments and will not be shown in the final output. --> <!-- If you are a collaborator, please add labels and assign other collaborators for a review. --> ### Describe the issue: - [x] Bug ### Issue long description: Subsurface fails to sync with cloud storage. A red banner shows "creating commit failed". ### Operating system: archlinux ### Subsurface version: Self compiled binary: Subsurface v4.7.8-289-g9611e92cf0d9, built with libdivecomputer v0.7.0-devel-Subsurface-NG (e97a47cca55973199715df0f818b4955e60d3a31) built with Qt Version 5.11.0, runtime from Qt Version 5.11.0 built with libgit2 0.27.0 Also happens with Appimage: Subsurface v4.7.8, built with libdivecomputer v0.7.0-devel-Subsurface-branch (d02f1c3cdcc6a04d085538578d872ec6e3282382) built with Qt Version 5.9.3, runtime from Qt Version 5.9.3 built with libgit2 0.26.0 Also happens with Archlinux package: Subsurface v, built with libdivecomputer v0.7.0-devel-Subsurface-branch () built with Qt Version 5.11.0, runtime from Qt Version 5.11.0 built with libgit2 0.27.0 ### Steps to reproduce: 1) open local logbook *.ssrf 2) select file-> save to cloud storage 3) red banner appears, "creating commit failed", new dives are not saved to cloud ### Current behavior: No new commit is created in ./subsurface/cloudstorage/hash/ ### Expected behavior: Create a new commit with the changes from the local file and push it to the cloud server. ### Additional information: Command line log: ... cloud URL set as "https://cloud.subsurface-divelog.org//git/___@gmail.com[___@gmail.com]" Saving cloud storage to: "https://cloud.subsurface-divelog.org//git/___@gmail.com[__@gmail.com]" git storage: Synchronising data file git storage: Sync with cloud storage git storage: Successful cloud connection, fetch remote git storage: Done syncing with cloud storage git storage: Preparing to save data git storage: Start saving data git storage: Start saving dives git storage: Done creating local cache ![failed](https://user-images.githubusercontent.com/6360942/40850387-c2140236-65c4-11e8-9f3b-57467decb770.png) seems to be related to https://github.com/Subsurface-divelog/subsurface/blob/2209054226f463a3fd0b88a34e9dcd18e15b363b/core/save-git.c#L1271
True
Syncing with cloud/ Creating commit fails - <!-- Lines like this one are comments and will not be shown in the final output. --> <!-- If you are a collaborator, please add labels and assign other collaborators for a review. --> ### Describe the issue: - [x] Bug ### Issue long description: Subsurface fails to sync with cloud storage. A red banner shows "creating commit failed". ### Operating system: archlinux ### Subsurface version: Self compiled binary: Subsurface v4.7.8-289-g9611e92cf0d9, built with libdivecomputer v0.7.0-devel-Subsurface-NG (e97a47cca55973199715df0f818b4955e60d3a31) built with Qt Version 5.11.0, runtime from Qt Version 5.11.0 built with libgit2 0.27.0 Also happens with Appimage: Subsurface v4.7.8, built with libdivecomputer v0.7.0-devel-Subsurface-branch (d02f1c3cdcc6a04d085538578d872ec6e3282382) built with Qt Version 5.9.3, runtime from Qt Version 5.9.3 built with libgit2 0.26.0 Also happens with Archlinux package: Subsurface v, built with libdivecomputer v0.7.0-devel-Subsurface-branch () built with Qt Version 5.11.0, runtime from Qt Version 5.11.0 built with libgit2 0.27.0 ### Steps to reproduce: 1) open local logbook *.ssrf 2) select file-> save to cloud storage 3) red banner appears, "creating commit failed", new dives are not saved to cloud ### Current behavior: No new commit is created in ./subsurface/cloudstorage/hash/ ### Expected behavior: Create a new commit with the changes from the local file and push it to the cloud server. ### Additional information: Command line log: ... cloud URL set as "https://cloud.subsurface-divelog.org//git/___@gmail.com[___@gmail.com]" Saving cloud storage to: "https://cloud.subsurface-divelog.org//git/___@gmail.com[__@gmail.com]" git storage: Synchronising data file git storage: Sync with cloud storage git storage: Successful cloud connection, fetch remote git storage: Done syncing with cloud storage git storage: Preparing to save data git storage: Start saving data git storage: Start saving dives git storage: Done creating local cache ![failed](https://user-images.githubusercontent.com/6360942/40850387-c2140236-65c4-11e8-9f3b-57467decb770.png) seems to be related to https://github.com/Subsurface-divelog/subsurface/blob/2209054226f463a3fd0b88a34e9dcd18e15b363b/core/save-git.c#L1271
main
syncing with cloud creating commit fails describe the issue bug issue long description subsurface fails to sync with cloud storage a red banner shows creating commit failed operating system archlinux subsurface version self compiled binary subsurface built with libdivecomputer devel subsurface ng built with qt version runtime from qt version built with also happens with appimage subsurface built with libdivecomputer devel subsurface branch built with qt version runtime from qt version built with also happens with archlinux package subsurface v built with libdivecomputer devel subsurface branch built with qt version runtime from qt version built with steps to reproduce open local logbook ssrf select file save to cloud storage red banner appears creating commit failed new dives are not saved to cloud current behavior no new commit is created in subsurface cloudstorage hash expected behavior create a new commit with the changes from the local file and push it to the cloud server additional information command line log cloud url set as saving cloud storage to git storage synchronising data file git storage sync with cloud storage git storage successful cloud connection fetch remote git storage done syncing with cloud storage git storage preparing to save data git storage start saving data git storage start saving dives git storage done creating local cache seems to be related to
1
3,212
12,288,607,697
IssuesEvent
2020-05-09 17:30:01
short-d/short
https://api.github.com/repos/short-d/short
opened
[Refactor] Add foreign keys for `user` referenced tables.
maintainability
There are a lot of DB tables that hold the `user_id` reference which leads to data integrity violation, as anytime a `user` record can be deleted or modified. **Your solution** After the migration process (primary key user_email -> user_id) for `user` table to be completed, it requires adding foreign key constraints in referenced tables.
True
[Refactor] Add foreign keys for `user` referenced tables. - There are a lot of DB tables that hold the `user_id` reference which leads to data integrity violation, as anytime a `user` record can be deleted or modified. **Your solution** After the migration process (primary key user_email -> user_id) for `user` table to be completed, it requires adding foreign key constraints in referenced tables.
main
add foreign keys for user referenced tables there are a lot of db tables that hold the user id reference which leads to data integrity violation as anytime a user record can be deleted or modified your solution after the migration process primary key user email user id for user table to be completed it requires adding foreign key constraints in referenced tables
1
4,463
23,253,711,311
IssuesEvent
2022-08-04 07:22:10
JENOT-ANT/ENIGMA
https://api.github.com/repos/JENOT-ANT/ENIGMA
opened
Konsolowy interfejs użytkownika
dev task(s) waiting-for-maintainers
Trzebaby dodać jakieś menu użytkownika z opcjami typu: - szyfruj z pliku - szyfruj w czasie rzeczywistym - konfiguracja maszyny - itp.
True
Konsolowy interfejs użytkownika - Trzebaby dodać jakieś menu użytkownika z opcjami typu: - szyfruj z pliku - szyfruj w czasie rzeczywistym - konfiguracja maszyny - itp.
main
konsolowy interfejs użytkownika trzebaby dodać jakieś menu użytkownika z opcjami typu szyfruj z pliku szyfruj w czasie rzeczywistym konfiguracja maszyny itp
1
896
4,554,432,795
IssuesEvent
2016-09-13 09:25:25
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
eos_template: TypeError: load_config() got an unexpected keyword argument 'session'
affects_2.2 bug_report networking waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME eos_template ##### ANSIBLE VERSION ``` ansible --version ansible 2.2.0 (devel 9868117d1f) last updated 2016/09/12 20:18:28 (GMT +100) lib/ansible/modules/core: (devel 432ee70da1) last updated 2016/09/12 20:19:14 (GMT +100) lib/ansible/modules/extras: (devel 67a1bebbd3) last updated 2016/09/12 12:05:15 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY https://github.com/ansible/ansible-modules-core/commit/e464599632340f247c593f2db770be9782f7dec5 changed `module.config.load_config` to pass in `session` however that doesn't seem to be in https://github.com/ansible/ansible/blame/devel/lib/ansible/module_utils/netcfg.py#L56 @privateip When you ran `ansible-playbook -vvv eos.yaml ` Did you have other local changes ##### STEPS TO REPRODUCE ``` ansible-playbook -vvv eos.yaml -i ../inventory-testnetwork -e "limit_to=eos_template" ``` ##### EXPECTED RESULTS Tests to pass ##### ACTUAL RESULTS ``` TASK [test_eos_template : configure device with config] ************************ task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_eos_template/tests/cli/force.yaml:14 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/eos/eos_template.py <veos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <veos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" && echo ansible-tmp-1473707985.23-162291198567673="` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" ) && sleep 0' <veos01> PUT /tmp/tmpsPqOzL TO /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py <veos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/ /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py && sleep 0' <veos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 215, in <module> main() File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 205, in main commit=True) File "/tmp/ansible_aJfsia/ansible_modlib.zip/ansible/module_utils/netcfg.py", line 58, in load_config TypeError: load_config() got an unexpected keyword argument 'session' ```
True
eos_template: TypeError: load_config() got an unexpected keyword argument 'session' - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME eos_template ##### ANSIBLE VERSION ``` ansible --version ansible 2.2.0 (devel 9868117d1f) last updated 2016/09/12 20:18:28 (GMT +100) lib/ansible/modules/core: (devel 432ee70da1) last updated 2016/09/12 20:19:14 (GMT +100) lib/ansible/modules/extras: (devel 67a1bebbd3) last updated 2016/09/12 12:05:15 (GMT +100) ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY https://github.com/ansible/ansible-modules-core/commit/e464599632340f247c593f2db770be9782f7dec5 changed `module.config.load_config` to pass in `session` however that doesn't seem to be in https://github.com/ansible/ansible/blame/devel/lib/ansible/module_utils/netcfg.py#L56 @privateip When you ran `ansible-playbook -vvv eos.yaml ` Did you have other local changes ##### STEPS TO REPRODUCE ``` ansible-playbook -vvv eos.yaml -i ../inventory-testnetwork -e "limit_to=eos_template" ``` ##### EXPECTED RESULTS Tests to pass ##### ACTUAL RESULTS ``` TASK [test_eos_template : configure device with config] ************************ task path: /home/johnb/git/ansible-inc/test-network-modules/roles/test_eos_template/tests/cli/force.yaml:14 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/eos/eos_template.py <veos01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <veos01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" && echo ansible-tmp-1473707985.23-162291198567673="` echo $HOME/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673 `" ) && sleep 0' <veos01> PUT /tmp/tmpsPqOzL TO /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py <veos01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/ /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py && sleep 0' <veos01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/eos_template.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1473707985.23-162291198567673/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 215, in <module> main() File "/tmp/ansible_aJfsia/ansible_module_eos_template.py", line 205, in main commit=True) File "/tmp/ansible_aJfsia/ansible_modlib.zip/ansible/module_utils/netcfg.py", line 58, in load_config TypeError: load_config() got an unexpected keyword argument 'session' ```
main
eos template typeerror load config got an unexpected keyword argument session issue type bug report component name eos template ansible version ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt configuration os environment summary changed module config load config to pass in session however that doesn t seem to be in privateip when you ran ansible playbook vvv eos yaml did you have other local changes steps to reproduce ansible playbook vvv eos yaml i inventory testnetwork e limit to eos template expected results tests to pass actual results task task path home johnb git ansible inc test network modules roles test eos template tests cli force yaml using module file home johnb git ansible inc ansible lib ansible modules core network eos eos template py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpspqozl to home johnb ansible tmp ansible tmp eos template py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp eos template py sleep exec bin sh c python home johnb ansible tmp ansible tmp eos template py rm rf home johnb ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible ajfsia ansible module eos template py line in main file tmp ansible ajfsia ansible module eos template py line in main commit true file tmp ansible ajfsia ansible modlib zip ansible module utils netcfg py line in load config typeerror load config got an unexpected keyword argument session
1
365,823
25,554,076,383
IssuesEvent
2022-11-30 04:05:46
simonw/datasette
https://api.github.com/repos/simonw/datasette
closed
Interactive demo of Datasette 1.0 write APIs
enhancement documentation ops
I'm going to try to get this working on https://latest.datasette.io/ - it already has a way for people to sign in as root, but none of the databases there are writable. So I'm going to build a plugin which adds a writable named in-memory database. And some kind of mechanism for clearing out that database on a regular basis - maybe tables in that database get deleted automatically an hour after they are created? (Would be neat to display their time-left-until-deleted too)
1.0
Interactive demo of Datasette 1.0 write APIs - I'm going to try to get this working on https://latest.datasette.io/ - it already has a way for people to sign in as root, but none of the databases there are writable. So I'm going to build a plugin which adds a writable named in-memory database. And some kind of mechanism for clearing out that database on a regular basis - maybe tables in that database get deleted automatically an hour after they are created? (Would be neat to display their time-left-until-deleted too)
non_main
interactive demo of datasette write apis i m going to try to get this working on it already has a way for people to sign in as root but none of the databases there are writable so i m going to build a plugin which adds a writable named in memory database and some kind of mechanism for clearing out that database on a regular basis maybe tables in that database get deleted automatically an hour after they are created would be neat to display their time left until deleted too
0
1,138
4,998,877,174
IssuesEvent
2016-12-09 21:19:46
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2_vpc: Must specify resource_tags when deleting(!) VPC
affects_2.1 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_vpc ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> ##### SUMMARY According to [the doc](https://docs.ansible.com/ansible/ec2_vpc_module.html) you don't have to specify the `resource_tags` parameter when deleting a VPC. However, this module barfs when this parameter is missing. ##### STEPS TO REPRODUCE This is my task: <!--- Paste example playbooks or commands between quotes below --> ``` - name: "Destroy PREPROD VPC" ec2_vpc: region: "{{ region }}" state: absent vpc_id: "{{ preprod_vpc.vpcs[0].id }}" Name: "{{ owner }}_vpc_{{ env }}" Env: "{{ owner }}_{{ env }}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS I expected that the task would be executed successfully. ##### ACTUAL RESULTS `ec2_vpc` complained as follows: <!--- Paste verbatim command output between quotes below --> ``` missing required arguments: resource_tags ``` After I added the parameter the call was successful.
True
ec2_vpc: Must specify resource_tags when deleting(!) VPC - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ec2_vpc ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> ##### SUMMARY According to [the doc](https://docs.ansible.com/ansible/ec2_vpc_module.html) you don't have to specify the `resource_tags` parameter when deleting a VPC. However, this module barfs when this parameter is missing. ##### STEPS TO REPRODUCE This is my task: <!--- Paste example playbooks or commands between quotes below --> ``` - name: "Destroy PREPROD VPC" ec2_vpc: region: "{{ region }}" state: absent vpc_id: "{{ preprod_vpc.vpcs[0].id }}" Name: "{{ owner }}_vpc_{{ env }}" Env: "{{ owner }}_{{ env }}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS I expected that the task would be executed successfully. ##### ACTUAL RESULTS `ec2_vpc` complained as follows: <!--- Paste verbatim command output between quotes below --> ``` missing required arguments: resource_tags ``` After I added the parameter the call was successful.
main
vpc must specify resource tags when deleting vpc issue type bug report component name vpc ansible version ansible config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific summary according to you don t have to specify the resource tags parameter when deleting a vpc however this module barfs when this parameter is missing steps to reproduce this is my task name destroy preprod vpc vpc region region state absent vpc id preprod vpc vpcs id name owner vpc env env owner env expected results i expected that the task would be executed successfully actual results vpc complained as follows missing required arguments resource tags after i added the parameter the call was successful
1
35,133
12,311,802,245
IssuesEvent
2020-05-12 13:01:10
vondenstein/gitblast.com
https://api.github.com/repos/vondenstein/gitblast.com
opened
CVE-2019-6286 (Medium) detected in opennms-opennms-source-22.0.1-1, node-sass-4.13.1.tgz
security vulnerability
## CVE-2019-6286 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.13.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gitblast.com/package.json</p> <p>Path to vulnerable library: /gitblast.com/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.13.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vondenstein/gitblast.com/commit/9ba0bd32c0c48dfe3dc35a01cf167196a1e46913">9ba0bd32c0c48dfe3dc35a01cf167196a1e46913</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::skip_over_scopes in prelexer.hpp when called from Sass::Parser::parse_import(), a similar issue to CVE-2018-11693. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6286>CVE-2019-6286</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286</a></p> <p>Release Date: 2019-08-06</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-6286 (Medium) detected in opennms-opennms-source-22.0.1-1, node-sass-4.13.1.tgz - ## CVE-2019-6286 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>node-sass-4.13.1.tgz</b></p></summary> <p> <details><summary><b>node-sass-4.13.1.tgz</b></p></summary> <p>Wrapper around libsass</p> <p>Library home page: <a href="https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz">https://registry.npmjs.org/node-sass/-/node-sass-4.13.1.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gitblast.com/package.json</p> <p>Path to vulnerable library: /gitblast.com/node_modules/node-sass/package.json</p> <p> Dependency Hierarchy: - :x: **node-sass-4.13.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vondenstein/gitblast.com/commit/9ba0bd32c0c48dfe3dc35a01cf167196a1e46913">9ba0bd32c0c48dfe3dc35a01cf167196a1e46913</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In LibSass 3.5.5, a heap-based buffer over-read exists in Sass::Prelexer::skip_over_scopes in prelexer.hpp when called from Sass::Parser::parse_import(), a similar issue to CVE-2018-11693. <p>Publish Date: 2019-01-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-6286>CVE-2019-6286</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-6286</a></p> <p>Release Date: 2019-08-06</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in opennms opennms source node sass tgz cve medium severity vulnerability vulnerable libraries node sass tgz node sass tgz wrapper around libsass library home page a href path to dependency file tmp ws scm gitblast com package json path to vulnerable library gitblast com node modules node sass package json dependency hierarchy x node sass tgz vulnerable library found in head commit a href vulnerability details in libsass a heap based buffer over read exists in sass prelexer skip over scopes in prelexer hpp when called from sass parser parse import a similar issue to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
3,768
15,833,528,410
IssuesEvent
2021-04-06 15:44:22
ethersphere/bee-status
https://api.github.com/repos/ethersphere/bee-status
closed
Add Repository Description
need:maintainers-input type:issue
Currently the repo doesn't contain a description which may make it harder for people to discern what the project is.
True
Add Repository Description - Currently the repo doesn't contain a description which may make it harder for people to discern what the project is.
main
add repository description currently the repo doesn t contain a description which may make it harder for people to discern what the project is
1
1,557
6,572,253,565
IssuesEvent
2017-09-11 00:39:03
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
Jira module is not sending valid json when operation = edit
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME jira (module) ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION Default configuration - no changes ##### OS / ENVIRONMENT Tested on Mac OS - however seems be global (not related with local platform) ##### SUMMARY According documentation jira module should allow create / modify items in Jira. Creating tickets is working fine, adding comments are working fine - however when operation == 'edit', module complain with error. ##### STEPS TO REPRODUCE This can be reproduce with `netcat` command and `jira` task. Running `ansible -vvvv` not helps in this case - not displaying outgoing json string which is sending to Jira URL. Repeat steps: On console 1 (this allows display json which is sending without sending this to real JIRA url) ``` # Run netcat and listen for connections nc -l 8080 ``` On another console run following playbook: ``` # Playbook: - name: Updating ticket on Jira (testing) hosts: localhost gather_facts: no vars: - soft : 'Ansible' - url : 'http://127.0.0.1:8080' # Server: nc -l 8080 - issue_test : 'Test' vars_files: # Copy 'common_jira' to 'custom_jira' and adjust with valid credentials - [ "custom_jira.yml", "common_jira.yml" ] tasks: - name: Edit issue - use arguments with fields jira: username: "{{ jira_user.name }}" password: "{{ jira_user.pass }}" uri: "{{ url }}" operation: "edit" issue: "200" register: issue_edit ignore_errors: yes args: fields: - labels: - autocreated - ansible - yaml # Create file with variables: custom_jira.yml --- jira_user: name: 'put_your_login_here' pass: 'put_your_pass_here' # Create basic hosts.txt (Empty - local host will be added automatically) # Run command ansible-playbook -i hosts.txt playbook.yml ``` ##### EXPECTED RESULTS After debugging with Jira following Json should be created during sending ``` {"fields": {"labels": ["autocreated", "ansible"]}} ``` Module should connect with JIRA via REST API and edit issue based on above fields. ##### ACTUAL RESULTS ``` TASK [Edit issue - use arguments with fields] ********************************** fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "HTTP Error 400: Bad Request"} # JSON: Difference of inverted commas. {"fields": "{'labels': ['autocreated', 'ansible']}"} ``` Runnning ansible -vvvv not shows outgoing json - this is why netcat was used to see what 'server' is receiving. Last info: Result from Jira based on wrong json: ``` HTTP/1.1 400 Bad Request ``` Result from Jira based on valid json: ``` HTTP/1.1 204 No Content # And Jira issue were updated with valid fields ```
True
Jira module is not sending valid json when operation = edit - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME jira (module) ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION Default configuration - no changes ##### OS / ENVIRONMENT Tested on Mac OS - however seems be global (not related with local platform) ##### SUMMARY According documentation jira module should allow create / modify items in Jira. Creating tickets is working fine, adding comments are working fine - however when operation == 'edit', module complain with error. ##### STEPS TO REPRODUCE This can be reproduce with `netcat` command and `jira` task. Running `ansible -vvvv` not helps in this case - not displaying outgoing json string which is sending to Jira URL. Repeat steps: On console 1 (this allows display json which is sending without sending this to real JIRA url) ``` # Run netcat and listen for connections nc -l 8080 ``` On another console run following playbook: ``` # Playbook: - name: Updating ticket on Jira (testing) hosts: localhost gather_facts: no vars: - soft : 'Ansible' - url : 'http://127.0.0.1:8080' # Server: nc -l 8080 - issue_test : 'Test' vars_files: # Copy 'common_jira' to 'custom_jira' and adjust with valid credentials - [ "custom_jira.yml", "common_jira.yml" ] tasks: - name: Edit issue - use arguments with fields jira: username: "{{ jira_user.name }}" password: "{{ jira_user.pass }}" uri: "{{ url }}" operation: "edit" issue: "200" register: issue_edit ignore_errors: yes args: fields: - labels: - autocreated - ansible - yaml # Create file with variables: custom_jira.yml --- jira_user: name: 'put_your_login_here' pass: 'put_your_pass_here' # Create basic hosts.txt (Empty - local host will be added automatically) # Run command ansible-playbook -i hosts.txt playbook.yml ``` ##### EXPECTED RESULTS After debugging with Jira following Json should be created during sending ``` {"fields": {"labels": ["autocreated", "ansible"]}} ``` Module should connect with JIRA via REST API and edit issue based on above fields. ##### ACTUAL RESULTS ``` TASK [Edit issue - use arguments with fields] ********************************** fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "HTTP Error 400: Bad Request"} # JSON: Difference of inverted commas. {"fields": "{'labels': ['autocreated', 'ansible']}"} ``` Runnning ansible -vvvv not shows outgoing json - this is why netcat was used to see what 'server' is receiving. Last info: Result from Jira based on wrong json: ``` HTTP/1.1 400 Bad Request ``` Result from Jira based on valid json: ``` HTTP/1.1 204 No Content # And Jira issue were updated with valid fields ```
main
jira module is not sending valid json when operation edit issue type bug report component name jira module ansible version ansible config file configured module search path default w o overrides configuration default configuration no changes os environment tested on mac os however seems be global not related with local platform summary according documentation jira module should allow create modify items in jira creating tickets is working fine adding comments are working fine however when operation edit module complain with error steps to reproduce this can be reproduce with netcat command and jira task running ansible vvvv not helps in this case not displaying outgoing json string which is sending to jira url repeat steps on console this allows display json which is sending without sending this to real jira url run netcat and listen for connections nc l on another console run following playbook playbook name updating ticket on jira testing hosts localhost gather facts no vars soft ansible url  server nc l issue test test vars files copy common jira to custom jira and adjust with valid credentials tasks name edit issue use arguments with fields jira username jira user name password jira user pass uri url operation edit issue register issue edit ignore errors yes args fields labels autocreated ansible yaml create file with variables custom jira yml jira user name put your login here pass put your pass here  create basic hosts txt empty local host will be added automatically  run command ansible playbook i hosts txt playbook yml expected results after debugging with jira following json should be created during sending fields labels module should connect with jira via rest api and edit issue based on above fields actual results task fatal failed changed false failed true msg http error bad request json difference of inverted commas fields labels runnning ansible vvvv not shows outgoing json this is why netcat was used to see what server is receiving last info result from jira based on wrong json http bad request result from jira based on valid json http no content  and jira issue were updated with valid fields
1
815,328
30,546,736,723
IssuesEvent
2023-07-20 04:59:54
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Sample web app breaks with below logs when the tomcat directory exists in a location where there is any space
Priority/High Severity/Major bug Affected/5.11.0 samples
**Describe the issue:** Sample web app breaks with below logs when **tomcat** (parent) directory exists in a location where there is any space. ``` 18-Jul-2023 19:09:05.981 INFO [main] org.wso2.sample.identity.jks.JKSLoader.contextInitialized Setting trust store path to : /Users/ratnajothy/Desktop/apache-tomcat%20%20%20%20%209.0.41/webapps/pickup-dispatch/WEB-INF/classes/wso2carbon.jks 18-Jul-2023 19:09:06.104 SEVERE [main] org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse Error while retrieving response. java.net.SocketException: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext) at javax.net.ssl.DefaultSSLSocketFactory.throwException(SSLSocketFactory.java:248) at javax.net.ssl.DefaultSSLSocketFactory.createSocket(SSLSocketFactory.java:255) at sun.net.www.protocol.https.HttpsClient.createSocket(HttpsClient.java:405) at sun.net.NetworkClient.doConnect(NetworkClient.java:162) at sun.net.www.http.HttpClient.openServer(HttpClient.java:463) at sun.net.www.http.HttpClient.openServer(HttpClient.java:558) at sun.net.www.protocol.https.HttpsClient.<init>(HttpsClient.java:264) at sun.net.www.protocol.https.HttpsClient.New(HttpsClient.java:367) at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.getNewHttpClient(AbstractDelegateHttpsURLConnection.java:203) at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1162) at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1056) at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:189) at sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1340) at sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1315) at sun.net.www.protocol.https.HttpsURLConnectionImpl.getOutputStream(HttpsURLConnectionImpl.java:264) at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:261) at org.wso2.samples.claims.manager.ClaimManagerProxy.getOIDCDialectNodes(ClaimManagerProxy.java:192) at org.wso2.samples.claims.manager.ClaimManagerProxy.<init>(ClaimManagerProxy.java:74) at org.wso2.sample.identity.oauth2.SampleContextEventListener.contextInitialized(SampleContextEventListener.java:50) at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4768) at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5230) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:728) at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:700) at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:696) at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1024) at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1911) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75) at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:112) at org.apache.catalina.startup.HostConfig.deployWARs(HostConfig.java:825) at org.apache.catalina.startup.HostConfig.deployApps(HostConfig.java:475) at org.apache.catalina.startup.HostConfig.start(HostConfig.java:1618) at org.apache.catalina.startup.HostConfig.lifecycleEvent(HostConfig.java:319) at org.apache.catalina.util.LifecycleBase.fireLifecycleEvent(LifecycleBase.java:123) at org.apache.catalina.util.LifecycleBase.setStateInternal(LifecycleBase.java:423) at org.apache.catalina.util.LifecycleBase.setState(LifecycleBase.java:366) at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:948) at org.apache.catalina.core.StandardHost.startInternal(StandardHost.java:835) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1398) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1388) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75) at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:134) at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:921) at org.apache.catalina.core.StandardEngine.startInternal(StandardEngine.java:263) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.StandardService.startInternal(StandardService.java:437) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.StandardServer.startInternal(StandardServer.java:934) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.startup.Catalina.start(Catalina.java:772) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:345) at org.apache.catalina.startup.Bootstrap.main(Bootstrap.java:476) Caused by: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext) at java.security.Provider$Service.newInstance(Provider.java:1617) at sun.security.jca.GetInstance.getInstance(GetInstance.java:236) at sun.security.jca.GetInstance.getInstance(GetInstance.java:164) at javax.net.ssl.SSLContext.getInstance(SSLContext.java:156) at javax.net.ssl.SSLContext.getDefault(SSLContext.java:96) at javax.net.ssl.SSLSocketFactory.getDefault(SSLSocketFactory.java:122) at javax.net.ssl.HttpsURLConnection.getDefaultSSLSocketFactory(HttpsURLConnection.java:332) at javax.net.ssl.HttpsURLConnection.<init>(HttpsURLConnection.java:289) at sun.net.www.protocol.https.HttpsURLConnectionImpl.<init>(HttpsURLConnectionImpl.java:99) at sun.net.www.protocol.https.Handler.openConnection(Handler.java:62) at sun.net.www.protocol.https.Handler.openConnection(Handler.java:57) at java.net.URL.openConnection(URL.java:1002) at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:251) ... 44 more Caused by: java.security.KeyStoreException: problem accessing trust store at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:73) at javax.net.ssl.TrustManagerFactory.init(TrustManagerFactory.java:250) at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.getTrustManagers(SSLContextImpl.java:1041) at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.<clinit>(SSLContextImpl.java:1011) at sun.security.ssl.SSLContextImpl$DefaultSSLContext.<init>(SSLContextImpl.java:1186) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at java.security.Provider$Service.newInstance(Provider.java:1595) ... 56 more Caused by: java.io.IOException: Keystore was tampered with, or password was incorrect at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:792) at sun.security.provider.JavaKeyStore$JKS.engineLoad(JavaKeyStore.java:57) at sun.security.provider.KeyStoreDelegator.engineLoad(KeyStoreDelegator.java:224) at sun.security.provider.JavaKeyStore$DualFormatJKS.engineLoad(JavaKeyStore.java:71) at java.security.KeyStore.load(KeyStore.java:1445) at sun.security.ssl.TrustStoreManager$TrustAnchorManager.loadKeyStore(TrustStoreManager.java:365) at sun.security.ssl.TrustStoreManager$TrustAnchorManager.getTrustedCerts(TrustStoreManager.java:313) at sun.security.ssl.TrustStoreManager.getTrustedCerts(TrustStoreManager.java:55) at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:49) ... 65 more Caused by: java.security.UnrecoverableKeyException: Password verification failed at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:790) ... 73 more [Fatal Error] :-1:-1: Premature end of file. ``` **How to reproduce:** 1. Rename the tomcat directory by giving a space in its name 2. Make sure there are no other webapps except for the Pickup-manager or Pickup-dispatch 3. Pickup-manager and Pickup-dispatch fails to deploy successfully **Expected behaviour:** <!-- A clear and concise description of what you expected to happen. --> Should work despite having spaces in the located path of the tomcat directory. **Suggestion** Fix the logic in obtaining resource path as URL when setting Trust Store Path : https://github.com/wso2/samples-is/blob/345db6b9228396d6ac74a8eb7dcb441dde7574cf/sso-samples/oidc-sso-sample/oidc-jks-loader/src/main/org/wso2/sample/identity/jks/JKSLoader.java#L58 _Fix should be for Pickup-manager and dispatch (both)_ **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: ....., 5.10.0, 5.11.0 - OS: Windows, Linux, Mac (Verified this issue with ORA JDK 8 and Tomcat 9.0.41)
1.0
Sample web app breaks with below logs when the tomcat directory exists in a location where there is any space - **Describe the issue:** Sample web app breaks with below logs when **tomcat** (parent) directory exists in a location where there is any space. ``` 18-Jul-2023 19:09:05.981 INFO [main] org.wso2.sample.identity.jks.JKSLoader.contextInitialized Setting trust store path to : /Users/ratnajothy/Desktop/apache-tomcat%20%20%20%20%209.0.41/webapps/pickup-dispatch/WEB-INF/classes/wso2carbon.jks 18-Jul-2023 19:09:06.104 SEVERE [main] org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse Error while retrieving response. java.net.SocketException: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext) at javax.net.ssl.DefaultSSLSocketFactory.throwException(SSLSocketFactory.java:248) at javax.net.ssl.DefaultSSLSocketFactory.createSocket(SSLSocketFactory.java:255) at sun.net.www.protocol.https.HttpsClient.createSocket(HttpsClient.java:405) at sun.net.NetworkClient.doConnect(NetworkClient.java:162) at sun.net.www.http.HttpClient.openServer(HttpClient.java:463) at sun.net.www.http.HttpClient.openServer(HttpClient.java:558) at sun.net.www.protocol.https.HttpsClient.<init>(HttpsClient.java:264) at sun.net.www.protocol.https.HttpsClient.New(HttpsClient.java:367) at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.getNewHttpClient(AbstractDelegateHttpsURLConnection.java:203) at sun.net.www.protocol.http.HttpURLConnection.plainConnect0(HttpURLConnection.java:1162) at sun.net.www.protocol.http.HttpURLConnection.plainConnect(HttpURLConnection.java:1056) at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:189) at sun.net.www.protocol.http.HttpURLConnection.getOutputStream0(HttpURLConnection.java:1340) at sun.net.www.protocol.http.HttpURLConnection.getOutputStream(HttpURLConnection.java:1315) at sun.net.www.protocol.https.HttpsURLConnectionImpl.getOutputStream(HttpsURLConnectionImpl.java:264) at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:261) at org.wso2.samples.claims.manager.ClaimManagerProxy.getOIDCDialectNodes(ClaimManagerProxy.java:192) at org.wso2.samples.claims.manager.ClaimManagerProxy.<init>(ClaimManagerProxy.java:74) at org.wso2.sample.identity.oauth2.SampleContextEventListener.contextInitialized(SampleContextEventListener.java:50) at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:4768) at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5230) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:728) at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:700) at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:696) at org.apache.catalina.startup.HostConfig.deployWAR(HostConfig.java:1024) at org.apache.catalina.startup.HostConfig$DeployWar.run(HostConfig.java:1911) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75) at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:112) at org.apache.catalina.startup.HostConfig.deployWARs(HostConfig.java:825) at org.apache.catalina.startup.HostConfig.deployApps(HostConfig.java:475) at org.apache.catalina.startup.HostConfig.start(HostConfig.java:1618) at org.apache.catalina.startup.HostConfig.lifecycleEvent(HostConfig.java:319) at org.apache.catalina.util.LifecycleBase.fireLifecycleEvent(LifecycleBase.java:123) at org.apache.catalina.util.LifecycleBase.setStateInternal(LifecycleBase.java:423) at org.apache.catalina.util.LifecycleBase.setState(LifecycleBase.java:366) at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:948) at org.apache.catalina.core.StandardHost.startInternal(StandardHost.java:835) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1398) at org.apache.catalina.core.ContainerBase$StartChild.call(ContainerBase.java:1388) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at org.apache.tomcat.util.threads.InlineExecutorService.execute(InlineExecutorService.java:75) at java.util.concurrent.AbstractExecutorService.submit(AbstractExecutorService.java:134) at org.apache.catalina.core.ContainerBase.startInternal(ContainerBase.java:921) at org.apache.catalina.core.StandardEngine.startInternal(StandardEngine.java:263) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.StandardService.startInternal(StandardService.java:437) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.core.StandardServer.startInternal(StandardServer.java:934) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:183) at org.apache.catalina.startup.Catalina.start(Catalina.java:772) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.catalina.startup.Bootstrap.start(Bootstrap.java:345) at org.apache.catalina.startup.Bootstrap.main(Bootstrap.java:476) Caused by: java.security.NoSuchAlgorithmException: Error constructing implementation (algorithm: Default, provider: SunJSSE, class: sun.security.ssl.SSLContextImpl$DefaultSSLContext) at java.security.Provider$Service.newInstance(Provider.java:1617) at sun.security.jca.GetInstance.getInstance(GetInstance.java:236) at sun.security.jca.GetInstance.getInstance(GetInstance.java:164) at javax.net.ssl.SSLContext.getInstance(SSLContext.java:156) at javax.net.ssl.SSLContext.getDefault(SSLContext.java:96) at javax.net.ssl.SSLSocketFactory.getDefault(SSLSocketFactory.java:122) at javax.net.ssl.HttpsURLConnection.getDefaultSSLSocketFactory(HttpsURLConnection.java:332) at javax.net.ssl.HttpsURLConnection.<init>(HttpsURLConnection.java:289) at sun.net.www.protocol.https.HttpsURLConnectionImpl.<init>(HttpsURLConnectionImpl.java:99) at sun.net.www.protocol.https.Handler.openConnection(Handler.java:62) at sun.net.www.protocol.https.Handler.openConnection(Handler.java:57) at java.net.URL.openConnection(URL.java:1002) at org.wso2.samples.claims.manager.ClaimManagerProxy.getClaimManagementResponse(ClaimManagerProxy.java:251) ... 44 more Caused by: java.security.KeyStoreException: problem accessing trust store at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:73) at javax.net.ssl.TrustManagerFactory.init(TrustManagerFactory.java:250) at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.getTrustManagers(SSLContextImpl.java:1041) at sun.security.ssl.SSLContextImpl$DefaultManagersHolder.<clinit>(SSLContextImpl.java:1011) at sun.security.ssl.SSLContextImpl$DefaultSSLContext.<init>(SSLContextImpl.java:1186) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:423) at java.security.Provider$Service.newInstance(Provider.java:1595) ... 56 more Caused by: java.io.IOException: Keystore was tampered with, or password was incorrect at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:792) at sun.security.provider.JavaKeyStore$JKS.engineLoad(JavaKeyStore.java:57) at sun.security.provider.KeyStoreDelegator.engineLoad(KeyStoreDelegator.java:224) at sun.security.provider.JavaKeyStore$DualFormatJKS.engineLoad(JavaKeyStore.java:71) at java.security.KeyStore.load(KeyStore.java:1445) at sun.security.ssl.TrustStoreManager$TrustAnchorManager.loadKeyStore(TrustStoreManager.java:365) at sun.security.ssl.TrustStoreManager$TrustAnchorManager.getTrustedCerts(TrustStoreManager.java:313) at sun.security.ssl.TrustStoreManager.getTrustedCerts(TrustStoreManager.java:55) at sun.security.ssl.TrustManagerFactoryImpl.engineInit(TrustManagerFactoryImpl.java:49) ... 65 more Caused by: java.security.UnrecoverableKeyException: Password verification failed at sun.security.provider.JavaKeyStore.engineLoad(JavaKeyStore.java:790) ... 73 more [Fatal Error] :-1:-1: Premature end of file. ``` **How to reproduce:** 1. Rename the tomcat directory by giving a space in its name 2. Make sure there are no other webapps except for the Pickup-manager or Pickup-dispatch 3. Pickup-manager and Pickup-dispatch fails to deploy successfully **Expected behaviour:** <!-- A clear and concise description of what you expected to happen. --> Should work despite having spaces in the located path of the tomcat directory. **Suggestion** Fix the logic in obtaining resource path as URL when setting Trust Store Path : https://github.com/wso2/samples-is/blob/345db6b9228396d6ac74a8eb7dcb441dde7574cf/sso-samples/oidc-sso-sample/oidc-jks-loader/src/main/org/wso2/sample/identity/jks/JKSLoader.java#L58 _Fix should be for Pickup-manager and dispatch (both)_ **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: ....., 5.10.0, 5.11.0 - OS: Windows, Linux, Mac (Verified this issue with ORA JDK 8 and Tomcat 9.0.41)
non_main
sample web app breaks with below logs when the tomcat directory exists in a location where there is any space describe the issue sample web app breaks with below logs when tomcat parent directory exists in a location where there is any space jul info org sample identity jks jksloader contextinitialized setting trust store path to users ratnajothy desktop apache tomcat webapps pickup dispatch web inf classes jks jul severe org samples claims manager claimmanagerproxy getclaimmanagementresponse error while retrieving response java net socketexception java security nosuchalgorithmexception error constructing implementation algorithm default provider sunjsse class sun security ssl sslcontextimpl defaultsslcontext at javax net ssl defaultsslsocketfactory throwexception sslsocketfactory java at javax net ssl defaultsslsocketfactory createsocket sslsocketfactory java at sun net at sun net networkclient doconnect networkclient java at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at sun net at org samples claims manager claimmanagerproxy getclaimmanagementresponse claimmanagerproxy java at org samples claims manager claimmanagerproxy getoidcdialectnodes claimmanagerproxy java at org samples claims manager claimmanagerproxy claimmanagerproxy java at org sample identity samplecontexteventlistener contextinitialized samplecontexteventlistener java at org apache catalina core standardcontext listenerstart standardcontext java at org apache catalina core standardcontext startinternal standardcontext java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core containerbase addchildinternal containerbase java at org apache catalina core containerbase addchild containerbase java at org apache catalina core standardhost addchild standardhost java at org apache catalina startup hostconfig deploywar hostconfig java at org apache catalina startup hostconfig deploywar run hostconfig java at java util concurrent executors runnableadapter call executors java at java util concurrent futuretask run futuretask java at org apache tomcat util threads inlineexecutorservice execute inlineexecutorservice java at java util concurrent abstractexecutorservice submit abstractexecutorservice java at org apache catalina startup hostconfig deploywars hostconfig java at org apache catalina startup hostconfig deployapps hostconfig java at org apache catalina startup hostconfig start hostconfig java at org apache catalina startup hostconfig lifecycleevent hostconfig java at org apache catalina util lifecyclebase firelifecycleevent lifecyclebase java at org apache catalina util lifecyclebase setstateinternal lifecyclebase java at org apache catalina util lifecyclebase setstate lifecyclebase java at org apache catalina core containerbase startinternal containerbase java at org apache catalina core standardhost startinternal standardhost java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core containerbase startchild call containerbase java at org apache catalina core containerbase startchild call containerbase java at java util concurrent futuretask run futuretask java at org apache tomcat util threads inlineexecutorservice execute inlineexecutorservice java at java util concurrent abstractexecutorservice submit abstractexecutorservice java at org apache catalina core containerbase startinternal containerbase java at org apache catalina core standardengine startinternal standardengine java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core standardservice startinternal standardservice java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina core standardserver startinternal standardserver java at org apache catalina util lifecyclebase start lifecyclebase java at org apache catalina startup catalina start catalina java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache catalina startup bootstrap start bootstrap java at org apache catalina startup bootstrap main bootstrap java caused by java security nosuchalgorithmexception error constructing implementation algorithm default provider sunjsse class sun security ssl sslcontextimpl defaultsslcontext at java security provider service newinstance provider java at sun security jca getinstance getinstance getinstance java at sun security jca getinstance getinstance getinstance java at javax net ssl sslcontext getinstance sslcontext java at javax net ssl sslcontext getdefault sslcontext java at javax net ssl sslsocketfactory getdefault sslsocketfactory java at javax net ssl httpsurlconnection getdefaultsslsocketfactory httpsurlconnection java at javax net ssl httpsurlconnection httpsurlconnection java at sun net at sun net at sun net at java net url openconnection url java at org samples claims manager claimmanagerproxy getclaimmanagementresponse claimmanagerproxy java more caused by java security keystoreexception problem accessing trust store at sun security ssl trustmanagerfactoryimpl engineinit trustmanagerfactoryimpl java at javax net ssl trustmanagerfactory init trustmanagerfactory java at sun security ssl sslcontextimpl defaultmanagersholder gettrustmanagers sslcontextimpl java at sun security ssl sslcontextimpl defaultmanagersholder sslcontextimpl java at sun security ssl sslcontextimpl defaultsslcontext sslcontextimpl java at sun reflect nativeconstructoraccessorimpl native method at sun reflect nativeconstructoraccessorimpl newinstance nativeconstructoraccessorimpl java at sun reflect delegatingconstructoraccessorimpl newinstance delegatingconstructoraccessorimpl java at java lang reflect constructor newinstance constructor java at java security provider service newinstance provider java more caused by java io ioexception keystore was tampered with or password was incorrect at sun security provider javakeystore engineload javakeystore java at sun security provider javakeystore jks engineload javakeystore java at sun security provider keystoredelegator engineload keystoredelegator java at sun security provider javakeystore dualformatjks engineload javakeystore java at java security keystore load keystore java at sun security ssl truststoremanager trustanchormanager loadkeystore truststoremanager java at sun security ssl truststoremanager trustanchormanager gettrustedcerts truststoremanager java at sun security ssl truststoremanager gettrustedcerts truststoremanager java at sun security ssl trustmanagerfactoryimpl engineinit trustmanagerfactoryimpl java more caused by java security unrecoverablekeyexception password verification failed at sun security provider javakeystore engineload javakeystore java more premature end of file how to reproduce rename the tomcat directory by giving a space in its name make sure there are no other webapps except for the pickup manager or pickup dispatch pickup manager and pickup dispatch fails to deploy successfully expected behaviour should work despite having spaces in the located path of the tomcat directory suggestion fix the logic in obtaining resource path as url when setting trust store path fix should be for pickup manager and dispatch both environment information please complete the following information remove any unnecessary fields product version os windows linux mac verified this issue with ora jdk and tomcat
0
244,921
18,769,439,701
IssuesEvent
2021-11-06 15:06:07
AY2122S1-CS2103T-T15-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-T15-2/tp
closed
[PE-D] Visuals in UG
type.Documentation
Perhaps could have included more visuals for things like task and event commands that aren't covered in persons command. <!--session: 1635494514989-f88cdb01-2d06-45de-8a15-2a40c14ffa30--> <!--Version: Web v3.4.1--> ------------- Labels: `type.DocumentationBug` `severity.VeryLow` original: Shruthi0310/ped#14
1.0
[PE-D] Visuals in UG - Perhaps could have included more visuals for things like task and event commands that aren't covered in persons command. <!--session: 1635494514989-f88cdb01-2d06-45de-8a15-2a40c14ffa30--> <!--Version: Web v3.4.1--> ------------- Labels: `type.DocumentationBug` `severity.VeryLow` original: Shruthi0310/ped#14
non_main
visuals in ug perhaps could have included more visuals for things like task and event commands that aren t covered in persons command labels type documentationbug severity verylow original ped
0
1,169
5,087,807,764
IssuesEvent
2016-12-31 09:47:21
openwrt/packages
https://api.github.com/repos/openwrt/packages
closed
transmission has wrong maintainer
waiting for maintainer
net/transmission still has the maintainer from the old packages feed, which does not match the pull request initiator. @mstorchak since you added the package, please change the maintainer to yourself.
True
transmission has wrong maintainer - net/transmission still has the maintainer from the old packages feed, which does not match the pull request initiator. @mstorchak since you added the package, please change the maintainer to yourself.
main
transmission has wrong maintainer net transmission still has the maintainer from the old packages feed which does not match the pull request initiator mstorchak since you added the package please change the maintainer to yourself
1
142,567
19,099,352,334
IssuesEvent
2021-11-29 20:26:23
harrinry/stream-chat-react
https://api.github.com/repos/harrinry/stream-chat-react
opened
CVE-2021-32723 (Medium) detected in prismjs-1.22.0.tgz
security vulnerability
## CVE-2021-32723 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.22.0.tgz</b></p></summary> <p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p> <p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz</a></p> <p>Path to dependency file: stream-chat-react/package.json</p> <p>Path to vulnerable library: stream-chat-react/node_modules/prismjs/package.json</p> <p> Dependency Hierarchy: - react-styleguidist-11.1.4.tgz (Root Library) - :x: **prismjs-1.22.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/harrinry/stream-chat-react/commit/a80df0ece7581dc566b7c05d5f5ae6ecc66b730f">a80df0ece7581dc566b7c05d5f5ae6ecc66b730f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text. <p>Publish Date: 2021-06-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723>CVE-2021-32723</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg">https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg</a></p> <p>Release Date: 2021-06-28</p> <p>Fix Resolution: prismjs - 1.24.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.22.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-styleguidist:11.1.4;prismjs:1.22.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"prismjs - 1.24.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32723","vulnerabilityDetails":"Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2021-32723 (Medium) detected in prismjs-1.22.0.tgz - ## CVE-2021-32723 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.22.0.tgz</b></p></summary> <p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p> <p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.22.0.tgz</a></p> <p>Path to dependency file: stream-chat-react/package.json</p> <p>Path to vulnerable library: stream-chat-react/node_modules/prismjs/package.json</p> <p> Dependency Hierarchy: - react-styleguidist-11.1.4.tgz (Root Library) - :x: **prismjs-1.22.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/harrinry/stream-chat-react/commit/a80df0ece7581dc566b7c05d5f5ae6ecc66b730f">a80df0ece7581dc566b7c05d5f5ae6ecc66b730f</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text. <p>Publish Date: 2021-06-28 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723>CVE-2021-32723</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg">https://github.com/PrismJS/prism/security/advisories/GHSA-gj77-59wh-66hg</a></p> <p>Release Date: 2021-06-28</p> <p>Fix Resolution: prismjs - 1.24.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.22.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-styleguidist:11.1.4;prismjs:1.22.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"prismjs - 1.24.0","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-32723","vulnerabilityDetails":"Prism is a syntax highlighting library. Some languages before 1.24.0 are vulnerable to Regular Expression Denial of Service (ReDoS). When Prism is used to highlight untrusted (user-given) text, an attacker can craft a string that will take a very very long time to highlight. This problem has been fixed in Prism v1.24. As a workaround, do not use ASCIIDoc or ERB to highlight untrusted text. Other languages are not affected and can be used to highlight untrusted text.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-32723","cvss3Severity":"medium","cvss3Score":"6.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"Required","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in prismjs tgz cve medium severity vulnerability vulnerable library prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file stream chat react package json path to vulnerable library stream chat react node modules prismjs package json dependency hierarchy react styleguidist tgz root library x prismjs tgz vulnerable library found in head commit a href found in base branch master vulnerability details prism is a syntax highlighting library some languages before are vulnerable to regular expression denial of service redos when prism is used to highlight untrusted user given text an attacker can craft a string that will take a very very long time to highlight this problem has been fixed in prism as a workaround do not use asciidoc or erb to highlight untrusted text other languages are not affected and can be used to highlight untrusted text publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution prismjs isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree react styleguidist prismjs isminimumfixversionavailable true minimumfixversion prismjs isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails prism is a syntax highlighting library some languages before are vulnerable to regular expression denial of service redos when prism is used to highlight untrusted user given text an attacker can craft a string that will take a very very long time to highlight this problem has been fixed in prism as a workaround do not use asciidoc or erb to highlight untrusted text other languages are not affected and can be used to highlight untrusted text vulnerabilityurl
0
943
4,674,565,765
IssuesEvent
2016-10-07 02:00:43
tomchentw/react-google-maps
https://api.github.com/repos/tomchentw/react-google-maps
closed
InfoWindow clears on update
CALL_FOR_MAINTAINERS
Hi guys, I'm currently using `react-google-maps` to create a public transit map and I'm using the `InfoWindow` component to display stop timetables. When I open the `InfoWindow` component through a marker, the window opens as normal, however, I am using JQuery and Ajax to load the data from a server into the React element. When the Ajax request finishes and the data is stored in the React element state, the `InfoWindow` clears and empties. My code is here: https://gist.github.com/uniquoooo/c80e6bda7bc901d8a53753c2d10e15a1 I'm not sure why this is happening, could someone look at my code and see if I am doing anything wrong? Thanks
True
InfoWindow clears on update - Hi guys, I'm currently using `react-google-maps` to create a public transit map and I'm using the `InfoWindow` component to display stop timetables. When I open the `InfoWindow` component through a marker, the window opens as normal, however, I am using JQuery and Ajax to load the data from a server into the React element. When the Ajax request finishes and the data is stored in the React element state, the `InfoWindow` clears and empties. My code is here: https://gist.github.com/uniquoooo/c80e6bda7bc901d8a53753c2d10e15a1 I'm not sure why this is happening, could someone look at my code and see if I am doing anything wrong? Thanks
main
infowindow clears on update hi guys i m currently using react google maps to create a public transit map and i m using the infowindow component to display stop timetables when i open the infowindow component through a marker the window opens as normal however i am using jquery and ajax to load the data from a server into the react element when the ajax request finishes and the data is stored in the react element state the infowindow clears and empties my code is here i m not sure why this is happening could someone look at my code and see if i am doing anything wrong thanks
1
545,403
15,949,885,119
IssuesEvent
2021-04-15 08:00:01
mobigen/IRIS-BigData-Platform
https://api.github.com/repos/mobigen/IRIS-BigData-Platform
closed
[Studio][개발요청] 변수에 저장된 text type 값들을 "예제1", "예제2", "예제3" 처럼 따옴표와 콤마로 만드는 함수 또는 명령어
#DSL #Studio Priority: P0 Status: Backlog
## 기능 요청 ## b-iris.mobigen.com demo 계정 - studio : EDU_예제_그룹_여집합구하기 http://b-iris.mobigen.com:80/studio/exported/89d7a85f1c784c24ac0306f366dfc80623435f48882e47ae9704e66a76b26d5e 에서 - "common_data_3" <img width="479" alt="스크린샷 2021-03-22 오전 8 29 48" src="https://user-images.githubusercontent.com/5126497/111924633-e9884600-8ae8-11eb-87cf-d76ef35b8177.png"> - "common_data_2" <img width="478" alt="스크린샷 2021-03-22 오전 8 31 48" src="https://user-images.githubusercontent.com/5126497/111924683-0fade600-8ae9-11eb-9020-c18d4e37921f.png"> - 원하는 구문 ``` * | sql "SELECT explode(array_except( array( 'setosa','virginica' ), array( 'versicolor','virginica' ) ) ) as my_uniqSpecies from angora" ``` - 변수명에서는 문자열에 따옴표가 있지만, studio 에서 사용할 때는 **따옴표가 삭제** 됩니다. 그래서 nunjucks 로 for loop 구문을 써서 따옴표를 붙이는 작업을 해야 합니다. ``` * | {% set my_string = '' %} {% for i in common_data_3.results -%} {% set my_string = my_string ~ "'" ~ i ~ "'," %} {%- endfor %} {% set my_string = my_string.slice(0,-1) %} {% set my_string2 = '' %} {% for i in common_data_2.results -%} {% set my_string2 = my_string2 ~ "'" ~ i ~ "'," %} {%- endfor %} {% set my_string2 = my_string2.slice(0,-1) %} sql "SELECT explode(array_except( array( {{ my_string }} ), array( {{ my_string2 }} ) ) ) as my_uniqSpecies from angora" ``` ## 원하는 솔루션 설명 ## nunjucks 없이 변수에 TYPE : text 이면 따옴표가 붙은 채로 사용할 수 있는 기능 검색어에서 - 현재 ``` * | where A in ( ${common_data_3.results} ) => * | where A in ( setosa,virginica ) ``` - 요청 ``` * | where A in ( ${common_data_3.results} ) => * | where A in ( 'setosa', 'virginica' ) ``` 로 사용할 수 있는 방법이 필요합니다. ## 고려한 다른 대안 ## { 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. } ## 기타 ## 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다.
1.0
[Studio][개발요청] 변수에 저장된 text type 값들을 "예제1", "예제2", "예제3" 처럼 따옴표와 콤마로 만드는 함수 또는 명령어 - ## 기능 요청 ## b-iris.mobigen.com demo 계정 - studio : EDU_예제_그룹_여집합구하기 http://b-iris.mobigen.com:80/studio/exported/89d7a85f1c784c24ac0306f366dfc80623435f48882e47ae9704e66a76b26d5e 에서 - "common_data_3" <img width="479" alt="스크린샷 2021-03-22 오전 8 29 48" src="https://user-images.githubusercontent.com/5126497/111924633-e9884600-8ae8-11eb-87cf-d76ef35b8177.png"> - "common_data_2" <img width="478" alt="스크린샷 2021-03-22 오전 8 31 48" src="https://user-images.githubusercontent.com/5126497/111924683-0fade600-8ae9-11eb-9020-c18d4e37921f.png"> - 원하는 구문 ``` * | sql "SELECT explode(array_except( array( 'setosa','virginica' ), array( 'versicolor','virginica' ) ) ) as my_uniqSpecies from angora" ``` - 변수명에서는 문자열에 따옴표가 있지만, studio 에서 사용할 때는 **따옴표가 삭제** 됩니다. 그래서 nunjucks 로 for loop 구문을 써서 따옴표를 붙이는 작업을 해야 합니다. ``` * | {% set my_string = '' %} {% for i in common_data_3.results -%} {% set my_string = my_string ~ "'" ~ i ~ "'," %} {%- endfor %} {% set my_string = my_string.slice(0,-1) %} {% set my_string2 = '' %} {% for i in common_data_2.results -%} {% set my_string2 = my_string2 ~ "'" ~ i ~ "'," %} {%- endfor %} {% set my_string2 = my_string2.slice(0,-1) %} sql "SELECT explode(array_except( array( {{ my_string }} ), array( {{ my_string2 }} ) ) ) as my_uniqSpecies from angora" ``` ## 원하는 솔루션 설명 ## nunjucks 없이 변수에 TYPE : text 이면 따옴표가 붙은 채로 사용할 수 있는 기능 검색어에서 - 현재 ``` * | where A in ( ${common_data_3.results} ) => * | where A in ( setosa,virginica ) ``` - 요청 ``` * | where A in ( ${common_data_3.results} ) => * | where A in ( 'setosa', 'virginica' ) ``` 로 사용할 수 있는 방법이 필요합니다. ## 고려한 다른 대안 ## { 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다. } ## 기타 ## 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다.
non_main
변수에 저장된 text type 값들을 처럼 따옴표와 콤마로 만드는 함수 또는 명령어 기능 요청 b iris mobigen com demo 계정 studio edu 예제 그룹 여집합구하기 에서 common data img width alt 스크린샷 오전 src common data img width alt 스크린샷 오전 src 원하는 구문 sql select explode array except array setosa virginica array versicolor virginica as my uniqspecies from angora 변수명에서는 문자열에 따옴표가 있지만 studio 에서 사용할 때는 따옴표가 삭제 됩니다 그래서 nunjucks 로 for loop 구문을 써서 따옴표를 붙이는 작업을 해야 합니다 set my string for i in common data results set my string my string i endfor set my string my string slice set my for i in common data results set my my i endfor set my my slice sql select explode array except array my string array my as my uniqspecies from angora 원하는 솔루션 설명 nunjucks 없이 변수에 type text 이면 따옴표가 붙은 채로 사용할 수 있는 기능 검색어에서 현재 where a in common data results where a in setosa virginica 요청 where a in common data results where a in setosa virginica 로 사용할 수 있는 방법이 필요합니다 고려한 다른 대안 고려한 대체 솔루션이나 기능에 대한 명확하고 간결한 설명 부탁드립니다 기타 기능 요청에 대한 다른 의견 또는 스크린샷을 여기에 모두 추가 부탁드립니다
0
35
2,582,640,188
IssuesEvent
2015-02-15 13:50:23
0robustus1/savage
https://api.github.com/repos/0robustus1/savage
opened
switch test-framework
maintainability
I don't really like the standard unit-testing syntax. We should take a look at other clojure libraries. Maybe try out Midje?
True
switch test-framework - I don't really like the standard unit-testing syntax. We should take a look at other clojure libraries. Maybe try out Midje?
main
switch test framework i don t really like the standard unit testing syntax we should take a look at other clojure libraries maybe try out midje
1
270,465
8,460,817,487
IssuesEvent
2018-10-22 19:57:14
the-better-gibgab/gibgab-frontend
https://api.github.com/repos/the-better-gibgab/gibgab-frontend
opened
View Post
Priority: High Status: Available
### View Post As a user I want to view the content of a post So that I can interact with the social network #### Acceptance Criteria Given : Post data When : The user chooses a post Then : Post and comment data is displayed to the user #### Story Points Points: 4
1.0
View Post - ### View Post As a user I want to view the content of a post So that I can interact with the social network #### Acceptance Criteria Given : Post data When : The user chooses a post Then : Post and comment data is displayed to the user #### Story Points Points: 4
non_main
view post view post as a user i want to view the content of a post so that i can interact with the social network acceptance criteria given post data when the user chooses a post then post and comment data is displayed to the user story points points
0
5,219
26,479,630,348
IssuesEvent
2023-01-17 13:49:16
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Simplify the CI/CD workflows to use the builtin Maven cache feature in setup-java action that now supports it
bug maintainability CI/CD
We can now remove the actions/cache steps and replace them with a config setting under `setup-java` action. It automatically uses actions/cache already under the hood and will create the fileHash for the Maven `**/pom.xml` automatically. Documention: https://github.com/actions/setup-java/blob/main/README.md#caching-packages-dependencies Usage such as: ``` steps: - uses: actions/checkout@v3 - uses: actions/setup-java@v3 with: distribution: 'temurin' java-version: '17' cache: 'maven' - name: Build with Maven run: mvn -B package --file pom.xml ``` ### To Reproduce Steps to reproduce the behavior: 1. Run a PR workflow. ### Current Results Extra steps not need for `actions/cache` separately. ### Expected Behavior cleaner and less verbose build/test output in logs.
True
Simplify the CI/CD workflows to use the builtin Maven cache feature in setup-java action that now supports it - We can now remove the actions/cache steps and replace them with a config setting under `setup-java` action. It automatically uses actions/cache already under the hood and will create the fileHash for the Maven `**/pom.xml` automatically. Documention: https://github.com/actions/setup-java/blob/main/README.md#caching-packages-dependencies Usage such as: ``` steps: - uses: actions/checkout@v3 - uses: actions/setup-java@v3 with: distribution: 'temurin' java-version: '17' cache: 'maven' - name: Build with Maven run: mvn -B package --file pom.xml ``` ### To Reproduce Steps to reproduce the behavior: 1. Run a PR workflow. ### Current Results Extra steps not need for `actions/cache` separately. ### Expected Behavior cleaner and less verbose build/test output in logs.
main
simplify the ci cd workflows to use the builtin maven cache feature in setup java action that now supports it we can now remove the actions cache steps and replace them with a config setting under setup java action it automatically uses actions cache already under the hood and will create the filehash for the maven pom xml automatically documention usage such as steps uses actions checkout uses actions setup java with distribution temurin java version cache maven name build with maven run mvn b package file pom xml to reproduce steps to reproduce the behavior run a pr workflow current results extra steps not need for actions cache separately expected behavior cleaner and less verbose build test output in logs
1
4,739
24,460,271,323
IssuesEvent
2022-10-07 10:28:21
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
Numbers in templates are still getting localized
bug 🦠 engineering Maintain
This is caused by `L10N=True`: even though we removed `USE_THOUSAND_SEPARATOR`, which turns this behaviour off for English/default, the `L10N` setting "overrules" that and turns on silent number reformatting with even less control. There is the `{{ value | unlocalize }}` template filter from the `l10n` tags that we can use, but it would be fantastic if we can find out how to actually turns this off for every locale, except where `intcomma` is explicitly used.
True
Numbers in templates are still getting localized - This is caused by `L10N=True`: even though we removed `USE_THOUSAND_SEPARATOR`, which turns this behaviour off for English/default, the `L10N` setting "overrules" that and turns on silent number reformatting with even less control. There is the `{{ value | unlocalize }}` template filter from the `l10n` tags that we can use, but it would be fantastic if we can find out how to actually turns this off for every locale, except where `intcomma` is explicitly used.
main
numbers in templates are still getting localized this is caused by true even though we removed use thousand separator which turns this behaviour off for english default the setting overrules that and turns on silent number reformatting with even less control there is the value unlocalize template filter from the tags that we can use but it would be fantastic if we can find out how to actually turns this off for every locale except where intcomma is explicitly used
1
157,373
12,371,339,342
IssuesEvent
2020-05-18 18:23:14
fitbenchmarking/fitbenchmarking
https://api.github.com/repos/fitbenchmarking/fitbenchmarking
opened
Better windows testing in test_parsers.py
Maintenance Priority - low Testing
**Which module and class/method/function does this relate to?** This relates to test_parsers.py. Tests on windows limited to `formats = ['nist']` Extend this to all combinations that are supported by basic install (i.e. pip install .) in some way Relates to #526
1.0
Better windows testing in test_parsers.py - **Which module and class/method/function does this relate to?** This relates to test_parsers.py. Tests on windows limited to `formats = ['nist']` Extend this to all combinations that are supported by basic install (i.e. pip install .) in some way Relates to #526
non_main
better windows testing in test parsers py which module and class method function does this relate to this relates to test parsers py tests on windows limited to formats extend this to all combinations that are supported by basic install i e pip install in some way relates to
0
99,949
12,491,085,395
IssuesEvent
2020-06-01 02:47:25
WestpacGEL/GEL
https://api.github.com/repos/WestpacGEL/GEL
closed
Component: Modal
KTM-Design bug modal
## Features - [x] Sizing - [x] Examples - [ ] Jest tests - [ ] Cypress tests - [x] Docs ## To discuss - Whether separate modal header and body components are necessary - Use of GEL Button for the close icon ## Feedback from Kate's review - **modal/default** - Doesn't seem to have all styling applied, ie Hero coloured header separator line. _JPS: This has now been fixed... In the Modal PR (https://github.com/WestpacGEL/GEL/pull/199), awaiting merge) ## Jonny review - Currently in @jaortiz ’s PR https://github.com/WestpacGEL/GEL/pull/199. Will wait before giving to Intopia. _JPS: now merged_
1.0
Component: Modal - ## Features - [x] Sizing - [x] Examples - [ ] Jest tests - [ ] Cypress tests - [x] Docs ## To discuss - Whether separate modal header and body components are necessary - Use of GEL Button for the close icon ## Feedback from Kate's review - **modal/default** - Doesn't seem to have all styling applied, ie Hero coloured header separator line. _JPS: This has now been fixed... In the Modal PR (https://github.com/WestpacGEL/GEL/pull/199), awaiting merge) ## Jonny review - Currently in @jaortiz ’s PR https://github.com/WestpacGEL/GEL/pull/199. Will wait before giving to Intopia. _JPS: now merged_
non_main
component modal features sizing examples jest tests cypress tests docs to discuss whether separate modal header and body components are necessary use of gel button for the close icon feedback from kate s review modal default doesn t seem to have all styling applied ie hero coloured header separator line jps this has now been fixed in the modal pr awaiting merge jonny review currently in jaortiz ’s pr will wait before giving to intopia jps now merged
0
3,674
15,036,029,509
IssuesEvent
2021-02-02 14:48:38
IITIDIDX597/sp_2021_team1
https://api.github.com/repos/IITIDIDX597/sp_2021_team1
opened
Tagging articles for better search
Epic: 5 Maintaining the system Story Week 3
**Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care. **Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform. **Sub-Hill Statements:** 1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments. ### **Story Details:** As an: administrator I want: to be able to tag the article with various labels according to the topic So that: it's easier for people to search
True
Tagging articles for better search - **Project Goal:** S Lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way, while at the same time foster deeper learning experiences in order to deliver better AbilityLab Patient care. **Hill Statement:** Individual Clinicians can reference relevant, continuously evolving information for their patient's therapy needs to self-manage their approach & patient care plan development in a single platform. **Sub-Hill Statements:** 1. The learning platform will be routinely updated with S Lab's own research advancements, as well as outside discoveries and best practices developed for rehabilitation treatments. ### **Story Details:** As an: administrator I want: to be able to tag the article with various labels according to the topic So that: it's easier for people to search
main
tagging articles for better search project goal s lab is a tailored integrative learning and collaboration platform for clinicians that combines the latest research and tacit knowledge gained from experience in a practical way while at the same time foster deeper learning experiences in order to deliver better abilitylab patient care hill statement individual clinicians can reference relevant continuously evolving information for their patient s therapy needs to self manage their approach patient care plan development in a single platform sub hill statements the learning platform will be routinely updated with s lab s own research advancements as well as outside discoveries and best practices developed for rehabilitation treatments story details as an administrator i want to be able to tag the article with various labels according to the topic so that it s easier for people to search
1
2,887
10,319,603,894
IssuesEvent
2019-08-30 17:59:53
backdrop-ops/contrib
https://api.github.com/repos/backdrop-ops/contrib
closed
First contrib module: simple EU cookies
Maintainer application
I wrote my first module for backdrop. Displays a warning about the use of cookies. It is simply an implementation of jquery module: cookiebar. I would like other people to take advantage of this work. I would appreciate it got in the repository modules backdrop. The link to the module: https://github.com/davidvina/simple_EU_cookies best regards, and excuse me for my poor English.
True
First contrib module: simple EU cookies - I wrote my first module for backdrop. Displays a warning about the use of cookies. It is simply an implementation of jquery module: cookiebar. I would like other people to take advantage of this work. I would appreciate it got in the repository modules backdrop. The link to the module: https://github.com/davidvina/simple_EU_cookies best regards, and excuse me for my poor English.
main
first contrib module simple eu cookies i wrote my first module for backdrop displays a warning about the use of cookies it is simply an implementation of jquery module cookiebar i would like other people to take advantage of this work i would appreciate it got in the repository modules backdrop the link to the module best regards and excuse me for my poor english
1
4,444
23,085,904,417
IssuesEvent
2022-07-26 11:22:52
software-mansion/react-native-reanimated
https://api.github.com/repos/software-mansion/react-native-reanimated
opened
Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from` on `export * as default from './Animated';`
needs-review maintainer-issue
<!-- NOTE: please submit only bug reports here, any new questions or feature requests should be submitted in Discussions: https://github.com/software-mansion/react-native-reanimated/discussions --> ## Description <!-- Tell us what's happening here. --> Starting from Reanimated 2.9.0, **if you forget to add Reanimated's Babel plugin**, instead of a user-friendly error message, the following error will appear: ```sh error: node_modules/react-native-reanimated/src/index.ts: /Users/tomekzaw/RNOS/Issue3397/node_modules/react-native-reanimated/src/index.ts: Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from`. 5 | export * from './reanimated1'; 6 | export * from './reanimated2'; > 7 | export * as default from './Animated'; | ^^^^^^^^^^^^ 8 | ``` | Android | iOS | |:-:|:-:| | <img src="https://user-images.githubusercontent.com/20516055/180992073-0c43b86b-84bb-4cc0-8a0f-fae2cd89beda.png" alt="iOS error" height="700" /> | <img src="https://user-images.githubusercontent.com/20516055/180992066-da1f2227-7dc0-46e8-9e1b-55eb0bd095f0.png" alt="Android error" height="700" /> | ![funny meme](https://media.makeameme.org/created/but-we-have-5b164d.jpg) ## Solution 1. Add Reanimated's babel plugin to your `babel.config.js`: ```diff module.exports = { presets: ['module:metro-react-native-babel-preset'], + plugins: ['react-native-reanimated/plugin'], }; ``` 2. Clear application cache (depending on your workflow or favourite package manager): * `yarn start --reset-cache` * `npm start -- --reset-cache` * `expo start -c` 3. Reload app See [installation docs](https://docs.swmansion.com/react-native-reanimated/docs/fundamentals/installation#babel-plugin) for more details.
True
Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from` on `export * as default from './Animated';` - <!-- NOTE: please submit only bug reports here, any new questions or feature requests should be submitted in Discussions: https://github.com/software-mansion/react-native-reanimated/discussions --> ## Description <!-- Tell us what's happening here. --> Starting from Reanimated 2.9.0, **if you forget to add Reanimated's Babel plugin**, instead of a user-friendly error message, the following error will appear: ```sh error: node_modules/react-native-reanimated/src/index.ts: /Users/tomekzaw/RNOS/Issue3397/node_modules/react-native-reanimated/src/index.ts: Export namespace should be first transformed by `@babel/plugin-proposal-export-namespace-from`. 5 | export * from './reanimated1'; 6 | export * from './reanimated2'; > 7 | export * as default from './Animated'; | ^^^^^^^^^^^^ 8 | ``` | Android | iOS | |:-:|:-:| | <img src="https://user-images.githubusercontent.com/20516055/180992073-0c43b86b-84bb-4cc0-8a0f-fae2cd89beda.png" alt="iOS error" height="700" /> | <img src="https://user-images.githubusercontent.com/20516055/180992066-da1f2227-7dc0-46e8-9e1b-55eb0bd095f0.png" alt="Android error" height="700" /> | ![funny meme](https://media.makeameme.org/created/but-we-have-5b164d.jpg) ## Solution 1. Add Reanimated's babel plugin to your `babel.config.js`: ```diff module.exports = { presets: ['module:metro-react-native-babel-preset'], + plugins: ['react-native-reanimated/plugin'], }; ``` 2. Clear application cache (depending on your workflow or favourite package manager): * `yarn start --reset-cache` * `npm start -- --reset-cache` * `expo start -c` 3. Reload app See [installation docs](https://docs.swmansion.com/react-native-reanimated/docs/fundamentals/installation#babel-plugin) for more details.
main
export namespace should be first transformed by babel plugin proposal export namespace from on export as default from animated note please submit only bug reports here any new questions or feature requests should be submitted in discussions description tell us what s happening here starting from reanimated if you forget to add reanimated s babel plugin instead of a user friendly error message the following error will appear sh error node modules react native reanimated src index ts users tomekzaw rnos node modules react native reanimated src index ts export namespace should be first transformed by babel plugin proposal export namespace from export from export from export as default from animated android ios solution add reanimated s babel plugin to your babel config js diff module exports presets plugins clear application cache depending on your workflow or favourite package manager yarn start reset cache npm start reset cache expo start c reload app see for more details
1
99,562
4,056,886,409
IssuesEvent
2016-05-24 20:11:22
ndoit/fenrir
https://api.github.com/repos/ndoit/fenrir
closed
Partial word search not working
Priority 2
When you type ahead you see results. Try typing "Empl", you see: ![image](https://cloud.githubusercontent.com/assets/407661/15408809/0ce30f84-1de0-11e6-894a-d98a880c67d7.png) When you hit enter you see: ![image](https://cloud.githubusercontent.com/assets/407661/15408822/1e2b2a10-1de0-11e6-89bc-1726264f958f.png)
1.0
Partial word search not working - When you type ahead you see results. Try typing "Empl", you see: ![image](https://cloud.githubusercontent.com/assets/407661/15408809/0ce30f84-1de0-11e6-894a-d98a880c67d7.png) When you hit enter you see: ![image](https://cloud.githubusercontent.com/assets/407661/15408822/1e2b2a10-1de0-11e6-89bc-1726264f958f.png)
non_main
partial word search not working when you type ahead you see results try typing empl you see when you hit enter you see
0
430,797
30,200,593,346
IssuesEvent
2023-07-05 05:06:53
sulton-max/backend.caching
https://api.github.com/repos/sulton-max/backend.caching
opened
Add In Memory Cache documentation
documentation
### Task Add In Memory Cache documentation including * When to use it * What is it * Why to use it * How to use it
1.0
Add In Memory Cache documentation - ### Task Add In Memory Cache documentation including * When to use it * What is it * Why to use it * How to use it
non_main
add in memory cache documentation task add in memory cache documentation including when to use it what is it why to use it how to use it
0
63,860
15,726,854,261
IssuesEvent
2021-03-29 11:54:31
rticommunity/rticonnextdds-examples
https://api.github.com/repos/rticommunity/rticonnextdds-examples
closed
Port flat_data_latency C++11 example to CMake
build documentation enhancement good first issue style
The example `/examples/connext_dds/flat_data_latency/c++11` is missing the new `README.md` template and the `CMakeLists.txt`.
1.0
Port flat_data_latency C++11 example to CMake - The example `/examples/connext_dds/flat_data_latency/c++11` is missing the new `README.md` template and the `CMakeLists.txt`.
non_main
port flat data latency c example to cmake the example examples connext dds flat data latency c is missing the new readme md template and the cmakelists txt
0
4,841
24,955,745,275
IssuesEvent
2022-11-01 11:33:58
Lissy93/dashy
https://api.github.com/repos/Lissy93/dashy
closed
[QUESTION] Can I use actual domain not second level domain to fetch favicon?
🤷‍♂️ Question 👤 Awaiting Maintainer Response
### Question hi, I found dashy's favicon api always use sencod-level domain to fetch a favicon, just like google.com for google. Is there any method to use the actual domain to fetch a favicon,no matter it domain level,just like docs.google.com for google Doc ### Category Using Icons ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
True
[QUESTION] Can I use actual domain not second level domain to fetch favicon? - ### Question hi, I found dashy's favicon api always use sencod-level domain to fetch a favicon, just like google.com for google. Is there any method to use the actual domain to fetch a favicon,no matter it domain level,just like docs.google.com for google Doc ### Category Using Icons ### Please tick the boxes - [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy (check the first two digits of the version number) - [X] You've checked that this [question hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue) - [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide - [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct)
main
can i use actual domain not second level domain to fetch favicon question hi i found dashy s favicon api always use sencod level domain to fetch a favicon just like google com for google is there any method to use the actual domain to fetch a favicon,no matter it domain level,just like docs google com for google doc category using icons please tick the boxes you are using a version of dashy check the first two digits of the version number you ve checked that this you ve checked the and guide you agree to the
1
276
3,041,430,828
IssuesEvent
2015-08-07 21:16:07
angular/universal
https://api.github.com/repos/angular/universal
opened
Move preboot tests back to modules/preboot/test
module: preboot.js type: RepoMaintainerExperience
e2e tests are global and remain in root while module tests should remain within the module
True
Move preboot tests back to modules/preboot/test - e2e tests are global and remain in root while module tests should remain within the module
main
move preboot tests back to modules preboot test tests are global and remain in root while module tests should remain within the module
1
5,797
30,708,962,495
IssuesEvent
2023-07-27 08:27:31
precice/precice
https://api.github.com/repos/precice/precice
opened
Refactor B-Spline interpolation into a separate function
maintainability
**Please describe the problem you are trying to solve.** We are currently using B-Spline interpolation in waveforms provided by Eigen. This is implemented as part of the `time::Storage`. The interface is clear though: `(timestamps, samples, timeToSample) -> sample` We should extract this into a separate file, and test it. **Additional context** #1746
True
Refactor B-Spline interpolation into a separate function - **Please describe the problem you are trying to solve.** We are currently using B-Spline interpolation in waveforms provided by Eigen. This is implemented as part of the `time::Storage`. The interface is clear though: `(timestamps, samples, timeToSample) -> sample` We should extract this into a separate file, and test it. **Additional context** #1746
main
refactor b spline interpolation into a separate function please describe the problem you are trying to solve we are currently using b spline interpolation in waveforms provided by eigen this is implemented as part of the time storage the interface is clear though timestamps samples timetosample sample we should extract this into a separate file and test it additional context
1
4,971
25,540,480,841
IssuesEvent
2022-11-29 14:59:45
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Provide targeted error messages when tables & columns involved in an exploration are no longer present
type: bug work: backend work: frontend restricted: maintainers status: started
* Create a new exploration with a base_table and some initial columns * Delete some of the columns in the table directly - Notice that the API responds with a 500 error code * Delete the table which is the base_table of the query - The query also gets deleted - We need to show these in the warning message while prompting for user confirmation * Delete a linked table whose columns we use as initial_columns in the query - The API responds with a 500 Expected: * The frontend needs to show the columns and highlight the ones that are no longer present * The API response needs to provide both the column structure and targetted errors
True
Provide targeted error messages when tables & columns involved in an exploration are no longer present - * Create a new exploration with a base_table and some initial columns * Delete some of the columns in the table directly - Notice that the API responds with a 500 error code * Delete the table which is the base_table of the query - The query also gets deleted - We need to show these in the warning message while prompting for user confirmation * Delete a linked table whose columns we use as initial_columns in the query - The API responds with a 500 Expected: * The frontend needs to show the columns and highlight the ones that are no longer present * The API response needs to provide both the column structure and targetted errors
main
provide targeted error messages when tables columns involved in an exploration are no longer present create a new exploration with a base table and some initial columns delete some of the columns in the table directly notice that the api responds with a error code delete the table which is the base table of the query the query also gets deleted we need to show these in the warning message while prompting for user confirmation delete a linked table whose columns we use as initial columns in the query the api responds with a expected the frontend needs to show the columns and highlight the ones that are no longer present the api response needs to provide both the column structure and targetted errors
1
3,993
18,495,026,731
IssuesEvent
2021-10-19 07:43:50
camunda-cloud/zeebe
https://api.github.com/repos/camunda-cloud/zeebe
closed
Define a way to manage the state access
Type: Maintenance Impact: Performance Impact: Maintainability
**Description** Define a way how to * avoid multiple reads/writes of one state entity within one BPMN step. * check invariants * track state changes for debugging Related to #4227
True
Define a way to manage the state access - **Description** Define a way how to * avoid multiple reads/writes of one state entity within one BPMN step. * check invariants * track state changes for debugging Related to #4227
main
define a way to manage the state access description define a way how to avoid multiple reads writes of one state entity within one bpmn step check invariants track state changes for debugging related to
1
597,560
18,166,487,258
IssuesEvent
2021-09-27 15:05:43
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
opened
Expose cleanup flag in oneprovisoin
Type: Feature Status: Accepted Priority: Normal Category: Provision
**Description** Delete a provision fails if trying to delete a provision that contains virtual machines and images. To avoid this issue you need to use --cleanup flag in oneprovision, but this is not exposed through the GUI. **Use case** Erase a provision from the GUI **Interface Changes** Add a checkbox on the delete dialog to activate the flag. **Additional Context** Add any other context or screenshots about the feature request here. Or any other alternative you have considered to addressed this new feature. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
Expose cleanup flag in oneprovisoin - **Description** Delete a provision fails if trying to delete a provision that contains virtual machines and images. To avoid this issue you need to use --cleanup flag in oneprovision, but this is not exposed through the GUI. **Use case** Erase a provision from the GUI **Interface Changes** Add a checkbox on the delete dialog to activate the flag. **Additional Context** Add any other context or screenshots about the feature request here. Or any other alternative you have considered to addressed this new feature. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
non_main
expose cleanup flag in oneprovisoin description delete a provision fails if trying to delete a provision that contains virtual machines and images to avoid this issue you need to use cleanup flag in oneprovision but this is not exposed through the gui use case erase a provision from the gui interface changes add a checkbox on the delete dialog to activate the flag additional context add any other context or screenshots about the feature request here or any other alternative you have considered to addressed this new feature progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
0
91,686
8,316,500,842
IssuesEvent
2018-09-25 09:13:37
Microsoft/AzureStorageExplorer
https://api.github.com/repos/Microsoft/AzureStorageExplorer
opened
Display a tedious message when opening blobs/files
testing
**Storage Explorer Version**: 1.5.0 **Platform/OS Version**: Windows 10/ Linux Ubuntu 16.04/ MacOS High Sierra **Architecture**: ia32/x64 **Build Number**: 20180924.1 **Commit**: 1a75b88f **Regression From**: Previous release 1.4.2(20180917.1) #### Steps to Reproduce: #### 1. Open one blob container then upload a blob to it. 2. Try to open the uploaded blob -> Check the displayed message in Activities. #### Expected Experience: #### The tedious message disappears after opening the blob/file successfully. #### Actual Experience: #### Display a tedious message in Activities. ![tedious](https://user-images.githubusercontent.com/34729022/46002087-99a5a100-c09c-11e8-99a3-a753856e9673.png)
1.0
Display a tedious message when opening blobs/files - **Storage Explorer Version**: 1.5.0 **Platform/OS Version**: Windows 10/ Linux Ubuntu 16.04/ MacOS High Sierra **Architecture**: ia32/x64 **Build Number**: 20180924.1 **Commit**: 1a75b88f **Regression From**: Previous release 1.4.2(20180917.1) #### Steps to Reproduce: #### 1. Open one blob container then upload a blob to it. 2. Try to open the uploaded blob -> Check the displayed message in Activities. #### Expected Experience: #### The tedious message disappears after opening the blob/file successfully. #### Actual Experience: #### Display a tedious message in Activities. ![tedious](https://user-images.githubusercontent.com/34729022/46002087-99a5a100-c09c-11e8-99a3-a753856e9673.png)
non_main
display a tedious message when opening blobs files storage explorer version platform os version windows linux ubuntu macos high sierra architecture build number commit regression from previous release steps to reproduce open one blob container then upload a blob to it try to open the uploaded blob check the displayed message in activities expected experience the tedious message disappears after opening the blob file successfully actual experience display a tedious message in activities
0
2,036
6,848,588,102
IssuesEvent
2017-11-13 19:02:29
tgstation/tgstation-server
https://api.github.com/repos/tgstation/tgstation-server
closed
Console mode for the service/Seperate actual hosting from service integration
Feature Request Maintainability Issue Work In Progress
Would make debugging much easier
True
Console mode for the service/Seperate actual hosting from service integration - Would make debugging much easier
main
console mode for the service seperate actual hosting from service integration would make debugging much easier
1
176,124
14,564,131,811
IssuesEvent
2020-12-17 04:17:43
lammps/lammps
https://api.github.com/repos/lammps/lammps
opened
Review use of THERMO_ENERGY in `Fix::setmask()` and methods of computing of the conserved quantity in MD
documentation maintenance
There are some inconsistencies and unusual behavior due to choices made a long time ago, that need to be reviewed and possibly some changes made. For details, please see the discussion on PR #2515 Further discussion results and proposed changes should be documented here.
1.0
Review use of THERMO_ENERGY in `Fix::setmask()` and methods of computing of the conserved quantity in MD - There are some inconsistencies and unusual behavior due to choices made a long time ago, that need to be reviewed and possibly some changes made. For details, please see the discussion on PR #2515 Further discussion results and proposed changes should be documented here.
non_main
review use of thermo energy in fix setmask and methods of computing of the conserved quantity in md there are some inconsistencies and unusual behavior due to choices made a long time ago that need to be reviewed and possibly some changes made for details please see the discussion on pr further discussion results and proposed changes should be documented here
0
894
4,553,934,052
IssuesEvent
2016-09-13 07:35:41
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker: 'reloaded' state recreates container everytime if volumes-from is used
affects_2.0 bug_report cloud docker P3 waiting_on_maintainer
Hi, according to the docs it should recreate the container if parameters changed: > "reloaded" asserts that all matching containers are running and restarts any that have any images or configuration out of date. But if you run this simple play multiple times, ansible will recreate the container every time even though no parameters have changed: ``` - hosts: 127.0.0.1 connection: local tasks: - docker: image=busybox name=test-data volumes=/data command=/bin/true state=present - docker: image: ubuntu name: test state: reloaded command: "nc -l -k 2342" volumes_from: test-data ``` In general state reloaded doesn't seem to work properly. See also #1129 which I just opened for a different issue with this.
True
docker: 'reloaded' state recreates container everytime if volumes-from is used - Hi, according to the docs it should recreate the container if parameters changed: > "reloaded" asserts that all matching containers are running and restarts any that have any images or configuration out of date. But if you run this simple play multiple times, ansible will recreate the container every time even though no parameters have changed: ``` - hosts: 127.0.0.1 connection: local tasks: - docker: image=busybox name=test-data volumes=/data command=/bin/true state=present - docker: image: ubuntu name: test state: reloaded command: "nc -l -k 2342" volumes_from: test-data ``` In general state reloaded doesn't seem to work properly. See also #1129 which I just opened for a different issue with this.
main
docker reloaded state recreates container everytime if volumes from is used hi according to the docs it should recreate the container if parameters changed reloaded asserts that all matching containers are running and restarts any that have any images or configuration out of date but if you run this simple play multiple times ansible will recreate the container every time even though no parameters have changed hosts connection local tasks docker image busybox name test data volumes data command bin true state present docker image ubuntu name test state reloaded command nc l k volumes from test data in general state reloaded doesn t seem to work properly see also which i just opened for a different issue with this
1
3,373
13,057,324,638
IssuesEvent
2020-07-30 07:05:29
short-d/short
https://api.github.com/repos/short-d/short
closed
[Refactor] Replace mustParseTime with must.Time across the codebase
maintainability
**What is frustrating you?** The existing tests use [mustParseTime](https://github.com/short-d/short/blob/master/backend/app/adapter/sqldb/main_test.go#L54) helper to convert human readable time in the test case to help test readers understand the test case. However, it does NOT inherently support returning `Time pointer`, preventing time string being directly embed into the test case. https://github.com/short-d/short/pull/965 is adding `must.Time()` to support point return value. This lead to code duplication & inconsistency. **Your solution** Remove or refactor `mustParseTime` across the codebase.
True
[Refactor] Replace mustParseTime with must.Time across the codebase - **What is frustrating you?** The existing tests use [mustParseTime](https://github.com/short-d/short/blob/master/backend/app/adapter/sqldb/main_test.go#L54) helper to convert human readable time in the test case to help test readers understand the test case. However, it does NOT inherently support returning `Time pointer`, preventing time string being directly embed into the test case. https://github.com/short-d/short/pull/965 is adding `must.Time()` to support point return value. This lead to code duplication & inconsistency. **Your solution** Remove or refactor `mustParseTime` across the codebase.
main
replace mustparsetime with must time across the codebase what is frustrating you the existing tests use helper to convert human readable time in the test case to help test readers understand the test case however it does not inherently support returning time pointer preventing time string being directly embed into the test case is adding must time to support point return value this lead to code duplication inconsistency your solution remove or refactor mustparsetime across the codebase
1
373,283
26,047,476,668
IssuesEvent
2022-12-22 15:33:14
arcanus55/neodigm55
https://api.github.com/repos/arcanus55/neodigm55
closed
Enchanted CTA | Support Material Design icons (google font) via neodigm-icon element
documentation enhancement
Adding a neodigm-icon element within the button text should display an inline icon. Create a wiki recipe.
1.0
Enchanted CTA | Support Material Design icons (google font) via neodigm-icon element - Adding a neodigm-icon element within the button text should display an inline icon. Create a wiki recipe.
non_main
enchanted cta support material design icons google font via neodigm icon element adding a neodigm icon element within the button text should display an inline icon create a wiki recipe
0
279,372
30,702,524,042
IssuesEvent
2023-07-27 01:37:35
Satheesh575555/linux-4.1.15
https://api.github.com/repos/Satheesh575555/linux-4.1.15
closed
CVE-2019-19061 (High) detected in linuxlinux-4.6 - autoclosed
Mend: dependency security vulnerability
## CVE-2019-19061 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15/commit/951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1">951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/iio/imu/adis_buffer.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the adis_update_scan_mode_burst() function in drivers/iio/imu/adis_buffer.c in the Linux kernel before 5.3.9 allows attackers to cause a denial of service (memory consumption), aka CID-9c0530e898f3. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19061>CVE-2019-19061</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061</a></p> <p>Release Date: 2020-09-25</p> <p>Fix Resolution: v5.4-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-19061 (High) detected in linuxlinux-4.6 - autoclosed - ## CVE-2019-19061 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/linux-4.1.15/commit/951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1">951a6fe29b85bb7a6493c21ded9c3151b6a6c8f1</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/iio/imu/adis_buffer.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A memory leak in the adis_update_scan_mode_burst() function in drivers/iio/imu/adis_buffer.c in the Linux kernel before 5.3.9 allows attackers to cause a denial of service (memory consumption), aka CID-9c0530e898f3. <p>Publish Date: 2019-11-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19061>CVE-2019-19061</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19061</a></p> <p>Release Date: 2020-09-25</p> <p>Fix Resolution: v5.4-rc3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers iio imu adis buffer c vulnerability details a memory leak in the adis update scan mode burst function in drivers iio imu adis buffer c in the linux kernel before allows attackers to cause a denial of service memory consumption aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
4,059
18,981,825,719
IssuesEvent
2021-11-21 02:10:09
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Change set is throwing an error while deploying the template.
area/installation blocked/close-if-inactive stage/waiting-for-release maintainer/need-followup
<!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. --> ### Description: <!-- Briefly describe the bug you are facing.--> When I am deploying the template using SAM deploy. It is giving an error : **botocore.exceptions.ParamValidationError: Parameter validation failed: Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken** I am not using either of **--disable-rollback | --no-disable-rollback** while deploying ### Steps to reproduce: <!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) --> The command I am using : sam deploy --template-file packaged.yaml --stack-name <name of stack> --s3-bucket <name of bucket> --capabilities CAPABILITY_IAM --region $AWS_DEFAULT_REGION --parameter-overrides <list of parameters> ### Observed result: <!-- Please provide command output with `--debug` flag set. --> Error with --debug flag: Changeset created successfully. *** 2021-10-27 09:29:38,186 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'fb75737e-a71f-40ff-b874-f9c6d5b53f53', 'installationId': '47f3640c-0c6d-4159-9dac-4f18e6526b4d', 'sessionId': 'a629f491-1932-4811-b3b4-7532d1cdceb3', 'executionEnvironment': 'GitLab', 'ci': True, 'pyversion': '3.8.12', 'samcliVersion': '1.34.1', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': 'eu-west-1', 'commandName': 'sam deploy', 'duration': 16611, 'exitReason': 'ParamValidationError', 'exitCode': 255}}]} 2021-10-27 09:29:38,810 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1) Traceback (most recent call last): File "/usr/local/bin/sam", line 8, in <module> sys.exit(cli()) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/lib/cli_validation/image_repository_validation.py", line 85, in wrapped return func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/click/decorators.py", line 73, in new_func return ctx.invoke(f, obj, *args, **kwargs) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 157, in wrapped raise exception # pylint: disable=raising-bad-type File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 122, in wrapped return_value = func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/lib/utils/version_checker.py", line 41, in wrapped actual_result = func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/cli/main.py", line 87, in wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 163, in cli do_cli( File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 317, in do_cli deploy_context.run() File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 161, in run return self.deploy( File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 265, in deploy self.deployer.execute_changeset(result["Id"], stack_name, disable_rollback) File "/usr/local/lib/python3.8/site-packages/samcli/lib/deploy/deployer.py", line 333, in execute_changeset return self._client.execute_change_set( File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 386, in _api_call return self._make_api_call(operation_name, kwargs) File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 677, in _make_api_call request_dict = self._convert_to_request_dict( File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 725, in _convert_to_request_dict request_dict = self._serializer.serialize_to_request( File "/usr/local/lib/python3.8/site-packages/botocore/validate.py", line 319, in serialize_to_request raise ParamValidationError(report=report.generate_report()) botocore.exceptions.ParamValidationError: Parameter validation failed: Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken ### Expected result: <!-- Describe what you expected. --> ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Linux 2. `sam --version`: 1.34.1 3. AWS region: eu-west-1 `Add --debug flag to command you are running`
True
Change set is throwing an error while deploying the template. - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. --> ### Description: <!-- Briefly describe the bug you are facing.--> When I am deploying the template using SAM deploy. It is giving an error : **botocore.exceptions.ParamValidationError: Parameter validation failed: Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken** I am not using either of **--disable-rollback | --no-disable-rollback** while deploying ### Steps to reproduce: <!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) --> The command I am using : sam deploy --template-file packaged.yaml --stack-name <name of stack> --s3-bucket <name of bucket> --capabilities CAPABILITY_IAM --region $AWS_DEFAULT_REGION --parameter-overrides <list of parameters> ### Observed result: <!-- Please provide command output with `--debug` flag set. --> Error with --debug flag: Changeset created successfully. *** 2021-10-27 09:29:38,186 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'fb75737e-a71f-40ff-b874-f9c6d5b53f53', 'installationId': '47f3640c-0c6d-4159-9dac-4f18e6526b4d', 'sessionId': 'a629f491-1932-4811-b3b4-7532d1cdceb3', 'executionEnvironment': 'GitLab', 'ci': True, 'pyversion': '3.8.12', 'samcliVersion': '1.34.1', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': 'eu-west-1', 'commandName': 'sam deploy', 'duration': 16611, 'exitReason': 'ParamValidationError', 'exitCode': 255}}]} 2021-10-27 09:29:38,810 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1) Traceback (most recent call last): File "/usr/local/bin/sam", line 8, in <module> sys.exit(cli()) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/lib/cli_validation/image_repository_validation.py", line 85, in wrapped return func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/click/decorators.py", line 73, in new_func return ctx.invoke(f, obj, *args, **kwargs) File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 157, in wrapped raise exception # pylint: disable=raising-bad-type File "/usr/local/lib/python3.8/site-packages/samcli/lib/telemetry/metric.py", line 122, in wrapped return_value = func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/lib/utils/version_checker.py", line 41, in wrapped actual_result = func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/cli/main.py", line 87, in wrapper return func(*args, **kwargs) File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 163, in cli do_cli( File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/command.py", line 317, in do_cli deploy_context.run() File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 161, in run return self.deploy( File "/usr/local/lib/python3.8/site-packages/samcli/commands/deploy/deploy_context.py", line 265, in deploy self.deployer.execute_changeset(result["Id"], stack_name, disable_rollback) File "/usr/local/lib/python3.8/site-packages/samcli/lib/deploy/deployer.py", line 333, in execute_changeset return self._client.execute_change_set( File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 386, in _api_call return self._make_api_call(operation_name, kwargs) File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 677, in _make_api_call request_dict = self._convert_to_request_dict( File "/usr/local/lib/python3.8/site-packages/botocore/client.py", line 725, in _convert_to_request_dict request_dict = self._serializer.serialize_to_request( File "/usr/local/lib/python3.8/site-packages/botocore/validate.py", line 319, in serialize_to_request raise ParamValidationError(report=report.generate_report()) botocore.exceptions.ParamValidationError: Parameter validation failed: Unknown parameter in input: "DisableRollback", must be one of: ChangeSetName, StackName, ClientRequestToken ### Expected result: <!-- Describe what you expected. --> ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Linux 2. `sam --version`: 1.34.1 3. AWS region: eu-west-1 `Add --debug flag to command you are running`
main
change set is throwing an error while deploying the template make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description when i am deploying the template using sam deploy it is giving an error botocore exceptions paramvalidationerror parameter validation failed unknown parameter in input disablerollback must be one of changesetname stackname clientrequesttoken i am not using either of disable rollback no disable rollback while deploying steps to reproduce the command i am using sam deploy template file packaged yaml stack name bucket capabilities capability iam region aws default region parameter overrides observed result error with debug flag changeset created successfully sending telemetry metrics httpsconnectionpool host aws serverless tools telemetry us west amazonaws com port read timed out read timeout traceback most recent call last file usr local bin sam line in sys exit cli file usr local lib site packages click core py line in call return self main args kwargs file usr local lib site packages click core py line in main rv self invoke ctx file usr local lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file usr local lib site packages click core py line in invoke return ctx invoke self callback ctx params file usr local lib site packages click core py line in invoke return callback args kwargs file usr local lib site packages samcli lib cli validation image repository validation py line in wrapped return func args kwargs file usr local lib site packages click decorators py line in new func return ctx invoke f obj args kwargs file usr local lib site packages click core py line in invoke return callback args kwargs file usr local lib site packages samcli lib telemetry metric py line in wrapped raise exception pylint disable raising bad type file usr local lib site packages samcli lib telemetry metric py line in wrapped return value func args kwargs file usr local lib site packages samcli lib utils version checker py line in wrapped actual result func args kwargs file usr local lib site packages samcli cli main py line in wrapper return func args kwargs file usr local lib site packages samcli commands deploy command py line in cli do cli file usr local lib site packages samcli commands deploy command py line in do cli deploy context run file usr local lib site packages samcli commands deploy deploy context py line in run return self deploy file usr local lib site packages samcli commands deploy deploy context py line in deploy self deployer execute changeset result stack name disable rollback file usr local lib site packages samcli lib deploy deployer py line in execute changeset return self client execute change set file usr local lib site packages botocore client py line in api call return self make api call operation name kwargs file usr local lib site packages botocore client py line in make api call request dict self convert to request dict file usr local lib site packages botocore client py line in convert to request dict request dict self serializer serialize to request file usr local lib site packages botocore validate py line in serialize to request raise paramvalidationerror report report generate report botocore exceptions paramvalidationerror parameter validation failed unknown parameter in input disablerollback must be one of changesetname stackname clientrequesttoken expected result additional environment details ex windows mac amazon linux etc os linux sam version aws region eu west add debug flag to command you are running
1
319,213
23,760,456,735
IssuesEvent
2022-09-01 08:26:11
Danilo-Danijel-team/sms_sender
https://api.github.com/repos/Danilo-Danijel-team/sms_sender
closed
Create postman documentation tests for created server routes.
documentation test
- [x] Create tests based on what services are returning from their API responses. Some of the routes that will be present are listed in #10 issue. - [x] Create tests before @danilojezernik will start doing #10 issue, there must be postman documentation tests prepared that will serve as a testing environment for server development. Those tests will @danilojezernik use while developing server logic. Be aware that this list of routes and requirements will change many times in the future, and you will add new tests and correct existing ones over many iterations.
1.0
Create postman documentation tests for created server routes. - - [x] Create tests based on what services are returning from their API responses. Some of the routes that will be present are listed in #10 issue. - [x] Create tests before @danilojezernik will start doing #10 issue, there must be postman documentation tests prepared that will serve as a testing environment for server development. Those tests will @danilojezernik use while developing server logic. Be aware that this list of routes and requirements will change many times in the future, and you will add new tests and correct existing ones over many iterations.
non_main
create postman documentation tests for created server routes create tests based on what services are returning from their api responses some of the routes that will be present are listed in issue create tests before danilojezernik will start doing issue there must be postman documentation tests prepared that will serve as a testing environment for server development those tests will danilojezernik use while developing server logic be aware that this list of routes and requirements will change many times in the future and you will add new tests and correct existing ones over many iterations
0
2,724
9,612,818,489
IssuesEvent
2019-05-13 09:47:37
precice/precice
https://api.github.com/repos/precice/precice
opened
Refactor static ids
maintainability
preCICE currently has to manage a set of `static` IDs due to some prior design decisions regarding simpler object creation. The additional bookkeeping, however, is error-prone and already lead to issues #378. Affected classes are: 1) Participant via `_participantsSize` 2) Mesh via `_managePropertyID` 3) Data via `_dataCount` These static IDs should be refactored out of the affected classes into the objects responsible for the generation of IDed objects. The object lifetime of the generators should then correctly handle the bookkeeping. These generators are most likely going to be configuration objects.
True
Refactor static ids - preCICE currently has to manage a set of `static` IDs due to some prior design decisions regarding simpler object creation. The additional bookkeeping, however, is error-prone and already lead to issues #378. Affected classes are: 1) Participant via `_participantsSize` 2) Mesh via `_managePropertyID` 3) Data via `_dataCount` These static IDs should be refactored out of the affected classes into the objects responsible for the generation of IDed objects. The object lifetime of the generators should then correctly handle the bookkeeping. These generators are most likely going to be configuration objects.
main
refactor static ids precice currently has to manage a set of static ids due to some prior design decisions regarding simpler object creation the additional bookkeeping however is error prone and already lead to issues affected classes are participant via participantssize mesh via managepropertyid data via datacount these static ids should be refactored out of the affected classes into the objects responsible for the generation of ided objects the object lifetime of the generators should then correctly handle the bookkeeping these generators are most likely going to be configuration objects
1
2,603
3,916,938,910
IssuesEvent
2016-04-21 05:29:30
steveworley/drush-audit
https://api.github.com/repos/steveworley/drush-audit
closed
Login security is enabled and configured
enhancement security
- [Login security](https://www.drupal.org/project/login_security), recommend for projects if it is not available - If the module exists ensure that it is enabled and that it is configured correctly
True
Login security is enabled and configured - - [Login security](https://www.drupal.org/project/login_security), recommend for projects if it is not available - If the module exists ensure that it is enabled and that it is configured correctly
non_main
login security is enabled and configured recommend for projects if it is not available if the module exists ensure that it is enabled and that it is configured correctly
0
88
2,608,261,350
IssuesEvent
2015-02-26 03:56:33
mranney/node_pcap
https://api.github.com/repos/mranney/node_pcap
closed
Setup travis CI build which runs tests and hints code
maintainance
Having a travis ci build really helps in determining the state of pull requests. - [x] Setup travis - [x] Run hinter on travis - [x] Run unit tests on travis - [x] Setup coveralls
True
Setup travis CI build which runs tests and hints code - Having a travis ci build really helps in determining the state of pull requests. - [x] Setup travis - [x] Run hinter on travis - [x] Run unit tests on travis - [x] Setup coveralls
main
setup travis ci build which runs tests and hints code having a travis ci build really helps in determining the state of pull requests setup travis run hinter on travis run unit tests on travis setup coveralls
1
2,310
8,279,120,360
IssuesEvent
2018-09-18 01:18:27
spacetelescope/wfc3tools
https://api.github.com/repos/spacetelescope/wfc3tools
opened
TST: Add real tests and put them on Jenkins/Artifactory
maintainance
For working examples, see `hstcal`, `acstools`, `stistools`, or `calcos`.
True
TST: Add real tests and put them on Jenkins/Artifactory - For working examples, see `hstcal`, `acstools`, `stistools`, or `calcos`.
main
tst add real tests and put them on jenkins artifactory for working examples see hstcal acstools stistools or calcos
1
4,596
23,831,872,185
IssuesEvent
2022-09-05 22:23:23
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
opened
MAINT: improve auto-labeling
maintainability
I should mention that the approach to auto-labeling that we currently use for PRs and stuff was rejected upstream and replaced with something that is supposed to work better re: not mutating/removing labels humans already placed on PRs: https://github.com/scipy/scipy/pull/16870 We may want to look into that--I'm almost certain I've seen the auto-labeling doing that kind of annoying thing of removing a human label over here as well.
True
MAINT: improve auto-labeling - I should mention that the approach to auto-labeling that we currently use for PRs and stuff was rejected upstream and replaced with something that is supposed to work better re: not mutating/removing labels humans already placed on PRs: https://github.com/scipy/scipy/pull/16870 We may want to look into that--I'm almost certain I've seen the auto-labeling doing that kind of annoying thing of removing a human label over here as well.
main
maint improve auto labeling i should mention that the approach to auto labeling that we currently use for prs and stuff was rejected upstream and replaced with something that is supposed to work better re not mutating removing labels humans already placed on prs we may want to look into that i m almost certain i ve seen the auto labeling doing that kind of annoying thing of removing a human label over here as well
1
2,580
8,773,417,146
IssuesEvent
2018-12-18 16:48:41
arcticicestudio/nord-docs
https://api.github.com/repos/arcticicestudio/nord-docs
opened
Google Analytics
context-workflow scope-maintainability scope-quality scope-stability type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167256-14bbd380-02e9-11e9-8aca-a31baf745cd8.png" width="20%"/></p> > Associated epics: #86 This issue documents the implementation of [Google Analytics][ga-mark] like documented in the [“Analytics & Statistics” design concept][gh-86]. <p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167593-c824c800-02e9-11e9-9b70-84b6fc40c05f.png " width="20%"/></p> The main tool to collect and analyze data will be [Google Analytics][ga-mark]. It is a stable and proven service with a lot of useful configurable features and a reliable persistence. _Nord Docs_ will use the latest and recommended [gtag.js][gdev-ga-gtag] library that optionally allows, next to Google Analytics itself, the integration of almost all Google Marketing services like e.g. [Google Tag Manager][gdev-tm]. The library will be integrated through [gatsby-plugin-google-gtag][gh-gb-p-ga-tag]. ## Tasks - [ ] Install required packages: - [gatsby-plugin-google-gtag][npm-gp-gtag] - [ ] Implement required internal constants. - [ ] Implement the plugin configuration. [g-sup-anonip]: https://support.google.com/analytics/answer/2763052 [gh-gb-p-ga-tag]: https://github.com/gatsbyjs/gatsby/tree/master/packages/gatsby-plugin-google-gtag [gh-86]: https://github.com/arcticicestudio/nord-docs/issues/86 [ga-mark]: https://marketingplatform.google.com/about/analytics [gdev-ga-gtag]: https://developers.google.com/analytics/devguides/collection/gtagjs [gdev-tm]: https://developers.google.com/tag-manager [wiki-a]: https://en.wikipedia.org/wiki/Analytics [wiki-s]: https://en.wikipedia.org/wiki/Statistics [wiki-dnt]: https://en.wikipedia.org/wiki/Do_Not_Track [npm-gp-gtag]: https://www.npmjs.com/package/gatsby-plugin-google-gtag
True
Google Analytics - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167256-14bbd380-02e9-11e9-8aca-a31baf745cd8.png" width="20%"/></p> > Associated epics: #86 This issue documents the implementation of [Google Analytics][ga-mark] like documented in the [“Analytics & Statistics” design concept][gh-86]. <p align="center"><img src="https://user-images.githubusercontent.com/7836623/50167593-c824c800-02e9-11e9-9b70-84b6fc40c05f.png " width="20%"/></p> The main tool to collect and analyze data will be [Google Analytics][ga-mark]. It is a stable and proven service with a lot of useful configurable features and a reliable persistence. _Nord Docs_ will use the latest and recommended [gtag.js][gdev-ga-gtag] library that optionally allows, next to Google Analytics itself, the integration of almost all Google Marketing services like e.g. [Google Tag Manager][gdev-tm]. The library will be integrated through [gatsby-plugin-google-gtag][gh-gb-p-ga-tag]. ## Tasks - [ ] Install required packages: - [gatsby-plugin-google-gtag][npm-gp-gtag] - [ ] Implement required internal constants. - [ ] Implement the plugin configuration. [g-sup-anonip]: https://support.google.com/analytics/answer/2763052 [gh-gb-p-ga-tag]: https://github.com/gatsbyjs/gatsby/tree/master/packages/gatsby-plugin-google-gtag [gh-86]: https://github.com/arcticicestudio/nord-docs/issues/86 [ga-mark]: https://marketingplatform.google.com/about/analytics [gdev-ga-gtag]: https://developers.google.com/analytics/devguides/collection/gtagjs [gdev-tm]: https://developers.google.com/tag-manager [wiki-a]: https://en.wikipedia.org/wiki/Analytics [wiki-s]: https://en.wikipedia.org/wiki/Statistics [wiki-dnt]: https://en.wikipedia.org/wiki/Do_Not_Track [npm-gp-gtag]: https://www.npmjs.com/package/gatsby-plugin-google-gtag
main
google analytics associated epics this issue documents the implementation of like documented in the the main tool to collect and analyze data will be it is a stable and proven service with a lot of useful configurable features and a reliable persistence nord docs will use the latest and recommended library that optionally allows next to google analytics itself the integration of almost all google marketing services like e g the library will be integrated through tasks install required packages implement required internal constants implement the plugin configuration
1
1,971
6,694,170,457
IssuesEvent
2017-10-10 00:04:14
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Amazon: localised amazon results
Maintainer Input Requested
Currently, the results are returned from "amazon.com" (by default). It would be **useful** if we could return **localized results**. For example, the user located in Germany should see results from "amazon.de" instead of "amazon.com". --- IA Page: http://duck.co/ia/view/products [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @bsstoner
True
Amazon: localised amazon results - Currently, the results are returned from "amazon.com" (by default). It would be **useful** if we could return **localized results**. For example, the user located in Germany should see results from "amazon.de" instead of "amazon.com". --- IA Page: http://duck.co/ia/view/products [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @bsstoner
main
amazon localised amazon results currently the results are returned from amazon com by default it would be useful if we could return localized results for example the user located in germany should see results from amazon de instead of amazon com ia page bsstoner
1
4,947
25,455,552,139
IssuesEvent
2022-11-24 13:55:25
pace/bricks
https://api.github.com/repos/pace/bricks
closed
Remove the service argument
T::Maintainance
Remove the `service` argument in the command so that we can create new microservices with `pace new pay`
True
Remove the service argument - Remove the `service` argument in the command so that we can create new microservices with `pace new pay`
main
remove the service argument remove the service argument in the command so that we can create new microservices with pace new pay
1
2,555
8,695,162,685
IssuesEvent
2018-12-04 14:33:09
arcticicestudio/nord-docs
https://api.github.com/repos/arcticicestudio/nord-docs
closed
Theme Global Base Styles
context-ui scope-maintainability scope-ux type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/49328407-a4832300-f570-11e8-8613-21c78267b558.png" width="20%" /></p> > Associated epic: #51 This issue handles the implementation of the basic theme styles like font and background colors. They will be initially used in the global theme styles implemented in #53. ## Tasks - [x] Implement `background` and `font` CSS module styles. - [x] Integrate basic `background-color` and `color` in global CSS theme styles. - Include a `transition` for both attributes to ensure a smooth theme mode transition.
True
Theme Global Base Styles - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/49328407-a4832300-f570-11e8-8613-21c78267b558.png" width="20%" /></p> > Associated epic: #51 This issue handles the implementation of the basic theme styles like font and background colors. They will be initially used in the global theme styles implemented in #53. ## Tasks - [x] Implement `background` and `font` CSS module styles. - [x] Integrate basic `background-color` and `color` in global CSS theme styles. - Include a `transition` for both attributes to ensure a smooth theme mode transition.
main
theme global base styles associated epic this issue handles the implementation of the basic theme styles like font and background colors they will be initially used in the global theme styles implemented in tasks implement background and font css module styles integrate basic background color and color in global css theme styles include a transition for both attributes to ensure a smooth theme mode transition
1
2,927
10,452,286,943
IssuesEvent
2019-09-19 14:25:49
substrate-developer-hub/hacktoberfest
https://api.github.com/repos/substrate-developer-hub/hacktoberfest
closed
Publish Substrate Runtime Library Modules v1.0 to crates.io
m-prep m-wontfix t-maintaince
Required steps: - Raise a PR against the substrate 1.0 branch that: - [ ] clean up dependency tree of sr- and srml-modules (see `cargo tree`) - [ ] Updates the license, description and other missing fields for each module (see `cargo package` warnings) - [ ] pins down versions of locally used dependencies - Publish _(this can only be done by a Parity employee)_: - publish all srml-, sr- and dependent crates on crates.io (see `cargo publish`) - make sure [paritytech:core-devs](https://crates.io/teams/github:paritytech:core-devs) is one of the owners of the package on crates.io (see [`cargo owner`](https://doc.rust-lang.org/cargo/reference/publishing.html#cargo-owner)) - Raise a PR against [the templates repository](https://github.com/shawntabrizi/substrate-package), replacing all published crates with version numbers
True
Publish Substrate Runtime Library Modules v1.0 to crates.io - Required steps: - Raise a PR against the substrate 1.0 branch that: - [ ] clean up dependency tree of sr- and srml-modules (see `cargo tree`) - [ ] Updates the license, description and other missing fields for each module (see `cargo package` warnings) - [ ] pins down versions of locally used dependencies - Publish _(this can only be done by a Parity employee)_: - publish all srml-, sr- and dependent crates on crates.io (see `cargo publish`) - make sure [paritytech:core-devs](https://crates.io/teams/github:paritytech:core-devs) is one of the owners of the package on crates.io (see [`cargo owner`](https://doc.rust-lang.org/cargo/reference/publishing.html#cargo-owner)) - Raise a PR against [the templates repository](https://github.com/shawntabrizi/substrate-package), replacing all published crates with version numbers
main
publish substrate runtime library modules to crates io required steps raise a pr against the substrate branch that clean up dependency tree of sr and srml modules see cargo tree updates the license description and other missing fields for each module see cargo package warnings pins down versions of locally used dependencies publish this can only be done by a parity employee publish all srml sr and dependent crates on crates io see cargo publish make sure is one of the owners of the package on crates io see raise a pr against replacing all published crates with version numbers
1
188,781
14,475,410,351
IssuesEvent
2020-12-10 01:33:14
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
radondb/radon: src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go; 13 LoC
fresh small test
Found a possible issue in [radondb/radon](https://www.github.com/radondb/radon) at [src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable s used in defer or goroutine at line 154 [Click here to see the code in its original context.](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159) <details> <summary>Click here to show the 13 line(s) of Go which triggered the analyzer.</summary> ```go for _, s := range servers { wg.Add(1) go func() { defer wg.Done() var a int64 var b string err := s.db.QueryRow("SELECT a, b FROM tbl WHERE id=1").Scan(&a, &b) assert.Nil(t, err) assert.Equal(t, int64(1), a) assert.Equal(t, "hello world", b) }() } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: e3744d066da240cb8657ac95a44373495369df46
1.0
radondb/radon: src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go; 13 LoC - Found a possible issue in [radondb/radon](https://www.github.com/radondb/radon) at [src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable s used in defer or goroutine at line 154 [Click here to see the code in its original context.](https://github.com/radondb/radon/blob/e3744d066da240cb8657ac95a44373495369df46/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/server/server_test.go#L147-L159) <details> <summary>Click here to show the 13 line(s) of Go which triggered the analyzer.</summary> ```go for _, s := range servers { wg.Add(1) go func() { defer wg.Done() var a int64 var b string err := s.db.QueryRow("SELECT a, b FROM tbl WHERE id=1").Scan(&a, &b) assert.Nil(t, err) assert.Equal(t, int64(1), a) assert.Equal(t, "hello world", b) }() } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: e3744d066da240cb8657ac95a44373495369df46
non_main
radondb radon src vendor github com radondb shift vendor github com siddontang go mysql server server test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable s used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for s range servers wg add go func defer wg done var a var b string err s db queryrow select a b from tbl where id scan a b assert nil t err assert equal t a assert equal t hello world b leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
145,418
13,150,421,095
IssuesEvent
2020-08-09 11:27:00
JuliaReach/LazySets.jl
https://api.github.com/repos/JuliaReach/LazySets.jl
opened
Invalid local links in documentation
documentation :book: fix :face_with_head_bandage:
```julia ┌ Warning: invalid local link: unresolved path in lib/interfaces.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "is_right_turn") │ link.url = "utils.html#LazySets.Arrays.is_right_turn" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/LineSegment.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "plot_recipe(::LazySet{<:Real})") │ link.url = "interfaces.html#LazySets.plot_recipe-Union{Tuple{LazySet{N}}, Tuple{N}, Tuple{LazySet{N},N}} where N<:Real" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "_isbounded_unit_dimensions") │ link.url = "interfaces.html#LazySets._isbounded_unit_dimensions-Union{Tuple{LazySet{N}}, Tuple{N}} where N<:Real" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "_isbounded_stiemke") │ link.url = "interfaces.html#LazySets._isbounded_stiemke-Union{Tuple{HPolyhedron{N,VN} where VN<:AbstractArray{N,1}}, Tuple{N}} where N<:Real" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "_isbounded_stiemke") │ link.url ```
1.0
Invalid local links in documentation - ```julia ┌ Warning: invalid local link: unresolved path in lib/interfaces.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "is_right_turn") │ link.url = "utils.html#LazySets.Arrays.is_right_turn" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/LineSegment.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "plot_recipe(::LazySet{<:Real})") │ link.url = "interfaces.html#LazySets.plot_recipe-Union{Tuple{LazySet{N}}, Tuple{N}, Tuple{LazySet{N},N}} where N<:Real" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "_isbounded_unit_dimensions") │ link.url = "interfaces.html#LazySets._isbounded_unit_dimensions-Union{Tuple{LazySet{N}}, Tuple{N}} where N<:Real" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "_isbounded_stiemke") │ link.url = "interfaces.html#LazySets._isbounded_stiemke-Union{Tuple{HPolyhedron{N,VN} where VN<:AbstractArray{N,1}}, Tuple{N}} where N<:Real" └ @ Documenter.Writers.HTMLWriter ~/.julia/packages/Documenter/bRylW/src/Writers/HTMLWriter.jl:1725 ┌ Warning: invalid local link: unresolved path in lib/sets/HPolyhedron.md │ link.text = │ 1-element Array{Any,1}: │ Markdown.Code("", "_isbounded_stiemke") │ link.url ```
non_main
invalid local links in documentation julia ┌ warning invalid local link unresolved path in lib interfaces md │ link text │ element array any │ markdown code is right turn │ link url utils html lazysets arrays is right turn └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets linesegment md │ link text │ element array any │ markdown code plot recipe lazyset real │ link url interfaces html lazysets plot recipe union tuple lazyset n tuple n tuple lazyset n n where n real └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets hpolyhedron md │ link text │ element array any │ markdown code isbounded unit dimensions │ link url interfaces html lazysets isbounded unit dimensions union tuple lazyset n tuple n where n real └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets hpolyhedron md │ link text │ element array any │ markdown code isbounded stiemke │ link url interfaces html lazysets isbounded stiemke union tuple hpolyhedron n vn where vn abstractarray n tuple n where n real └ documenter writers htmlwriter julia packages documenter brylw src writers htmlwriter jl ┌ warning invalid local link unresolved path in lib sets hpolyhedron md │ link text │ element array any │ markdown code isbounded stiemke │ link url
0
345,122
30,789,683,064
IssuesEvent
2023-07-31 15:20:31
vedmack/yadcf
https://api.github.com/repos/vedmack/yadcf
closed
Android 12 / Chrome 11 / Fixed Headers
waiting for response / test page link
With fixedHeader : true and YADCF this bug happens when selecting a text input field. Once i disable fixedHeader it does not have the input bug. **video of issue in my app** https://drive.google.com/file/d/1DcZ2jp_WYvgcKpdqeddUO7_oX-eRKeUR/view **demo url to see issue** [https://www.otelconsulting.com/bug.html](https://www.otelconsulting.com/bug.html) **demo url with latest beta 0.9.4.beta.46** [https://www.otelconsulting.com/bug-094beta46.html](https://www.otelconsulting.com/bug-094beta46.html) **How to replicate** visit url and then click on the column filter input in the position column iOS latest and Chrome do not produce the same issue.
1.0
Android 12 / Chrome 11 / Fixed Headers - With fixedHeader : true and YADCF this bug happens when selecting a text input field. Once i disable fixedHeader it does not have the input bug. **video of issue in my app** https://drive.google.com/file/d/1DcZ2jp_WYvgcKpdqeddUO7_oX-eRKeUR/view **demo url to see issue** [https://www.otelconsulting.com/bug.html](https://www.otelconsulting.com/bug.html) **demo url with latest beta 0.9.4.beta.46** [https://www.otelconsulting.com/bug-094beta46.html](https://www.otelconsulting.com/bug-094beta46.html) **How to replicate** visit url and then click on the column filter input in the position column iOS latest and Chrome do not produce the same issue.
non_main
android chrome fixed headers with fixedheader true and yadcf this bug happens when selecting a text input field once i disable fixedheader it does not have the input bug video of issue in my app demo url to see issue demo url with latest beta beta how to replicate visit url and then click on the column filter input in the position column ios latest and chrome do not produce the same issue
0
529
3,925,714,914
IssuesEvent
2016-04-22 20:05:39
heiglandreas/authLdap
https://api.github.com/repos/heiglandreas/authLdap
closed
Check for already successfull authentication before "reauthenticating"
bug maintainer reply expected
This is an issue reported at https://wordpress.org/support/topic/authldap-doesnt-work-with-other-auth-plugins.
True
Check for already successfull authentication before "reauthenticating" - This is an issue reported at https://wordpress.org/support/topic/authldap-doesnt-work-with-other-auth-plugins.
main
check for already successfull authentication before reauthenticating this is an issue reported at
1
52,587
13,753,231,334
IssuesEvent
2020-10-06 15:23:46
flyingcircusio/nixpkgs
https://api.github.com/repos/flyingcircusio/nixpkgs
opened
Vulnerability roundup 9: bind-9.12.4-P2: 3 advisories [7.5]
1.severity: security
[search](https://search.nix.gsc.io/?q=bind&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=bind+in%3Apath&type=Code) * [ ] [CVE-2020-8623](https://nvd.nist.gov/vuln/detail/CVE-2020-8623) CVSSv3=7.5 (nixos-19.03) * [ ] [CVE-2020-8622](https://nvd.nist.gov/vuln/detail/CVE-2020-8622) CVSSv3=6.5 (nixos-19.03) * [ ] [CVE-2020-8624](https://nvd.nist.gov/vuln/detail/CVE-2020-8624) CVSSv3=4.3 (nixos-19.03) Scanned versions: nixos-19.03: 715db89e7d3.
True
Vulnerability roundup 9: bind-9.12.4-P2: 3 advisories [7.5] - [search](https://search.nix.gsc.io/?q=bind&i=fosho&repos=NixOS-nixpkgs), [files](https://github.com/NixOS/nixpkgs/search?utf8=%E2%9C%93&q=bind+in%3Apath&type=Code) * [ ] [CVE-2020-8623](https://nvd.nist.gov/vuln/detail/CVE-2020-8623) CVSSv3=7.5 (nixos-19.03) * [ ] [CVE-2020-8622](https://nvd.nist.gov/vuln/detail/CVE-2020-8622) CVSSv3=6.5 (nixos-19.03) * [ ] [CVE-2020-8624](https://nvd.nist.gov/vuln/detail/CVE-2020-8624) CVSSv3=4.3 (nixos-19.03) Scanned versions: nixos-19.03: 715db89e7d3.
non_main
vulnerability roundup bind advisories nixos nixos nixos scanned versions nixos
0
264
3,023,990,424
IssuesEvent
2015-08-02 04:06:44
daemonraco/toobasic
https://api.github.com/repos/daemonraco/toobasic
closed
BLOB defaults error on MySQL
bug Database Structure Maintainer
## Error What __TooBasic__ attempts to create a table using BLOB columns it failes on the default value.
True
BLOB defaults error on MySQL - ## Error What __TooBasic__ attempts to create a table using BLOB columns it failes on the default value.
main
blob defaults error on mysql error what toobasic attempts to create a table using blob columns it failes on the default value
1
441,022
30,765,734,901
IssuesEvent
2023-07-30 09:19:54
openhab/openhab-distro
https://api.github.com/repos/openhab/openhab-distro
closed
Migration guide from v 2.x.x to 3.x.x Breaking changes link is broken
documentation
Official docs https://www.openhab.org/docs/configuration/migration/ guides the user to the wiki page "Breaking Changes in openHAB 3" (https://github.com/openhab/openhab-distro/wiki/Breaking-Changes-in-openHAB-3), which in turn guides the user to another document (https://github.com/openhab/openhab-distro/releases/tag/3.0.0#breaking-changes-that-require-manual-interaction-after-the-upgrade). It has a quick link "Breaking Changes" which refers the document that does not exist: https://github.com/openhab/openhab-distro/blob/3.0.0/breaking-changes-that-require-manual-interaction-after-the-upgrade. As a result there is no way to see the breaking changes when migrating from v2.x.x to v3.x.x
1.0
Migration guide from v 2.x.x to 3.x.x Breaking changes link is broken - Official docs https://www.openhab.org/docs/configuration/migration/ guides the user to the wiki page "Breaking Changes in openHAB 3" (https://github.com/openhab/openhab-distro/wiki/Breaking-Changes-in-openHAB-3), which in turn guides the user to another document (https://github.com/openhab/openhab-distro/releases/tag/3.0.0#breaking-changes-that-require-manual-interaction-after-the-upgrade). It has a quick link "Breaking Changes" which refers the document that does not exist: https://github.com/openhab/openhab-distro/blob/3.0.0/breaking-changes-that-require-manual-interaction-after-the-upgrade. As a result there is no way to see the breaking changes when migrating from v2.x.x to v3.x.x
non_main
migration guide from v x x to x x breaking changes link is broken official docs guides the user to the wiki page breaking changes in openhab which in turn guides the user to another document it has a quick link breaking changes which refers the document that does not exist as a result there is no way to see the breaking changes when migrating from x x to x x
0
251,865
18,977,314,437
IssuesEvent
2021-11-20 07:44:10
corona-warn-app/cwa-wishlist
https://api.github.com/repos/corona-warn-app/cwa-wishlist
closed
Vaccination certificate from another country
documentation enhancement
<!-- Thanks for pointing us to missing information 🙌 ❤️ Before opening a new issue, please make sure that we do not have any duplicates already open. You can ensure this by searching the issue list for this repository. If there is a duplicate, please close your issue and add a comment to the existing issue instead. To browse existing issues by category, please see these overview issues: https://github.com/corona-warn-app/cwa-wishlist/issues/337 Specifically, please check if your suggestion has already been raised here: https://github.com/corona-warn-app/cwa-wishlist/issues/394 --> ## What is missing <!-- Outline the information that you would like to see added. Please be rather specific (e.g., not only 'more information about', but what exactly is missing). --> I couldn’t find if the app supports loading certificates generated in other EU countries, and if it’s possible to mix them up. For example, if one gets the first two doses in Germany and the booster in Italy or Spain. ## Why should it be included <!-- Which aspects of the corona warn app project cannot be properly understood without this information? --> Considering that we will probably get one dose every 6 to 12 months, more and more people will have a mix situation and would like to make sure that the app supports it. ## Where should it be included <!-- If you think the information should be part of a specific existing document, please let us know. --> Not sure.
1.0
Vaccination certificate from another country - <!-- Thanks for pointing us to missing information 🙌 ❤️ Before opening a new issue, please make sure that we do not have any duplicates already open. You can ensure this by searching the issue list for this repository. If there is a duplicate, please close your issue and add a comment to the existing issue instead. To browse existing issues by category, please see these overview issues: https://github.com/corona-warn-app/cwa-wishlist/issues/337 Specifically, please check if your suggestion has already been raised here: https://github.com/corona-warn-app/cwa-wishlist/issues/394 --> ## What is missing <!-- Outline the information that you would like to see added. Please be rather specific (e.g., not only 'more information about', but what exactly is missing). --> I couldn’t find if the app supports loading certificates generated in other EU countries, and if it’s possible to mix them up. For example, if one gets the first two doses in Germany and the booster in Italy or Spain. ## Why should it be included <!-- Which aspects of the corona warn app project cannot be properly understood without this information? --> Considering that we will probably get one dose every 6 to 12 months, more and more people will have a mix situation and would like to make sure that the app supports it. ## Where should it be included <!-- If you think the information should be part of a specific existing document, please let us know. --> Not sure.
non_main
vaccination certificate from another country thanks for pointing us to missing information 🙌 ❤️ before opening a new issue please make sure that we do not have any duplicates already open you can ensure this by searching the issue list for this repository if there is a duplicate please close your issue and add a comment to the existing issue instead to browse existing issues by category please see these overview issues specifically please check if your suggestion has already been raised here what is missing i couldn’t find if the app supports loading certificates generated in other eu countries and if it’s possible to mix them up for example if one gets the first two doses in germany and the booster in italy or spain why should it be included considering that we will probably get one dose every to months more and more people will have a mix situation and would like to make sure that the app supports it where should it be included not sure
0
5,201
26,440,582,887
IssuesEvent
2023-01-15 22:49:04
MarcusWolschon/osmeditor4android
https://api.github.com/repos/MarcusWolschon/osmeditor4android
opened
Rework GPX file loading
Maintainability
A lot of the GPX file loading code has no function anymore since everything is being done in setupLayers, this should be cleaned up.
True
Rework GPX file loading - A lot of the GPX file loading code has no function anymore since everything is being done in setupLayers, this should be cleaned up.
main
rework gpx file loading a lot of the gpx file loading code has no function anymore since everything is being done in setuplayers this should be cleaned up
1
4,170
19,985,454,752
IssuesEvent
2022-01-30 15:41:55
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
opened
[MAINTAIN] r-stan
maintain
<!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** <details> ``` rix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes] /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h: In instantiation of ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’: /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:300:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 1>’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:551:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 3>’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseBase.h:41:34: required from ‘class Eigen::DenseBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MatrixBase.h:48:34: required from ‘class Eigen::MatrixBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:37:34: required from ‘class Eigen::MapBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:221:34: [ skipping 20 instantiation contexts, use -ftemplate-backtrace-limit=0 to disable ] /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:251:64: required from ‘static void Eigen::internal::generic_product_impl<Lhs, Rhs, Eigen::DenseShape, Eigen::DenseShape, 6>::evalTo(Dst&, const Lhs&, const Rhs&) [with Dst = Eigen::Matrix<double, 1, 1, 0, 1, 1>; Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:124:75: required from ‘Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::product_evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes] g++ -std=gnu++14 -shared -L/usr/lib64/R/lib -Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now -o rstan.so Module.o chains.o init.o misc.o pointer-tools.o sparse_extractors.o stan_fit_base.o stan_fit_rccp.o stanc.o stan/lang/ast_def.o stan/lang/grammars/bare_type_grammar_inst.o stan/lang/grammars/block_var_decls_grammar_inst.o stan/lang/grammars/expression07_grammar_inst.o stan/lang/grammars/expression_grammar_inst.o stan/lang/grammars/functions_grammar_inst.o stan/lang/grammars/indexes_grammar_inst.o stan/lang/grammars/local_var_decls_grammar_inst.o stan/lang/grammars/program_grammar_inst.o stan/lang/grammars/semantic_actions_def.o stan/lang/grammars/statement_2_grammar_inst.o stan/lang/grammars/statement_grammar_inst.o stan/lang/grammars/term_grammar_inst.o stan/lang/grammars/whitespace_grammar_inst.o -L/usr/lib64/R/lib -lR ar -rs ../inst/lib//libStanServices.a stan_fit.o stan_fit_base.o ar: creating ../inst/lib//libStanServices.a installing to /build/r-rstan/src/00LOCK-rstan/00new/rstan/libs ** R ** inst ** byte-compile and prepare package for lazy loading Error in as.POSIXlt.character(x, tz, ...) : character string is not in a standard unambiguous format Error: unable to load R code in package ‘rstan’ Execution halted ERROR: lazy loading failed for package ‘rstan’ ``` </details> **Packages (please complete the following information):** - Package Name: r-rstan **Description** https://log.bioarchlinux.org/2022-01-28T13%3A17%3A39/r-rstan.log
True
[MAINTAIN] r-stan - <!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** <details> ``` rix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes] /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h: In instantiation of ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’: /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:300:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 1>’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:551:7: required from ‘class Eigen::DenseCoeffsBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 3>’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseBase.h:41:34: required from ‘class Eigen::DenseBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MatrixBase.h:48:34: required from ‘class Eigen::MatrixBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true> >’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:37:34: required from ‘class Eigen::MapBase<Eigen::Block<Eigen::Transpose<Eigen::Matrix<double, 1, -1> >, -1, 1, true>, 0>’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/MapBase.h:221:34: [ skipping 20 instantiation contexts, use -ftemplate-backtrace-limit=0 to disable ] /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:251:64: required from ‘static void Eigen::internal::generic_product_impl<Lhs, Rhs, Eigen::DenseShape, Eigen::DenseShape, 6>::evalTo(Dst&, const Lhs&, const Rhs&) [with Dst = Eigen::Matrix<double, 1, 1, 0, 1, 1>; Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:124:75: required from ‘Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::product_evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; int ProductTag = 6; LhsShape = Eigen::DenseShape; RhsShape = Eigen::DenseShape; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Rhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Rhs = Eigen::Matrix<double, -1, 1>; typename Eigen::internal::traits<typename Eigen::Product<Lhs, Rhs, Option>::Lhs>::Scalar = double; typename Eigen::Product<Lhs, Rhs, Option>::Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Eigen::internal::product_evaluator<Eigen::Product<Lhs, Rhs, Option>, ProductTag, LhsShape, RhsShape>::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/ProductEvaluators.h:35:90: required from ‘Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::evaluator(const XprType&) [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Options = 0; Eigen::internal::evaluator<Eigen::Product<Lhs, Rhs, Option> >::XprType = Eigen::Product<Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>, Eigen::Matrix<double, -1, 1>, 0>]’ /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/Product.h:132:22: required from ‘Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::operator const Scalar() const [with Lhs = Eigen::Product<Eigen::CwiseBinaryOp<Eigen::internal::scalar_product_op<double, double>, const Eigen::CwiseNullaryOp<Eigen::internal::scalar_constant_op<double>, const Eigen::Matrix<double, 1, -1> >, const Eigen::Transpose<Eigen::Matrix<double, -1, 1> > >, Eigen::Matrix<double, -1, -1>, 0>; Rhs = Eigen::Matrix<double, -1, 1>; int Option = 0; Eigen::internal::dense_product_base<Lhs, Rhs, Option, 6>::Scalar = double]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:23:56: required from ‘double stan::mcmc::dense_e_metric<Model, BaseRNG>::T(stan::mcmc::dense_e_point&) [with Model = stan::model::model_base; BaseRNG = boost::random::additive_combine_engine<boost::random::linear_congruential_engine<unsigned int, 40014, 0, 2147483563>, boost::random::linear_congruential_engine<unsigned int, 40692, 0, 2147483399> >]’ ./stan/mcmc/hmc/hamiltonians/dense_e_metric.hpp:22:10: required from here /usr/lib/R/library/RcppEigen/include/Eigen/src/Core/DenseCoeffsBase.h:55:30: warning: ignoring attributes on template argument ‘Eigen::internal::packet_traits<double>::type’ {aka ‘__m128d’} [-Wignored-attributes] g++ -std=gnu++14 -shared -L/usr/lib64/R/lib -Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now -o rstan.so Module.o chains.o init.o misc.o pointer-tools.o sparse_extractors.o stan_fit_base.o stan_fit_rccp.o stanc.o stan/lang/ast_def.o stan/lang/grammars/bare_type_grammar_inst.o stan/lang/grammars/block_var_decls_grammar_inst.o stan/lang/grammars/expression07_grammar_inst.o stan/lang/grammars/expression_grammar_inst.o stan/lang/grammars/functions_grammar_inst.o stan/lang/grammars/indexes_grammar_inst.o stan/lang/grammars/local_var_decls_grammar_inst.o stan/lang/grammars/program_grammar_inst.o stan/lang/grammars/semantic_actions_def.o stan/lang/grammars/statement_2_grammar_inst.o stan/lang/grammars/statement_grammar_inst.o stan/lang/grammars/term_grammar_inst.o stan/lang/grammars/whitespace_grammar_inst.o -L/usr/lib64/R/lib -lR ar -rs ../inst/lib//libStanServices.a stan_fit.o stan_fit_base.o ar: creating ../inst/lib//libStanServices.a installing to /build/r-rstan/src/00LOCK-rstan/00new/rstan/libs ** R ** inst ** byte-compile and prepare package for lazy loading Error in as.POSIXlt.character(x, tz, ...) : character string is not in a standard unambiguous format Error: unable to load R code in package ‘rstan’ Execution halted ERROR: lazy loading failed for package ‘rstan’ ``` </details> **Packages (please complete the following information):** - Package Name: r-rstan **Description** https://log.bioarchlinux.org/2022-01-28T13%3A17%3A39/r-rstan.log
main
r stan please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug rix eigen matrix rhs eigen matrix int options int producttag lhsshape eigen denseshape rhsshape eigen denseshape typename eigen internal traits rhs scalar double typename eigen product rhs eigen matrix typename eigen internal traits lhs scalar double typename eigen product lhs eigen product const eigen cwisenullaryop const eigen matrix const eigen transpose eigen matrix eigen internal product evaluator producttag lhsshape rhsshape xprtype eigen product const eigen cwisenullaryop const eigen matrix const eigen transpose eigen matrix eigen matrix ’ usr lib r library rcppeigen include eigen src core productevaluators h required from ‘eigen internal evaluator evaluator const xprtype ’ usr lib r library rcppeigen include eigen src core product h required from ‘eigen internal dense product base operator const scalar const ’ stan mcmc hmc hamiltonians dense e metric hpp required from ‘double stan mcmc dense e metric t stan mcmc dense e point ’ stan mcmc hmc hamiltonians dense e metric hpp required from here usr lib r library rcppeigen include eigen src core densecoeffsbase h warning ignoring attributes on template argument ‘eigen internal packet traits type’ aka ‘ ’ usr lib r library rcppeigen include eigen src core densecoeffsbase h in instantiation of ‘class eigen densecoeffsbase true ’ usr lib r library rcppeigen include eigen src core densecoeffsbase h required from ‘class eigen densecoeffsbase true ’ usr lib r library rcppeigen include eigen src core densecoeffsbase h required from ‘class eigen densecoeffsbase true ’ usr lib r library rcppeigen include eigen src core densebase h required from ‘class eigen densebase true ’ usr lib r library rcppeigen include eigen src core matrixbase h required from ‘class eigen matrixbase true ’ usr lib r library rcppeigen include eigen src core mapbase h required from ‘class eigen mapbase true ’ usr lib r library rcppeigen include eigen src core mapbase h usr lib r library rcppeigen include eigen src core productevaluators h required from ‘static void eigen internal generic product impl evalto dst const lhs const rhs ’ usr lib r library rcppeigen include eigen src core productevaluators h required from ‘eigen internal product evaluator producttag lhsshape rhsshape product evaluator const xprtype ’ usr lib r library rcppeigen include eigen src core productevaluators h required from ‘eigen internal evaluator evaluator const xprtype ’ usr lib r library rcppeigen include eigen src core product h required from ‘eigen internal dense product base operator const scalar const ’ stan mcmc hmc hamiltonians dense e metric hpp required from ‘double stan mcmc dense e metric t stan mcmc dense e point ’ stan mcmc hmc hamiltonians dense e metric hpp required from here usr lib r library rcppeigen include eigen src core densecoeffsbase h warning ignoring attributes on template argument ‘eigen internal packet traits type’ aka ‘ ’ g std gnu shared l usr r lib wl sort common as needed z relro z now o rstan so module o chains o init o misc o pointer tools o sparse extractors o stan fit base o stan fit rccp o stanc o stan lang ast def o stan lang grammars bare type grammar inst o stan lang grammars block var decls grammar inst o stan lang grammars grammar inst o stan lang grammars expression grammar inst o stan lang grammars functions grammar inst o stan lang grammars indexes grammar inst o stan lang grammars local var decls grammar inst o stan lang grammars program grammar inst o stan lang grammars semantic actions def o stan lang grammars statement grammar inst o stan lang grammars statement grammar inst o stan lang grammars term grammar inst o stan lang grammars whitespace grammar inst o l usr r lib lr ar rs inst lib libstanservices a stan fit o stan fit base o ar creating inst lib libstanservices a installing to build r rstan src rstan rstan libs r inst byte compile and prepare package for lazy loading error in as posixlt character x tz character string is not in a standard unambiguous format error unable to load r code in package ‘rstan’ execution halted error lazy loading failed for package ‘rstan’ packages please complete the following information package name r rstan description
1
1,535
6,572,225,490
IssuesEvent
2017-09-11 00:17:08
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
nmcli: conn_name required
affects_2.1 bug_report docs_report networking waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME nmcli ##### ANSIBLE VERSION ``` # latest head of 2016-04-16 ansible 2.1.0 config file = configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT CentOS Linux release 7.2.1511 (Core) ##### SUMMARY Conn Name Required althouth stated otherwise in documentation: > Where conn_name will be the name used to call the connection. **when not provided a default name is generated: `<type>[-<ifname>][-<num>]`** ##### STEPS TO REPRODUCE ``` - name: Configure network adapter nmcli: type=ethernet ip4="192.168.1.1/16" ifname="enp0s1" autoconnect=yes state=present ``` ##### EXPECTED RESULTS OK ##### ACTUAL RESULTS ``` FAILED! => {"changed": false, "failed": true, "msg": "missing required arguments: conn_name"} ```
True
nmcli: conn_name required - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME nmcli ##### ANSIBLE VERSION ``` # latest head of 2016-04-16 ansible 2.1.0 config file = configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT CentOS Linux release 7.2.1511 (Core) ##### SUMMARY Conn Name Required althouth stated otherwise in documentation: > Where conn_name will be the name used to call the connection. **when not provided a default name is generated: `<type>[-<ifname>][-<num>]`** ##### STEPS TO REPRODUCE ``` - name: Configure network adapter nmcli: type=ethernet ip4="192.168.1.1/16" ifname="enp0s1" autoconnect=yes state=present ``` ##### EXPECTED RESULTS OK ##### ACTUAL RESULTS ``` FAILED! => {"changed": false, "failed": true, "msg": "missing required arguments: conn_name"} ```
main
nmcli conn name required issue type bug report component name nmcli ansible version latest head of ansible config file configured module search path default w o overrides os environment centos linux release core summary conn name required althouth stated otherwise in documentation where conn name will be the name used to call the connection when not provided a default name is generated steps to reproduce name configure network adapter nmcli type ethernet ifname autoconnect yes state present expected results ok actual results failed changed false failed true msg missing required arguments conn name
1
2,925
10,427,223,538
IssuesEvent
2019-09-16 19:25:38
NaluKit/nalu
https://api.github.com/repos/NaluKit/nalu
closed
Remove errorRoute implementation
maintainance
With the new upcoming error event feature, it gets clearer, that supporting both features will end up in a lot of more code. Some methods gets inconsistent and make the framework harder to use. Due to this fact, the errorRoute implementation gets removed. The error implementation of Nalu v1.0.0 can be easily reused. There will be an example how to do this.
True
Remove errorRoute implementation - With the new upcoming error event feature, it gets clearer, that supporting both features will end up in a lot of more code. Some methods gets inconsistent and make the framework harder to use. Due to this fact, the errorRoute implementation gets removed. The error implementation of Nalu v1.0.0 can be easily reused. There will be an example how to do this.
main
remove errorroute implementation with the new upcoming error event feature it gets clearer that supporting both features will end up in a lot of more code some methods gets inconsistent and make the framework harder to use due to this fact the errorroute implementation gets removed the error implementation of nalu can be easily reused there will be an example how to do this
1
94,561
10,831,668,028
IssuesEvent
2019-11-11 08:53:25
p-amonpitakpun/clock-drawing-dnn
https://api.github.com/repos/p-amonpitakpun/clock-drawing-dnn
closed
Metric Result Exporting.
documentation enhancement question
How can we export the result of the classifier so that we can use in the presentation and document (such as putting it in README) ?
1.0
Metric Result Exporting. - How can we export the result of the classifier so that we can use in the presentation and document (such as putting it in README) ?
non_main
metric result exporting how can we export the result of the classifier so that we can use in the presentation and document such as putting it in readme
0
550,703
16,130,585,677
IssuesEvent
2021-04-29 03:39:24
rainlab/translate-plugin
https://api.github.com/repos/rainlab/translate-plugin
closed
Back-end locale switcher single vs. multi field default behavior
Priority: Low Type: Discussion
There's been some question in our shop about whether from a UX perspective the back-end locale switcher should default to switching all translatable fields (the current CTRL/CMD click behavior), with the CTRL/CMD click being used to change only one field. I wanted to pose this question to see if there's agreement on that front, or if we'd need to customize if we decide that's the behavior we want to offer our clients.
1.0
Back-end locale switcher single vs. multi field default behavior - There's been some question in our shop about whether from a UX perspective the back-end locale switcher should default to switching all translatable fields (the current CTRL/CMD click behavior), with the CTRL/CMD click being used to change only one field. I wanted to pose this question to see if there's agreement on that front, or if we'd need to customize if we decide that's the behavior we want to offer our clients.
non_main
back end locale switcher single vs multi field default behavior there s been some question in our shop about whether from a ux perspective the back end locale switcher should default to switching all translatable fields the current ctrl cmd click behavior with the ctrl cmd click being used to change only one field i wanted to pose this question to see if there s agreement on that front or if we d need to customize if we decide that s the behavior we want to offer our clients
0
570
4,047,242,928
IssuesEvent
2016-05-23 03:35:00
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Rust Cargo Packages: should accept the term "crate"
Maintainer Input Requested Suggestion Triggering
The unit of compilation in Rust is a crate, and packages can only contain a single library crate, so when searching, people are frequently going to want to search for a "rust crate" or "cargo crate". You should be able to use "cargo crate time" in the same way you would use "cargo package time". ------ IA Page: http://duck.co/ia/view/rust_cargo [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @TomBebbington
True
Rust Cargo Packages: should accept the term "crate" - The unit of compilation in Rust is a crate, and packages can only contain a single library crate, so when searching, people are frequently going to want to search for a "rust crate" or "cargo crate". You should be able to use "cargo crate time" in the same way you would use "cargo package time". ------ IA Page: http://duck.co/ia/view/rust_cargo [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @TomBebbington
main
rust cargo packages should accept the term crate the unit of compilation in rust is a crate and packages can only contain a single library crate so when searching people are frequently going to want to search for a rust crate or cargo crate you should be able to use cargo crate time in the same way you would use cargo package time ia page tombebbington
1
403
3,460,612,650
IssuesEvent
2015-12-19 09:33:04
espeak-ng/espeak-ng
https://api.github.com/repos/espeak-ng/espeak-ng
closed
Remove redundant `end of ...` comments
maintainability resolved/fixed
Various functions inconsistently have `{//===...` at the start and `} // end of ...` comments. These make the code harder to read and should be removed.
True
Remove redundant `end of ...` comments - Various functions inconsistently have `{//===...` at the start and `} // end of ...` comments. These make the code harder to read and should be removed.
main
remove redundant end of comments various functions inconsistently have at the start and end of comments these make the code harder to read and should be removed
1
1,370
5,933,073,146
IssuesEvent
2017-05-24 10:59:41
aroberge/reeborg
https://api.github.com/repos/aroberge/reeborg
closed
Remove RUR.CURRENT_WORLD almost everywhere
easier to maintain
The world state is put in a global variable named `RUR.CURRENT_WORLD`. If possible, it should be removed from almost all functions, except when it needs to be created. Elsewhere, an exported function should be used to access it.
True
Remove RUR.CURRENT_WORLD almost everywhere - The world state is put in a global variable named `RUR.CURRENT_WORLD`. If possible, it should be removed from almost all functions, except when it needs to be created. Elsewhere, an exported function should be used to access it.
main
remove rur current world almost everywhere the world state is put in a global variable named rur current world if possible it should be removed from almost all functions except when it needs to be created elsewhere an exported function should be used to access it
1
20,469
15,571,286,835
IssuesEvent
2021-03-17 04:34:08
SanderMertens/flecs
https://api.github.com/repos/SanderMertens/flecs
opened
Update API & query parser to new entity-relation terminology/notation
enhancement usability
**Describe the problem you are trying to solve.** With the upcoming new query implementation (#295) the primary use case for traits will be as a way to store relationships between entities. The terminology used by the current API does not make this obvious or intuitive. Additionally, there are a few inconsistencies between the C and C++ API that need to be addressed (the C API uses the reverse order of specifying traits as the C++ API). **Describe the solution you'd like** Modify the API to use terminology as described here: https://github.com/SanderMertens/flecs/discussions/344
True
Update API & query parser to new entity-relation terminology/notation - **Describe the problem you are trying to solve.** With the upcoming new query implementation (#295) the primary use case for traits will be as a way to store relationships between entities. The terminology used by the current API does not make this obvious or intuitive. Additionally, there are a few inconsistencies between the C and C++ API that need to be addressed (the C API uses the reverse order of specifying traits as the C++ API). **Describe the solution you'd like** Modify the API to use terminology as described here: https://github.com/SanderMertens/flecs/discussions/344
non_main
update api query parser to new entity relation terminology notation describe the problem you are trying to solve with the upcoming new query implementation the primary use case for traits will be as a way to store relationships between entities the terminology used by the current api does not make this obvious or intuitive additionally there are a few inconsistencies between the c and c api that need to be addressed the c api uses the reverse order of specifying traits as the c api describe the solution you d like modify the api to use terminology as described here
0
3,835
16,686,917,247
IssuesEvent
2021-06-08 08:59:00
keptn/community
https://api.github.com/repos/keptn/community
closed
REQUEST: Maintainer membership for @grabnerandi @AloisReitbauer @danielkhan
membership:maintainer status:approved
### Multi-request for Keptn maintainer team Please note, that these people are already maintainers. This issue is created for bookkeeping purposes. * Andreas Grabner @grabnerandi * Alois Reitbauer @AloisReitbauer * Daniel Khan @danielkhan ### Requirements - [x] We have reviewed the community membership guidelines (https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md) - [x] We have enabled 2FA on my GitHub account. See https://github.com/settings/security - [x] We have subscribed to the [Keptn Slack channel](http://slack.keptn.sh/) - [x] We are actively contributing to 1 or more Keptn subprojects in terms of guiding the direction of the project - [x] We have two sponsors that meet the sponsor requirements listed in the community membership guidelines. Among other requirements, sponsors must be approvers or maintainers of at least one repository in the organization - [x] I have spoken to my sponsors ahead of this application, and they have agreed to sponsor my application ### Sponsors <!-- Replace (at) with the `@` sign --> - @christian-kreuzberger-dtx - @johannes-b Each sponsor should reply to this issue with the comment "*I support*". Please remember, it is an applicant's responsibility to get their sponsors' confirmation before submitting the request. ### List of contributions to the Keptn project The aforementioned members have been actively contributing to the Keptn project for several months and are driving the direction of the project. Their contributions range from design decisions, technical guidance, roadmap, and even code contributions to the project.
True
REQUEST: Maintainer membership for @grabnerandi @AloisReitbauer @danielkhan - ### Multi-request for Keptn maintainer team Please note, that these people are already maintainers. This issue is created for bookkeeping purposes. * Andreas Grabner @grabnerandi * Alois Reitbauer @AloisReitbauer * Daniel Khan @danielkhan ### Requirements - [x] We have reviewed the community membership guidelines (https://github.com/keptn/community/blob/master/COMMUNITY_MEMBERSHIP.md) - [x] We have enabled 2FA on my GitHub account. See https://github.com/settings/security - [x] We have subscribed to the [Keptn Slack channel](http://slack.keptn.sh/) - [x] We are actively contributing to 1 or more Keptn subprojects in terms of guiding the direction of the project - [x] We have two sponsors that meet the sponsor requirements listed in the community membership guidelines. Among other requirements, sponsors must be approvers or maintainers of at least one repository in the organization - [x] I have spoken to my sponsors ahead of this application, and they have agreed to sponsor my application ### Sponsors <!-- Replace (at) with the `@` sign --> - @christian-kreuzberger-dtx - @johannes-b Each sponsor should reply to this issue with the comment "*I support*". Please remember, it is an applicant's responsibility to get their sponsors' confirmation before submitting the request. ### List of contributions to the Keptn project The aforementioned members have been actively contributing to the Keptn project for several months and are driving the direction of the project. Their contributions range from design decisions, technical guidance, roadmap, and even code contributions to the project.
main
request maintainer membership for grabnerandi aloisreitbauer danielkhan multi request for keptn maintainer team please note that these people are already maintainers this issue is created for bookkeeping purposes andreas grabner grabnerandi alois reitbauer aloisreitbauer daniel khan danielkhan requirements we have reviewed the community membership guidelines we have enabled on my github account see we have subscribed to the we are actively contributing to or more keptn subprojects in terms of guiding the direction of the project we have two sponsors that meet the sponsor requirements listed in the community membership guidelines among other requirements sponsors must be approvers or maintainers of at least one repository in the organization i have spoken to my sponsors ahead of this application and they have agreed to sponsor my application sponsors christian kreuzberger dtx johannes b each sponsor should reply to this issue with the comment i support please remember it is an applicant s responsibility to get their sponsors confirmation before submitting the request list of contributions to the keptn project the aforementioned members have been actively contributing to the keptn project for several months and are driving the direction of the project their contributions range from design decisions technical guidance roadmap and even code contributions to the project
1
353,817
25,137,220,389
IssuesEvent
2022-11-09 19:39:34
Merck/gsDesign2
https://api.github.com/repos/Merck/gsDesign2
closed
Remove `simtrial` dependency
documentation
Could we include required functions from `simtrial` to make `gsDesign2` self-contained? cc: @nanxstats
1.0
Remove `simtrial` dependency - Could we include required functions from `simtrial` to make `gsDesign2` self-contained? cc: @nanxstats
non_main
remove simtrial dependency could we include required functions from simtrial to make self contained cc nanxstats
0
11,986
3,247,994,598
IssuesEvent
2015-10-17 00:10:07
realm/realm-cocoa
https://api.github.com/repos/realm/realm-cocoa
closed
RealmTests.testCompact fails on iOS 7 device
bug P2 tests
This happens consistently with the 64-bit iOS 7 device I just added to CI: https://ci.realm.io/job/objc_pr/2112/configuration=Release,swift_version=2.0,target=ios-device-objc/testReport/junit/(root)/RealmTests/testCompact/ Maybe the unit test assumes more than it should about the behavior of compaction? We have 3 options: 1. fix the code/test 2. disable the test for iOS 7 devices 3. remove the iOS 7 device from CI Option 1 is of course ideal, but if no one will be able to dig into this seriously for a bit, I propose option 2.
1.0
RealmTests.testCompact fails on iOS 7 device - This happens consistently with the 64-bit iOS 7 device I just added to CI: https://ci.realm.io/job/objc_pr/2112/configuration=Release,swift_version=2.0,target=ios-device-objc/testReport/junit/(root)/RealmTests/testCompact/ Maybe the unit test assumes more than it should about the behavior of compaction? We have 3 options: 1. fix the code/test 2. disable the test for iOS 7 devices 3. remove the iOS 7 device from CI Option 1 is of course ideal, but if no one will be able to dig into this seriously for a bit, I propose option 2.
non_main
realmtests testcompact fails on ios device this happens consistently with the bit ios device i just added to ci maybe the unit test assumes more than it should about the behavior of compaction we have options fix the code test disable the test for ios devices remove the ios device from ci option is of course ideal but if no one will be able to dig into this seriously for a bit i propose option
0
66,739
3,257,697,271
IssuesEvent
2015-10-20 18:58:29
openshift/origin
https://api.github.com/repos/openshift/origin
closed
oc deploy should allow the user to view the logs of the latest deployment
component/deployments kind/enhancement priority/P2
`oc deploy --logs` should show the logs of the current deployment (if available). It maybe should support looking at older logs.
1.0
oc deploy should allow the user to view the logs of the latest deployment - `oc deploy --logs` should show the logs of the current deployment (if available). It maybe should support looking at older logs.
non_main
oc deploy should allow the user to view the logs of the latest deployment oc deploy logs should show the logs of the current deployment if available it maybe should support looking at older logs
0
672,456
22,826,905,249
IssuesEvent
2022-07-12 09:23:02
bedita/manager
https://api.github.com/repos/bedita/manager
closed
Data Modeling Object Type modify
feature Priority - Normal UI / UX
Provide the following enhancements: - [x] Single custom property Hide/Show button should populate `Hidden` field properly (https://github.com/bedita/manager/pull/810) - [x] Single custom property Delete button should make an ajax call to api delete proper endpoint (https://github.com/bedita/manager/pull/812) - [x] Parent name should be an abstract objects only select combo (https://github.com/bedita/manager/pull/815) - [x] Table should be a select combo (static list for now, result of an api call in the future) (https://github.com/bedita/manager/pull/815) - [x] Associations should be a list of checkboxes (static list for now: DateRanges, Streams, Categories, Tags; there could be an extra value from api: show it as readonly checked) (https://github.com/bedita/manager/pull/816) - [x] Single custom property Type should be a property_types select combo (https://github.com/bedita/manager/pull/816)
1.0
Data Modeling Object Type modify - Provide the following enhancements: - [x] Single custom property Hide/Show button should populate `Hidden` field properly (https://github.com/bedita/manager/pull/810) - [x] Single custom property Delete button should make an ajax call to api delete proper endpoint (https://github.com/bedita/manager/pull/812) - [x] Parent name should be an abstract objects only select combo (https://github.com/bedita/manager/pull/815) - [x] Table should be a select combo (static list for now, result of an api call in the future) (https://github.com/bedita/manager/pull/815) - [x] Associations should be a list of checkboxes (static list for now: DateRanges, Streams, Categories, Tags; there could be an extra value from api: show it as readonly checked) (https://github.com/bedita/manager/pull/816) - [x] Single custom property Type should be a property_types select combo (https://github.com/bedita/manager/pull/816)
non_main
data modeling object type modify provide the following enhancements single custom property hide show button should populate hidden field properly single custom property delete button should make an ajax call to api delete proper endpoint parent name should be an abstract objects only select combo table should be a select combo static list for now result of an api call in the future associations should be a list of checkboxes static list for now dateranges streams categories tags there could be an extra value from api show it as readonly checked single custom property type should be a property types select combo
0
714,562
24,566,453,789
IssuesEvent
2022-10-13 03:49:17
AY2223S1-CS2113-T17-1/tp
https://api.github.com/repos/AY2223S1-CS2113-T17-1/tp
closed
[List] As an AOM, I can view the details of a passenger
type.Story priority.High
so that I am able to have an overview of the passenger list in terminal 1. Vignesh: Class creation (With accompanying methods) Ivan: implement in main class Due Date: 11th Oct 2022 (Tuesday)
1.0
[List] As an AOM, I can view the details of a passenger - so that I am able to have an overview of the passenger list in terminal 1. Vignesh: Class creation (With accompanying methods) Ivan: implement in main class Due Date: 11th Oct 2022 (Tuesday)
non_main
as an aom i can view the details of a passenger so that i am able to have an overview of the passenger list in terminal vignesh class creation with accompanying methods ivan implement in main class due date oct tuesday
0
149,010
13,253,535,284
IssuesEvent
2020-08-20 07:46:43
nolar/kopf
https://api.github.com/repos/nolar/kopf
opened
Fork Kopf
documentation
Kopf has a new GitHub home: * https://github.com/nolar/kopf. Announcement (a link to go around paywall): * https://medium.com/@nolar/kopf-is-forked-cdca40026ea7?source=friends_link&sk=d476cc32ec728382d664506d2cf08b69
1.0
Fork Kopf - Kopf has a new GitHub home: * https://github.com/nolar/kopf. Announcement (a link to go around paywall): * https://medium.com/@nolar/kopf-is-forked-cdca40026ea7?source=friends_link&sk=d476cc32ec728382d664506d2cf08b69
non_main
fork kopf kopf has a new github home announcement a link to go around paywall
0
2,418
8,580,906,017
IssuesEvent
2018-11-13 13:22:34
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
ec2_ami_copy: When using tag_equality with wait, error occurs
affects_2.8 aws bug cloud has_pr module needs_info needs_maintainer support:community traceback
<!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Explain the problem briefly below --> When using the ec2_ami_copy module with the tag_equality and wait parameters set to true, if the AMI exists, an error occurs. Unhandled exception. (local variable 'image_id' referenced before assignment) ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below --> ec2_ami_copy.py ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ansible 2.7.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ansible/custom_modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible python version = 2.7.5 (default, May 31 2018, 09:41:32) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> When tag_equality and wait parameters are set and AMI exists, the problem presents itself. <!--- Paste example playbooks or commands between quotes below --> ```yaml ec2_ami_copy_pdx: source_region: "us-east-1" source_image_id: "ami-0a65asfd4" name: TestAMI tags: Name: "TestAMI" tag_equality: yes wait: yes ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> The ec2_ami_copy module should not create a new AMI and should recognize that one exists. ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> An error message occurs when running the module. <!--- Paste verbatim command output between quotes --> ```paste below The full traceback is: WARNING: The below traceback may *not* be related to the actual failure. File "/tmp/ansible_ec2_ami_copy_payload_bJvVsP/__main__.py", line 193, in copy_image ImageIds=[image_id], fatal: [localhost]: FAILED! => changed=false invocation: module_args: aws_access_key: null aws_secret_key: null description: null ec2_url: null encrypted: false name: TestAMI profile: null region: null security_token: null source_image_id: ami-0a65asfd4 source_region: us-east-1 tag_equality: true tags: Name: TestAMI validate_certs: true wait: true wait_timeout: 600 msg: Unhandled exception. (local variable 'image_id' referenced before assignment) ```
True
ec2_ami_copy: When using tag_equality with wait, error occurs - <!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Explain the problem briefly below --> When using the ec2_ami_copy module with the tag_equality and wait parameters set to true, if the AMI exists, an error occurs. Unhandled exception. (local variable 'image_id' referenced before assignment) ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below --> ec2_ami_copy.py ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ansible 2.7.0 config file = /etc/ansible/ansible.cfg configured module search path = [u'/etc/ansible/custom_modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /usr/bin/ansible python version = 2.7.5 (default, May 31 2018, 09:41:32) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] ``` ##### OS / ENVIRONMENT <!--- Provide all relevant information below, e.g. target OS versions, network device firmware, etc. --> ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> When tag_equality and wait parameters are set and AMI exists, the problem presents itself. <!--- Paste example playbooks or commands between quotes below --> ```yaml ec2_ami_copy_pdx: source_region: "us-east-1" source_image_id: "ami-0a65asfd4" name: TestAMI tags: Name: "TestAMI" tag_equality: yes wait: yes ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- Describe what you expected to happen when running the steps above --> The ec2_ami_copy module should not create a new AMI and should recognize that one exists. ##### ACTUAL RESULTS <!--- Describe what actually happened. If possible run with extra verbosity (-vvvv) --> An error message occurs when running the module. <!--- Paste verbatim command output between quotes --> ```paste below The full traceback is: WARNING: The below traceback may *not* be related to the actual failure. File "/tmp/ansible_ec2_ami_copy_payload_bJvVsP/__main__.py", line 193, in copy_image ImageIds=[image_id], fatal: [localhost]: FAILED! => changed=false invocation: module_args: aws_access_key: null aws_secret_key: null description: null ec2_url: null encrypted: false name: TestAMI profile: null region: null security_token: null source_image_id: ami-0a65asfd4 source_region: us-east-1 tag_equality: true tags: Name: TestAMI validate_certs: true wait: true wait_timeout: 600 msg: Unhandled exception. (local variable 'image_id' referenced before assignment) ```
main
ami copy when using tag equality with wait error occurs summary when using the ami copy module with the tag equality and wait parameters set to true if the ami exists an error occurs unhandled exception local variable image id referenced before assignment issue type bug report component name ami copy py ansible version paste below configuration paste below ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible python version default may os environment steps to reproduce when tag equality and wait parameters are set and ami exists the problem presents itself yaml ami copy pdx source region us east source image id ami name testami tags name testami tag equality yes wait yes expected results the ami copy module should not create a new ami and should recognize that one exists actual results an error message occurs when running the module paste below the full traceback is warning the below traceback may not be related to the actual failure file tmp ansible ami copy payload bjvvsp main py line in copy image imageids fatal failed changed false invocation module args aws access key null aws secret key null description null url null encrypted false name testami profile null region null security token null source image id ami source region us east tag equality true tags name testami validate certs true wait true wait timeout msg unhandled exception local variable image id referenced before assignment
1
56,111
13,757,517,391
IssuesEvent
2020-10-06 21:48:37
sitmun/sitmun-backend-core
https://api.github.com/repos/sitmun/sitmun-backend-core
closed
Fix generation of API documentation
bug build
Generation of documentation does not work as expected. Swagger.json file is generated at the root of the project rather than at the build folder. In addition tests are run twice in Travis-CI because `convertSwagger2markup` always depends on tests.
1.0
Fix generation of API documentation - Generation of documentation does not work as expected. Swagger.json file is generated at the root of the project rather than at the build folder. In addition tests are run twice in Travis-CI because `convertSwagger2markup` always depends on tests.
non_main
fix generation of api documentation generation of documentation does not work as expected swagger json file is generated at the root of the project rather than at the build folder in addition tests are run twice in travis ci because always depends on tests
0
924
4,627,883,043
IssuesEvent
2016-09-28 00:59:40
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
junos_command: ValueError: need more than 1 value to unpack
affects_2.2 bug_report in progress networking P1 waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME junos_command ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel eb33ed4219) last updated 2016/09/27 09:18:44 (GMT +100) lib/ansible/modules/core: (devel c03697c81e) last updated 2016/09/27 09:18:49 (GMT +100) lib/ansible/modules/extras: (devel 119bc466be) last updated 2016/09/27 09:18:51 (GMT +100) ``` plus the following patch ``` diff --git a/lib/ansible/module_utils/junos.py b/lib/ansible/module_utils/junos.py index 9cf36da..d3e3cb2 100644 --- a/lib/ansible/module_utils/junos.py +++ b/lib/ansible/module_utils/junos.py @@ -171,7 +171,7 @@ class Netconf(object): return ele def load_config(self, config, commit=False, replace=False, confirm=None, - comment=None, config_format='text'): + comment=None, config_format='text', **kwargs): if replace: merge = False ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY Issue may need fixing in `module_utils/netcli.py`, though I'm triggering it from `junos_command`, so raising it in ansible-modules-core. ##### STEPS TO REPRODUCE ```yaml - name: test gt operator junos_command: commands: - show version - show interfaces fxp0 format: xml wait_for: - "stdout[1]['interface-information']['physical-interface'].mtu[0]" provider: "{{ netconf }}" register: result ``` ##### EXPECTED RESULTS ##### ACTUAL RESULTS ```yaml TASK [test_junos_command : test gt operator] *********************************** task path: /home/johnb/git/ansible-inc/test-junos/roles/test_junos_command/tests/netconf_xml/greaterthan.yaml:4 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/junos/junos_command.py <vsrx01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <vsrx01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" && echo ansible-tmp-1475001874.51-177586194093154="` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" ) && sleep 0' <vsrx01> PUT /tmp/tmpbDvPHR TO /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py <vsrx01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/ /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py && sleep 0' <vsrx01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 304, in <module> main() File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 269, in main runner.add_conditional(item) File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 153, in add_conditional File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 195, in __init__ ValueError: need more than 1 value to unpack fatal: [vsrx01]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "junos_command" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 304, in <module>\n main()\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 269, in main\n runner.add_conditional(item)\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 153, in add_conditional\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 195, in __init__\nValueError: need more than 1 value to unpack\n", "module_stdout": "", "msg": "MODULE FAILURE" } ```
True
junos_command: ValueError: need more than 1 value to unpack - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME junos_command ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel eb33ed4219) last updated 2016/09/27 09:18:44 (GMT +100) lib/ansible/modules/core: (devel c03697c81e) last updated 2016/09/27 09:18:49 (GMT +100) lib/ansible/modules/extras: (devel 119bc466be) last updated 2016/09/27 09:18:51 (GMT +100) ``` plus the following patch ``` diff --git a/lib/ansible/module_utils/junos.py b/lib/ansible/module_utils/junos.py index 9cf36da..d3e3cb2 100644 --- a/lib/ansible/module_utils/junos.py +++ b/lib/ansible/module_utils/junos.py @@ -171,7 +171,7 @@ class Netconf(object): return ele def load_config(self, config, commit=False, replace=False, confirm=None, - comment=None, config_format='text'): + comment=None, config_format='text', **kwargs): if replace: merge = False ``` ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY Issue may need fixing in `module_utils/netcli.py`, though I'm triggering it from `junos_command`, so raising it in ansible-modules-core. ##### STEPS TO REPRODUCE ```yaml - name: test gt operator junos_command: commands: - show version - show interfaces fxp0 format: xml wait_for: - "stdout[1]['interface-information']['physical-interface'].mtu[0]" provider: "{{ netconf }}" register: result ``` ##### EXPECTED RESULTS ##### ACTUAL RESULTS ```yaml TASK [test_junos_command : test gt operator] *********************************** task path: /home/johnb/git/ansible-inc/test-junos/roles/test_junos_command/tests/netconf_xml/greaterthan.yaml:4 Using module file /home/johnb/git/ansible-inc/ansible/lib/ansible/modules/core/network/junos/junos_command.py <vsrx01> ESTABLISH LOCAL CONNECTION FOR USER: johnb <vsrx01> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" && echo ansible-tmp-1475001874.51-177586194093154="` echo $HOME/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154 `" ) && sleep 0' <vsrx01> PUT /tmp/tmpbDvPHR TO /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py <vsrx01> EXEC /bin/sh -c 'chmod u+x /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/ /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py && sleep 0' <vsrx01> EXEC /bin/sh -c 'python /home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/junos_command.py; rm -rf "/home/johnb/.ansible/tmp/ansible-tmp-1475001874.51-177586194093154/" > /dev/null 2>&1 && sleep 0' An exception occurred during task execution. The full traceback is: Traceback (most recent call last): File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 304, in <module> main() File "/tmp/ansible_rnv3rr/ansible_module_junos_command.py", line 269, in main runner.add_conditional(item) File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 153, in add_conditional File "/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py", line 195, in __init__ ValueError: need more than 1 value to unpack fatal: [vsrx01]: FAILED! => { "changed": false, "failed": true, "invocation": { "module_name": "junos_command" }, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 304, in <module>\n main()\n File \"/tmp/ansible_rnv3rr/ansible_module_junos_command.py\", line 269, in main\n runner.add_conditional(item)\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 153, in add_conditional\n File \"/tmp/ansible_rnv3rr/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 195, in __init__\nValueError: need more than 1 value to unpack\n", "module_stdout": "", "msg": "MODULE FAILURE" } ```
main
junos command valueerror need more than value to unpack issue type bug report component name junos command ansible version ansible devel last updated gmt lib ansible modules core devel last updated gmt lib ansible modules extras devel last updated gmt plus the following patch diff git a lib ansible module utils junos py b lib ansible module utils junos py index a lib ansible module utils junos py b lib ansible module utils junos py class netconf object return ele def load config self config commit false replace false confirm none comment none config format text comment none config format text kwargs if replace merge false configuration os environment summary issue may need fixing in module utils netcli py though i m triggering it from junos command so raising it in ansible modules core steps to reproduce yaml name test gt operator junos command commands show version show interfaces format xml wait for stdout mtu provider netconf register result expected results actual results yaml task task path home johnb git ansible inc test junos roles test junos command tests netconf xml greaterthan yaml using module file home johnb git ansible inc ansible lib ansible modules core network junos junos command py establish local connection for user johnb exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpbdvphr to home johnb ansible tmp ansible tmp junos command py exec bin sh c chmod u x home johnb ansible tmp ansible tmp home johnb ansible tmp ansible tmp junos command py sleep exec bin sh c python home johnb ansible tmp ansible tmp junos command py rm rf home johnb ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible ansible module junos command py line in main file tmp ansible ansible module junos command py line in main runner add conditional item file tmp ansible ansible modlib zip ansible module utils netcli py line in add conditional file tmp ansible ansible modlib zip ansible module utils netcli py line in init valueerror need more than value to unpack fatal failed changed false failed true invocation module name junos command module stderr traceback most recent call last n file tmp ansible ansible module junos command py line in n main n file tmp ansible ansible module junos command py line in main n runner add conditional item n file tmp ansible ansible modlib zip ansible module utils netcli py line in add conditional n file tmp ansible ansible modlib zip ansible module utils netcli py line in init nvalueerror need more than value to unpack n module stdout msg module failure
1
3,342
12,962,193,174
IssuesEvent
2020-07-20 16:46:08
short-d/short
https://api.github.com/repos/short-d/short
closed
[Refactor] Update user short link relation in Usecase
maintainability
**What is frustrating you?** When updating the alias of a short link, the relation of user short link is updated only because they are joint in postgres database but it is not updated in updater usecase https://github.com/short-d/short/blob/b945aac5192b0f4bb674fd6dfdbdb7d7b1f05fd3/backend/app/usecase/shortlink/updater.go#L80-L81 **Your solution** Update user short link relation in Usecase
True
[Refactor] Update user short link relation in Usecase - **What is frustrating you?** When updating the alias of a short link, the relation of user short link is updated only because they are joint in postgres database but it is not updated in updater usecase https://github.com/short-d/short/blob/b945aac5192b0f4bb674fd6dfdbdb7d7b1f05fd3/backend/app/usecase/shortlink/updater.go#L80-L81 **Your solution** Update user short link relation in Usecase
main
update user short link relation in usecase what is frustrating you when updating the alias of a short link the relation of user short link is updated only because they are joint in postgres database but it is not updated in updater usecase your solution update user short link relation in usecase
1
4,096
19,323,198,690
IssuesEvent
2021-12-14 08:37:45
WarenGonzaga/daisy.js
https://api.github.com/repos/WarenGonzaga/daisy.js
opened
move site to docs folder
chore maintainers only tweak
It is more reasonable to move the demo site to the docs folder instead. 👌
True
move site to docs folder - It is more reasonable to move the demo site to the docs folder instead. 👌
main
move site to docs folder it is more reasonable to move the demo site to the docs folder instead 👌
1
5,613
28,095,593,858
IssuesEvent
2023-03-30 15:34:57
MozillaFoundation/foundation.mozilla.org
https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org
opened
Implement security relevant upgrades
engineering maintain needs grooming security
## Description See this list of security relevant upgraded: https://github.com/MozillaFoundation/foundation.mozilla.org/security/dependabot ## Acceptance criteria - [ ] All security relevant upgrades have been implemented.
True
Implement security relevant upgrades - ## Description See this list of security relevant upgraded: https://github.com/MozillaFoundation/foundation.mozilla.org/security/dependabot ## Acceptance criteria - [ ] All security relevant upgrades have been implemented.
main
implement security relevant upgrades description see this list of security relevant upgraded acceptance criteria all security relevant upgrades have been implemented
1
1,477
6,404,174,472
IssuesEvent
2017-08-07 01:23:43
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
microsoft-office uninstall does not remove .app files from /Applications
awaiting maintainer feedback
#### General troubleshooting steps - [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [X] None of the templates was appropriate for my issue, or I’m not sure. - [X] I ran `brew update-reset && brew update` and retried my command. - [X] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue `brew cask zap microsoft-office` should get rid of the apps, but the apps (microsoft-office is a suite) remain on /Applications/. #### Output of your command with `--verbose --debug` ``` [I] gtklocker@schwarz ~> brew cask zap microsoft-office --verbose --debug ==> Zapping Cask microsoft-office ==> Implied "brew cask uninstall microsoft-office" ==> Un-installing artifacts ==> Determining which artifacts are present in Cask microsoft-office ==> 3 artifact/s defined #<Hbc::Artifact::Uninstall:0x007fe6ad109b90> #<Hbc::Artifact::Pkg:0x007fe6ad109758> #<Hbc::Artifact::Zap:0x007fe6ad109460> ==> Un-installing artifact of class Hbc::Artifact::Uninstall ==> Running uninstall process for microsoft-office; your password may be necessary ==> Removing launchctl service com.microsoft.autoupdate.helpertool ==> Executing: ["/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"] ==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"] Password: ==> Removing launchctl service com.microsoft.office.licensing.helper ==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensing.helper"] ==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensing.helper"] ==> Removing launchctl service com.microsoft.office.licensingV2.helper ==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"] ==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"] ==> Uninstalling packages: ==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.package.*"] ==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.pkg.licensing"] ==> Dispatching zap stanza ==> Running zap process for microsoft-office; your password may be necessary ==> Removing files: ~/Library/Application Scripts/com.microsoft.Excel ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Office365ServiceV2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Outlook ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Powerpoint ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Word ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.errorreporting ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.onenote.mac ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.excel.sfl ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.powerpoint.sfl ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.word.sfl ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate.fba ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/com.microsoft.autoupdate.fba ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/com.microsoft.autoupdate2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Excel ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Office365ServiceV2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Outlook ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Powerpoint ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Word ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.errorreporting ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.onenote.mac ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Cookies/com.microsoft.autoupdate.fba.binarycookies ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Cookies/com.microsoft.autoupdate2.binarycookies ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Group Containers/UBF8T346G9.Office ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Group Containers/UBF8T346G9.OfficeOsfWebHost ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Group Containers/UBF8T346G9.ms ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.Excel.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.Powerpoint.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.Word.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.autoupdate.fba.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.autoupdate2.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Saved Application State/com.microsoft.autoupdate2.savedState ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Saved Application State/com.microsoft.office.setupassistant.savedState ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ==> Removing directories if empty: ~/Library/Caches/Microsoft/uls ~/Library/Caches/Microsoft ==> Removing all staged versions of Cask 'microsoft-office' ==> Purging all staged versions of Cask microsoft-office [I] gtklocker@schwarz ~> ls -d1 /Applications/Microsoft\ * /Applications/Microsoft Excel.app /Applications/Microsoft OneNote.app /Applications/Microsoft Outlook.app /Applications/Microsoft PowerPoint.app /Applications/Microsoft Word.app ``` #### Output of `brew cask doctor` ``` [I] gtklocker@schwarz ~> brew cask doctor ==> Homebrew-Cask Version Homebrew-Cask 1.3.0-39-gf57a172 caskroom/homebrew-cask (git revision ebc89; last commit 2017-08-07) ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Cached Downloads ~/Library/Caches/Homebrew/Cask (47 files, 3.5GB) ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3676 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1107 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (160 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin16 /Library/Ruby/Site/2.0.0/universal-darwin16 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16 ==> Environment Variables LANG="en_GB.UTF-8" PATH="/usr/local/bin:/usr/local/opt/fzf/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/usr/local/bin/fish" ```
True
microsoft-office uninstall does not remove .app files from /Applications - #### General troubleshooting steps - [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [X] None of the templates was appropriate for my issue, or I’m not sure. - [X] I ran `brew update-reset && brew update` and retried my command. - [X] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue `brew cask zap microsoft-office` should get rid of the apps, but the apps (microsoft-office is a suite) remain on /Applications/. #### Output of your command with `--verbose --debug` ``` [I] gtklocker@schwarz ~> brew cask zap microsoft-office --verbose --debug ==> Zapping Cask microsoft-office ==> Implied "brew cask uninstall microsoft-office" ==> Un-installing artifacts ==> Determining which artifacts are present in Cask microsoft-office ==> 3 artifact/s defined #<Hbc::Artifact::Uninstall:0x007fe6ad109b90> #<Hbc::Artifact::Pkg:0x007fe6ad109758> #<Hbc::Artifact::Zap:0x007fe6ad109460> ==> Un-installing artifact of class Hbc::Artifact::Uninstall ==> Running uninstall process for microsoft-office; your password may be necessary ==> Removing launchctl service com.microsoft.autoupdate.helpertool ==> Executing: ["/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"] ==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.autoupdate.helpertool"] Password: ==> Removing launchctl service com.microsoft.office.licensing.helper ==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensing.helper"] ==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensing.helper"] ==> Removing launchctl service com.microsoft.office.licensingV2.helper ==> Executing: ["/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"] ==> Executing: ["/usr/bin/sudo", "-E", "--", "/bin/launchctl", "list", "com.microsoft.office.licensingV2.helper"] ==> Uninstalling packages: ==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.package.*"] ==> Executing: ["/usr/sbin/pkgutil", "--pkgs=com.microsoft.pkg.licensing"] ==> Dispatching zap stanza ==> Running zap process for microsoft-office; your password may be necessary ==> Removing files: ~/Library/Application Scripts/com.microsoft.Excel ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Office365ServiceV2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Outlook ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Powerpoint ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.Word ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.errorreporting ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Scripts/com.microsoft.onenote.mac ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.excel.sfl ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.powerpoint.sfl ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Application Support/com.apple.sharedfilelist/com.apple.LSSharedFileList.ApplicationRecentDocuments/com.microsoft.word.sfl ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate.fba ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/Microsoft/uls/com.microsoft.autoupdate2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/com.microsoft.autoupdate.fba ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Caches/com.microsoft.autoupdate2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Excel ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Office365ServiceV2 ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Outlook ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Powerpoint ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.Word ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.errorreporting ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Containers/com.microsoft.onenote.mac ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Cookies/com.microsoft.autoupdate.fba.binarycookies ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Cookies/com.microsoft.autoupdate2.binarycookies ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Group Containers/UBF8T346G9.Office ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Group Containers/UBF8T346G9.OfficeOsfWebHost ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Group Containers/UBF8T346G9.ms ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.Excel.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.Powerpoint.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.Word.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.autoupdate.fba.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Preferences/com.microsoft.autoupdate2.plist ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Saved Application State/com.microsoft.autoupdate2.savedState ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ~/Library/Saved Application State/com.microsoft.office.setupassistant.savedState ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/bin/xargs", "-0", "--", "/bin/rm", "-r", "-f", "--"] ==> Removing directories if empty: ~/Library/Caches/Microsoft/uls ~/Library/Caches/Microsoft ==> Removing all staged versions of Cask 'microsoft-office' ==> Purging all staged versions of Cask microsoft-office [I] gtklocker@schwarz ~> ls -d1 /Applications/Microsoft\ * /Applications/Microsoft Excel.app /Applications/Microsoft OneNote.app /Applications/Microsoft Outlook.app /Applications/Microsoft PowerPoint.app /Applications/Microsoft Word.app ``` #### Output of `brew cask doctor` ``` [I] gtklocker@schwarz ~> brew cask doctor ==> Homebrew-Cask Version Homebrew-Cask 1.3.0-39-gf57a172 caskroom/homebrew-cask (git revision ebc89; last commit 2017-08-07) ==> Homebrew-Cask Install Location <NONE> ==> Homebrew-Cask Staging Location /usr/local/Caskroom ==> Homebrew-Cask Cached Downloads ~/Library/Caches/Homebrew/Cask (47 files, 3.5GB) ==> Homebrew-Cask Taps: /usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask (3676 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts (1107 casks) /usr/local/Homebrew/Library/Taps/caskroom/homebrew-versions (160 casks) ==> Contents of $LOAD_PATH /usr/local/Homebrew/Library/Homebrew/cask/lib /usr/local/Homebrew/Library/Homebrew /Library/Ruby/Site/2.0.0 /Library/Ruby/Site/2.0.0/x86_64-darwin16 /Library/Ruby/Site/2.0.0/universal-darwin16 /Library/Ruby/Site /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16 /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16 ==> Environment Variables LANG="en_GB.UTF-8" PATH="/usr/local/bin:/usr/local/opt/fzf/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/Homebrew/Library/Homebrew/shims/scm" SHELL="/usr/local/bin/fish" ```
main
microsoft office uninstall does not remove app files from applications general troubleshooting steps i have checked the instructions for or before opening the issue none of the templates was appropriate for my issue or i’m not sure i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i understand that description of issue brew cask zap microsoft office should get rid of the apps but the apps microsoft office is a suite remain on applications output of your command with verbose debug gtklocker schwarz brew cask zap microsoft office verbose debug zapping cask microsoft office implied brew cask uninstall microsoft office un installing artifacts determining which artifacts are present in cask microsoft office artifact s defined un installing artifact of class hbc artifact uninstall running uninstall process for microsoft office your password may be necessary removing launchctl service com microsoft autoupdate helpertool executing executing password removing launchctl service com microsoft office licensing helper executing executing removing launchctl service com microsoft office helper executing executing uninstalling packages executing executing dispatching zap stanza running zap process for microsoft office your password may be necessary removing files library application scripts com microsoft excel executing library application scripts com microsoft executing library application scripts com microsoft outlook executing library application scripts com microsoft powerpoint executing library application scripts com microsoft word executing library application scripts com microsoft errorreporting executing library application scripts com microsoft onenote mac executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft excel sfl executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft powerpoint sfl executing library application support com apple sharedfilelist com apple lssharedfilelist applicationrecentdocuments com microsoft word sfl executing library caches microsoft uls com microsoft autoupdate fba executing library caches microsoft uls com microsoft executing library caches com microsoft autoupdate fba executing library caches com microsoft executing library containers com microsoft excel executing library containers com microsoft executing library containers com microsoft outlook executing library containers com microsoft powerpoint executing library containers com microsoft word executing library containers com microsoft errorreporting executing library containers com microsoft onenote mac executing library cookies com microsoft autoupdate fba binarycookies executing library cookies com microsoft binarycookies executing library group containers office executing library group containers officeosfwebhost executing library group containers ms executing library preferences com microsoft excel plist executing library preferences com microsoft powerpoint plist executing library preferences com microsoft word plist executing library preferences com microsoft autoupdate fba plist executing library preferences com microsoft plist executing library saved application state com microsoft savedstate executing library saved application state com microsoft office setupassistant savedstate executing removing directories if empty library caches microsoft uls library caches microsoft removing all staged versions of cask microsoft office purging all staged versions of cask microsoft office gtklocker schwarz ls applications microsoft applications microsoft excel app applications microsoft onenote app applications microsoft outlook app applications microsoft powerpoint app applications microsoft word app output of brew cask doctor gtklocker schwarz brew cask doctor homebrew cask version homebrew cask caskroom homebrew cask git revision last commit homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask cached downloads library caches homebrew cask files homebrew cask taps usr local homebrew library taps caskroom homebrew cask casks usr local homebrew library taps caskroom homebrew fonts casks usr local homebrew library taps caskroom homebrew versions casks contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal environment variables lang en gb utf path usr local bin usr local opt fzf bin usr bin bin usr sbin sbin opt bin usr local homebrew library homebrew shims scm shell usr local bin fish
1
506,425
14,664,989,563
IssuesEvent
2020-12-29 13:17:23
teamforus/forus
https://api.github.com/repos/teamforus/forus
closed
Time based criteria, prevalidations adjustments
Epic Priority: Must have project-107 project-108
## Main asssignee: @ ## Context/goal: - [ ] Fund configuration for how long a record_type is trusted - All record_types from criteria should have the trusted period (if you want to activate voucher directly) - Describe all edge cases if trusted period is different: - One record is valid, other is not - All records are invalid - [ ] create a prevalidation -> date of creation of prevalidation should added to record/record_validation - [ ] Requester should be able to get a new voucher when they don’t have an active voucher. - [ ] Use the expiration date of the voucher itself, ignore the expiration date of the fund. - Math.min(fund.expiration_date, voucher.expiration_date); - [ ] Replace or update ‘creation’/’trusted’ date of a prevalidation - [ ] Soft deleting a single prevalidation from table
1.0
Time based criteria, prevalidations adjustments - ## Main asssignee: @ ## Context/goal: - [ ] Fund configuration for how long a record_type is trusted - All record_types from criteria should have the trusted period (if you want to activate voucher directly) - Describe all edge cases if trusted period is different: - One record is valid, other is not - All records are invalid - [ ] create a prevalidation -> date of creation of prevalidation should added to record/record_validation - [ ] Requester should be able to get a new voucher when they don’t have an active voucher. - [ ] Use the expiration date of the voucher itself, ignore the expiration date of the fund. - Math.min(fund.expiration_date, voucher.expiration_date); - [ ] Replace or update ‘creation’/’trusted’ date of a prevalidation - [ ] Soft deleting a single prevalidation from table
non_main
time based criteria prevalidations adjustments main asssignee context goal fund configuration for how long a record type is trusted all record types from criteria should have the trusted period if you want to activate voucher directly describe all edge cases if trusted period is different one record is valid other is not all records are invalid create a prevalidation date of creation of prevalidation should added to record record validation requester should be able to get a new voucher when they don’t have an active voucher use the expiration date of the voucher itself ignore the expiration date of the fund math min fund expiration date voucher expiration date replace or update ‘creation’ ’trusted’ date of a prevalidation soft deleting a single prevalidation from table
0
2,308
8,271,873,700
IssuesEvent
2018-09-16 14:14:45
react-navigation/react-navigation
https://api.github.com/repos/react-navigation/react-navigation
reopened
Params are not passed from Stack navigator to Stack navigator if they are children of SwitchNavigator
needs action from maintainer
## The issue tracker is reserved for bug reports only. this is the bug only previous same bugs that seem are not fixed yet https://github.com/react-navigation/react-navigation/issues/143 https://github.com/react-navigation/react-navigation/issues/3787 ### Current Behavior Params are not passed from one screen to another. Screens are initial screens of different Stack navigators. **Config** ``` SwitchNavigator LoadingScreen: Component App: StackNavigator ProfileScreen DashboardScreen SignInFlow: StackNavigator LoginScreen // from this screen (and vice versa) LikeScreen SignUpFlow: StackNavigator SignUpScreen // to this screen (and vice versa) TCScreen ``` **Flow** From `LoginScreen` I try to navigate to `SignUpScreen` AND pass params. As you can see above, these screens belong to different stacks. Those stacks are children of `SwitchNavigator`. **Issue** Since `react-navigation v1` renders screen twice if I navigate from stack to stack (it is infamous bug of v1), I have to navigate to stack itself, not its particular screen. ```javascript // params are passed OK, but RN renders screen twice. So it is not an option this.props.navigation.navigate('LoginScreen', { ...this.state.data }) } // RN renders screen once as expected, but params are not being passed this.props.navigation.navigate('SignInFlow', { ...this.state.data }) } ``` ### Expected Behavior I get params in initial screen as if I was navigating to it, not to it's parent `Stacknavigator` ### How to reproduce Same way ### Your Environment | software | version | ---------------- | ------- | react-navigation | 1.5.8 | react-native |0.54.2 | node |8.11.3 | npm or yarn |5.6.0
True
Params are not passed from Stack navigator to Stack navigator if they are children of SwitchNavigator - ## The issue tracker is reserved for bug reports only. this is the bug only previous same bugs that seem are not fixed yet https://github.com/react-navigation/react-navigation/issues/143 https://github.com/react-navigation/react-navigation/issues/3787 ### Current Behavior Params are not passed from one screen to another. Screens are initial screens of different Stack navigators. **Config** ``` SwitchNavigator LoadingScreen: Component App: StackNavigator ProfileScreen DashboardScreen SignInFlow: StackNavigator LoginScreen // from this screen (and vice versa) LikeScreen SignUpFlow: StackNavigator SignUpScreen // to this screen (and vice versa) TCScreen ``` **Flow** From `LoginScreen` I try to navigate to `SignUpScreen` AND pass params. As you can see above, these screens belong to different stacks. Those stacks are children of `SwitchNavigator`. **Issue** Since `react-navigation v1` renders screen twice if I navigate from stack to stack (it is infamous bug of v1), I have to navigate to stack itself, not its particular screen. ```javascript // params are passed OK, but RN renders screen twice. So it is not an option this.props.navigation.navigate('LoginScreen', { ...this.state.data }) } // RN renders screen once as expected, but params are not being passed this.props.navigation.navigate('SignInFlow', { ...this.state.data }) } ``` ### Expected Behavior I get params in initial screen as if I was navigating to it, not to it's parent `Stacknavigator` ### How to reproduce Same way ### Your Environment | software | version | ---------------- | ------- | react-navigation | 1.5.8 | react-native |0.54.2 | node |8.11.3 | npm or yarn |5.6.0
main
params are not passed from stack navigator to stack navigator if they are children of switchnavigator the issue tracker is reserved for bug reports only this is the bug only previous same bugs that seem are not fixed yet current behavior params are not passed from one screen to another screens are initial screens of different stack navigators config switchnavigator loadingscreen component app stacknavigator profilescreen dashboardscreen signinflow stacknavigator loginscreen from this screen and vice versa likescreen signupflow stacknavigator signupscreen to this screen and vice versa tcscreen flow from loginscreen i try to navigate to signupscreen and pass params as you can see above these screens belong to different stacks those stacks are children of switchnavigator issue since react navigation renders screen twice if i navigate from stack to stack it is infamous bug of i have to navigate to stack itself not its particular screen javascript params are passed ok but rn renders screen twice so it is not an option this props navigation navigate loginscreen this state data rn renders screen once as expected but params are not being passed this props navigation navigate signinflow this state data expected behavior i get params in initial screen as if i was navigating to it not to it s parent stacknavigator how to reproduce same way your environment software version react navigation react native node npm or yarn
1
1,508
6,523,561,302
IssuesEvent
2017-08-29 09:09:50
Kristinita/Erics-Green-Room
https://api.github.com/repos/Kristinita/Erics-Green-Room
opened
[Feature request] Улучшения дизайна комнаты
need-maintainer
### 1. Моё программное обеспечение На котором сделаны скриншоты. + Windows 10 Enterprise LTSB 64-bit EN, + Firefox 55.0.3 (64-bit). ### 2. Кнопка «Выйти» ![Выйти](http://i.imgur.com/lpPzZTY.png) Полагаю, не очень удачно расположена, перекрывая скроллбар, верхнее меню и список пользователей. ### 3. Полоса под полем ответа ![Голубая полоса](http://i.imgur.com/bfYZdes.png) Не понимаю предназначение голубой полосы под полем ответа. Если она действительно не несёт никаких функций, можно её убрать, как занимающую лишнее место. ### 4. Настройка цвета в комнатах #### 1. Желательно Предоставить пользователям возможность выбора предпочтительного для них цвета бэкграунда. #### 2. Альтернатива Добавить в бэкграунд зелёный оттенок, поскольку комната называется «Зелёной комнатой Эрика». Сейчас фон при заходе через браузер бело-голубой. Спасибо.
True
[Feature request] Улучшения дизайна комнаты - ### 1. Моё программное обеспечение На котором сделаны скриншоты. + Windows 10 Enterprise LTSB 64-bit EN, + Firefox 55.0.3 (64-bit). ### 2. Кнопка «Выйти» ![Выйти](http://i.imgur.com/lpPzZTY.png) Полагаю, не очень удачно расположена, перекрывая скроллбар, верхнее меню и список пользователей. ### 3. Полоса под полем ответа ![Голубая полоса](http://i.imgur.com/bfYZdes.png) Не понимаю предназначение голубой полосы под полем ответа. Если она действительно не несёт никаких функций, можно её убрать, как занимающую лишнее место. ### 4. Настройка цвета в комнатах #### 1. Желательно Предоставить пользователям возможность выбора предпочтительного для них цвета бэкграунда. #### 2. Альтернатива Добавить в бэкграунд зелёный оттенок, поскольку комната называется «Зелёной комнатой Эрика». Сейчас фон при заходе через браузер бело-голубой. Спасибо.
main
улучшения дизайна комнаты моё программное обеспечение на котором сделаны скриншоты windows enterprise ltsb bit en firefox bit кнопка «выйти» полагаю не очень удачно расположена перекрывая скроллбар верхнее меню и список пользователей полоса под полем ответа не понимаю предназначение голубой полосы под полем ответа если она действительно не несёт никаких функций можно её убрать как занимающую лишнее место настройка цвета в комнатах желательно предоставить пользователям возможность выбора предпочтительного для них цвета бэкграунда альтернатива добавить в бэкграунд зелёный оттенок поскольку комната называется «зелёной комнатой эрика» сейчас фон при заходе через браузер бело голубой спасибо
1
5,690
29,952,334,921
IssuesEvent
2023-06-23 03:04:18
spicetify/spicetify-themes
https://api.github.com/repos/spicetify/spicetify-themes
closed
[Dreary] Playlists not able to scroll
☠️ unmaintained
**Describe the bug** There is no scroll bar on the playlists section nor is there an outline on them even though I enabled `spicetify config sidebar_config 1` **To Reproduce** Steps to reproduce the behavior: Not sure but I can send my installed things if needed **Expected behavior** Be able to scroll through my playlists in the sidebar **Screenshots** ![image](https://user-images.githubusercontent.com/41766284/232498358-e6631022-0e1c-4fda-9d1d-cf504f6c2581.png) **Logs** ```console Failed to load resource: the server responded with a status of 404 () extension.js:1 Initializing Spicetify Marketplace v0.8.5 extension.js:1 Loaded Marketplace extension hidePodcasts.js?time=1681738153179:1 HidePodcasts: Loaded xpui.js:1 cosmos-connector: fetchAdForSlot endpoint=%o sp://ads/v1/ads/hpto hidePodcasts.js?time=1681738153179:5 / <section class=​"main-shelf-shelf Shelf" aria-label=​"Jump back in">​…​</section>​ flex hidePodcasts.js?time=1681738153179:5 Tagging carousel: Your shows vendor~xpui.js:29 NotFoundError: GET request to sp://core-playlist/v1/playlist/spotify%3Aplaylist%3A37i9dQZF1EYkqdzj48dyYq/?responseFormat=protobufJson&start=0&length=0 failed with status code 404 at l.onSuccess (xpui.js:1:2001835) at onSuccess (xpui.js:1:2001649) onError @ vendor~xpui.js:29 raw.githubusercontent.com/3raxton/spicetify-custom-apps-and-extensions/main/manifest.json:1 Failed to load resource: the server responded with a status of 404 () raw.githubusercontent.com/TheRandomLabs/Scoop-Spotify/master/manifest.json:1 Failed to load resource: the server responded with a status of 404 () DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/vendor~xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/9692.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/8934.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-hpto.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui-desktop-modals.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-v2.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/7542.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED extension.js:1 GET https://raw.githubusercontent.com/catppuccin/spicetify/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/itsmeow/Spicetify-Canvas/master/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/getomni/spicetify/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/schnensch0/ziro/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404 ``` **Specifics (please complete the following information):** - OS: [Windows] - Spotify version [Spotify for Windows 1.2.9.743.g85d9593d] - Spicetify version [Spicetify v2.17.2] - Dreary ![image](https://user-images.githubusercontent.com/41766284/232499645-1074879a-a7f1-42e2-9c99-21db1cc0afbf.png)
True
[Dreary] Playlists not able to scroll - **Describe the bug** There is no scroll bar on the playlists section nor is there an outline on them even though I enabled `spicetify config sidebar_config 1` **To Reproduce** Steps to reproduce the behavior: Not sure but I can send my installed things if needed **Expected behavior** Be able to scroll through my playlists in the sidebar **Screenshots** ![image](https://user-images.githubusercontent.com/41766284/232498358-e6631022-0e1c-4fda-9d1d-cf504f6c2581.png) **Logs** ```console Failed to load resource: the server responded with a status of 404 () extension.js:1 Initializing Spicetify Marketplace v0.8.5 extension.js:1 Loaded Marketplace extension hidePodcasts.js?time=1681738153179:1 HidePodcasts: Loaded xpui.js:1 cosmos-connector: fetchAdForSlot endpoint=%o sp://ads/v1/ads/hpto hidePodcasts.js?time=1681738153179:5 / <section class=​"main-shelf-shelf Shelf" aria-label=​"Jump back in">​…​</section>​ flex hidePodcasts.js?time=1681738153179:5 Tagging carousel: Your shows vendor~xpui.js:29 NotFoundError: GET request to sp://core-playlist/v1/playlist/spotify%3Aplaylist%3A37i9dQZF1EYkqdzj48dyYq/?responseFormat=protobufJson&start=0&length=0 failed with status code 404 at l.onSuccess (xpui.js:1:2001835) at onSuccess (xpui.js:1:2001649) onError @ vendor~xpui.js:29 raw.githubusercontent.com/3raxton/spicetify-custom-apps-and-extensions/main/manifest.json:1 Failed to load resource: the server responded with a status of 404 () raw.githubusercontent.com/TheRandomLabs/Scoop-Spotify/master/manifest.json:1 Failed to load resource: the server responded with a status of 404 () DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/vendor~xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/9692.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/8934.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-hpto.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/xpui-desktop-modals.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/home-v2.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED DevTools failed to load source map: Could not load content for https://xpui.app.spotify.com/7542.js.map: Connection error: net::ERR_NAME_NOT_RESOLVED extension.js:1 GET https://raw.githubusercontent.com/catppuccin/spicetify/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/itsmeow/Spicetify-Canvas/master/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/getomni/spicetify/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/schnensch0/ziro/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404 R @ extension.js:1 (anonymous) @ extension.js:1 (anonymous) @ extension.js:1 extension.js:1 GET https://raw.githubusercontent.com/Darkempire78/Fast-BlockTheSpot-And-Spicetify-Update/main/manifest.json 404 ``` **Specifics (please complete the following information):** - OS: [Windows] - Spotify version [Spotify for Windows 1.2.9.743.g85d9593d] - Spicetify version [Spicetify v2.17.2] - Dreary ![image](https://user-images.githubusercontent.com/41766284/232499645-1074879a-a7f1-42e2-9c99-21db1cc0afbf.png)
main
playlists not able to scroll describe the bug there is no scroll bar on the playlists section nor is there an outline on them even though i enabled spicetify config sidebar config to reproduce steps to reproduce the behavior not sure but i can send my installed things if needed expected behavior be able to scroll through my playlists in the sidebar screenshots logs console failed to load resource the server responded with a status of extension js initializing spicetify marketplace extension js loaded marketplace extension hidepodcasts js time hidepodcasts loaded xpui js cosmos connector fetchadforslot endpoint o sp ads ads hpto hidepodcasts js time ​…​ ​ flex hidepodcasts js time tagging carousel your shows vendor xpui js notfounderror get request to sp core playlist playlist spotify responseformat protobufjson start length failed with status code at l onsuccess xpui js at onsuccess xpui js onerror vendor xpui js raw githubusercontent com spicetify custom apps and extensions main manifest json failed to load resource the server responded with a status of raw githubusercontent com therandomlabs scoop spotify master manifest json failed to load resource the server responded with a status of devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved devtools failed to load source map could not load content for connection error net err name not resolved extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get r extension js anonymous extension js anonymous extension js extension js get specifics please complete the following information os spotify version spotify for windows spicetify version dreary
1
119,227
17,604,737,773
IssuesEvent
2021-08-17 15:40:37
Dima2021/NodeGoat
https://api.github.com/repos/Dima2021/NodeGoat
closed
WS-2020-0344 (High) detected in is-my-json-valid-2.19.0.tgz, is-my-json-valid-2.15.0.tgz - autoclosed
security vulnerability
## WS-2020-0344 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>is-my-json-valid-2.19.0.tgz</b>, <b>is-my-json-valid-2.15.0.tgz</b></p></summary> <p> <details><summary><b>is-my-json-valid-2.19.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.19.0.tgz** (Vulnerable Library) </details> <details><summary><b>is-my-json-valid-2.15.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Dima2021/NodeGoat/commit/3b25a27fed9a418e332bfca71ab656c04a404f1e">3b25a27fed9a418e332bfca71ab656c04a404f1e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function. <p>Publish Date: 2020-06-09 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p> <p>Release Date: 2020-06-09</p> <p>Fix Resolution: is-my-json-valid - 2.20.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.19.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;har-validator:2.0.6;is-my-json-valid:2.19.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"},{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0344","vulnerabilityDetails":"Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
WS-2020-0344 (High) detected in is-my-json-valid-2.19.0.tgz, is-my-json-valid-2.15.0.tgz - autoclosed - ## WS-2020-0344 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>is-my-json-valid-2.19.0.tgz</b>, <b>is-my-json-valid-2.15.0.tgz</b></p></summary> <p> <details><summary><b>is-my-json-valid-2.19.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.19.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-retire-0.3.12.tgz (Root Library) - request-2.67.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.19.0.tgz** (Vulnerable Library) </details> <details><summary><b>is-my-json-valid-2.15.0.tgz</b></p></summary> <p>A JSONSchema validator that uses code generation to be extremely fast</p> <p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.15.0.tgz</a></p> <p>Path to dependency file: NodeGoat/package.json</p> <p>Path to vulnerable library: NodeGoat/node_modules/npm/node_modules/request/node_modules/har-validator/node_modules/is-my-json-valid/package.json</p> <p> Dependency Hierarchy: - grunt-npm-install-0.3.1.tgz (Root Library) - npm-3.10.10.tgz - request-2.75.0.tgz - har-validator-2.0.6.tgz - :x: **is-my-json-valid-2.15.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Dima2021/NodeGoat/commit/3b25a27fed9a418e332bfca71ab656c04a404f1e">3b25a27fed9a418e332bfca71ab656c04a404f1e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function. <p>Publish Date: 2020-06-09 <p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb">https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb</a></p> <p>Release Date: 2020-06-09</p> <p>Fix Resolution: is-my-json-valid - 2.20.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.19.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-retire:0.3.12;request:2.67.0;har-validator:2.0.6;is-my-json-valid:2.19.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"},{"packageType":"javascript/Node.js","packageName":"is-my-json-valid","packageVersion":"2.15.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-npm-install:0.3.1;npm:3.10.10;request:2.75.0;har-validator:2.0.6;is-my-json-valid:2.15.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"is-my-json-valid - 2.20.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2020-0344","vulnerabilityDetails":"Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.","vulnerabilityUrl":"https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_main
ws high detected in is my json valid tgz is my json valid tgz autoclosed ws high severity vulnerability vulnerable libraries is my json valid tgz is my json valid tgz is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules is my json valid package json dependency hierarchy grunt retire tgz root library request tgz har validator tgz x is my json valid tgz vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file nodegoat package json path to vulnerable library nodegoat node modules npm node modules request node modules har validator node modules is my json valid package json dependency hierarchy grunt npm install tgz root library npm tgz request tgz har validator tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch master vulnerability details arbitrary code execution vulnerability was found in is my json valid before via the fromatname function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution is my json valid isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt retire request har validator is my json valid isminimumfixversionavailable true minimumfixversion is my json valid packagetype javascript node js packagename is my json valid packageversion packagefilepaths istransitivedependency true dependencytree grunt npm install npm request har validator is my json valid isminimumfixversionavailable true minimumfixversion is my json valid basebranches vulnerabilityidentifier ws vulnerabilitydetails arbitrary code execution vulnerability was found in is my json valid before via the fromatname function vulnerabilityurl
0
234,643
18,011,518,911
IssuesEvent
2021-09-16 09:08:58
gardener/documentation
https://api.github.com/repos/gardener/documentation
closed
Local preview doesn't work correctly
kind/bug component/documentation priority/2
**What happened**: With the shift to `docforge` some folders were removed from documentation because they are fetched by `docforge`. Consequently when doing `make serve` for local preview they are missing form the website because the step with docforge build has not been executed. The only way round that is to use docforge, before doing local preview, similar to what we do in the automated central build. That unfortunately, makes it difficult to support live changes, but is still a way to preview before submitting. **What you expected to happen**: Supply the documentation manifest for which we want to see preview and have it built and shown for us. /priority critical
1.0
Local preview doesn't work correctly - **What happened**: With the shift to `docforge` some folders were removed from documentation because they are fetched by `docforge`. Consequently when doing `make serve` for local preview they are missing form the website because the step with docforge build has not been executed. The only way round that is to use docforge, before doing local preview, similar to what we do in the automated central build. That unfortunately, makes it difficult to support live changes, but is still a way to preview before submitting. **What you expected to happen**: Supply the documentation manifest for which we want to see preview and have it built and shown for us. /priority critical
non_main
local preview doesn t work correctly what happened with the shift to docforge some folders were removed from documentation because they are fetched by docforge consequently when doing make serve for local preview they are missing form the website because the step with docforge build has not been executed the only way round that is to use docforge before doing local preview similar to what we do in the automated central build that unfortunately makes it difficult to support live changes but is still a way to preview before submitting what you expected to happen supply the documentation manifest for which we want to see preview and have it built and shown for us priority critical
0
5,318
26,839,240,214
IssuesEvent
2023-02-02 22:23:05
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
add deploy --outputs-file option, like AWS CDK
type/feature area/deploy stage/pm-review maintainer/need-followup
### Describe your idea/feature/enhancement I wish SAM CLI would have an `--outputs-file` optional CLI argument for `sam deploy`, like the one `cdk deploy` has, see [the CDK docs](https://docs.aws.amazon.com/cdk/latest/guide/cli.html#w109aac23b7c33c13). Right now, it only prints the outputs in an ASCII table, mixed in with all the other stuff that goes to stdout, which of course is not machine readable, forcing a user to have to write a separate program to query CloudFormation to get the outputs.
True
add deploy --outputs-file option, like AWS CDK - ### Describe your idea/feature/enhancement I wish SAM CLI would have an `--outputs-file` optional CLI argument for `sam deploy`, like the one `cdk deploy` has, see [the CDK docs](https://docs.aws.amazon.com/cdk/latest/guide/cli.html#w109aac23b7c33c13). Right now, it only prints the outputs in an ASCII table, mixed in with all the other stuff that goes to stdout, which of course is not machine readable, forcing a user to have to write a separate program to query CloudFormation to get the outputs.
main
add deploy outputs file option like aws cdk describe your idea feature enhancement i wish sam cli would have an outputs file optional cli argument for sam deploy like the one cdk deploy has see right now it only prints the outputs in an ascii table mixed in with all the other stuff that goes to stdout which of course is not machine readable forcing a user to have to write a separate program to query cloudformation to get the outputs
1
809,644
30,203,303,886
IssuesEvent
2023-07-05 07:43:59
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Filter box in table has horizontal scroll
Type:Bug Priority:P3 Querying/Parameters & Variables .Frontend Visualization/Tables .Team/42 :milky_way:
### Describe the bug When filtering a field that has long strings, the filter box creates an unnecessary horizontal scroll. ### To Reproduce 1. Sample database -> People 2. Password (long strings field): filter by this column 3. Is: "8bee" or any known value (dont copy and paste) 4. Horizontal scroll <img width="544" alt="Screenshot 2023-06-07 at 5 43 39 PM" src="https://github.com/metabase/metabase/assets/132273646/6ef8a49a-b8b6-45f8-bdfa-14b700d5a90d"> ### Expected behavior No horizontal scroll ### Logs _No response_ ### Information about your Metabase installation ```JSON v1.46.4 ``` ### Severity Minor ### Additional context _No response_
1.0
Filter box in table has horizontal scroll - ### Describe the bug When filtering a field that has long strings, the filter box creates an unnecessary horizontal scroll. ### To Reproduce 1. Sample database -> People 2. Password (long strings field): filter by this column 3. Is: "8bee" or any known value (dont copy and paste) 4. Horizontal scroll <img width="544" alt="Screenshot 2023-06-07 at 5 43 39 PM" src="https://github.com/metabase/metabase/assets/132273646/6ef8a49a-b8b6-45f8-bdfa-14b700d5a90d"> ### Expected behavior No horizontal scroll ### Logs _No response_ ### Information about your Metabase installation ```JSON v1.46.4 ``` ### Severity Minor ### Additional context _No response_
non_main
filter box in table has horizontal scroll describe the bug when filtering a field that has long strings the filter box creates an unnecessary horizontal scroll to reproduce sample database people password long strings field filter by this column is or any known value dont copy and paste horizontal scroll img width alt screenshot at pm src expected behavior no horizontal scroll logs no response information about your metabase installation json severity minor additional context no response
0
99,815
4,065,670,387
IssuesEvent
2016-05-26 12:17:27
Taeir/ContextProject-MIGI2
https://api.github.com/repos/Taeir/ContextProject-MIGI2
closed
Refactoring - Refactoring Packages
Priority A
The packages needs to be clear and correct to improve the structure of the code. This is a subtask of #207 - Refactoring
1.0
Refactoring - Refactoring Packages - The packages needs to be clear and correct to improve the structure of the code. This is a subtask of #207 - Refactoring
non_main
refactoring refactoring packages the packages needs to be clear and correct to improve the structure of the code this is a subtask of refactoring
0
23
2,524,070,209
IssuesEvent
2015-01-20 15:33:41
simplesamlphp/simplesamlphp
https://api.github.com/repos/simplesamlphp/simplesamlphp
closed
Remove backwards-compatibility code in SimpleSAML_SessionHandlerStore
enhancement maintainability started
The `loadSession()` method has some backwards-compatible code that must go away.
True
Remove backwards-compatibility code in SimpleSAML_SessionHandlerStore - The `loadSession()` method has some backwards-compatible code that must go away.
main
remove backwards compatibility code in simplesaml sessionhandlerstore the loadsession method has some backwards compatible code that must go away
1
1,554
6,572,249,517
IssuesEvent
2017-09-11 00:35:58
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
New monit release breaks monit module
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME monitoring/monit ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### SUMMARY Since Monit changed the output of the `summary` command in version 5.18, the monit module is not able to parse it. ##### STEPS TO REPRODUCE Running the following task: ``` - name: Start server monit: name=server state=monitored ``` with the corresponding monit service definition in place, results in ``` fatal: [xhost]: FAILED! => { "changed": false, "failed": true, "msg": "server process not presently configured with monit", "name": "server", "state": "monitored" } ``` This happens due to the change of the output of the `monit summary` command: before: ``` The Monit daemon 5.17.1 uptime: 0m System 'xhost' Running Process 'server' Running ``` after upgrade to Monit 5.18: ``` Monit uptime: 56m Service Name Status Type xhost Running System server Running Process ``` Notice the additional header and the re-ordered columns.
True
New monit release breaks monit module - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME monitoring/monit ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### SUMMARY Since Monit changed the output of the `summary` command in version 5.18, the monit module is not able to parse it. ##### STEPS TO REPRODUCE Running the following task: ``` - name: Start server monit: name=server state=monitored ``` with the corresponding monit service definition in place, results in ``` fatal: [xhost]: FAILED! => { "changed": false, "failed": true, "msg": "server process not presently configured with monit", "name": "server", "state": "monitored" } ``` This happens due to the change of the output of the `monit summary` command: before: ``` The Monit daemon 5.17.1 uptime: 0m System 'xhost' Running Process 'server' Running ``` after upgrade to Monit 5.18: ``` Monit uptime: 56m Service Name Status Type xhost Running System server Running Process ``` Notice the additional header and the re-ordered columns.
main
new monit release breaks monit module issue type bug report component name monitoring monit ansible version ansible summary since monit changed the output of the summary command in version the monit module is not able to parse it steps to reproduce running the following task name start server monit name server state monitored with the corresponding monit service definition in place results in fatal failed changed false failed true msg server process not presently configured with monit name server state monitored this happens due to the change of the output of the monit summary command before the monit daemon uptime system xhost running process server running after upgrade to monit monit uptime service name status type xhost running system server running process notice the additional header and the re ordered columns
1
1,896
6,577,539,092
IssuesEvent
2017-09-12 01:37:11
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ansible-modules-core/cloud/amazon/ec2_vpc.py tag issue - feature request
affects_2.0 aws cloud feature_idea waiting_on_maintainer
##### Issue Type: - Feature Idea ##### Plugin Name: - ec2_vpc.py ##### Ansible Version: ansible 2.0.0.2 config file = /home/naslanidis/code/ag-vpc-management/ansible.cfg configured module search path = /usr/share/ansible/modules ##### Ansible Configuration: N/A ##### Environment: Fedora 22 x64 ##### Summary: When creating VPC's using this module, if any tag is changed or added a completely new VPC is created rather than the tags being added or updated. Looking at the documentation, all tags are used along with the cidr range to uniquely identify an existing VPC. In my opinion this should just be a NAME tag and other tags should be able to be added, changed and removed for an existing VPC ##### Steps To Reproduce: 1. Create a VPC using this module with a number of tags in place 2. Add a new tag (e.g. cost code), and rerun the script 3. A new duplicate VPC is created rather than adding the cost code tag to the existing VPC ``` - name: test out vpc module 1st run though ec2_vpc: state: present region: ap-southeast-2 profile: non_prod cidr_block: 10.33.0.0/16 resource_tags: Name: AAG-VPC-33 Owner: ABC register: created_vpc ``` ``` - name: test out vpc module 2nd run though ec2_vpc: state: present region: ap-southeast-2 profile: non_prod cidr_block: 10.33.0.0/16 resource_tags: Name: AAG-VPC-33 Owner: ABC Cost_code: 3 register: created_vpc ``` ##### Expected Results: The existing VPC should be updated with the new Cost_code tag. ##### Actual Results: A new VPC is created that duplicates the first VPC.
True
ansible-modules-core/cloud/amazon/ec2_vpc.py tag issue - feature request - ##### Issue Type: - Feature Idea ##### Plugin Name: - ec2_vpc.py ##### Ansible Version: ansible 2.0.0.2 config file = /home/naslanidis/code/ag-vpc-management/ansible.cfg configured module search path = /usr/share/ansible/modules ##### Ansible Configuration: N/A ##### Environment: Fedora 22 x64 ##### Summary: When creating VPC's using this module, if any tag is changed or added a completely new VPC is created rather than the tags being added or updated. Looking at the documentation, all tags are used along with the cidr range to uniquely identify an existing VPC. In my opinion this should just be a NAME tag and other tags should be able to be added, changed and removed for an existing VPC ##### Steps To Reproduce: 1. Create a VPC using this module with a number of tags in place 2. Add a new tag (e.g. cost code), and rerun the script 3. A new duplicate VPC is created rather than adding the cost code tag to the existing VPC ``` - name: test out vpc module 1st run though ec2_vpc: state: present region: ap-southeast-2 profile: non_prod cidr_block: 10.33.0.0/16 resource_tags: Name: AAG-VPC-33 Owner: ABC register: created_vpc ``` ``` - name: test out vpc module 2nd run though ec2_vpc: state: present region: ap-southeast-2 profile: non_prod cidr_block: 10.33.0.0/16 resource_tags: Name: AAG-VPC-33 Owner: ABC Cost_code: 3 register: created_vpc ``` ##### Expected Results: The existing VPC should be updated with the new Cost_code tag. ##### Actual Results: A new VPC is created that duplicates the first VPC.
main
ansible modules core cloud amazon vpc py tag issue feature request issue type feature idea plugin name vpc py ansible version ansible config file home naslanidis code ag vpc management ansible cfg configured module search path usr share ansible modules ansible configuration n a environment fedora summary when creating vpc s using this module if any tag is changed or added a completely new vpc is created rather than the tags being added or updated looking at the documentation all tags are used along with the cidr range to uniquely identify an existing vpc in my opinion this should just be a name tag and other tags should be able to be added changed and removed for an existing vpc steps to reproduce create a vpc using this module with a number of tags in place add a new tag e g cost code and rerun the script a new duplicate vpc is created rather than adding the cost code tag to the existing vpc name test out vpc module run though vpc state present region ap southeast profile non prod cidr block resource tags name aag vpc owner abc register created vpc name test out vpc module run though vpc state present region ap southeast profile non prod cidr block resource tags name aag vpc owner abc cost code register created vpc expected results the existing vpc should be updated with the new cost code tag actual results a new vpc is created that duplicates the first vpc
1