Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
1,366
5,892,287,134
IssuesEvent
2017-05-17 19:05:49
tethysplatform/tethys
https://api.github.com/repos/tethysplatform/tethys
closed
Upgrade OWSLIB to 0.9.0
bug maintain dependencies tethys_wps
Change requirement and fix bugs that are introduced. Specifically, the Parleys Creek Management and GSSHA Index Map Editor App need to be fixed to support it.
True
Upgrade OWSLIB to 0.9.0 - Change requirement and fix bugs that are introduced. Specifically, the Parleys Creek Management and GSSHA Index Map Editor App need to be fixed to support it.
main
upgrade owslib to change requirement and fix bugs that are introduced specifically the parleys creek management and gssha index map editor app need to be fixed to support it
1
1,245
5,308,976,616
IssuesEvent
2017-02-12 04:01:56
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
unable shutdown vmware guest. only power off
affects_2.2 bug_report cloud vmware waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> vmware_guest ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` 2.2.0 ``` ##### SUMMARY <!--- Explain the problem briefly --> Unable to shutdown/restart vmware guest using the Guest Operations Manager ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> These methods are implemented by https://github.com/ViaSat/ansible-vsphere
True
unable shutdown vmware guest. only power off - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> vmware_guest ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` 2.2.0 ``` ##### SUMMARY <!--- Explain the problem briefly --> Unable to shutdown/restart vmware guest using the Guest Operations Manager ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> These methods are implemented by https://github.com/ViaSat/ansible-vsphere
main
unable shutdown vmware guest only power off issue type bug report feature idea component name vmware guest ansible version summary unable to shutdown restart vmware guest using the guest operations manager expected results these methods are implemented by
1
34,160
4,892,420,626
IssuesEvent
2016-11-18 19:40:19
ThoNohT/NohBoard
https://api.github.com/repos/ThoNohT/NohBoard
closed
Mapping combination keys to a single key
Feature Request Testing
This is a feature request. As an option, let the user map certain keyboard combination keys that, when pressed simultaneously, are mapped to a single key. For example, the **ctrl,** **alt** and **e** keys could be mapped to a **€** key instead of lighting up all three keys. Particularly useful for custom keyboards.
1.0
Mapping combination keys to a single key - This is a feature request. As an option, let the user map certain keyboard combination keys that, when pressed simultaneously, are mapped to a single key. For example, the **ctrl,** **alt** and **e** keys could be mapped to a **€** key instead of lighting up all three keys. Particularly useful for custom keyboards.
non_main
mapping combination keys to a single key this is a feature request as an option let the user map certain keyboard combination keys that when pressed simultaneously are mapped to a single key for example the ctrl alt and e keys could be mapped to a € key instead of lighting up all three keys particularly useful for custom keyboards
0
2,022
6,757,633,087
IssuesEvent
2017-10-24 11:34:23
Kristinita/Erics-Green-Room
https://api.github.com/repos/Kristinita/Erics-Green-Room
closed
[Feature request] Автоматический вход на сайт
need-maintainer web
### 1. Запрос Было бы неплохо, если б игроки могли автоматически заходить на Альфа-хаб под своими логином и паролем. ### 2. Аргументация Экономия времени. Не нужно будет перед началом каждого сеанса нажимать кнопку <kbd>Войти</kbd>. ### 3. Шаги для воспроизведения Игрок заходит на **<http://alfavika.ru/wchat.html>**. ### 4. Желаемое поведение Игрок может сразу может приступать к игре под своим логином. ### 5. Актуальное поведение Игроку нужно вводить логин с паролем (если в браузере есть возможность автосохранения, можно и не вводить) и нажать на <kbd>Войти</kbd> (а это обязательно вне зависимости от браузера). ### 6. Программное окружение Не похоже, что актуальное поведение — баг. Проверял на чистых и последних на момент написания данного issue версиях браузеров: + Firefox 56.0 (64-bit), + Chrome 58.0.3029.81 (64-bit). Спасибо.
True
[Feature request] Автоматический вход на сайт - ### 1. Запрос Было бы неплохо, если б игроки могли автоматически заходить на Альфа-хаб под своими логином и паролем. ### 2. Аргументация Экономия времени. Не нужно будет перед началом каждого сеанса нажимать кнопку <kbd>Войти</kbd>. ### 3. Шаги для воспроизведения Игрок заходит на **<http://alfavika.ru/wchat.html>**. ### 4. Желаемое поведение Игрок может сразу может приступать к игре под своим логином. ### 5. Актуальное поведение Игроку нужно вводить логин с паролем (если в браузере есть возможность автосохранения, можно и не вводить) и нажать на <kbd>Войти</kbd> (а это обязательно вне зависимости от браузера). ### 6. Программное окружение Не похоже, что актуальное поведение — баг. Проверял на чистых и последних на момент написания данного issue версиях браузеров: + Firefox 56.0 (64-bit), + Chrome 58.0.3029.81 (64-bit). Спасибо.
main
автоматический вход на сайт запрос было бы неплохо если б игроки могли автоматически заходить на альфа хаб под своими логином и паролем аргументация экономия времени не нужно будет перед началом каждого сеанса нажимать кнопку войти шаги для воспроизведения игрок заходит на желаемое поведение игрок может сразу может приступать к игре под своим логином актуальное поведение игроку нужно вводить логин с паролем если в браузере есть возможность автосохранения можно и не вводить и нажать на войти а это обязательно вне зависимости от браузера программное окружение не похоже что актуальное поведение — баг проверял на чистых и последних на момент написания данного issue версиях браузеров firefox bit chrome bit спасибо
1
1,681
6,574,141,818
IssuesEvent
2017-09-11 11:40:32
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Failure when adding a program using supervisorctl which disables supervisorctl on managed host
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME supervisorctl ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.2.0 config file = /home/mike/installer configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> [defaults] host_key_checking = False private_key_file = /home/mike/.ssh/id_rsa retry_files_enabled = False [ssh_connection] ssh_args = -o ControlPersist=60s -q scp_if_ssh = True ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> OS running Ansible: Linux OS being managed: Linux ##### SUMMARY <!--- Explain the problem briefly --> - I have a program _foobar_ that exists in Git. - supervisord is currently running _foobar_ on my managed machine. - I have written a playbook that checks out the code for _foobar_ from Git, builds it on the Ansible machine, rsync's it to my managed machine (along side other supervisord managed programs). This all works well. - Here's the part where I suffer **occasional** and **intermittent** failures. My playbook attempts to reload _foobar_ under supervisord using the supervisorctl module. - 90% - 95% of the time, the program is reloaded and restarted correctly, and Ansible does not report any failures. - However, on occasion, Ansible will fail: ``` fatal: [example.com]: FAILED! => {"changed": false, "cmd": "usr/bin/supervisorctl -c /home/mike/services/supervisord.conf reread", "failed": true, "msg": "", "rc": 2, "stderr": "", "stdout": "error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567 n", ""stdout_lines": ["error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567"]} ``` - When I log onto the managed host, using the `ps` command, I can see that supervisord is still running and that all of the managed Java programs are still running, but I can no longer run `supervisorctl -c supervisord.conf -i` on the managed machine (as it returns _refused connection_). - Additionally, it appears that the programs being managed under supervisord are restarting infinitely as their PID values continue increasing. - Once I `kill` the supervisord process, the managed programs cease restarting and terminate completely. - This essentially means every once and a while, when I have a failure like this, I have to log on to each host, kill supervisord, and restart it again. After I've done that, the deployments and subsequent Ansible supervisorctl commands continue working again for an arbitrary period of time. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> - Start a supervisord managed program. - Attempt to start and stop the service from the Ansible host. Here are the commands I'm using from the Ansible side: ``` # This is where the failure occurs. - name: add programs under the group name foobar to supervisord if necessary supervisorctl: name="foobar:" state=present config="/home/mike/services/supervisord.conf" - name: restart programs under the group name foobar supervisorctl: name="foobar:" state=restarted config="/home/mike/services/supervisord.conf" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The supervisord managed service should either: - If it is currently running under supervisord, reload the service, and restart it. - If it is not running under supervisord, add the service to supervisord, and start it. ##### ACTUAL RESULTS As I mentioned previously, it appears that supervisorctl can no longer communicate with the supervisord instance running on the managed host and it is either caused by or related to the exception that Ansible reports : ``` fatal: [example.com]: FAILED! => {"changed": false, "cmd": "usr/bin/supervisorctl -c /home/mike/services/supervisord.conf reread", "failed": true, "msg": "", "rc": 2, "stderr": "", "stdout": "error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567 n", ""stdout_lines": ["error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567"]} ```
True
Failure when adding a program using supervisorctl which disables supervisorctl on managed host - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME supervisorctl ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.2.0 config file = /home/mike/installer configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> [defaults] host_key_checking = False private_key_file = /home/mike/.ssh/id_rsa retry_files_enabled = False [ssh_connection] ssh_args = -o ControlPersist=60s -q scp_if_ssh = True ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> OS running Ansible: Linux OS being managed: Linux ##### SUMMARY <!--- Explain the problem briefly --> - I have a program _foobar_ that exists in Git. - supervisord is currently running _foobar_ on my managed machine. - I have written a playbook that checks out the code for _foobar_ from Git, builds it on the Ansible machine, rsync's it to my managed machine (along side other supervisord managed programs). This all works well. - Here's the part where I suffer **occasional** and **intermittent** failures. My playbook attempts to reload _foobar_ under supervisord using the supervisorctl module. - 90% - 95% of the time, the program is reloaded and restarted correctly, and Ansible does not report any failures. - However, on occasion, Ansible will fail: ``` fatal: [example.com]: FAILED! => {"changed": false, "cmd": "usr/bin/supervisorctl -c /home/mike/services/supervisord.conf reread", "failed": true, "msg": "", "rc": 2, "stderr": "", "stdout": "error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567 n", ""stdout_lines": ["error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567"]} ``` - When I log onto the managed host, using the `ps` command, I can see that supervisord is still running and that all of the managed Java programs are still running, but I can no longer run `supervisorctl -c supervisord.conf -i` on the managed machine (as it returns _refused connection_). - Additionally, it appears that the programs being managed under supervisord are restarting infinitely as their PID values continue increasing. - Once I `kill` the supervisord process, the managed programs cease restarting and terminate completely. - This essentially means every once and a while, when I have a failure like this, I have to log on to each host, kill supervisord, and restart it again. After I've done that, the deployments and subsequent Ansible supervisorctl commands continue working again for an arbitrary period of time. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> - Start a supervisord managed program. - Attempt to start and stop the service from the Ansible host. Here are the commands I'm using from the Ansible side: ``` # This is where the failure occurs. - name: add programs under the group name foobar to supervisord if necessary supervisorctl: name="foobar:" state=present config="/home/mike/services/supervisord.conf" - name: restart programs under the group name foobar supervisorctl: name="foobar:" state=restarted config="/home/mike/services/supervisord.conf" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> The supervisord managed service should either: - If it is currently running under supervisord, reload the service, and restart it. - If it is not running under supervisord, add the service to supervisord, and start it. ##### ACTUAL RESULTS As I mentioned previously, it appears that supervisorctl can no longer communicate with the supervisord instance running on the managed host and it is either caused by or related to the exception that Ansible reports : ``` fatal: [example.com]: FAILED! => {"changed": false, "cmd": "usr/bin/supervisorctl -c /home/mike/services/supervisord.conf reread", "failed": true, "msg": "", "rc": 2, "stderr": "", "stdout": "error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567 n", ""stdout_lines": ["error: <class 'socket.error'>, [Errno 111] Connection refused: file: /usr/lib64/python2.6/socket.py line: 567"]} ```
main
failure when adding a program using supervisorctl which disables supervisorctl on managed host issue type bug report component name supervisorctl ansible version ansible config file home mike installer configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables host key checking false private key file home mike ssh id rsa retry files enabled false ssh args o controlpersist q scp if ssh true os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific os running ansible linux os being managed linux summary i have a program foobar that exists in git supervisord is currently running foobar on my managed machine i have written a playbook that checks out the code for foobar from git builds it on the ansible machine rsync s it to my managed machine along side other supervisord managed programs this all works well here s the part where i suffer occasional and intermittent failures my playbook attempts to reload foobar under supervisord using the supervisorctl module of the time the program is reloaded and restarted correctly and ansible does not report any failures however on occasion ansible will fail fatal failed changed false cmd usr bin supervisorctl c home mike services supervisord conf reread failed true msg rc stderr stdout error connection refused file usr socket py line n stdout lines connection refused file usr socket py line when i log onto the managed host using the ps command i can see that supervisord is still running and that all of the managed java programs are still running but i can no longer run supervisorctl c supervisord conf i on the managed machine as it returns refused connection additionally it appears that the programs being managed under supervisord are restarting infinitely as their pid values continue increasing once i kill the supervisord process the managed programs cease restarting and terminate completely this essentially means every once and a while when i have a failure like this i have to log on to each host kill supervisord and restart it again after i ve done that the deployments and subsequent ansible supervisorctl commands continue working again for an arbitrary period of time steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used start a supervisord managed program attempt to start and stop the service from the ansible host here are the commands i m using from the ansible side this is where the failure occurs name add programs under the group name foobar to supervisord if necessary supervisorctl name foobar state present config home mike services supervisord conf name restart programs under the group name foobar supervisorctl name foobar state restarted config home mike services supervisord conf expected results the supervisord managed service should either if it is currently running under supervisord reload the service and restart it if it is not running under supervisord add the service to supervisord and start it actual results as i mentioned previously it appears that supervisorctl can no longer communicate with the supervisord instance running on the managed host and it is either caused by or related to the exception that ansible reports fatal failed changed false cmd usr bin supervisorctl c home mike services supervisord conf reread failed true msg rc stderr stdout error connection refused file usr socket py line n stdout lines connection refused file usr socket py line
1
889
4,553,127,777
IssuesEvent
2016-09-13 02:46:27
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
lxd_container module can not have remote as part of container name
affects_2.2 bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME lxd_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 (devel f4237b2151) last updated 2016/08/16 23:49:16 (GMT +200) lib/ansible/modules/core: (detached HEAD 45c1ae0ac1) last updated 2016/08/16 23:49:20 (GMT +200) lib/ansible/modules/extras: (detached HEAD a6b34973a8) last updated 2016/08/16 23:49:20 (GMT +200) config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION default ##### OS / ENVIRONMENT Ubuntu 16.04 on both local and remote ##### SUMMARY I want to create a container on a remote lxd server. The server has been registered locally with `lxc remote add nuc1 <ip>`. ##### STEPS TO REPRODUCE `ansible-playbook test.yml` ###### test.yml ``` - hosts: localhost connection: local tasks: - name: create container test1 lxd_container: name: "nuc1:test1" state: started source: type: image mode: pull server: https://cloud-images.ubuntu.com/daily protocol: simplestreams alias: "16.04" architecture: x86_64 ``` ##### EXPECTED RESULTS I expected the container `test1` to have been launched on remote `nuc1`, essentially executing the command `lxc launch ubuntu-daily:16.04 nuc1:test1` ##### ACTUAL RESULTS <!--- Paste verbatim command output between quotes below --> ``` No config file found; using defaults Loaded callback default of type stdout, v2.0 PLAYBOOK: test.yml ************************************************************* 1 plays in test.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* Using module file /home/magne/src/ansible/lib/ansible/modules/core/system/setup.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: magne <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055 `" && echo ansible-tmp-1471409462.83-210640866662055="` echo $HOME/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055 `" ) && sleep 0' <127.0.0.1> PUT /tmp/magne/tmpkYE_4L TO /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/setup.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/ /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/setup.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/setup.py; rm -rf "/home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [create container test1] ************************************************** task path: /home/magne/development/ansible/t/test.yml:5 Using module file /home/magne/src/ansible/lib/ansible/modules/extras/cloud/lxd/lxd_container.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: magne <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257 `" && echo ansible-tmp-1471409464.58-225577408881257="` echo $HOME/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257 `" ) && sleep 0' <127.0.0.1> PUT /tmp/magne/tmpkmiVeV TO /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/lxd_container.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/ /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/lxd_container.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/lxd_container.py; rm -rf "/home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/" > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => { "actions": [], "changed": false, "failed": true, "invocation": { "module_args": { "architecture": null, "cert_file": "/home/magne/.config/lxc/client.crt", "config": null, "description": null, "devices": null, "ephemeral": null, "force_stop": false, "key_file": "/home/magne/.config/lxc/client.key", "name": "nuc1:test1", "profiles": null, "source": { "alias": "16.04", "architecture": "x86_64", "mode": "pull", "protocol": "simplestreams", "server": "https://cloud-images.ubuntu.com/daily", "type": "image" }, "state": "started", "timeout": 30, "trust_password": null, "url": "unix:/var/lib/lxd/unix.socket", "wait_for_ipv4_addresses": false }, "module_name": "lxd_container" }, "logs": [ { "request": { "json": null, "method": "GET", "timeout": null, "url": "/1.0/containers/nuc1:test1" }, "response": { "json": { "error": "not found", "error_code": 404, "type": "error" } }, "type": "sent request" }, { "request": { "json": { "name": "nuc1:test1", "source": { "alias": "16.04", "architecture": "x86_64", "mode": "pull", "protocol": "simplestreams", "server": "https://cloud-images.ubuntu.com/daily", "type": "image" } }, "method": "POST", "timeout": null, "url": "/1.0/containers" }, "response": { "json": { "metadata": { "class": "task", "created_at": "2016-08-17T06:51:04.837933973+02:00", "err": "", "id": "8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8", "may_cancel": false, "metadata": null, "resources": { "containers": [ "/1.0/containers/nuc1:test1" ] }, "status": "Running", "status_code": 103, "updated_at": "2016-08-17T06:51:04.837933973+02:00" }, "operation": "/1.0/operations/8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8", "status": "Operation created", "status_code": 100, "type": "async" } }, "type": "sent request" }, { "request": { "json": null, "method": "GET", "timeout": null, "url": "/1.0/operations/8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8/wait" }, "response": { "json": { "metadata": { "class": "task", "created_at": "2016-08-17T06:51:04.837933973+02:00", "err": "Container name isn't a valid hostname.", "id": "8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8", "may_cancel": false, "metadata": null, "resources": { "containers": [ "/1.0/containers/nuc1:test1" ] }, "status": "Failure", "status_code": 400, "updated_at": "2016-08-17T06:51:04.837933973+02:00" }, "status": "Success", "status_code": 200, "type": "sync" } }, "type": "sent request" } ], "msg": "Container name isn't a valid hostname." } to retry, use: --limit @test.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ```
True
lxd_container module can not have remote as part of container name - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME lxd_container ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0 (devel f4237b2151) last updated 2016/08/16 23:49:16 (GMT +200) lib/ansible/modules/core: (detached HEAD 45c1ae0ac1) last updated 2016/08/16 23:49:20 (GMT +200) lib/ansible/modules/extras: (detached HEAD a6b34973a8) last updated 2016/08/16 23:49:20 (GMT +200) config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION default ##### OS / ENVIRONMENT Ubuntu 16.04 on both local and remote ##### SUMMARY I want to create a container on a remote lxd server. The server has been registered locally with `lxc remote add nuc1 <ip>`. ##### STEPS TO REPRODUCE `ansible-playbook test.yml` ###### test.yml ``` - hosts: localhost connection: local tasks: - name: create container test1 lxd_container: name: "nuc1:test1" state: started source: type: image mode: pull server: https://cloud-images.ubuntu.com/daily protocol: simplestreams alias: "16.04" architecture: x86_64 ``` ##### EXPECTED RESULTS I expected the container `test1` to have been launched on remote `nuc1`, essentially executing the command `lxc launch ubuntu-daily:16.04 nuc1:test1` ##### ACTUAL RESULTS <!--- Paste verbatim command output between quotes below --> ``` No config file found; using defaults Loaded callback default of type stdout, v2.0 PLAYBOOK: test.yml ************************************************************* 1 plays in test.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* Using module file /home/magne/src/ansible/lib/ansible/modules/core/system/setup.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: magne <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055 `" && echo ansible-tmp-1471409462.83-210640866662055="` echo $HOME/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055 `" ) && sleep 0' <127.0.0.1> PUT /tmp/magne/tmpkYE_4L TO /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/setup.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/ /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/setup.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/setup.py; rm -rf "/home/magne/.ansible/tmp/ansible-tmp-1471409462.83-210640866662055/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [create container test1] ************************************************** task path: /home/magne/development/ansible/t/test.yml:5 Using module file /home/magne/src/ansible/lib/ansible/modules/extras/cloud/lxd/lxd_container.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: magne <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257 `" && echo ansible-tmp-1471409464.58-225577408881257="` echo $HOME/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257 `" ) && sleep 0' <127.0.0.1> PUT /tmp/magne/tmpkmiVeV TO /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/lxd_container.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/ /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/lxd_container.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/lxd_container.py; rm -rf "/home/magne/.ansible/tmp/ansible-tmp-1471409464.58-225577408881257/" > /dev/null 2>&1 && sleep 0' fatal: [localhost]: FAILED! => { "actions": [], "changed": false, "failed": true, "invocation": { "module_args": { "architecture": null, "cert_file": "/home/magne/.config/lxc/client.crt", "config": null, "description": null, "devices": null, "ephemeral": null, "force_stop": false, "key_file": "/home/magne/.config/lxc/client.key", "name": "nuc1:test1", "profiles": null, "source": { "alias": "16.04", "architecture": "x86_64", "mode": "pull", "protocol": "simplestreams", "server": "https://cloud-images.ubuntu.com/daily", "type": "image" }, "state": "started", "timeout": 30, "trust_password": null, "url": "unix:/var/lib/lxd/unix.socket", "wait_for_ipv4_addresses": false }, "module_name": "lxd_container" }, "logs": [ { "request": { "json": null, "method": "GET", "timeout": null, "url": "/1.0/containers/nuc1:test1" }, "response": { "json": { "error": "not found", "error_code": 404, "type": "error" } }, "type": "sent request" }, { "request": { "json": { "name": "nuc1:test1", "source": { "alias": "16.04", "architecture": "x86_64", "mode": "pull", "protocol": "simplestreams", "server": "https://cloud-images.ubuntu.com/daily", "type": "image" } }, "method": "POST", "timeout": null, "url": "/1.0/containers" }, "response": { "json": { "metadata": { "class": "task", "created_at": "2016-08-17T06:51:04.837933973+02:00", "err": "", "id": "8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8", "may_cancel": false, "metadata": null, "resources": { "containers": [ "/1.0/containers/nuc1:test1" ] }, "status": "Running", "status_code": 103, "updated_at": "2016-08-17T06:51:04.837933973+02:00" }, "operation": "/1.0/operations/8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8", "status": "Operation created", "status_code": 100, "type": "async" } }, "type": "sent request" }, { "request": { "json": null, "method": "GET", "timeout": null, "url": "/1.0/operations/8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8/wait" }, "response": { "json": { "metadata": { "class": "task", "created_at": "2016-08-17T06:51:04.837933973+02:00", "err": "Container name isn't a valid hostname.", "id": "8f17ad34-7c9f-4f08-a90b-c2a69ed68fe8", "may_cancel": false, "metadata": null, "resources": { "containers": [ "/1.0/containers/nuc1:test1" ] }, "status": "Failure", "status_code": 400, "updated_at": "2016-08-17T06:51:04.837933973+02:00" }, "status": "Success", "status_code": 200, "type": "sync" } }, "type": "sent request" } ], "msg": "Container name isn't a valid hostname." } to retry, use: --limit @test.retry PLAY RECAP ********************************************************************* localhost : ok=1 changed=0 unreachable=0 failed=1 ```
main
lxd container module can not have remote as part of container name issue type bug report component name lxd container ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file configured module search path default w o overrides configuration default os environment ubuntu on both local and remote summary i want to create a container on a remote lxd server the server has been registered locally with lxc remote add steps to reproduce ansible playbook test yml test yml hosts localhost connection local tasks name create container lxd container name state started source type image mode pull server protocol simplestreams alias architecture expected results i expected the container to have been launched on remote essentially executing the command lxc launch ubuntu daily actual results no config file found using defaults loaded callback default of type stdout playbook test yml plays in test yml play task using module file home magne src ansible lib ansible modules core system setup py establish local connection for user magne exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp magne tmpkye to home magne ansible tmp ansible tmp setup py exec bin sh c chmod u x home magne ansible tmp ansible tmp home magne ansible tmp ansible tmp setup py sleep exec bin sh c usr bin python home magne ansible tmp ansible tmp setup py rm rf home magne ansible tmp ansible tmp dev null sleep ok task task path home magne development ansible t test yml using module file home magne src ansible lib ansible modules extras cloud lxd lxd container py establish local connection for user magne exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp magne tmpkmivev to home magne ansible tmp ansible tmp lxd container py exec bin sh c chmod u x home magne ansible tmp ansible tmp home magne ansible tmp ansible tmp lxd container py sleep exec bin sh c usr bin python home magne ansible tmp ansible tmp lxd container py rm rf home magne ansible tmp ansible tmp dev null sleep fatal failed actions changed false failed true invocation module args architecture null cert file home magne config lxc client crt config null description null devices null ephemeral null force stop false key file home magne config lxc client key name profiles null source alias architecture mode pull protocol simplestreams server type image state started timeout trust password null url unix var lib lxd unix socket wait for addresses false module name lxd container logs request json null method get timeout null url containers response json error not found error code type error type sent request request json name source alias architecture mode pull protocol simplestreams server type image method post timeout null url containers response json metadata class task created at err id may cancel false metadata null resources containers containers status running status code updated at operation operations status operation created status code type async type sent request request json null method get timeout null url operations wait response json metadata class task created at err container name isn t a valid hostname id may cancel false metadata null resources containers containers status failure status code updated at status success status code type sync type sent request msg container name isn t a valid hostname to retry use limit test retry play recap localhost ok changed unreachable failed
1
3,592
2,610,065,405
IssuesEvent
2015-02-26 18:19:14
chrsmith/jsjsj122
https://api.github.com/repos/chrsmith/jsjsj122
opened
临海治疗前列腺炎哪里好
auto-migrated Priority-Medium Type-Defect
``` 临海治疗前列腺炎哪里好【台州五洲生殖医院】24小时健康咨 询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州 市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108� ��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109 、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:20
1.0
临海治疗前列腺炎哪里好 - ``` 临海治疗前列腺炎哪里好【台州五洲生殖医院】24小时健康咨 询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:台州 市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104、108� ��118、198及椒江一金清公交车直达枫南小区,乘坐107、105、109 、112、901、 902公交车到星星广场下车,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 ``` ----- Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 8:20
non_main
临海治疗前列腺炎哪里好 临海治疗前列腺炎哪里好【台州五洲生殖医院】 询热线 微信号tzwzszyy 医院地址 台州 (枫南大转盘旁)乘车线路 、 � �� 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at
0
4,192
20,378,915,573
IssuesEvent
2022-02-21 18:46:58
HPCL/code-analysis
https://api.github.com/repos/HPCL/code-analysis
closed
CWE-1064 Invokable Control Element with Signature Containing an Excessive Number of Parameters
CLAIMED ISO/IEC 5055:2021 OperationSignature WEAKNESS CATEGORY: MAINTAINABILITY
**Usage Name** Excessive parameterization **Reference** [https://cwe.mitre.org/data/definitions/1064](https://cwe.mitre.org/data/definitions/1064) **Roles** - the *OperationSignature* **Detection Patterns** - 8.2.124 ASCQM Limit Number of Parameters
True
CWE-1064 Invokable Control Element with Signature Containing an Excessive Number of Parameters - **Usage Name** Excessive parameterization **Reference** [https://cwe.mitre.org/data/definitions/1064](https://cwe.mitre.org/data/definitions/1064) **Roles** - the *OperationSignature* **Detection Patterns** - 8.2.124 ASCQM Limit Number of Parameters
main
cwe invokable control element with signature containing an excessive number of parameters usage name excessive parameterization reference roles the operationsignature detection patterns ascqm limit number of parameters
1
11,622
9,322,391,146
IssuesEvent
2019-03-27 08:02:26
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
closed
Set up deployment of SBL Bridge
infrastructure sbl sbl-bridge team-steam
**Functional architect/designer:** @-mention **Technical architect:** @-mention **Description** The platform integration application needs to be deployed on the portal servers. **Technical considerations** Input (beyond tasks) on how the user story should be solved can be put here. **Acceptance criterea** - When Portal services are deployed the Platform integration application should be deployed - The platform integration application should also be deployed on SBL dev environment **Tasks** - [ ] Update deployment scripts so Platform Integration application is deployed on the Portal servers
1.0
Set up deployment of SBL Bridge - **Functional architect/designer:** @-mention **Technical architect:** @-mention **Description** The platform integration application needs to be deployed on the portal servers. **Technical considerations** Input (beyond tasks) on how the user story should be solved can be put here. **Acceptance criterea** - When Portal services are deployed the Platform integration application should be deployed - The platform integration application should also be deployed on SBL dev environment **Tasks** - [ ] Update deployment scripts so Platform Integration application is deployed on the Portal servers
non_main
set up deployment of sbl bridge functional architect designer mention technical architect mention description the platform integration application needs to be deployed on the portal servers technical considerations input beyond tasks on how the user story should be solved can be put here acceptance criterea when portal services are deployed the platform integration application should be deployed the platform integration application should also be deployed on sbl dev environment tasks update deployment scripts so platform integration application is deployed on the portal servers
0
5,002
25,731,267,735
IssuesEvent
2022-12-07 20:32:13
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Use display options in record summaries
type: bug work: frontend status: ready restricted: maintainers
## Steps to reproduce 1. Configure a table to include a date column within its record summary. 1. Expect the date to be formatted according to the display options set for that column. 1. Observe instead that the date is formatted in ISO.
True
Use display options in record summaries - ## Steps to reproduce 1. Configure a table to include a date column within its record summary. 1. Expect the date to be formatted according to the display options set for that column. 1. Observe instead that the date is formatted in ISO.
main
use display options in record summaries steps to reproduce configure a table to include a date column within its record summary expect the date to be formatted according to the display options set for that column observe instead that the date is formatted in iso
1
93,585
8,434,950,525
IssuesEvent
2018-10-17 11:47:29
vgstation-coders/vgstation13
https://api.github.com/repos/vgstation-coders/vgstation13
closed
computer frame constrction problems
Bug / Fix Needs Moar Testing
(WEB REPORT BY: pvak REMOTE: 172.93.109.202:7777) >Revision (Should be above if you're viewing this from ingame!) >General description of the issue if you fuck up one of the steps to build a computer like moving away while wrenching the computer frame becomes unusable, you cant wrench it again or anything, same for other steps that require you to stand still >What you expected to happen being able to construct a computer >What actually happened computer frames that cant be interacted with >Steps to reproduce if possible make a frame, start wrenching , move away, try wrench (same problem for other steps)
1.0
computer frame constrction problems - (WEB REPORT BY: pvak REMOTE: 172.93.109.202:7777) >Revision (Should be above if you're viewing this from ingame!) >General description of the issue if you fuck up one of the steps to build a computer like moving away while wrenching the computer frame becomes unusable, you cant wrench it again or anything, same for other steps that require you to stand still >What you expected to happen being able to construct a computer >What actually happened computer frames that cant be interacted with >Steps to reproduce if possible make a frame, start wrenching , move away, try wrench (same problem for other steps)
non_main
computer frame constrction problems web report by pvak remote revision should be above if you re viewing this from ingame general description of the issue if you fuck up one of the steps to build a computer like moving away while wrenching the computer frame becomes unusable you cant wrench it again or anything same for other steps that require you to stand still what you expected to happen being able to construct a computer what actually happened computer frames that cant be interacted with steps to reproduce if possible make a frame start wrenching move away try wrench same problem for other steps
0
212,959
16,490,246,837
IssuesEvent
2021-05-25 01:56:30
ObliqueNET/Server
https://api.github.com/repos/ObliqueNET/Server
closed
[Sponge] Master ticket on mobs and Heill
bug landlord needs testing oblique sponge
Around Heill, you cant attack mobs... that works for passive mobs like cows, but you cant attack some hostile mobs as well. Some notable mentions are Shulkers and Zombie Villagers. While regular old Bow equipped skeleton and zombie can be attacked, these mobs although hostile, cannot be damaged. (Instead they vanish but still hit you).
1.0
[Sponge] Master ticket on mobs and Heill - Around Heill, you cant attack mobs... that works for passive mobs like cows, but you cant attack some hostile mobs as well. Some notable mentions are Shulkers and Zombie Villagers. While regular old Bow equipped skeleton and zombie can be attacked, these mobs although hostile, cannot be damaged. (Instead they vanish but still hit you).
non_main
master ticket on mobs and heill around heill you cant attack mobs that works for passive mobs like cows but you cant attack some hostile mobs as well some notable mentions are shulkers and zombie villagers while regular old bow equipped skeleton and zombie can be attacked these mobs although hostile cannot be damaged instead they vanish but still hit you
0
5,714
30,200,974,595
IssuesEvent
2023-07-05 05:38:56
jupyter-naas/awesome-notebooks
https://api.github.com/repos/jupyter-naas/awesome-notebooks
closed
OS - Get folder stats
templates maintainer
This notebook will get the stats of a folder and its content. It is usefull for organizations to get a quick overview of the size of a folder and its content.
True
OS - Get folder stats - This notebook will get the stats of a folder and its content. It is usefull for organizations to get a quick overview of the size of a folder and its content.
main
os get folder stats this notebook will get the stats of a folder and its content it is usefull for organizations to get a quick overview of the size of a folder and its content
1
158,288
12,412,127,949
IssuesEvent
2020-05-22 09:56:25
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
(warning) Assignment of function parameter has no effect outside the function. Did you forget derefe
bug cppcheck static analysis testing triage
```yaml { "id": 1, "title": "(warning) Assignment of function parameter has no effect outside the function. Did you forget derefe", "type": "bug", "description": "[src/opencv3/3rdparty/libjasper/jp2_cod.c:869]: (warning) Assignment of function parameter has no effect outside the function. Did you forget dereferencing it?", "cwe": "None", "cve": "None", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "system": "src/opencv3/3rdparty/libjasper/jp2_cod.c", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "N/A", "architectural-location": "N/A", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-22 (09:56)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-22 (09:56)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "", "reproduction": "See artifacts below (if available)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
(warning) Assignment of function parameter has no effect outside the function. Did you forget derefe - ```yaml { "id": 1, "title": "(warning) Assignment of function parameter has no effect outside the function. Did you forget derefe", "type": "bug", "description": "[src/opencv3/3rdparty/libjasper/jp2_cod.c:869]: (warning) Assignment of function parameter has no effect outside the function. Did you forget dereferencing it?", "cwe": "None", "cve": "None", "keywords": [ "cppcheck", "static analysis", "testing", "triage", "bug" ], "system": "src/opencv3/3rdparty/libjasper/jp2_cod.c", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "N/A", "architectural-location": "N/A", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-05-22 (09:56)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-05-22 (09:56)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "", "reproduction": "See artifacts below (if available)", "reproduction-image": "gitlab.com/aliasrobotics/offensive/alurity/pipelines/active/pipeline_ros_kinetic/-/jobs/563367426/artifacts/download" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_main
warning assignment of function parameter has no effect outside the function did you forget derefe yaml id title warning assignment of function parameter has no effect outside the function did you forget derefe type bug description warning assignment of function parameter has no effect outside the function did you forget dereferencing it cwe none cve none keywords cppcheck static analysis testing triage bug system src libjasper cod c vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity n a architectural location n a application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace reproduction see artifacts below if available reproduction image gitlab com aliasrobotics offensive alurity pipelines active pipeline ros kinetic jobs artifacts download exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
5,610
28,090,071,804
IssuesEvent
2023-03-30 12:30:57
Windham-High-School/CubeServer
https://api.github.com/repos/Windham-High-School/CubeServer
closed
Problems with Arm Arch
bug docker maintainability
Output of `journalctl -xe`: ``` Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | iptables v1.8.8 (nf_tables): unknown option "--dport" Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | iptables v1.8.8 (nf_tables): unknown option "--dport" Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | iptables v1.8.8 (nf_tables): unknown option "--dport" Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | [-] Error creating iptables rules Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | [*] Setting wlan0 settings Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | ifdown: interface wlan0 not configured Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | ip: RTNETLINK answers: File exists Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | WARNING: MongoDB 5.0+ requires ARMv8.2-A or higher, and your current system does not appear to implement any of the common features fo> Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | see https://jira.mongodb.org/browse/SERVER-55178 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | see also https://en.wikichip.org/wiki/arm/armv8#ARMv8_Extensions_and_Processor_Features Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | see also https://github.com/docker-library/mongo/issues/485#issuecomment-970864306 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | [+] Configuration successful! Services will start now Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Internet Systems Consortium DHCP Server 4.4.3-P1 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Copyright 2004-2022 Internet Systems Consortium. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | All rights reserved. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | For info, please visit https://www.isc.org/software/dhcp/ Jan 18 17:04:07 cubeserver systemd-networkd[618]: br-1d35c8049ea6: Gained IPv6LL Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | rfkill: Cannot open RFKILL control device Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Config file: /etc/dhcp/dhcpd.conf Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Database file: /var/lib/dhcp/dhcpd.leases Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | PID file: /run/dhcp/dhcpd.pid Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | wlan0: interface state UNINITIALIZED->COUNTRY_UPDATE Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Wrote 0 leases to leases file. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Listening on LPF/wlan0/b8:27:eb:5c:d5:62/192.168.252.0/24 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Sending on LPF/wlan0/b8:27:eb:5c:d5:62/192.168.252.0/24 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Sending on Socket/fallback/fallback-net Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Can't create PID file /run/dhcp/dhcpd.pid: No such file or directory. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Server starting service. Jan 18 17:04:07 cubeserver systemd-networkd[618]: wlan0: Gained carrier Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | wlan0: interface state COUNTRY_UPDATE->ENABLED Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | wlan0: AP-ENABLED Jan 18 17:04:08 cubeserver systemd-networkd[618]: vethd733872: Gained IPv6LL ```
True
Problems with Arm Arch - Output of `journalctl -xe`: ``` Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | iptables v1.8.8 (nf_tables): unknown option "--dport" Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | iptables v1.8.8 (nf_tables): unknown option "--dport" Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | iptables v1.8.8 (nf_tables): unknown option "--dport" Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Try `iptables -h' or 'iptables --help' for more information. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | [-] Error creating iptables rules Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | [*] Setting wlan0 settings Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | ifdown: interface wlan0 not configured Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | ip: RTNETLINK answers: File exists Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | WARNING: MongoDB 5.0+ requires ARMv8.2-A or higher, and your current system does not appear to implement any of the common features fo> Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | see https://jira.mongodb.org/browse/SERVER-55178 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | see also https://en.wikichip.org/wiki/arm/armv8#ARMv8_Extensions_and_Processor_Features Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | see also https://github.com/docker-library/mongo/issues/485#issuecomment-970864306 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-mongodb | Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | [+] Configuration successful! Services will start now Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Internet Systems Consortium DHCP Server 4.4.3-P1 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Copyright 2004-2022 Internet Systems Consortium. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | All rights reserved. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | For info, please visit https://www.isc.org/software/dhcp/ Jan 18 17:04:07 cubeserver systemd-networkd[618]: br-1d35c8049ea6: Gained IPv6LL Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | rfkill: Cannot open RFKILL control device Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Config file: /etc/dhcp/dhcpd.conf Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Database file: /var/lib/dhcp/dhcpd.leases Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | PID file: /run/dhcp/dhcpd.pid Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | wlan0: interface state UNINITIALIZED->COUNTRY_UPDATE Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Wrote 0 leases to leases file. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Listening on LPF/wlan0/b8:27:eb:5c:d5:62/192.168.252.0/24 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Sending on LPF/wlan0/b8:27:eb:5c:d5:62/192.168.252.0/24 Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Sending on Socket/fallback/fallback-net Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Can't create PID file /run/dhcp/dhcpd.pid: No such file or directory. Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | Server starting service. Jan 18 17:04:07 cubeserver systemd-networkd[618]: wlan0: Gained carrier Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | wlan0: interface state COUNTRY_UPDATE->ENABLED Jan 18 17:04:07 cubeserver docker[2366]: CubeServer-accesspoint | wlan0: AP-ENABLED Jan 18 17:04:08 cubeserver systemd-networkd[618]: vethd733872: Gained IPv6LL ```
main
problems with arm arch output of journalctl xe jan cubeserver docker cubeserver accesspoint try iptables h or iptables help for more information jan cubeserver docker cubeserver accesspoint iptables nf tables unknown option dport jan cubeserver docker cubeserver accesspoint try iptables h or iptables help for more information jan cubeserver docker cubeserver accesspoint iptables nf tables unknown option dport jan cubeserver docker cubeserver accesspoint try iptables h or iptables help for more information jan cubeserver docker cubeserver accesspoint iptables nf tables unknown option dport jan cubeserver docker cubeserver accesspoint try iptables h or iptables help for more information jan cubeserver docker cubeserver accesspoint error creating iptables rules jan cubeserver docker cubeserver accesspoint setting settings jan cubeserver docker cubeserver accesspoint ifdown interface not configured jan cubeserver docker cubeserver accesspoint ip rtnetlink answers file exists jan cubeserver docker cubeserver mongodb jan cubeserver docker cubeserver mongodb warning mongodb requires a or higher and your current system does not appear to implement any of the common features fo jan cubeserver docker cubeserver mongodb see jan cubeserver docker cubeserver mongodb see also jan cubeserver docker cubeserver mongodb see also jan cubeserver docker cubeserver mongodb jan cubeserver docker cubeserver accesspoint configuration successful services will start now jan cubeserver docker cubeserver accesspoint internet systems consortium dhcp server jan cubeserver docker cubeserver accesspoint copyright internet systems consortium jan cubeserver docker cubeserver accesspoint all rights reserved jan cubeserver docker cubeserver accesspoint for info please visit jan cubeserver systemd networkd br gained jan cubeserver docker cubeserver accesspoint rfkill cannot open rfkill control device jan cubeserver docker cubeserver accesspoint config file etc dhcp dhcpd conf jan cubeserver docker cubeserver accesspoint database file var lib dhcp dhcpd leases jan cubeserver docker cubeserver accesspoint pid file run dhcp dhcpd pid jan cubeserver docker cubeserver accesspoint interface state uninitialized country update jan cubeserver docker cubeserver accesspoint wrote leases to leases file jan cubeserver docker cubeserver accesspoint listening on lpf eb jan cubeserver docker cubeserver accesspoint sending on lpf eb jan cubeserver docker cubeserver accesspoint sending on socket fallback fallback net jan cubeserver docker cubeserver accesspoint can t create pid file run dhcp dhcpd pid no such file or directory jan cubeserver docker cubeserver accesspoint server starting service jan cubeserver systemd networkd gained carrier jan cubeserver docker cubeserver accesspoint interface state country update enabled jan cubeserver docker cubeserver accesspoint ap enabled jan cubeserver systemd networkd gained
1
580,859
17,268,799,318
IssuesEvent
2021-07-22 16:52:46
open-apparel-registry/open-apparel-registry
https://api.github.com/repos/open-apparel-registry/open-apparel-registry
closed
Geocoding reports only pulling data started in July
+ bug high priority task 025
## Overview We are only seeing data in the geocoding reports from July 2021 forward. We'd like to see this data looking back as well, at least to the beginning of this year but ideally as far back as we can go. We are trying to wrap up our midyear reporting so if this can be rectified soon, that would be fantastic. These are the reports: https://openapparel.org/admin/reports/recent_monthly_geocoding_time_with_queue/ https://openapparel.org/admin/reports/recent_monthly_geocoding_time_without_queue/ https://openapparel.org/admin/reports/recent_weekly_geocoding_time_with_queue/ https://openapparel.org/admin/reports/recent_weekly_geocoding_time_without_queue/ ### Expected Behavior We would expect to see data for previous months/weeks in this report but only see it for one week/month. ### Actual Behavior We are only seeing one week/month of data. ### Steps to Reproduce Check out those reports! ### Demo ### Additional context cc @katieashaw @isaacokunato
1.0
Geocoding reports only pulling data started in July - ## Overview We are only seeing data in the geocoding reports from July 2021 forward. We'd like to see this data looking back as well, at least to the beginning of this year but ideally as far back as we can go. We are trying to wrap up our midyear reporting so if this can be rectified soon, that would be fantastic. These are the reports: https://openapparel.org/admin/reports/recent_monthly_geocoding_time_with_queue/ https://openapparel.org/admin/reports/recent_monthly_geocoding_time_without_queue/ https://openapparel.org/admin/reports/recent_weekly_geocoding_time_with_queue/ https://openapparel.org/admin/reports/recent_weekly_geocoding_time_without_queue/ ### Expected Behavior We would expect to see data for previous months/weeks in this report but only see it for one week/month. ### Actual Behavior We are only seeing one week/month of data. ### Steps to Reproduce Check out those reports! ### Demo ### Additional context cc @katieashaw @isaacokunato
non_main
geocoding reports only pulling data started in july overview we are only seeing data in the geocoding reports from july forward we d like to see this data looking back as well at least to the beginning of this year but ideally as far back as we can go we are trying to wrap up our midyear reporting so if this can be rectified soon that would be fantastic these are the reports expected behavior we would expect to see data for previous months weeks in this report but only see it for one week month actual behavior we are only seeing one week month of data steps to reproduce check out those reports demo additional context cc katieashaw isaacokunato
0
3,982
18,344,784,069
IssuesEvent
2021-10-08 03:51:00
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
AWS SAM-CLI The security token included in the request is invalid.
stage/bug-repro maintainer/need-response
### Description: I am creating my first Lambda Function using AWS Toolkit and SAM-CLI and I am facing two main issues: 1. When I use a workaround and this function is getting information from AWS System Manager Parameter Store: An error occurred (UnrecognizedClientException) when calling the GetParameter operation: The security token included in the request is invalid. I use Windows 10, Visual Studio Code, SAM CLI version 0.44 (Tried to update using the install but no success) and Python 3.8. ### Steps to reproduce: Below is the YAML. ``` AWSTemplateFormatVersion: '2010-09-09' Transform: AWS::Serverless-2016-10-31 Description: > research-gam-download-revenuereport Sample SAM Template for research-gam-download-revenuereport # More info about Globals: https://github.com/awslabs/serverless-application-model/blob/master/docs/globals.rst Globals: Function: Timeout: 3 Resources: HelloWorldFunction: Type: AWS::Serverless::Function # More info about Function Resource: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#awsserverlessfunction Properties: Role: arn:aws:iam::123568192875:role/10-research-insights-sb-lambdaiam CodeUri: hello_world/ Handler: lambda_function.lambda_handler Runtime: python3.8 Events: HelloWorld: Type: Api # More info about API Event Source: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#api Properties: Path: /hello Method: get Layers: - !Ref Test Test: Type: AWS::Serverless::LayerVersion Properties: CompatibleRuntimes: - python3.6 - python3.7 - python3.8 ContentUri: ../../../../../ten/ten_package.zip Description: research-insights-sb-tenlayer LayerName: tenlayer LicenseInfo: MIT ``` ### Observed result: Error 1: An error occurred (UnrecognizedClientException) when calling the GetParameter operation: The security token included in the request is invalid. ### Expected result: I was expecting it would connect to other AWS resources as it is already with the credentials ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows 10 2. `sam --version`: 0.44 3. AWS region: ap-southeast-2
True
AWS SAM-CLI The security token included in the request is invalid. - ### Description: I am creating my first Lambda Function using AWS Toolkit and SAM-CLI and I am facing two main issues: 1. When I use a workaround and this function is getting information from AWS System Manager Parameter Store: An error occurred (UnrecognizedClientException) when calling the GetParameter operation: The security token included in the request is invalid. I use Windows 10, Visual Studio Code, SAM CLI version 0.44 (Tried to update using the install but no success) and Python 3.8. ### Steps to reproduce: Below is the YAML. ``` AWSTemplateFormatVersion: '2010-09-09' Transform: AWS::Serverless-2016-10-31 Description: > research-gam-download-revenuereport Sample SAM Template for research-gam-download-revenuereport # More info about Globals: https://github.com/awslabs/serverless-application-model/blob/master/docs/globals.rst Globals: Function: Timeout: 3 Resources: HelloWorldFunction: Type: AWS::Serverless::Function # More info about Function Resource: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#awsserverlessfunction Properties: Role: arn:aws:iam::123568192875:role/10-research-insights-sb-lambdaiam CodeUri: hello_world/ Handler: lambda_function.lambda_handler Runtime: python3.8 Events: HelloWorld: Type: Api # More info about API Event Source: https://github.com/awslabs/serverless-application-model/blob/master/versions/2016-10-31.md#api Properties: Path: /hello Method: get Layers: - !Ref Test Test: Type: AWS::Serverless::LayerVersion Properties: CompatibleRuntimes: - python3.6 - python3.7 - python3.8 ContentUri: ../../../../../ten/ten_package.zip Description: research-insights-sb-tenlayer LayerName: tenlayer LicenseInfo: MIT ``` ### Observed result: Error 1: An error occurred (UnrecognizedClientException) when calling the GetParameter operation: The security token included in the request is invalid. ### Expected result: I was expecting it would connect to other AWS resources as it is already with the credentials ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows 10 2. `sam --version`: 0.44 3. AWS region: ap-southeast-2
main
aws sam cli the security token included in the request is invalid description i am creating my first lambda function using aws toolkit and sam cli and i am facing two main issues when i use a workaround and this function is getting information from aws system manager parameter store an error occurred unrecognizedclientexception when calling the getparameter operation the security token included in the request is invalid i use windows visual studio code sam cli version tried to update using the install but no success and python steps to reproduce below is the yaml awstemplateformatversion transform aws serverless description research gam download revenuereport sample sam template for research gam download revenuereport more info about globals globals function timeout resources helloworldfunction type aws serverless function more info about function resource properties role arn aws iam role research insights sb lambdaiam codeuri hello world handler lambda function lambda handler runtime events helloworld type api more info about api event source properties path hello method get layers ref test test type aws serverless layerversion properties compatibleruntimes contenturi ten ten package zip description research insights sb tenlayer layername tenlayer licenseinfo mit observed result error an error occurred unrecognizedclientexception when calling the getparameter operation the security token included in the request is invalid expected result i was expecting it would connect to other aws resources as it is already with the credentials additional environment details ex windows mac amazon linux etc os windows sam version aws region ap southeast
1
54,266
3,062,156,051
IssuesEvent
2015-08-16 09:25:41
valnet/valuenetwork
https://api.github.com/repos/valnet/valuenetwork
opened
Map agent locations
enhancement priority
Now that we got maps working again, let's put agents on the map. We've created a bit of a mess here because we got Locations that go on the map, but they are connected only to resources. Agents have an address field, that is not mapped. And they also have a primary_location field that is a Location, and would go on the map, but it's not on AgentCreateForm. So we could either put the address field on the map (which would require geocoding when entered), or put the primary_location on the AgentCreateForm. Problem with primary_location is that it's an extra set of steps (to create a location, and then add it to the agent). Maybe better? when they enter the address, we could geocode it, create a Location, and make that the primary_location. Might want to rethink Locations altogether...
1.0
Map agent locations - Now that we got maps working again, let's put agents on the map. We've created a bit of a mess here because we got Locations that go on the map, but they are connected only to resources. Agents have an address field, that is not mapped. And they also have a primary_location field that is a Location, and would go on the map, but it's not on AgentCreateForm. So we could either put the address field on the map (which would require geocoding when entered), or put the primary_location on the AgentCreateForm. Problem with primary_location is that it's an extra set of steps (to create a location, and then add it to the agent). Maybe better? when they enter the address, we could geocode it, create a Location, and make that the primary_location. Might want to rethink Locations altogether...
non_main
map agent locations now that we got maps working again let s put agents on the map we ve created a bit of a mess here because we got locations that go on the map but they are connected only to resources agents have an address field that is not mapped and they also have a primary location field that is a location and would go on the map but it s not on agentcreateform so we could either put the address field on the map which would require geocoding when entered or put the primary location on the agentcreateform problem with primary location is that it s an extra set of steps to create a location and then add it to the agent maybe better when they enter the address we could geocode it create a location and make that the primary location might want to rethink locations altogether
0
2,422
8,606,880,284
IssuesEvent
2018-11-17 16:41:12
arcticicestudio/nord-docs
https://api.github.com/repos/arcticicestudio/nord-docs
closed
lint-staged
context-workflow scope-maintainability type-task
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/48658851-01e38400-ea49-11e8-911e-d859eefe6dd5.png" width="25%" /></p> > Must be resolved **after** #14 #15 #16 > Must be resolved **before** #18 Integrate [lint-staged][gh-lint-staged] to run linters against staged Git files to prevent to add code that violates any style guide into the code base. <p align="center"><img src="https://raw.githubusercontent.com/okonet/lint-staged/master/screenshots/lint-staged-prettier.gif" width="80%" /></p> ### Configuration The configuration file `lint-staged.config.js` will be placed in the project root and includes the command that should be run for matching file extensions (globs). It will include at least the three following entries with the same order as listed here: 1. `prettier --list-different` - Run Prettier (#16) against `*.{js,json,jsx,mdx,yml}` to ensure all files are formatted correctly. The `--list-different` prints the found files that are not conform to the Prettier configuration. 2. `eslint` - Run ESLint (#15) against `*.{js,jsx}` to ensure all JavaScript files are compliant to the style guide after being formatted with Prettier. 3. `remark --no-stdout` - Run remark-lint (#14) against `*.md` to ensure all Markdown files are compliant to the style guide. The `--no-stdout` flag suppresses the output of the parsed file content. ## Tasks - [x] Install [lint-staged][npm-lint-staged] package. - [x] Implement `lint-staged.config.js` configuration file. [gh-lint-staged]: https://github.com/okonet/lint-staged [npm-lint-staged]: https://www.npmjs.com/package/lint-staged
True
lint-staged - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48658851-01e38400-ea49-11e8-911e-d859eefe6dd5.png" width="25%" /></p> > Must be resolved **after** #14 #15 #16 > Must be resolved **before** #18 Integrate [lint-staged][gh-lint-staged] to run linters against staged Git files to prevent to add code that violates any style guide into the code base. <p align="center"><img src="https://raw.githubusercontent.com/okonet/lint-staged/master/screenshots/lint-staged-prettier.gif" width="80%" /></p> ### Configuration The configuration file `lint-staged.config.js` will be placed in the project root and includes the command that should be run for matching file extensions (globs). It will include at least the three following entries with the same order as listed here: 1. `prettier --list-different` - Run Prettier (#16) against `*.{js,json,jsx,mdx,yml}` to ensure all files are formatted correctly. The `--list-different` prints the found files that are not conform to the Prettier configuration. 2. `eslint` - Run ESLint (#15) against `*.{js,jsx}` to ensure all JavaScript files are compliant to the style guide after being formatted with Prettier. 3. `remark --no-stdout` - Run remark-lint (#14) against `*.md` to ensure all Markdown files are compliant to the style guide. The `--no-stdout` flag suppresses the output of the parsed file content. ## Tasks - [x] Install [lint-staged][npm-lint-staged] package. - [x] Implement `lint-staged.config.js` configuration file. [gh-lint-staged]: https://github.com/okonet/lint-staged [npm-lint-staged]: https://www.npmjs.com/package/lint-staged
main
lint staged must be resolved after must be resolved before integrate to run linters against staged git files to prevent to add code that violates any style guide into the code base configuration the configuration file lint staged config js will be placed in the project root and includes the command that should be run for matching file extensions globs it will include at least the three following entries with the same order as listed here prettier list different run prettier against js json jsx mdx yml to ensure all files are formatted correctly the list different prints the found files that are not conform to the prettier configuration eslint run eslint against js jsx to ensure all javascript files are compliant to the style guide after being formatted with prettier remark no stdout run remark lint against md to ensure all markdown files are compliant to the style guide the no stdout flag suppresses the output of the parsed file content tasks install package implement lint staged config js configuration file
1
4,723
24,375,280,075
IssuesEvent
2022-10-03 23:57:16
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
init an internal lambda function?
area/examples area/init stage/pm-review maintainer/need-followup
`sam init --runtime go1.x` appears to create an API gateway type of function. What happens if I want to create a function that gets triggered by SNS or some direct lambda call (with JSON payload) instead??
True
init an internal lambda function? - `sam init --runtime go1.x` appears to create an API gateway type of function. What happens if I want to create a function that gets triggered by SNS or some direct lambda call (with JSON payload) instead??
main
init an internal lambda function sam init runtime x appears to create an api gateway type of function what happens if i want to create a function that gets triggered by sns or some direct lambda call with json payload instead
1
367,888
10,862,653,078
IssuesEvent
2019-11-14 13:43:57
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
eu.roadmunk.com - Only Chrome is supported
browser-firefox engine-gecko priority-normal severity-critical type-unsupported
<!-- @browser: Firefox 66.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:66.0) Gecko/20100101 Firefox/66.0 --> <!-- @reported_with: web --> **URL**: https://eu.roadmunk.com/ **Browser / Version**: Firefox 66.0 **Operating System**: Mac OS X 10.12 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Roadmunk is damaging the openness of the web - blocks all browsers except Google Chrome **Steps to Reproduce**: Opened the url in Firefox (Vivaldi,Safari,etc) none are accepted other than Google Chrome. Particularly gauling in this case is that Vivaldi is a Chromium based browser and likely supports whatever proprietary features Roadmunk felt were necessary - but they block it anyway. This sort of behaviour should be called out and the companies shamed. In the light of Edge dropping their own rendering engine in favour of becoming yet another Chromium clone the diversity of the web is at risk and we could end up with another ie6 world where Google calls all the shots. [![Screenshot Description](https://webcompat.com/uploads/2019/4/dfc78383-77cd-4221-bcc7-af39bdde689f-thumb.jpg)](https://webcompat.com/uploads/2019/4/dfc78383-77cd-4221-bcc7-af39bdde689f.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
eu.roadmunk.com - Only Chrome is supported - <!-- @browser: Firefox 66.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:66.0) Gecko/20100101 Firefox/66.0 --> <!-- @reported_with: web --> **URL**: https://eu.roadmunk.com/ **Browser / Version**: Firefox 66.0 **Operating System**: Mac OS X 10.12 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: Roadmunk is damaging the openness of the web - blocks all browsers except Google Chrome **Steps to Reproduce**: Opened the url in Firefox (Vivaldi,Safari,etc) none are accepted other than Google Chrome. Particularly gauling in this case is that Vivaldi is a Chromium based browser and likely supports whatever proprietary features Roadmunk felt were necessary - but they block it anyway. This sort of behaviour should be called out and the companies shamed. In the light of Edge dropping their own rendering engine in favour of becoming yet another Chromium clone the diversity of the web is at risk and we could end up with another ie6 world where Google calls all the shots. [![Screenshot Description](https://webcompat.com/uploads/2019/4/dfc78383-77cd-4221-bcc7-af39bdde689f-thumb.jpg)](https://webcompat.com/uploads/2019/4/dfc78383-77cd-4221-bcc7-af39bdde689f.jpg) <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
eu roadmunk com only chrome is supported url browser version firefox operating system mac os x tested another browser yes problem type site is not usable description roadmunk is damaging the openness of the web blocks all browsers except google chrome steps to reproduce opened the url in firefox vivaldi safari etc none are accepted other than google chrome particularly gauling in this case is that vivaldi is a chromium based browser and likely supports whatever proprietary features roadmunk felt were necessary but they block it anyway this sort of behaviour should be called out and the companies shamed in the light of edge dropping their own rendering engine in favour of becoming yet another chromium clone the diversity of the web is at risk and we could end up with another world where google calls all the shots browser configuration none from with ❤️
0
4,288
21,565,934,721
IssuesEvent
2022-05-01 21:21:49
pypiserver/pypiserver
https://api.github.com/repos/pypiserver/pypiserver
closed
Preparing new `pypiserver` release
type.Maintainance status.CRITICAL
> TO BE ELABORATED. # Purpose Hey everyone, I wanted to make a public strategy for the upcoming release collected in this issue as, unfortunately, it is taking me longer than I have anticipated. I'm getting there but would like to apologize for the delays. I'm promising to get it up and running. Here I would like to make a draft roadmap of things I'm working on to get it ready. Any help and suggestions on this are very welcome! ## Timeline - [x] Receive the release credentials and access to DockerHub and Pypi - [x] Prepare a low-labor release process (in progress) - [x] Setup the release candidate preparation flow (#417) - [x] Setup the release CI/CD pipeline (#423) - [x] Get a new release published - [ ] Document updated release process
True
Preparing new `pypiserver` release - > TO BE ELABORATED. # Purpose Hey everyone, I wanted to make a public strategy for the upcoming release collected in this issue as, unfortunately, it is taking me longer than I have anticipated. I'm getting there but would like to apologize for the delays. I'm promising to get it up and running. Here I would like to make a draft roadmap of things I'm working on to get it ready. Any help and suggestions on this are very welcome! ## Timeline - [x] Receive the release credentials and access to DockerHub and Pypi - [x] Prepare a low-labor release process (in progress) - [x] Setup the release candidate preparation flow (#417) - [x] Setup the release CI/CD pipeline (#423) - [x] Get a new release published - [ ] Document updated release process
main
preparing new pypiserver release to be elaborated purpose hey everyone i wanted to make a public strategy for the upcoming release collected in this issue as unfortunately it is taking me longer than i have anticipated i m getting there but would like to apologize for the delays i m promising to get it up and running here i would like to make a draft roadmap of things i m working on to get it ready any help and suggestions on this are very welcome timeline receive the release credentials and access to dockerhub and pypi prepare a low labor release process in progress setup the release candidate preparation flow setup the release ci cd pipeline get a new release published document updated release process
1
1,791
6,575,887,132
IssuesEvent
2017-09-11 17:42:52
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
apt force with deb doesn't force
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME apt ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /home/jenkins/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Ubuntu ##### SUMMARY Trying to force install a deb file doesn't (always) work. ##### STEPS TO REPRODUCE Custom built nginx deb and install it. Then change some of the included modules, and create another custom deb for that nginx. Now try install it via ansible and force. ``` - name: install nginx pagespeed apt: > deb={{ nginx_pagespeed_deb_url }} dpkg_options='force-confold,force-confdef' force=yes ``` I suspect that because the deb has the same header (and version) as the existing deb ansible is refusing to say changed. The contents of the deb has changed with regards to what modules were compiled in, but the version is the same. ##### EXPECTED RESULTS Expected deb to be installed as if I did `dpkg -i file.deb` And to see ansible emit the changed message. Changed. ##### ACTUAL RESULTS ``` ok: [host2.example.org] => {"changed": false, "diff": "", "invocation": {"module_args": {"allow_unauthenticated": false, "autoremove": false, "cache_valid_time": null, "deb": "/tmp/ansible_mj0640/nginx_1.11.1-1_amd64.deb", "default_release": null, "dpkg_options": "force-confold,force-confdef", "force": true, "install_recommends": null, "only_upgrade": false, "package": null, "purge": false, "state": "present", "update_cache": false, "upgrade": null}, "module_name": "apt"}, "stderr": "", "stdout": "", "stdout_lines": []} ```
True
apt force with deb doesn't force - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME apt ##### ANSIBLE VERSION ``` ansible 2.1.1.0 config file = /home/jenkins/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Ubuntu ##### SUMMARY Trying to force install a deb file doesn't (always) work. ##### STEPS TO REPRODUCE Custom built nginx deb and install it. Then change some of the included modules, and create another custom deb for that nginx. Now try install it via ansible and force. ``` - name: install nginx pagespeed apt: > deb={{ nginx_pagespeed_deb_url }} dpkg_options='force-confold,force-confdef' force=yes ``` I suspect that because the deb has the same header (and version) as the existing deb ansible is refusing to say changed. The contents of the deb has changed with regards to what modules were compiled in, but the version is the same. ##### EXPECTED RESULTS Expected deb to be installed as if I did `dpkg -i file.deb` And to see ansible emit the changed message. Changed. ##### ACTUAL RESULTS ``` ok: [host2.example.org] => {"changed": false, "diff": "", "invocation": {"module_args": {"allow_unauthenticated": false, "autoremove": false, "cache_valid_time": null, "deb": "/tmp/ansible_mj0640/nginx_1.11.1-1_amd64.deb", "default_release": null, "dpkg_options": "force-confold,force-confdef", "force": true, "install_recommends": null, "only_upgrade": false, "package": null, "purge": false, "state": "present", "update_cache": false, "upgrade": null}, "module_name": "apt"}, "stderr": "", "stdout": "", "stdout_lines": []} ```
main
apt force with deb doesn t force issue type bug report component name apt ansible version ansible config file home jenkins ansible ansible cfg configured module search path default w o overrides configuration os environment ubuntu summary trying to force install a deb file doesn t always work steps to reproduce custom built nginx deb and install it then change some of the included modules and create another custom deb for that nginx now try install it via ansible and force name install nginx pagespeed apt deb nginx pagespeed deb url dpkg options force confold force confdef force yes i suspect that because the deb has the same header and version as the existing deb ansible is refusing to say changed the contents of the deb has changed with regards to what modules were compiled in but the version is the same expected results expected deb to be installed as if i did dpkg i file deb and to see ansible emit the changed message changed actual results ok changed false diff invocation module args allow unauthenticated false autoremove false cache valid time null deb tmp ansible nginx deb default release null dpkg options force confold force confdef force true install recommends null only upgrade false package null purge false state present update cache false upgrade null module name apt stderr stdout stdout lines
1
5,768
30,572,661,171
IssuesEvent
2023-07-21 00:36:38
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Synchronize cell "active" and "focused" states
type: enhancement work: frontend status: ready restricted: maintainers
## Current behavior - Every cell in the sheet has the following state: - **active** or **inactive** No more than one cell can be active at a time, and we do our best (imperatively) to ensure that we always have one active cell. - **selected** or **unselected** Multiple cells can be selected within the sheet. We do our best (imperatively) to ensure that the active cell is also selected. - **focused** or **unfocused** No more than one cell can be focused at a time. But it's quite common to have _zero_ cells focused -- and necessary so that the browser can focus other UI on the page. We have rather airtight logic to ensure that a _focused_ cell is always _active_. - The user can interact with a table cell via the keyboard in several ways (e.g. `Enter` to enter edit mode, arrow keys to move the active cell, etc) -- but **only when the cell is _focused_**. If you click on some other UI like empty space above the table or a button somewhere else, the cell loses focus and those keyboard actions no longer have any effect even though the cell may still be _active_, but not the other way around. - We indicate that a cell is _active_ by displaying a thick blue border if the cell allows editing (or displaying a grey border if the cell does not allow editing). - We have no indicator for the cell's _focused_ state. ## Problems with the current behavior - I don't care whether a cell is _active_ -- I care whether a cell is _focused_. I can't identify any special behavior that an active-unfocused cell has over an inactive-unfocused cell. I can't do anything with it if it's not focused. Maybe I'm overlooking something here though. - The current behavior has the effect of making me unsure whether I can use the keyboard to interact with the cell, so I subconsciously end up using the mouse more. ## Comparisons with other products - **Google Sheets** and **LibreOffice Calc** both behave very similarly to our current behavior. While that may be an argument for keeping it, I still don't like it. - **AirTable** almost entirely synchronizes their active cell with their focused cell, but they also discard the entire selection (and active cell) when the user focuses on other UI. I like the simplicity of this behavior, but it wouldn't be compatible with our table inspector design because we need to maintain a selection of cells while user is focusing on UI within the table inspector. ## Desired behavior - Cell selection would work the same as it currently does, but we would eliminate the active/inactive state, keeping only focused/unfocused. When a cell is focused it gets a thick blue border (or grey if disabled). When a cell is unfocused it doesn't have the border. That's actually the behavior I built within the Record Selector, and I think it works well. @kgodey @pavish @rajatvijay @ghislaineguerin I'm curious what you think about this idea. It seems like a change that would probably be easy to make.
True
Synchronize cell "active" and "focused" states - ## Current behavior - Every cell in the sheet has the following state: - **active** or **inactive** No more than one cell can be active at a time, and we do our best (imperatively) to ensure that we always have one active cell. - **selected** or **unselected** Multiple cells can be selected within the sheet. We do our best (imperatively) to ensure that the active cell is also selected. - **focused** or **unfocused** No more than one cell can be focused at a time. But it's quite common to have _zero_ cells focused -- and necessary so that the browser can focus other UI on the page. We have rather airtight logic to ensure that a _focused_ cell is always _active_. - The user can interact with a table cell via the keyboard in several ways (e.g. `Enter` to enter edit mode, arrow keys to move the active cell, etc) -- but **only when the cell is _focused_**. If you click on some other UI like empty space above the table or a button somewhere else, the cell loses focus and those keyboard actions no longer have any effect even though the cell may still be _active_, but not the other way around. - We indicate that a cell is _active_ by displaying a thick blue border if the cell allows editing (or displaying a grey border if the cell does not allow editing). - We have no indicator for the cell's _focused_ state. ## Problems with the current behavior - I don't care whether a cell is _active_ -- I care whether a cell is _focused_. I can't identify any special behavior that an active-unfocused cell has over an inactive-unfocused cell. I can't do anything with it if it's not focused. Maybe I'm overlooking something here though. - The current behavior has the effect of making me unsure whether I can use the keyboard to interact with the cell, so I subconsciously end up using the mouse more. ## Comparisons with other products - **Google Sheets** and **LibreOffice Calc** both behave very similarly to our current behavior. While that may be an argument for keeping it, I still don't like it. - **AirTable** almost entirely synchronizes their active cell with their focused cell, but they also discard the entire selection (and active cell) when the user focuses on other UI. I like the simplicity of this behavior, but it wouldn't be compatible with our table inspector design because we need to maintain a selection of cells while user is focusing on UI within the table inspector. ## Desired behavior - Cell selection would work the same as it currently does, but we would eliminate the active/inactive state, keeping only focused/unfocused. When a cell is focused it gets a thick blue border (or grey if disabled). When a cell is unfocused it doesn't have the border. That's actually the behavior I built within the Record Selector, and I think it works well. @kgodey @pavish @rajatvijay @ghislaineguerin I'm curious what you think about this idea. It seems like a change that would probably be easy to make.
main
synchronize cell active and focused states current behavior every cell in the sheet has the following state active or inactive no more than one cell can be active at a time and we do our best imperatively to ensure that we always have one active cell selected or unselected multiple cells can be selected within the sheet we do our best imperatively to ensure that the active cell is also selected focused or unfocused no more than one cell can be focused at a time but it s quite common to have zero cells focused and necessary so that the browser can focus other ui on the page we have rather airtight logic to ensure that a focused cell is always active the user can interact with a table cell via the keyboard in several ways e g enter to enter edit mode arrow keys to move the active cell etc but only when the cell is focused if you click on some other ui like empty space above the table or a button somewhere else the cell loses focus and those keyboard actions no longer have any effect even though the cell may still be active but not the other way around we indicate that a cell is active by displaying a thick blue border if the cell allows editing or displaying a grey border if the cell does not allow editing we have no indicator for the cell s focused state problems with the current behavior i don t care whether a cell is active i care whether a cell is focused i can t identify any special behavior that an active unfocused cell has over an inactive unfocused cell i can t do anything with it if it s not focused maybe i m overlooking something here though the current behavior has the effect of making me unsure whether i can use the keyboard to interact with the cell so i subconsciously end up using the mouse more comparisons with other products google sheets and libreoffice calc both behave very similarly to our current behavior while that may be an argument for keeping it i still don t like it airtable almost entirely synchronizes their active cell with their focused cell but they also discard the entire selection and active cell when the user focuses on other ui i like the simplicity of this behavior but it wouldn t be compatible with our table inspector design because we need to maintain a selection of cells while user is focusing on ui within the table inspector desired behavior cell selection would work the same as it currently does but we would eliminate the active inactive state keeping only focused unfocused when a cell is focused it gets a thick blue border or grey if disabled when a cell is unfocused it doesn t have the border that s actually the behavior i built within the record selector and i think it works well kgodey pavish rajatvijay ghislaineguerin i m curious what you think about this idea it seems like a change that would probably be easy to make
1
621
10,916,568,248
IssuesEvent
2019-11-21 13:37:19
Crevette/Test
https://api.github.com/repos/Crevette/Test
opened
Migration ECM Prod-EU - 17.2
DB - peopleask Deployment Env - prod-eu
Could you deploy the 17.2 version of the SQL for ECM on Prod-EU. Thank you !
1.0
Migration ECM Prod-EU - 17.2 - Could you deploy the 17.2 version of the SQL for ECM on Prod-EU. Thank you !
non_main
migration ecm prod eu could you deploy the version of the sql for ecm on prod eu thank you
0
359,496
25,240,866,736
IssuesEvent
2022-11-15 07:17:23
einsteinpy/einsteinpy
https://api.github.com/repos/einsteinpy/einsteinpy
opened
Add new example Jupyter notebooks / edit old ones
good first issue documentation
### 🐞 **Problem** We need new [example notebooks](https://github.com/einsteinpy/einsteinpy/tree/main/docs/source/examples) or edits to older notebooks to highlight the feature additions and API changes that EinsteinPy has seen over the years. Recently, two PRs added features that could use a notebook demonstrating them. This issue will track multiple features of EinsteinPy that should be properly documented. - [ ] PR #625 added support for custom metrics functions, outside of electro-vacuum (EV) solutions. We need one or two examples showcasing this (preferably with a non-EV metric). - [ ] PR #612 (issue #610) improved scaling in geodesic plots. Most of the related example notebook were updated (See discussion in #612), but [one](https://github.com/einsteinpy/einsteinpy/blob/main/docs/source/examples/Visualizing%20Precession%20in%20Schwarzschild%20Spacetime.ipynb) had to be left out at the time due to merge conflicts. It can be edited now. - [ ] *More to be added* ***If you are interested in contributing towards this issue, please do not hesitate to ask for clarifications.*** ### 📋 **Steps to solve the problem** If you are a new contributor, see this: https://docs.einsteinpy.org/en/latest/dev_guide.html. * Comment below about what you've started working on. * Add, commit, push your changes. * Submit a pull request. * Add this in the comments: - `Addresses #<put issue number here>` if you are partially fixing the issue. - `Fixes #<put issue number here>` if you are completely fixing the issue. * Ask for a review in comments section of pull request. * Celebrate your contribution to this project 🎉.
1.0
Add new example Jupyter notebooks / edit old ones - ### 🐞 **Problem** We need new [example notebooks](https://github.com/einsteinpy/einsteinpy/tree/main/docs/source/examples) or edits to older notebooks to highlight the feature additions and API changes that EinsteinPy has seen over the years. Recently, two PRs added features that could use a notebook demonstrating them. This issue will track multiple features of EinsteinPy that should be properly documented. - [ ] PR #625 added support for custom metrics functions, outside of electro-vacuum (EV) solutions. We need one or two examples showcasing this (preferably with a non-EV metric). - [ ] PR #612 (issue #610) improved scaling in geodesic plots. Most of the related example notebook were updated (See discussion in #612), but [one](https://github.com/einsteinpy/einsteinpy/blob/main/docs/source/examples/Visualizing%20Precession%20in%20Schwarzschild%20Spacetime.ipynb) had to be left out at the time due to merge conflicts. It can be edited now. - [ ] *More to be added* ***If you are interested in contributing towards this issue, please do not hesitate to ask for clarifications.*** ### 📋 **Steps to solve the problem** If you are a new contributor, see this: https://docs.einsteinpy.org/en/latest/dev_guide.html. * Comment below about what you've started working on. * Add, commit, push your changes. * Submit a pull request. * Add this in the comments: - `Addresses #<put issue number here>` if you are partially fixing the issue. - `Fixes #<put issue number here>` if you are completely fixing the issue. * Ask for a review in comments section of pull request. * Celebrate your contribution to this project 🎉.
non_main
add new example jupyter notebooks edit old ones 🐞 problem we need new or edits to older notebooks to highlight the feature additions and api changes that einsteinpy has seen over the years recently two prs added features that could use a notebook demonstrating them this issue will track multiple features of einsteinpy that should be properly documented pr added support for custom metrics functions outside of electro vacuum ev solutions we need one or two examples showcasing this preferably with a non ev metric pr issue improved scaling in geodesic plots most of the related example notebook were updated see discussion in but had to be left out at the time due to merge conflicts it can be edited now more to be added if you are interested in contributing towards this issue please do not hesitate to ask for clarifications 📋 steps to solve the problem if you are a new contributor see this comment below about what you ve started working on add commit push your changes submit a pull request add this in the comments addresses if you are partially fixing the issue fixes if you are completely fixing the issue ask for a review in comments section of pull request celebrate your contribution to this project 🎉
0
140,164
11,303,466,227
IssuesEvent
2020-01-17 20:11:25
ni/nimi-python
https://api.github.com/repos/ni/nimi-python
closed
Eliminate the need to run GNU make while executing system tests
priority-medium test
nimi-bot builds the module installers (wheels) and installs them before it runs system tests. In order to do so, GNU make is involved which means that nimi-bot needs to have mingw installed. We are no longer recommending / supporting [MinGW](http://www.mingw.org). Customers should use WSL. So one alternative is WSL. But WSL image with 32-bit Python is not easy to find. Also WSL solution would involve 2 different python interpreters (one of Windows and one for WSL). We think it's a better solution to completely eliminate the need for running GNU make. This will simplify the setup of nimi-bot altogether. We could have a way to build the installers outside of GNU make, or we could simply install from `generated/` using setup.py with a small loss of test coverage (we would not be running the wheels anymore). Either way it's better than requiring MinGW.
1.0
Eliminate the need to run GNU make while executing system tests - nimi-bot builds the module installers (wheels) and installs them before it runs system tests. In order to do so, GNU make is involved which means that nimi-bot needs to have mingw installed. We are no longer recommending / supporting [MinGW](http://www.mingw.org). Customers should use WSL. So one alternative is WSL. But WSL image with 32-bit Python is not easy to find. Also WSL solution would involve 2 different python interpreters (one of Windows and one for WSL). We think it's a better solution to completely eliminate the need for running GNU make. This will simplify the setup of nimi-bot altogether. We could have a way to build the installers outside of GNU make, or we could simply install from `generated/` using setup.py with a small loss of test coverage (we would not be running the wheels anymore). Either way it's better than requiring MinGW.
non_main
eliminate the need to run gnu make while executing system tests nimi bot builds the module installers wheels and installs them before it runs system tests in order to do so gnu make is involved which means that nimi bot needs to have mingw installed we are no longer recommending supporting customers should use wsl so one alternative is wsl but wsl image with bit python is not easy to find also wsl solution would involve different python interpreters one of windows and one for wsl we think it s a better solution to completely eliminate the need for running gnu make this will simplify the setup of nimi bot altogether we could have a way to build the installers outside of gnu make or we could simply install from generated using setup py with a small loss of test coverage we would not be running the wheels anymore either way it s better than requiring mingw
0
67,283
14,861,164,822
IssuesEvent
2021-01-18 22:09:21
gate5/test2
https://api.github.com/repos/gate5/test2
opened
CVE-2020-36182 (High) detected in jackson-databind-2.9.10.7.jar
security vulnerability
## CVE-2020-36182 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: test2/target/BookStore-6.0.1/WEB-INF/lib/jackson-databind-2.9.10.7.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.7/jackson-databind-2.9.10.7.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gate5/test2/commit/c3cd1a15653981e818c45c95cf12d16b04d6cf85">c3cd1a15653981e818c45c95cf12d16b04d6cf85</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.cpdsadapter.DriverAdapterCPDS. <p>Publish Date: 2021-01-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36182>CVE-2020-36182</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/3004">https://github.com/FasterXML/jackson-databind/issues/3004</a></p> <p>Release Date: 2021-01-07</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p>
True
CVE-2020-36182 (High) detected in jackson-databind-2.9.10.7.jar - ## CVE-2020-36182 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.7.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: test2/target/BookStore-6.0.1/WEB-INF/lib/jackson-databind-2.9.10.7.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.10.7/jackson-databind-2.9.10.7.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.7.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/gate5/test2/commit/c3cd1a15653981e818c45c95cf12d16b04d6cf85">c3cd1a15653981e818c45c95cf12d16b04d6cf85</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp2.cpdsadapter.DriverAdapterCPDS. <p>Publish Date: 2021-01-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36182>CVE-2020-36182</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/3004">https://github.com/FasterXML/jackson-databind/issues/3004</a></p> <p>Release Date: 2021-01-07</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p>
non_main
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library target bookstore web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp cpdsadapter driveradaptercpds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind
0
2,032
6,812,452,205
IssuesEvent
2017-11-06 03:11:54
dgets/CPURecord
https://api.github.com/repos/dgets/CPURecord
opened
Other architectures crashing
bug help wanted maintainability
Luckily, I had all of the VMs on my computer down, for backup purposes, for awhile here. If it wasn't for that, and not having the USB cable to connect my 'droid to the computer, I might not have noticed how other architectures, and/or /proc layouts, are making the app hang. For instance, I switched from my _LG G6_ (four coar AArch64 rev. 1 -- physical device), to an emulated _Nexus 5X_ (api 26), and now I can't even get the stats box to populate. Need to track down where things like this might be coming from and add cases to deal.
True
Other architectures crashing - Luckily, I had all of the VMs on my computer down, for backup purposes, for awhile here. If it wasn't for that, and not having the USB cable to connect my 'droid to the computer, I might not have noticed how other architectures, and/or /proc layouts, are making the app hang. For instance, I switched from my _LG G6_ (four coar AArch64 rev. 1 -- physical device), to an emulated _Nexus 5X_ (api 26), and now I can't even get the stats box to populate. Need to track down where things like this might be coming from and add cases to deal.
main
other architectures crashing luckily i had all of the vms on my computer down for backup purposes for awhile here if it wasn t for that and not having the usb cable to connect my droid to the computer i might not have noticed how other architectures and or proc layouts are making the app hang for instance i switched from my lg four coar rev physical device to an emulated nexus api and now i can t even get the stats box to populate need to track down where things like this might be coming from and add cases to deal
1
3,628
14,676,409,957
IssuesEvent
2020-12-30 20:07:06
DLR-RM/stable-baselines3
https://api.github.com/repos/DLR-RM/stable-baselines3
closed
[Question] Models keep using CPU for training
Maintainers on vacation question
Do I need specific versions of TF2.x/CUDA/cuDNN to be able to train on GPU? I tried running some of the CoLab notebooks and the model did pick up the GPU on creation, but when I tried it locally it keeps using the CPU for training. I have successfully trained tf models on my GPU before and tf does see it when calling device_lib.list_local_devices(). I have also tried passing device='cuda' when creating the model (specifically SAC) but still defaults to using the CPU. I am using the following versions: - Nvidia Drivers 460.79 - TensorFlow 2.4.0 - CUDA 11.0 - cuDNN 8.0.4
True
[Question] Models keep using CPU for training - Do I need specific versions of TF2.x/CUDA/cuDNN to be able to train on GPU? I tried running some of the CoLab notebooks and the model did pick up the GPU on creation, but when I tried it locally it keeps using the CPU for training. I have successfully trained tf models on my GPU before and tf does see it when calling device_lib.list_local_devices(). I have also tried passing device='cuda' when creating the model (specifically SAC) but still defaults to using the CPU. I am using the following versions: - Nvidia Drivers 460.79 - TensorFlow 2.4.0 - CUDA 11.0 - cuDNN 8.0.4
main
models keep using cpu for training do i need specific versions of x cuda cudnn to be able to train on gpu i tried running some of the colab notebooks and the model did pick up the gpu on creation but when i tried it locally it keeps using the cpu for training i have successfully trained tf models on my gpu before and tf does see it when calling device lib list local devices i have also tried passing device cuda when creating the model specifically sac but still defaults to using the cpu i am using the following versions nvidia drivers tensorflow cuda cudnn
1
64,592
8,749,393,615
IssuesEvent
2018-12-13 16:17:40
webhintio/hint
https://api.github.com/repos/webhintio/hint
closed
Update titles and other tweaks of some of the docs
area:documentation difficulty:easy good first issue
Now that we are publishing everything in the website (extensions, parsers, connectors, etc.) we should look into updating some of the names. E.g.: * All formatter, parser and connector titles are `XXXX formatter/parser/connector` which looks weird in the menu. We can probably drop the last part. ![image](https://user-images.githubusercontent.com/606594/48727204-7273d780-ebe6-11e8-9e2c-7af9982b7b60.png) * VS Code extension is `webhint extension for Visual Studio Code`. In this case I'd just say `VS Code` or something similar. Also we should add a bit more information and maybe add a gif or similar. @antross @alrra thoughts?
1.0
Update titles and other tweaks of some of the docs - Now that we are publishing everything in the website (extensions, parsers, connectors, etc.) we should look into updating some of the names. E.g.: * All formatter, parser and connector titles are `XXXX formatter/parser/connector` which looks weird in the menu. We can probably drop the last part. ![image](https://user-images.githubusercontent.com/606594/48727204-7273d780-ebe6-11e8-9e2c-7af9982b7b60.png) * VS Code extension is `webhint extension for Visual Studio Code`. In this case I'd just say `VS Code` or something similar. Also we should add a bit more information and maybe add a gif or similar. @antross @alrra thoughts?
non_main
update titles and other tweaks of some of the docs now that we are publishing everything in the website extensions parsers connectors etc we should look into updating some of the names e g all formatter parser and connector titles are xxxx formatter parser connector which looks weird in the menu we can probably drop the last part vs code extension is webhint extension for visual studio code in this case i d just say vs code or something similar also we should add a bit more information and maybe add a gif or similar antross alrra thoughts
0
142,757
19,102,988,314
IssuesEvent
2021-11-30 01:52:26
Nehamaefi/Efigit
https://api.github.com/repos/Nehamaefi/Efigit
closed
CVE-2018-12022 (High) detected in jackson-databind-2.9.4.jar - autoclosed
security vulnerability
## CVE-2018-12022 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Efigit/apps/rest-showcase/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar,epository/com/fasterxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Nehamaefi/Efigit/commit/72c969e9db891da76ca4ea40803b5d450c212b27">72c969e9db891da76ca4ea40803b5d450c212b27</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022>CVE-2018-12022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: 2.7.9.4, 2.8.11.2, 2.9.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.4","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.7.9.4, 2.8.11.2, 2.9.6"}],"vulnerabilityIdentifier":"CVE-2018-12022","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2018-12022 (High) detected in jackson-databind-2.9.4.jar - autoclosed - ## CVE-2018-12022 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.4.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: Efigit/apps/rest-showcase/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar,epository/com/fasterxml/jackson/core/jackson-databind/2.9.4/jackson-databind-2.9.4.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Nehamaefi/Efigit/commit/72c969e9db891da76ca4ea40803b5d450c212b27">72c969e9db891da76ca4ea40803b5d450c212b27</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-03-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022>CVE-2018-12022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-12022</a></p> <p>Release Date: 2019-03-21</p> <p>Fix Resolution: 2.7.9.4, 2.8.11.2, 2.9.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.4","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.7.9.4, 2.8.11.2, 2.9.6"}],"vulnerabilityIdentifier":"CVE-2018-12022","vulnerabilityDetails":"An issue was discovered in FasterXML jackson-databind prior to 2.7.9.4, 2.8.11.2, and 2.9.6. When Default Typing is enabled (either globally or for a specific property), the service has the Jodd-db jar (for database access for the Jodd framework) in the classpath, and an attacker can provide an LDAP service to access, it is possible to make the service execute a malicious payload.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-12022","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_main
cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file efigit apps rest showcase pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar epository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details an issue was discovered in fasterxml jackson databind prior to and when default typing is enabled either globally or for a specific property the service has the jodd db jar for database access for the jodd framework in the classpath and an attacker can provide an ldap service to access it is possible to make the service execute a malicious payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in fasterxml jackson databind prior to and when default typing is enabled either globally or for a specific property the service has the jodd db jar for database access for the jodd framework in the classpath and an attacker can provide an ldap service to access it is possible to make the service execute a malicious payload vulnerabilityurl
0
1,658
6,574,047,651
IssuesEvent
2017-09-11 11:14:42
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Elasticache: Use of cache security groups is not permitted in this API version for your account
affects_2.2 aws bug_report cloud waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> elasticache ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION [defaults] hostfile = hosts <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> macOS Sierra ##### SUMMARY <!--- Explain the problem briefly --> I cannot create a cluster elasticache like the example http://docs.ansible.com/ansible/elasticache_module.html ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> ansible-playbook example.yml <!--- Paste example playbooks or commands between quotes below --> ``` - elasticache: name: "test-please-delete" state: present engine: memcached cache_engine_version: 1.4.14 node_type: cache.m1.small num_nodes: 1 cache_port: 11211 cache_security_groups: - default region: us-east-1 zone: us-east-1d ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> cluster created ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Use of cache security groups is not permitted in this API version for your account <!--- Paste verbatim command output between quotes below --> ``` ```
True
Elasticache: Use of cache security groups is not permitted in this API version for your account - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> elasticache ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.2.0.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION [defaults] hostfile = hosts <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> macOS Sierra ##### SUMMARY <!--- Explain the problem briefly --> I cannot create a cluster elasticache like the example http://docs.ansible.com/ansible/elasticache_module.html ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> ansible-playbook example.yml <!--- Paste example playbooks or commands between quotes below --> ``` - elasticache: name: "test-please-delete" state: present engine: memcached cache_engine_version: 1.4.14 node_type: cache.m1.small num_nodes: 1 cache_port: 11211 cache_security_groups: - default region: us-east-1 zone: us-east-1d ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> cluster created ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Use of cache security groups is not permitted in this API version for your account <!--- Paste verbatim command output between quotes below --> ``` ```
main
elasticache use of cache security groups is not permitted in this api version for your account issue type bug report component name elasticache ansible version ansible config file configured module search path default w o overrides configuration hostfile hosts mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific macos sierra summary i cannot create a cluster elasticache like the example steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used ansible playbook example yml elasticache name test please delete state present engine memcached cache engine version node type cache small num nodes cache port cache security groups default region us east zone us east expected results cluster created actual results use of cache security groups is not permitted in this api version for your account
1
208,043
23,546,524,862
IssuesEvent
2022-08-21 07:33:23
ghc-dev/3069454_1867
https://api.github.com/repos/ghc-dev/3069454_1867
opened
save-server-1.0.7.tgz: 1 vulnerabilities (highest severity is: 7.5)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>save-server-1.0.7.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bcrypt/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/3069454_1867/commit/aad3283625febcf6c3feede90effeab26c015909">aad3283625febcf6c3feede90effeab26c015909</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-7689](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7689) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | bcrypt-4.0.1.tgz | Transitive | 1.1.0 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7689</summary> ### Vulnerable Library - <b>bcrypt-4.0.1.tgz</b></p> <p>A bcrypt library for NodeJS.</p> <p>Library home page: <a href="https://registry.npmjs.org/bcrypt/-/bcrypt-4.0.1.tgz">https://registry.npmjs.org/bcrypt/-/bcrypt-4.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bcrypt/package.json</p> <p> Dependency Hierarchy: - save-server-1.0.7.tgz (Root Library) - :x: **bcrypt-4.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/3069454_1867/commit/aad3283625febcf6c3feede90effeab26c015909">aad3283625febcf6c3feede90effeab26c015909</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Data is truncated wrong when its length is greater than 255 bytes. <p>Publish Date: 2020-07-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7689>CVE-2020-7689</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7689">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7689</a></p> <p>Release Date: 2020-07-01</p> <p>Fix Resolution (bcrypt): 5.0.0</p> <p>Direct dependency fix Resolution (save-server): 1.1.0</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details> *** <p>In order to enable automatic remediation for this issue, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
True
save-server-1.0.7.tgz: 1 vulnerabilities (highest severity is: 7.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>save-server-1.0.7.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bcrypt/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/3069454_1867/commit/aad3283625febcf6c3feede90effeab26c015909">aad3283625febcf6c3feede90effeab26c015909</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-7689](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7689) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | bcrypt-4.0.1.tgz | Transitive | 1.1.0 | &#9989; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-7689</summary> ### Vulnerable Library - <b>bcrypt-4.0.1.tgz</b></p> <p>A bcrypt library for NodeJS.</p> <p>Library home page: <a href="https://registry.npmjs.org/bcrypt/-/bcrypt-4.0.1.tgz">https://registry.npmjs.org/bcrypt/-/bcrypt-4.0.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/bcrypt/package.json</p> <p> Dependency Hierarchy: - save-server-1.0.7.tgz (Root Library) - :x: **bcrypt-4.0.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ghc-dev/3069454_1867/commit/aad3283625febcf6c3feede90effeab26c015909">aad3283625febcf6c3feede90effeab26c015909</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Data is truncated wrong when its length is greater than 255 bytes. <p>Publish Date: 2020-07-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7689>CVE-2020-7689</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7689">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7689</a></p> <p>Release Date: 2020-07-01</p> <p>Fix Resolution (bcrypt): 5.0.0</p> <p>Direct dependency fix Resolution (save-server): 1.1.0</p> </p> <p></p> <p>In order to enable automatic remediation, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p> </details> *** <p>In order to enable automatic remediation for this issue, please create <a target="_blank" href="https://docs.mend.io/bundle/integrations/page/mend_for_github_com.html#MendforGitHub.com-RemediateSettings(remediateSettings)">workflow rules</a></p>
non_main
save server tgz vulnerabilities highest severity is vulnerable library save server tgz path to dependency file package json path to vulnerable library node modules bcrypt package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high bcrypt tgz transitive details cve vulnerable library bcrypt tgz a bcrypt library for nodejs library home page a href path to dependency file package json path to vulnerable library node modules bcrypt package json dependency hierarchy save server tgz root library x bcrypt tgz vulnerable library found in head commit a href found in base branch main vulnerability details data is truncated wrong when its length is greater than bytes publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution bcrypt direct dependency fix resolution save server in order to enable automatic remediation please create in order to enable automatic remediation for this issue please create
0
2,034
6,846,664,052
IssuesEvent
2017-11-13 12:57:37
dgets/CPURecord
https://api.github.com/repos/dgets/CPURecord
opened
*Details and *Record superclass potential & implementation
enhancement help wanted maintainability
It might be a good idea to look at the structures of _CPUDetails_ and _CPURecord_, and see how much common methodology there is going to be in the implementation of _BattDetails_ and _BattRecord_. Superclasses would save some time and make things a lot easier to maintain.
True
*Details and *Record superclass potential & implementation - It might be a good idea to look at the structures of _CPUDetails_ and _CPURecord_, and see how much common methodology there is going to be in the implementation of _BattDetails_ and _BattRecord_. Superclasses would save some time and make things a lot easier to maintain.
main
details and record superclass potential implementation it might be a good idea to look at the structures of cpudetails and cpurecord and see how much common methodology there is going to be in the implementation of battdetails and battrecord superclasses would save some time and make things a lot easier to maintain
1
42,394
5,435,613,360
IssuesEvent
2017-03-05 18:28:11
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
"service_protocol discovery port forwarding - default port" is flaky on Linux?
dev: tests dev: tool regression
https://build.chromium.org/p/client.flutter/builders/Linux/builds/2860/steps/test.dart/logs/stdio Not sure if this test recently became flaky? ``` 00:29 +106 -1: service_protocol discovery port forwarding - default port Expected: <54777> Actual: <54779> package:test expect test/protocol_discovery_test.dart 86:7 main.<fn>.<fn> test/src/context.dart 81:26 testUsingContext.<fn>.<fn> package:flutter_tools/src/base/context.dart 76:26 AppContext._run package:flutter_tools/src/base/context.dart 66:13 AppContext.runInZone.<fn> dart:async runZoned package:flutter_tools/src/base/context.dart 65:12 AppContext.runInZone test/src/context.dart 72:32 testUsingContext.<fn> ===== asynchronous gap =========================== dart:async _Completer.completeError test/protocol_discovery_test.dart 91:6 main.<fn>.<fn> ===== asynchronous gap =========================== dart:async _asyncThenWrapperHelper test/protocol_discovery_test.dart main.<fn>.<fn> test/src/context.dart 81:26 testUsingContext.<fn>.<fn> package:flutter_tools/src/base/context.dart 76:26 AppContext._run ===== asynchronous gap =========================== dart:async Future.Future.microtask package:flutter_tools/src/base/context.dart AppContext._run package:flutter_tools/src/base/context.dart 66:13 AppContext.runInZone.<fn> dart:async runZoned package:flutter_tools/src/base/context.dart 65:12 AppContext.runInZone test/src/context.dart 72:32 testUsingContext.<fn> ===== asynchronous gap =========================== dart:async Future.Future.microtask test/src/context.dart testUsingContext.<fn> package:test Declarer.test.<fn>.<fn> ```
1.0
"service_protocol discovery port forwarding - default port" is flaky on Linux? - https://build.chromium.org/p/client.flutter/builders/Linux/builds/2860/steps/test.dart/logs/stdio Not sure if this test recently became flaky? ``` 00:29 +106 -1: service_protocol discovery port forwarding - default port Expected: <54777> Actual: <54779> package:test expect test/protocol_discovery_test.dart 86:7 main.<fn>.<fn> test/src/context.dart 81:26 testUsingContext.<fn>.<fn> package:flutter_tools/src/base/context.dart 76:26 AppContext._run package:flutter_tools/src/base/context.dart 66:13 AppContext.runInZone.<fn> dart:async runZoned package:flutter_tools/src/base/context.dart 65:12 AppContext.runInZone test/src/context.dart 72:32 testUsingContext.<fn> ===== asynchronous gap =========================== dart:async _Completer.completeError test/protocol_discovery_test.dart 91:6 main.<fn>.<fn> ===== asynchronous gap =========================== dart:async _asyncThenWrapperHelper test/protocol_discovery_test.dart main.<fn>.<fn> test/src/context.dart 81:26 testUsingContext.<fn>.<fn> package:flutter_tools/src/base/context.dart 76:26 AppContext._run ===== asynchronous gap =========================== dart:async Future.Future.microtask package:flutter_tools/src/base/context.dart AppContext._run package:flutter_tools/src/base/context.dart 66:13 AppContext.runInZone.<fn> dart:async runZoned package:flutter_tools/src/base/context.dart 65:12 AppContext.runInZone test/src/context.dart 72:32 testUsingContext.<fn> ===== asynchronous gap =========================== dart:async Future.Future.microtask test/src/context.dart testUsingContext.<fn> package:test Declarer.test.<fn>.<fn> ```
non_main
service protocol discovery port forwarding default port is flaky on linux not sure if this test recently became flaky service protocol discovery port forwarding default port expected actual package test expect test protocol discovery test dart main test src context dart testusingcontext package flutter tools src base context dart appcontext run package flutter tools src base context dart appcontext runinzone dart async runzoned package flutter tools src base context dart appcontext runinzone test src context dart testusingcontext asynchronous gap dart async completer completeerror test protocol discovery test dart main asynchronous gap dart async asyncthenwrapperhelper test protocol discovery test dart main test src context dart testusingcontext package flutter tools src base context dart appcontext run asynchronous gap dart async future future microtask package flutter tools src base context dart appcontext run package flutter tools src base context dart appcontext runinzone dart async runzoned package flutter tools src base context dart appcontext runinzone test src context dart testusingcontext asynchronous gap dart async future future microtask test src context dart testusingcontext package test declarer test
0
3,441
13,211,584,094
IssuesEvent
2020-08-16 00:17:58
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Terraform: Add support for create destroy plans
affects_2.10 bot_closed cloud collection collection:community.general feature has_pr module needs_collection_redirect needs_maintainer support:community
<!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Terraform support `-destroy` flag in `plan` command. This should be supported by `planned` state. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> <!--- Paste example playbooks or commands between quotes below --> ```yaml - terraform: project_path: ./terraform state: planned state_file: tfplan destroy: true ``` <!--- HINT: You can also paste gist.github.com links for larger files -->
True
Terraform: Add support for create destroy plans - <!--- Verify first that your feature was not already discussed on GitHub --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY <!--- Describe the new feature/improvement briefly below --> Terraform support `-destroy` flag in `plan` command. This should be supported by `planned` state. ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME <!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure --> ##### ADDITIONAL INFORMATION <!--- Describe how the feature would be used, why it is needed and what it would solve --> <!--- Paste example playbooks or commands between quotes below --> ```yaml - terraform: project_path: ./terraform state: planned state_file: tfplan destroy: true ``` <!--- HINT: You can also paste gist.github.com links for larger files -->
main
terraform add support for create destroy plans summary terraform support destroy flag in plan command this should be supported by planned state issue type feature idea component name additional information yaml terraform project path terraform state planned state file tfplan destroy true
1
1,394
6,025,334,938
IssuesEvent
2017-06-08 08:25:47
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_iis_webbinding: ssl_flags property [FR]
affects_2.3 feature_idea waiting_on_maintainer windows
##### ISSUE TYPE Feature Idea ##### COMPONENT NAME win_iis_webbinding module ##### ANSIBLE VERSION N/A ##### SUMMARY Hi. `ssl_flags` property of web binding corresponds to SNI. More info on [stackoverflow](http://stackoverflow.com/questions/26485834/register-a-https-binding-with-sni-enabled-in-iis-8-with-powershell-script) It would be nice to be able to edit it. Thanks.
True
win_iis_webbinding: ssl_flags property [FR] - ##### ISSUE TYPE Feature Idea ##### COMPONENT NAME win_iis_webbinding module ##### ANSIBLE VERSION N/A ##### SUMMARY Hi. `ssl_flags` property of web binding corresponds to SNI. More info on [stackoverflow](http://stackoverflow.com/questions/26485834/register-a-https-binding-with-sni-enabled-in-iis-8-with-powershell-script) It would be nice to be able to edit it. Thanks.
main
win iis webbinding ssl flags property issue type feature idea component name win iis webbinding module ansible version n a summary hi ssl flags property of web binding corresponds to sni more info on it would be nice to be able to edit it thanks
1
68,636
21,773,226,244
IssuesEvent
2022-05-13 11:13:28
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
can't build on Slackware GNU/Linux: builds for Debian!
T-Defect X-Needs-Info
### Steps to reproduce yarn install yarn run fetch --importkey yarn run fetch --cfgdir "" yarn run build ### Outcome #### What did you expect? SlackBuild a txz. #### What happened instead? d@cosmos:~/math/src/network/chat/element-desktop-1.10.12$ yarn install yarn run fetch --importkey yarn run fetch --cfgdir "" yarn run build yarn install v1.22.10 [1/4] Resolving packages... [2/4] Fetching packages... info fsevents@2.3.2: The platform "linux" is incompatible with this module. info "fsevents@2.3.2" is an optional dependency and failed compatibility check. Excluding it from installation. info dmg-license@1.0.11: The platform "linux" is incompatible with this module. info "dmg-license@1.0.11" is an optional dependency and failed compatibility check. Excluding it from installation. info iconv-corefoundation@1.1.7: The platform "linux" is incompatible with this module. info "iconv-corefoundation@1.1.7" is an optional dependency and failed compatibility check. Excluding it from installation. [3/4] Linking dependencies... warning " > eslint-plugin-matrix-org@0.4.0" has unmet peer dependency "eslint-plugin-jsx-a11y@^6.5.1". warning " > ts-node@10.7.0" has unmet peer dependency "@types/node@*". [4/4] Building fresh packages... yarn run v1.22.10 $ yarn run mkdirs && node scripts/fetch-package.js --importkey $ mkdirp packages deploys Key imported! yarn run v1.22.10 $ yarn run mkdirs && node scripts/fetch-package.js --cfgdir '' $ mkdirp packages deploys Downloading https://github.com/vector-im/element-web/releases/download/v1.10.12/element-v1.10.12.tar.gz... Downloading https://github.com/vector-im/element-web/releases/download/v1.10.12/element-v1.10.12.tar.gz.asc... packages/element-v1.10.12.tar.gz downloaded and verified Skipping config file Pack deploys/element-v1.10.12 -> webapp.asar Done! yarn run v1.22.10 $ yarn run build:ts && yarn run build:res && electron-builder $ tsc $ node scripts/copy-res.js • electron-builder version=22.11.4 os=5.17.6 • loaded configuration file=package.json ("build" field) • writing effective config file=dist/builder-effective-config.yaml • packaging platform=linux arch=x64 electron=17.4.0 appOutDir=dist/linux-unpacked • building target=deb arch=x64 file=dist/element-desktop_1.10.12_amd64.deb ### Operating system Slackware64 15.1-current ### Application version element-desktop-1.10.12 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? Yes
1.0
can't build on Slackware GNU/Linux: builds for Debian! - ### Steps to reproduce yarn install yarn run fetch --importkey yarn run fetch --cfgdir "" yarn run build ### Outcome #### What did you expect? SlackBuild a txz. #### What happened instead? d@cosmos:~/math/src/network/chat/element-desktop-1.10.12$ yarn install yarn run fetch --importkey yarn run fetch --cfgdir "" yarn run build yarn install v1.22.10 [1/4] Resolving packages... [2/4] Fetching packages... info fsevents@2.3.2: The platform "linux" is incompatible with this module. info "fsevents@2.3.2" is an optional dependency and failed compatibility check. Excluding it from installation. info dmg-license@1.0.11: The platform "linux" is incompatible with this module. info "dmg-license@1.0.11" is an optional dependency and failed compatibility check. Excluding it from installation. info iconv-corefoundation@1.1.7: The platform "linux" is incompatible with this module. info "iconv-corefoundation@1.1.7" is an optional dependency and failed compatibility check. Excluding it from installation. [3/4] Linking dependencies... warning " > eslint-plugin-matrix-org@0.4.0" has unmet peer dependency "eslint-plugin-jsx-a11y@^6.5.1". warning " > ts-node@10.7.0" has unmet peer dependency "@types/node@*". [4/4] Building fresh packages... yarn run v1.22.10 $ yarn run mkdirs && node scripts/fetch-package.js --importkey $ mkdirp packages deploys Key imported! yarn run v1.22.10 $ yarn run mkdirs && node scripts/fetch-package.js --cfgdir '' $ mkdirp packages deploys Downloading https://github.com/vector-im/element-web/releases/download/v1.10.12/element-v1.10.12.tar.gz... Downloading https://github.com/vector-im/element-web/releases/download/v1.10.12/element-v1.10.12.tar.gz.asc... packages/element-v1.10.12.tar.gz downloaded and verified Skipping config file Pack deploys/element-v1.10.12 -> webapp.asar Done! yarn run v1.22.10 $ yarn run build:ts && yarn run build:res && electron-builder $ tsc $ node scripts/copy-res.js • electron-builder version=22.11.4 os=5.17.6 • loaded configuration file=package.json ("build" field) • writing effective config file=dist/builder-effective-config.yaml • packaging platform=linux arch=x64 electron=17.4.0 appOutDir=dist/linux-unpacked • building target=deb arch=x64 file=dist/element-desktop_1.10.12_amd64.deb ### Operating system Slackware64 15.1-current ### Application version element-desktop-1.10.12 ### How did you install the app? _No response_ ### Homeserver _No response_ ### Will you send logs? Yes
non_main
can t build on slackware gnu linux builds for debian steps to reproduce yarn install yarn run fetch importkey yarn run fetch cfgdir yarn run build outcome what did you expect slackbuild a txz what happened instead d cosmos math src network chat element desktop yarn install yarn run fetch importkey yarn run fetch cfgdir yarn run build yarn install resolving packages fetching packages info fsevents the platform linux is incompatible with this module info fsevents is an optional dependency and failed compatibility check excluding it from installation info dmg license the platform linux is incompatible with this module info dmg license is an optional dependency and failed compatibility check excluding it from installation info iconv corefoundation the platform linux is incompatible with this module info iconv corefoundation is an optional dependency and failed compatibility check excluding it from installation linking dependencies warning eslint plugin matrix org has unmet peer dependency eslint plugin jsx warning ts node has unmet peer dependency types node building fresh packages yarn run yarn run mkdirs node scripts fetch package js importkey mkdirp packages deploys key imported yarn run yarn run mkdirs node scripts fetch package js cfgdir mkdirp packages deploys downloading downloading packages element tar gz downloaded and verified skipping config file pack deploys element webapp asar done yarn run yarn run build ts yarn run build res electron builder tsc node scripts copy res js • electron builder version os • loaded configuration file package json build field • writing effective config file dist builder effective config yaml • packaging platform linux arch electron appoutdir dist linux unpacked • building target deb arch file dist element desktop deb operating system current application version element desktop how did you install the app no response homeserver no response will you send logs yes
0
200,295
15,097,686,561
IssuesEvent
2021-02-07 19:39:09
UCSD/campus-mobile
https://api.github.com/repos/UCSD/campus-mobile
opened
Add in-app build info
Needs Tester
Time Tracking: MA-95 Add in-app build info to assist in tracking down issues in test and live environments This is done when: - [x] App name, version, and build number are available in-app on the Profile tab - [ ] Testing PASS in QA - [ ] Testing PASS in PROD
1.0
Add in-app build info - Time Tracking: MA-95 Add in-app build info to assist in tracking down issues in test and live environments This is done when: - [x] App name, version, and build number are available in-app on the Profile tab - [ ] Testing PASS in QA - [ ] Testing PASS in PROD
non_main
add in app build info time tracking ma add in app build info to assist in tracking down issues in test and live environments this is done when app name version and build number are available in app on the profile tab testing pass in qa testing pass in prod
0
43,425
11,714,013,030
IssuesEvent
2020-03-09 11:27:13
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
Outdated documentation for AMBER TRJReader
Component-Docs Format-Amber defect
**Expected behavior** https://github.com/MDAnalysis/mdanalysis/blob/e12958ca0e6c5700563e236e21023c2b696489b7/package/MDAnalysis/coordinates/TRJ.py#L114-L124 https://github.com/MDAnalysis/mdanalysis/blob/e12958ca0e6c5700563e236e21023c2b696489b7/package/MDAnalysis/coordinates/TRJ.py#L208-L209 I cannot directly access frames of the trajectory by indexing. **Actual behavior** I can directly access frames of the trajectory by indexing. **Code to reproduce the behavior** Show us how to reproduce the failiure. If you can, use trajectory files from the test data. ``` pycon In [28]: import MDAnalysis as mda In [29]: from MDAnalysis.tests.datafiles import PRM, TRJ In [30]: trj = mda.Universe(PRM, TRJ) /Users/lily/pydev/mdanalysis/package/MDAnalysis/topology/TOPParser.py:270: UserWarning: ATOMIC_NUMBER record not found, guessing atom elements based on their atom types warnings.warn(msg) In [34]: trj.trajectory[2] Out[34]: < Timestep 2 with unit cell dimensions [0. 0. 0. 0. 0. 0.] > In [35]: t2 = trj.atoms.positions In [36]: trj.trajectory[5] Out[36]: < Timestep 5 with unit cell dimensions [0. 0. 0. 0. 0. 0.] > In [37]: t5 = trj.atoms.positions In [41]: np.array_equal(t2, t5) Out[41]: False ``` **Currently version of MDAnalysis** - Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.20.1 - Which version of Python (`python -V`)? 3.7.3 - Which operating system? MacOS Just checking that I'm understanding this limitation correctly and can leave it out of the user guide.
1.0
Outdated documentation for AMBER TRJReader - **Expected behavior** https://github.com/MDAnalysis/mdanalysis/blob/e12958ca0e6c5700563e236e21023c2b696489b7/package/MDAnalysis/coordinates/TRJ.py#L114-L124 https://github.com/MDAnalysis/mdanalysis/blob/e12958ca0e6c5700563e236e21023c2b696489b7/package/MDAnalysis/coordinates/TRJ.py#L208-L209 I cannot directly access frames of the trajectory by indexing. **Actual behavior** I can directly access frames of the trajectory by indexing. **Code to reproduce the behavior** Show us how to reproduce the failiure. If you can, use trajectory files from the test data. ``` pycon In [28]: import MDAnalysis as mda In [29]: from MDAnalysis.tests.datafiles import PRM, TRJ In [30]: trj = mda.Universe(PRM, TRJ) /Users/lily/pydev/mdanalysis/package/MDAnalysis/topology/TOPParser.py:270: UserWarning: ATOMIC_NUMBER record not found, guessing atom elements based on their atom types warnings.warn(msg) In [34]: trj.trajectory[2] Out[34]: < Timestep 2 with unit cell dimensions [0. 0. 0. 0. 0. 0.] > In [35]: t2 = trj.atoms.positions In [36]: trj.trajectory[5] Out[36]: < Timestep 5 with unit cell dimensions [0. 0. 0. 0. 0. 0.] > In [37]: t5 = trj.atoms.positions In [41]: np.array_equal(t2, t5) Out[41]: False ``` **Currently version of MDAnalysis** - Which version are you using? (run `python -c "import MDAnalysis as mda; print(mda.__version__)"`) 0.20.1 - Which version of Python (`python -V`)? 3.7.3 - Which operating system? MacOS Just checking that I'm understanding this limitation correctly and can leave it out of the user guide.
non_main
outdated documentation for amber trjreader expected behavior i cannot directly access frames of the trajectory by indexing actual behavior i can directly access frames of the trajectory by indexing code to reproduce the behavior show us how to reproduce the failiure if you can use trajectory files from the test data pycon in import mdanalysis as mda in from mdanalysis tests datafiles import prm trj in trj mda universe prm trj users lily pydev mdanalysis package mdanalysis topology topparser py userwarning atomic number record not found guessing atom elements based on their atom types warnings warn msg in trj trajectory out in trj atoms positions in trj trajectory out in trj atoms positions in np array equal out false currently version of mdanalysis which version are you using run python c import mdanalysis as mda print mda version which version of python python v which operating system macos just checking that i m understanding this limitation correctly and can leave it out of the user guide
0
143,590
5,520,609,949
IssuesEvent
2017-03-19 07:24:20
CS2103Jan2017-W13-B2/main
https://api.github.com/repos/CS2103Jan2017-W13-B2/main
closed
Add Mark command
priority.high type.epic
Add Mark command to set task status to done or undone Example: Mark 1 done Parameters: Mark INDEX_NUMBER STATUS
1.0
Add Mark command - Add Mark command to set task status to done or undone Example: Mark 1 done Parameters: Mark INDEX_NUMBER STATUS
non_main
add mark command add mark command to set task status to done or undone example mark done parameters mark index number status
0
824,807
31,224,492,486
IssuesEvent
2023-08-19 00:26:41
juno-fx/report
https://api.github.com/repos/juno-fx/report
opened
Mars: Specify Project Bucket Prefix
high priority
Each project is isolated by prefix on the Atlas bucket. This is governed by the IAM policy which will be defined in this ticket https://github.com/juno-fx/report/issues/146.
1.0
Mars: Specify Project Bucket Prefix - Each project is isolated by prefix on the Atlas bucket. This is governed by the IAM policy which will be defined in this ticket https://github.com/juno-fx/report/issues/146.
non_main
mars specify project bucket prefix each project is isolated by prefix on the atlas bucket this is governed by the iam policy which will be defined in this ticket
0
5,791
30,687,925,913
IssuesEvent
2023-07-26 13:30:34
precice/precice
https://api.github.com/repos/precice/precice
closed
Remove Broyden acceleration
maintainability good first issue breaking change
This acceleration type was added for the sole purpose of comparing it to the different quasi-Newton variants. There is no use case where the single-vector Broyden is superior to IQN-ILS or IQN-IMVJ. On the other hand, the configuration of the method is a bit ugly as all quasi-Newton parameters are required as `BroydenAcceleration` inherits from `BaseQNAcceleration`. Those parameters are not used, however. Therefore, I would suggest to simply remove the method.
True
Remove Broyden acceleration - This acceleration type was added for the sole purpose of comparing it to the different quasi-Newton variants. There is no use case where the single-vector Broyden is superior to IQN-ILS or IQN-IMVJ. On the other hand, the configuration of the method is a bit ugly as all quasi-Newton parameters are required as `BroydenAcceleration` inherits from `BaseQNAcceleration`. Those parameters are not used, however. Therefore, I would suggest to simply remove the method.
main
remove broyden acceleration this acceleration type was added for the sole purpose of comparing it to the different quasi newton variants there is no use case where the single vector broyden is superior to iqn ils or iqn imvj on the other hand the configuration of the method is a bit ugly as all quasi newton parameters are required as broydenacceleration inherits from baseqnacceleration those parameters are not used however therefore i would suggest to simply remove the method
1
258,434
27,564,016,056
IssuesEvent
2023-03-08 01:22:48
attesch/zencart
https://api.github.com/repos/attesch/zencart
opened
CVE-2017-20165 (High) detected in debug-2.3.3.tgz, debug-2.2.0.tgz
Mend: dependency security vulnerability
## CVE-2017-20165 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.3.3.tgz</b>, <b>debug-2.2.0.tgz</b></p></summary> <p> <details><summary><b>debug-2.3.3.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p> <p>Path to dependency file: /zencart/admin/includes/template/javascript/gridstack.js-master/package.json</p> <p>Path to vulnerable library: /admin/includes/template/javascript/gridstack.js-master/node_modules/socket.io-adapter/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - socket.io-1.7.3.tgz - :x: **debug-2.3.3.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: /zencart/admin/includes/template/javascript/gridstack.js-master/package.json</p> <p>Path to vulnerable library: /admin/includes/template/javascript/gridstack.js-master/node_modules/socket.io-parser/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - socket.io-1.7.3.tgz - socket.io-parser-2.3.1.tgz - :x: **debug-2.2.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability classified as problematic has been found in debug-js debug up to 3.0.x. This affects the function useColors of the file src/node.js. The manipulation of the argument str leads to inefficient regular expression complexity. Upgrading to version 3.1.0 is able to address this issue. The name of the patch is c38a0166c266a679c8de012d4eaccec3f944e685. It is recommended to upgrade the affected component. The identifier VDB-217665 was assigned to this vulnerability. <p>Publish Date: 2023-01-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-20165>CVE-2017-20165</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-9vvw-cc9w-f27h">https://github.com/advisories/GHSA-9vvw-cc9w-f27h</a></p> <p>Release Date: 2023-01-09</p> <p>Fix Resolution (debug): 2.6.9</p> <p>Direct dependency fix Resolution (karma): 2.0.0</p><p>Fix Resolution (debug): 2.6.9</p> <p>Direct dependency fix Resolution (karma): 2.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-20165 (High) detected in debug-2.3.3.tgz, debug-2.2.0.tgz - ## CVE-2017-20165 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>debug-2.3.3.tgz</b>, <b>debug-2.2.0.tgz</b></p></summary> <p> <details><summary><b>debug-2.3.3.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.3.3.tgz">https://registry.npmjs.org/debug/-/debug-2.3.3.tgz</a></p> <p>Path to dependency file: /zencart/admin/includes/template/javascript/gridstack.js-master/package.json</p> <p>Path to vulnerable library: /admin/includes/template/javascript/gridstack.js-master/node_modules/socket.io-adapter/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - socket.io-1.7.3.tgz - :x: **debug-2.3.3.tgz** (Vulnerable Library) </details> <details><summary><b>debug-2.2.0.tgz</b></p></summary> <p>small debugging utility</p> <p>Library home page: <a href="https://registry.npmjs.org/debug/-/debug-2.2.0.tgz">https://registry.npmjs.org/debug/-/debug-2.2.0.tgz</a></p> <p>Path to dependency file: /zencart/admin/includes/template/javascript/gridstack.js-master/package.json</p> <p>Path to vulnerable library: /admin/includes/template/javascript/gridstack.js-master/node_modules/socket.io-parser/node_modules/debug/package.json</p> <p> Dependency Hierarchy: - karma-1.7.1.tgz (Root Library) - socket.io-1.7.3.tgz - socket.io-parser-2.3.1.tgz - :x: **debug-2.2.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability classified as problematic has been found in debug-js debug up to 3.0.x. This affects the function useColors of the file src/node.js. The manipulation of the argument str leads to inefficient regular expression complexity. Upgrading to version 3.1.0 is able to address this issue. The name of the patch is c38a0166c266a679c8de012d4eaccec3f944e685. It is recommended to upgrade the affected component. The identifier VDB-217665 was assigned to this vulnerability. <p>Publish Date: 2023-01-09 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-20165>CVE-2017-20165</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-9vvw-cc9w-f27h">https://github.com/advisories/GHSA-9vvw-cc9w-f27h</a></p> <p>Release Date: 2023-01-09</p> <p>Fix Resolution (debug): 2.6.9</p> <p>Direct dependency fix Resolution (karma): 2.0.0</p><p>Fix Resolution (debug): 2.6.9</p> <p>Direct dependency fix Resolution (karma): 2.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in debug tgz debug tgz cve high severity vulnerability vulnerable libraries debug tgz debug tgz debug tgz small debugging utility library home page a href path to dependency file zencart admin includes template javascript gridstack js master package json path to vulnerable library admin includes template javascript gridstack js master node modules socket io adapter node modules debug package json dependency hierarchy karma tgz root library socket io tgz x debug tgz vulnerable library debug tgz small debugging utility library home page a href path to dependency file zencart admin includes template javascript gridstack js master package json path to vulnerable library admin includes template javascript gridstack js master node modules socket io parser node modules debug package json dependency hierarchy karma tgz root library socket io tgz socket io parser tgz x debug tgz vulnerable library vulnerability details a vulnerability classified as problematic has been found in debug js debug up to x this affects the function usecolors of the file src node js the manipulation of the argument str leads to inefficient regular expression complexity upgrading to version is able to address this issue the name of the patch is it is recommended to upgrade the affected component the identifier vdb was assigned to this vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution debug direct dependency fix resolution karma fix resolution debug direct dependency fix resolution karma step up your open source security game with mend
0
55,886
14,072,609,667
IssuesEvent
2020-11-04 02:20:12
junghanlee/juice-shop
https://api.github.com/repos/junghanlee/juice-shop
opened
CVE-2015-9235 (High) detected in jsonwebtoken-0.4.0.tgz
security vulnerability
## CVE-2015-9235 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-0.4.0.tgz</b></p></summary> <p>JSON Web Token implementation (symmetric and asymmetric)</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-0.4.0.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-0.4.0.tgz</a></p> <p>Path to dependency file: juice-shop/package.json</p> <p>Path to vulnerable library: juice-shop/node_modules/jsonwebtoken/package.json</p> <p> Dependency Hierarchy: - :x: **jsonwebtoken-0.4.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/junghanlee/juice-shop/commit/765fa79934265f7322f82b9e91a33ec80e92457d">765fa79934265f7322f82b9e91a33ec80e92457d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jsonwebtoken node module before 4.2.2 it is possible for an attacker to bypass verification when a token digitally signed with an asymmetric key (RS/ES family) of algorithms but instead the attacker send a token digitally signed with a symmetric algorithm (HS* family). <p>Publish Date: 2018-05-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9235>CVE-2015-9235</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/17">https://nodesecurity.io/advisories/17</a></p> <p>Release Date: 2015-04-01</p> <p>Fix Resolution: Update to version 4.2.2 or later.</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2015-9235 (High) detected in jsonwebtoken-0.4.0.tgz - ## CVE-2015-9235 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jsonwebtoken-0.4.0.tgz</b></p></summary> <p>JSON Web Token implementation (symmetric and asymmetric)</p> <p>Library home page: <a href="https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-0.4.0.tgz">https://registry.npmjs.org/jsonwebtoken/-/jsonwebtoken-0.4.0.tgz</a></p> <p>Path to dependency file: juice-shop/package.json</p> <p>Path to vulnerable library: juice-shop/node_modules/jsonwebtoken/package.json</p> <p> Dependency Hierarchy: - :x: **jsonwebtoken-0.4.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/junghanlee/juice-shop/commit/765fa79934265f7322f82b9e91a33ec80e92457d">765fa79934265f7322f82b9e91a33ec80e92457d</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jsonwebtoken node module before 4.2.2 it is possible for an attacker to bypass verification when a token digitally signed with an asymmetric key (RS/ES family) of algorithms but instead the attacker send a token digitally signed with a symmetric algorithm (HS* family). <p>Publish Date: 2018-05-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2015-9235>CVE-2015-9235</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/17">https://nodesecurity.io/advisories/17</a></p> <p>Release Date: 2015-04-01</p> <p>Fix Resolution: Update to version 4.2.2 or later.</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in jsonwebtoken tgz cve high severity vulnerability vulnerable library jsonwebtoken tgz json web token implementation symmetric and asymmetric library home page a href path to dependency file juice shop package json path to vulnerable library juice shop node modules jsonwebtoken package json dependency hierarchy x jsonwebtoken tgz vulnerable library found in head commit a href found in base branch master vulnerability details in jsonwebtoken node module before it is possible for an attacker to bypass verification when a token digitally signed with an asymmetric key rs es family of algorithms but instead the attacker send a token digitally signed with a symmetric algorithm hs family publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution update to version or later step up your open source security game with whitesource
0
35,946
9,691,016,116
IssuesEvent
2019-05-24 10:02:11
Lundalogik/lip
https://api.github.com/repos/Lundalogik/lip
opened
Support for Chromium
bug package builder
The package builder does not work with Chromium. This needs to be fixed since the desktop client will move completely to chromium within a few releases. Tricky situation: We want the package builder to work also for customers not using Chromium yet. Or can we say that once we fix support for Chromium we will release a new major version and customers using older DC will not be able to use the package builder?
1.0
Support for Chromium - The package builder does not work with Chromium. This needs to be fixed since the desktop client will move completely to chromium within a few releases. Tricky situation: We want the package builder to work also for customers not using Chromium yet. Or can we say that once we fix support for Chromium we will release a new major version and customers using older DC will not be able to use the package builder?
non_main
support for chromium the package builder does not work with chromium this needs to be fixed since the desktop client will move completely to chromium within a few releases tricky situation we want the package builder to work also for customers not using chromium yet or can we say that once we fix support for chromium we will release a new major version and customers using older dc will not be able to use the package builder
0
179,987
13,914,789,203
IssuesEvent
2020-10-20 22:56:06
daniel-norris/neu_ui
https://api.github.com/repos/daniel-norris/neu_ui
closed
Improve Typography component test to 100% coverage
hacktoberfest tests
**Is your feature request related to a problem? Please describe.** Test coverage across the application is low. We need to build confidence that the components have the expected behaviour that we want and to help mitigate any regression in the future. **Describe the solution you'd like** We need to implement better test coverage across the library. Ideally each component should be accompanied by a test case written using Jest. We need to improve the test coverage for the Typography component, currently implemented using Jest, to 100% coverage of the component. It currently stands at ~45%. More info on Jest can be found here (https://jestjs.io/). ![image](https://user-images.githubusercontent.com/58988735/96526095-8f9efe00-1274-11eb-97ab-d1c61d377053.png) For examples of how this is done, take a look at existing test cases in the library. An example would be the CardHeader or Input components. You can run your tests using npm run test or to see test coverage across the library npm run test:cov. This is part of epic #19.
1.0
Improve Typography component test to 100% coverage - **Is your feature request related to a problem? Please describe.** Test coverage across the application is low. We need to build confidence that the components have the expected behaviour that we want and to help mitigate any regression in the future. **Describe the solution you'd like** We need to implement better test coverage across the library. Ideally each component should be accompanied by a test case written using Jest. We need to improve the test coverage for the Typography component, currently implemented using Jest, to 100% coverage of the component. It currently stands at ~45%. More info on Jest can be found here (https://jestjs.io/). ![image](https://user-images.githubusercontent.com/58988735/96526095-8f9efe00-1274-11eb-97ab-d1c61d377053.png) For examples of how this is done, take a look at existing test cases in the library. An example would be the CardHeader or Input components. You can run your tests using npm run test or to see test coverage across the library npm run test:cov. This is part of epic #19.
non_main
improve typography component test to coverage is your feature request related to a problem please describe test coverage across the application is low we need to build confidence that the components have the expected behaviour that we want and to help mitigate any regression in the future describe the solution you d like we need to implement better test coverage across the library ideally each component should be accompanied by a test case written using jest we need to improve the test coverage for the typography component currently implemented using jest to coverage of the component it currently stands at more info on jest can be found here for examples of how this is done take a look at existing test cases in the library an example would be the cardheader or input components you can run your tests using npm run test or to see test coverage across the library npm run test cov this is part of epic
0
435,463
30,501,973,725
IssuesEvent
2023-07-18 14:28:02
usnistgov/dioptra
https://api.github.com/repos/usnistgov/dioptra
opened
Add Release instructions for merging dev to main into a RELEASE.md file
documentation
## Definition of Done - [ ] Instructions for what to do to update the main branch from dev are written up in Markdown format and stored in a RELEASE.md file
1.0
Add Release instructions for merging dev to main into a RELEASE.md file - ## Definition of Done - [ ] Instructions for what to do to update the main branch from dev are written up in Markdown format and stored in a RELEASE.md file
non_main
add release instructions for merging dev to main into a release md file definition of done instructions for what to do to update the main branch from dev are written up in markdown format and stored in a release md file
0
156,331
13,647,041,039
IssuesEvent
2020-09-26 01:14:07
sflynlang/sflynlang-bot
https://api.github.com/repos/sflynlang/sflynlang-bot
opened
Avatar Preview [Command]
documentation enhancement good first issue hacktoberfest
If you knows only the basic about Discord bots, you can contribute doing a command to view your avatar. The only thing thaat is required: Make it in a Message Embed. ## Resources * [Message Embed (Discord.js Guide)](https://discordjs.guide/popular-topics/embeds.html) * [Message Embed (Discord.js Official Guide)](https://discord.js.org/#/docs/main/stable/examples/embed) * [Avatars Command (Discord.js Official Guide)](https://discord.js.org/#/docs/main/stable/examples/avatars) ## Information * Username * Avatar Preview
1.0
Avatar Preview [Command] - If you knows only the basic about Discord bots, you can contribute doing a command to view your avatar. The only thing thaat is required: Make it in a Message Embed. ## Resources * [Message Embed (Discord.js Guide)](https://discordjs.guide/popular-topics/embeds.html) * [Message Embed (Discord.js Official Guide)](https://discord.js.org/#/docs/main/stable/examples/embed) * [Avatars Command (Discord.js Official Guide)](https://discord.js.org/#/docs/main/stable/examples/avatars) ## Information * Username * Avatar Preview
non_main
avatar preview if you knows only the basic about discord bots you can contribute doing a command to view your avatar the only thing thaat is required make it in a message embed resources information username avatar preview
0
12,123
7,785,009,441
IssuesEvent
2018-06-06 14:46:01
LiskHQ/lisk-hub
https://api.github.com/repos/LiskHQ/lisk-hub
closed
Cache delegate names lookup in localStorage
*medium performance
To be used in - vote launch protocol - vote transaction detail - deleagte detail
True
Cache delegate names lookup in localStorage - To be used in - vote launch protocol - vote transaction detail - deleagte detail
non_main
cache delegate names lookup in localstorage to be used in vote launch protocol vote transaction detail deleagte detail
0
2,932
10,512,201,922
IssuesEvent
2019-09-27 17:17:46
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
closed
Add API function to create VSIB operands
Maintainability OpSys-x64 OpSys-x86
Feature request to implement a function that creates a VSIB addressing memory operand. Something like` opnd_create_vsib(reg_id_t base_reg, reg_id_t vector_reg, int scale, int disp)` would be great.
True
Add API function to create VSIB operands - Feature request to implement a function that creates a VSIB addressing memory operand. Something like` opnd_create_vsib(reg_id_t base_reg, reg_id_t vector_reg, int scale, int disp)` would be great.
main
add api function to create vsib operands feature request to implement a function that creates a vsib addressing memory operand something like opnd create vsib reg id t base reg reg id t vector reg int scale int disp would be great
1
1,529
6,572,221,390
IssuesEvent
2017-09-11 00:13:54
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
lvg module does not allow a pesize < 2 MiB
affects_2.0 bug_report waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> lvg module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT Ansible host: Ubuntu 15.10 Remote host: Ubuntu 14.04 ##### SUMMARY The lvg module has a `pesize` parameter which maps to `vgcreate`'s `--physicalextentsize` parameter. The man page for `vgcreate` says that the value for this parameter... ``` must be a power of 2 of at least 1 sector (where the sector size is the largest sector size of the PVs currently used in the VG) or, if not a power of 2, at least 128KiB. [...] The smallest PE is 1KiB ``` However, the lvg module's implementation of this parameter is far more limited: ``` The size of the physical extent in megabytes. Must be a power of 2. ``` The practical upshot of this is that you can't use the lvg module to create a volume group if you want the volume group's physical extent size to be less than 2 MiB. There are many use cases for smaller extents. A common one might be a VG which will contain a thin-provisioned LV. If your data consists mainly of smaller files, extents smaller than the default will be more space-efficient when dealing with thin LVs created from LV snapshots, for example. But the core of this bug is the inflexibility of the `pesize` parameter as it stands. Ideally, the parameter should either accept the same unit suffixes as `vgcreate` does, or be eliminated altogether. (It's curious that among the numerous parameters, `--physicalextentsize` was singled out for exposure as an lvg module parameter.) It is also unfortunately the case that `pesize` cannot be overridden by `vg_options`. Attempting to do so results in the following error: ``` fatal: [gfs-02.eng.example.net]: FAILED! => {"changed": false, "err": " Option -s/--physicalextentsize may not be repeated.\n Error during parsing of command line.\n", "failed": true, "msg": "Creating volume group 'data-vg' failed", "rc": 3} ``` I have omitted the "steps to reproduce" (etc) sections because this is more of a design issue than clear-cut "this thingy is broken" bug. So in summary, any _one_ of these would fix this bug: - The `pesize` parameter behave the same as described in the `vgcreate` man page - Allow `pesize` to be overridden by `--physicalextentsize` in `vg_options` - Removal of the `pesize` parameter from the lvg module
True
lvg module does not allow a pesize < 2 MiB - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> lvg module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT Ansible host: Ubuntu 15.10 Remote host: Ubuntu 14.04 ##### SUMMARY The lvg module has a `pesize` parameter which maps to `vgcreate`'s `--physicalextentsize` parameter. The man page for `vgcreate` says that the value for this parameter... ``` must be a power of 2 of at least 1 sector (where the sector size is the largest sector size of the PVs currently used in the VG) or, if not a power of 2, at least 128KiB. [...] The smallest PE is 1KiB ``` However, the lvg module's implementation of this parameter is far more limited: ``` The size of the physical extent in megabytes. Must be a power of 2. ``` The practical upshot of this is that you can't use the lvg module to create a volume group if you want the volume group's physical extent size to be less than 2 MiB. There are many use cases for smaller extents. A common one might be a VG which will contain a thin-provisioned LV. If your data consists mainly of smaller files, extents smaller than the default will be more space-efficient when dealing with thin LVs created from LV snapshots, for example. But the core of this bug is the inflexibility of the `pesize` parameter as it stands. Ideally, the parameter should either accept the same unit suffixes as `vgcreate` does, or be eliminated altogether. (It's curious that among the numerous parameters, `--physicalextentsize` was singled out for exposure as an lvg module parameter.) It is also unfortunately the case that `pesize` cannot be overridden by `vg_options`. Attempting to do so results in the following error: ``` fatal: [gfs-02.eng.example.net]: FAILED! => {"changed": false, "err": " Option -s/--physicalextentsize may not be repeated.\n Error during parsing of command line.\n", "failed": true, "msg": "Creating volume group 'data-vg' failed", "rc": 3} ``` I have omitted the "steps to reproduce" (etc) sections because this is more of a design issue than clear-cut "this thingy is broken" bug. So in summary, any _one_ of these would fix this bug: - The `pesize` parameter behave the same as described in the `vgcreate` man page - Allow `pesize` to be overridden by `--physicalextentsize` in `vg_options` - Removal of the `pesize` parameter from the lvg module
main
lvg module does not allow a pesize issue type bug report component name lvg module ansible version ansible config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment ansible host ubuntu remote host ubuntu summary the lvg module has a pesize parameter which maps to vgcreate s physicalextentsize parameter the man page for vgcreate says that the value for this parameter must be a power of of at least sector where the sector size is the largest sector size of the pvs currently used in the vg or if not a power of at least the smallest pe is however the lvg module s implementation of this parameter is far more limited the size of the physical extent in megabytes must be a power of the practical upshot of this is that you can t use the lvg module to create a volume group if you want the volume group s physical extent size to be less than mib there are many use cases for smaller extents a common one might be a vg which will contain a thin provisioned lv if your data consists mainly of smaller files extents smaller than the default will be more space efficient when dealing with thin lvs created from lv snapshots for example but the core of this bug is the inflexibility of the pesize parameter as it stands ideally the parameter should either accept the same unit suffixes as vgcreate does or be eliminated altogether it s curious that among the numerous parameters physicalextentsize was singled out for exposure as an lvg module parameter it is also unfortunately the case that pesize cannot be overridden by vg options attempting to do so results in the following error fatal failed changed false err option s physicalextentsize may not be repeated n error during parsing of command line n failed true msg creating volume group data vg failed rc i have omitted the steps to reproduce etc sections because this is more of a design issue than clear cut this thingy is broken bug so in summary any one of these would fix this bug the pesize parameter behave the same as described in the vgcreate man page allow pesize to be overridden by physicalextentsize in vg options removal of the pesize parameter from the lvg module
1
1,634
6,572,657,916
IssuesEvent
2017-09-11 04:08:54
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Fetch Silently Fails To Fetch File
affects_2.0 bug_report P2 waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME fetch module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Running from Amazon Linux 2016.03 release Target is Windows Server 2012 R2 (recent patches applied) ##### SUMMARY Fetching an 80mbyte file from Windows Server silently fails. Ansible does not report an error, but the file never arrives. (Fetching the same file from a remote Linux server works fine, btw) ##### STEPS TO REPRODUCE ``` - name: fetch deliverable fetch: src="{{ topleveldir }}\ult_{{ hostvars['localhost']['version'] }}_{{ hostvars['localhost']['soft'] }}_{{ hostvars['localhost']['target'] }}.zip" dest=/home/ec2-user/files flat=yes fail_on_missing=yes - name: Output the name of the archive debug: msg="archive - ult_{{ hostvars['localhost']['version'] }}_{{ hostvars['localhost']['soft'] }}_{{ hostvars['localhost']['target'] }}.zip" ``` ##### EXPECTED RESULTS I expect the file to be placed on the control machine. ##### ACTUAL RESULTS The file never arrives but the results don't indicate a problem (afaict). ``` TASK [fetch deliverable] ******************************************************* task path: /build_scripts/build_noc_cust_win.yml:204 <52.36.54.12> ESTABLISH WINRM CONNECTION FOR USER: Administrator on PORT 5986 TO 52.36.54.12 <52.36.54.12> WINRM CONNECT: transport=ssl endpoint=https://52.36.54.12:5986/wsman <52.36.54.12> EXEC Set-StrictMode -Version Latest If (Test-Path -PathType Leaf "C:\apps\ult_master_noc_win2012.zip") { $sp = new-object -TypeName System.Security.Cryptography.SHA1CryptoServiceProvider; $fp = [System.IO.File]::Open("C:\apps\ult_master_noc_win2012.zip", [System.IO.Filemode]::Open, [System.IO.FileAccess]::Read); [System.BitConverter]::ToString($sp.ComputeHash($fp)).Replace("-", "").ToLower(); $fp.Dispose(); } ElseIf (Test-Path -PathType Container "C:\apps\ult_master_noc_win2012.zip") { Write-Host "3"; } Else { Write-Host "1"; } <52.36.54.12> WINRM OPEN SHELL: E5D85769-78F7-40DF-9A5F-B98304A49455 <52.36.54.12> WINRM EXEC u'PowerShell' [u'-NoProfile', u'-NonInteractive', u'-ExecutionPolicy', u'Unrestricted', u'-EncodedCommand', u'UwBlAHQALQBTAHQAcgBpAGMAdABNAG8AZABlACAALQBWAGUAcgBzAGkAbwBuACAATABhAHQAZQBzAHQACgBJAGYAIAAoAFQAZQBzAHQALQBQAGEAdABoACAALQBQAGEAdABoAFQAeQBwAGUAIABMAGUAYQBmACAAIgBDADoAXABjAGEAcABzAGUAbgB0AGEAXAB1AGwAdAByAGEAdwByAGEAcABfAG0AYQBzAHQAZQByAF8AbgBvAGUAdABsAF8AdwBpAG4AMgAwADEAMgAuAHoAaQBwACIAKQAKAHsACgAkAHMAcAAgAD0AIABuAGUAdwAtAG8AYgBqAGUAYwB0ACAALQBUAHkAcABlAE4AYQBtAGUAIABTAHkAcwB0AGUAbQAuAFMAZQBjAHUAcgBpAHQAeQAuAEMAcgB5AHAAdABvAGcAcgBhAHAAaAB5AC4AUwBIAEEAMQBDAHIAeQBwAHQAbwBTAGUAcgB2AGkAYwBlAFAAcgBvAHYAaQBkAGUAcgA7AAoAJABmAHAAIAA9ACAAWwBTAHkAcwB0AGUAbQAuAEkATwAuAEYAaQBsAGUAXQA6dABhAFwAdQBsAHQAcgBhAHcAcgBhAHAAXwBtAGEAcwB0AGUAcgBfAG4AbwBlAHQAbABfAHcAaQBuADIAMAAxADIALgB6AGkAcAAiACwAIABbAFMAeQBzAHQAZQBtAC4ASQBPAC4ARgBpAGwAZQBtAG8AZABlAF0AOgA6AE8AcABlAG4ALAAgAFsAUwB5AHMAdABlAG0ALgBJAE8ALgBGAGkAbABlAEEAYwBjAGUAcwBzAF0AOgA6AFIAZQBhAGQAKQA7AAoAWwBTAHkAcwB0AGUAbQAuAEIAaQB0AEMAbwBuAHYAZQByAHQAZQByAF0AOgA6AFQAbwBTAHQAcgBpAG4AZwAoACQAcwBwAC4AQwBvAG0AcAB1AHQAZQBIAGEAcwBoACgAJABmAHAAKQApAC4AUgBlAHAAbABhAGMAZQAoACIALQAiACwAIAAiACIAKQAuAFQAbwBMAG8AdwBlAHIAKAApADsACgAkAGYAcAAuAEQAaQBzAHAAbwBzAGUAKAApADsACgB9AAoARQBsAAAXwBtAGEAcwB0AGUAcgBfAG4AbwBlAHQAbABfAHcAaQBuADIAMAAxADIALgB6AGkAcAAiACkACgB7AAoAVwByAGkAdABlAC0ASABvAHMAdAAgACIAMwAiADsACgB9AAoARQBsAHMAZQAKAHsACgBXAHIAaQB0AGUALQBIAG8AcwB0ACAAIgAxACIAOwAKAH0A'] <52.36.54.12> WINRM RESULT u'<Response code 0, out "22d5bc624a1e6d0ce116", err "">' <52.36.54.12> WINRM STDOUT 22d5bc624a1e6d0ce1162a60ab04f49c70ac3dc8 <52.36.54.12> WINRM STDERR <52.36.54.12> FETCH "C:\apps\ult_master_noc_win2012.zip" TO "/home/ec2-user/files" <52.36.54.12> WINRM FETCH "C:\apps\ult_master_noc_win2012.zip" to "/home/ec2-user/files" (offset=0) <52.36.54.12> WINRM EXEC 'PowerShell' ['-NoProfile', '-NonInteractive', '-ExecutionPolicy', 'Unrestricted', '-EncodedCommand', 'UwBlAHQALQBTAHQAcgBpAGMAdABNAG8AZABlACAALQBWAGUAcgBzAGkAbwBuACAATABhAHQAZQBzAHQACgBJAGYAIAAoAFQAZQBzAHQALQBQAGEAdABoACAALQBQAGEAdABoAFQAeQBwAGUAIABMAGUAYQBmACAAIgBDADoAXABjAGEAcABzAGUAbgB0AGEAXAB1AGwAdAByAGEAdwByAGEAcABfAG0AYQBzAHQAZQByAFBuADIAMAAxADIALgB6AGkAcAAiACkAOwAKACQAcwB0AHIAZQBhAG0ALgBTAGUAZQBrACgAMAAsACAAWwBTAHkAcwB0AGUAbQAuAEkATwAuAFMAZQBlAGsATwByAGkAZwBpAG4AXQA6ADoAQgBlAGcAaQBuACkAIAB8ACAATwB1AHQALQBOAHUAbABsADsACgAkAGIAdQBmAGYAZQByACAAPQAgAE4AZQB3AC0ATwBiAGoAZQBjAHQAIABCAHkAdABlAFsAXQAgADUAMgA0ADIAOAA4ADsACgAkAGIAeQB0AGUAcwBSAGUAYQBkACAAPQAgACQAcwB0AHIAZQBhAG0ALgBSAGUAYQBkACgAJABiAHUAZgBmAGUAcgAsACAAMAAsACAANQAyADQAMgA4ADgAKQA7AAoAJABiAHkAdABlAHMAIAA9ACAAJABiAHUAZgBmAGUAcgBbADAALgAuACgAJABiAHkAdABlAHMAUgBlAGEAZAAtADEAKQBdADsACgBbAFMAeQBzAHQAZQBtAC4AQwBvAG4AdgBlAHIAdABdADoAOgBUAG8AQgBhAHMAZQA2ADQAUwB0AHIAaQBuAGcAKAAkAGIAeQB0AGUAcwApADsACgAkAHMAdAByAGUAYQBtAC4AMgAuAHoAaQBwACIAKQAKAHsACgBXAHIAaQB0AGUALQBIAG8AcwB0ACAAIgBbAEQASQBSAF0AIgA7AAoAfQAKAEUAbABzAGUACgB7AAoAVwByAGkAdABlAC0ARQByAHIAbwByACAAIgBDADoAXABjAGEAcABzAGUAbgB0AGEAXAB1AGwAdAByAGEAdwByAGEAcABfAG0AYQBzAHQAZQByAF8AbgBvAGUAdABsAF8AdwBpAG4AMgAwADEAMgAuAHoAaQBwACAAZABvAGUAcwAgAG4AbwB0ACAAZQB4AGkAcwB0ACIAOwAKAEUAeABpAHQAIAAxADsACgB9AA=='] <52.36.54.12> WINRM RESULT u'<Response code 0, out "UEsDBBQAAAAAALsogkgA", err "">' <52.36.54.12> WINRM STDOUT [Long encoded part deleted] <52.36.54.12> WINRM STDERR <52.36.54.12> WINRM CLOSE SHELL: E5D85769-78F7-40DF-9A5F-B98304A49455 changed: [52.36.10.20] => {"changed": true, "checksum": null, "dest": "/home/ec2-user/files", "invocation": {"module_args": {"dest": "/home/ec2-user/files", "fail_on_missing": "yes", "flat": "yes", "src": "C:\\apps\\ult_master_noc_win2012.zip"}, "module_name": "fetch"}, "md5sum": null, "remote_checksum": "22d5bc624a1e6d0ce1162a60ab04f49c70ac3dc8", "remote_md5sum": null} TASK [Output the name of the archive] ****************************************** task path: /build_scripts/build_noetl_cust_win.yml:208 ok: [52.36.10.20] => { "msg": "archive - ult_master_noc_win2012.zip" } ```
True
Fetch Silently Fails To Fetch File - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME fetch module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION ##### OS / ENVIRONMENT Running from Amazon Linux 2016.03 release Target is Windows Server 2012 R2 (recent patches applied) ##### SUMMARY Fetching an 80mbyte file from Windows Server silently fails. Ansible does not report an error, but the file never arrives. (Fetching the same file from a remote Linux server works fine, btw) ##### STEPS TO REPRODUCE ``` - name: fetch deliverable fetch: src="{{ topleveldir }}\ult_{{ hostvars['localhost']['version'] }}_{{ hostvars['localhost']['soft'] }}_{{ hostvars['localhost']['target'] }}.zip" dest=/home/ec2-user/files flat=yes fail_on_missing=yes - name: Output the name of the archive debug: msg="archive - ult_{{ hostvars['localhost']['version'] }}_{{ hostvars['localhost']['soft'] }}_{{ hostvars['localhost']['target'] }}.zip" ``` ##### EXPECTED RESULTS I expect the file to be placed on the control machine. ##### ACTUAL RESULTS The file never arrives but the results don't indicate a problem (afaict). ``` TASK [fetch deliverable] ******************************************************* task path: /build_scripts/build_noc_cust_win.yml:204 <52.36.54.12> ESTABLISH WINRM CONNECTION FOR USER: Administrator on PORT 5986 TO 52.36.54.12 <52.36.54.12> WINRM CONNECT: transport=ssl endpoint=https://52.36.54.12:5986/wsman <52.36.54.12> EXEC Set-StrictMode -Version Latest If (Test-Path -PathType Leaf "C:\apps\ult_master_noc_win2012.zip") { $sp = new-object -TypeName System.Security.Cryptography.SHA1CryptoServiceProvider; $fp = [System.IO.File]::Open("C:\apps\ult_master_noc_win2012.zip", [System.IO.Filemode]::Open, [System.IO.FileAccess]::Read); [System.BitConverter]::ToString($sp.ComputeHash($fp)).Replace("-", "").ToLower(); $fp.Dispose(); } ElseIf (Test-Path -PathType Container "C:\apps\ult_master_noc_win2012.zip") { Write-Host "3"; } Else { Write-Host "1"; } <52.36.54.12> WINRM OPEN SHELL: E5D85769-78F7-40DF-9A5F-B98304A49455 <52.36.54.12> WINRM EXEC u'PowerShell' [u'-NoProfile', u'-NonInteractive', u'-ExecutionPolicy', u'Unrestricted', u'-EncodedCommand', u'UwBlAHQALQBTAHQAcgBpAGMAdABNAG8AZABlACAALQBWAGUAcgBzAGkAbwBuACAATABhAHQAZQBzAHQACgBJAGYAIAAoAFQAZQBzAHQALQBQAGEAdABoACAALQBQAGEAdABoAFQAeQBwAGUAIABMAGUAYQBmACAAIgBDADoAXABjAGEAcABzAGUAbgB0AGEAXAB1AGwAdAByAGEAdwByAGEAcABfAG0AYQBzAHQAZQByAF8AbgBvAGUAdABsAF8AdwBpAG4AMgAwADEAMgAuAHoAaQBwACIAKQAKAHsACgAkAHMAcAAgAD0AIABuAGUAdwAtAG8AYgBqAGUAYwB0ACAALQBUAHkAcABlAE4AYQBtAGUAIABTAHkAcwB0AGUAbQAuAFMAZQBjAHUAcgBpAHQAeQAuAEMAcgB5AHAAdABvAGcAcgBhAHAAaAB5AC4AUwBIAEEAMQBDAHIAeQBwAHQAbwBTAGUAcgB2AGkAYwBlAFAAcgBvAHYAaQBkAGUAcgA7AAoAJABmAHAAIAA9ACAAWwBTAHkAcwB0AGUAbQAuAEkATwAuAEYAaQBsAGUAXQA6dABhAFwAdQBsAHQAcgBhAHcAcgBhAHAAXwBtAGEAcwB0AGUAcgBfAG4AbwBlAHQAbABfAHcAaQBuADIAMAAxADIALgB6AGkAcAAiACwAIABbAFMAeQBzAHQAZQBtAC4ASQBPAC4ARgBpAGwAZQBtAG8AZABlAF0AOgA6AE8AcABlAG4ALAAgAFsAUwB5AHMAdABlAG0ALgBJAE8ALgBGAGkAbABlAEEAYwBjAGUAcwBzAF0AOgA6AFIAZQBhAGQAKQA7AAoAWwBTAHkAcwB0AGUAbQAuAEIAaQB0AEMAbwBuAHYAZQByAHQAZQByAF0AOgA6AFQAbwBTAHQAcgBpAG4AZwAoACQAcwBwAC4AQwBvAG0AcAB1AHQAZQBIAGEAcwBoACgAJABmAHAAKQApAC4AUgBlAHAAbABhAGMAZQAoACIALQAiACwAIAAiACIAKQAuAFQAbwBMAG8AdwBlAHIAKAApADsACgAkAGYAcAAuAEQAaQBzAHAAbwBzAGUAKAApADsACgB9AAoARQBsAAAXwBtAGEAcwB0AGUAcgBfAG4AbwBlAHQAbABfAHcAaQBuADIAMAAxADIALgB6AGkAcAAiACkACgB7AAoAVwByAGkAdABlAC0ASABvAHMAdAAgACIAMwAiADsACgB9AAoARQBsAHMAZQAKAHsACgBXAHIAaQB0AGUALQBIAG8AcwB0ACAAIgAxACIAOwAKAH0A'] <52.36.54.12> WINRM RESULT u'<Response code 0, out "22d5bc624a1e6d0ce116", err "">' <52.36.54.12> WINRM STDOUT 22d5bc624a1e6d0ce1162a60ab04f49c70ac3dc8 <52.36.54.12> WINRM STDERR <52.36.54.12> FETCH "C:\apps\ult_master_noc_win2012.zip" TO "/home/ec2-user/files" <52.36.54.12> WINRM FETCH "C:\apps\ult_master_noc_win2012.zip" to "/home/ec2-user/files" (offset=0) <52.36.54.12> WINRM EXEC 'PowerShell' ['-NoProfile', '-NonInteractive', '-ExecutionPolicy', 'Unrestricted', '-EncodedCommand', 'UwBlAHQALQBTAHQAcgBpAGMAdABNAG8AZABlACAALQBWAGUAcgBzAGkAbwBuACAATABhAHQAZQBzAHQACgBJAGYAIAAoAFQAZQBzAHQALQBQAGEAdABoACAALQBQAGEAdABoAFQAeQBwAGUAIABMAGUAYQBmACAAIgBDADoAXABjAGEAcABzAGUAbgB0AGEAXAB1AGwAdAByAGEAdwByAGEAcABfAG0AYQBzAHQAZQByAFBuADIAMAAxADIALgB6AGkAcAAiACkAOwAKACQAcwB0AHIAZQBhAG0ALgBTAGUAZQBrACgAMAAsACAAWwBTAHkAcwB0AGUAbQAuAEkATwAuAFMAZQBlAGsATwByAGkAZwBpAG4AXQA6ADoAQgBlAGcAaQBuACkAIAB8ACAATwB1AHQALQBOAHUAbABsADsACgAkAGIAdQBmAGYAZQByACAAPQAgAE4AZQB3AC0ATwBiAGoAZQBjAHQAIABCAHkAdABlAFsAXQAgADUAMgA0ADIAOAA4ADsACgAkAGIAeQB0AGUAcwBSAGUAYQBkACAAPQAgACQAcwB0AHIAZQBhAG0ALgBSAGUAYQBkACgAJABiAHUAZgBmAGUAcgAsACAAMAAsACAANQAyADQAMgA4ADgAKQA7AAoAJABiAHkAdABlAHMAIAA9ACAAJABiAHUAZgBmAGUAcgBbADAALgAuACgAJABiAHkAdABlAHMAUgBlAGEAZAAtADEAKQBdADsACgBbAFMAeQBzAHQAZQBtAC4AQwBvAG4AdgBlAHIAdABdADoAOgBUAG8AQgBhAHMAZQA2ADQAUwB0AHIAaQBuAGcAKAAkAGIAeQB0AGUAcwApADsACgAkAHMAdAByAGUAYQBtAC4AMgAuAHoAaQBwACIAKQAKAHsACgBXAHIAaQB0AGUALQBIAG8AcwB0ACAAIgBbAEQASQBSAF0AIgA7AAoAfQAKAEUAbABzAGUACgB7AAoAVwByAGkAdABlAC0ARQByAHIAbwByACAAIgBDADoAXABjAGEAcABzAGUAbgB0AGEAXAB1AGwAdAByAGEAdwByAGEAcABfAG0AYQBzAHQAZQByAF8AbgBvAGUAdABsAF8AdwBpAG4AMgAwADEAMgAuAHoAaQBwACAAZABvAGUAcwAgAG4AbwB0ACAAZQB4AGkAcwB0ACIAOwAKAEUAeABpAHQAIAAxADsACgB9AA=='] <52.36.54.12> WINRM RESULT u'<Response code 0, out "UEsDBBQAAAAAALsogkgA", err "">' <52.36.54.12> WINRM STDOUT [Long encoded part deleted] <52.36.54.12> WINRM STDERR <52.36.54.12> WINRM CLOSE SHELL: E5D85769-78F7-40DF-9A5F-B98304A49455 changed: [52.36.10.20] => {"changed": true, "checksum": null, "dest": "/home/ec2-user/files", "invocation": {"module_args": {"dest": "/home/ec2-user/files", "fail_on_missing": "yes", "flat": "yes", "src": "C:\\apps\\ult_master_noc_win2012.zip"}, "module_name": "fetch"}, "md5sum": null, "remote_checksum": "22d5bc624a1e6d0ce1162a60ab04f49c70ac3dc8", "remote_md5sum": null} TASK [Output the name of the archive] ****************************************** task path: /build_scripts/build_noetl_cust_win.yml:208 ok: [52.36.10.20] => { "msg": "archive - ult_master_noc_win2012.zip" } ```
main
fetch silently fails to fetch file issue type bug report component name fetch module ansible version ansible config file configured module search path default w o overrides configuration os environment running from amazon linux release target is windows server recent patches applied summary fetching an file from windows server silently fails ansible does not report an error but the file never arrives fetching the same file from a remote linux server works fine btw steps to reproduce name fetch deliverable fetch src topleveldir ult hostvars hostvars hostvars zip dest home user files flat yes fail on missing yes name output the name of the archive debug msg archive ult hostvars hostvars hostvars zip expected results i expect the file to be placed on the control machine actual results the file never arrives but the results don t indicate a problem afaict task task path build scripts build noc cust win yml establish winrm connection for user administrator on port to winrm connect transport ssl endpoint exec set strictmode version latest if test path pathtype leaf c apps ult master noc zip sp new object typename system security cryptography fp open c apps ult master noc zip open read tostring sp computehash fp replace tolower fp dispose elseif test path pathtype container c apps ult master noc zip write host else write host winrm open shell winrm exec u powershell winrm result u winrm stdout winrm stderr fetch c apps ult master noc zip to home user files winrm fetch c apps ult master noc zip to home user files offset winrm exec powershell winrm result u winrm stdout winrm stderr winrm close shell changed changed true checksum null dest home user files invocation module args dest home user files fail on missing yes flat yes src c apps ult master noc zip module name fetch null remote checksum remote null task task path build scripts build noetl cust win yml ok msg archive ult master noc zip
1
4,915
25,260,907,488
IssuesEvent
2022-11-15 22:38:54
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Improve UX after columns are extracted to a new table
type: enhancement work: frontend status: ready restricted: maintainers
## Current behavior - We don't provide a great experience for the user after extracting columns to a new table. - The new column ends up ordered at the end. - The new column may be horizontally scrolled out of view (especially given that it's placed at the end of the table). - The new column is not selected. - The user doesn't have any indication of success, other than the modal closing. ## Desired behavior - The new FK column is selected. - The new FK is horizontally scrolled into view. ## Design questions - Other than the "desired behavior" above, is there anything _else_ we want to do to communicate success to the user? Do we want to display a toast message? ## Other _eventually_ desired behavior, not covered by this ticket (Mentioned here for the sake of completeness) - Our [specs](https://github.com/centerofci/mathesar-wiki/blob/master/design/specs/column-extraction.md) also require the new FK to be positioned in the table where the old columns were. Notwithstanding the ambiguity arising from non-adjacent old columns, merely sorting the new column at all will already require some [formidable work](https://github.com/centerofci/mathesar/issues/1584). I don't want to attempt to address that within this ticket, but I'm highlighting this problem to make it clearer why we have the other problems.
True
Improve UX after columns are extracted to a new table - ## Current behavior - We don't provide a great experience for the user after extracting columns to a new table. - The new column ends up ordered at the end. - The new column may be horizontally scrolled out of view (especially given that it's placed at the end of the table). - The new column is not selected. - The user doesn't have any indication of success, other than the modal closing. ## Desired behavior - The new FK column is selected. - The new FK is horizontally scrolled into view. ## Design questions - Other than the "desired behavior" above, is there anything _else_ we want to do to communicate success to the user? Do we want to display a toast message? ## Other _eventually_ desired behavior, not covered by this ticket (Mentioned here for the sake of completeness) - Our [specs](https://github.com/centerofci/mathesar-wiki/blob/master/design/specs/column-extraction.md) also require the new FK to be positioned in the table where the old columns were. Notwithstanding the ambiguity arising from non-adjacent old columns, merely sorting the new column at all will already require some [formidable work](https://github.com/centerofci/mathesar/issues/1584). I don't want to attempt to address that within this ticket, but I'm highlighting this problem to make it clearer why we have the other problems.
main
improve ux after columns are extracted to a new table current behavior we don t provide a great experience for the user after extracting columns to a new table the new column ends up ordered at the end the new column may be horizontally scrolled out of view especially given that it s placed at the end of the table the new column is not selected the user doesn t have any indication of success other than the modal closing desired behavior the new fk column is selected the new fk is horizontally scrolled into view design questions other than the desired behavior above is there anything else we want to do to communicate success to the user do we want to display a toast message other eventually desired behavior not covered by this ticket mentioned here for the sake of completeness our also require the new fk to be positioned in the table where the old columns were notwithstanding the ambiguity arising from non adjacent old columns merely sorting the new column at all will already require some i don t want to attempt to address that within this ticket but i m highlighting this problem to make it clearer why we have the other problems
1
6,985
2,610,320,649
IssuesEvent
2015-02-26 19:43:26
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Text
auto-migrated Priority-Medium Type-Defect
``` Edit: Information about gungans being a threat to CIS forces when doing a land battle says 'missing' ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 8 May 2011 at 4:15
1.0
Text - ``` Edit: Information about gungans being a threat to CIS forces when doing a land battle says 'missing' ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 8 May 2011 at 4:15
non_main
text edit information about gungans being a threat to cis forces when doing a land battle says missing original issue reported on code google com by gmail com on may at
0
7,924
10,127,838,291
IssuesEvent
2019-08-01 11:15:28
konsolas/AAC-Issues
https://api.github.com/repos/konsolas/AAC-Issues
closed
AAC server crash. (LP Verbose)
1. question 3. compatibility 4. alt.-solution
Using LuckPerms (https://github.com/lucko/LuckPerms) verbose spams the entire chat. (Permission checks are constant). Further more it lags the server to the point of a crash. (This has been an issue in the past). _/lp verbose on valese_ Regular Player Verbose: https://i.imgur.com/XAeHtNF.png _/lp verbose on andrewkm_ Admin Verbose: https://i.imgur.com/QXIM745.png Would be lovely if you can fix this, along side LuckPerms, as verbose is very important for debugging permission issues on production servers, and AAC confliction makes this system useless. PS: Perhaps @lucko can shed some light as well. ### Core information **Server version**: 1.12.2 & 1.13.2 **AAC version**: 4.0.5 **AAC config**: 4.0.5|1-Default-Celebrimbor **ProtocolLib version**: 4.4.0 Build 425 **Plugin list**: LuckPerms 4.4.2.7 (Remaining on request via Discord - andrewkm-0001) ### Special environment information **ViaVersion version**: 2.1.4 Build 361 **ProtocolSupport version**: none **ViaBackwards version**: none **BungeeCord plugins**: none
True
AAC server crash. (LP Verbose) - Using LuckPerms (https://github.com/lucko/LuckPerms) verbose spams the entire chat. (Permission checks are constant). Further more it lags the server to the point of a crash. (This has been an issue in the past). _/lp verbose on valese_ Regular Player Verbose: https://i.imgur.com/XAeHtNF.png _/lp verbose on andrewkm_ Admin Verbose: https://i.imgur.com/QXIM745.png Would be lovely if you can fix this, along side LuckPerms, as verbose is very important for debugging permission issues on production servers, and AAC confliction makes this system useless. PS: Perhaps @lucko can shed some light as well. ### Core information **Server version**: 1.12.2 & 1.13.2 **AAC version**: 4.0.5 **AAC config**: 4.0.5|1-Default-Celebrimbor **ProtocolLib version**: 4.4.0 Build 425 **Plugin list**: LuckPerms 4.4.2.7 (Remaining on request via Discord - andrewkm-0001) ### Special environment information **ViaVersion version**: 2.1.4 Build 361 **ProtocolSupport version**: none **ViaBackwards version**: none **BungeeCord plugins**: none
non_main
aac server crash lp verbose using luckperms verbose spams the entire chat permission checks are constant further more it lags the server to the point of a crash this has been an issue in the past lp verbose on valese regular player verbose lp verbose on andrewkm admin verbose would be lovely if you can fix this along side luckperms as verbose is very important for debugging permission issues on production servers and aac confliction makes this system useless ps perhaps lucko can shed some light as well core information server version aac version aac config default celebrimbor protocollib version build plugin list luckperms remaining on request via discord andrewkm special environment information viaversion version build protocolsupport version none viabackwards version none bungeecord plugins none
0
440,104
12,693,285,741
IssuesEvent
2020-06-22 02:46:19
GoogleContainerTools/skaffold
https://api.github.com/repos/GoogleContainerTools/skaffold
closed
skaffold debug with nodejs duplicates ENV values
area/debug kind/bug priority/p1
### Expected behavior When running skaffold debug with NodeJS application deployed by kustomize, I expect the env variables from overlays will not be duplicated or overwritten ### Actual behavior The resulting deployed pod template had duplicate ENV variables. For `valueFrom: secretKeyRef` the additional blank ENV variable actually overwrites the previous value. ![Screen Shot 2020-06-19 at 10 32 21 AM](https://user-images.githubusercontent.com/15863509/85090726-6ca60400-b218-11ea-88b6-f9c509940a28.png) ### Information - Skaffold version: v1.11.0 - Kustomize version {Version:3.6.1 GitCommit:c97fa946d576eb6ed559f17f2ac43b3b5a8d5dbd BuildDate:2020-05-27T23:38:55+01:00 GoOs:darwin GoArch:amd64} - OS: MacOS 10.14.6 - Contents of skaffold.yaml: ```yaml apiVersion: skaffold/v2beta5 kind: Config build: artifacts: - image: skaffold-debug-nodejs context: nodejs deploy: kustomize: {} ``` ### Steps to reproduce the behavior Based on the integration test data in repo: https://github.com/GoogleContainerTools/skaffold/tree/master/integration/testdata/debug 1. a clonable repository: https://github.com/legopin/skaffold-debug-nodejs-env-reproduction.git 2. `skaffold debug` 3. kubectl get deployment nodejs -o yaml
1.0
skaffold debug with nodejs duplicates ENV values - ### Expected behavior When running skaffold debug with NodeJS application deployed by kustomize, I expect the env variables from overlays will not be duplicated or overwritten ### Actual behavior The resulting deployed pod template had duplicate ENV variables. For `valueFrom: secretKeyRef` the additional blank ENV variable actually overwrites the previous value. ![Screen Shot 2020-06-19 at 10 32 21 AM](https://user-images.githubusercontent.com/15863509/85090726-6ca60400-b218-11ea-88b6-f9c509940a28.png) ### Information - Skaffold version: v1.11.0 - Kustomize version {Version:3.6.1 GitCommit:c97fa946d576eb6ed559f17f2ac43b3b5a8d5dbd BuildDate:2020-05-27T23:38:55+01:00 GoOs:darwin GoArch:amd64} - OS: MacOS 10.14.6 - Contents of skaffold.yaml: ```yaml apiVersion: skaffold/v2beta5 kind: Config build: artifacts: - image: skaffold-debug-nodejs context: nodejs deploy: kustomize: {} ``` ### Steps to reproduce the behavior Based on the integration test data in repo: https://github.com/GoogleContainerTools/skaffold/tree/master/integration/testdata/debug 1. a clonable repository: https://github.com/legopin/skaffold-debug-nodejs-env-reproduction.git 2. `skaffold debug` 3. kubectl get deployment nodejs -o yaml
non_main
skaffold debug with nodejs duplicates env values expected behavior when running skaffold debug with nodejs application deployed by kustomize i expect the env variables from overlays will not be duplicated or overwritten actual behavior the resulting deployed pod template had duplicate env variables for valuefrom secretkeyref the additional blank env variable actually overwrites the previous value information skaffold version kustomize version version gitcommit builddate goos darwin goarch os macos contents of skaffold yaml yaml apiversion skaffold kind config build artifacts image skaffold debug nodejs context nodejs deploy kustomize steps to reproduce the behavior based on the integration test data in repo a clonable repository skaffold debug kubectl get deployment nodejs o yaml
0
108,912
23,677,537,774
IssuesEvent
2022-08-28 10:09:18
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[3.0] Repeated Media fields in a repeatable subform loose live preview rendering
No Code Attached Yet
Hello! ### Steps to reproduce the issue Create a form xml with a repeatable subform field element. Within this subform use a media field with preview enabled. ### Expected result When repeating that subform field, every repeated media field preview should display the preview straight away after media selection ### Actual result It works only for the first media field , the one that has not been repeated. Other media fields would only display the preview after the form has been saved ### System information (as much as possible) J3.9.19 - PHP 7.3.12 ### Additional comments
1.0
[3.0] Repeated Media fields in a repeatable subform loose live preview rendering - Hello! ### Steps to reproduce the issue Create a form xml with a repeatable subform field element. Within this subform use a media field with preview enabled. ### Expected result When repeating that subform field, every repeated media field preview should display the preview straight away after media selection ### Actual result It works only for the first media field , the one that has not been repeated. Other media fields would only display the preview after the form has been saved ### System information (as much as possible) J3.9.19 - PHP 7.3.12 ### Additional comments
non_main
repeated media fields in a repeatable subform loose live preview rendering hello steps to reproduce the issue create a form xml with a repeatable subform field element within this subform use a media field with preview enabled expected result when repeating that subform field every repeated media field preview should display the preview straight away after media selection actual result it works only for the first media field the one that has not been repeated other media fields would only display the preview after the form has been saved system information as much as possible php additional comments
0
126,713
17,970,625,131
IssuesEvent
2021-09-14 01:10:49
Watemlifts/redash
https://api.github.com/repos/Watemlifts/redash
opened
WS-2021-0369 (Medium) detected in sqlparse-0.2.4-py2.py3-none-any.whl
security vulnerability
## WS-2021-0369 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sqlparse-0.2.4-py2.py3-none-any.whl</b></p></summary> <p>Non-validating SQL parser</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/65/85/20bdd72f4537cf2c4d5d005368d502b2f464ede22982e724a82c86268eda/sqlparse-0.2.4-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/65/85/20bdd72f4537cf2c4d5d005368d502b2f464ede22982e724a82c86268eda/sqlparse-0.2.4-py2.py3-none-any.whl</a></p> <p>Path to dependency file: redash/requirements.txt</p> <p>Path to vulnerable library: redash/requirements.txt</p> <p> Dependency Hierarchy: - :x: **sqlparse-0.2.4-py2.py3-none-any.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> StripComments filter contains a regular expression that is vulnerable to ReDOS (Regular Expression Denial of Service) The formatter function that strips comments from a SQL contains a regular expression that is vulnerable to ReDoS (Regular Expression Denial of Service). The regular expression may cause exponential backtracking on strings containing many repetitions of '\r\n' in SQL comments. <p>Publish Date: 2021-09-10 <p>URL: <a href=https://github.com/andialbrecht/sqlparse/commit/8238a9e450ed1524e40cb3a8b0b3c00606903aeb>WS-2021-0369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-p5w8-wqhj-9hhf">https://github.com/advisories/GHSA-p5w8-wqhj-9hhf</a></p> <p>Release Date: 2021-09-10</p> <p>Fix Resolution: sqlparse - 0.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2021-0369 (Medium) detected in sqlparse-0.2.4-py2.py3-none-any.whl - ## WS-2021-0369 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sqlparse-0.2.4-py2.py3-none-any.whl</b></p></summary> <p>Non-validating SQL parser</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/65/85/20bdd72f4537cf2c4d5d005368d502b2f464ede22982e724a82c86268eda/sqlparse-0.2.4-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/65/85/20bdd72f4537cf2c4d5d005368d502b2f464ede22982e724a82c86268eda/sqlparse-0.2.4-py2.py3-none-any.whl</a></p> <p>Path to dependency file: redash/requirements.txt</p> <p>Path to vulnerable library: redash/requirements.txt</p> <p> Dependency Hierarchy: - :x: **sqlparse-0.2.4-py2.py3-none-any.whl** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> StripComments filter contains a regular expression that is vulnerable to ReDOS (Regular Expression Denial of Service) The formatter function that strips comments from a SQL contains a regular expression that is vulnerable to ReDoS (Regular Expression Denial of Service). The regular expression may cause exponential backtracking on strings containing many repetitions of '\r\n' in SQL comments. <p>Publish Date: 2021-09-10 <p>URL: <a href=https://github.com/andialbrecht/sqlparse/commit/8238a9e450ed1524e40cb3a8b0b3c00606903aeb>WS-2021-0369</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-p5w8-wqhj-9hhf">https://github.com/advisories/GHSA-p5w8-wqhj-9hhf</a></p> <p>Release Date: 2021-09-10</p> <p>Fix Resolution: sqlparse - 0.4.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
ws medium detected in sqlparse none any whl ws medium severity vulnerability vulnerable library sqlparse none any whl non validating sql parser library home page a href path to dependency file redash requirements txt path to vulnerable library redash requirements txt dependency hierarchy x sqlparse none any whl vulnerable library vulnerability details stripcomments filter contains a regular expression that is vulnerable to redos regular expression denial of service the formatter function that strips comments from a sql contains a regular expression that is vulnerable to redos regular expression denial of service the regular expression may cause exponential backtracking on strings containing many repetitions of r n in sql comments publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sqlparse step up your open source security game with whitesource
0
44,738
13,079,412,003
IssuesEvent
2020-08-01 03:13:20
fufunoyu/example-pip-travis
https://api.github.com/repos/fufunoyu/example-pip-travis
opened
CVE-2020-10994 (Medium) detected in Pillow-3.2.0.tar.gz
security vulnerability
## CVE-2020-10994 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-3.2.0.tar.gz</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e2/af/0a3981fffc5cd43078eb8b1057702e0dd2d5771e5aaa36cbd140e32f8473/Pillow-3.2.0.tar.gz">https://files.pythonhosted.org/packages/e2/af/0a3981fffc5cd43078eb8b1057702e0dd2d5771e5aaa36cbd140e32f8473/Pillow-3.2.0.tar.gz</a></p> <p>Path to dependency file: /tmp/ws-scm/example-pip-travis/requirements.txt</p> <p>Path to vulnerable library: /tmp/ws-scm/example-pip-travis/requirements.txt</p> <p> Dependency Hierarchy: - image-1.5.5.tar.gz (Root Library) - :x: **Pillow-3.2.0.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fufunoyu/example-pip-travis/commit/6393ea41be5cd8cdab87f9e6f950e76cd82e60c4">6393ea41be5cd8cdab87f9e6f950e76cd82e60c4</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In libImaging/Jpeg2KDecode.c in Pillow before 7.1.0, there are multiple out-of-bounds reads via a crafted JP2 file. <p>Publish Date: 2020-06-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10994>CVE-2020-10994</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8">https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8</a></p> <p>Release Date: 2020-06-25</p> <p>Fix Resolution: 7.1.0</p> </p> </details> <p></p>
True
CVE-2020-10994 (Medium) detected in Pillow-3.2.0.tar.gz - ## CVE-2020-10994 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-3.2.0.tar.gz</b></p></summary> <p>Python Imaging Library (Fork)</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/e2/af/0a3981fffc5cd43078eb8b1057702e0dd2d5771e5aaa36cbd140e32f8473/Pillow-3.2.0.tar.gz">https://files.pythonhosted.org/packages/e2/af/0a3981fffc5cd43078eb8b1057702e0dd2d5771e5aaa36cbd140e32f8473/Pillow-3.2.0.tar.gz</a></p> <p>Path to dependency file: /tmp/ws-scm/example-pip-travis/requirements.txt</p> <p>Path to vulnerable library: /tmp/ws-scm/example-pip-travis/requirements.txt</p> <p> Dependency Hierarchy: - image-1.5.5.tar.gz (Root Library) - :x: **Pillow-3.2.0.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fufunoyu/example-pip-travis/commit/6393ea41be5cd8cdab87f9e6f950e76cd82e60c4">6393ea41be5cd8cdab87f9e6f950e76cd82e60c4</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In libImaging/Jpeg2KDecode.c in Pillow before 7.1.0, there are multiple out-of-bounds reads via a crafted JP2 file. <p>Publish Date: 2020-06-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10994>CVE-2020-10994</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8">https://github.com/python-pillow/Pillow/commit/41b554bc56982ee4f30238a7677c0f4ff90a73a8</a></p> <p>Release Date: 2020-06-25</p> <p>Fix Resolution: 7.1.0</p> </p> </details> <p></p>
non_main
cve medium detected in pillow tar gz cve medium severity vulnerability vulnerable library pillow tar gz python imaging library fork library home page a href path to dependency file tmp ws scm example pip travis requirements txt path to vulnerable library tmp ws scm example pip travis requirements txt dependency hierarchy image tar gz root library x pillow tar gz vulnerable library found in head commit a href vulnerability details in libimaging c in pillow before there are multiple out of bounds reads via a crafted file publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
86,845
24,969,536,390
IssuesEvent
2022-11-01 23:00:23
NVIDIA/spark-rapids
https://api.github.com/repos/NVIDIA/spark-rapids
reopened
[SPARK-40066][SQL] ANSI mode: always return null on invalid access to map column
bug build P0 audit_3.4.0
#### Context - Spark 3.4: https://github.com/apache/spark/commit/95cbd61857 - Followup: https://github.com/apache/spark/commit/fbc0edac58 - Other related error formatting auditing issues: https://github.com/NVIDIA/spark-rapids/issues/6293 #### What changes were proposed in SPARK pull request? Change the syntax of map column access under ANSI mode: always return null results instead of throwing `MAP_KEY_DOES_NOT_EXIST` errors. This PR also remove an internal `spark.sql.ansi.strictIndexOperator`. #### Why are the changes needed in Spark? PostgreSQL/Snowflake/Biguqery and all of them returns null return on map(json) key not exists. #### Does this PR introduce _any_ user-facing change? Yes. Since Spark 3.4, when ANSI SQL mode(configuration `spark.sql.ansi.enabled`) is on, Spark SQL always returns NULL result on getting a map value with a non-existing key. In Spark 3.3 or earlier, there will be an error. #### Why it might affect RAPIDS? Yes, change of behavior in Ansi Mode. - shims for 3.4+ need to be updated. #### Impact on Testing? Yes, need to update integration tests to match Spark Behavior #### Requires Doc update? No
1.0
[SPARK-40066][SQL] ANSI mode: always return null on invalid access to map column - #### Context - Spark 3.4: https://github.com/apache/spark/commit/95cbd61857 - Followup: https://github.com/apache/spark/commit/fbc0edac58 - Other related error formatting auditing issues: https://github.com/NVIDIA/spark-rapids/issues/6293 #### What changes were proposed in SPARK pull request? Change the syntax of map column access under ANSI mode: always return null results instead of throwing `MAP_KEY_DOES_NOT_EXIST` errors. This PR also remove an internal `spark.sql.ansi.strictIndexOperator`. #### Why are the changes needed in Spark? PostgreSQL/Snowflake/Biguqery and all of them returns null return on map(json) key not exists. #### Does this PR introduce _any_ user-facing change? Yes. Since Spark 3.4, when ANSI SQL mode(configuration `spark.sql.ansi.enabled`) is on, Spark SQL always returns NULL result on getting a map value with a non-existing key. In Spark 3.3 or earlier, there will be an error. #### Why it might affect RAPIDS? Yes, change of behavior in Ansi Mode. - shims for 3.4+ need to be updated. #### Impact on Testing? Yes, need to update integration tests to match Spark Behavior #### Requires Doc update? No
non_main
ansi mode always return null on invalid access to map column context spark followup other related error formatting auditing issues what changes were proposed in spark pull request change the syntax of map column access under ansi mode always return null results instead of throwing map key does not exist errors this pr also remove an internal spark sql ansi strictindexoperator why are the changes needed in spark postgresql snowflake biguqery and all of them returns null return on map json key not exists does this pr introduce any user facing change yes since spark when ansi sql mode configuration spark sql ansi enabled is on spark sql always returns null result on getting a map value with a non existing key in spark or earlier there will be an error why it might affect rapids yes change of behavior in ansi mode shims for need to be updated impact on testing yes need to update integration tests to match spark behavior requires doc update no
0
83,530
3,636,965,593
IssuesEvent
2016-02-12 08:14:14
cwrc/HuViz
https://api.github.com/repos/cwrc/HuViz
closed
point HuViz at the latest Orlando Ontology as hosted on GitHub (3 hrs)
Dataset DEVELOPMENT Functionality Priority - high
Have users of HuViz get their version of the ontology from the appropriate GitHub repo directly. The file needs to be in .ttl format for HuViz to be able to parse it. Communicate with Rob requesting a parallel version being included in the repo.
1.0
point HuViz at the latest Orlando Ontology as hosted on GitHub (3 hrs) - Have users of HuViz get their version of the ontology from the appropriate GitHub repo directly. The file needs to be in .ttl format for HuViz to be able to parse it. Communicate with Rob requesting a parallel version being included in the repo.
non_main
point huviz at the latest orlando ontology as hosted on github hrs have users of huviz get their version of the ontology from the appropriate github repo directly the file needs to be in ttl format for huviz to be able to parse it communicate with rob requesting a parallel version being included in the repo
0
18,480
2,615,171,941
IssuesEvent
2015-03-01 06:53:55
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
I'm
auto-migrated Priority-P2 Type-Bug
``` Please describe the issue: Please provide any additional information below. ``` Original issue reported on code.google.com by `russell.lr44` on 12 Jan 2012 at 3:14
1.0
I'm - ``` Please describe the issue: Please provide any additional information below. ``` Original issue reported on code.google.com by `russell.lr44` on 12 Jan 2012 at 3:14
non_main
i m please describe the issue please provide any additional information below original issue reported on code google com by russell on jan at
0
3,749
15,781,716,699
IssuesEvent
2021-04-01 11:47:49
svengreb/tmpl
https://api.github.com/repos/svengreb/tmpl
opened
Dependency handling with lockfiles
context-techstack context-workflow scope-compatibility scope-maintainability scope-stability target-base type-task
The usage of dependency lockfiles like [`package-lock.json`][npm-docs-v7-lockfile] or [`yarn.lock`][yarn-docs-lock] has always been a controversial topic where opinions go in different directions. On one side many project maintainers tend to argue that is helps to achieve deterministic build results, but on the side it might also hide problems when any later versions of a used dependency, or its transitive dependencies, is not compatible with the own project anymore. I‘ve investigated a lot of time into research again to finally find a solution that works for my projects. In short, the result is to go with the rule that is also used by many large-scale projects: Do **not use lockfiles for multi-consumer projects like libraries** but **only for single-consumer projects like applications**. Therefore the currently committed `yarn.lock` file will be removed since this makes no sense for a repository template anyway. The the sections below for some more details about how to decide to use a lockfile or not. ### When to use lockfiles The clear advantage of lockfiles are reproducible builds and the persistence of a running project state. They ensure that a project artifact can be rebuild at anytime using the exact same dependencies, resulting in the exact same artifact, even when the project was not updated in years. This applies to projects that are focused on building a **end-to-end experience like applications and other end-user products**. These are the advantages listed in the official npm documentation about `package-lock.json` files: > - Describe a single representation of a dependency tree such that teammates, deployments, and continuous integration are guaranteed to install exactly the same dependencies. > - Provide a facility for users to "time-travel" to previous states of node_modules without having to commit the directory itself. > - Facilitate greater visibility of tree changes through readable source control diffs. > - Optimize the installation process by allowing npm to skip repeated metadata resolutions for previously-installed packages. > - As of npm `v7`, lockfiles include enough information to gain a complete picture of the package tree, reducing the need to read package.json files, and allowing for significant performance improvements. Like mentioned, npm `v7` comes with a lot of advantages and the [team recommends to commit the file into project repositories][npm-blog-v7_keep_lock]: - the lockfile has enough information to describe the precise package tree all by itself. - the lockfile maps the packages to their information by their relative location to the root (instead of their name). - the npm CLI uses `yarn.lock` lockfiles if available, as a source of package metadata and resolution guidance when there is missing information, knowing that the `package-lock.json` is the authoritative definition. - `yarn.lock` lockfiles cannot completely replace npm’s lockfile since the current implementation doesn’t have enough information needed for the complete npm functionality. - the npm CLI uses a “hidden lockfile“ placed inside the `node_module` directory that helps to avoid repeated package tree reading. Another points is that in end-user projects dependencies in `package.json` files are pinned often instead of using [SemVer range selectors][npm-webapp-semver] like `^` (latest minor-only) or `~` (latest patch only). In such cases a lockfile helps to keep control about transitive dependencies and persist projects states in time. ### When to avoid lockfiles Even though [the _Yarn_ team published a blog post in 2016][yarn-blog-lockfiles] that states to always commit the `yarn.lock` file, regardless of the project type, this advice was not adopted by every project and some “real-world scenarios“ often showed that this decision was justified. There are [blog posts that summarize when not to use a lockfile][medium-gajus-stop_lockfile] where even [Yarn maintainers reply with comments that claim the opposite][devto-gajus-stop_lockfile-comment-yarn_maintainer], but over the time more and more projects went away from using One argument is that [lockfiles are important to enure that library contributors in 10 years still know what was the last confirmed set of packages which worked as expected][tw-arcanis-1164229994165559299-comment-19], but this can almost be ignored in a ecosystem like Node that changes almost every day. Another important point is to mention that the usage of [lockfiles were also a attack surface to inject malicious dependencies][snyk-blog-lockfile]. Due to the large size of lockfiles, it is also often a challenge for project maintainers to review and validate a lockfile in pull requests are so they are often [ignored and blindly trusted][tw-bcrypt-1208950722097598465]. The community is still not of one opinion and I guess this will never change, but [learning about the experience of well-known maintainers][gh-sindresorhus/ama-479#c-310661514] and [popular projects][gh-airbnb/javascript-2409] is often a good way to find the own decision. In conclusion, the usage of lockfiles in a non-end-user project can be well summarized with [“just postponing the inevitable breakage“][tw-renovatebot-1163789817492230144]: <p align="center"><img src="https://user-images.githubusercontent.com/13448100/113289644-c3388680-92f0-11eb-9a0b-d710c78edb92.png" width="350" /></p> [devto-gajus-stop_lockfile-comment-yarn_maintainer]: https://dev.to/arcanis/comment/fo33 [devto-gajus-stop_lockfile]: https://dev.to/gajus/stop-using-package-lock-json-or-yarn-lock-3ddi [gh-airbnb/javascript-2409]: https://github.com/airbnb/javascript/issues/2409 [gh-sindresorhus/ama-479#c-310661514]: https://github.com/sindresorhus/ama/issues/479#issuecomment-310661514 [npm-blog-v7_keep_lock]: https://blog.npmjs.org/post/621733939456933888/npm-v7-series-why-keep-package-lockjson.html [npm-docs-v7-lockfile]: https://docs.npmjs.com/cli/v7/configuring-npm/package-lock-json [npm-webapp-semver]: https://semver.npmjs.com [snyk-blog-lockfile]: https://snyk.io/blog/why-npm-lockfiles-can-be-a-security-blindspot-for-injecting-malicious-modules [tw-arcanis-1164229994165559299-comment-19]: https://twitter.com/arcanis/status/1164229994165559299?s=19 [tw-bcrypt-1208950722097598465]: https://twitter.com/bcrypt/status/1208950722097598465 [tw-renovatebot-1163789817492230144]: https://twitter.com/renovatebot/status/1163789817492230144 [yarn-blog-lockfiles]: https://classic.yarnpkg.com/blog/2016/11/24/lockfiles-for-all [yarn-docs-lock]: https://classic.yarnpkg.com/en/docs/yarn-lock
True
Dependency handling with lockfiles - The usage of dependency lockfiles like [`package-lock.json`][npm-docs-v7-lockfile] or [`yarn.lock`][yarn-docs-lock] has always been a controversial topic where opinions go in different directions. On one side many project maintainers tend to argue that is helps to achieve deterministic build results, but on the side it might also hide problems when any later versions of a used dependency, or its transitive dependencies, is not compatible with the own project anymore. I‘ve investigated a lot of time into research again to finally find a solution that works for my projects. In short, the result is to go with the rule that is also used by many large-scale projects: Do **not use lockfiles for multi-consumer projects like libraries** but **only for single-consumer projects like applications**. Therefore the currently committed `yarn.lock` file will be removed since this makes no sense for a repository template anyway. The the sections below for some more details about how to decide to use a lockfile or not. ### When to use lockfiles The clear advantage of lockfiles are reproducible builds and the persistence of a running project state. They ensure that a project artifact can be rebuild at anytime using the exact same dependencies, resulting in the exact same artifact, even when the project was not updated in years. This applies to projects that are focused on building a **end-to-end experience like applications and other end-user products**. These are the advantages listed in the official npm documentation about `package-lock.json` files: > - Describe a single representation of a dependency tree such that teammates, deployments, and continuous integration are guaranteed to install exactly the same dependencies. > - Provide a facility for users to "time-travel" to previous states of node_modules without having to commit the directory itself. > - Facilitate greater visibility of tree changes through readable source control diffs. > - Optimize the installation process by allowing npm to skip repeated metadata resolutions for previously-installed packages. > - As of npm `v7`, lockfiles include enough information to gain a complete picture of the package tree, reducing the need to read package.json files, and allowing for significant performance improvements. Like mentioned, npm `v7` comes with a lot of advantages and the [team recommends to commit the file into project repositories][npm-blog-v7_keep_lock]: - the lockfile has enough information to describe the precise package tree all by itself. - the lockfile maps the packages to their information by their relative location to the root (instead of their name). - the npm CLI uses `yarn.lock` lockfiles if available, as a source of package metadata and resolution guidance when there is missing information, knowing that the `package-lock.json` is the authoritative definition. - `yarn.lock` lockfiles cannot completely replace npm’s lockfile since the current implementation doesn’t have enough information needed for the complete npm functionality. - the npm CLI uses a “hidden lockfile“ placed inside the `node_module` directory that helps to avoid repeated package tree reading. Another points is that in end-user projects dependencies in `package.json` files are pinned often instead of using [SemVer range selectors][npm-webapp-semver] like `^` (latest minor-only) or `~` (latest patch only). In such cases a lockfile helps to keep control about transitive dependencies and persist projects states in time. ### When to avoid lockfiles Even though [the _Yarn_ team published a blog post in 2016][yarn-blog-lockfiles] that states to always commit the `yarn.lock` file, regardless of the project type, this advice was not adopted by every project and some “real-world scenarios“ often showed that this decision was justified. There are [blog posts that summarize when not to use a lockfile][medium-gajus-stop_lockfile] where even [Yarn maintainers reply with comments that claim the opposite][devto-gajus-stop_lockfile-comment-yarn_maintainer], but over the time more and more projects went away from using One argument is that [lockfiles are important to enure that library contributors in 10 years still know what was the last confirmed set of packages which worked as expected][tw-arcanis-1164229994165559299-comment-19], but this can almost be ignored in a ecosystem like Node that changes almost every day. Another important point is to mention that the usage of [lockfiles were also a attack surface to inject malicious dependencies][snyk-blog-lockfile]. Due to the large size of lockfiles, it is also often a challenge for project maintainers to review and validate a lockfile in pull requests are so they are often [ignored and blindly trusted][tw-bcrypt-1208950722097598465]. The community is still not of one opinion and I guess this will never change, but [learning about the experience of well-known maintainers][gh-sindresorhus/ama-479#c-310661514] and [popular projects][gh-airbnb/javascript-2409] is often a good way to find the own decision. In conclusion, the usage of lockfiles in a non-end-user project can be well summarized with [“just postponing the inevitable breakage“][tw-renovatebot-1163789817492230144]: <p align="center"><img src="https://user-images.githubusercontent.com/13448100/113289644-c3388680-92f0-11eb-9a0b-d710c78edb92.png" width="350" /></p> [devto-gajus-stop_lockfile-comment-yarn_maintainer]: https://dev.to/arcanis/comment/fo33 [devto-gajus-stop_lockfile]: https://dev.to/gajus/stop-using-package-lock-json-or-yarn-lock-3ddi [gh-airbnb/javascript-2409]: https://github.com/airbnb/javascript/issues/2409 [gh-sindresorhus/ama-479#c-310661514]: https://github.com/sindresorhus/ama/issues/479#issuecomment-310661514 [npm-blog-v7_keep_lock]: https://blog.npmjs.org/post/621733939456933888/npm-v7-series-why-keep-package-lockjson.html [npm-docs-v7-lockfile]: https://docs.npmjs.com/cli/v7/configuring-npm/package-lock-json [npm-webapp-semver]: https://semver.npmjs.com [snyk-blog-lockfile]: https://snyk.io/blog/why-npm-lockfiles-can-be-a-security-blindspot-for-injecting-malicious-modules [tw-arcanis-1164229994165559299-comment-19]: https://twitter.com/arcanis/status/1164229994165559299?s=19 [tw-bcrypt-1208950722097598465]: https://twitter.com/bcrypt/status/1208950722097598465 [tw-renovatebot-1163789817492230144]: https://twitter.com/renovatebot/status/1163789817492230144 [yarn-blog-lockfiles]: https://classic.yarnpkg.com/blog/2016/11/24/lockfiles-for-all [yarn-docs-lock]: https://classic.yarnpkg.com/en/docs/yarn-lock
main
dependency handling with lockfiles the usage of dependency lockfiles like or has always been a controversial topic where opinions go in different directions on one side many project maintainers tend to argue that is helps to achieve deterministic build results but on the side it might also hide problems when any later versions of a used dependency or its transitive dependencies is not compatible with the own project anymore i‘ve investigated a lot of time into research again to finally find a solution that works for my projects in short the result is to go with the rule that is also used by many large scale projects do not use lockfiles for multi consumer projects like libraries but only for single consumer projects like applications therefore the currently committed yarn lock file will be removed since this makes no sense for a repository template anyway the the sections below for some more details about how to decide to use a lockfile or not when to use lockfiles the clear advantage of lockfiles are reproducible builds and the persistence of a running project state they ensure that a project artifact can be rebuild at anytime using the exact same dependencies resulting in the exact same artifact even when the project was not updated in years this applies to projects that are focused on building a end to end experience like applications and other end user products these are the advantages listed in the official npm documentation about package lock json files describe a single representation of a dependency tree such that teammates deployments and continuous integration are guaranteed to install exactly the same dependencies provide a facility for users to time travel to previous states of node modules without having to commit the directory itself facilitate greater visibility of tree changes through readable source control diffs optimize the installation process by allowing npm to skip repeated metadata resolutions for previously installed packages as of npm lockfiles include enough information to gain a complete picture of the package tree reducing the need to read package json files and allowing for significant performance improvements like mentioned npm comes with a lot of advantages and the the lockfile has enough information to describe the precise package tree all by itself the lockfile maps the packages to their information by their relative location to the root instead of their name the npm cli uses yarn lock lockfiles if available as a source of package metadata and resolution guidance when there is missing information knowing that the package lock json is the authoritative definition yarn lock lockfiles cannot completely replace npm’s lockfile since the current implementation doesn’t have enough information needed for the complete npm functionality the npm cli uses a “hidden lockfile“ placed inside the node module directory that helps to avoid repeated package tree reading another points is that in end user projects dependencies in package json files are pinned often instead of using like latest minor only or latest patch only in such cases a lockfile helps to keep control about transitive dependencies and persist projects states in time when to avoid lockfiles even though that states to always commit the yarn lock file regardless of the project type this advice was not adopted by every project and some “real world scenarios“ often showed that this decision was justified there are where even but over the time more and more projects went away from using one argument is that but this can almost be ignored in a ecosystem like node that changes almost every day another important point is to mention that the usage of due to the large size of lockfiles it is also often a challenge for project maintainers to review and validate a lockfile in pull requests are so they are often the community is still not of one opinion and i guess this will never change but and is often a good way to find the own decision in conclusion the usage of lockfiles in a non end user project can be well summarized with
1
1,119
4,989,596,269
IssuesEvent
2016-12-08 12:26:47
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_container doesn't allow setting log_options without setting log_driver
affects_2.2 bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel 28feba2fb3) last updated 2016/09/14 12:25:32 (GMT +200) lib/ansible/modules/core: (detached HEAD ae6992bf8c) last updated 2016/09/14 12:25:32 (GMT +200) lib/ansible/modules/extras: (detached HEAD afd0b23836) last updated 2016/09/14 12:25:32 (GMT +200) ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY Docker uses the log driver 'json-file' by default (also stated in the docker_container documentation). If you want to provide log_options (e.g. for log rotation) you have to explicitly set the log_driver to json-file, else this module attribute is ignored, although it should IMHO assume the Docker default. This behavior is probably caused by the [following code](https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker_container.py#L1091) ``` def _parse_log_config(self): ''' Create a LogConfig object ''' if self.log_driver is None: return None ``` ##### STEPS TO REPRODUCE ``` docker_container: name: mongodb image: bitnami/mongodb log_options: max-size: "512m" max-file: "5" ``` ##### EXPECTED RESULTS When you check the specs of the running container (docker inspect) you see should see the passed log options. ``` "LogConfig": { "Type": "json-file", "Config": { "max-file": "5", "max-size": "512m" } ``` ##### ACTUAL RESULTS When you check the specs of the running container you see that the log config block is empty. ``` "LogConfig": { "Type": "json-file", "Config": {} }, ```
True
docker_container doesn't allow setting log_options without setting log_driver - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.2.0 (devel 28feba2fb3) last updated 2016/09/14 12:25:32 (GMT +200) lib/ansible/modules/core: (detached HEAD ae6992bf8c) last updated 2016/09/14 12:25:32 (GMT +200) lib/ansible/modules/extras: (detached HEAD afd0b23836) last updated 2016/09/14 12:25:32 (GMT +200) ``` ##### CONFIGURATION N/A ##### OS / ENVIRONMENT N/A ##### SUMMARY Docker uses the log driver 'json-file' by default (also stated in the docker_container documentation). If you want to provide log_options (e.g. for log rotation) you have to explicitly set the log_driver to json-file, else this module attribute is ignored, although it should IMHO assume the Docker default. This behavior is probably caused by the [following code](https://github.com/ansible/ansible-modules-core/blob/devel/cloud/docker/docker_container.py#L1091) ``` def _parse_log_config(self): ''' Create a LogConfig object ''' if self.log_driver is None: return None ``` ##### STEPS TO REPRODUCE ``` docker_container: name: mongodb image: bitnami/mongodb log_options: max-size: "512m" max-file: "5" ``` ##### EXPECTED RESULTS When you check the specs of the running container (docker inspect) you see should see the passed log options. ``` "LogConfig": { "Type": "json-file", "Config": { "max-file": "5", "max-size": "512m" } ``` ##### ACTUAL RESULTS When you check the specs of the running container you see that the log config block is empty. ``` "LogConfig": { "Type": "json-file", "Config": {} }, ```
main
docker container doesn t allow setting log options without setting log driver issue type bug report component name docker container ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt configuration n a os environment n a summary docker uses the log driver json file by default also stated in the docker container documentation if you want to provide log options e g for log rotation you have to explicitly set the log driver to json file else this module attribute is ignored although it should imho assume the docker default this behavior is probably caused by the def parse log config self create a logconfig object if self log driver is none return none steps to reproduce docker container name mongodb image bitnami mongodb log options max size max file expected results when you check the specs of the running container docker inspect you see should see the passed log options logconfig type json file config max file max size actual results when you check the specs of the running container you see that the log config block is empty logconfig type json file config
1
3,397
13,170,797,432
IssuesEvent
2020-08-11 15:39:33
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Proxmox : netif and ip_address incompatible with proxmox 4.0
affects_2.3 bug cloud collection collection:community.libvirt module needs_collection_redirect support:community virt waiting_on_maintainer
_From @Jdsilberzahn On March 4 2016 1247_ ##### ISSUE TYPE Bug Report ##### COMPONENT NAME proxmox ##### ANSIBLE VERSION 2.3 ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY Plugin name: proxmox when using netif / ip_address with LXC containers (pve 4), it results in a 400 Bad Request Error Quick and dirty workaround: - task call : netif='name=eth0,hwaddr=52:4A:5E:26:58:D8,ip=192.168.15.147/24,gw=192.168.15.1,bridge=vmbr0' - in proxmox.py (ligne 349 - create_instance) : net0 = module.params['netif'], _Copied from original issue: ansible/ansible-modules-extras#1789_
True
Proxmox : netif and ip_address incompatible with proxmox 4.0 - _From @Jdsilberzahn On March 4 2016 1247_ ##### ISSUE TYPE Bug Report ##### COMPONENT NAME proxmox ##### ANSIBLE VERSION 2.3 ##### CONFIGURATION ##### OS / ENVIRONMENT ##### SUMMARY Plugin name: proxmox when using netif / ip_address with LXC containers (pve 4), it results in a 400 Bad Request Error Quick and dirty workaround: - task call : netif='name=eth0,hwaddr=52:4A:5E:26:58:D8,ip=192.168.15.147/24,gw=192.168.15.1,bridge=vmbr0' - in proxmox.py (ligne 349 - create_instance) : net0 = module.params['netif'], _Copied from original issue: ansible/ansible-modules-extras#1789_
main
proxmox netif and ip address incompatible with proxmox from jdsilberzahn on march issue type bug report component name proxmox ansible version configuration os environment summary plugin name proxmox when using netif ip address with lxc containers pve it results in a bad request error quick and dirty workaround task call netif name hwaddr ip gw bridge in proxmox py ligne create instance module params copied from original issue ansible ansible modules extras
1
5,825
30,841,799,125
IssuesEvent
2023-08-02 11:10:44
cncf/glossary
https://api.github.com/repos/cncf/glossary
closed
Reviews for non-English `i18n/xx.toml` are requested to repo maintainers
maintainers
- Example: #2184 - Suggested solution: - move each `i18n/xx.toml` to somewhere else, like `data/i18n/xx/xx.toml` (Ref: https://github.com/kubernetes/website/tree/main/data/i18n/ko) - create symbolic link (`i18n/xx.toml` -> `data/i18n/xx/xx.toml`) - populate `CODEOWNERS` files for each `data/i18n/xx/` directory Note: if we are to adopt this, then we need to manage & update `data/i18n/xx/xx.toml` files when the root `CODEOWNERS` file gets changed, to make them synced well.
True
Reviews for non-English `i18n/xx.toml` are requested to repo maintainers - - Example: #2184 - Suggested solution: - move each `i18n/xx.toml` to somewhere else, like `data/i18n/xx/xx.toml` (Ref: https://github.com/kubernetes/website/tree/main/data/i18n/ko) - create symbolic link (`i18n/xx.toml` -> `data/i18n/xx/xx.toml`) - populate `CODEOWNERS` files for each `data/i18n/xx/` directory Note: if we are to adopt this, then we need to manage & update `data/i18n/xx/xx.toml` files when the root `CODEOWNERS` file gets changed, to make them synced well.
main
reviews for non english xx toml are requested to repo maintainers example suggested solution move each xx toml to somewhere else like data xx xx toml ref create symbolic link xx toml data xx xx toml populate codeowners files for each data xx directory note if we are to adopt this then we need to manage update data xx xx toml files when the root codeowners file gets changed to make them synced well
1
5,590
28,013,558,502
IssuesEvent
2023-03-27 20:33:15
bazelbuild/intellij
https://api.github.com/repos/bazelbuild/intellij
closed
Disable "Expand Sync to Working Set" by default
type: feature request product: IntelliJ topic: sync awaiting-maintainer
I want to bring to attention one scenario that I experiences which confused me and my investigation into it. **TLDR:** I propose to change the default for [`BlazeUserSettings#expandSyncToWorkingSet `](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/settings/BlazeUserSettings.java#L72) from `true` to `false` because IMO it can confuse users more then help them. In my case I have a huge mono repository where we practice fork driven approach (each engineer has a fork to work on and we submit PRs to the main repository). I have two origins in my Git repository: `origin` and `fork`. For example, I have a feature branch `fork/fedor/feature-1` which was checkout from `fork/master`. I did some development, opened a PR, got some feedback after a few days and now I'm rebasing my feature branch on top of the latest changes from `origin/master`. Now when I run Sync from IntelliJ I see that the plugin **syncs every single target in the repository**. After some investigation I found the root cause: because my `fork/master` hasn't been rebased as well this causes the "working set" to contain all files changed between `fork/fedor/feature-1` and the outdated `fork/master` instead of `origin/master`. [Here](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/vcs/git/GitBlazeVcsHandler.java#L53-L66) is how `GitBlazeVcsHandler` calculates the working set of files. The troubling part is how the [SHA is calculated in `getUpstreamSha`](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/vcs/git/GitBlazeVcsHandler.java#L123-L141). I propose to change the default for [`BlazeUserSettings#expandSyncToWorkingSet`](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/settings/BlazeUserSettings.java#L72) from `true` to `false`. It seems this automatic expansion of the target set based on changed files can bring more harm than benefits for use cases similar to mine. Unfortunately I wasn't able to find more information about motivation for `expandSyncToWorkingSet` feature because this feature predates the publicly available Git history.
True
Disable "Expand Sync to Working Set" by default - I want to bring to attention one scenario that I experiences which confused me and my investigation into it. **TLDR:** I propose to change the default for [`BlazeUserSettings#expandSyncToWorkingSet `](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/settings/BlazeUserSettings.java#L72) from `true` to `false` because IMO it can confuse users more then help them. In my case I have a huge mono repository where we practice fork driven approach (each engineer has a fork to work on and we submit PRs to the main repository). I have two origins in my Git repository: `origin` and `fork`. For example, I have a feature branch `fork/fedor/feature-1` which was checkout from `fork/master`. I did some development, opened a PR, got some feedback after a few days and now I'm rebasing my feature branch on top of the latest changes from `origin/master`. Now when I run Sync from IntelliJ I see that the plugin **syncs every single target in the repository**. After some investigation I found the root cause: because my `fork/master` hasn't been rebased as well this causes the "working set" to contain all files changed between `fork/fedor/feature-1` and the outdated `fork/master` instead of `origin/master`. [Here](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/vcs/git/GitBlazeVcsHandler.java#L53-L66) is how `GitBlazeVcsHandler` calculates the working set of files. The troubling part is how the [SHA is calculated in `getUpstreamSha`](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/vcs/git/GitBlazeVcsHandler.java#L123-L141). I propose to change the default for [`BlazeUserSettings#expandSyncToWorkingSet`](https://github.com/bazelbuild/intellij/blob/4c91d565062e152cace081bf3bfc8860008dc145/base/src/com/google/idea/blaze/base/settings/BlazeUserSettings.java#L72) from `true` to `false`. It seems this automatic expansion of the target set based on changed files can bring more harm than benefits for use cases similar to mine. Unfortunately I wasn't able to find more information about motivation for `expandSyncToWorkingSet` feature because this feature predates the publicly available Git history.
main
disable expand sync to working set by default i want to bring to attention one scenario that i experiences which confused me and my investigation into it tldr i propose to change the default for from true to false because imo it can confuse users more then help them in my case i have a huge mono repository where we practice fork driven approach each engineer has a fork to work on and we submit prs to the main repository i have two origins in my git repository origin and fork for example i have a feature branch fork fedor feature which was checkout from fork master i did some development opened a pr got some feedback after a few days and now i m rebasing my feature branch on top of the latest changes from origin master now when i run sync from intellij i see that the plugin syncs every single target in the repository after some investigation i found the root cause because my fork master hasn t been rebased as well this causes the working set to contain all files changed between fork fedor feature and the outdated fork master instead of origin master is how gitblazevcshandler calculates the working set of files the troubling part is how the i propose to change the default for from true to false it seems this automatic expansion of the target set based on changed files can bring more harm than benefits for use cases similar to mine unfortunately i wasn t able to find more information about motivation for expandsynctoworkingset feature because this feature predates the publicly available git history
1
122,828
4,845,790,164
IssuesEvent
2016-11-10 09:35:32
rism-ch/muscat
https://api.github.com/repos/rism-ch/muscat
closed
Adding a script to mark the Bach Digital titles as unpublished
First priority
We do not have this in the MarcXML, we need an additional post-script
1.0
Adding a script to mark the Bach Digital titles as unpublished - We do not have this in the MarcXML, we need an additional post-script
non_main
adding a script to mark the bach digital titles as unpublished we do not have this in the marcxml we need an additional post script
0
542,688
15,864,915,908
IssuesEvent
2021-04-08 14:13:50
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
opened
scandi_adventurer_region_trigger Error
:beetle: bug :beetle: :exclamation: priority high good first issue
<!-- **DO NOT REMOVE PRE-EXISTING LINES** ------------------------------------------------------------------------------------------------------------ --> **Your mod version is:** 8377e6400368feee2a27088843f3f62cf44311dd **What expansions do you have installed?** NL **Are you using any submods/mods? If so, which?** Nope **Please explain your issue in as much detail as possible:** ``` [18:12:05][jomini_script_system.cpp:169]: Script system error! Error: geographical_region trigger [ Wrong scope for trigger: character, expected province ] Script location: file: file: common/script_values/00_war_values.txt line: 99; scandi_adventurer_region_trigger line: 2 ``` **Steps to reproduce the issue:** Open `error.log` **Upload an attachment below: .zip of your save, or screenshots:**
1.0
scandi_adventurer_region_trigger Error - <!-- **DO NOT REMOVE PRE-EXISTING LINES** ------------------------------------------------------------------------------------------------------------ --> **Your mod version is:** 8377e6400368feee2a27088843f3f62cf44311dd **What expansions do you have installed?** NL **Are you using any submods/mods? If so, which?** Nope **Please explain your issue in as much detail as possible:** ``` [18:12:05][jomini_script_system.cpp:169]: Script system error! Error: geographical_region trigger [ Wrong scope for trigger: character, expected province ] Script location: file: file: common/script_values/00_war_values.txt line: 99; scandi_adventurer_region_trigger line: 2 ``` **Steps to reproduce the issue:** Open `error.log` **Upload an attachment below: .zip of your save, or screenshots:**
non_main
scandi adventurer region trigger error do not remove pre existing lines your mod version is what expansions do you have installed nl are you using any submods mods if so which nope please explain your issue in as much detail as possible script system error error geographical region trigger script location file file common script values war values txt line scandi adventurer region trigger line steps to reproduce the issue open error log upload an attachment below zip of your save or screenshots
0
763
4,364,160,894
IssuesEvent
2016-08-03 05:05:31
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_container doesn't (can't) easily work on Atomic and similar OS's
bug_report cloud docker feature_idea waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION none ##### OS / ENVIRONMENT ansible - Fedora 23 targets - RHEL Atomic 7.2 ##### SUMMARY docker_container doesn't work for Atomic Hosts (or any OS styled similarly) docker_container looks like it requires docker-py to be installed on the target systems for the plugin to work. You can't (easily) do that in an OS like Atomic Host. I couldn't find an issue to track making this work. Since this is new to 2.1 I'm assuming it's the path forward. Is there a plan to get this to work without target-based dependencies being needed? ##### STEPS TO REPRODUCE ``` - name: create glusterfs server container docker_container: name: gluster-fs image: registry.access.redhat.com/rhgs3/rhgs-server-rhel7 privileged: yes network_mode: host volumes_from: glusterdata volumes: - /mnt/brick1:/mnt/container_brick1[:z] ``` ##### EXPECTED RESULTS the play should run ##### ACTUAL RESULTS the play errors out looking for a dependency that isn't there ``` fatal: [kube2]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to import docker-py - cannot import name NotFound. Try `pip install docker-py`"} ```
True
docker_container doesn't (can't) easily work on Atomic and similar OS's - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION none ##### OS / ENVIRONMENT ansible - Fedora 23 targets - RHEL Atomic 7.2 ##### SUMMARY docker_container doesn't work for Atomic Hosts (or any OS styled similarly) docker_container looks like it requires docker-py to be installed on the target systems for the plugin to work. You can't (easily) do that in an OS like Atomic Host. I couldn't find an issue to track making this work. Since this is new to 2.1 I'm assuming it's the path forward. Is there a plan to get this to work without target-based dependencies being needed? ##### STEPS TO REPRODUCE ``` - name: create glusterfs server container docker_container: name: gluster-fs image: registry.access.redhat.com/rhgs3/rhgs-server-rhel7 privileged: yes network_mode: host volumes_from: glusterdata volumes: - /mnt/brick1:/mnt/container_brick1[:z] ``` ##### EXPECTED RESULTS the play should run ##### ACTUAL RESULTS the play errors out looking for a dependency that isn't there ``` fatal: [kube2]: FAILED! => {"changed": false, "failed": true, "msg": "Failed to import docker-py - cannot import name NotFound. Try `pip install docker-py`"} ```
main
docker container doesn t can t easily work on atomic and similar os s issue type bug report component name docker container ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration none os environment ansible fedora targets rhel atomic summary docker container doesn t work for atomic hosts or any os styled similarly docker container looks like it requires docker py to be installed on the target systems for the plugin to work you can t easily do that in an os like atomic host i couldn t find an issue to track making this work since this is new to i m assuming it s the path forward is there a plan to get this to work without target based dependencies being needed steps to reproduce name create glusterfs server container docker container name gluster fs image registry access redhat com rhgs server privileged yes network mode host volumes from glusterdata volumes mnt mnt container expected results the play should run actual results the play errors out looking for a dependency that isn t there fatal failed changed false failed true msg failed to import docker py cannot import name notfound try pip install docker py
1
37,044
6,565,203,564
IssuesEvent
2017-09-08 07:14:13
minishift/minishift
https://api.github.com/repos/minishift/minishift
opened
Default addons should be specified by name
component/documentation kind/bug priority/major
No defaults addons are specified in the doc, [minishift install addons](https://docs.openshift.org/latest/minishift/command-ref/minishift_addons_install.html) section. IMO it should be something like ``` --defaults If true, installs all Minishift default add-ons (default add-ons anyuid, admin-user, xpaas, registry-route) ```
1.0
Default addons should be specified by name - No defaults addons are specified in the doc, [minishift install addons](https://docs.openshift.org/latest/minishift/command-ref/minishift_addons_install.html) section. IMO it should be something like ``` --defaults If true, installs all Minishift default add-ons (default add-ons anyuid, admin-user, xpaas, registry-route) ```
non_main
default addons should be specified by name no defaults addons are specified in the doc section imo it should be something like defaults if true installs all minishift default add ons default add ons anyuid admin user xpaas registry route
0
299,480
22,608,391,468
IssuesEvent
2022-06-29 15:01:26
gefyrahq/gefyra
https://api.github.com/repos/gefyrahq/gefyra
closed
Add guide for docker desktop
documentation
The getting started guide does only partially work for docker desktop since there is no open port or ingress installed. Some docs for that would probably really help.
1.0
Add guide for docker desktop - The getting started guide does only partially work for docker desktop since there is no open port or ingress installed. Some docs for that would probably really help.
non_main
add guide for docker desktop the getting started guide does only partially work for docker desktop since there is no open port or ingress installed some docs for that would probably really help
0
146,988
11,765,753,295
IssuesEvent
2020-03-14 18:51:45
yobnytech/queueone_issues_tracker
https://api.github.com/repos/yobnytech/queueone_issues_tracker
closed
Customer name not adding correctly while adding a walkin
qone_partner_web_app test and close
When we add a walkin from partner web, the customer name given as "first-name last-name" But it appears in the queue as "first-name last-name first-name last-name"
1.0
Customer name not adding correctly while adding a walkin - When we add a walkin from partner web, the customer name given as "first-name last-name" But it appears in the queue as "first-name last-name first-name last-name"
non_main
customer name not adding correctly while adding a walkin when we add a walkin from partner web the customer name given as first name last name but it appears in the queue as first name last name first name last name
0
1,832
6,577,362,042
IssuesEvent
2017-09-12 00:22:54
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
django_manage createsuperuser does not seem idempotent
affects_2.0 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME django_manage ##### ANSIBLE VERSION ``` ansible 2.0.0.2 ``` ##### OS / ENVIRONMENT OSX El Captain / 10.11.5, managing CentOS 7.1 ##### SUMMARY Issuing command createsuperuser ends up in error after second invocation. ##### STEPS TO REPRODUCE Using this task, hard coding username/mail does not yeald in better effect ``` - name: Django add superusers django_manage: app_path: "{{ seeder_home }}/Seeder" virtualenv: "{{ seeder_virtualenv }}" command: "createsuperuser --noinput --username={{ item.name }} --email={{ item.mail }}" # Not idempotent, probably bug in django_manage module with_items: "{{ seeder_admins }}" tags: django ignore_errors: yes ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS It should not try to create user again ##### ACTUAL RESULTS ``` stderr: Traceback (most recent call last):\n File \"/opt/virtualenv/seeder/lib/python3.4/site-packages/django/db/backends/utils.py\", line 64, in execute\n return self.cursor.execute(sql, params)\npsycopg2.IntegrityError: duplicate key value violates unique constraint \"auth_user_username_key\"\nDETAIL: Key (username)=(rudolf) already exists. ```
True
django_manage createsuperuser does not seem idempotent - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME django_manage ##### ANSIBLE VERSION ``` ansible 2.0.0.2 ``` ##### OS / ENVIRONMENT OSX El Captain / 10.11.5, managing CentOS 7.1 ##### SUMMARY Issuing command createsuperuser ends up in error after second invocation. ##### STEPS TO REPRODUCE Using this task, hard coding username/mail does not yeald in better effect ``` - name: Django add superusers django_manage: app_path: "{{ seeder_home }}/Seeder" virtualenv: "{{ seeder_virtualenv }}" command: "createsuperuser --noinput --username={{ item.name }} --email={{ item.mail }}" # Not idempotent, probably bug in django_manage module with_items: "{{ seeder_admins }}" tags: django ignore_errors: yes ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS It should not try to create user again ##### ACTUAL RESULTS ``` stderr: Traceback (most recent call last):\n File \"/opt/virtualenv/seeder/lib/python3.4/site-packages/django/db/backends/utils.py\", line 64, in execute\n return self.cursor.execute(sql, params)\npsycopg2.IntegrityError: duplicate key value violates unique constraint \"auth_user_username_key\"\nDETAIL: Key (username)=(rudolf) already exists. ```
main
django manage createsuperuser does not seem idempotent issue type bug report component name django manage ansible version ansible os environment osx el captain managing centos summary issuing command createsuperuser ends up in error after second invocation steps to reproduce using this task hard coding username mail does not yeald in better effect name django add superusers django manage app path seeder home seeder virtualenv seeder virtualenv command createsuperuser noinput username item name email item mail not idempotent probably bug in django manage module with items seeder admins tags django ignore errors yes expected results it should not try to create user again actual results stderr traceback most recent call last n file opt virtualenv seeder lib site packages django db backends utils py line in execute n return self cursor execute sql params integrityerror duplicate key value violates unique constraint auth user username key ndetail key username rudolf already exists
1
796,273
28,104,693,935
IssuesEvent
2023-03-30 22:54:29
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Hide "trust token" from devtools
priority/P4 QA/Yes release-notes/exclude OS/Desktop
As we don't support trust token, we shouldn't reveal the item on dev tools Internal thread: https://bravesoftware.slack.com/archives/C7VLGSR55/p1675788916119129 --- ### Steps to reproduce * Open dev tools > "Application" tab * "Trust token" item is visible ### Expected <img width="333" alt="image" src="https://user-images.githubusercontent.com/5474642/217700088-4b20635a-337d-4bf9-8323-1d846346d425.png"> ### Actual <img width="310" alt="Screenshot 2023-02-09 at 11 18 08 AM" src="https://user-images.githubusercontent.com/5474642/217700239-5ea0131b-ede5-48c0-b60a-98a03260d601.png">
1.0
Hide "trust token" from devtools - As we don't support trust token, we shouldn't reveal the item on dev tools Internal thread: https://bravesoftware.slack.com/archives/C7VLGSR55/p1675788916119129 --- ### Steps to reproduce * Open dev tools > "Application" tab * "Trust token" item is visible ### Expected <img width="333" alt="image" src="https://user-images.githubusercontent.com/5474642/217700088-4b20635a-337d-4bf9-8323-1d846346d425.png"> ### Actual <img width="310" alt="Screenshot 2023-02-09 at 11 18 08 AM" src="https://user-images.githubusercontent.com/5474642/217700239-5ea0131b-ede5-48c0-b60a-98a03260d601.png">
non_main
hide trust token from devtools as we don t support trust token we shouldn t reveal the item on dev tools internal thread steps to reproduce open dev tools application tab trust token item is visible expected img width alt image src actual img width alt screenshot at am src
0
9,022
12,530,876,847
IssuesEvent
2020-06-04 13:44:48
ISISScientificComputing/autoreduce
https://api.github.com/repos/ISISScientificComputing/autoreduce
closed
Build.log not picking up errors
👤 Developer Requirement 🔑 Internal
The `build.log` file that is shows the output of building autoreduction using the setup.py commands does not pick up all the error cases correctly. Instances we have seen for this so far include: * Installing icat on windows when python is not in the path` Build.log should return a FULL report of all std out / std err from the console
1.0
Build.log not picking up errors - The `build.log` file that is shows the output of building autoreduction using the setup.py commands does not pick up all the error cases correctly. Instances we have seen for this so far include: * Installing icat on windows when python is not in the path` Build.log should return a FULL report of all std out / std err from the console
non_main
build log not picking up errors the build log file that is shows the output of building autoreduction using the setup py commands does not pick up all the error cases correctly instances we have seen for this so far include installing icat on windows when python is not in the path build log should return a full report of all std out std err from the console
0
18,349
5,624,816,235
IssuesEvent
2017-04-04 17:55:23
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Filters set in com_content stay active in com_menu
No Code Attached Yet
### Steps to reproduce the issue ![comcontent-filters](https://cloud.githubusercontent.com/assets/4589318/23233597/a2d885aa-f94f-11e6-9265-d92cee26e7f8.gif) ### Expected result In com_menu when creating new article have no filters active ### Actual result Filters set prior in com_content are active in com_menu when creating a new article by menu item ### System information (as much as possible) ### Additional comments
1.0
Filters set in com_content stay active in com_menu - ### Steps to reproduce the issue ![comcontent-filters](https://cloud.githubusercontent.com/assets/4589318/23233597/a2d885aa-f94f-11e6-9265-d92cee26e7f8.gif) ### Expected result In com_menu when creating new article have no filters active ### Actual result Filters set prior in com_content are active in com_menu when creating a new article by menu item ### System information (as much as possible) ### Additional comments
non_main
filters set in com content stay active in com menu steps to reproduce the issue expected result in com menu when creating new article have no filters active actual result filters set prior in com content are active in com menu when creating a new article by menu item system information as much as possible additional comments
0
3,967
18,151,844,822
IssuesEvent
2021-09-26 11:58:58
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Label pull requests based on the labels of the issues they close
enhancement maintainability
We do not use labels much on PRs so far, but it could be useful to do that: - to help reviewers find PRs they could be interested in - to make up for poor PR titles which do not describe their contents well - to make it possible to draft release notes automatically from the Git commit log (#2359) Since we are already tagging issues, and pull requests are normally linked to the issues they solve, we should be able to automatically tag PRs by copying those tags.
True
Label pull requests based on the labels of the issues they close - We do not use labels much on PRs so far, but it could be useful to do that: - to help reviewers find PRs they could be interested in - to make up for poor PR titles which do not describe their contents well - to make it possible to draft release notes automatically from the Git commit log (#2359) Since we are already tagging issues, and pull requests are normally linked to the issues they solve, we should be able to automatically tag PRs by copying those tags.
main
label pull requests based on the labels of the issues they close we do not use labels much on prs so far but it could be useful to do that to help reviewers find prs they could be interested in to make up for poor pr titles which do not describe their contents well to make it possible to draft release notes automatically from the git commit log since we are already tagging issues and pull requests are normally linked to the issues they solve we should be able to automatically tag prs by copying those tags
1
16,457
21,335,642,785
IssuesEvent
2022-04-18 14:16:36
lbryio/scribe
https://api.github.com/repos/lbryio/scribe
opened
MemoryError (OOM) in advance block
type: bug area: block processor
```2022-04-18 12:14:38,547 ERROR scribe.service.BlockchainProcessorService:1629: error in block processor loop: Traceback (most recent call last): File "/home/lbry/scribe/blockchain/service.py", line 1625, in process_blocks_and_mempool_forever raise err File "/home/lbry/scribe/blockchain/service.py", line 1620, in process_blocks_and_mempool_forever await self.check_and_advance_blocks(blocks) File "/home/lbry/scribe/blockchain/service.py", line 212, in check_and_advance_blocks txo_count = await self.run_in_thread_with_lock(self.advance_block, block) File "/home/lbry/scribe/blockchain/service.py", line 140, in run_in_thread_with_lock return await asyncio.shield(run_in_thread_locked()) File "/home/lbry/scribe/blockchain/service.py", line 139, in run_in_thread_locked return await asyncio.get_event_loop().run_in_executor(self._executor, func, *args) File "/usr/lib/python3.9/concurrent/futures/thread.py", line 52, in run result = self.fn(*self.args, **self.kwargs) File "/home/lbry/scribe/blockchain/service.py", line 1441, in advance_block self.db.prefix_db.commit(self.height, self.tip) File "/home/lbry/scribe/db/interface.py", line 209, in commit undo_ops = self._op_stack.get_undo_ops() File "/home/lbry/scribe/db/revertable.py", line 163, in get_undo_ops return b''.join(op.invert().pack() for op in reversed(self)) MemoryError ``` This error can be recovered from by restarting scribe.
1.0
MemoryError (OOM) in advance block - ```2022-04-18 12:14:38,547 ERROR scribe.service.BlockchainProcessorService:1629: error in block processor loop: Traceback (most recent call last): File "/home/lbry/scribe/blockchain/service.py", line 1625, in process_blocks_and_mempool_forever raise err File "/home/lbry/scribe/blockchain/service.py", line 1620, in process_blocks_and_mempool_forever await self.check_and_advance_blocks(blocks) File "/home/lbry/scribe/blockchain/service.py", line 212, in check_and_advance_blocks txo_count = await self.run_in_thread_with_lock(self.advance_block, block) File "/home/lbry/scribe/blockchain/service.py", line 140, in run_in_thread_with_lock return await asyncio.shield(run_in_thread_locked()) File "/home/lbry/scribe/blockchain/service.py", line 139, in run_in_thread_locked return await asyncio.get_event_loop().run_in_executor(self._executor, func, *args) File "/usr/lib/python3.9/concurrent/futures/thread.py", line 52, in run result = self.fn(*self.args, **self.kwargs) File "/home/lbry/scribe/blockchain/service.py", line 1441, in advance_block self.db.prefix_db.commit(self.height, self.tip) File "/home/lbry/scribe/db/interface.py", line 209, in commit undo_ops = self._op_stack.get_undo_ops() File "/home/lbry/scribe/db/revertable.py", line 163, in get_undo_ops return b''.join(op.invert().pack() for op in reversed(self)) MemoryError ``` This error can be recovered from by restarting scribe.
non_main
memoryerror oom in advance block error scribe service blockchainprocessorservice error in block processor loop traceback most recent call last file home lbry scribe blockchain service py line in process blocks and mempool forever raise err file home lbry scribe blockchain service py line in process blocks and mempool forever await self check and advance blocks blocks file home lbry scribe blockchain service py line in check and advance blocks txo count await self run in thread with lock self advance block block file home lbry scribe blockchain service py line in run in thread with lock return await asyncio shield run in thread locked file home lbry scribe blockchain service py line in run in thread locked return await asyncio get event loop run in executor self executor func args file usr lib concurrent futures thread py line in run result self fn self args self kwargs file home lbry scribe blockchain service py line in advance block self db prefix db commit self height self tip file home lbry scribe db interface py line in commit undo ops self op stack get undo ops file home lbry scribe db revertable py line in get undo ops return b join op invert pack for op in reversed self memoryerror this error can be recovered from by restarting scribe
0
5,072
25,960,182,761
IssuesEvent
2022-12-18 19:59:34
cran-task-views/Hydrology
https://api.github.com/repos/cran-task-views/Hydrology
closed
Package 'EcoHydRology' has been archived on CRAN for more than 60 days
maintainer-contacted to-be-removed
Package [EcoHydRology](https://CRAN.R-project.org/package=EcoHydRology) is currently listed in CRAN Task View [Hydrology](https://CRAN.R-project.org/view=Hydrology) but the package has actually been archived for more than 60 days on CRAN. Often this indicates that the package is currently not sufficiently actively maintained and should be excluded from the task view. Alternatively, you might also consider reaching out to the authors of the package and encourage (or even help) them to bring the package back to CRAN. In any case, the situation should be resolved in the next four weeks. If the package does not seem to be brought back to CRAN, please exclude it from the task view.
True
Package 'EcoHydRology' has been archived on CRAN for more than 60 days - Package [EcoHydRology](https://CRAN.R-project.org/package=EcoHydRology) is currently listed in CRAN Task View [Hydrology](https://CRAN.R-project.org/view=Hydrology) but the package has actually been archived for more than 60 days on CRAN. Often this indicates that the package is currently not sufficiently actively maintained and should be excluded from the task view. Alternatively, you might also consider reaching out to the authors of the package and encourage (or even help) them to bring the package back to CRAN. In any case, the situation should be resolved in the next four weeks. If the package does not seem to be brought back to CRAN, please exclude it from the task view.
main
package ecohydrology has been archived on cran for more than days package is currently listed in cran task view but the package has actually been archived for more than days on cran often this indicates that the package is currently not sufficiently actively maintained and should be excluded from the task view alternatively you might also consider reaching out to the authors of the package and encourage or even help them to bring the package back to cran in any case the situation should be resolved in the next four weeks if the package does not seem to be brought back to cran please exclude it from the task view
1
4,668
24,133,931,384
IssuesEvent
2022-09-21 09:41:15
beyarkay/eskom-calendar
https://api.github.com/repos/beyarkay/eskom-calendar
closed
City Power schedule issue
bug waiting-on-maintainer missing-area-schedule
Hi there, Thank you for a great product. It seems there is a discrepancy for city-power-11 or I am doing something wrong. If I compare CityPowers schedule, there are four blocks of load shedding for this date, but only one block appears in the calendar. https://www.citypower.co.za/customers/Pages/Load_Shedding_Downloads.aspx ![image](https://user-images.githubusercontent.com/7857734/191454315-ba3942bf-9e45-473a-8710-59919a66be63.png) ![image](https://user-images.githubusercontent.com/7857734/191454362-96962e00-f7a2-4362-8b6a-20eb5a8145d1.png) Is there anything I can do?
True
City Power schedule issue - Hi there, Thank you for a great product. It seems there is a discrepancy for city-power-11 or I am doing something wrong. If I compare CityPowers schedule, there are four blocks of load shedding for this date, but only one block appears in the calendar. https://www.citypower.co.za/customers/Pages/Load_Shedding_Downloads.aspx ![image](https://user-images.githubusercontent.com/7857734/191454315-ba3942bf-9e45-473a-8710-59919a66be63.png) ![image](https://user-images.githubusercontent.com/7857734/191454362-96962e00-f7a2-4362-8b6a-20eb5a8145d1.png) Is there anything I can do?
main
city power schedule issue hi there thank you for a great product it seems there is a discrepancy for city power or i am doing something wrong if i compare citypowers schedule there are four blocks of load shedding for this date but only one block appears in the calendar is there anything i can do
1
151,493
5,821,167,834
IssuesEvent
2017-05-06 02:28:17
dhowe/AdNauseam
https://api.github.com/repos/dhowe/AdNauseam
closed
Slider goes back to the default position after observing ad details
PRIORITY: Medium
To reproduce: 1.Open the vault 2.Change the slider 3.Click an ad to observe the detail 4.Back to the vault window, the slider is in full range again
1.0
Slider goes back to the default position after observing ad details - To reproduce: 1.Open the vault 2.Change the slider 3.Click an ad to observe the detail 4.Back to the vault window, the slider is in full range again
non_main
slider goes back to the default position after observing ad details to reproduce open the vault change the slider click an ad to observe the detail back to the vault window the slider is in full range again
0
5,091
7,009,700,844
IssuesEvent
2017-12-19 20:09:05
PATRIC3/patric3_website
https://api.github.com/repos/PATRIC3/patric3_website
opened
Genome annotation needs more checking for genbank data
Service: Annotation
If one submits a genbank file that is missing org name you get mysterious crashes. This should be checked in the new preflight tool.
1.0
Genome annotation needs more checking for genbank data - If one submits a genbank file that is missing org name you get mysterious crashes. This should be checked in the new preflight tool.
non_main
genome annotation needs more checking for genbank data if one submits a genbank file that is missing org name you get mysterious crashes this should be checked in the new preflight tool
0
2,910
10,345,870,478
IssuesEvent
2019-09-04 14:16:31
kensho-technologies/graphql-compiler
https://api.github.com/repos/kensho-technologies/graphql-compiler
opened
The "slow" pytest mark is not properly defined and generates a warning
maintainer quality-of-life
Running tests with `pytest` generates the below: ```PytestUnknownMarkWarning: Unknown pytest.mark.slow - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/latest/mark.html```
True
The "slow" pytest mark is not properly defined and generates a warning - Running tests with `pytest` generates the below: ```PytestUnknownMarkWarning: Unknown pytest.mark.slow - is this a typo? You can register custom marks to avoid this warning - for details, see https://docs.pytest.org/en/latest/mark.html```
main
the slow pytest mark is not properly defined and generates a warning running tests with pytest generates the below pytestunknownmarkwarning unknown pytest mark slow is this a typo you can register custom marks to avoid this warning for details see
1
124,474
17,772,575,486
IssuesEvent
2021-08-30 15:12:41
kapseliboi/energy-futures-vis-avenir-energetique
https://api.github.com/repos/kapseliboi/energy-futures-vis-avenir-energetique
opened
CVE-2020-7789 (Medium) detected in node-notifier-5.4.2.tgz
security vulnerability
## CVE-2020-7789 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-notifier-5.4.2.tgz</b></p></summary> <p>A Node.js module for sending notifications on native Mac, Windows (post and pre 8) and Linux (or Growl as fallback)</p> <p>Library home page: <a href="https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.2.tgz">https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.2.tgz</a></p> <p>Path to dependency file: energy-futures-vis-avenir-energetique/package.json</p> <p>Path to vulnerable library: energy-futures-vis-avenir-energetique/node_modules/node-notifier/package.json</p> <p> Dependency Hierarchy: - jest-24.9.0.tgz (Root Library) - jest-cli-24.9.0.tgz - core-24.9.0.tgz - reporters-24.9.0.tgz - :x: **node-notifier-5.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/energy-futures-vis-avenir-energetique/commit/907b3c15edb7159764857453edc4f32b2432cdd4">907b3c15edb7159764857453edc4f32b2432cdd4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package node-notifier before 9.0.0. It allows an attacker to run arbitrary commands on Linux machines due to the options params not being sanitised when being passed an array. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789>CVE-2020-7789</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: 9.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7789 (Medium) detected in node-notifier-5.4.2.tgz - ## CVE-2020-7789 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-notifier-5.4.2.tgz</b></p></summary> <p>A Node.js module for sending notifications on native Mac, Windows (post and pre 8) and Linux (or Growl as fallback)</p> <p>Library home page: <a href="https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.2.tgz">https://registry.npmjs.org/node-notifier/-/node-notifier-5.4.2.tgz</a></p> <p>Path to dependency file: energy-futures-vis-avenir-energetique/package.json</p> <p>Path to vulnerable library: energy-futures-vis-avenir-energetique/node_modules/node-notifier/package.json</p> <p> Dependency Hierarchy: - jest-24.9.0.tgz (Root Library) - jest-cli-24.9.0.tgz - core-24.9.0.tgz - reporters-24.9.0.tgz - :x: **node-notifier-5.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/energy-futures-vis-avenir-energetique/commit/907b3c15edb7159764857453edc4f32b2432cdd4">907b3c15edb7159764857453edc4f32b2432cdd4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package node-notifier before 9.0.0. It allows an attacker to run arbitrary commands on Linux machines due to the options params not being sanitised when being passed an array. <p>Publish Date: 2020-12-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789>CVE-2020-7789</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7789</a></p> <p>Release Date: 2020-12-11</p> <p>Fix Resolution: 9.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in node notifier tgz cve medium severity vulnerability vulnerable library node notifier tgz a node js module for sending notifications on native mac windows post and pre and linux or growl as fallback library home page a href path to dependency file energy futures vis avenir energetique package json path to vulnerable library energy futures vis avenir energetique node modules node notifier package json dependency hierarchy jest tgz root library jest cli tgz core tgz reporters tgz x node notifier tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package node notifier before it allows an attacker to run arbitrary commands on linux machines due to the options params not being sanitised when being passed an array publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
104,956
13,147,938,716
IssuesEvent
2020-08-08 18:26:52
mrkirkmorgan/song-builder
https://api.github.com/repos/mrkirkmorgan/song-builder
opened
User Authentication Process
design rollout
This ticket covers creating an authentication process by which users can log into their account and access their information/music.
1.0
User Authentication Process - This ticket covers creating an authentication process by which users can log into their account and access their information/music.
non_main
user authentication process this ticket covers creating an authentication process by which users can log into their account and access their information music
0
2,075
7,029,374,999
IssuesEvent
2017-12-25 23:22:00
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
We have two competing standards for decks of cards
Maintainability/Hinders improvements
![avatar_134682b6a0bb_128](https://cloud.githubusercontent.com/assets/4176358/18365720/cad6836e-75e1-11e6-93c9-c647e58732c0.png) From left to right: /obj/item/toy/cards/deck, /obj/item/weapon/deck, and /obj/item/toy/cards/deck/cas All offer benefits and draw backs, but it's kinda confusing we have both of these.
True
We have two competing standards for decks of cards - ![avatar_134682b6a0bb_128](https://cloud.githubusercontent.com/assets/4176358/18365720/cad6836e-75e1-11e6-93c9-c647e58732c0.png) From left to right: /obj/item/toy/cards/deck, /obj/item/weapon/deck, and /obj/item/toy/cards/deck/cas All offer benefits and draw backs, but it's kinda confusing we have both of these.
main
we have two competing standards for decks of cards from left to right obj item toy cards deck obj item weapon deck and obj item toy cards deck cas all offer benefits and draw backs but it s kinda confusing we have both of these
1
5,458
27,301,825,799
IssuesEvent
2023-02-24 03:06:16
Bernz322/mfe-core
https://api.github.com/repos/Bernz322/mfe-core
opened
[🐛 BUG] : Translation with i18next does not work in production
bug help wanted maintainers p2
**Describe the bug** i18next translation does not work in production but working in dev environment. **To Reproduce** Steps to reproduce the behavior: 1. Go to [hosted site](https://microfrontend-core-prod.vercel.app). 2. Click on 'Buttons with different languages'. 3. Rendered typography does not change languages. 4. See console logs with error 'i18next::translator: missingKey en/de/jp translation' **Expected behavior** When a specific language button is clicked, the typography present in the page should be translated to the specific language chosen. **Screenshots** - In Production: ![image](https://user-images.githubusercontent.com/49061484/221081915-adde6eab-48fa-4231-995a-93a27278c31b.png) ![image](https://user-images.githubusercontent.com/49061484/221081927-25c01ae8-447a-4fd5-9d0e-d314ac697bf2.png) - In Dev ![image](https://user-images.githubusercontent.com/49061484/221081871-14f42372-22d7-4e86-9c29-5c3138419077.png) ![image](https://user-images.githubusercontent.com/49061484/221081879-9791bc61-2178-4332-8755-b5422ed3b627.png) ![image](https://user-images.githubusercontent.com/49061484/221081886-10012b98-0db3-4b13-8aea-e1048182ad3c.png) **Desktop (please complete the following information):** - OS: [Windows 10] - Browser [Opera, Chrome, Edge] - Version [All mentioned browsers are up to the latest version to this writing] **Additional context** In dev, all translation.jsson is fetched and can be navigated to the specific file path, however in production, it is not being fetched. I suspect that the project it is not being built properly and translation.json are not bundled with it.
True
[🐛 BUG] : Translation with i18next does not work in production - **Describe the bug** i18next translation does not work in production but working in dev environment. **To Reproduce** Steps to reproduce the behavior: 1. Go to [hosted site](https://microfrontend-core-prod.vercel.app). 2. Click on 'Buttons with different languages'. 3. Rendered typography does not change languages. 4. See console logs with error 'i18next::translator: missingKey en/de/jp translation' **Expected behavior** When a specific language button is clicked, the typography present in the page should be translated to the specific language chosen. **Screenshots** - In Production: ![image](https://user-images.githubusercontent.com/49061484/221081915-adde6eab-48fa-4231-995a-93a27278c31b.png) ![image](https://user-images.githubusercontent.com/49061484/221081927-25c01ae8-447a-4fd5-9d0e-d314ac697bf2.png) - In Dev ![image](https://user-images.githubusercontent.com/49061484/221081871-14f42372-22d7-4e86-9c29-5c3138419077.png) ![image](https://user-images.githubusercontent.com/49061484/221081879-9791bc61-2178-4332-8755-b5422ed3b627.png) ![image](https://user-images.githubusercontent.com/49061484/221081886-10012b98-0db3-4b13-8aea-e1048182ad3c.png) **Desktop (please complete the following information):** - OS: [Windows 10] - Browser [Opera, Chrome, Edge] - Version [All mentioned browsers are up to the latest version to this writing] **Additional context** In dev, all translation.jsson is fetched and can be navigated to the specific file path, however in production, it is not being fetched. I suspect that the project it is not being built properly and translation.json are not bundled with it.
main
translation with does not work in production describe the bug translation does not work in production but working in dev environment to reproduce steps to reproduce the behavior go to click on buttons with different languages rendered typography does not change languages see console logs with error translator missingkey en de jp translation expected behavior when a specific language button is clicked the typography present in the page should be translated to the specific language chosen screenshots in production in dev desktop please complete the following information os browser version additional context in dev all translation jsson is fetched and can be navigated to the specific file path however in production it is not being fetched i suspect that the project it is not being built properly and translation json are not bundled with it
1
1,589
6,572,368,261
IssuesEvent
2017-09-11 01:46:30
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
iam_cert module occasionally returns BotoServerError 404 NotFound
affects_2.3 aws bug_report cloud waiting_on_maintainer
As part of our automation testing environment we create and destroy IAM server certs in rapid succession. Occasionally we encounter a race condition where a server certificate is present on AWS at the time that IAMConnection.get_all_server_certs() is executed. But the certificate is destroyed on AWS just before IAMConnection.get_server_certificate(certificate-name-that-no-longer-exists) is executed. This results in the following error: ``` boto.exception.BotoServerError: BotoServerError: 404 Not Found Sender NoSuchEntity The Server Certificate with name certificate-that-no-longer-exists cannot be found. ```
True
iam_cert module occasionally returns BotoServerError 404 NotFound - As part of our automation testing environment we create and destroy IAM server certs in rapid succession. Occasionally we encounter a race condition where a server certificate is present on AWS at the time that IAMConnection.get_all_server_certs() is executed. But the certificate is destroyed on AWS just before IAMConnection.get_server_certificate(certificate-name-that-no-longer-exists) is executed. This results in the following error: ``` boto.exception.BotoServerError: BotoServerError: 404 Not Found Sender NoSuchEntity The Server Certificate with name certificate-that-no-longer-exists cannot be found. ```
main
iam cert module occasionally returns botoservererror notfound as part of our automation testing environment we create and destroy iam server certs in rapid succession occasionally we encounter a race condition where a server certificate is present on aws at the time that iamconnection get all server certs is executed but the certificate is destroyed on aws just before iamconnection get server certificate certificate name that no longer exists is executed this results in the following error boto exception botoservererror botoservererror not found sender nosuchentity the server certificate with name certificate that no longer exists cannot be found
1
3,753
15,787,230,923
IssuesEvent
2021-04-01 18:54:39
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
Unreadable text on InlineNotification
status: needs more info 🤷‍♀️ status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬 type: bug 🐛
## What package(s) are you using? - [x ] `carbon-components 10.29.0` - [ x] `carbon-components-react 7.29.0` ## Detailed description Text of `InlineNotification` in darkmode is unreadable when using `lowContrast` mode. ## Additional information dark mode ![image](https://user-images.githubusercontent.com/43094706/112666151-2727fe80-8e5c-11eb-8260-79e2f29105c3.png) light mode ![image](https://user-images.githubusercontent.com/43094706/112666213-3909a180-8e5c-11eb-9305-119ea772fec2.png)
True
Unreadable text on InlineNotification - ## What package(s) are you using? - [x ] `carbon-components 10.29.0` - [ x] `carbon-components-react 7.29.0` ## Detailed description Text of `InlineNotification` in darkmode is unreadable when using `lowContrast` mode. ## Additional information dark mode ![image](https://user-images.githubusercontent.com/43094706/112666151-2727fe80-8e5c-11eb-8260-79e2f29105c3.png) light mode ![image](https://user-images.githubusercontent.com/43094706/112666213-3909a180-8e5c-11eb-9305-119ea772fec2.png)
main
unreadable text on inlinenotification what package s are you using carbon components carbon components react detailed description text of inlinenotification in darkmode is unreadable when using lowcontrast mode additional information dark mode light mode
1
4,053
18,946,436,660
IssuesEvent
2021-11-18 10:39:58
perun-network/poly-go
https://api.github.com/repos/perun-network/poly-go
closed
Update linter config
CI/CD :construction_worker_man: quality/maintainability :gem:
Update the linter config to better reflect our current programming style. Analogous to https://github.com/hyperledger-labs/go-perun/pull/260.
True
Update linter config - Update the linter config to better reflect our current programming style. Analogous to https://github.com/hyperledger-labs/go-perun/pull/260.
main
update linter config update the linter config to better reflect our current programming style analogous to
1
3,400
13,181,782,126
IssuesEvent
2020-08-12 14:49:22
duo-labs/cloudmapper
https://api.github.com/repos/duo-labs/cloudmapper
closed
Support adding notes to objects in mapview
map unmaintained_functionality
As a user, it would be of great help if there is an option to enter notes for an object that can be stored as a tag (add new, edit existing) if the user has appropriate permissions.
True
Support adding notes to objects in mapview - As a user, it would be of great help if there is an option to enter notes for an object that can be stored as a tag (add new, edit existing) if the user has appropriate permissions.
main
support adding notes to objects in mapview as a user it would be of great help if there is an option to enter notes for an object that can be stored as a tag add new edit existing if the user has appropriate permissions
1
436,353
30,547,476,157
IssuesEvent
2023-07-20 05:50:24
owncloud/ocis-charts
https://api.github.com/repos/owncloud/ocis-charts
opened
secrets documentation
Priority:p2-high Topic:Documentation
https://doc.owncloud.com/ocis/next/deployment/container/orchestration/orchestration.html#secrets does not reflect the current situation and needs to be updated. There was a major break in the way how this chart can be used because it now generates secrets automatically
1.0
secrets documentation - https://doc.owncloud.com/ocis/next/deployment/container/orchestration/orchestration.html#secrets does not reflect the current situation and needs to be updated. There was a major break in the way how this chart can be used because it now generates secrets automatically
non_main
secrets documentation does not reflect the current situation and needs to be updated there was a major break in the way how this chart can be used because it now generates secrets automatically
0
47,423
10,091,167,078
IssuesEvent
2019-07-26 13:38:26
graphql/graphiql
https://api.github.com/repos/graphql/graphiql
closed
No intellisense for __typename
codemirror-graphql enhancement
If I want to use `__typename` _on any Type_ in a query, it doesn't show up in the 'intellisense-autocomplete-popup'. This is inconsistent, because top-level introspection fields like `__type` and `__schema` do show up in the popup. ![typename](https://user-images.githubusercontent.com/852069/27988025-908af7a0-6418-11e7-985c-b488bcbdd271.gif)
1.0
No intellisense for __typename - If I want to use `__typename` _on any Type_ in a query, it doesn't show up in the 'intellisense-autocomplete-popup'. This is inconsistent, because top-level introspection fields like `__type` and `__schema` do show up in the popup. ![typename](https://user-images.githubusercontent.com/852069/27988025-908af7a0-6418-11e7-985c-b488bcbdd271.gif)
non_main
no intellisense for typename if i want to use typename on any type in a query it doesn t show up in the intellisense autocomplete popup this is inconsistent because top level introspection fields like type and schema do show up in the popup
0
1,130
4,998,415,563
IssuesEvent
2016-12-09 19:47:07
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
replace: TypeError when run under Python 3
affects_2.2 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - files/replace ##### ANSIBLE VERSION ``` ansible 2.2.0.0 (detached HEAD cdec853e37) last updated 2016/12/01 10:23:30 (GMT +200) lib/ansible/modules/core: (detached HEAD fe9c56a003) last updated 2016/12/01 10:24:42 (GMT +200) lib/ansible/modules/extras: (detached HEAD f564e77a08) last updated 2016/12/01 10:24:42 (GMT +200) config file = configured module search path = ['/Users/per/Projects/servers/submodules/ansible/library'] ``` ##### CONFIGURATION - ansible_python_interpreter=/usr/bin/python3 ##### OS / ENVIRONMENT - Local: MacOS - Remote: Ubuntu 16.04, with Python 3.5.2 ##### SUMMARY Replace fails with a TypeError when run under Python 3 ##### STEPS TO REPRODUCE ``` ansible *** -m replace \ -a "dest=/etc/lsb-release regexp=nomatchfound replace=nomatchfound" ``` ##### EXPECTED RESULTS No replacement (since regex doesn't match). ##### ACTUAL RESULTS Command fails with error: `TypeError: cannot use a string pattern on a bytes-like object` ``` *** | FAILED! => { "changed": false, "failed": true, "module_stderr": "Shared connection to *** closed.\r\n", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 180, in <module>\r\n main()\r\n File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 153, in main\r\n result = re.subn(mre, params['replace'], contents, 0)\r\n File \"/usr/lib/python3.5/re.py\", line 193, in subn\r\n return _compile(pattern, flags).subn(repl, string, count)\r\nTypeError: cannot use a string pattern on a bytes-like object\r\n", "msg": "MODULE FAILURE" } ``` Readable traceback (from above): ``` Traceback (most recent call last): File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 180, in <module> main() File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 153, in main result = re.subn(mre, params['replace'], contents, 0) File \"/usr/lib/python3.5/re.py\", line 193, in subn return _compile(pattern, flags).subn(repl, string, count) TypeError: cannot use a string pattern on a bytes-like object ```
True
replace: TypeError when run under Python 3 - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME - files/replace ##### ANSIBLE VERSION ``` ansible 2.2.0.0 (detached HEAD cdec853e37) last updated 2016/12/01 10:23:30 (GMT +200) lib/ansible/modules/core: (detached HEAD fe9c56a003) last updated 2016/12/01 10:24:42 (GMT +200) lib/ansible/modules/extras: (detached HEAD f564e77a08) last updated 2016/12/01 10:24:42 (GMT +200) config file = configured module search path = ['/Users/per/Projects/servers/submodules/ansible/library'] ``` ##### CONFIGURATION - ansible_python_interpreter=/usr/bin/python3 ##### OS / ENVIRONMENT - Local: MacOS - Remote: Ubuntu 16.04, with Python 3.5.2 ##### SUMMARY Replace fails with a TypeError when run under Python 3 ##### STEPS TO REPRODUCE ``` ansible *** -m replace \ -a "dest=/etc/lsb-release regexp=nomatchfound replace=nomatchfound" ``` ##### EXPECTED RESULTS No replacement (since regex doesn't match). ##### ACTUAL RESULTS Command fails with error: `TypeError: cannot use a string pattern on a bytes-like object` ``` *** | FAILED! => { "changed": false, "failed": true, "module_stderr": "Shared connection to *** closed.\r\n", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 180, in <module>\r\n main()\r\n File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 153, in main\r\n result = re.subn(mre, params['replace'], contents, 0)\r\n File \"/usr/lib/python3.5/re.py\", line 193, in subn\r\n return _compile(pattern, flags).subn(repl, string, count)\r\nTypeError: cannot use a string pattern on a bytes-like object\r\n", "msg": "MODULE FAILURE" } ``` Readable traceback (from above): ``` Traceback (most recent call last): File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 180, in <module> main() File \"/tmp/ansible_1wjonocj/ansible_module_replace.py\", line 153, in main result = re.subn(mre, params['replace'], contents, 0) File \"/usr/lib/python3.5/re.py\", line 193, in subn return _compile(pattern, flags).subn(repl, string, count) TypeError: cannot use a string pattern on a bytes-like object ```
main
replace typeerror when run under python issue type bug report component name files replace ansible version ansible detached head last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file configured module search path configuration ansible python interpreter usr bin os environment local macos remote ubuntu with python summary replace fails with a typeerror when run under python steps to reproduce ansible m replace a dest etc lsb release regexp nomatchfound replace nomatchfound expected results no replacement since regex doesn t match actual results command fails with error typeerror cannot use a string pattern on a bytes like object failed changed false failed true module stderr shared connection to closed r n module stdout traceback most recent call last r n file tmp ansible ansible module replace py line in r n main r n file tmp ansible ansible module replace py line in main r n result re subn mre params contents r n file usr lib re py line in subn r n return compile pattern flags subn repl string count r ntypeerror cannot use a string pattern on a bytes like object r n msg module failure readable traceback from above traceback most recent call last file tmp ansible ansible module replace py line in main file tmp ansible ansible module replace py line in main result re subn mre params contents file usr lib re py line in subn return compile pattern flags subn repl string count typeerror cannot use a string pattern on a bytes like object
1
37,257
8,244,822,763
IssuesEvent
2018-09-11 07:47:48
Chisel-Team/Chisel
https://api.github.com/repos/Chisel-Team/Chisel
closed
[1.12] java.lang.NullPointerException - when disabling blocks in config
bug-code complete-code
Forge: 14.23.2.2623 Chisel: 0.2.0.31 Crash Report: https://pastebin.com/XJwyqE6C
2.0
[1.12] java.lang.NullPointerException - when disabling blocks in config - Forge: 14.23.2.2623 Chisel: 0.2.0.31 Crash Report: https://pastebin.com/XJwyqE6C
non_main
java lang nullpointerexception when disabling blocks in config forge chisel crash report
0
13,121
8,798,225,060
IssuesEvent
2018-12-24 05:53:02
amaredeus/brianpatino.venfo.xyz
https://api.github.com/repos/amaredeus/brianpatino.venfo.xyz
closed
Fix Bootstrap vulnerability
security
Address bootstrap identified vulnerability CVE-2018-14041 (https://nvd.nist.gov/vuln/detail/CVE-2018-14041) Vulnerable versions: < 4.1.2 Patched version: 4.1.2 In Bootstrap before 4.1.2, XSS is possible in the data-target property of scrollspy. This is similar to CVE-2018-14042.
True
Fix Bootstrap vulnerability - Address bootstrap identified vulnerability CVE-2018-14041 (https://nvd.nist.gov/vuln/detail/CVE-2018-14041) Vulnerable versions: < 4.1.2 Patched version: 4.1.2 In Bootstrap before 4.1.2, XSS is possible in the data-target property of scrollspy. This is similar to CVE-2018-14042.
non_main
fix bootstrap vulnerability address bootstrap identified vulnerability cve vulnerable versions patched version in bootstrap before xss is possible in the data target property of scrollspy this is similar to cve
0
786
4,389,629,573
IssuesEvent
2016-08-08 22:52:48
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Feature Idea: Amazon AWS IAM module, set password policy
aws cloud feature_idea waiting_on_maintainer
##### Issue Type: - Feature Idea ##### Plugin Name: IAM policy plugin ##### Ansible Version: ``` ansible 2.0.1.0 ``` ##### Ansible Configuration: ##### Environment: Using: Ubuntu 15.10 Wily Werewolf ##### Summary: I would like to set the Amazon AWS IAM password policy, using Ansible. http://docs.aws.amazon.com/cli/latest/reference/iam/update-account-password-policy.html ##### Steps To Reproduce: ##### Expected Results: Amazon AWS IAM password policy will be updated. ##### Actual Results:
True
Feature Idea: Amazon AWS IAM module, set password policy - ##### Issue Type: - Feature Idea ##### Plugin Name: IAM policy plugin ##### Ansible Version: ``` ansible 2.0.1.0 ``` ##### Ansible Configuration: ##### Environment: Using: Ubuntu 15.10 Wily Werewolf ##### Summary: I would like to set the Amazon AWS IAM password policy, using Ansible. http://docs.aws.amazon.com/cli/latest/reference/iam/update-account-password-policy.html ##### Steps To Reproduce: ##### Expected Results: Amazon AWS IAM password policy will be updated. ##### Actual Results:
main
feature idea amazon aws iam module set password policy issue type feature idea plugin name iam policy plugin ansible version ansible ansible configuration environment using ubuntu wily werewolf summary i would like to set the amazon aws iam password policy using ansible steps to reproduce expected results amazon aws iam password policy will be updated actual results
1