Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
957
labels
stringlengths
4
795
body
stringlengths
1
259k
index
stringclasses
12 values
text_combine
stringlengths
96
259k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
461,166
13,224,536,776
IssuesEvent
2020-08-17 19:20:16
grey-software/Twitter-Focus
https://api.github.com/repos/grey-software/Twitter-Focus
opened
Inprove the README
Domain: Dev Experience Priority: Medium Type: Maintenance
Consult https://github.com/grey-software/LinkedInFocus - [ ] Add webstore link - [ ] Add local dev instructions - [ ] Add screenshots or gif - [ ] Add logo
1.0
Inprove the README - Consult https://github.com/grey-software/LinkedInFocus - [ ] Add webstore link - [ ] Add local dev instructions - [ ] Add screenshots or gif - [ ] Add logo
priority
inprove the readme consult add webstore link add local dev instructions add screenshots or gif add logo
1
146,794
5,627,950,302
IssuesEvent
2017-04-05 03:58:50
dnGrep/dnGrep
https://api.github.com/repos/dnGrep/dnGrep
closed
no supports for gb2312/utf8 encoding?
bug imported Priority-Medium
_From [mfm...@sina.com](https://code.google.com/u/107484051280825188802/) on February 25, 2013 23:15:51_ it turns out to be no supports for greping gb2312/utf8 encoded text. 1. dngrep did find the occurs, but linenos and positions of the result are very wrong 2. chinese characters displayed in result pane are unreadable, while they are good in preview pane. 3. converting gb2312/utf8 encoded text file to unicode encoding will solve this problem, but it is a tough work and not feasible 3. env: dngrep version 2.7.1, win7 x64(simplified chinese) _Original issue: http://code.google.com/p/dngrep/issues/detail?id=177_
1.0
no supports for gb2312/utf8 encoding? - _From [mfm...@sina.com](https://code.google.com/u/107484051280825188802/) on February 25, 2013 23:15:51_ it turns out to be no supports for greping gb2312/utf8 encoded text. 1. dngrep did find the occurs, but linenos and positions of the result are very wrong 2. chinese characters displayed in result pane are unreadable, while they are good in preview pane. 3. converting gb2312/utf8 encoded text file to unicode encoding will solve this problem, but it is a tough work and not feasible 3. env: dngrep version 2.7.1, win7 x64(simplified chinese) _Original issue: http://code.google.com/p/dngrep/issues/detail?id=177_
priority
no supports for encoding from on february it turns out to be no supports for greping encoded text dngrep did find the occurs but linenos and positions of the result are very wrong chinese characters displayed in result pane are unreadable while they are good in preview pane converting encoded text file to unicode encoding will solve this problem but it is a tough work and not feasible env dngrep version simplified chinese original issue
1
405,428
11,873,120,850
IssuesEvent
2020-03-26 16:49:01
netdata/netdata
https://api.github.com/repos/netdata/netdata
closed
During the agent installation, if the ACLK fails to be built, show an error message to the user
ACLK internal priority/high priority/medium
#### Summary Ensure that the requirement in product#282 are met. - [x] If ACLK error reporting not set yet by #8051 define it with this issue - [x] If ACLK build fails make it prominent to the user so he knows about it. It should not be just small line easily overlooked in log. @amoss will speak to @jacekkolasa about this separately. - [x] Netdata should log on startup it is build without ACLK - [x] Report failure to Cloud to same endpoint #8051 - [x] Respect DO_NOT_TRACK environment variable (Some of this will be covered by PR 8025 but there are fresh requests at the bottom of the discussion).
2.0
During the agent installation, if the ACLK fails to be built, show an error message to the user - #### Summary Ensure that the requirement in product#282 are met. - [x] If ACLK error reporting not set yet by #8051 define it with this issue - [x] If ACLK build fails make it prominent to the user so he knows about it. It should not be just small line easily overlooked in log. @amoss will speak to @jacekkolasa about this separately. - [x] Netdata should log on startup it is build without ACLK - [x] Report failure to Cloud to same endpoint #8051 - [x] Respect DO_NOT_TRACK environment variable (Some of this will be covered by PR 8025 but there are fresh requests at the bottom of the discussion).
priority
during the agent installation if the aclk fails to be built show an error message to the user summary ensure that the requirement in product are met if aclk error reporting not set yet by define it with this issue if aclk build fails make it prominent to the user so he knows about it it should not be just small line easily overlooked in log amoss will speak to jacekkolasa about this separately netdata should log on startup it is build without aclk report failure to cloud to same endpoint respect do not track environment variable some of this will be covered by pr but there are fresh requests at the bottom of the discussion
1
632,748
20,205,862,237
IssuesEvent
2022-02-11 20:16:28
nens/threedi-api-qgis-client
https://api.github.com/repos/nens/threedi-api-qgis-client
closed
Download results if user has access to simulation results but not to schematisation
⏰ Priority: 3. Medium
This currently results in the following error: ![image](https://user-images.githubusercontent.com/26224299/153220671-5f78d066-8c2c-406a-b81e-8ef4e8a0b287.png) Desired behaviour: if schematisation is not found, let user choose where to store the results (Save as dialog)
1.0
Download results if user has access to simulation results but not to schematisation - This currently results in the following error: ![image](https://user-images.githubusercontent.com/26224299/153220671-5f78d066-8c2c-406a-b81e-8ef4e8a0b287.png) Desired behaviour: if schematisation is not found, let user choose where to store the results (Save as dialog)
priority
download results if user has access to simulation results but not to schematisation this currently results in the following error desired behaviour if schematisation is not found let user choose where to store the results save as dialog
1
69,655
3,309,741,825
IssuesEvent
2015-11-05 03:18:40
cs2103aug2015-t14-2j/main
https://api.github.com/repos/cs2103aug2015-t14-2j/main
closed
As a user i want to have a GUI
priority.medium
so that I can visualize the different tasks on my screen and better conceptualize my schedule
1.0
As a user i want to have a GUI - so that I can visualize the different tasks on my screen and better conceptualize my schedule
priority
as a user i want to have a gui so that i can visualize the different tasks on my screen and better conceptualize my schedule
1
479,370
13,795,364,105
IssuesEvent
2020-10-09 17:53:18
medic/cht-core
https://api.github.com/repos/medic/cht-core
closed
Update `admin` to be a standalone app
Priority: 2 - Medium Type: Technical issue
**Describe the issue** Currently, the admin app requires a significantly large number of webapp files (services, filters, directives, redux actions/reducers/selectors). This becomes a problem with the migration to Angular 10, where the files imported from webapp will, most likely, be unusable by admin. **Describe the improvement you'd like** "Duplicate" all files that admin app requires from the webapp folder, preserving git line history as much as possible. (this includes their tests) **Describe alternatives you've considered** We could postpone this migration under the assumption that we may be able to use post-angular-10-migration webapp files as we do now, but it's a risk.
1.0
Update `admin` to be a standalone app - **Describe the issue** Currently, the admin app requires a significantly large number of webapp files (services, filters, directives, redux actions/reducers/selectors). This becomes a problem with the migration to Angular 10, where the files imported from webapp will, most likely, be unusable by admin. **Describe the improvement you'd like** "Duplicate" all files that admin app requires from the webapp folder, preserving git line history as much as possible. (this includes their tests) **Describe alternatives you've considered** We could postpone this migration under the assumption that we may be able to use post-angular-10-migration webapp files as we do now, but it's a risk.
priority
update admin to be a standalone app describe the issue currently the admin app requires a significantly large number of webapp files services filters directives redux actions reducers selectors this becomes a problem with the migration to angular where the files imported from webapp will most likely be unusable by admin describe the improvement you d like duplicate all files that admin app requires from the webapp folder preserving git line history as much as possible this includes their tests describe alternatives you ve considered we could postpone this migration under the assumption that we may be able to use post angular migration webapp files as we do now but it s a risk
1
148,749
5,695,880,666
IssuesEvent
2017-04-16 04:23:56
tootsuite/mastodon
https://api.github.com/repos/tootsuite/mastodon
closed
Videos in unsupported codecs cause UI problems
bug priority - medium ui
Viewing a toot containing a video in a format your browser doesn't support causes usability issues with the UI: * there is a blank space where the thumbnail would be (see the first toot in the following screenshot) ![img_0217](https://cloud.githubusercontent.com/assets/3615519/24889007/20c9d9b2-1e2d-11e7-8e39-643eb8d17af5.PNG) * clicking on the space covers the screen in a lightbox, but it's empty other than a close button ![img_0218](https://cloud.githubusercontent.com/assets/3615519/24889014/35fcfd64-1e2d-11e7-915b-78acc6e2599a.PNG) These screenshots were taken of a toot containing a WebM video, on iOS, which doesn't support WebM video encoding in the browser (due to lack of hardware acceleration of that format). I understand that it's not feasible to convert every video uploaded to Mastodon to compatible formats for every device, but I'm wondering if it's possible to detect the unsupported format and either hide the thumbnail preview and/or display a message or icon indicating the format isn't supported (in any case, the link should still be clickable to download the file because another app might support it even if the browser doesn't). * * * * - [X] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
1.0
Videos in unsupported codecs cause UI problems - Viewing a toot containing a video in a format your browser doesn't support causes usability issues with the UI: * there is a blank space where the thumbnail would be (see the first toot in the following screenshot) ![img_0217](https://cloud.githubusercontent.com/assets/3615519/24889007/20c9d9b2-1e2d-11e7-8e39-643eb8d17af5.PNG) * clicking on the space covers the screen in a lightbox, but it's empty other than a close button ![img_0218](https://cloud.githubusercontent.com/assets/3615519/24889014/35fcfd64-1e2d-11e7-915b-78acc6e2599a.PNG) These screenshots were taken of a toot containing a WebM video, on iOS, which doesn't support WebM video encoding in the browser (due to lack of hardware acceleration of that format). I understand that it's not feasible to convert every video uploaded to Mastodon to compatible formats for every device, but I'm wondering if it's possible to detect the unsupported format and either hide the thumbnail preview and/or display a message or icon indicating the format isn't supported (in any case, the link should still be clickable to download the file because another app might support it even if the browser doesn't). * * * * - [X] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
priority
videos in unsupported codecs cause ui problems viewing a toot containing a video in a format your browser doesn t support causes usability issues with the ui there is a blank space where the thumbnail would be see the first toot in the following screenshot clicking on the space covers the screen in a lightbox but it s empty other than a close button these screenshots were taken of a toot containing a webm video on ios which doesn t support webm video encoding in the browser due to lack of hardware acceleration of that format i understand that it s not feasible to convert every video uploaded to mastodon to compatible formats for every device but i m wondering if it s possible to detect the unsupported format and either hide the thumbnail preview and or display a message or icon indicating the format isn t supported in any case the link should still be clickable to download the file because another app might support it even if the browser doesn t i searched or browsed the repo’s other issues to ensure this is not a duplicate
1
232,593
7,667,799,349
IssuesEvent
2018-05-14 00:47:04
hackoregon/civic-devops
https://api.github.com/repos/hackoregon/civic-devops
closed
Cannot `ssh` into Bastion Host in newly created stack
Priority: medium bug
This issue was previously mentioned at issue #87 @MikeTheCanuck suspects that the security group assigned to the Bastion Host only allows certain IP addresses incoming. The below screenshot showed that this host allows `tcp/22` from anywhere but both Mike and me got `Connect time out` ![image](https://user-images.githubusercontent.com/6059347/39965428-52cb5f90-564d-11e8-96d3-ca838bc50e4f.png)
1.0
Cannot `ssh` into Bastion Host in newly created stack - This issue was previously mentioned at issue #87 @MikeTheCanuck suspects that the security group assigned to the Bastion Host only allows certain IP addresses incoming. The below screenshot showed that this host allows `tcp/22` from anywhere but both Mike and me got `Connect time out` ![image](https://user-images.githubusercontent.com/6059347/39965428-52cb5f90-564d-11e8-96d3-ca838bc50e4f.png)
priority
cannot ssh into bastion host in newly created stack this issue was previously mentioned at issue mikethecanuck suspects that the security group assigned to the bastion host only allows certain ip addresses incoming the below screenshot showed that this host allows tcp from anywhere but both mike and me got connect time out
1
413,916
12,093,352,708
IssuesEvent
2020-04-19 19:17:09
samiha-rahman/soen390
https://api.github.com/repos/samiha-rahman/soen390
closed
US-64: Select classroom as start or end destination
epic 4 priority: medium user story
As a user, I would like to be able to select a classroom as a start or end destination. This story is created from a requirement extracted from #31 to reduce the scope of that story. **Associated epic:** #20 **Acceptance criteria:** Step # | Execution Procedure or Input | Expected Results/Outputs | Passed/Failed --- | --- | --- | --- 1 | Select a room | Should provide option to select designated room as starting point or end destination |
1.0
US-64: Select classroom as start or end destination - As a user, I would like to be able to select a classroom as a start or end destination. This story is created from a requirement extracted from #31 to reduce the scope of that story. **Associated epic:** #20 **Acceptance criteria:** Step # | Execution Procedure or Input | Expected Results/Outputs | Passed/Failed --- | --- | --- | --- 1 | Select a room | Should provide option to select designated room as starting point or end destination |
priority
us select classroom as start or end destination as a user i would like to be able to select a classroom as a start or end destination this story is created from a requirement extracted from to reduce the scope of that story associated epic acceptance criteria step execution procedure or input expected results outputs passed failed select a room should provide option to select designated room as starting point or end destination
1
422,945
12,288,633,361
IssuesEvent
2020-05-09 17:37:10
HabitRPG/habitica
https://api.github.com/repos/HabitRPG/habitica
closed
Username formatting breaks markdown for links in posts
help wanted priority: medium section: Guilds section: Tavern Chat
![image](https://user-images.githubusercontent.com/8144640/68435223-bde61380-01b2-11ea-9add-19ebab1ceb81.png) > "Something I already suspected but just confirmed: Putting a mention in a link markdown breaks the link the person is trying to add (I've had to use a different "@" glyph in the first code demonstration to avoid triggering it expanding the actual mention's code there...) > > Typing the following: > > [@citrusella really likes Habitica!](https://habitica.com) > > produces > > [@citrusella really likes Habitica!](https://habitica.com) > > which the website likely sees in the following way: > > [[@citrusella](/profile/2d6ef231-50b4-4a22-90e7-45eb97147a2c) really likes Habitica!](https://habitica.com)" (Image included since obviously one can't demonstrate in the same way on github! In other words, including @username in markdown links breaks it because the @mention formatting takes precedence.
1.0
Username formatting breaks markdown for links in posts - ![image](https://user-images.githubusercontent.com/8144640/68435223-bde61380-01b2-11ea-9add-19ebab1ceb81.png) > "Something I already suspected but just confirmed: Putting a mention in a link markdown breaks the link the person is trying to add (I've had to use a different "@" glyph in the first code demonstration to avoid triggering it expanding the actual mention's code there...) > > Typing the following: > > [@citrusella really likes Habitica!](https://habitica.com) > > produces > > [@citrusella really likes Habitica!](https://habitica.com) > > which the website likely sees in the following way: > > [[@citrusella](/profile/2d6ef231-50b4-4a22-90e7-45eb97147a2c) really likes Habitica!](https://habitica.com)" (Image included since obviously one can't demonstrate in the same way on github! In other words, including @username in markdown links breaks it because the @mention formatting takes precedence.
priority
username formatting breaks markdown for links in posts something i already suspected but just confirmed putting a mention in a link markdown breaks the link the person is trying to add i ve had to use a different glyph in the first code demonstration to avoid triggering it expanding the actual mention s code there typing the following produces which the website likely sees in the following way profile really likes habitica image included since obviously one can t demonstrate in the same way on github in other words including username in markdown links breaks it because the mention formatting takes precedence
1
776,099
27,246,883,097
IssuesEvent
2023-02-22 03:16:28
ansible-collections/azure
https://api.github.com/repos/ansible-collections/azure
closed
azure_rm_virtualmachine fails in AzureChinaCloud
has_pr medium_priority
##### SUMMARY something called by azure_rm_virtualmachine does not obey AZURE_CLOUD_ENVIRONMENT and attempts to connect to AzureCloud when operating on AzureChinaCloud hosts ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME azure_rm_virtualmachine ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible-playbook [core 2.12.5.post0] config file = /ansible/ansible.cfg configured module search path = ['/home/runner/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.8/site-packages/ansible ansible collection location = /home/runner/.ansible/collections:/ansible/collections executable location = /usr/local/bin/ansible-playbook python version = 3.8.13 (default, Jun 24 2022, 15:27:57) [GCC 8.5.0 20210514 (Red Hat 8.5.0-13)] jinja version = 3.1.2 libyaml = True ``` ##### COLLECTION VERSION <!--- Paste verbatim output from "ansible-galaxy collection list <namespace>.<collection>" between the quotes for example: ansible-galaxy collection list community.general --> ```paste below - name: azure.azcollection version: '==1.14.0' ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ANY_ERRORS_FATAL(/ansible/ansible.cfg) = False COLLECTIONS_PATHS(/ansible/ansible.cfg) = ['/home/runner/.ansible/collections', '/ansible/collections'] DEFAULT_FORKS(/ansible/ansible.cfg) = 50 DEFAULT_GATHER_SUBSET(/ansible/ansible.cfg) = ['all'] DEFAULT_GATHER_TIMEOUT(/ansible/ansible.cfg) = 20 DEFAULT_HASH_BEHAVIOUR(/ansible/ansible.cfg) = merge DEFAULT_TIMEOUT(/ansible/ansible.cfg) = 20 DEFAULT_TRANSPORT(/ansible/ansible.cfg) = smart HOST_KEY_CHECKING(/ansible/ansible.cfg) = False INVENTORY_ENABLED(/ansible/ansible.cfg) = ['vue.azure.azure_rm', 'script', 'yaml', 'ini'] PERSISTENT_COMMAND_TIMEOUT(/ansible/ansible.cfg) = 20 ``` ##### OS / ENVIRONMENT running on a recent awx-ee image that has had additional packages/modules installed ##### STEPS TO REPRODUCE example one: setting vm tags ```yaml - name: Set patch_version tags azure.azcollection.azure_rm_virtualmachine: resource_group: "{{ resource_group }}" name: "{{ name }}" tags: patch_version: "{{ ansible_date_time.iso8601 }}" zones: "{{ availability_zone }}" delegate_to: localhost when: - "'patch_version' in tags" - tags['patch_version'] == 'none' ``` example two: rebooting a vm via azure_rm_virtualmachine: ``` - name: Tell Azure to reboot this VM (so UAC setting change takes effect) azure.azcollection.azure_rm_virtualmachine: resource_group: "{{ resource_group }}" name: "{{ name }}" restarted: yes zones: "{{ availability_zone }}" delegate_to: localhost connection: local ``` ##### EXPECTED RESULTS VM tags set, or VM rebooted, no fatal errors ##### ACTUAL RESULTS ```paste below TASK [windows : Set patch_version tags] *************************************************************************************************************** task path: /ansible/roles/windows/tasks/patching_apply.yml:32 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: root <localhost> EXEC /bin/sh -c 'echo ~root && sleep 0' <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" && echo ansible-tmp-1671211081.1372511-135-112305740467344="` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" ) && sleep 0' Using module file /home/runner/.ansible/collections/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py <localhost> PUT /home/runner/.ansible/tmp/ansible-local-64whuibyec/tmpa3z1tq85 TO /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py <localhost> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0' <localhost> EXEC /bin/sh -c '/usr/bin/python3 /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0' <localhost> EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ > /dev/null 2>&1 && sleep 0' fatal: [************-vm0 -> localhost]: FAILED! => { "changed": false, "module_stderr": "ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: ****\r\nTimestamp: 2022-12-16 17:18:11Z\nTraceback (most recent call last):\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module>\n _ansiballz_main()\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main\n invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module\n runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path),\n File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module\n return _run_module_code(code, init_globals, run_name, mod_spec)\n File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code\n _run_code(code, mod_globals, init_globals,\n File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code\n exec(code, run_globals)\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module>\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module\n File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer\n return func(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run\n return first_node.send(pipeline_request) # type: ignore\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n [Previous line repeated 2 more times]\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send\n self.on_request(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request\n self._token = self._credential.get_token(*self._scopes)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token\n token = self._request_token(*scopes, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper\n return fn(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token\n raise ClientAuthenticationError(message=message)\nazure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: *****\r\nTimestamp: 2022-12-16 17:18:11Z\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1 } ``` decoded: ``` ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant. Trace ID: **** Correlation ID: **** Timestamp: 2022-12-16 16:08:27Z Traceback (most recent call last): File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module> _ansiballz_main() File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS) File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path), File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module return _run_module_code(code, init_globals, run_name, mod_spec) File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code _run_code(code, mod_globals, init_globals, File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code exec(code, run_globals) File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module> File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__ File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__ File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer return func(*args, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run return first_node.send(pipeline_request) # type: ignore File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send response = self.next.send(request) [Previous line repeated 2 more times] File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send self.on_request(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request self._token = self._credential.get_token(*self._scopes) File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token token = self._request_token(*scopes, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper return fn(*args, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token raise ClientAuthenticationError(message=message) azure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant. ```
1.0
azure_rm_virtualmachine fails in AzureChinaCloud - ##### SUMMARY something called by azure_rm_virtualmachine does not obey AZURE_CLOUD_ENVIRONMENT and attempts to connect to AzureCloud when operating on AzureChinaCloud hosts ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME azure_rm_virtualmachine ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible-playbook [core 2.12.5.post0] config file = /ansible/ansible.cfg configured module search path = ['/home/runner/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/local/lib/python3.8/site-packages/ansible ansible collection location = /home/runner/.ansible/collections:/ansible/collections executable location = /usr/local/bin/ansible-playbook python version = 3.8.13 (default, Jun 24 2022, 15:27:57) [GCC 8.5.0 20210514 (Red Hat 8.5.0-13)] jinja version = 3.1.2 libyaml = True ``` ##### COLLECTION VERSION <!--- Paste verbatim output from "ansible-galaxy collection list <namespace>.<collection>" between the quotes for example: ansible-galaxy collection list community.general --> ```paste below - name: azure.azcollection version: '==1.14.0' ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below ANY_ERRORS_FATAL(/ansible/ansible.cfg) = False COLLECTIONS_PATHS(/ansible/ansible.cfg) = ['/home/runner/.ansible/collections', '/ansible/collections'] DEFAULT_FORKS(/ansible/ansible.cfg) = 50 DEFAULT_GATHER_SUBSET(/ansible/ansible.cfg) = ['all'] DEFAULT_GATHER_TIMEOUT(/ansible/ansible.cfg) = 20 DEFAULT_HASH_BEHAVIOUR(/ansible/ansible.cfg) = merge DEFAULT_TIMEOUT(/ansible/ansible.cfg) = 20 DEFAULT_TRANSPORT(/ansible/ansible.cfg) = smart HOST_KEY_CHECKING(/ansible/ansible.cfg) = False INVENTORY_ENABLED(/ansible/ansible.cfg) = ['vue.azure.azure_rm', 'script', 'yaml', 'ini'] PERSISTENT_COMMAND_TIMEOUT(/ansible/ansible.cfg) = 20 ``` ##### OS / ENVIRONMENT running on a recent awx-ee image that has had additional packages/modules installed ##### STEPS TO REPRODUCE example one: setting vm tags ```yaml - name: Set patch_version tags azure.azcollection.azure_rm_virtualmachine: resource_group: "{{ resource_group }}" name: "{{ name }}" tags: patch_version: "{{ ansible_date_time.iso8601 }}" zones: "{{ availability_zone }}" delegate_to: localhost when: - "'patch_version' in tags" - tags['patch_version'] == 'none' ``` example two: rebooting a vm via azure_rm_virtualmachine: ``` - name: Tell Azure to reboot this VM (so UAC setting change takes effect) azure.azcollection.azure_rm_virtualmachine: resource_group: "{{ resource_group }}" name: "{{ name }}" restarted: yes zones: "{{ availability_zone }}" delegate_to: localhost connection: local ``` ##### EXPECTED RESULTS VM tags set, or VM rebooted, no fatal errors ##### ACTUAL RESULTS ```paste below TASK [windows : Set patch_version tags] *************************************************************************************************************** task path: /ansible/roles/windows/tasks/patching_apply.yml:32 <localhost> ESTABLISH LOCAL CONNECTION FOR USER: root <localhost> EXEC /bin/sh -c 'echo ~root && sleep 0' <localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" && echo ansible-tmp-1671211081.1372511-135-112305740467344="` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" ) && sleep 0' Using module file /home/runner/.ansible/collections/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py <localhost> PUT /home/runner/.ansible/tmp/ansible-local-64whuibyec/tmpa3z1tq85 TO /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py <localhost> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0' <localhost> EXEC /bin/sh -c '/usr/bin/python3 /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0' <localhost> EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ > /dev/null 2>&1 && sleep 0' fatal: [************-vm0 -> localhost]: FAILED! => { "changed": false, "module_stderr": "ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: ****\r\nTimestamp: 2022-12-16 17:18:11Z\nTraceback (most recent call last):\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module>\n _ansiballz_main()\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main\n invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module\n runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path),\n File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module\n return _run_module_code(code, init_globals, run_name, mod_spec)\n File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code\n _run_code(code, mod_globals, init_globals,\n File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code\n exec(code, run_globals)\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module>\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module\n File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer\n return func(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run\n return first_node.send(pipeline_request) # type: ignore\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n [Previous line repeated 2 more times]\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send\n self.on_request(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request\n self._token = self._credential.get_token(*self._scopes)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token\n token = self._request_token(*scopes, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper\n return fn(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token\n raise ClientAuthenticationError(message=message)\nazure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: *****\r\nTimestamp: 2022-12-16 17:18:11Z\n", "module_stdout": "", "msg": "MODULE FAILURE\nSee stdout/stderr for the exact error", "rc": 1 } ``` decoded: ``` ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant. Trace ID: **** Correlation ID: **** Timestamp: 2022-12-16 16:08:27Z Traceback (most recent call last): File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module> _ansiballz_main() File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS) File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path), File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module return _run_module_code(code, init_globals, run_name, mod_spec) File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code _run_code(code, mod_globals, init_globals, File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code exec(code, run_globals) File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module> File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__ File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__ File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer return func(*args, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run return first_node.send(pipeline_request) # type: ignore File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send response = self.next.send(request) [Previous line repeated 2 more times] File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send response = self.next.send(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send self.on_request(request) File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request self._token = self._credential.get_token(*self._scopes) File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token token = self._request_token(*scopes, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper return fn(*args, **kwargs) File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token raise ClientAuthenticationError(message=message) azure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant. ```
priority
azure rm virtualmachine fails in azurechinacloud summary something called by azure rm virtualmachine does not obey azure cloud environment and attempts to connect to azurecloud when operating on azurechinacloud hosts issue type bug report component name azure rm virtualmachine ansible version paste below ansible playbook config file ansible ansible cfg configured module search path ansible python module location usr local lib site packages ansible ansible collection location home runner ansible collections ansible collections executable location usr local bin ansible playbook python version default jun jinja version libyaml true collection version between the quotes for example ansible galaxy collection list community general paste below name azure azcollection version configuration paste below any errors fatal ansible ansible cfg false collections paths ansible ansible cfg default forks ansible ansible cfg default gather subset ansible ansible cfg default gather timeout ansible ansible cfg default hash behaviour ansible ansible cfg merge default timeout ansible ansible cfg default transport ansible ansible cfg smart host key checking ansible ansible cfg false inventory enabled ansible ansible cfg persistent command timeout ansible ansible cfg os environment running on a recent awx ee image that has had additional packages modules installed steps to reproduce example one setting vm tags yaml name set patch version tags azure azcollection azure rm virtualmachine resource group resource group name name tags patch version ansible date time zones availability zone delegate to localhost when patch version in tags tags none example two rebooting a vm via azure rm virtualmachine name tell azure to reboot this vm so uac setting change takes effect azure azcollection azure rm virtualmachine resource group resource group name name restarted yes zones availability zone delegate to localhost connection local expected results vm tags set or vm rebooted no fatal errors actual results paste below task task path ansible roles windows tasks patching apply yml establish local connection for user root exec bin sh c echo root sleep exec bin sh c umask mkdir p echo root ansible tmp mkdir echo root ansible tmp ansible tmp echo ansible tmp echo root ansible tmp ansible tmp sleep using module file home runner ansible collections ansible collections azure azcollection plugins modules azure rm virtualmachine py put home runner ansible tmp ansible local to root ansible tmp ansible tmp ansiballz azure rm virtualmachine py exec bin sh c chmod u x root ansible tmp ansible tmp root ansible tmp ansible tmp ansiballz azure rm virtualmachine py sleep exec bin sh c usr bin root ansible tmp ansible tmp ansiballz azure rm virtualmachine py sleep exec bin sh c rm f r root ansible tmp ansible tmp dev null sleep fatal failed changed false module stderr clientsecretcredential get token failed authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant r ntrace id r ncorrelation id r ntimestamp ntraceback most recent call last n file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in n ansiballz main n file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in ansiballz main n invoke module zipped mod temp path ansiballz params n file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in invoke module n runpy run module mod name ansible collections azure azcollection plugins modules azure rm virtualmachine init globals dict module fqn ansible collections azure azcollection plugins modules azure rm virtualmachine modlib path modlib path n file usr runpy py line in run module n return run module code code init globals run name mod spec n file usr runpy py line in run module code n run code code mod globals init globals n file usr runpy py line in run code n exec code run globals n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in main n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in init n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins module utils azure rm common py line in init n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in exec module n file usr local lib site packages azure core tracing decorator py line in wrapper use tracer n return func args kwargs n file usr local lib site packages azure mgmt compute operations virtual machines operations py line in get n pipeline response self client pipeline run request stream false kwargs n file usr local lib site packages azure core pipeline base py line in run n return first node send pipeline request type ignore n file usr local lib site packages azure core pipeline base py line in send n response self next send request n file usr local lib site packages azure core pipeline base py line in send n response self next send request n file usr local lib site packages azure core pipeline base py line in send n response self next send request n n file usr local lib site packages azure mgmt core policies base py line in send n response self next send request n file usr local lib site packages azure core pipeline policies redirect py line in send n response self next send request n file usr local lib site packages azure core pipeline policies retry py line in send n response self next send request n file usr local lib site packages azure core pipeline policies authentication py line in send n self on request request n file usr local lib site packages azure core pipeline policies authentication py line in on request n self token self credential get token self scopes n file usr local lib site packages azure identity internal get token mixin py line in get token n token self request token scopes kwargs n file usr local lib site packages azure identity internal decorators py line in wrapper n return fn args kwargs n file usr local lib site packages azure identity internal client credential base py line in request token n raise clientauthenticationerror message message nazure core exceptions clientauthenticationerror authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant r ntrace id r ncorrelation id r ntimestamp n module stdout msg module failure nsee stdout stderr for the exact error rc decoded clientsecretcredential get token failed authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant trace id correlation id timestamp traceback most recent call last file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in ansiballz main file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in ansiballz main invoke module zipped mod temp path ansiballz params file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in invoke module runpy run module mod name ansible collections azure azcollection plugins modules azure rm virtualmachine init globals dict module fqn ansible collections azure azcollection plugins modules azure rm virtualmachine modlib path modlib path file usr runpy py line in run module return run module code code init globals run name mod spec file usr runpy py line in run module code run code code mod globals init globals file usr runpy py line in run code exec code run globals file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in main file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in init file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins module utils azure rm common py line in init file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in exec module file usr local lib site packages azure core tracing decorator py line in wrapper use tracer return func args kwargs file usr local lib site packages azure mgmt compute operations virtual machines operations py line in get pipeline response self client pipeline run request stream false kwargs file usr local lib site packages azure core pipeline base py line in run return first node send pipeline request type ignore file usr local lib site packages azure core pipeline base py line in send response self next send request file usr local lib site packages azure core pipeline base py line in send response self next send request file usr local lib site packages azure core pipeline base py line in send response self next send request file usr local lib site packages azure mgmt core policies base py line in send response self next send request file usr local lib site packages azure core pipeline policies redirect py line in send response self next send request file usr local lib site packages azure core pipeline policies retry py line in send response self next send request file usr local lib site packages azure core pipeline policies authentication py line in send self on request request file usr local lib site packages azure core pipeline policies authentication py line in on request self token self credential get token self scopes file usr local lib site packages azure identity internal get token mixin py line in get token token self request token scopes kwargs file usr local lib site packages azure identity internal decorators py line in wrapper return fn args kwargs file usr local lib site packages azure identity internal client credential base py line in request token raise clientauthenticationerror message message azure core exceptions clientauthenticationerror authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant
1
812,561
30,342,069,154
IssuesEvent
2023-07-11 13:20:27
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID: 318645] Out-of-bounds access in subsys/bluetooth/controller/ll_sw/ull_adv_aux.c
bug priority: medium area: Bluetooth Coverity area: Bluetooth Controller
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c#L842 Category: Memory - corruptions Function: `ll_adv_aux_sr_data_set` Component: Bluetooth CID: [318645](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=318645) Details: https://github.com/zephyrproject-rtos/zephyr/blob/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c ``` 828 829 if (op == BT_HCI_LE_EXT_ADV_OP_INTERM_FRAG || 830 op == BT_HCI_LE_EXT_ADV_OP_LAST_FRAG) { 831 /* Append fragment to existing data */ 832 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA | 833 ULL_ADV_PDU_HDR_FIELD_AD_DATA_APPEND; >>> CID 318645: (OVERRUN) >>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20. 834 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu, 835 hdr_add_fields, 836 0, 837 hdr_data); 838 } else { 839 /* Add AD Data and remove any prior presence of Aux Ptr */ 806 */ 807 *val_ptr++ = len; 808 (void)memcpy(val_ptr, &data, sizeof(data)); 809 } 810 811 /* Trigger DID update */ >>> CID 318645: (OVERRUN) >>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20. 812 err = ull_adv_aux_hdr_set_clear(adv, hdr_add_fields, 0U, 813 hdr_data, &pri_idx, &sec_idx); 814 if (err) { 815 return err; 816 } 817 836 0, 837 hdr_data); 838 } else { 839 /* Add AD Data and remove any prior presence of Aux Ptr */ 840 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA | 841 ULL_ADV_PDU_HDR_FIELD_AD_DATA; >>> CID 318645: (OVERRUN) >>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20. 842 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu, 843 hdr_add_fields, 844 ULL_ADV_PDU_HDR_FIELD_AUX_PTR, 845 hdr_data); 846 } 847 #if defined(CONFIG_BT_CTLR_ADV_AUX_PDU_LINK) ``` For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_OVERRUN). ([CWE-119](http://cwe.mitre.org/data/definitions/119.html)) Please fix or provide comments in coverity using the link: https://scan9.scan.coverity.com/reports.htm#v29271/p12996 Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID: 318645] Out-of-bounds access in subsys/bluetooth/controller/ll_sw/ull_adv_aux.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c#L842 Category: Memory - corruptions Function: `ll_adv_aux_sr_data_set` Component: Bluetooth CID: [318645](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=318645) Details: https://github.com/zephyrproject-rtos/zephyr/blob/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c ``` 828 829 if (op == BT_HCI_LE_EXT_ADV_OP_INTERM_FRAG || 830 op == BT_HCI_LE_EXT_ADV_OP_LAST_FRAG) { 831 /* Append fragment to existing data */ 832 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA | 833 ULL_ADV_PDU_HDR_FIELD_AD_DATA_APPEND; >>> CID 318645: (OVERRUN) >>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20. 834 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu, 835 hdr_add_fields, 836 0, 837 hdr_data); 838 } else { 839 /* Add AD Data and remove any prior presence of Aux Ptr */ 806 */ 807 *val_ptr++ = len; 808 (void)memcpy(val_ptr, &data, sizeof(data)); 809 } 810 811 /* Trigger DID update */ >>> CID 318645: (OVERRUN) >>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20. 812 err = ull_adv_aux_hdr_set_clear(adv, hdr_add_fields, 0U, 813 hdr_data, &pri_idx, &sec_idx); 814 if (err) { 815 return err; 816 } 817 836 0, 837 hdr_data); 838 } else { 839 /* Add AD Data and remove any prior presence of Aux Ptr */ 840 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA | 841 ULL_ADV_PDU_HDR_FIELD_AD_DATA; >>> CID 318645: (OVERRUN) >>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20. 842 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu, 843 hdr_add_fields, 844 ULL_ADV_PDU_HDR_FIELD_AUX_PTR, 845 hdr_data); 846 } 847 #if defined(CONFIG_BT_CTLR_ADV_AUX_PDU_LINK) ``` For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_OVERRUN). ([CWE-119](http://cwe.mitre.org/data/definitions/119.html)) Please fix or provide comments in coverity using the link: https://scan9.scan.coverity.com/reports.htm#v29271/p12996 Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
priority
out of bounds access in subsys bluetooth controller ll sw ull adv aux c static code scan issues found in file category memory corruptions function ll adv aux sr data set component bluetooth cid details if op bt hci le ext adv op interm frag op bt hci le ext adv op last frag append fragment to existing data hdr add fields ull adv pdu hdr field adva ull adv pdu hdr field ad data append cid overrun overrunning array hdr data of bytes by passing it to a function which accesses it at byte offset err ull adv aux pdu set clear adv sr pdu prev sr pdu hdr add fields hdr data else add ad data and remove any prior presence of aux ptr val ptr len void memcpy val ptr data sizeof data trigger did update cid overrun overrunning array hdr data of bytes by passing it to a function which accesses it at byte offset err ull adv aux hdr set clear adv hdr add fields hdr data pri idx sec idx if err return err hdr data else add ad data and remove any prior presence of aux ptr hdr add fields ull adv pdu hdr field adva ull adv pdu hdr field ad data cid overrun overrunning array hdr data of bytes by passing it to a function which accesses it at byte offset err ull adv aux pdu set clear adv sr pdu prev sr pdu hdr add fields ull adv pdu hdr field aux ptr hdr data if defined config bt ctlr adv aux pdu link for more information about the violation check the please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
1
85,953
3,700,889,459
IssuesEvent
2016-02-29 10:38:45
OCHA-DAP/hdx-ckan
https://api.github.com/repos/OCHA-DAP/hdx-ckan
closed
New contribute flow: Horizontal scrollbar showing
bug New Contribute Flow Priority-Medium
I tested in Linux - Firefox and Chrome in FullHD, and in Windows in Firefox: ![scrollbar](https://cloud.githubusercontent.com/assets/6586348/12813794/0d469918-cb43-11e5-8988-e077686ae36b.png) Notice the scrollbar at the bottom
1.0
New contribute flow: Horizontal scrollbar showing - I tested in Linux - Firefox and Chrome in FullHD, and in Windows in Firefox: ![scrollbar](https://cloud.githubusercontent.com/assets/6586348/12813794/0d469918-cb43-11e5-8988-e077686ae36b.png) Notice the scrollbar at the bottom
priority
new contribute flow horizontal scrollbar showing i tested in linux firefox and chrome in fullhd and in windows in firefox notice the scrollbar at the bottom
1
424,564
12,313,098,666
IssuesEvent
2020-05-12 14:50:56
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
closed
Come up with a way to handle reviews for the feature replacement workflow
Category: Core Priority: Medium Status: New/Undefined Type: Feature
Two types of reviews could be useful during feature replacement. The first would be the regular conflate review and another could be flagging reviews for linear features snapped post conflation (not yet added; could group reviews by connected ways to reduce the number of them). Since our output is a changeset and the command works outside of the UI review workflow, there is currently no way to handle reviews and they are dropped completely.
1.0
Come up with a way to handle reviews for the feature replacement workflow - Two types of reviews could be useful during feature replacement. The first would be the regular conflate review and another could be flagging reviews for linear features snapped post conflation (not yet added; could group reviews by connected ways to reduce the number of them). Since our output is a changeset and the command works outside of the UI review workflow, there is currently no way to handle reviews and they are dropped completely.
priority
come up with a way to handle reviews for the feature replacement workflow two types of reviews could be useful during feature replacement the first would be the regular conflate review and another could be flagging reviews for linear features snapped post conflation not yet added could group reviews by connected ways to reduce the number of them since our output is a changeset and the command works outside of the ui review workflow there is currently no way to handle reviews and they are dropped completely
1
679,463
23,233,245,396
IssuesEvent
2022-08-03 09:28:25
owncloud/web
https://api.github.com/repos/owncloud/web
closed
Selected item glues on bottom & scrolls for ⬆️ key up
Type:Bug Priority:p3-medium GA-Blocker
### Steps to reproduce 1. Login to https://ocis.ocis-web.latest.owncloud.works/ 2. upload ~100 files into a single folder 3. select one of the 100 files 4. press key down ⬇️ until scolling starts 5. press key ⬆️ until scrolling starts 6. Selected item glues on bottom and scrolls on every keystroke https://user-images.githubusercontent.com/26610733/180446452-a27ec8b3-ac8d-45aa-9635-4e21cd9132b4.mp4 ### Expected behaviour selection should "walk" to the top without scrolling, then start scrolling https://user-images.githubusercontent.com/26610733/180445840-e7deb84f-da62-41c6-b625-156e48980383.mp4 ### Actual behaviour Selected item glues on bottom and scrolls on every keystroke
1.0
Selected item glues on bottom & scrolls for ⬆️ key up - ### Steps to reproduce 1. Login to https://ocis.ocis-web.latest.owncloud.works/ 2. upload ~100 files into a single folder 3. select one of the 100 files 4. press key down ⬇️ until scolling starts 5. press key ⬆️ until scrolling starts 6. Selected item glues on bottom and scrolls on every keystroke https://user-images.githubusercontent.com/26610733/180446452-a27ec8b3-ac8d-45aa-9635-4e21cd9132b4.mp4 ### Expected behaviour selection should "walk" to the top without scrolling, then start scrolling https://user-images.githubusercontent.com/26610733/180445840-e7deb84f-da62-41c6-b625-156e48980383.mp4 ### Actual behaviour Selected item glues on bottom and scrolls on every keystroke
priority
selected item glues on bottom scrolls for ⬆️ key up steps to reproduce login to upload files into a single folder select one of the files press key down ⬇️ until scolling starts press key ⬆️ until scrolling starts selected item glues on bottom and scrolls on every keystroke expected behaviour selection should walk to the top without scrolling then start scrolling actual behaviour selected item glues on bottom and scrolls on every keystroke
1
205,052
7,093,594,910
IssuesEvent
2018-01-12 21:12:42
certificate-helper/TLS-Inspector
https://api.github.com/repos/certificate-helper/TLS-Inspector
closed
Limit the number of redirects TLS Inspector will follow
CertificateKit bug easy medium priority merged
**Affected Version:** Since 1.6.0 **Is this a Test Flight version or the App Store version?** App Store **Device and iOS Version:** All **What steps will reproduce the problem?** 1. Navigate to a web page that triggers a redirect loop **What is the expected output?** A specific warning about a redirect loop **What do you see instead?** Long loading then eventually timeout **Please provide any additional information below.**
1.0
Limit the number of redirects TLS Inspector will follow - **Affected Version:** Since 1.6.0 **Is this a Test Flight version or the App Store version?** App Store **Device and iOS Version:** All **What steps will reproduce the problem?** 1. Navigate to a web page that triggers a redirect loop **What is the expected output?** A specific warning about a redirect loop **What do you see instead?** Long loading then eventually timeout **Please provide any additional information below.**
priority
limit the number of redirects tls inspector will follow affected version since is this a test flight version or the app store version app store device and ios version all what steps will reproduce the problem navigate to a web page that triggers a redirect loop what is the expected output a specific warning about a redirect loop what do you see instead long loading then eventually timeout please provide any additional information below
1
134,710
5,232,910,807
IssuesEvent
2017-01-30 11:03:23
openworm/behavioral_syntax
https://api.github.com/repos/openworm/behavioral_syntax
closed
consider alternatives to compression(MDL)
medium priority
Following Andre's presentation at the OpenWorm journal club today it might be a good idea to look into using: 1) Hidden Markov Models 2) Statistical Analysis: use p-values to see whether there's any information gain from using n grams vs (n-1) grams 3) Use Hierarchical Markov Models
1.0
consider alternatives to compression(MDL) - Following Andre's presentation at the OpenWorm journal club today it might be a good idea to look into using: 1) Hidden Markov Models 2) Statistical Analysis: use p-values to see whether there's any information gain from using n grams vs (n-1) grams 3) Use Hierarchical Markov Models
priority
consider alternatives to compression mdl following andre s presentation at the openworm journal club today it might be a good idea to look into using hidden markov models statistical analysis use p values to see whether there s any information gain from using n grams vs n grams use hierarchical markov models
1
623,214
19,663,333,600
IssuesEvent
2022-01-10 19:25:46
ScottUK/ladojrp-issues
https://api.github.com/repos/ScottUK/ladojrp-issues
closed
PA System For Police
Class: enhancement Priority: medium Scope: scripts
**Describe the feature you'd like implemented** There should be a PA system for police to use in their vehicles so if they need to yell for a vehicle to pull over or pull closer to the side of the road they can do so by using that implemented PA system.
1.0
PA System For Police - **Describe the feature you'd like implemented** There should be a PA system for police to use in their vehicles so if they need to yell for a vehicle to pull over or pull closer to the side of the road they can do so by using that implemented PA system.
priority
pa system for police describe the feature you d like implemented there should be a pa system for police to use in their vehicles so if they need to yell for a vehicle to pull over or pull closer to the side of the road they can do so by using that implemented pa system
1
437,075
12,559,884,624
IssuesEvent
2020-06-07 20:20:57
DuckBoss/JJMumbleBot
https://api.github.com/repos/DuckBoss/JJMumbleBot
closed
Better documentation
Medium Priority Next Release Wiki
Hey, I really like your project! For a better and quicker understanding of your projects and the installation process if would be really helpful to add some sample files (not a image) and improve the quickstart with more details and in depth examples.
1.0
Better documentation - Hey, I really like your project! For a better and quicker understanding of your projects and the installation process if would be really helpful to add some sample files (not a image) and improve the quickstart with more details and in depth examples.
priority
better documentation hey i really like your project for a better and quicker understanding of your projects and the installation process if would be really helpful to add some sample files not a image and improve the quickstart with more details and in depth examples
1
211,567
7,202,254,780
IssuesEvent
2018-02-06 02:49:34
Albert481/myTrolley
https://api.github.com/repos/Albert481/myTrolley
closed
Report trolley condition
Medium NTUC Shoppers Priority Normal
report condition of trolley, provide comment box to specify which part if necessary
1.0
Report trolley condition - report condition of trolley, provide comment box to specify which part if necessary
priority
report trolley condition report condition of trolley provide comment box to specify which part if necessary
1
218,543
7,331,716,185
IssuesEvent
2018-03-05 14:23:41
Mapita/shorter
https://api.github.com/repos/Mapita/shorter
opened
Use Travis CI to automatically run unit tests before approving PRs
effort: low meta: blocked meta: next release priority: medium type: integration
Travis CI should be used to run tests for the master branch and before merging PRs. Blocked by https://github.com/Mapita/shorter/issues/4
1.0
Use Travis CI to automatically run unit tests before approving PRs - Travis CI should be used to run tests for the master branch and before merging PRs. Blocked by https://github.com/Mapita/shorter/issues/4
priority
use travis ci to automatically run unit tests before approving prs travis ci should be used to run tests for the master branch and before merging prs blocked by
1
382,258
11,303,041,845
IssuesEvent
2020-01-17 19:07:34
chicagopython/chipy.org
https://api.github.com/repos/chicagopython/chipy.org
closed
Remove Maybe option from RSVP selection
priority: medium status: revision needed type: enhancement
As we will be using the website going forward, let's remove the ambiguity of Maybe from RSVP options. I'm gonna take this one. Will be a good way to dig into the codebase.
1.0
Remove Maybe option from RSVP selection - As we will be using the website going forward, let's remove the ambiguity of Maybe from RSVP options. I'm gonna take this one. Will be a good way to dig into the codebase.
priority
remove maybe option from rsvp selection as we will be using the website going forward let s remove the ambiguity of maybe from rsvp options i m gonna take this one will be a good way to dig into the codebase
1
696,676
23,911,107,177
IssuesEvent
2022-09-09 08:17:45
Chatterino/chatterino2
https://api.github.com/repos/Chatterino/chatterino2
closed
Deprecation of IRC commands
enhancement Platform: Twitch Priority: Medium Deprecation: Twitch IRC Commands
On or around Feb 18th, 2023, Twitch will deprecate all commands currently executed through IRC, `excluding /disconnect & /me` More information on this can be found [here.](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486) This will mean the following commands will need to be transferred from IRC to Helix, in Chatterino: <details> <summary>Command List</summary> | Completed | Command | Documentation | | ----------- | ----------- | ----------- | | ❌ | `/announce` | https://dev.twitch.tv/docs/api/reference#send-chat-announcement | | ❌ | `/ban` | https://dev.twitch.tv/docs/api/reference#ban-user ³| | ❌ | `/unban` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴| | ❌ | `/clear` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹| | ❌ | `/color` | https://dev.twitch.tv/docs/api/reference#update-user-chat-color | | ❌ | `/commercial` | https://dev.twitch.tv/docs/api/reference#start-commercial | | ❌ | `/delete` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹| | ❌ | `/emoteonly` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/emoteonlyoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/followers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/followersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ✅ | `/marker` | https://dev.twitch.tv/docs/api/reference#create-stream-marker | | ❌ | `/mod` | https://dev.twitch.tv/docs/api/reference#add-channel-moderator | | ❌ | `/mods` | https://dev.twitch.tv/docs/api/reference#get-moderators ⁵| | ❌ | `/unmod` | https://dev.twitch.tv/docs/api/reference#remove-channel-moderator | | ❌ | `/raid` | https://dev.twitch.tv/docs/api/reference#start-a-raid ⁵| | ❌ | `/unraid` | https://dev.twitch.tv/docs/api/reference#cancel-a-raid ⁵| | ❌ | `/slow` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/slowoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/subscribers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/subscribersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/timeout` | https://dev.twitch.tv/docs/api/reference#ban-user ³| | ❌ | `/untimeout` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴| | ❌ | `/uniquechat` | https://dev.twitch.tv/docs/api/reference#update-chat-settings | | ❌ | `/uniquechatoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings | | ❌ | `/vip` | https://dev.twitch.tv/docs/api/reference#add-channel-vip | | ❌ | `/vips` | https://dev.twitch.tv/docs/api/reference#get-vips ⁵| | ❌ | `/unvip` | https://dev.twitch.tv/docs/api/reference#remove-channel-vip | | ❌ | `/w` | https://dev.twitch.tv/docs/api/reference#send-whisper ⁶| 1. `/clear` & `/delete` use the same API call 2. `/emoteonly`, `/emoteonlyoff`, `/followers`, `/followersoff`, `/slow`, `/slowoff`, `/subscribers`, `/subscribersoff`, `/uniquechat`, `/uniquechatoff` all use the same API call 3. `/ban` & `/timeout` use the same API call 4. `/unban` & `/untimeout` use the same API call 5. `/mods`, `/vips`, `/raid`, and `/unraid` are now broadcaster only commands 6. `/w` now requires Twitch phone number verification `This is not the same as 2FA` <!--- here are the emojis you'll need to update this: ✅❌ ---> </details> The newly required scopes have already been handled in Chatterino/website#77 As noted from the list above, the following commands will need custom error handling - `/mods` - Will need a custom error for non-broadcasters stating they are no longer able use this command. - `/vips` - "" - `/raid` - "" - `/unraid` - "" - `/w` - Will require custom error handling when receiving back an error code 401, which will indicate the user has not verified their phone number with Twitch. - Likely will point to the whisper section of the wiki, which will need to be updated showing where to add said phone number. `settings -> security`
1.0
Deprecation of IRC commands - On or around Feb 18th, 2023, Twitch will deprecate all commands currently executed through IRC, `excluding /disconnect & /me` More information on this can be found [here.](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486) This will mean the following commands will need to be transferred from IRC to Helix, in Chatterino: <details> <summary>Command List</summary> | Completed | Command | Documentation | | ----------- | ----------- | ----------- | | ❌ | `/announce` | https://dev.twitch.tv/docs/api/reference#send-chat-announcement | | ❌ | `/ban` | https://dev.twitch.tv/docs/api/reference#ban-user ³| | ❌ | `/unban` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴| | ❌ | `/clear` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹| | ❌ | `/color` | https://dev.twitch.tv/docs/api/reference#update-user-chat-color | | ❌ | `/commercial` | https://dev.twitch.tv/docs/api/reference#start-commercial | | ❌ | `/delete` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹| | ❌ | `/emoteonly` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/emoteonlyoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/followers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/followersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ✅ | `/marker` | https://dev.twitch.tv/docs/api/reference#create-stream-marker | | ❌ | `/mod` | https://dev.twitch.tv/docs/api/reference#add-channel-moderator | | ❌ | `/mods` | https://dev.twitch.tv/docs/api/reference#get-moderators ⁵| | ❌ | `/unmod` | https://dev.twitch.tv/docs/api/reference#remove-channel-moderator | | ❌ | `/raid` | https://dev.twitch.tv/docs/api/reference#start-a-raid ⁵| | ❌ | `/unraid` | https://dev.twitch.tv/docs/api/reference#cancel-a-raid ⁵| | ❌ | `/slow` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/slowoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/subscribers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/subscribersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²| | ❌ | `/timeout` | https://dev.twitch.tv/docs/api/reference#ban-user ³| | ❌ | `/untimeout` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴| | ❌ | `/uniquechat` | https://dev.twitch.tv/docs/api/reference#update-chat-settings | | ❌ | `/uniquechatoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings | | ❌ | `/vip` | https://dev.twitch.tv/docs/api/reference#add-channel-vip | | ❌ | `/vips` | https://dev.twitch.tv/docs/api/reference#get-vips ⁵| | ❌ | `/unvip` | https://dev.twitch.tv/docs/api/reference#remove-channel-vip | | ❌ | `/w` | https://dev.twitch.tv/docs/api/reference#send-whisper ⁶| 1. `/clear` & `/delete` use the same API call 2. `/emoteonly`, `/emoteonlyoff`, `/followers`, `/followersoff`, `/slow`, `/slowoff`, `/subscribers`, `/subscribersoff`, `/uniquechat`, `/uniquechatoff` all use the same API call 3. `/ban` & `/timeout` use the same API call 4. `/unban` & `/untimeout` use the same API call 5. `/mods`, `/vips`, `/raid`, and `/unraid` are now broadcaster only commands 6. `/w` now requires Twitch phone number verification `This is not the same as 2FA` <!--- here are the emojis you'll need to update this: ✅❌ ---> </details> The newly required scopes have already been handled in Chatterino/website#77 As noted from the list above, the following commands will need custom error handling - `/mods` - Will need a custom error for non-broadcasters stating they are no longer able use this command. - `/vips` - "" - `/raid` - "" - `/unraid` - "" - `/w` - Will require custom error handling when receiving back an error code 401, which will indicate the user has not verified their phone number with Twitch. - Likely will point to the whisper section of the wiki, which will need to be updated showing where to add said phone number. `settings -> security`
priority
deprecation of irc commands on or around feb twitch will deprecate all commands currently executed through irc excluding disconnect me more information on this can be found this will mean the following commands will need to be transferred from irc to helix in chatterino command list completed command documentation ❌ announce ❌ ban ³ ❌ unban ⁴ ❌ clear ¹ ❌ color ❌ commercial ❌ delete ¹ ❌ emoteonly ² ❌ emoteonlyoff ² ❌ followers ² ❌ followersoff ² ✅ marker ❌ mod ❌ mods ⁵ ❌ unmod ❌ raid ⁵ ❌ unraid ⁵ ❌ slow ² ❌ slowoff ² ❌ subscribers ² ❌ subscribersoff ² ❌ timeout ³ ❌ untimeout ⁴ ❌ uniquechat ❌ uniquechatoff ❌ vip ❌ vips ⁵ ❌ unvip ❌ w ⁶ clear delete use the same api call emoteonly emoteonlyoff followers followersoff slow slowoff subscribers subscribersoff uniquechat uniquechatoff all use the same api call ban timeout use the same api call unban untimeout use the same api call mods vips raid and unraid are now broadcaster only commands w now requires twitch phone number verification this is not the same as the newly required scopes have already been handled in chatterino website as noted from the list above the following commands will need custom error handling mods will need a custom error for non broadcasters stating they are no longer able use this command vips raid unraid w will require custom error handling when receiving back an error code which will indicate the user has not verified their phone number with twitch likely will point to the whisper section of the wiki which will need to be updated showing where to add said phone number settings security
1
781,477
27,439,114,080
IssuesEvent
2023-03-02 09:45:30
vaticle/typedb-behaviour
https://api.github.com/repos/vaticle/typedb-behaviour
opened
Improve user management features
type: feature priority: medium
## Problem to Solve Our user management feature is fairly bare-bones right now, with TypeDB Cluster carrying the bulk of the weight for testing user management. ## Current Workaround We test much of user management with TypeDB Cluster integration tests. ## Proposed Solution We should expand the test suite and the available steps that can be executed during a test. These will be more stateful and akin to our tests which open and close sessions and transactions.
1.0
Improve user management features - ## Problem to Solve Our user management feature is fairly bare-bones right now, with TypeDB Cluster carrying the bulk of the weight for testing user management. ## Current Workaround We test much of user management with TypeDB Cluster integration tests. ## Proposed Solution We should expand the test suite and the available steps that can be executed during a test. These will be more stateful and akin to our tests which open and close sessions and transactions.
priority
improve user management features problem to solve our user management feature is fairly bare bones right now with typedb cluster carrying the bulk of the weight for testing user management current workaround we test much of user management with typedb cluster integration tests proposed solution we should expand the test suite and the available steps that can be executed during a test these will be more stateful and akin to our tests which open and close sessions and transactions
1
60,919
3,135,692,769
IssuesEvent
2015-09-10 16:19:32
PeerSay/Atlas
https://api.github.com/repos/PeerSay/Atlas
closed
When all topic's requirements are undefined (i.e. '?') - the overall topic grade should be undefined
Priority: Medium
If all topic grades are '?' then the overall of the topic should show '?' and not 0.
1.0
When all topic's requirements are undefined (i.e. '?') - the overall topic grade should be undefined - If all topic grades are '?' then the overall of the topic should show '?' and not 0.
priority
when all topic s requirements are undefined i e the overall topic grade should be undefined if all topic grades are then the overall of the topic should show and not
1
748,700
26,133,662,416
IssuesEvent
2022-12-29 09:16:21
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
closed
[Quest] Victory At Death's Breach!
NPC Quest - Cataclysm (1-60) Priority: Medium Status: Confirmed Bug Report from Discord
Unintendet OP — 12/10/2022 6:54 PM Quest = Victory At Death's Breach! ... Maybe I was to stupid for it but the Teleport didn't port me Up, or the Fly-guy to Fly Up is missing.
1.0
[Quest] Victory At Death's Breach! - Unintendet OP — 12/10/2022 6:54 PM Quest = Victory At Death's Breach! ... Maybe I was to stupid for it but the Teleport didn't port me Up, or the Fly-guy to Fly Up is missing.
priority
victory at death s breach unintendet op — pm quest victory at death s breach maybe i was to stupid for it but the teleport didn t port me up or the fly guy to fly up is missing
1
351,954
10,525,703,995
IssuesEvent
2019-09-30 15:33:47
forceworkbench/forceworkbench
https://api.github.com/repos/forceworkbench/forceworkbench
closed
Show realtionship name instead of Type for relationship query headers
Component-Query Priority-Medium Scheduled-Backlog bug imported
_Original author: ryan.bra...@gmail.com (February 01, 2009 11:11:47)_ For a query like [SELECT Opportunity.LastModifiedBy.Name FROM OpportunityContactRole], don't show Opportunity.User.Name. Not sure if this is going to be possible because of the way the SOAP client parses the tags _Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=146_
1.0
Show realtionship name instead of Type for relationship query headers - _Original author: ryan.bra...@gmail.com (February 01, 2009 11:11:47)_ For a query like [SELECT Opportunity.LastModifiedBy.Name FROM OpportunityContactRole], don't show Opportunity.User.Name. Not sure if this is going to be possible because of the way the SOAP client parses the tags _Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=146_
priority
show realtionship name instead of type for relationship query headers original author ryan bra gmail com february for a query like select opportunity lastmodifiedby name from opportunitycontactrole don t show opportunity user name not sure if this is going to be possible because of the way the soap client parses the tags original issue
1
351,921
10,525,701,180
IssuesEvent
2019-09-30 15:33:26
forceworkbench/forceworkbench
https://api.github.com/repos/forceworkbench/forceworkbench
closed
add SOQL Pagination support to query builder
Component-Query Priority-Medium Scheduled-Backlog enhancement imported
_Original author: ryan.bra...@gmail.com (August 05, 2012 00:22:33)_ SOQL Pagination—Generally Available SOQL Pagination enables you to specify the starting row of the result set your SOQL query returns. You use SOQL Pagination via the SOQL OFFSET clause. Using SOQL OFFSET is helpful for paging through large result sets when you need to quickly jump to a particular subset of the entire results. As the offset calculation is done on the server and only the result subset is returned, using OFFSET is more efficient than retrieving the full result set and then filtering the results locally. For example, this SOQL query returns a result set that skips the first 10 rows: SELECT Name FROM Merchandise__c WHERE Price__c &gt; 5.0 ORDER BY Name LIMIT 100 OFFSET 10 The maximum offset is 2,000 rows. Requesting an offset greater than 2,000 will result in a NUMBER_OUTSIDE_VALID_RANGE error. _Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=584_
1.0
add SOQL Pagination support to query builder - _Original author: ryan.bra...@gmail.com (August 05, 2012 00:22:33)_ SOQL Pagination—Generally Available SOQL Pagination enables you to specify the starting row of the result set your SOQL query returns. You use SOQL Pagination via the SOQL OFFSET clause. Using SOQL OFFSET is helpful for paging through large result sets when you need to quickly jump to a particular subset of the entire results. As the offset calculation is done on the server and only the result subset is returned, using OFFSET is more efficient than retrieving the full result set and then filtering the results locally. For example, this SOQL query returns a result set that skips the first 10 rows: SELECT Name FROM Merchandise__c WHERE Price__c &gt; 5.0 ORDER BY Name LIMIT 100 OFFSET 10 The maximum offset is 2,000 rows. Requesting an offset greater than 2,000 will result in a NUMBER_OUTSIDE_VALID_RANGE error. _Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=584_
priority
add soql pagination support to query builder original author ryan bra gmail com august soql pagination—generally available soql pagination enables you to specify the starting row of the result set your soql query returns you use soql pagination via the soql offset clause using soql offset is helpful for paging through large result sets when you need to quickly jump to a particular subset of the entire results as the offset calculation is done on the server and only the result subset is returned using offset is more efficient than retrieving the full result set and then filtering the results locally for example this soql query returns a result set that skips the first rows select name from merchandise c where price c gt order by name limit offset the maximum offset is rows requesting an offset greater than will result in a number outside valid range error original issue
1
397,247
11,725,717,475
IssuesEvent
2020-03-10 13:26:32
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio-ui] Add a link from Studio to acknowledgements page in docs
priority: medium task
**Is your feature request related to a problem? Please describe.** Add a link from Studio to acknowledgements page in docs **Describe the solution you'd like** Please add a link from the `Help` -> `About` screen in Studio with the link pointing to https://docs.craftercms.org/en/3.1/acknowledgements/index.html (Something like the screen below, the message is not centered, but that's the idea =) ) ![Screen Shot 2020-03-04 at 2 05 11 PM](https://user-images.githubusercontent.com/25483966/75913780-89f9c280-5e21-11ea-8530-e9e0e514f03f.png) `Crafter CMS is made possible by these other_open source software projects_ ` where `_open source software projects_` points to the acknowledgements in the docs **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
[studio-ui] Add a link from Studio to acknowledgements page in docs - **Is your feature request related to a problem? Please describe.** Add a link from Studio to acknowledgements page in docs **Describe the solution you'd like** Please add a link from the `Help` -> `About` screen in Studio with the link pointing to https://docs.craftercms.org/en/3.1/acknowledgements/index.html (Something like the screen below, the message is not centered, but that's the idea =) ) ![Screen Shot 2020-03-04 at 2 05 11 PM](https://user-images.githubusercontent.com/25483966/75913780-89f9c280-5e21-11ea-8530-e9e0e514f03f.png) `Crafter CMS is made possible by these other_open source software projects_ ` where `_open source software projects_` points to the acknowledgements in the docs **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
priority
add a link from studio to acknowledgements page in docs is your feature request related to a problem please describe add a link from studio to acknowledgements page in docs describe the solution you d like please add a link from the help about screen in studio with the link pointing to something like the screen below the message is not centered but that s the idea crafter cms is made possible by these other open source software projects where open source software projects points to the acknowledgements in the docs describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
1
758,289
26,548,974,730
IssuesEvent
2023-01-20 05:01:38
apimatic/go-core-runtime
https://api.github.com/repos/apimatic/go-core-runtime
opened
[FEATURE REQUEST] - Add support for Retries and Backoff
enhancement priority-medium
Go's native HTTP client does not support Retrying. Retrying allows an application to recover from transient failures while trying to reach out to a resource. So adding support for Retries and Backoff will only improve the API calls made.
1.0
[FEATURE REQUEST] - Add support for Retries and Backoff - Go's native HTTP client does not support Retrying. Retrying allows an application to recover from transient failures while trying to reach out to a resource. So adding support for Retries and Backoff will only improve the API calls made.
priority
add support for retries and backoff go s native http client does not support retrying retrying allows an application to recover from transient failures while trying to reach out to a resource so adding support for retries and backoff will only improve the api calls made
1
435,337
12,534,423,650
IssuesEvent
2020-06-04 19:24:24
Energy-Innovation/eps-us
https://api.github.com/repos/Energy-Innovation/eps-us
opened
Allow carbon tax rate to be set separately for district heat & hydrogen sector, instead of inheriting rate from industry sector
medium priority
The district heat and hydrogen sector uses the same fuels as the industry sector. When the carbon tax lever was designed, it was assumed the tax was levied upstream on these fuels, and so the rate seen by industrial facilities and district heat/hydrogen facilities would be the same. But some countries consider district heat (at least, maybe also hydrogen supply) to be part of the energy sector and may exempt them from carbon taxes or tax them differently than the industry sector. Change the carbon tax lever so it is set separately for the district heat and hydrogen supply sector, rather than inheriting the industry sector's carbon tax rate.
1.0
Allow carbon tax rate to be set separately for district heat & hydrogen sector, instead of inheriting rate from industry sector - The district heat and hydrogen sector uses the same fuels as the industry sector. When the carbon tax lever was designed, it was assumed the tax was levied upstream on these fuels, and so the rate seen by industrial facilities and district heat/hydrogen facilities would be the same. But some countries consider district heat (at least, maybe also hydrogen supply) to be part of the energy sector and may exempt them from carbon taxes or tax them differently than the industry sector. Change the carbon tax lever so it is set separately for the district heat and hydrogen supply sector, rather than inheriting the industry sector's carbon tax rate.
priority
allow carbon tax rate to be set separately for district heat hydrogen sector instead of inheriting rate from industry sector the district heat and hydrogen sector uses the same fuels as the industry sector when the carbon tax lever was designed it was assumed the tax was levied upstream on these fuels and so the rate seen by industrial facilities and district heat hydrogen facilities would be the same but some countries consider district heat at least maybe also hydrogen supply to be part of the energy sector and may exempt them from carbon taxes or tax them differently than the industry sector change the carbon tax lever so it is set separately for the district heat and hydrogen supply sector rather than inheriting the industry sector s carbon tax rate
1
280,066
8,677,745,211
IssuesEvent
2018-11-30 17:40:38
Killeroo/PowerPing
https://api.github.com/repos/Killeroo/PowerPing
closed
Fix Control C event handler
Bug Medium Priority
Event C handler currently not working properly, works with delay/lag Responsiveness tested in: - _Sending Normally_ (requires more testing) - Graph - _Flood_ (requires more testing) - Scan
1.0
Fix Control C event handler - Event C handler currently not working properly, works with delay/lag Responsiveness tested in: - _Sending Normally_ (requires more testing) - Graph - _Flood_ (requires more testing) - Scan
priority
fix control c event handler event c handler currently not working properly works with delay lag responsiveness tested in sending normally requires more testing graph flood requires more testing scan
1
416,585
12,148,927,724
IssuesEvent
2020-04-24 15:20:09
scality/metalk8s
https://api.github.com/repos/scality/metalk8s
opened
Add solution get in python CLI
complexity:easy complexity:medium kind:enhancement priority:low topic:solutions
**Component**: 'salt', 'cli', 'solution' **Why this is needed**: Centralized CLI **What should be done**: MetalK8s solution get command should be part of `metalk8sctl` python CLI ``` metalk8sctl solution get <name> ``` This command will just show information about an available solution. Name argument is not mandatory, if no name then all available and activated solutions are listed. NOTE: We may want to handle regexp matching (not mandatory for first iteration) **Implementation proposal** (strongly recommended): This command will just retrieve all solutions information according to the CLI arguments. Check design documentation about Centralized CLI for more detail.
1.0
Add solution get in python CLI - **Component**: 'salt', 'cli', 'solution' **Why this is needed**: Centralized CLI **What should be done**: MetalK8s solution get command should be part of `metalk8sctl` python CLI ``` metalk8sctl solution get <name> ``` This command will just show information about an available solution. Name argument is not mandatory, if no name then all available and activated solutions are listed. NOTE: We may want to handle regexp matching (not mandatory for first iteration) **Implementation proposal** (strongly recommended): This command will just retrieve all solutions information according to the CLI arguments. Check design documentation about Centralized CLI for more detail.
priority
add solution get in python cli component salt cli solution why this is needed centralized cli what should be done solution get command should be part of python cli solution get this command will just show information about an available solution name argument is not mandatory if no name then all available and activated solutions are listed note we may want to handle regexp matching not mandatory for first iteration implementation proposal strongly recommended this command will just retrieve all solutions information according to the cli arguments check design documentation about centralized cli for more detail
1
463,542
13,283,460,503
IssuesEvent
2020-08-24 03:17:51
mandarineorg/mandarinets
https://api.github.com/repos/mandarineorg/mandarinets
closed
Add Support for Middleware targetting methods instead of controllers
Medium Priority
**[Context]** Right now, in order to use Mandarine's middleware, you need to create a `@Middleware` component targetting a RegExp of a route. What we want to do now is to add a feature for the developer to add middleware to the method. This can be located under the options of the Http method ex: `@GET(route, options)` where options would have a middleware field, or we could also create a decorator `UseMiddleware`.
1.0
Add Support for Middleware targetting methods instead of controllers - **[Context]** Right now, in order to use Mandarine's middleware, you need to create a `@Middleware` component targetting a RegExp of a route. What we want to do now is to add a feature for the developer to add middleware to the method. This can be located under the options of the Http method ex: `@GET(route, options)` where options would have a middleware field, or we could also create a decorator `UseMiddleware`.
priority
add support for middleware targetting methods instead of controllers right now in order to use mandarine s middleware you need to create a middleware component targetting a regexp of a route what we want to do now is to add a feature for the developer to add middleware to the method this can be located under the options of the http method ex get route options where options would have a middleware field or we could also create a decorator usemiddleware
1
541,200
15,823,075,090
IssuesEvent
2021-04-05 23:48:22
dtcenter/METplus
https://api.github.com/repos/dtcenter/METplus
reopened
Apply TCPairs, TCStat, CyclonePlotter, GridStat to EMC 2020 Data
component: use case configuration component: use case wrapper priority: medium requestor: NCAR requestor: NOAA/other type: enhancement type: task
*Replace italics below with details for this issue.* ## Describe the Task ## *Learn how to use TCPairs, TCStat, CyclonePlotter, and GridStat in the context of cyclones, using existing sample data from GitHub* ### Time Estimate ### *Estimate the amount of work required here.* *3 calendar days of work spread out over December.* *Future work will either extend this task or create another one* ### Sub-Issues ### Consider breaking the task down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### * John O will provide more data probably in January.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [x] Select **engineer(s)** or **no engineer** required - [x] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label - [ ] Select **milestone** to next major version milestone or "Future Versions" ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Task Checklist ## See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues** - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
1.0
Apply TCPairs, TCStat, CyclonePlotter, GridStat to EMC 2020 Data - *Replace italics below with details for this issue.* ## Describe the Task ## *Learn how to use TCPairs, TCStat, CyclonePlotter, and GridStat in the context of cyclones, using existing sample data from GitHub* ### Time Estimate ### *Estimate the amount of work required here.* *3 calendar days of work spread out over December.* *Future work will either extend this task or create another one* ### Sub-Issues ### Consider breaking the task down into sub-issues. - [ ] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### * John O will provide more data probably in January.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [x] Select **engineer(s)** or **no engineer** required - [x] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label - [ ] Select **milestone** to next major version milestone or "Future Versions" ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Task Checklist ## See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues** - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
priority
apply tcpairs tcstat cycloneplotter gridstat to emc data replace italics below with details for this issue describe the task learn how to use tcpairs tcstat cycloneplotter and gridstat in the context of cyclones using existing sample data from github time estimate estimate the amount of work required here calendar days of work spread out over december future work will either extend this task or create another one sub issues consider breaking the task down into sub issues add a checkbox for each sub issue here relevant deadlines john o will provide more data probably in january funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone review projects and select relevant repository and organization ones or add alert need project assignment label select milestone to next major version milestone or future versions define related issue s consider the impact to the other metplus components task checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s project s milestone and linked issues iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
1
33,380
2,764,522,937
IssuesEvent
2015-04-29 15:50:26
IQSS/dataverse
https://api.github.com/repos/IQSS/dataverse
closed
Ability to Publish Dataverse when Trying to Publish Dataset (if user has permission to do so)
Priority: Medium Status: QA Type: Feature
--- Author Name: **Elizabeth Quigley** (@eaquigley) Original Redmine Issue: 4065, https://redmine.hmdc.harvard.edu/issues/4065 Original Date: 2014-06-03 --- Need to add in functionality so that a user can publish a dataverse and a dataset at the same time if they are trying to publish a dataset in an unpublished dataverse. This will relate to permissions as the system will be able to identify that a user has access to publish the dataverse. If not, then we need messaging that states why they can't publish the dataset.
1.0
Ability to Publish Dataverse when Trying to Publish Dataset (if user has permission to do so) - --- Author Name: **Elizabeth Quigley** (@eaquigley) Original Redmine Issue: 4065, https://redmine.hmdc.harvard.edu/issues/4065 Original Date: 2014-06-03 --- Need to add in functionality so that a user can publish a dataverse and a dataset at the same time if they are trying to publish a dataset in an unpublished dataverse. This will relate to permissions as the system will be able to identify that a user has access to publish the dataverse. If not, then we need messaging that states why they can't publish the dataset.
priority
ability to publish dataverse when trying to publish dataset if user has permission to do so author name elizabeth quigley eaquigley original redmine issue original date need to add in functionality so that a user can publish a dataverse and a dataset at the same time if they are trying to publish a dataset in an unpublished dataverse this will relate to permissions as the system will be able to identify that a user has access to publish the dataverse if not then we need messaging that states why they can t publish the dataset
1
423,766
12,301,890,926
IssuesEvent
2020-05-11 16:05:04
wri/gfw-mapbuilder
https://api.github.com/repos/wri/gfw-mapbuilder
closed
Loading state for analysis tab
4.x Upgrade medium priority
Clicking on poly on the map and going to analysis tab should have a better loading state. Maybe a spinner would be good? Right now it is just saying "geostore is registering". See the spinner which shows when we click "Run analsysis" for the vega chart.
1.0
Loading state for analysis tab - Clicking on poly on the map and going to analysis tab should have a better loading state. Maybe a spinner would be good? Right now it is just saying "geostore is registering". See the spinner which shows when we click "Run analsysis" for the vega chart.
priority
loading state for analysis tab clicking on poly on the map and going to analysis tab should have a better loading state maybe a spinner would be good right now it is just saying geostore is registering see the spinner which shows when we click run analsysis for the vega chart
1
818,328
30,683,813,084
IssuesEvent
2023-07-26 10:56:03
netdata/netdata-cloud
https://api.github.com/repos/netdata/netdata-cloud
closed
[Bug]: overview search bar looks misalligned
bug priority/medium cloud-frontend
### Bug description ![image](https://github.com/netdata/netdata-cloud/assets/2178292/22a35ac0-88ef-4e17-a79c-c3fa75fb5be7) ### Expected behavior like it was before ### Steps to reproduce 1. go [here](https://app.netdata.cloud/spaces/netdata-demo/rooms/all-nodes/overview#metrics_correlation=false&after=-900&before=0&utc=Europe%2FLondon&offset=%2B1&timezoneName=Edinburgh%2C%20London&modal=&modalTab=&d8a4e0c5-7c79-4145-900e-83a9f06fcb6a--chartName=menu_system) ### Screenshots _No response_ ### Error Logs _No response_ ### Desktop OS: [e.g. iOS] Browser [e.g. chrome, safari] Browser Version [e.g. 22] ### Additional context _No response_
1.0
[Bug]: overview search bar looks misalligned - ### Bug description ![image](https://github.com/netdata/netdata-cloud/assets/2178292/22a35ac0-88ef-4e17-a79c-c3fa75fb5be7) ### Expected behavior like it was before ### Steps to reproduce 1. go [here](https://app.netdata.cloud/spaces/netdata-demo/rooms/all-nodes/overview#metrics_correlation=false&after=-900&before=0&utc=Europe%2FLondon&offset=%2B1&timezoneName=Edinburgh%2C%20London&modal=&modalTab=&d8a4e0c5-7c79-4145-900e-83a9f06fcb6a--chartName=menu_system) ### Screenshots _No response_ ### Error Logs _No response_ ### Desktop OS: [e.g. iOS] Browser [e.g. chrome, safari] Browser Version [e.g. 22] ### Additional context _No response_
priority
overview search bar looks misalligned bug description expected behavior like it was before steps to reproduce go screenshots no response error logs no response desktop os browser browser version additional context no response
1
346,966
10,422,231,303
IssuesEvent
2019-09-16 08:32:26
zdnscloud/singlecloud
https://api.github.com/repos/zdnscloud/singlecloud
opened
lvm存储被删除后, 依然会出现在storageclass列表里
bug priority: Medium
分支: add-cluster-scversion 1. 创建一个名称为lvmstorage的lvm存储, 2. 删除这个存储, 3. 创建一个名称为lvm的lvm存储, 4. 删除这个存储, 结果: get storageclass 时, 被删除的lvm和lvmstorage会出现在列表里. ![image](https://user-images.githubusercontent.com/47511655/64943737-2265b000-d89f-11e9-8373-74105f658651.png) ![image](https://user-images.githubusercontent.com/47511655/64943853-6062d400-d89f-11e9-9855-ab5ce7b7d51b.png) ![image](https://user-images.githubusercontent.com/47511655/64943728-1b3ea200-d89f-11e9-8ba7-a7683f882d93.png)
1.0
lvm存储被删除后, 依然会出现在storageclass列表里 - 分支: add-cluster-scversion 1. 创建一个名称为lvmstorage的lvm存储, 2. 删除这个存储, 3. 创建一个名称为lvm的lvm存储, 4. 删除这个存储, 结果: get storageclass 时, 被删除的lvm和lvmstorage会出现在列表里. ![image](https://user-images.githubusercontent.com/47511655/64943737-2265b000-d89f-11e9-8373-74105f658651.png) ![image](https://user-images.githubusercontent.com/47511655/64943853-6062d400-d89f-11e9-9855-ab5ce7b7d51b.png) ![image](https://user-images.githubusercontent.com/47511655/64943728-1b3ea200-d89f-11e9-8ba7-a7683f882d93.png)
priority
lvm存储被删除后 依然会出现在storageclass列表里 分支 add cluster scversion 创建一个名称为lvmstorage的lvm存储 删除这个存储 创建一个名称为lvm的lvm存储 删除这个存储 结果 get storageclass 时 被删除的lvm和lvmstorage会出现在列表里
1
46,456
2,957,660,074
IssuesEvent
2015-07-08 17:27:05
chef/chef
https://api.github.com/repos/chef/chef
closed
MacOSX Service provider - Running services are not restarted
Bug Chef Core Medium Priority
### Version: 12.2.1 ### Environment: MacOSX Any cookbook that uses the provider service/macosx.rb source: https://github.com/chef/chef/blob/master/lib/chef/provider/service/macosx.rb ### Scenario: Restart services when they are running or stopped. It just restart the service when it is stopped. However when it is running just stop it. ### Steps to Reproduce: Use the provider service in some cookbook and run it in a MacOSX. Example: service "SOME-SERVICE" do action :restart end ### Expected Result: Expect to the service be restarted. ### Actual Result: When the service is running, it is stopped but not started. Debug message: [2015-04-01T14:41:25+01:00] DEBUG: service[thoughtworks.go-agent-1] already running, not starting [2015-04-01T14:41:25+01:00] INFO: service[thoughtworks.go-agent-1] restarted When the service is stopped, it is started correctly. ### Problem in source: When start_service and stop_service methods are called it is checked if the attribute 'running' is 'true' or 'false' respectively. However this attribute is defined at the beginning of the service instance definition and not updated if it the service is stopped or started. When you restart the service when it is running, the flag is defined as "true" and the stop_service stops the service. However the start_service doesn't do nothing because the 'running' attribute is still 'true' (not updated). ### Proposed solution: Remove in the start_service method (line 81 source code) the conditional: if @current_resource.running Chef::Log.debug("#{@new_resource} already running, not starting") And to be consistent remove also in the stop_service method (line 93 source code) the conditional: unless @current_resource.running Chef::Log.debug("#{@new_resource} not running, not stopping")
1.0
MacOSX Service provider - Running services are not restarted - ### Version: 12.2.1 ### Environment: MacOSX Any cookbook that uses the provider service/macosx.rb source: https://github.com/chef/chef/blob/master/lib/chef/provider/service/macosx.rb ### Scenario: Restart services when they are running or stopped. It just restart the service when it is stopped. However when it is running just stop it. ### Steps to Reproduce: Use the provider service in some cookbook and run it in a MacOSX. Example: service "SOME-SERVICE" do action :restart end ### Expected Result: Expect to the service be restarted. ### Actual Result: When the service is running, it is stopped but not started. Debug message: [2015-04-01T14:41:25+01:00] DEBUG: service[thoughtworks.go-agent-1] already running, not starting [2015-04-01T14:41:25+01:00] INFO: service[thoughtworks.go-agent-1] restarted When the service is stopped, it is started correctly. ### Problem in source: When start_service and stop_service methods are called it is checked if the attribute 'running' is 'true' or 'false' respectively. However this attribute is defined at the beginning of the service instance definition and not updated if it the service is stopped or started. When you restart the service when it is running, the flag is defined as "true" and the stop_service stops the service. However the start_service doesn't do nothing because the 'running' attribute is still 'true' (not updated). ### Proposed solution: Remove in the start_service method (line 81 source code) the conditional: if @current_resource.running Chef::Log.debug("#{@new_resource} already running, not starting") And to be consistent remove also in the stop_service method (line 93 source code) the conditional: unless @current_resource.running Chef::Log.debug("#{@new_resource} not running, not stopping")
priority
macosx service provider running services are not restarted version environment macosx any cookbook that uses the provider service macosx rb source scenario restart services when they are running or stopped it just restart the service when it is stopped however when it is running just stop it steps to reproduce use the provider service in some cookbook and run it in a macosx example service some service do action restart end expected result expect to the service be restarted actual result when the service is running it is stopped but not started debug message debug service already running not starting info service restarted when the service is stopped it is started correctly problem in source when start service and stop service methods are called it is checked if the attribute running is true or false respectively however this attribute is defined at the beginning of the service instance definition and not updated if it the service is stopped or started when you restart the service when it is running the flag is defined as true and the stop service stops the service however the start service doesn t do nothing because the running attribute is still true not updated proposed solution remove in the start service method line source code the conditional if current resource running chef log debug new resource already running not starting and to be consistent remove also in the stop service method line source code the conditional unless current resource running chef log debug new resource not running not stopping
1
502,031
14,538,545,208
IssuesEvent
2020-12-15 10:37:03
bdecon/econ_data
https://api.github.com/repos/bdecon/econ_data
opened
bd CPS: New variable DISCTYPE
new variable priority: medium
The BLS definition for discouraged workers covers only those who want a job, looked for one in the last 12 months, are available to take one, but don't think there is one available for them. In the NILFREASON variable I've defined discouraged workers as anyone who wants a job but isn't in the labor force. To make the two match, I want to create a new variable called DISCTYPE that is defined as: `Discouraged` where `PRDISC == 1`, `Marginally attached` where `PRDISC == 2`, `No recent search` where `PRJOBSEA == 5`, and `Unavailable` where `PRDISC == 3`. It will be defined only for those with `NILFREASON == "Discouraged"`. I can then remove some of the other variables, MRGNATT and PRDISC, for example.
1.0
bd CPS: New variable DISCTYPE - The BLS definition for discouraged workers covers only those who want a job, looked for one in the last 12 months, are available to take one, but don't think there is one available for them. In the NILFREASON variable I've defined discouraged workers as anyone who wants a job but isn't in the labor force. To make the two match, I want to create a new variable called DISCTYPE that is defined as: `Discouraged` where `PRDISC == 1`, `Marginally attached` where `PRDISC == 2`, `No recent search` where `PRJOBSEA == 5`, and `Unavailable` where `PRDISC == 3`. It will be defined only for those with `NILFREASON == "Discouraged"`. I can then remove some of the other variables, MRGNATT and PRDISC, for example.
priority
bd cps new variable disctype the bls definition for discouraged workers covers only those who want a job looked for one in the last months are available to take one but don t think there is one available for them in the nilfreason variable i ve defined discouraged workers as anyone who wants a job but isn t in the labor force to make the two match i want to create a new variable called disctype that is defined as discouraged where prdisc marginally attached where prdisc no recent search where prjobsea and unavailable where prdisc it will be defined only for those with nilfreason discouraged i can then remove some of the other variables mrgnatt and prdisc for example
1
251,724
8,026,010,231
IssuesEvent
2018-07-27 01:13:40
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
USER ISSUE: To many Skillpoints
Medium Priority
**Version:** 0.7.2.3 beta **Steps to Reproduce:** I got 11.914 Skillpoint in ine day... i didn't change the server options... **Expected behavior:** I should have about 30 skillpoints.... **Actual behavior:**
1.0
USER ISSUE: To many Skillpoints - **Version:** 0.7.2.3 beta **Steps to Reproduce:** I got 11.914 Skillpoint in ine day... i didn't change the server options... **Expected behavior:** I should have about 30 skillpoints.... **Actual behavior:**
priority
user issue to many skillpoints version beta steps to reproduce i got skillpoint in ine day i didn t change the server options expected behavior i should have about skillpoints actual behavior
1
711,788
24,475,399,382
IssuesEvent
2022-10-08 05:06:21
roq-trading/roq-issues
https://api.github.com/repos/roq-trading/roq-issues
closed
[roq-server] Now using --cache_dir instead of --auth_cache_dir
chore medium priority
Authentication tokens will be written to `<auth_cache_dir>/<name>/auth/`. This change was done to align with changes needed to cache configuration history, see * #287
1.0
[roq-server] Now using --cache_dir instead of --auth_cache_dir - Authentication tokens will be written to `<auth_cache_dir>/<name>/auth/`. This change was done to align with changes needed to cache configuration history, see * #287
priority
now using cache dir instead of auth cache dir authentication tokens will be written to auth this change was done to align with changes needed to cache configuration history see
1
167,215
6,334,526,072
IssuesEvent
2017-07-26 16:50:59
ACEmulator/ACE
https://api.github.com/repos/ACEmulator/ACE
closed
Background maintenance thread
priority:medium
We need some sort of low priority thread to run and periodically perform routine maintenance tasks such as removing characters marked for deletion.
1.0
Background maintenance thread - We need some sort of low priority thread to run and periodically perform routine maintenance tasks such as removing characters marked for deletion.
priority
background maintenance thread we need some sort of low priority thread to run and periodically perform routine maintenance tasks such as removing characters marked for deletion
1
722,263
24,856,607,169
IssuesEvent
2022-10-27 03:15:15
JasonBock/Rocks
https://api.github.com/repos/JasonBock/Rocks
opened
Projected Types With Open Generics Are Not Handled Correctly
bug Medium Priority
To reproduce: ```csharp using Rocks; using System; public interface IPixel { } public interface IPixel<TSelf> : IPixel, IEquatable<TSelf> where TSelf : unmanaged, IPixel<TSelf> { } public interface IUseSpanWithOpenGeneric { void From<TSourcePixel>( ReadOnlySpan<TSourcePixel> sourcePixels) where TSourcePixel : unmanaged, IPixel<TSourcePixel>; } public static class Test { public static void Go() { var expectations = Rock.Create<IUseSpanWithOpenGeneric>(); } } ``` This will lead to the following errors: ``` ``` This was found on ImageSharp's `PixelOperation` type.
1.0
Projected Types With Open Generics Are Not Handled Correctly - To reproduce: ```csharp using Rocks; using System; public interface IPixel { } public interface IPixel<TSelf> : IPixel, IEquatable<TSelf> where TSelf : unmanaged, IPixel<TSelf> { } public interface IUseSpanWithOpenGeneric { void From<TSourcePixel>( ReadOnlySpan<TSourcePixel> sourcePixels) where TSourcePixel : unmanaged, IPixel<TSourcePixel>; } public static class Test { public static void Go() { var expectations = Rock.Create<IUseSpanWithOpenGeneric>(); } } ``` This will lead to the following errors: ``` ``` This was found on ImageSharp's `PixelOperation` type.
priority
projected types with open generics are not handled correctly to reproduce csharp using rocks using system public interface ipixel public interface ipixel ipixel iequatable where tself unmanaged ipixel public interface iusespanwithopengeneric void from readonlyspan sourcepixels where tsourcepixel unmanaged ipixel public static class test public static void go var expectations rock create this will lead to the following errors this was found on imagesharp s pixeloperation type
1
674,516
23,054,367,233
IssuesEvent
2022-07-25 02:05:43
FEeasy404/GameUs
https://api.github.com/repos/FEeasy404/GameUs
closed
게시글 작성 페이지 및 게시물 업로드 구현
✨Feat 🖐Priority: Medium
## 추가 기능 설명 3.1.11 게시글 작성 페이지 - 게시글을 작성할 수 있는 페이지로, 하단 메뉴바에서 게시글 작성 을 클릭하면 표시됩니다. - 글이 입력되거나 사진이 업로드 되면 업로드 버튼이 활성화되고, 버튼을 누르면 게시글이 업로드됩니다. ## 할 일 - [x] 게시글 작성 페이지 UI - [x] 하단 메뉴바 클릭시 게시글 작성 페이지 뜨도록 구현 - [x] 글 입력시 업로드 버튼 활성화 구현 - [x] 버튼 클릭시 게시글 업로드 구현 ## ETC
1.0
게시글 작성 페이지 및 게시물 업로드 구현 - ## 추가 기능 설명 3.1.11 게시글 작성 페이지 - 게시글을 작성할 수 있는 페이지로, 하단 메뉴바에서 게시글 작성 을 클릭하면 표시됩니다. - 글이 입력되거나 사진이 업로드 되면 업로드 버튼이 활성화되고, 버튼을 누르면 게시글이 업로드됩니다. ## 할 일 - [x] 게시글 작성 페이지 UI - [x] 하단 메뉴바 클릭시 게시글 작성 페이지 뜨도록 구현 - [x] 글 입력시 업로드 버튼 활성화 구현 - [x] 버튼 클릭시 게시글 업로드 구현 ## ETC
priority
게시글 작성 페이지 및 게시물 업로드 구현 추가 기능 설명 게시글 작성 페이지 게시글을 작성할 수 있는 페이지로 하단 메뉴바에서 게시글 작성 을 클릭하면 표시됩니다 글이 입력되거나 사진이 업로드 되면 업로드 버튼이 활성화되고 버튼을 누르면 게시글이 업로드됩니다 할 일 게시글 작성 페이지 ui 하단 메뉴바 클릭시 게시글 작성 페이지 뜨도록 구현 글 입력시 업로드 버튼 활성화 구현 버튼 클릭시 게시글 업로드 구현 etc
1
802,699
29,044,455,652
IssuesEvent
2023-05-13 11:25:15
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Framing (borders.c): can't enter custom ratio in dropdown
priority: medium reproduce: confirmed scope: UI bug: pending release notes: pending
### Describe the bug The tooltip for the Aspect dropdown (not the Aspect ratio slider) says one can enter a custom ratio in the form w:h. This is not actually possible. While one *can* enter a text string instead of selecting a predefined entry, the text is not actually evaluated - it is displayed for the dropdown but the aspect ratio slider is not updated. ### Steps to reproduce 1. Open the Framing module on an image 2. Click (or right-click) on "aspect" 3. Type a value - decimal number, w:h, or w/h 4. See the typed text displayed when the dropdown collapses 5. Note that the aspect ratio slider has not changed ### Expected behavior The behavior should match the tooltip, or custom text should be disabled for the dropdown and the tooltip for the aspect ratio slider updated. A one-line change in common/calculator.c would permit the bauhaus to accept "7:6" as well as "7/6" for ratios. ### Logfile | Screenshot | Screencast _No response_ ### Commit _No response_ ### Where did you install darktable from? self compiled ### darktable version master 4.3.0+2056 ### What OS are you using? Linux ### What is the version of your OS? Mageia 8.1 ### Describe your system? _No response_ ### Are you using OpenCL GPU in darktable? None ### If yes, what is the GPU card and driver? _No response_ ### Please provide additional context if applicable. You can attach files too, but might need to rename to .txt or .zip _No response_
1.0
Framing (borders.c): can't enter custom ratio in dropdown - ### Describe the bug The tooltip for the Aspect dropdown (not the Aspect ratio slider) says one can enter a custom ratio in the form w:h. This is not actually possible. While one *can* enter a text string instead of selecting a predefined entry, the text is not actually evaluated - it is displayed for the dropdown but the aspect ratio slider is not updated. ### Steps to reproduce 1. Open the Framing module on an image 2. Click (or right-click) on "aspect" 3. Type a value - decimal number, w:h, or w/h 4. See the typed text displayed when the dropdown collapses 5. Note that the aspect ratio slider has not changed ### Expected behavior The behavior should match the tooltip, or custom text should be disabled for the dropdown and the tooltip for the aspect ratio slider updated. A one-line change in common/calculator.c would permit the bauhaus to accept "7:6" as well as "7/6" for ratios. ### Logfile | Screenshot | Screencast _No response_ ### Commit _No response_ ### Where did you install darktable from? self compiled ### darktable version master 4.3.0+2056 ### What OS are you using? Linux ### What is the version of your OS? Mageia 8.1 ### Describe your system? _No response_ ### Are you using OpenCL GPU in darktable? None ### If yes, what is the GPU card and driver? _No response_ ### Please provide additional context if applicable. You can attach files too, but might need to rename to .txt or .zip _No response_
priority
framing borders c can t enter custom ratio in dropdown describe the bug the tooltip for the aspect dropdown not the aspect ratio slider says one can enter a custom ratio in the form w h this is not actually possible while one can enter a text string instead of selecting a predefined entry the text is not actually evaluated it is displayed for the dropdown but the aspect ratio slider is not updated steps to reproduce open the framing module on an image click or right click on aspect type a value decimal number w h or w h see the typed text displayed when the dropdown collapses note that the aspect ratio slider has not changed expected behavior the behavior should match the tooltip or custom text should be disabled for the dropdown and the tooltip for the aspect ratio slider updated a one line change in common calculator c would permit the bauhaus to accept as well as for ratios logfile screenshot screencast no response commit no response where did you install darktable from self compiled darktable version master what os are you using linux what is the version of your os mageia describe your system no response are you using opencl gpu in darktable none if yes what is the gpu card and driver no response please provide additional context if applicable you can attach files too but might need to rename to txt or zip no response
1
649,629
21,316,757,672
IssuesEvent
2022-04-16 12:16:26
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Extra error reported in console when converting topic to HTML5 with args.filter
bug priority/medium plugin/html5 stale
Publishing any DITA topic to HTML5 using an args.filter pointing to some ditaval file results in an extra error in the console output: [xslt] I/O error reported by XML parser processing file:/${dita.input.valfile.url}: \${dita.input.valfile.url} (The system cannot find the file specified) This seems to be the case because this ANT target: <target name="html5.topic.init" unless="noMap"> which populates the "dita.input.valfile.url" param is executed only when DITA Maps are published, and not individual topics.
1.0
Extra error reported in console when converting topic to HTML5 with args.filter - Publishing any DITA topic to HTML5 using an args.filter pointing to some ditaval file results in an extra error in the console output: [xslt] I/O error reported by XML parser processing file:/${dita.input.valfile.url}: \${dita.input.valfile.url} (The system cannot find the file specified) This seems to be the case because this ANT target: <target name="html5.topic.init" unless="noMap"> which populates the "dita.input.valfile.url" param is executed only when DITA Maps are published, and not individual topics.
priority
extra error reported in console when converting topic to with args filter publishing any dita topic to using an args filter pointing to some ditaval file results in an extra error in the console output i o error reported by xml parser processing file dita input valfile url dita input valfile url the system cannot find the file specified this seems to be the case because this ant target which populates the dita input valfile url param is executed only when dita maps are published and not individual topics
1
624,024
19,684,775,856
IssuesEvent
2022-01-11 20:44:44
GameFreedomGG/Sindragosa
https://api.github.com/repos/GameFreedomGG/Sindragosa
closed
[Profession][Item] Mining - Thorium Nodes & Arcane Crystals
Item Priority: Medium Professions Status: Confirmed Bug (Awaiting Development)
Decription: Player reported that drops for Thorium are messed up. Currently nearly all the time Arcane Crystal's drop, but Large Opals, Blue Sapphires etc should drop too. How it works: Only Arcane Crystal's are dropping. How it should work: Arcane 10% drop rate. Blue Sapphire 3%. Huge Emerald 3%. Large Opal 3%. Azerothian Diamond 3%. Source (you should point out proofs of your report, please give us some source): https://wotlk-twinhead.twinstar.cz/?object=175404
1.0
[Profession][Item] Mining - Thorium Nodes & Arcane Crystals - Decription: Player reported that drops for Thorium are messed up. Currently nearly all the time Arcane Crystal's drop, but Large Opals, Blue Sapphires etc should drop too. How it works: Only Arcane Crystal's are dropping. How it should work: Arcane 10% drop rate. Blue Sapphire 3%. Huge Emerald 3%. Large Opal 3%. Azerothian Diamond 3%. Source (you should point out proofs of your report, please give us some source): https://wotlk-twinhead.twinstar.cz/?object=175404
priority
mining thorium nodes arcane crystals decription player reported that drops for thorium are messed up currently nearly all the time arcane crystal s drop but large opals blue sapphires etc should drop too how it works only arcane crystal s are dropping how it should work arcane drop rate blue sapphire huge emerald large opal azerothian diamond source you should point out proofs of your report please give us some source
1
798,636
28,291,784,807
IssuesEvent
2023-04-09 09:52:21
KDT3-final-project-team2/backend
https://api.github.com/repos/KDT3-final-project-team2/backend
closed
[Fix] 기업회원 지원자 목록에 지원서id 추가
For: API Priority: Medium Status: In Progress
## 🔨개발 할 기능 지원자 목록 출력 DTO에 지원서 id 추가하기 ## 🧩 세부 기능 - [ ] DTO에 applicationID 추가하기 ## 📖 참고 사항
1.0
[Fix] 기업회원 지원자 목록에 지원서id 추가 - ## 🔨개발 할 기능 지원자 목록 출력 DTO에 지원서 id 추가하기 ## 🧩 세부 기능 - [ ] DTO에 applicationID 추가하기 ## 📖 참고 사항
priority
기업회원 지원자 목록에 지원서id 추가 🔨개발 할 기능 지원자 목록 출력 dto에 지원서 id 추가하기 🧩 세부 기능 dto에 applicationid 추가하기 📖 참고 사항
1
708,063
24,328,922,804
IssuesEvent
2022-09-30 17:23:51
kleros/kleros-v2
https://api.github.com/repos/kleros/kleros-v2
closed
Minor bug: after changing a court's minStake, jurors with a smaller stake are still drawn
Priority: Medium Type: Bug :bug: Package: Contracts
The issue also exists in v1.
1.0
Minor bug: after changing a court's minStake, jurors with a smaller stake are still drawn - The issue also exists in v1.
priority
minor bug after changing a court s minstake jurors with a smaller stake are still drawn the issue also exists in
1
190,983
6,824,470,437
IssuesEvent
2017-11-08 06:23:41
certificate-helper/TLS-Inspector
https://api.github.com/repos/certificate-helper/TLS-Inspector
closed
App URI protocol broken
bug medium priority merged
**Affected Version:** Current **Is this a Test Flight version or the App Store version?** App Store **Device and iOS Version:** All **What steps will reproduce the problem?** 1. Use test app to open `certinspector://inspect/google.com` **What is the expected output?** Google.com shows **What do you see instead?** Nothing **Please provide any additional information below.** Should remove feature.
1.0
App URI protocol broken - **Affected Version:** Current **Is this a Test Flight version or the App Store version?** App Store **Device and iOS Version:** All **What steps will reproduce the problem?** 1. Use test app to open `certinspector://inspect/google.com` **What is the expected output?** Google.com shows **What do you see instead?** Nothing **Please provide any additional information below.** Should remove feature.
priority
app uri protocol broken affected version current is this a test flight version or the app store version app store device and ios version all what steps will reproduce the problem use test app to open certinspector inspect google com what is the expected output google com shows what do you see instead nothing please provide any additional information below should remove feature
1
548,712
16,074,198,331
IssuesEvent
2021-04-25 02:55:58
rich-iannone/pointblank
https://api.github.com/repos/rich-iannone/pointblank
opened
Provide styled console output when using `yaml_exec()`
Difficulty: [2] Intermediate Effort: [2] Medium Priority: [3] High Type: ★ Enhancement
When using `yaml_exec()` to process YAML agents and informants *en masse*, it would be nice to be notified in the console about what happened (during interactive sessions). We can use {cli}-formatted messages like elsewhere in the package. Also, the function should invisibly return *something* about what was written. Right now, it always returns `NULL`.
1.0
Provide styled console output when using `yaml_exec()` - When using `yaml_exec()` to process YAML agents and informants *en masse*, it would be nice to be notified in the console about what happened (during interactive sessions). We can use {cli}-formatted messages like elsewhere in the package. Also, the function should invisibly return *something* about what was written. Right now, it always returns `NULL`.
priority
provide styled console output when using yaml exec when using yaml exec to process yaml agents and informants en masse it would be nice to be notified in the console about what happened during interactive sessions we can use cli formatted messages like elsewhere in the package also the function should invisibly return something about what was written right now it always returns null
1
188,843
6,782,488,138
IssuesEvent
2017-10-30 08:20:55
compodoc/compodoc
https://api.github.com/repos/compodoc/compodoc
closed
[BUG] Includes more files than it should
1. Type: Bug Priority: Medium Status: Completed Time: ~1 hour
<!-- > Please follow the issue template below for bug reports and queries. > For issue, start the label of the title with [BUG] > For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant. --> ##### **Overview of the issue** This is my tsconfig: ``` { "compilerOptions": { "target": "es5", "module": "commonjs", "moduleResolution": "node", "sourceMap": true, "emitDecoratorMetadata": true, "experimentalDecorators": true, "lib": [ "es2015", "dom" ], "rootDir": ".", "outDir": "dist" }, "include": [ "app/**/*.ts" ], "exclude": [ "app/main.prod.ts" ] } ``` But compodoc still includes files from dist/ (the outDir) and other directories in the root directory, which leads to all my modules, components, etc. being included two times in the docs. The expected behavior is that compodoc only includes the files specified in "files" or "include" if either of those are specified. That's the way the typescript compiler handles it. ##### **Operating System, Node.js, npm, compodoc version(s)** Ubuntu 17.04, Node v8.5.0, npm v5.3.0, compodoc v1.0.1 ##### **Angular configuration, a `package.json` file in the root folder** Here is my project: https://github.com/tradity/tradity-client/ ##### **Compodoc installed globally or locally ?** compodoc is installed locally
1.0
[BUG] Includes more files than it should - <!-- > Please follow the issue template below for bug reports and queries. > For issue, start the label of the title with [BUG] > For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant. --> ##### **Overview of the issue** This is my tsconfig: ``` { "compilerOptions": { "target": "es5", "module": "commonjs", "moduleResolution": "node", "sourceMap": true, "emitDecoratorMetadata": true, "experimentalDecorators": true, "lib": [ "es2015", "dom" ], "rootDir": ".", "outDir": "dist" }, "include": [ "app/**/*.ts" ], "exclude": [ "app/main.prod.ts" ] } ``` But compodoc still includes files from dist/ (the outDir) and other directories in the root directory, which leads to all my modules, components, etc. being included two times in the docs. The expected behavior is that compodoc only includes the files specified in "files" or "include" if either of those are specified. That's the way the typescript compiler handles it. ##### **Operating System, Node.js, npm, compodoc version(s)** Ubuntu 17.04, Node v8.5.0, npm v5.3.0, compodoc v1.0.1 ##### **Angular configuration, a `package.json` file in the root folder** Here is my project: https://github.com/tradity/tradity-client/ ##### **Compodoc installed globally or locally ?** compodoc is installed locally
priority
includes more files than it should please follow the issue template below for bug reports and queries for issue start the label of the title with for feature requests start the label of the title with and explain your use case and ideas clearly below you can remove sections which are not relevant overview of the issue this is my tsconfig compileroptions target module commonjs moduleresolution node sourcemap true emitdecoratormetadata true experimentaldecorators true lib rootdir outdir dist include app ts exclude app main prod ts but compodoc still includes files from dist the outdir and other directories in the root directory which leads to all my modules components etc being included two times in the docs the expected behavior is that compodoc only includes the files specified in files or include if either of those are specified that s the way the typescript compiler handles it operating system node js npm compodoc version s ubuntu node npm compodoc angular configuration a package json file in the root folder here is my project compodoc installed globally or locally compodoc is installed locally
1
4,047
2,544,727,167
IssuesEvent
2015-01-29 12:25:13
pychess/pychess
https://api.github.com/repos/pychess/pychess
closed
Chessrules
Component-Docs imported Milestone-Release1.0 Priority-Medium Type-Task Usability
_From [lobais](https://code.google.com/u/lobais/) on September 06, 2006 16:42:39_ We should have a help item. Other than simple uihelp, it should also contain the rules and a small toturial. http://en.wikipedia.org/wiki/Chess has a nice table of the most simple answer to "what is chess" _Original issue: http://code.google.com/p/pychess/issues/detail?id=37_
1.0
Chessrules - _From [lobais](https://code.google.com/u/lobais/) on September 06, 2006 16:42:39_ We should have a help item. Other than simple uihelp, it should also contain the rules and a small toturial. http://en.wikipedia.org/wiki/Chess has a nice table of the most simple answer to "what is chess" _Original issue: http://code.google.com/p/pychess/issues/detail?id=37_
priority
chessrules from on september we should have a help item other than simple uihelp it should also contain the rules and a small toturial has a nice table of the most simple answer to what is chess original issue
1
236,174
7,747,213,742
IssuesEvent
2018-05-30 01:57:13
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
closed
Disabled transitions still run
Priority: 2 - Medium Status: 1 - Triaged Type: Bug
**Steps to reproduce**: - disable the `default_response` transition - submit an unstructured SMS **What should happen**: - the `default_responses` transition should not run on the doc **What actually happens**: - the `default_responses` transition does run on the doc See example [here](https://standard-release.dev.medicmobile.org/medic/410ca5bf6c30c2cfa17f4e8e0ad8ab5c): `"transitions":{"default_responses":{"last_rev":2,"seq":674080,"ok":true}}` ``` 2018-05-29T20:29:17.063Z - info: info: Reloading configuration 2018-05-29T20:29:17.852Z - info: info: Loading transition "maintain_info_document" 2018-05-29T20:29:17.854Z - info: info: Loading transition "update_clinics" 2018-05-29T20:29:17.855Z - info: info: Loading transition "registration" 2018-05-29T20:29:17.861Z - info: info: Loading transition "accept_patient_reports" 2018-05-29T20:29:17.861Z - info: info: Loading transition "generate_patient_id_on_people" 2018-05-29T20:29:17.862Z - info: warn: Disabled transition "default_responses" 2018-05-29T20:29:17.862Z - info: warn: Disabled transition "update_sent_by" 2018-05-29T20:29:17.863Z - info: warn: Disabled transition "update_sent_forms" 2018-05-29T20:29:17.863Z - info: warn: Disabled transition "conditional_alerts" 2018-05-29T20:29:17.863Z - info: warn: Disabled transition "multi_report_alerts" 2018-05-29T20:29:17.864Z - info: info: Loading transition "update_notifications" 2018-05-29T20:29:17.864Z - info: warn: Disabled transition "update_scheduled_reports" 2018-05-29T20:29:17.865Z - info: warn: Disabled transition "resolve_pending" 2018-05-29T20:29:45.394Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad88dad seq 674064 2018-05-29T20:29:49.822Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89af9 seq 674068 2018-05-29T20:29:53.152Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89d49 seq 674072 2018-05-29T20:30:21.240Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8aace seq 674076 2018-05-29T20:32:00.006Z - info: info: checking schedule again in 5 minutes 2018-05-29T20:32:08.386Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8ab5c seq 674080 ``` This may help towards #4539. Seen on `standard-release.dev` running `2.14.3-beta.2`
1.0
Disabled transitions still run - **Steps to reproduce**: - disable the `default_response` transition - submit an unstructured SMS **What should happen**: - the `default_responses` transition should not run on the doc **What actually happens**: - the `default_responses` transition does run on the doc See example [here](https://standard-release.dev.medicmobile.org/medic/410ca5bf6c30c2cfa17f4e8e0ad8ab5c): `"transitions":{"default_responses":{"last_rev":2,"seq":674080,"ok":true}}` ``` 2018-05-29T20:29:17.063Z - info: info: Reloading configuration 2018-05-29T20:29:17.852Z - info: info: Loading transition "maintain_info_document" 2018-05-29T20:29:17.854Z - info: info: Loading transition "update_clinics" 2018-05-29T20:29:17.855Z - info: info: Loading transition "registration" 2018-05-29T20:29:17.861Z - info: info: Loading transition "accept_patient_reports" 2018-05-29T20:29:17.861Z - info: info: Loading transition "generate_patient_id_on_people" 2018-05-29T20:29:17.862Z - info: warn: Disabled transition "default_responses" 2018-05-29T20:29:17.862Z - info: warn: Disabled transition "update_sent_by" 2018-05-29T20:29:17.863Z - info: warn: Disabled transition "update_sent_forms" 2018-05-29T20:29:17.863Z - info: warn: Disabled transition "conditional_alerts" 2018-05-29T20:29:17.863Z - info: warn: Disabled transition "multi_report_alerts" 2018-05-29T20:29:17.864Z - info: info: Loading transition "update_notifications" 2018-05-29T20:29:17.864Z - info: warn: Disabled transition "update_scheduled_reports" 2018-05-29T20:29:17.865Z - info: warn: Disabled transition "resolve_pending" 2018-05-29T20:29:45.394Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad88dad seq 674064 2018-05-29T20:29:49.822Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89af9 seq 674068 2018-05-29T20:29:53.152Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89d49 seq 674072 2018-05-29T20:30:21.240Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8aace seq 674076 2018-05-29T20:32:00.006Z - info: info: checking schedule again in 5 minutes 2018-05-29T20:32:08.386Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8ab5c seq 674080 ``` This may help towards #4539. Seen on `standard-release.dev` running `2.14.3-beta.2`
priority
disabled transitions still run steps to reproduce disable the default response transition submit an unstructured sms what should happen the default responses transition should not run on the doc what actually happens the default responses transition does run on the doc see example transitions default responses last rev seq ok true info info reloading configuration info info loading transition maintain info document info info loading transition update clinics info info loading transition registration info info loading transition accept patient reports info info loading transition generate patient id on people info warn disabled transition default responses info warn disabled transition update sent by info warn disabled transition update sent forms info warn disabled transition conditional alerts info warn disabled transition multi report alerts info info loading transition update notifications info warn disabled transition update scheduled reports info warn disabled transition resolve pending info info saved changes on doc seq info info saved changes on doc seq info info saved changes on doc seq info info saved changes on doc seq info info checking schedule again in minutes info info saved changes on doc seq this may help towards seen on standard release dev running beta
1
528,223
15,362,148,046
IssuesEvent
2021-03-01 19:05:08
thehyve/delphyne
https://api.github.com/repos/thehyve/delphyne
opened
Cleanup old source_vocabulary_ids in STCM_version
STCM medium priority
If I change the vocabulary ids in `stcm_versions.tsv` (say the custom vocabularies I was referencing have been deleted and I want to use different ones), the old vocabularies are not automatically removed from the `SourceToConceptMapVersion` table in the database. Once the constraints are re-applied, I get a FK constraints error because the deleted vocabularies are still being referenced. Solution: implement stcm version cleanup.
1.0
Cleanup old source_vocabulary_ids in STCM_version - If I change the vocabulary ids in `stcm_versions.tsv` (say the custom vocabularies I was referencing have been deleted and I want to use different ones), the old vocabularies are not automatically removed from the `SourceToConceptMapVersion` table in the database. Once the constraints are re-applied, I get a FK constraints error because the deleted vocabularies are still being referenced. Solution: implement stcm version cleanup.
priority
cleanup old source vocabulary ids in stcm version if i change the vocabulary ids in stcm versions tsv say the custom vocabularies i was referencing have been deleted and i want to use different ones the old vocabularies are not automatically removed from the sourcetoconceptmapversion table in the database once the constraints are re applied i get a fk constraints error because the deleted vocabularies are still being referenced solution implement stcm version cleanup
1
643,190
20,925,544,918
IssuesEvent
2022-03-24 22:19:50
cds-snc/resources-ressources
https://api.github.com/repos/cds-snc/resources-ressources
closed
Prototype Version 1
Medium Priority | Priorité moyenne
# Summary [PR for prototype](https://github.com/cds-snc/resources-ressources/pull/6 ) First pass at a coded prototype, based on figma designs. **Done when**: The prototype can be visited through a link
1.0
Prototype Version 1 - # Summary [PR for prototype](https://github.com/cds-snc/resources-ressources/pull/6 ) First pass at a coded prototype, based on figma designs. **Done when**: The prototype can be visited through a link
priority
prototype version summary first pass at a coded prototype based on figma designs done when the prototype can be visited through a link
1
683,326
23,377,282,952
IssuesEvent
2022-08-11 05:29:55
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
closed
Cannot update versions in multi-line `replace` block in go.mod
type:bug priority-3-medium manager:gomod status:ready reproduction:confirmed
### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 31.81.3, also 32.26.2 ### Please select which platform you are using if self-hosting. github.com ### If you're self-hosting Renovate, tell us what version of the platform you run. ubuntu-latest ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug renovate does not update versions in a multi-line `replace` block in a go.mod file. It only works for the single line replace. I created an example repo [here](https://github.com/avorima/renovate-gomod-example). Not updated: https://github.com/avorima/renovate-gomod-example/runs/5413565230?check_suite_focus=true ``` replace ( k8s.io/client-go => k8s.io/client-go v0.21.9 sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7 ) ``` Updated: https://github.com/avorima/renovate-gomod-example/runs/5413625323?check_suite_focus=true ``` replace k8s.io/client-go => k8s.io/client-go v0.21.9 replace sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7 ``` ### Relevant debug logs _No response_ ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
1.0
Cannot update versions in multi-line `replace` block in go.mod - ### How are you running Renovate? Self-hosted ### If you're self-hosting Renovate, tell us what version of Renovate you run. 31.81.3, also 32.26.2 ### Please select which platform you are using if self-hosting. github.com ### If you're self-hosting Renovate, tell us what version of the platform you run. ubuntu-latest ### Was this something which used to work for you, and then stopped? I never saw this working ### Describe the bug renovate does not update versions in a multi-line `replace` block in a go.mod file. It only works for the single line replace. I created an example repo [here](https://github.com/avorima/renovate-gomod-example). Not updated: https://github.com/avorima/renovate-gomod-example/runs/5413565230?check_suite_focus=true ``` replace ( k8s.io/client-go => k8s.io/client-go v0.21.9 sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7 ) ``` Updated: https://github.com/avorima/renovate-gomod-example/runs/5413625323?check_suite_focus=true ``` replace k8s.io/client-go => k8s.io/client-go v0.21.9 replace sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7 ``` ### Relevant debug logs _No response_ ### Have you created a minimal reproduction repository? I have linked to a minimal reproduction repository in the bug description
priority
cannot update versions in multi line replace block in go mod how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run also please select which platform you are using if self hosting github com if you re self hosting renovate tell us what version of the platform you run ubuntu latest was this something which used to work for you and then stopped i never saw this working describe the bug renovate does not update versions in a multi line replace block in a go mod file it only works for the single line replace i created an example repo not updated replace io client go io client go sigs io controller runtime sigs io controller runtime updated replace io client go io client go replace sigs io controller runtime sigs io controller runtime relevant debug logs no response have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description
1
333,421
10,121,981,188
IssuesEvent
2019-07-31 16:49:55
salesagility/SuiteCRM
https://api.github.com/repos/salesagility/SuiteCRM
closed
Cookie path is not respected if globally set
Fix Proposed Medium Priority Resolved: Next Release bug
<!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue When logging in, the ```ck_login_id_20``` and ```ck_login_language_20``` cookies are set using the ```setCookie``` wrapper in SugarApplication.php, if the path has been changed using ```session.cookie_path``` then it is ignored and defaults to '/' #### Expected Behavior It should respect any settings made in PHP config and set to the correct value #### Actual Behavior The cookie path is set to '/' because of the function signature #### Possible Fix Change the sig from: ```php public static function setCookie( $name, $value, $expire = 0, $path = '/', $domain = null, $secure = false, $httponly = true ) { ``` To ```php public static function setCookie( $name, $value, $expire = 0, $path = null, $domain = null, $secure = false, $httponly = true ) { ``` #### Steps to Reproduce 1. Install suitecrm inside a folder (/alpha) 2. Add the following to .htaccess ``` php_value session.cookie_path /alpha ``` 3. Login and check path for cookies #### Context I run multiple CRM's in differnet folders under a single domain (crm.domain/aplha, crm.domain/beta etc) and other cookies are set correctly, but these ones are not. #### Your Environment * SuiteCRM Version used: Version 7.11.1 / Sugar Version 6.5.25 (Build 344) * Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Firefox * Environment name and version (e.g. MySQL, PHP 7): PHP 7.2.14, MySQL 8 * Operating System and version (e.g Ubuntu 16.04): CentOS 7
1.0
Cookie path is not respected if globally set - <!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue When logging in, the ```ck_login_id_20``` and ```ck_login_language_20``` cookies are set using the ```setCookie``` wrapper in SugarApplication.php, if the path has been changed using ```session.cookie_path``` then it is ignored and defaults to '/' #### Expected Behavior It should respect any settings made in PHP config and set to the correct value #### Actual Behavior The cookie path is set to '/' because of the function signature #### Possible Fix Change the sig from: ```php public static function setCookie( $name, $value, $expire = 0, $path = '/', $domain = null, $secure = false, $httponly = true ) { ``` To ```php public static function setCookie( $name, $value, $expire = 0, $path = null, $domain = null, $secure = false, $httponly = true ) { ``` #### Steps to Reproduce 1. Install suitecrm inside a folder (/alpha) 2. Add the following to .htaccess ``` php_value session.cookie_path /alpha ``` 3. Login and check path for cookies #### Context I run multiple CRM's in differnet folders under a single domain (crm.domain/aplha, crm.domain/beta etc) and other cookies are set correctly, but these ones are not. #### Your Environment * SuiteCRM Version used: Version 7.11.1 / Sugar Version 6.5.25 (Build 344) * Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Firefox * Environment name and version (e.g. MySQL, PHP 7): PHP 7.2.14, MySQL 8 * Operating System and version (e.g Ubuntu 16.04): CentOS 7
priority
cookie path is not respected if globally set issue when logging in the ck login id and ck login language cookies are set using the setcookie wrapper in sugarapplication php if the path has been changed using session cookie path then it is ignored and defaults to expected behavior it should respect any settings made in php config and set to the correct value actual behavior the cookie path is set to because of the function signature possible fix change the sig from php public static function setcookie name value expire path domain null secure false httponly true to php public static function setcookie name value expire path null domain null secure false httponly true steps to reproduce install suitecrm inside a folder alpha add the following to htaccess php value session cookie path alpha login and check path for cookies context i run multiple crm s in differnet folders under a single domain crm domain aplha crm domain beta etc and other cookies are set correctly but these ones are not your environment suitecrm version used version sugar version build browser name and version e g chrome version bit firefox environment name and version e g mysql php php mysql operating system and version e g ubuntu centos
1
534,913
15,651,867,434
IssuesEvent
2021-03-23 10:41:55
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
opened
REST API - In POST method on the same endpoint is not working
bug priority: medium
**Describe the bug** We can perform a GET on /wp-json/buddyboss/v1/topics as well as many other endpoints that respond with a GET request. This validates that the authentication is working. But when we attempt to prefer a POST or Patch on the same endpoint it is NOT successful. **To Reproduce** When we attempt to prefer a POST or Patch on the same endpoint it is NOT successful. https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503. **Expected behavior** when I attempt to prefer a POST or Patch on the same endpoint it is need work. **Screenshots** ​https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503. **Support ticket links** https://secure.helpscout.net/conversation/1453624113/130633?folderId=4265312
1.0
REST API - In POST method on the same endpoint is not working - **Describe the bug** We can perform a GET on /wp-json/buddyboss/v1/topics as well as many other endpoints that respond with a GET request. This validates that the authentication is working. But when we attempt to prefer a POST or Patch on the same endpoint it is NOT successful. **To Reproduce** When we attempt to prefer a POST or Patch on the same endpoint it is NOT successful. https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503. **Expected behavior** when I attempt to prefer a POST or Patch on the same endpoint it is need work. **Screenshots** ​https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503. **Support ticket links** https://secure.helpscout.net/conversation/1453624113/130633?folderId=4265312
priority
rest api in post method on the same endpoint is not working describe the bug we can perform a get on wp json buddyboss topics as well as many other endpoints that respond with a get request this validates that the authentication is working but when we attempt to prefer a post or patch on the same endpoint it is not successful to reproduce when we attempt to prefer a post or patch on the same endpoint it is not successful expected behavior when i attempt to prefer a post or patch on the same endpoint it is need work screenshots ​ support ticket links
1
77,288
3,506,336,821
IssuesEvent
2016-01-08 05:52:12
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Arenas and battlegrounds not even working (BB #336)
Category: Instances migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** henhouse **Original Date:** 01.11.2010 12:27:31 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/336 <hr> You can join arenas but the doors will never open. Battlegrounds you can join but you cannot see anyone in them, the doors open there.
1.0
Arenas and battlegrounds not even working (BB #336) - This issue was migrated from bitbucket. **Original Reporter:** henhouse **Original Date:** 01.11.2010 12:27:31 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/336 <hr> You can join arenas but the doors will never open. Battlegrounds you can join but you cannot see anyone in them, the doors open there.
priority
arenas and battlegrounds not even working bb this issue was migrated from bitbucket original reporter henhouse original date gmt original priority major original type bug original state resolved direct link you can join arenas but the doors will never open battlegrounds you can join but you cannot see anyone in them the doors open there
1
498,792
14,433,225,147
IssuesEvent
2020-12-07 04:15:00
SD2E/experimental-intent-parser
https://api.github.com/repos/SD2E/experimental-intent-parser
closed
“Failed to access SynBioHub” when running “Suggest Additions by Spelling”
bug medium priority
When running “Add to spellchecker dictionary” from“Suggest Additions by Spelling”, IP will report “Failed to access SynBioHub”. What should happen here is each term that a users indicates “add to spellcheck dictionary,” is expected to appear in the dictionary. IP is then expected to store those english words so that it knows which term has already been added to the dictionary.
1.0
“Failed to access SynBioHub” when running “Suggest Additions by Spelling” - When running “Add to spellchecker dictionary” from“Suggest Additions by Spelling”, IP will report “Failed to access SynBioHub”. What should happen here is each term that a users indicates “add to spellcheck dictionary,” is expected to appear in the dictionary. IP is then expected to store those english words so that it knows which term has already been added to the dictionary.
priority
“failed to access synbiohub” when running “suggest additions by spelling” when running “add to spellchecker dictionary” from“suggest additions by spelling” ip will report “failed to access synbiohub” what should happen here is each term that a users indicates “add to spellcheck dictionary ” is expected to appear in the dictionary ip is then expected to store those english words so that it knows which term has already been added to the dictionary
1
472,013
13,614,694,339
IssuesEvent
2020-09-23 13:34:28
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio] Ability to Unpublish the Published Content
CI new feature priority: medium wontfix
As a content author, I should be able to unpublish the published content. At the moment if we want to do this, we have to delete the content item/items and republish again. The solution I like would be an option to unpublish a published content item/items So we can still keep the same item in authoring.
1.0
[studio] Ability to Unpublish the Published Content - As a content author, I should be able to unpublish the published content. At the moment if we want to do this, we have to delete the content item/items and republish again. The solution I like would be an option to unpublish a published content item/items So we can still keep the same item in authoring.
priority
ability to unpublish the published content as a content author i should be able to unpublish the published content at the moment if we want to do this we have to delete the content item items and republish again the solution i like would be an option to unpublish a published content item items so we can still keep the same item in authoring
1
741,216
25,784,460,021
IssuesEvent
2022-12-09 18:57:19
EESSI/eessi-bot-software-layer
https://api.github.com/repos/EESSI/eessi-bot-software-layer
opened
Improve tarball handling
enhancement difficulty:medium priority:medium
Currently the script `EESSI-pilot-install-software.sh` of the software-layer repository creates a tarball with the name `eessi-VERSION-COMPONENT-OS-ARCH-TIMESTAMP.tar.gz` and the script `eessi-upload-to-staging` of the eessi-bot-software-layer repository creates a metadata file for a tarball that is uploaded to an S3 bucket. The format of the filename is used a several places. Any change to this may break functionality at different places (in the bot here and elsewhere, e.g., the autoingestion script running on the Stratum 0). We should come up with a more robust handling of the information needed. See discussion at https://github.com/EESSI/eessi-bot-software-layer/pull/62#discussion_r1023955186
1.0
Improve tarball handling - Currently the script `EESSI-pilot-install-software.sh` of the software-layer repository creates a tarball with the name `eessi-VERSION-COMPONENT-OS-ARCH-TIMESTAMP.tar.gz` and the script `eessi-upload-to-staging` of the eessi-bot-software-layer repository creates a metadata file for a tarball that is uploaded to an S3 bucket. The format of the filename is used a several places. Any change to this may break functionality at different places (in the bot here and elsewhere, e.g., the autoingestion script running on the Stratum 0). We should come up with a more robust handling of the information needed. See discussion at https://github.com/EESSI/eessi-bot-software-layer/pull/62#discussion_r1023955186
priority
improve tarball handling currently the script eessi pilot install software sh of the software layer repository creates a tarball with the name eessi version component os arch timestamp tar gz and the script eessi upload to staging of the eessi bot software layer repository creates a metadata file for a tarball that is uploaded to an bucket the format of the filename is used a several places any change to this may break functionality at different places in the bot here and elsewhere e g the autoingestion script running on the stratum we should come up with a more robust handling of the information needed see discussion at
1
618,895
19,490,531,951
IssuesEvent
2021-12-27 05:07:54
itsmeow/betteranimalsplus
https://api.github.com/repos/itsmeow/betteranimalsplus
opened
Fabric crash when loading into a server with fresh client near reindeer
crash priority:medium side:client mc:1.16 platform:fabric mc:1.17 mc:1.18
## Describe what you were doing before the crash If a player connects to a Fabric dedicated server on a fresh client that has not joined a singleplayer world or opened the Mods menu, and loads in next to reindeer, it causes a crash, because the client configuration has not initialized. **Workaround**: Load into a singleplayer world, or open the Mods menu and click BAP's config (w/ Mod Menu and Cloth Config installed) ## To Reproduce Steps to reproduce the behavior (if possible): 1. Delete `betteranimalsplus-client.json5` 2. Start game 3. Click Multiplayer 4. Connect to a server w/ BAP while nearby reindeer 5. See crash ## Crash Log (REQUIRED) https://pastebin.com/T1c7Wk93 ## Versions - Minecraft Version: 1.16.5-1.18.1 - Architectury API Version: N/A - Platform (Forge/Fabric): Fabric - Forge Version: N/A - Fabric Loader Version: 0.12.12 - Fabric API Version: N/A - Mod Version: 11.0.0/11.0.1
1.0
Fabric crash when loading into a server with fresh client near reindeer - ## Describe what you were doing before the crash If a player connects to a Fabric dedicated server on a fresh client that has not joined a singleplayer world or opened the Mods menu, and loads in next to reindeer, it causes a crash, because the client configuration has not initialized. **Workaround**: Load into a singleplayer world, or open the Mods menu and click BAP's config (w/ Mod Menu and Cloth Config installed) ## To Reproduce Steps to reproduce the behavior (if possible): 1. Delete `betteranimalsplus-client.json5` 2. Start game 3. Click Multiplayer 4. Connect to a server w/ BAP while nearby reindeer 5. See crash ## Crash Log (REQUIRED) https://pastebin.com/T1c7Wk93 ## Versions - Minecraft Version: 1.16.5-1.18.1 - Architectury API Version: N/A - Platform (Forge/Fabric): Fabric - Forge Version: N/A - Fabric Loader Version: 0.12.12 - Fabric API Version: N/A - Mod Version: 11.0.0/11.0.1
priority
fabric crash when loading into a server with fresh client near reindeer describe what you were doing before the crash if a player connects to a fabric dedicated server on a fresh client that has not joined a singleplayer world or opened the mods menu and loads in next to reindeer it causes a crash because the client configuration has not initialized workaround load into a singleplayer world or open the mods menu and click bap s config w mod menu and cloth config installed to reproduce steps to reproduce the behavior if possible delete betteranimalsplus client start game click multiplayer connect to a server w bap while nearby reindeer see crash crash log required versions minecraft version architectury api version n a platform forge fabric fabric forge version n a fabric loader version fabric api version n a mod version
1
355,653
10,583,239,116
IssuesEvent
2019-10-08 13:20:57
AbsaOSS/enceladus
https://api.github.com/repos/AbsaOSS/enceladus
closed
Add more options to control resources to helper scripts
feature priority: medium run scripts
## Background Running Standardization and Conformance on big files revealed options that are essential to run these jobs efficiently. ## Feature Need to expend `spark-submit` resource request options. Add support for the ollowing: - `--executor-cores` - `--conf spark.executor.memoryOverhead=2048` - `--conf spark.memory.fraction=0.1`
1.0
Add more options to control resources to helper scripts - ## Background Running Standardization and Conformance on big files revealed options that are essential to run these jobs efficiently. ## Feature Need to expend `spark-submit` resource request options. Add support for the ollowing: - `--executor-cores` - `--conf spark.executor.memoryOverhead=2048` - `--conf spark.memory.fraction=0.1`
priority
add more options to control resources to helper scripts background running standardization and conformance on big files revealed options that are essential to run these jobs efficiently feature need to expend spark submit resource request options add support for the ollowing executor cores conf spark executor memoryoverhead conf spark memory fraction
1
57,215
3,081,249,057
IssuesEvent
2015-08-22 14:40:06
bitfighter/bitfighter
https://api.github.com/repos/bitfighter/bitfighter
closed
Variable size testitems
020 bug duplicate imported Priority-Medium
_From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on June 01, 2014 05:37:26_ Do we want to add variable-size testItems? If so... do it! _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=437_
1.0
Variable size testitems - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on June 01, 2014 05:37:26_ Do we want to add variable-size testItems? If so... do it! _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=437_
priority
variable size testitems from on june do we want to add variable size testitems if so do it original issue
1
59,366
3,105,870,595
IssuesEvent
2015-08-31 23:32:44
urbit/urbit
https://api.github.com/repos/urbit/urbit
opened
New FP crashes with SIGABRT on some platforms
bug difficulty medium platform specific priority medium
@ohAitch and @galenwp report: ``` ~zod:dojo> (add:rd .~2 .~2)address 0xfffffffffffffffd out of loom! bail: oops Abort trap: 6 ``` More portability stuff. Looks OSX-specific, so I can't debug it because I have no access to OSX.
1.0
New FP crashes with SIGABRT on some platforms - @ohAitch and @galenwp report: ``` ~zod:dojo> (add:rd .~2 .~2)address 0xfffffffffffffffd out of loom! bail: oops Abort trap: 6 ``` More portability stuff. Looks OSX-specific, so I can't debug it because I have no access to OSX.
priority
new fp crashes with sigabrt on some platforms ohaitch and galenwp report zod dojo add rd address out of loom bail oops abort trap more portability stuff looks osx specific so i can t debug it because i have no access to osx
1
797,808
28,180,807,085
IssuesEvent
2023-04-04 02:09:03
nimblehq/ic-flutter-taher-toby
https://api.github.com/repos/nimblehq/ic-flutter-taher-toby
closed
Set up CD for iOS (TestFlight)
type : chore priority : medium @0.1.0 @0.2.0
## Why In order to have convenient testing and debugging, a proper CD must be added. **Note**: This has already been added in `ios_deploy_to_testflight.yml` (staging), but the one for production is still missing. ## Acceptance Criteria - Create a new project on App Store Connect - Add necessary environment variables to secrets - Every time we merge to `develop`, an iOS build has to be deployed to TestFlight (Staging) - Every time we merge to `main`, an iOS build has to be deployed to TestFlight (Production) - Rename `ios_deploy_to_testflight.yml` to `ios_deploy_staging.yml` - Remove `ios_deploy_to_app_store.yml` ## Design N/A ## Resources https://developer.apple.com/app-store-connect/
1.0
Set up CD for iOS (TestFlight) - ## Why In order to have convenient testing and debugging, a proper CD must be added. **Note**: This has already been added in `ios_deploy_to_testflight.yml` (staging), but the one for production is still missing. ## Acceptance Criteria - Create a new project on App Store Connect - Add necessary environment variables to secrets - Every time we merge to `develop`, an iOS build has to be deployed to TestFlight (Staging) - Every time we merge to `main`, an iOS build has to be deployed to TestFlight (Production) - Rename `ios_deploy_to_testflight.yml` to `ios_deploy_staging.yml` - Remove `ios_deploy_to_app_store.yml` ## Design N/A ## Resources https://developer.apple.com/app-store-connect/
priority
set up cd for ios testflight why in order to have convenient testing and debugging a proper cd must be added note this has already been added in ios deploy to testflight yml staging but the one for production is still missing acceptance criteria create a new project on app store connect add necessary environment variables to secrets every time we merge to develop an ios build has to be deployed to testflight staging every time we merge to main an ios build has to be deployed to testflight production rename ios deploy to testflight yml to ios deploy staging yml remove ios deploy to app store yml design n a resources
1
108,323
4,331,105,486
IssuesEvent
2016-07-26 22:16:05
SpeedCurve-Metrics/SpeedCurve
https://api.github.com/repos/SpeedCurve-Metrics/SpeedCurve
closed
checkbox to preserve UA string
priority medium status new type enhancement
Similar to public WebPageTest, have a checkbox to preserve original UA string (ie, do NOT add "PTST"). Some services (such as ads) serve different content if PTST is present. Note that this same behavior can be achieved by using scripting, although that scripted URL could only be run on one browser (since the UA string in the script is specific to a single browser).
1.0
checkbox to preserve UA string - Similar to public WebPageTest, have a checkbox to preserve original UA string (ie, do NOT add "PTST"). Some services (such as ads) serve different content if PTST is present. Note that this same behavior can be achieved by using scripting, although that scripted URL could only be run on one browser (since the UA string in the script is specific to a single browser).
priority
checkbox to preserve ua string similar to public webpagetest have a checkbox to preserve original ua string ie do not add ptst some services such as ads serve different content if ptst is present note that this same behavior can be achieved by using scripting although that scripted url could only be run on one browser since the ua string in the script is specific to a single browser
1
642,418
20,887,420,754
IssuesEvent
2022-03-23 07:25:23
giancarlopernudisegura/cmput404
https://api.github.com/repos/giancarlopernudisegura/cmput404
closed
[User Story]: Remove Nodes
Priority Level: Medium Implementation Difficulty: Medium User Story
### User Story As a server admin, I want to be able to remove nodes and stop sharing with them. ### Objects Inbox
1.0
[User Story]: Remove Nodes - ### User Story As a server admin, I want to be able to remove nodes and stop sharing with them. ### Objects Inbox
priority
remove nodes user story as a server admin i want to be able to remove nodes and stop sharing with them objects inbox
1
25,926
2,684,046,392
IssuesEvent
2015-03-28 16:10:22
ConEmu/old-issues
https://api.github.com/repos/ConEmu/old-issues
closed
не передаются нажатия кнопок в GUI приложения во вкладках ConEmu
1 star bug imported Priority-Medium
_From [SAnTA...@gmail.com](https://code.google.com/u/105830449921090467193/) on February 09, 2012 01:34:28_ Required information! OS version: Win7 SP1 x64 ConEmu version: build 120205 (7z version) Far version: Far Manager, version 3.0 (build 2434) x64 PuTTy version: 2012-02-09 (x86-32) Давно пользуюсь ConEmu и все терзала мысль о том чт оне хватает в ней PuTTy, а тут вдруг выясняю что работы по внедрению ведутся полным ходом. Решил попробовать. Запускается волшебно вот только на нажатия стрелок и на F1-F12 клавиши PuTTy не реагирует. Комбинации Ctrl+N и Ctrl+P в качестве замены стрелкам вверх и вниз, соответственно, работают, но как то с ними не очень удобно. _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=490_
1.0
не передаются нажатия кнопок в GUI приложения во вкладках ConEmu - _From [SAnTA...@gmail.com](https://code.google.com/u/105830449921090467193/) on February 09, 2012 01:34:28_ Required information! OS version: Win7 SP1 x64 ConEmu version: build 120205 (7z version) Far version: Far Manager, version 3.0 (build 2434) x64 PuTTy version: 2012-02-09 (x86-32) Давно пользуюсь ConEmu и все терзала мысль о том чт оне хватает в ней PuTTy, а тут вдруг выясняю что работы по внедрению ведутся полным ходом. Решил попробовать. Запускается волшебно вот только на нажатия стрелок и на F1-F12 клавиши PuTTy не реагирует. Комбинации Ctrl+N и Ctrl+P в качестве замены стрелкам вверх и вниз, соответственно, работают, но как то с ними не очень удобно. _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=490_
priority
не передаются нажатия кнопок в gui приложения во вкладках conemu from on february required information os version conemu version build version far version far manager version build putty version давно пользуюсь conemu и все терзала мысль о том чт оне хватает в ней putty а тут вдруг выясняю что работы по внедрению ведутся полным ходом решил попробовать запускается волшебно вот только на нажатия стрелок и на клавиши putty не реагирует комбинации ctrl n и ctrl p в качестве замены стрелкам вверх и вниз соответственно работают но как то с ними не очень удобно original issue
1
378,567
11,204,620,145
IssuesEvent
2020-01-05 07:39:57
AugurProject/augur
https://api.github.com/repos/AugurProject/augur
opened
Tradingview.com templates need time zone details.
Priority: Medium V2 Audit
TradingView.com supports user configurable time zone. The market description for these templates needs to specify that the user should use UTC time zone when determining the open/close price.
1.0
Tradingview.com templates need time zone details. - TradingView.com supports user configurable time zone. The market description for these templates needs to specify that the user should use UTC time zone when determining the open/close price.
priority
tradingview com templates need time zone details tradingview com supports user configurable time zone the market description for these templates needs to specify that the user should use utc time zone when determining the open close price
1
434,514
12,519,519,902
IssuesEvent
2020-06-03 14:33:33
Twin-Cities-Mutual-Aid/twin-cities-aid-distribution-locations
https://api.github.com/repos/Twin-Cities-Mutual-Aid/twin-cities-aid-distribution-locations
opened
Security concerns
Priority: Medium Type: Discussion
Wanted to open an issue so we can have an ongoing discussion for how to address security related to this project. A first step is captured in #27. My main concern in raising this is the possibility that this map could be used to target donation centers for harassment, but that's also just a risk inherent to the project. I'm sure there are other concerns as well. This doesn't necessarily need to be limited to technical solutions but anything that will help keep our volunteers safe and prevent any malicious attacks on the site.
1.0
Security concerns - Wanted to open an issue so we can have an ongoing discussion for how to address security related to this project. A first step is captured in #27. My main concern in raising this is the possibility that this map could be used to target donation centers for harassment, but that's also just a risk inherent to the project. I'm sure there are other concerns as well. This doesn't necessarily need to be limited to technical solutions but anything that will help keep our volunteers safe and prevent any malicious attacks on the site.
priority
security concerns wanted to open an issue so we can have an ongoing discussion for how to address security related to this project a first step is captured in my main concern in raising this is the possibility that this map could be used to target donation centers for harassment but that s also just a risk inherent to the project i m sure there are other concerns as well this doesn t necessarily need to be limited to technical solutions but anything that will help keep our volunteers safe and prevent any malicious attacks on the site
1
886
2,504,525,266
IssuesEvent
2015-01-10 09:15:30
Araq/Nim
https://api.github.com/repos/Araq/Nim
opened
Add file handle procedures to async modules
Medium Priority Stdlib
Aside from file handles, the async modules should also be able to deal with file handles/descriptors, such as reading and writing.
1.0
Add file handle procedures to async modules - Aside from file handles, the async modules should also be able to deal with file handles/descriptors, such as reading and writing.
priority
add file handle procedures to async modules aside from file handles the async modules should also be able to deal with file handles descriptors such as reading and writing
1
478,379
13,778,655,432
IssuesEvent
2020-10-08 12:46:47
level73/membernet
https://api.github.com/repos/level73/membernet
opened
backend: long load time confusing users as when manage entity is clicked, it opens new form instead of list
Platform: Membernet Priority: Medium Type: Enhancement
i am not sure if you can do anything about this, but this is causing confusion for new users that are not aware that htey need to let the page load before selecting the specific management entity for viewing/editing the list- as otherwise it opens a new entry... can we do something about this? at least a line of text at the top?
1.0
backend: long load time confusing users as when manage entity is clicked, it opens new form instead of list - i am not sure if you can do anything about this, but this is causing confusion for new users that are not aware that htey need to let the page load before selecting the specific management entity for viewing/editing the list- as otherwise it opens a new entry... can we do something about this? at least a line of text at the top?
priority
backend long load time confusing users as when manage entity is clicked it opens new form instead of list i am not sure if you can do anything about this but this is causing confusion for new users that are not aware that htey need to let the page load before selecting the specific management entity for viewing editing the list as otherwise it opens a new entry can we do something about this at least a line of text at the top
1
706,998
24,290,975,493
IssuesEvent
2022-09-29 05:56:10
awslabs/aws-saas-boost
https://api.github.com/repos/awslabs/aws-saas-boost
closed
Updating AppConfig deletes SaaS Boost created HostedZone
bug priority-medium application-management tenant-onboarding
If you have a SaaS Boost environment that automatically created a HostedZone because it was configured with a DomainName, updating the AppConfig afterwards (even if just updating something like a service description) will result in the HostedZone getting deleted. ### Reproduction Steps <!-- minimal amount of code that causes the bug (if possible) or a reference: --> 1. Stand up a SaaS Boost environment 2. Configure it with a Domain Name and 1 service 3. See that a HostedZone is created in Route53 with the provided Domain Name as its name and the description `"${DomainName} Public DNS Zone"` 4. Update the service description 5. See that Settings service updates the config and fires an "Application Configuration Changed" event in the `sb-${env}-settings-update-config` CloudWatch LogGroup 6. See that the Onboarding service finds the existing HostedZone and updates the base stack in `sb-${env}-onboarding-events` 7. See that the HostedZone is eventually deleted. ### What did you expect to happen? I expect that the HostedZone would stick around. ### What actually happened? The HostedZone was deleted. ### Environment - **AWS Region :** us-west-2 - **AWS SaaS Boost Version :** `main` at https://github.com/awslabs/aws-saas-boost/commit/176187a405f9ce574544745ed8ec01e0db7172dd - **Workload OS (Linux or Windows) :** N/A ### Other <!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, slack, etc --> This is because the condition on which the HostedZone is created is based on whether a HostedZone is passed (the intention here was to only create a HostedZone if one does not exist): [source](https://github.com/awslabs/aws-saas-boost/blob/main/resources/saas-boost-core.yaml#L64-L67) ``` # in resources/saas-boost-core.yaml Conditions: CreateHostedZone: !And - !Not [!Equals [!Ref DomainName, '']] - !Equals [!Ref HostedZone, ''] ``` However, the Onboarding Service sets the `HostedZone` parameter when updating the stack to whatever exists in Route53, which will be the Boost-created HostedZone if none already existed: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L1771-L1801) ``` String hostedZone = getExistingHostedZone(domainName); // If there's an existing hosted zone, we need to tell the AppConfig about it // Otherwise, if there's a domain name, CloudFormation will create a hosted zone // and the stack listener will tell AppConfig about the newly created one. if (Utils.isNotBlank(hostedZone)) { LOGGER.info("Publishing appConfig update event for Route53 hosted zone {}", hostedZone); Utils.publishEvent(eventBridge, SAAS_BOOST_EVENT_BUS, EVENT_SOURCE, "Application Configuration Resource Changed", Map.of("hostedZone", hostedZone)); } ... Parameter.builder().parameterKey("HostedZone").parameterValue(hostedZone).build(), ``` Using the `getExistingHostedZone` function: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L2188-L2226). It's worth noting that this is also where we only reuse existing HostedZones if SaaS Boost already created them, an issue brought up by users in https://github.com/awslabs/aws-saas-boost/discussions/299 and https://github.com/awslabs/aws-saas-boost/issues/257. ``` protected String getExistingHostedZone(String domainName) { String existingHostedZone = ""; if (Utils.isNotEmpty(domainName)) { String nextDnsName = null; String nextHostedZone = null; ListHostedZonesByNameResponse response; do { response = route53.listHostedZonesByName(ListHostedZonesByNameRequest.builder() .dnsName(nextDnsName) .hostedZoneId(nextHostedZone) .maxItems("100") .build() ); nextDnsName = response.nextDNSName(); nextHostedZone = response.nextHostedZoneId(); if (response.hasHostedZones()) { for (HostedZone hostedZone : response.hostedZones()) { // If there are multiple hosted zones for a given domain name, what should we do? // We could sort the response by "CallerReference" which appears to be a timestamp. // In the documentation, we can just tell people if they're suffering from // https://github.com/awslabs/aws-saas-boost/issues/74 to go clean things up manually first? if (hostedZone.name().startsWith(domainName) && hostedZone.config() != null && Boolean.FALSE.equals(hostedZone.config().privateZone())) { // Created by SaaS Boost CloudFormation? // TODO do we do this check? seems safest for now. if ((domainName + " Public DNS zone").equals(hostedZone.config().comment())) { LOGGER.info("Found existing hosted zone {} for domain {}", hostedZone, domainName); // Hosted zone id will be prefixed with /hostedzone/ existingHostedZone = hostedZone.id().replace("/hostedzone/", ""); break; } } } } } while (response.isTruncated()); } return existingHostedZone; } ``` When the Onboarding Service updates the stack passing in the created HostedZone id, the condition evaluates to `false` and CloudFormation deletes the HostedZone. One possible way to fix this is to view the existing stack resources for the core stack before passing template parameters. Pseudo-code follows: ``` if appConfig.domainName: stackParameters.domainName = appConfig.domainName stackResources = cfn.getStackResources(sb-env-core) if "hostedZone" not in stackResources && route53.hostedZoneExists(name=appConfig.domainName): // there exists a hostedZone in the user's account that works for this domain name // but SaaS Boost didn't create it. so add it in as a parameter: meaning the CloudFormation // condition will evaluate to false and no hostedZone will be created (and none will be deleted, // since it didn't already exist owned by this template) stackParameters.hostedZone = route53.hostedZone(name=appConfig.domainName).id else: // either hostedZone was created by SaaS Boost or there is no hosted zone in the user account // that works for this domain name. in either case we want the "CreateHostedZone" condition // to evaluate to true so either: // - the SaaS Boost created hosted zone won't be deleted // or // - a hosted zone will be created by SaaS Boost, since none exist // we do this by providing no value for the hostedZone parameter stackParameters.hostedZone = "" // the same as having no else in production code ``` --- This is :bug: Bug Report
1.0
Updating AppConfig deletes SaaS Boost created HostedZone - If you have a SaaS Boost environment that automatically created a HostedZone because it was configured with a DomainName, updating the AppConfig afterwards (even if just updating something like a service description) will result in the HostedZone getting deleted. ### Reproduction Steps <!-- minimal amount of code that causes the bug (if possible) or a reference: --> 1. Stand up a SaaS Boost environment 2. Configure it with a Domain Name and 1 service 3. See that a HostedZone is created in Route53 with the provided Domain Name as its name and the description `"${DomainName} Public DNS Zone"` 4. Update the service description 5. See that Settings service updates the config and fires an "Application Configuration Changed" event in the `sb-${env}-settings-update-config` CloudWatch LogGroup 6. See that the Onboarding service finds the existing HostedZone and updates the base stack in `sb-${env}-onboarding-events` 7. See that the HostedZone is eventually deleted. ### What did you expect to happen? I expect that the HostedZone would stick around. ### What actually happened? The HostedZone was deleted. ### Environment - **AWS Region :** us-west-2 - **AWS SaaS Boost Version :** `main` at https://github.com/awslabs/aws-saas-boost/commit/176187a405f9ce574544745ed8ec01e0db7172dd - **Workload OS (Linux or Windows) :** N/A ### Other <!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, slack, etc --> This is because the condition on which the HostedZone is created is based on whether a HostedZone is passed (the intention here was to only create a HostedZone if one does not exist): [source](https://github.com/awslabs/aws-saas-boost/blob/main/resources/saas-boost-core.yaml#L64-L67) ``` # in resources/saas-boost-core.yaml Conditions: CreateHostedZone: !And - !Not [!Equals [!Ref DomainName, '']] - !Equals [!Ref HostedZone, ''] ``` However, the Onboarding Service sets the `HostedZone` parameter when updating the stack to whatever exists in Route53, which will be the Boost-created HostedZone if none already existed: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L1771-L1801) ``` String hostedZone = getExistingHostedZone(domainName); // If there's an existing hosted zone, we need to tell the AppConfig about it // Otherwise, if there's a domain name, CloudFormation will create a hosted zone // and the stack listener will tell AppConfig about the newly created one. if (Utils.isNotBlank(hostedZone)) { LOGGER.info("Publishing appConfig update event for Route53 hosted zone {}", hostedZone); Utils.publishEvent(eventBridge, SAAS_BOOST_EVENT_BUS, EVENT_SOURCE, "Application Configuration Resource Changed", Map.of("hostedZone", hostedZone)); } ... Parameter.builder().parameterKey("HostedZone").parameterValue(hostedZone).build(), ``` Using the `getExistingHostedZone` function: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L2188-L2226). It's worth noting that this is also where we only reuse existing HostedZones if SaaS Boost already created them, an issue brought up by users in https://github.com/awslabs/aws-saas-boost/discussions/299 and https://github.com/awslabs/aws-saas-boost/issues/257. ``` protected String getExistingHostedZone(String domainName) { String existingHostedZone = ""; if (Utils.isNotEmpty(domainName)) { String nextDnsName = null; String nextHostedZone = null; ListHostedZonesByNameResponse response; do { response = route53.listHostedZonesByName(ListHostedZonesByNameRequest.builder() .dnsName(nextDnsName) .hostedZoneId(nextHostedZone) .maxItems("100") .build() ); nextDnsName = response.nextDNSName(); nextHostedZone = response.nextHostedZoneId(); if (response.hasHostedZones()) { for (HostedZone hostedZone : response.hostedZones()) { // If there are multiple hosted zones for a given domain name, what should we do? // We could sort the response by "CallerReference" which appears to be a timestamp. // In the documentation, we can just tell people if they're suffering from // https://github.com/awslabs/aws-saas-boost/issues/74 to go clean things up manually first? if (hostedZone.name().startsWith(domainName) && hostedZone.config() != null && Boolean.FALSE.equals(hostedZone.config().privateZone())) { // Created by SaaS Boost CloudFormation? // TODO do we do this check? seems safest for now. if ((domainName + " Public DNS zone").equals(hostedZone.config().comment())) { LOGGER.info("Found existing hosted zone {} for domain {}", hostedZone, domainName); // Hosted zone id will be prefixed with /hostedzone/ existingHostedZone = hostedZone.id().replace("/hostedzone/", ""); break; } } } } } while (response.isTruncated()); } return existingHostedZone; } ``` When the Onboarding Service updates the stack passing in the created HostedZone id, the condition evaluates to `false` and CloudFormation deletes the HostedZone. One possible way to fix this is to view the existing stack resources for the core stack before passing template parameters. Pseudo-code follows: ``` if appConfig.domainName: stackParameters.domainName = appConfig.domainName stackResources = cfn.getStackResources(sb-env-core) if "hostedZone" not in stackResources && route53.hostedZoneExists(name=appConfig.domainName): // there exists a hostedZone in the user's account that works for this domain name // but SaaS Boost didn't create it. so add it in as a parameter: meaning the CloudFormation // condition will evaluate to false and no hostedZone will be created (and none will be deleted, // since it didn't already exist owned by this template) stackParameters.hostedZone = route53.hostedZone(name=appConfig.domainName).id else: // either hostedZone was created by SaaS Boost or there is no hosted zone in the user account // that works for this domain name. in either case we want the "CreateHostedZone" condition // to evaluate to true so either: // - the SaaS Boost created hosted zone won't be deleted // or // - a hosted zone will be created by SaaS Boost, since none exist // we do this by providing no value for the hostedZone parameter stackParameters.hostedZone = "" // the same as having no else in production code ``` --- This is :bug: Bug Report
priority
updating appconfig deletes saas boost created hostedzone if you have a saas boost environment that automatically created a hostedzone because it was configured with a domainname updating the appconfig afterwards even if just updating something like a service description will result in the hostedzone getting deleted reproduction steps minimal amount of code that causes the bug if possible or a reference stand up a saas boost environment configure it with a domain name and service see that a hostedzone is created in with the provided domain name as its name and the description domainname public dns zone update the service description see that settings service updates the config and fires an application configuration changed event in the sb env settings update config cloudwatch loggroup see that the onboarding service finds the existing hostedzone and updates the base stack in sb env onboarding events see that the hostedzone is eventually deleted what did you expect to happen i expect that the hostedzone would stick around what actually happened the hostedzone was deleted environment aws region us west aws saas boost version main at workload os linux or windows n a other this is because the condition on which the hostedzone is created is based on whether a hostedzone is passed the intention here was to only create a hostedzone if one does not exist in resources saas boost core yaml conditions createhostedzone and not equals however the onboarding service sets the hostedzone parameter when updating the stack to whatever exists in which will be the boost created hostedzone if none already existed string hostedzone getexistinghostedzone domainname if there s an existing hosted zone we need to tell the appconfig about it otherwise if there s a domain name cloudformation will create a hosted zone and the stack listener will tell appconfig about the newly created one if utils isnotblank hostedzone logger info publishing appconfig update event for hosted zone hostedzone utils publishevent eventbridge saas boost event bus event source application configuration resource changed map of hostedzone hostedzone parameter builder parameterkey hostedzone parametervalue hostedzone build using the getexistinghostedzone function it s worth noting that this is also where we only reuse existing hostedzones if saas boost already created them an issue brought up by users in and protected string getexistinghostedzone string domainname string existinghostedzone if utils isnotempty domainname string nextdnsname null string nexthostedzone null listhostedzonesbynameresponse response do response listhostedzonesbyname listhostedzonesbynamerequest builder dnsname nextdnsname hostedzoneid nexthostedzone maxitems build nextdnsname response nextdnsname nexthostedzone response nexthostedzoneid if response hashostedzones for hostedzone hostedzone response hostedzones if there are multiple hosted zones for a given domain name what should we do we could sort the response by callerreference which appears to be a timestamp in the documentation we can just tell people if they re suffering from to go clean things up manually first if hostedzone name startswith domainname hostedzone config null boolean false equals hostedzone config privatezone created by saas boost cloudformation todo do we do this check seems safest for now if domainname public dns zone equals hostedzone config comment logger info found existing hosted zone for domain hostedzone domainname hosted zone id will be prefixed with hostedzone existinghostedzone hostedzone id replace hostedzone break while response istruncated return existinghostedzone when the onboarding service updates the stack passing in the created hostedzone id the condition evaluates to false and cloudformation deletes the hostedzone one possible way to fix this is to view the existing stack resources for the core stack before passing template parameters pseudo code follows if appconfig domainname stackparameters domainname appconfig domainname stackresources cfn getstackresources sb env core if hostedzone not in stackresources hostedzoneexists name appconfig domainname there exists a hostedzone in the user s account that works for this domain name but saas boost didn t create it so add it in as a parameter meaning the cloudformation condition will evaluate to false and no hostedzone will be created and none will be deleted since it didn t already exist owned by this template stackparameters hostedzone hostedzone name appconfig domainname id else either hostedzone was created by saas boost or there is no hosted zone in the user account that works for this domain name in either case we want the createhostedzone condition to evaluate to true so either the saas boost created hosted zone won t be deleted or a hosted zone will be created by saas boost since none exist we do this by providing no value for the hostedzone parameter stackparameters hostedzone the same as having no else in production code this is bug bug report
1
485,662
13,996,699,275
IssuesEvent
2020-10-28 06:30:16
AY2021S1-CS2113-T14-3/tp
https://api.github.com/repos/AY2021S1-CS2113-T14-3/tp
closed
Automatically create file directories and folders for modules
priority.Medium type.Story type.Task
As a student, I want to be able to automatically create file directories and folders on my computer (i.e. 2021S1 → CS2113 and this folder contains ‘Tutorial’ and ‘Lecture Notes’ folders) for the modules that I am taking this semester, so that I do not have to manually create them.
1.0
Automatically create file directories and folders for modules - As a student, I want to be able to automatically create file directories and folders on my computer (i.e. 2021S1 → CS2113 and this folder contains ‘Tutorial’ and ‘Lecture Notes’ folders) for the modules that I am taking this semester, so that I do not have to manually create them.
priority
automatically create file directories and folders for modules as a student i want to be able to automatically create file directories and folders on my computer i e → and this folder contains ‘tutorial’ and ‘lecture notes’ folders for the modules that i am taking this semester so that i do not have to manually create them
1
26,316
2,684,309,894
IssuesEvent
2015-03-28 21:16:31
ConEmu/old-issues
https://api.github.com/repos/ConEmu/old-issues
opened
ConEmuHk screws up output of long lines
1 star bug imported Priority-Medium
_From [bahamutzero8825](https://code.google.com/u/bahamutzero8825/) on March 04, 2013 22:04:57_ OS version: Windows 8 x64 ConEmu version: 130304 x64 Far version (if you are using Far Manager): not using Far I've been developing a program in Python where multiple threads and children process can write to the same terminal at the same time. This has generally not been an issue, but I have a couple logging calls that output long lines and ConEmuHk seems to mess up the Python interpreter (CPython x64 3.3.0 if that matters) and I get garbled output (separate lines get interleaved). If ConEmuHk is not injected or I am viewing a remote session over SSH, the output is fine. I noticed this in 130223, and it seems to be better in 130304 (the former mangled several lines, the latter only 2-3). *Steps to reproduction* 1. Enable ConEmuHk injection 2. Write more than one thing to the terminal at once. 3. Get garbled output. If you cannot reproduce this on your own, I will write a short Python script to reproduce it. _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=973_
1.0
ConEmuHk screws up output of long lines - _From [bahamutzero8825](https://code.google.com/u/bahamutzero8825/) on March 04, 2013 22:04:57_ OS version: Windows 8 x64 ConEmu version: 130304 x64 Far version (if you are using Far Manager): not using Far I've been developing a program in Python where multiple threads and children process can write to the same terminal at the same time. This has generally not been an issue, but I have a couple logging calls that output long lines and ConEmuHk seems to mess up the Python interpreter (CPython x64 3.3.0 if that matters) and I get garbled output (separate lines get interleaved). If ConEmuHk is not injected or I am viewing a remote session over SSH, the output is fine. I noticed this in 130223, and it seems to be better in 130304 (the former mangled several lines, the latter only 2-3). *Steps to reproduction* 1. Enable ConEmuHk injection 2. Write more than one thing to the terminal at once. 3. Get garbled output. If you cannot reproduce this on your own, I will write a short Python script to reproduce it. _Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=973_
priority
conemuhk screws up output of long lines from on march os version windows conemu version far version if you are using far manager not using far i ve been developing a program in python where multiple threads and children process can write to the same terminal at the same time this has generally not been an issue but i have a couple logging calls that output long lines and conemuhk seems to mess up the python interpreter cpython if that matters and i get garbled output separate lines get interleaved if conemuhk is not injected or i am viewing a remote session over ssh the output is fine i noticed this in and it seems to be better in the former mangled several lines the latter only steps to reproduction enable conemuhk injection write more than one thing to the terminal at once get garbled output if you cannot reproduce this on your own i will write a short python script to reproduce it original issue
1
309,294
9,466,471,626
IssuesEvent
2019-04-18 04:43:17
wso2/product-is
https://api.github.com/repos/wso2/product-is
reopened
When reCaptcha is enabled, multi option login steps are not shown for given number of failed attempts.
Affected/5.8.0-Alpha2 Complexity/Medium Component/Adaptive Auth Priority/High Severity/Critical Type/Bug
- Add a Service Provider (OIDC/SSO samples) - Enable reCaptcha (after 1 failed login) - Enable multi-option login (eg: basic auth as step 1 and Email OTP as step 2) - Add a new claim to store the failed attempts before login. - Use script based adaptive authentication and added the following code. ``` // This variable is used to define the number of invalid attempts allowed before prompting the second factor var invalidAttemptsToStepup = 2; var failedLoginAttemptsBeforeSuccessClaim= 'http://wso2.org/claims/identity/failedLoginAttemptsBeforeSuccess'; function onLoginRequest(context) { doLogin(context); } function doLogin(context) { executeStep(1, { onSuccess : function(context){ var user = context.steps[1].subject; if (isExceedInvalidAttempts(user)) { executeStep(2); } }, onFail : function(context) { // Retry the login.. doLogin(context); } }); } function isExceedInvalidAttempts(user) { if (user.localClaims[failedLoginAttemptsBeforeSuccessClaim] >= invalidAttemptsToStepup) { return true; } else { return false; } } ``` When reCaptcha is set to come up after 1 failed login attempt and Email OTP to come up after 2 failed login attempts, to get E-mail OTP (which is set as the 2nd step) it takes 5 or more invalid login attempts.
1.0
When reCaptcha is enabled, multi option login steps are not shown for given number of failed attempts. - - Add a Service Provider (OIDC/SSO samples) - Enable reCaptcha (after 1 failed login) - Enable multi-option login (eg: basic auth as step 1 and Email OTP as step 2) - Add a new claim to store the failed attempts before login. - Use script based adaptive authentication and added the following code. ``` // This variable is used to define the number of invalid attempts allowed before prompting the second factor var invalidAttemptsToStepup = 2; var failedLoginAttemptsBeforeSuccessClaim= 'http://wso2.org/claims/identity/failedLoginAttemptsBeforeSuccess'; function onLoginRequest(context) { doLogin(context); } function doLogin(context) { executeStep(1, { onSuccess : function(context){ var user = context.steps[1].subject; if (isExceedInvalidAttempts(user)) { executeStep(2); } }, onFail : function(context) { // Retry the login.. doLogin(context); } }); } function isExceedInvalidAttempts(user) { if (user.localClaims[failedLoginAttemptsBeforeSuccessClaim] >= invalidAttemptsToStepup) { return true; } else { return false; } } ``` When reCaptcha is set to come up after 1 failed login attempt and Email OTP to come up after 2 failed login attempts, to get E-mail OTP (which is set as the 2nd step) it takes 5 or more invalid login attempts.
priority
when recaptcha is enabled multi option login steps are not shown for given number of failed attempts add a service provider oidc sso samples enable recaptcha after failed login enable multi option login eg basic auth as step and email otp as step add a new claim to store the failed attempts before login use script based adaptive authentication and added the following code this variable is used to define the number of invalid attempts allowed before prompting the second factor var invalidattemptstostepup var failedloginattemptsbeforesuccessclaim function onloginrequest context dologin context function dologin context executestep onsuccess function context var user context steps subject if isexceedinvalidattempts user executestep onfail function context retry the login dologin context function isexceedinvalidattempts user if user localclaims invalidattemptstostepup return true else return false when recaptcha is set to come up after failed login attempt and email otp to come up after failed login attempts to get e mail otp which is set as the step it takes or more invalid login attempts
1
112,240
4,513,777,752
IssuesEvent
2016-09-04 13:48:30
thommoboy/There-are-no-brakes
https://api.github.com/repos/thommoboy/There-are-no-brakes
closed
Industry level: players and box pass through the wall if elevator shove them
bug Industrialists Priority Medium
if player stand someplace that elevator will come, elevator will shove and let player pass through the wall(even out of map)
1.0
Industry level: players and box pass through the wall if elevator shove them - if player stand someplace that elevator will come, elevator will shove and let player pass through the wall(even out of map)
priority
industry level players and box pass through the wall if elevator shove them if player stand someplace that elevator will come elevator will shove and let player pass through the wall even out of map
1
644,376
20,976,063,121
IssuesEvent
2022-03-28 15:18:48
msavastio/ZohoCreatorWidgetReact
https://api.github.com/repos/msavastio/ZohoCreatorWidgetReact
opened
Rework project audit UI
medium priority business priority
- [ ] Remove float and total w/ float from "As Engineered" - [ ] Add in Proposal Amount field when doing Opportunity => Closed Won - [ ] As Contracted now becomes as Engineered with Project Amount and Float - these are static values - [ ] As Engineered with Changes remains unchanged - [ ] Equipment budget will be original quotes +/- any change orders internal or external
2.0
Rework project audit UI - - [ ] Remove float and total w/ float from "As Engineered" - [ ] Add in Proposal Amount field when doing Opportunity => Closed Won - [ ] As Contracted now becomes as Engineered with Project Amount and Float - these are static values - [ ] As Engineered with Changes remains unchanged - [ ] Equipment budget will be original quotes +/- any change orders internal or external
priority
rework project audit ui remove float and total w float from as engineered add in proposal amount field when doing opportunity closed won as contracted now becomes as engineered with project amount and float these are static values as engineered with changes remains unchanged equipment budget will be original quotes any change orders internal or external
1
31,369
2,732,897,037
IssuesEvent
2015-04-17 10:04:07
tiku01/oryx-editor
https://api.github.com/repos/tiku01/oryx-editor
closed
I want to tag models as I save them
auto-migrated Priority-Medium Type-Enhancement
``` What steps will reproduce the problem? 1. Creating many models that belong to one project 2. when save the model 3. go to the repository everytime, refresh view, select model, save What is the expected output? I'd like to tag the models directly in the savescreen dialog What do you see instead? Otherwise it takes much longer and I need 4 steps to do it. Please provide any additional information below. ``` Original issue reported on code.google.com by `alexande...@googlemail.com` on 1 Dec 2008 at 11:50
1.0
I want to tag models as I save them - ``` What steps will reproduce the problem? 1. Creating many models that belong to one project 2. when save the model 3. go to the repository everytime, refresh view, select model, save What is the expected output? I'd like to tag the models directly in the savescreen dialog What do you see instead? Otherwise it takes much longer and I need 4 steps to do it. Please provide any additional information below. ``` Original issue reported on code.google.com by `alexande...@googlemail.com` on 1 Dec 2008 at 11:50
priority
i want to tag models as i save them what steps will reproduce the problem creating many models that belong to one project when save the model go to the repository everytime refresh view select model save what is the expected output i d like to tag the models directly in the savescreen dialog what do you see instead otherwise it takes much longer and i need steps to do it please provide any additional information below original issue reported on code google com by alexande googlemail com on dec at
1
487,357
14,045,141,402
IssuesEvent
2020-11-02 00:22:50
codidact/qpixel
https://api.github.com/repos/codidact/qpixel
opened
Markdown tries to apply block-level styling to comments, losing content
area: frontend complexity: unassessed priority: medium type: bug
https://meta.codidact.com/questions/278611 is a bug report about comments that begin with "1. " -- the number and period disappear. In comments, @luap42 says it's a more general problem with Markdown that looks like it's for block elements: > Furthermore, it appears that the issue isn't specifically with numbers followed by a dot but by everything that can cause Markdown to produce something (lists, headings, quotes, ...). The HTML is sanitized differently for comments, but Markdown doesn't know that and hence produces a list/header/comment. The sanitizer then removes the <blockquote> tag and hence it looks like the number (or dash or #) was removed. I think the right solution would be to make the Markdown generator know that we are in "inline Markdown" mode and hence not generate block-level stuff. — ‭luap42 is a ghost 👻 ‭ 14 days ago
1.0
Markdown tries to apply block-level styling to comments, losing content - https://meta.codidact.com/questions/278611 is a bug report about comments that begin with "1. " -- the number and period disappear. In comments, @luap42 says it's a more general problem with Markdown that looks like it's for block elements: > Furthermore, it appears that the issue isn't specifically with numbers followed by a dot but by everything that can cause Markdown to produce something (lists, headings, quotes, ...). The HTML is sanitized differently for comments, but Markdown doesn't know that and hence produces a list/header/comment. The sanitizer then removes the <blockquote> tag and hence it looks like the number (or dash or #) was removed. I think the right solution would be to make the Markdown generator know that we are in "inline Markdown" mode and hence not generate block-level stuff. — ‭luap42 is a ghost 👻 ‭ 14 days ago
priority
markdown tries to apply block level styling to comments losing content is a bug report about comments that begin with the number and period disappear in comments says it s a more general problem with markdown that looks like it s for block elements furthermore it appears that the issue isn t specifically with numbers followed by a dot but by everything that can cause markdown to produce something lists headings quotes the html is sanitized differently for comments but markdown doesn t know that and hence produces a list header comment the sanitizer then removes the tag and hence it looks like the number or dash or was removed i think the right solution would be to make the markdown generator know that we are in inline markdown mode and hence not generate block level stuff — ‭ is a ghost 👻 ‭ days ago
1
178,726
6,617,919,427
IssuesEvent
2017-09-21 05:19:02
za419/VersionControl
https://api.github.com/repos/za419/VersionControl
closed
Rename the project
enhancement help wanted Medium Priority
Proposed names: - Odin (by @kenellorando) - Seshat (the Egyptian goddess of wisdom, knowledge, and writing; a record keeper) - Hero[dotus] (the father of history)
1.0
Rename the project - Proposed names: - Odin (by @kenellorando) - Seshat (the Egyptian goddess of wisdom, knowledge, and writing; a record keeper) - Hero[dotus] (the father of history)
priority
rename the project proposed names odin by kenellorando seshat the egyptian goddess of wisdom knowledge and writing a record keeper hero the father of history
1
767,820
26,941,435,311
IssuesEvent
2023-02-08 02:46:16
Benjamin-Loison/YouTube-operational-API
https://api.github.com/repos/Benjamin-Loison/YouTube-operational-API
closed
Propose a Docker version?
enhancement help wanted discussion low priority medium
As found in [the forks of this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/network/members), could inspire myself of [this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/compare/main...ThomasMargnac:YouTube-operational-API:main) to add Docker support. **UPDATE:** this fork has been removed. Related to #132.
1.0
Propose a Docker version? - As found in [the forks of this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/network/members), could inspire myself of [this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/compare/main...ThomasMargnac:YouTube-operational-API:main) to add Docker support. **UPDATE:** this fork has been removed. Related to #132.
priority
propose a docker version as found in could inspire myself of to add docker support update this fork has been removed related to
1
182,261
6,668,425,077
IssuesEvent
2017-10-03 15:45:16
dmusican/Elegit
https://api.github.com/repos/dmusican/Elegit
opened
Elegit uses a lot of memory. Is this necessary?
performance priority medium
Elegit uses a lot of memory, which slows down itself and the computers it runs on. We should profile this to see where all that memory is going and if it can be reduced.
1.0
Elegit uses a lot of memory. Is this necessary? - Elegit uses a lot of memory, which slows down itself and the computers it runs on. We should profile this to see where all that memory is going and if it can be reduced.
priority
elegit uses a lot of memory is this necessary elegit uses a lot of memory which slows down itself and the computers it runs on we should profile this to see where all that memory is going and if it can be reduced
1
311,899
9,540,182,675
IssuesEvent
2019-04-30 18:51:02
ansible/awx
https://api.github.com/repos/ansible/awx
opened
Investigate SCIM support
component:api priority:medium state:needs_devel type:enhancement
##### ISSUE TYPE - Feature Idea ##### SUMMARY SCIM allows social-auth users (SAML, etc.) to automatically be provisioned *and deprovisioned* when necessary, allowing for more secure usage of social auth. django lib here: https://github.com/15five/django-scim2
1.0
Investigate SCIM support - ##### ISSUE TYPE - Feature Idea ##### SUMMARY SCIM allows social-auth users (SAML, etc.) to automatically be provisioned *and deprovisioned* when necessary, allowing for more secure usage of social auth. django lib here: https://github.com/15five/django-scim2
priority
investigate scim support issue type feature idea summary scim allows social auth users saml etc to automatically be provisioned and deprovisioned when necessary allowing for more secure usage of social auth django lib here
1
726,677
25,007,475,778
IssuesEvent
2022-11-03 13:00:35
bounswe/bounswe2022group2
https://api.github.com/repos/bounswe/bounswe2022group2
opened
Milestone 1: Review the Milestone
priority-medium status-new milestone
### Issue Description With the first customer presentation coming to an end, we have developed, deployed, and released the 0.1.0-alpha of our application, Learnify. As part of the **Milestone Group Review** part of the Milestone 1, we have planned to provide the **Review the Milestone** part as a collaborative work of three people: @xltvy, @egemenatikk, @surmelienes1 This issue is the main issue of the Review the Milestone part. Each responsible member will open their own sub-issues regarding the parts, they will work on and then link their sub-issues under this issue. The Milestone 1 can be reached from [here](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md). ### Step Details Steps that will be performed: - [ ] Every responsible member will create their sub-issues. - [ ] Every responsible will complete their parts. - [ ] All material will be added to the [report](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md). ### Final Actions After every sub-issue is resolved and our material is added to the milestone report, this issue will be closed. ### Responsible People - [ ] Bahrican Yesil - [X] Egemen Atik - [X] Altay Acar - [ ] Ezgi Aysel Batı - [X] Muhammed Enes Sürmeli - [ ] Onur Kömürcü - [ ] Mehmet Batuhan Çelik - [ ] Hasan Can Erol - [ ] Ecenur Sezer - [ ] Koray Tekin - [ ] Mehmet Gökay Yıldız ### Deadline of the Issue 03.11.2022 - Thursday - 23.59 ### Reviewer Egemen Atik ### Deadline for the Review 04.11.2022 - Friday - 16.00 ### Final To-Dos - [ ] Every responsible shared an info about the sections/parts s/he will perform. - [ ] Every responsible mentioned the sub-issue contains the details of his/her work, if a sub-issue is created. - [ ] Every responsible mentioned this issue in the description of his/her sub-issue, if a sub-issue is created.
1.0
Milestone 1: Review the Milestone - ### Issue Description With the first customer presentation coming to an end, we have developed, deployed, and released the 0.1.0-alpha of our application, Learnify. As part of the **Milestone Group Review** part of the Milestone 1, we have planned to provide the **Review the Milestone** part as a collaborative work of three people: @xltvy, @egemenatikk, @surmelienes1 This issue is the main issue of the Review the Milestone part. Each responsible member will open their own sub-issues regarding the parts, they will work on and then link their sub-issues under this issue. The Milestone 1 can be reached from [here](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md). ### Step Details Steps that will be performed: - [ ] Every responsible member will create their sub-issues. - [ ] Every responsible will complete their parts. - [ ] All material will be added to the [report](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md). ### Final Actions After every sub-issue is resolved and our material is added to the milestone report, this issue will be closed. ### Responsible People - [ ] Bahrican Yesil - [X] Egemen Atik - [X] Altay Acar - [ ] Ezgi Aysel Batı - [X] Muhammed Enes Sürmeli - [ ] Onur Kömürcü - [ ] Mehmet Batuhan Çelik - [ ] Hasan Can Erol - [ ] Ecenur Sezer - [ ] Koray Tekin - [ ] Mehmet Gökay Yıldız ### Deadline of the Issue 03.11.2022 - Thursday - 23.59 ### Reviewer Egemen Atik ### Deadline for the Review 04.11.2022 - Friday - 16.00 ### Final To-Dos - [ ] Every responsible shared an info about the sections/parts s/he will perform. - [ ] Every responsible mentioned the sub-issue contains the details of his/her work, if a sub-issue is created. - [ ] Every responsible mentioned this issue in the description of his/her sub-issue, if a sub-issue is created.
priority
milestone review the milestone issue description with the first customer presentation coming to an end we have developed deployed and released the alpha of our application learnify as part of the milestone group review part of the milestone we have planned to provide the review the milestone part as a collaborative work of three people xltvy egemenatikk this issue is the main issue of the review the milestone part each responsible member will open their own sub issues regarding the parts they will work on and then link their sub issues under this issue the milestone can be reached from step details steps that will be performed every responsible member will create their sub issues every responsible will complete their parts all material will be added to the final actions after every sub issue is resolved and our material is added to the milestone report this issue will be closed responsible people bahrican yesil egemen atik altay acar ezgi aysel batı muhammed enes sürmeli onur kömürcü mehmet batuhan çelik hasan can erol ecenur sezer koray tekin mehmet gökay yıldız deadline of the issue thursday reviewer egemen atik deadline for the review friday final to dos every responsible shared an info about the sections parts s he will perform every responsible mentioned the sub issue contains the details of his her work if a sub issue is created every responsible mentioned this issue in the description of his her sub issue if a sub issue is created
1
739,465
25,598,310,902
IssuesEvent
2022-12-01 17:54:12
zowe/imperative
https://api.github.com/repos/zowe/imperative
closed
Unexpected prompting error: 'prepared.args[option.name][0].toUpperCase is not a function'
bug priority-medium
Prompting code causes an error when a parameter is numeric by trying to use the string 'toUpperCase function. Even if the parameter is entered as a string yargs converts it to numeric.
1.0
Unexpected prompting error: 'prepared.args[option.name][0].toUpperCase is not a function' - Prompting code causes an error when a parameter is numeric by trying to use the string 'toUpperCase function. Even if the parameter is entered as a string yargs converts it to numeric.
priority
unexpected prompting error prepared args touppercase is not a function prompting code causes an error when a parameter is numeric by trying to use the string touppercase function even if the parameter is entered as a string yargs converts it to numeric
1
731,220
25,206,488,716
IssuesEvent
2022-11-13 18:48:34
Cheos137/ArmorpointsPlusplus
https://api.github.com/repos/Cheos137/ArmorpointsPlusplus
closed
[Feature Request, INTERNAL]: defer config caching to forge
feature request minecraft: 1.16 future update priority: medium minecraft: 1.18 minecraft: 1.19 forge
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description remove caching config implementations on forge versions that support caching by themselves ### Solution defer caching logic to forge to clean up code ### Alternatives _No response_ ### Additional Context _No response_
1.0
[Feature Request, INTERNAL]: defer config caching to forge - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description remove caching config implementations on forge versions that support caching by themselves ### Solution defer caching logic to forge to clean up code ### Alternatives _No response_ ### Additional Context _No response_
priority
defer config caching to forge is there an existing issue for this i have searched the existing issues description remove caching config implementations on forge versions that support caching by themselves solution defer caching logic to forge to clean up code alternatives no response additional context no response
1
100,404
4,087,156,659
IssuesEvent
2016-06-01 09:01:12
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
Passing method to macro causes seg fault
Medium Priority Semcheck
works when you only define the method: ```nimrod import macros type A = ref object macro testMacro(a: expr): stmt = result = newNimNode(nnkStmtList) let methDef = a.findChild(it.kind == nnkMethodDef) result.add methDef method test(x: A) {.testMacro.} = echo "hello" test(A()) ``` doesn't work if you don't define method: ```nimrod import macros type A = ref object macro testMacro(a: expr): stmt = discard method test(x: A) {.testMacro.} = echo "hello" ``` doesn't work if you add multiple statements to returned statement list: ```nimrod import macros type A = ref object macro testMacro(a: expr): stmt = result = newNimNode(nnkStmtList) let methDef = a.findChild(it.kind == nnkMethodDef) let methIdent = methDef.findChild(it.kind == nnkIdent) result.add methDef let procTyp = newNimNode(nnkProcTy).add(methDef.params).add(newEmptyNode()) let thisMethSym = genSym(nskLet, "thisMethod") let identDefs = newIdentDefs(thisMethSym, procTyp, methIdent) let letSection = newNimNode(nnkLetSection).add(identDefs) result.add letSection method test(x: A) {.testMacro.} = echo "hello" ```
1.0
Passing method to macro causes seg fault - works when you only define the method: ```nimrod import macros type A = ref object macro testMacro(a: expr): stmt = result = newNimNode(nnkStmtList) let methDef = a.findChild(it.kind == nnkMethodDef) result.add methDef method test(x: A) {.testMacro.} = echo "hello" test(A()) ``` doesn't work if you don't define method: ```nimrod import macros type A = ref object macro testMacro(a: expr): stmt = discard method test(x: A) {.testMacro.} = echo "hello" ``` doesn't work if you add multiple statements to returned statement list: ```nimrod import macros type A = ref object macro testMacro(a: expr): stmt = result = newNimNode(nnkStmtList) let methDef = a.findChild(it.kind == nnkMethodDef) let methIdent = methDef.findChild(it.kind == nnkIdent) result.add methDef let procTyp = newNimNode(nnkProcTy).add(methDef.params).add(newEmptyNode()) let thisMethSym = genSym(nskLet, "thisMethod") let identDefs = newIdentDefs(thisMethSym, procTyp, methIdent) let letSection = newNimNode(nnkLetSection).add(identDefs) result.add letSection method test(x: A) {.testMacro.} = echo "hello" ```
priority
passing method to macro causes seg fault works when you only define the method nimrod import macros type a ref object macro testmacro a expr stmt result newnimnode nnkstmtlist let methdef a findchild it kind nnkmethoddef result add methdef method test x a testmacro echo hello test a doesn t work if you don t define method nimrod import macros type a ref object macro testmacro a expr stmt discard method test x a testmacro echo hello doesn t work if you add multiple statements to returned statement list nimrod import macros type a ref object macro testmacro a expr stmt result newnimnode nnkstmtlist let methdef a findchild it kind nnkmethoddef let methident methdef findchild it kind nnkident result add methdef let proctyp newnimnode nnkprocty add methdef params add newemptynode let thismethsym gensym nsklet thismethod let identdefs newidentdefs thismethsym proctyp methident let letsection newnimnode nnkletsection add identdefs result add letsection method test x a testmacro echo hello
1
247,512
7,919,426,332
IssuesEvent
2018-07-04 16:46:56
Xceptance/neodymium-library
https://api.github.com/repos/Xceptance/neodymium-library
closed
Enhance NeodymiumCucumberRunner with @Browser annotation
Medium Priority doneInDevelop feature recheckedInDevelop
It would be nice to be able to add a @Browser annotation. This way it would reduce the duplication of test data tables in feature files.
1.0
Enhance NeodymiumCucumberRunner with @Browser annotation - It would be nice to be able to add a @Browser annotation. This way it would reduce the duplication of test data tables in feature files.
priority
enhance neodymiumcucumberrunner with browser annotation it would be nice to be able to add a browser annotation this way it would reduce the duplication of test data tables in feature files
1
770,570
27,045,629,794
IssuesEvent
2023-02-13 09:33:57
robotframework/robotframework
https://api.github.com/repos/robotframework/robotframework
closed
New `robot:flatten` tag for "flattening" keyword structures
enhancement priority: high effort: medium
Introduction ------------ With nested keyword structures, especially with recursive keyword calls and with WHILE and FOR loops, the log file can get hard do understand with many different nesting levels. Such nested structures also increase output.xml size, because even a simple keyword like ```robotframework *** Keywords *** Keyword Log Robot Log Framework ``` creates this much content: ```xml <kw name="Keyword"> <kw name="Log" library="BuiltIn"> <arg>Robot</arg> <doc>Logs the given message with the given level.</doc> <msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.663"/> </kw> <kw name="Log" library="BuiltIn"> <arg>Framework</arg> <doc>Logs the given message with the given level.</doc> <msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/> </kw> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/> </kw> ``` We have had `--flattenkeywords` option for "flattening" such structures since RF 2.8.2 (#1551) and it works great. When a keyword is flattened, its child keywords and control structures are removed otherwise, but all their messages are preserved. It doesn't affect output.xml generated during execution, but flattening happens when output.xml files are parsed and can save huge amounts of memory. When `--flattenkeywords` is used with Rebot, it is possible to create a new flattened output.xml. For example, the above structure is converted into this if `Keyowrd` is flattened: ``` <kw name="Keyword"> <doc>_*Content flattened.*_</doc> <msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg> <msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/> </kw> ``` Proposal -------- Flattening works based on keyword names and based on tags, but it needs to be activated separately from the command line. This issue proposes adding new built-in tag `robot:flatten` that activates this behavior automatically. Removing top level keywords from tests and leaving only their messages doesn't make sense, so `robot:flatten` should be usable only as a keyword tag. This functionality should work already during execution so that flattened keywords and control structures are never written to output.xml file. This avoid output.xml file growing big and is likely to also enhance the performance a bit. Open questions --------------- There are some open questions related to the design still: - [ ] Should `start/end_keyword` listener methods be called with flattened keywords? I believe not, but I don't feel too strongly about this. - [ ] Should we add *Content flattened* to keyword documentation like we do with `--flattenkeywords`? I believe not. There's the `robot:flatten` tag to indicate that anyway. - [ ] Should `--flattenkeywords` be changed to work during execution as well? I believe yes, but that requires a separate issue. - [ ] Should automatic TRACE level logging of arguments and return values of flattened keywords be disabled? I believe yes, but this isn't high priority. Possible future enhancements ------------------------------ `--flattenkeywords` allows flattening WHILE or FOR loops or all loop iterations. Something like that would be convenient with built-in tags as well. We could consider something like `robot:flatten:while` and `robot:flatten:iteration` to support that, but I believe that's something that can wait until future versions. Another alternative would be allowing tags with control structures as shown in the example below. This would require parser and model changes but could also have other use cases. That's certainly out of the scope of RF 6.1, though. ```robotframework *** Keywords *** Keyword WHILE True [Tags] robot:flatten Nested END ```
1.0
New `robot:flatten` tag for "flattening" keyword structures - Introduction ------------ With nested keyword structures, especially with recursive keyword calls and with WHILE and FOR loops, the log file can get hard do understand with many different nesting levels. Such nested structures also increase output.xml size, because even a simple keyword like ```robotframework *** Keywords *** Keyword Log Robot Log Framework ``` creates this much content: ```xml <kw name="Keyword"> <kw name="Log" library="BuiltIn"> <arg>Robot</arg> <doc>Logs the given message with the given level.</doc> <msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.663"/> </kw> <kw name="Log" library="BuiltIn"> <arg>Framework</arg> <doc>Logs the given message with the given level.</doc> <msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/> </kw> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/> </kw> ``` We have had `--flattenkeywords` option for "flattening" such structures since RF 2.8.2 (#1551) and it works great. When a keyword is flattened, its child keywords and control structures are removed otherwise, but all their messages are preserved. It doesn't affect output.xml generated during execution, but flattening happens when output.xml files are parsed and can save huge amounts of memory. When `--flattenkeywords` is used with Rebot, it is possible to create a new flattened output.xml. For example, the above structure is converted into this if `Keyowrd` is flattened: ``` <kw name="Keyword"> <doc>_*Content flattened.*_</doc> <msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg> <msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg> <status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/> </kw> ``` Proposal -------- Flattening works based on keyword names and based on tags, but it needs to be activated separately from the command line. This issue proposes adding new built-in tag `robot:flatten` that activates this behavior automatically. Removing top level keywords from tests and leaving only their messages doesn't make sense, so `robot:flatten` should be usable only as a keyword tag. This functionality should work already during execution so that flattened keywords and control structures are never written to output.xml file. This avoid output.xml file growing big and is likely to also enhance the performance a bit. Open questions --------------- There are some open questions related to the design still: - [ ] Should `start/end_keyword` listener methods be called with flattened keywords? I believe not, but I don't feel too strongly about this. - [ ] Should we add *Content flattened* to keyword documentation like we do with `--flattenkeywords`? I believe not. There's the `robot:flatten` tag to indicate that anyway. - [ ] Should `--flattenkeywords` be changed to work during execution as well? I believe yes, but that requires a separate issue. - [ ] Should automatic TRACE level logging of arguments and return values of flattened keywords be disabled? I believe yes, but this isn't high priority. Possible future enhancements ------------------------------ `--flattenkeywords` allows flattening WHILE or FOR loops or all loop iterations. Something like that would be convenient with built-in tags as well. We could consider something like `robot:flatten:while` and `robot:flatten:iteration` to support that, but I believe that's something that can wait until future versions. Another alternative would be allowing tags with control structures as shown in the example below. This would require parser and model changes but could also have other use cases. That's certainly out of the scope of RF 6.1, though. ```robotframework *** Keywords *** Keyword WHILE True [Tags] robot:flatten Nested END ```
priority
new robot flatten tag for flattening keyword structures introduction with nested keyword structures especially with recursive keyword calls and with while and for loops the log file can get hard do understand with many different nesting levels such nested structures also increase output xml size because even a simple keyword like robotframework keywords keyword log robot log framework creates this much content xml robot logs the given message with the given level robot framework logs the given message with the given level framework we have had flattenkeywords option for flattening such structures since rf and it works great when a keyword is flattened its child keywords and control structures are removed otherwise but all their messages are preserved it doesn t affect output xml generated during execution but flattening happens when output xml files are parsed and can save huge amounts of memory when flattenkeywords is used with rebot it is possible to create a new flattened output xml for example the above structure is converted into this if keyowrd is flattened content flattened robot framework proposal flattening works based on keyword names and based on tags but it needs to be activated separately from the command line this issue proposes adding new built in tag robot flatten that activates this behavior automatically removing top level keywords from tests and leaving only their messages doesn t make sense so robot flatten should be usable only as a keyword tag this functionality should work already during execution so that flattened keywords and control structures are never written to output xml file this avoid output xml file growing big and is likely to also enhance the performance a bit open questions there are some open questions related to the design still should start end keyword listener methods be called with flattened keywords i believe not but i don t feel too strongly about this should we add content flattened to keyword documentation like we do with flattenkeywords i believe not there s the robot flatten tag to indicate that anyway should flattenkeywords be changed to work during execution as well i believe yes but that requires a separate issue should automatic trace level logging of arguments and return values of flattened keywords be disabled i believe yes but this isn t high priority possible future enhancements flattenkeywords allows flattening while or for loops or all loop iterations something like that would be convenient with built in tags as well we could consider something like robot flatten while and robot flatten iteration to support that but i believe that s something that can wait until future versions another alternative would be allowing tags with control structures as shown in the example below this would require parser and model changes but could also have other use cases that s certainly out of the scope of rf though robotframework keywords keyword while true robot flatten nested end
1
759,956
26,620,080,245
IssuesEvent
2023-01-24 10:32:46
EddieHubCommunity/LinkFree
https://api.github.com/repos/EddieHubCommunity/LinkFree
closed
[DOCS] Wrong links in the FAQ
🏁 status: ready for dev 🟨 priority: medium 🛠 goal: fix good first issue 📄 aspect: text
### Description There are various broken links on the FAQ page (such as the single-user mode or Hacktoberfest pages) which should be fixed immediately. _Steps to contribute_ 1. Fork the repository [here](https://github.com/EddieHubCommunity/LinkFree/fork). 2. Create a new branch named `fix_faq_links`. 3. Open the `faqs.mdx` file, which is located at `pages/docs/`. 4. In line 9, change the single-user mode link to `advanced/single-user-mode`. We do not want to include `docs` as otherwise, it would lead to `docs/docs/...` which is invalid. 5. In line 15, change the Hacktoberfest link to `contributing/hacktoberfest`. 6. The Discord invitation link seems to be broken as well. Change it to `http://discord.eddiehub.org` (located in line 21). 7. Commit and push your changes. 8. Create a [pull request](https://github.com/EddieHubCommunity/LinkFree/compare). 9. You're done now! 🎉 You just have to wait until the maintainers check your PR. If you need any help, let us know about it and we'll gladly help you. 🙂 ### Screenshots _No response_ ### Additional information This is a good first issue. Please leave it open to first-time contributors and beginners.
1.0
[DOCS] Wrong links in the FAQ - ### Description There are various broken links on the FAQ page (such as the single-user mode or Hacktoberfest pages) which should be fixed immediately. _Steps to contribute_ 1. Fork the repository [here](https://github.com/EddieHubCommunity/LinkFree/fork). 2. Create a new branch named `fix_faq_links`. 3. Open the `faqs.mdx` file, which is located at `pages/docs/`. 4. In line 9, change the single-user mode link to `advanced/single-user-mode`. We do not want to include `docs` as otherwise, it would lead to `docs/docs/...` which is invalid. 5. In line 15, change the Hacktoberfest link to `contributing/hacktoberfest`. 6. The Discord invitation link seems to be broken as well. Change it to `http://discord.eddiehub.org` (located in line 21). 7. Commit and push your changes. 8. Create a [pull request](https://github.com/EddieHubCommunity/LinkFree/compare). 9. You're done now! 🎉 You just have to wait until the maintainers check your PR. If you need any help, let us know about it and we'll gladly help you. 🙂 ### Screenshots _No response_ ### Additional information This is a good first issue. Please leave it open to first-time contributors and beginners.
priority
wrong links in the faq description there are various broken links on the faq page such as the single user mode or hacktoberfest pages which should be fixed immediately steps to contribute fork the repository create a new branch named fix faq links open the faqs mdx file which is located at pages docs in line change the single user mode link to advanced single user mode we do not want to include docs as otherwise it would lead to docs docs which is invalid in line change the hacktoberfest link to contributing hacktoberfest the discord invitation link seems to be broken as well change it to located in line commit and push your changes create a you re done now 🎉 you just have to wait until the maintainers check your pr if you need any help let us know about it and we ll gladly help you 🙂 screenshots no response additional information this is a good first issue please leave it open to first time contributors and beginners
1
533,377
15,589,876,073
IssuesEvent
2021-03-18 08:39:14
sopra-fs21-group-02/server
https://api.github.com/repos/sopra-fs21-group-02/server
opened
Implement filtering for the map view
area:map priority:medium task
Make the map view consider the currently configured filters and only display elements that are wanted according to the filters. ## Estimate 6h ## User Story This task belongs to user story #3
1.0
Implement filtering for the map view - Make the map view consider the currently configured filters and only display elements that are wanted according to the filters. ## Estimate 6h ## User Story This task belongs to user story #3
priority
implement filtering for the map view make the map view consider the currently configured filters and only display elements that are wanted according to the filters estimate user story this task belongs to user story
1
253,215
8,052,926,704
IssuesEvent
2018-08-01 20:56:20
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
opened
Update the on/off handling to mark an individual as muted
Needs Triage Priority: 2 - Medium Type: Feature
This request is a continuation of the request in ticket #4767. It is not an immediate need, but if it is not too much extra work beyond 4767, it probably makes sense to do at the same time. Full details and additional screenshots can be found in the design spec: https://docs.google.com/document/d/1we_Qu1B4x8PuAi9pxiIu4MeXtdH1kY9W29kMidHThsc/edit?usp=sharing **Summary of individual muting** - We would like to make sure that the family muting is done in a way that can be logically and technically extended to individuals, even if that isn't actually implemented now. - Individuals can already be muted from their profiles using the features built for the “OFF” form, but we need to make an extension of that to also mark the patient muted and make the UI changes to their profile to reflect the “Muted” status - Generally speaking, there may be muted individuals in an un-muted household. However, there may NOT be un-muted individuals in a muted household. - If a user wants to unmute an individual in a muted household, the whole household (including all family members) will need to be unmuted. The user may then individually mute any other household members as needed. - Muting schedules attached to a CHW as a person in the app (in the case that they are fired, quit, or retire) will be technically available but considering the programmatic implications of this (re-assigning, performance metrics) is outside the scope of this MVP. **UI changes to the main People page list: none** - If an individual is muted from an individual profile form, then the greater family remains unmuted, and there is no change to that family’s appearance on the main People list ![screen shot 2018-08-01 at 1 55 23 pm](https://user-images.githubusercontent.com/26235957/43548386-9c7bfc74-9592-11e8-8576-7aaf9cf9ce84.png) **UI changes to the family profile** - If an individual is muted from an individual on/off form, then the greater family remains unmuted, and there is no change to that family’s appearance (family icons stays pink, no “Muted” text) - There should be a change made to the “Family Members” condition card. Any individuals that are muted within an unmuted family should be specially styled in this card so that they stand out from active unmuted family members. - Suggestion: The person’s name becomes grey colored, and instead of (or in addition to) the age on the second line, we also display the word “Muted”. See example above where Jon Bass is muted. ![screen shot 2018-08-01 at 1 55 33 pm](https://user-images.githubusercontent.com/26235957/43548399-a5a27c24-9592-11e8-8101-ad679c7dc70f.png) **UI changes to profiles** - Very similar to profile styling for death - The person/place icon becomes grey (#A7A9AC) - And we display the text “Muted” on the second line below the person/place name - Any forms that were previously available on the profile remain available - Instead of the “Mute person/place” form, the user now sees “Unmute person/place”
1.0
Update the on/off handling to mark an individual as muted - This request is a continuation of the request in ticket #4767. It is not an immediate need, but if it is not too much extra work beyond 4767, it probably makes sense to do at the same time. Full details and additional screenshots can be found in the design spec: https://docs.google.com/document/d/1we_Qu1B4x8PuAi9pxiIu4MeXtdH1kY9W29kMidHThsc/edit?usp=sharing **Summary of individual muting** - We would like to make sure that the family muting is done in a way that can be logically and technically extended to individuals, even if that isn't actually implemented now. - Individuals can already be muted from their profiles using the features built for the “OFF” form, but we need to make an extension of that to also mark the patient muted and make the UI changes to their profile to reflect the “Muted” status - Generally speaking, there may be muted individuals in an un-muted household. However, there may NOT be un-muted individuals in a muted household. - If a user wants to unmute an individual in a muted household, the whole household (including all family members) will need to be unmuted. The user may then individually mute any other household members as needed. - Muting schedules attached to a CHW as a person in the app (in the case that they are fired, quit, or retire) will be technically available but considering the programmatic implications of this (re-assigning, performance metrics) is outside the scope of this MVP. **UI changes to the main People page list: none** - If an individual is muted from an individual profile form, then the greater family remains unmuted, and there is no change to that family’s appearance on the main People list ![screen shot 2018-08-01 at 1 55 23 pm](https://user-images.githubusercontent.com/26235957/43548386-9c7bfc74-9592-11e8-8576-7aaf9cf9ce84.png) **UI changes to the family profile** - If an individual is muted from an individual on/off form, then the greater family remains unmuted, and there is no change to that family’s appearance (family icons stays pink, no “Muted” text) - There should be a change made to the “Family Members” condition card. Any individuals that are muted within an unmuted family should be specially styled in this card so that they stand out from active unmuted family members. - Suggestion: The person’s name becomes grey colored, and instead of (or in addition to) the age on the second line, we also display the word “Muted”. See example above where Jon Bass is muted. ![screen shot 2018-08-01 at 1 55 33 pm](https://user-images.githubusercontent.com/26235957/43548399-a5a27c24-9592-11e8-8101-ad679c7dc70f.png) **UI changes to profiles** - Very similar to profile styling for death - The person/place icon becomes grey (#A7A9AC) - And we display the text “Muted” on the second line below the person/place name - Any forms that were previously available on the profile remain available - Instead of the “Mute person/place” form, the user now sees “Unmute person/place”
priority
update the on off handling to mark an individual as muted this request is a continuation of the request in ticket it is not an immediate need but if it is not too much extra work beyond it probably makes sense to do at the same time full details and additional screenshots can be found in the design spec summary of individual muting we would like to make sure that the family muting is done in a way that can be logically and technically extended to individuals even if that isn t actually implemented now individuals can already be muted from their profiles using the features built for the “off” form but we need to make an extension of that to also mark the patient muted and make the ui changes to their profile to reflect the “muted” status generally speaking there may be muted individuals in an un muted household however there may not be un muted individuals in a muted household if a user wants to unmute an individual in a muted household the whole household including all family members will need to be unmuted the user may then individually mute any other household members as needed muting schedules attached to a chw as a person in the app in the case that they are fired quit or retire will be technically available but considering the programmatic implications of this re assigning performance metrics is outside the scope of this mvp ui changes to the main people page list none if an individual is muted from an individual profile form then the greater family remains unmuted and there is no change to that family’s appearance on the main people list ui changes to the family profile if an individual is muted from an individual on off form then the greater family remains unmuted and there is no change to that family’s appearance family icons stays pink no “muted” text there should be a change made to the “family members” condition card any individuals that are muted within an unmuted family should be specially styled in this card so that they stand out from active unmuted family members suggestion the person’s name becomes grey colored and instead of or in addition to the age on the second line we also display the word “muted” see example above where jon bass is muted ui changes to profiles very similar to profile styling for death the person place icon becomes grey and we display the text “muted” on the second line below the person place name any forms that were previously available on the profile remain available instead of the “mute person place” form the user now sees “unmute person place”
1
16,749
2,615,122,965
IssuesEvent
2015-03-01 05:50:24
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
opened
Need an example for google shopping API....Please
auto-migrated Priority-Medium Type-Sample
``` Which Google API and version (e.g. Google Calendar Data API version 2)?1.5 What format (e.g. JSON, Atom)?Json What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?ClientLogin Java environment (e.g. Java 6, Android 2.3, App Engine)?android 2.3 External references, such as API reference guide?GoogleShopping API Please provide any additional information below.Need an example for searching an API using android ..Please help me out ``` Original issue reported on code.google.com by `saikiran...@gmail.com` on 13 Oct 2011 at 11:05
1.0
Need an example for google shopping API....Please - ``` Which Google API and version (e.g. Google Calendar Data API version 2)?1.5 What format (e.g. JSON, Atom)?Json What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?ClientLogin Java environment (e.g. Java 6, Android 2.3, App Engine)?android 2.3 External references, such as API reference guide?GoogleShopping API Please provide any additional information below.Need an example for searching an API using android ..Please help me out ``` Original issue reported on code.google.com by `saikiran...@gmail.com` on 13 Oct 2011 at 11:05
priority
need an example for google shopping api please which google api and version e g google calendar data api version what format e g json atom json what authentation e g oauth oauth clientlogin clientlogin java environment e g java android app engine android external references such as api reference guide googleshopping api please provide any additional information below need an example for searching an api using android please help me out original issue reported on code google com by saikiran gmail com on oct at
1
358,687
10,631,312,269
IssuesEvent
2019-10-15 08:05:57
facebookresearch/nevergrad
https://api.github.com/repos/facebookresearch/nevergrad
opened
Consistent population-based algorithms
Difficulty: High Priority: Medium Type: Enhancement
Population based algorithms are written in a variety of ways while some common code could be extracted. Also, for now they don't allow asking more than the population, which can be cumbersome (especially if some evaluation died without providing a value: in this case an individual of the population will always be ignored)
1.0
Consistent population-based algorithms - Population based algorithms are written in a variety of ways while some common code could be extracted. Also, for now they don't allow asking more than the population, which can be cumbersome (especially if some evaluation died without providing a value: in this case an individual of the population will always be ignored)
priority
consistent population based algorithms population based algorithms are written in a variety of ways while some common code could be extracted also for now they don t allow asking more than the population which can be cumbersome especially if some evaluation died without providing a value in this case an individual of the population will always be ignored
1