Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
461,166 | 13,224,536,776 | IssuesEvent | 2020-08-17 19:20:16 | grey-software/Twitter-Focus | https://api.github.com/repos/grey-software/Twitter-Focus | opened | Inprove the README | Domain: Dev Experience Priority: Medium Type: Maintenance | Consult https://github.com/grey-software/LinkedInFocus
- [ ] Add webstore link
- [ ] Add local dev instructions
- [ ] Add screenshots or gif
- [ ] Add logo | 1.0 | Inprove the README - Consult https://github.com/grey-software/LinkedInFocus
- [ ] Add webstore link
- [ ] Add local dev instructions
- [ ] Add screenshots or gif
- [ ] Add logo | priority | inprove the readme consult add webstore link add local dev instructions add screenshots or gif add logo | 1 |
146,794 | 5,627,950,302 | IssuesEvent | 2017-04-05 03:58:50 | dnGrep/dnGrep | https://api.github.com/repos/dnGrep/dnGrep | closed | no supports for gb2312/utf8 encoding? | bug imported Priority-Medium | _From [mfm...@sina.com](https://code.google.com/u/107484051280825188802/) on February 25, 2013 23:15:51_
it turns out to be no supports for greping gb2312/utf8 encoded text.
1. dngrep did find the occurs, but linenos and positions of the result are very wrong
2. chinese characters displayed in result pane are unreadable, while they are good in preview pane.
3. converting gb2312/utf8 encoded text file to unicode encoding will solve this problem, but it is a tough work and not feasible
3. env: dngrep version 2.7.1, win7 x64(simplified chinese)
_Original issue: http://code.google.com/p/dngrep/issues/detail?id=177_
| 1.0 | no supports for gb2312/utf8 encoding? - _From [mfm...@sina.com](https://code.google.com/u/107484051280825188802/) on February 25, 2013 23:15:51_
it turns out to be no supports for greping gb2312/utf8 encoded text.
1. dngrep did find the occurs, but linenos and positions of the result are very wrong
2. chinese characters displayed in result pane are unreadable, while they are good in preview pane.
3. converting gb2312/utf8 encoded text file to unicode encoding will solve this problem, but it is a tough work and not feasible
3. env: dngrep version 2.7.1, win7 x64(simplified chinese)
_Original issue: http://code.google.com/p/dngrep/issues/detail?id=177_
| priority | no supports for encoding from on february it turns out to be no supports for greping encoded text dngrep did find the occurs but linenos and positions of the result are very wrong chinese characters displayed in result pane are unreadable while they are good in preview pane converting encoded text file to unicode encoding will solve this problem but it is a tough work and not feasible env dngrep version simplified chinese original issue | 1 |
405,428 | 11,873,120,850 | IssuesEvent | 2020-03-26 16:49:01 | netdata/netdata | https://api.github.com/repos/netdata/netdata | closed | During the agent installation, if the ACLK fails to be built, show an error message to the user | ACLK internal priority/high priority/medium | #### Summary
Ensure that the requirement in product#282 are met.
- [x] If ACLK error reporting not set yet by #8051 define it with this issue
- [x] If ACLK build fails make it prominent to the user so he knows about it. It should not be just small line easily overlooked in log. @amoss will speak to @jacekkolasa about this separately.
- [x] Netdata should log on startup it is build without ACLK
- [x] Report failure to Cloud to same endpoint #8051
- [x] Respect DO_NOT_TRACK environment variable
(Some of this will be covered by PR 8025 but there are fresh requests at the bottom of the discussion).
| 2.0 | During the agent installation, if the ACLK fails to be built, show an error message to the user - #### Summary
Ensure that the requirement in product#282 are met.
- [x] If ACLK error reporting not set yet by #8051 define it with this issue
- [x] If ACLK build fails make it prominent to the user so he knows about it. It should not be just small line easily overlooked in log. @amoss will speak to @jacekkolasa about this separately.
- [x] Netdata should log on startup it is build without ACLK
- [x] Report failure to Cloud to same endpoint #8051
- [x] Respect DO_NOT_TRACK environment variable
(Some of this will be covered by PR 8025 but there are fresh requests at the bottom of the discussion).
| priority | during the agent installation if the aclk fails to be built show an error message to the user summary ensure that the requirement in product are met if aclk error reporting not set yet by define it with this issue if aclk build fails make it prominent to the user so he knows about it it should not be just small line easily overlooked in log amoss will speak to jacekkolasa about this separately netdata should log on startup it is build without aclk report failure to cloud to same endpoint respect do not track environment variable some of this will be covered by pr but there are fresh requests at the bottom of the discussion | 1 |
632,748 | 20,205,862,237 | IssuesEvent | 2022-02-11 20:16:28 | nens/threedi-api-qgis-client | https://api.github.com/repos/nens/threedi-api-qgis-client | closed | Download results if user has access to simulation results but not to schematisation | ⏰ Priority: 3. Medium | This currently results in the following error:

Desired behaviour: if schematisation is not found, let user choose where to store the results (Save as dialog) | 1.0 | Download results if user has access to simulation results but not to schematisation - This currently results in the following error:

Desired behaviour: if schematisation is not found, let user choose where to store the results (Save as dialog) | priority | download results if user has access to simulation results but not to schematisation this currently results in the following error desired behaviour if schematisation is not found let user choose where to store the results save as dialog | 1 |
69,655 | 3,309,741,825 | IssuesEvent | 2015-11-05 03:18:40 | cs2103aug2015-t14-2j/main | https://api.github.com/repos/cs2103aug2015-t14-2j/main | closed | As a user i want to have a GUI | priority.medium | so that I can visualize the different tasks on my screen and better conceptualize my schedule | 1.0 | As a user i want to have a GUI - so that I can visualize the different tasks on my screen and better conceptualize my schedule | priority | as a user i want to have a gui so that i can visualize the different tasks on my screen and better conceptualize my schedule | 1 |
479,370 | 13,795,364,105 | IssuesEvent | 2020-10-09 17:53:18 | medic/cht-core | https://api.github.com/repos/medic/cht-core | closed | Update `admin` to be a standalone app | Priority: 2 - Medium Type: Technical issue | **Describe the issue**
Currently, the admin app requires a significantly large number of webapp files (services, filters, directives, redux actions/reducers/selectors).
This becomes a problem with the migration to Angular 10, where the files imported from webapp will, most likely, be unusable by admin.
**Describe the improvement you'd like**
"Duplicate" all files that admin app requires from the webapp folder, preserving git line history as much as possible. (this includes their tests)
**Describe alternatives you've considered**
We could postpone this migration under the assumption that we may be able to use post-angular-10-migration webapp files as we do now, but it's a risk.
| 1.0 | Update `admin` to be a standalone app - **Describe the issue**
Currently, the admin app requires a significantly large number of webapp files (services, filters, directives, redux actions/reducers/selectors).
This becomes a problem with the migration to Angular 10, where the files imported from webapp will, most likely, be unusable by admin.
**Describe the improvement you'd like**
"Duplicate" all files that admin app requires from the webapp folder, preserving git line history as much as possible. (this includes their tests)
**Describe alternatives you've considered**
We could postpone this migration under the assumption that we may be able to use post-angular-10-migration webapp files as we do now, but it's a risk.
| priority | update admin to be a standalone app describe the issue currently the admin app requires a significantly large number of webapp files services filters directives redux actions reducers selectors this becomes a problem with the migration to angular where the files imported from webapp will most likely be unusable by admin describe the improvement you d like duplicate all files that admin app requires from the webapp folder preserving git line history as much as possible this includes their tests describe alternatives you ve considered we could postpone this migration under the assumption that we may be able to use post angular migration webapp files as we do now but it s a risk | 1 |
148,749 | 5,695,880,666 | IssuesEvent | 2017-04-16 04:23:56 | tootsuite/mastodon | https://api.github.com/repos/tootsuite/mastodon | closed | Videos in unsupported codecs cause UI problems | bug priority - medium ui | Viewing a toot containing a video in a format your browser doesn't support causes usability issues with the UI:
* there is a blank space where the thumbnail would be (see the first toot in the following screenshot)

* clicking on the space covers the screen in a lightbox, but it's empty other than a close button

These screenshots were taken of a toot containing a WebM video, on iOS, which doesn't support WebM video encoding in the browser (due to lack of hardware acceleration of that format).
I understand that it's not feasible to convert every video uploaded to Mastodon to compatible formats for every device, but I'm wondering if it's possible to detect the unsupported format and either hide the thumbnail preview and/or display a message or icon indicating the format isn't supported (in any case, the link should still be clickable to download the file because another app might support it even if the browser doesn't).
* * * *
- [X] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
| 1.0 | Videos in unsupported codecs cause UI problems - Viewing a toot containing a video in a format your browser doesn't support causes usability issues with the UI:
* there is a blank space where the thumbnail would be (see the first toot in the following screenshot)

* clicking on the space covers the screen in a lightbox, but it's empty other than a close button

These screenshots were taken of a toot containing a WebM video, on iOS, which doesn't support WebM video encoding in the browser (due to lack of hardware acceleration of that format).
I understand that it's not feasible to convert every video uploaded to Mastodon to compatible formats for every device, but I'm wondering if it's possible to detect the unsupported format and either hide the thumbnail preview and/or display a message or icon indicating the format isn't supported (in any case, the link should still be clickable to download the file because another app might support it even if the browser doesn't).
* * * *
- [X] I searched or browsed the repo’s other issues to ensure this is not a duplicate.
| priority | videos in unsupported codecs cause ui problems viewing a toot containing a video in a format your browser doesn t support causes usability issues with the ui there is a blank space where the thumbnail would be see the first toot in the following screenshot clicking on the space covers the screen in a lightbox but it s empty other than a close button these screenshots were taken of a toot containing a webm video on ios which doesn t support webm video encoding in the browser due to lack of hardware acceleration of that format i understand that it s not feasible to convert every video uploaded to mastodon to compatible formats for every device but i m wondering if it s possible to detect the unsupported format and either hide the thumbnail preview and or display a message or icon indicating the format isn t supported in any case the link should still be clickable to download the file because another app might support it even if the browser doesn t i searched or browsed the repo’s other issues to ensure this is not a duplicate | 1 |
232,593 | 7,667,799,349 | IssuesEvent | 2018-05-14 00:47:04 | hackoregon/civic-devops | https://api.github.com/repos/hackoregon/civic-devops | closed | Cannot `ssh` into Bastion Host in newly created stack | Priority: medium bug | This issue was previously mentioned at issue #87
@MikeTheCanuck suspects that the security group assigned to the Bastion Host only allows certain IP addresses incoming. The below screenshot showed that this host allows `tcp/22` from anywhere but both Mike and me got `Connect time out`

| 1.0 | Cannot `ssh` into Bastion Host in newly created stack - This issue was previously mentioned at issue #87
@MikeTheCanuck suspects that the security group assigned to the Bastion Host only allows certain IP addresses incoming. The below screenshot showed that this host allows `tcp/22` from anywhere but both Mike and me got `Connect time out`

| priority | cannot ssh into bastion host in newly created stack this issue was previously mentioned at issue mikethecanuck suspects that the security group assigned to the bastion host only allows certain ip addresses incoming the below screenshot showed that this host allows tcp from anywhere but both mike and me got connect time out | 1 |
413,916 | 12,093,352,708 | IssuesEvent | 2020-04-19 19:17:09 | samiha-rahman/soen390 | https://api.github.com/repos/samiha-rahman/soen390 | closed | US-64: Select classroom as start or end destination | epic 4 priority: medium user story | As a user, I would like to be able to select a classroom as a start or end destination.
This story is created from a requirement extracted from #31 to reduce the scope of that story.
**Associated epic:** #20
**Acceptance criteria:**
Step # | Execution Procedure or Input | Expected Results/Outputs | Passed/Failed
--- | --- | --- | ---
1 | Select a room | Should provide option to select designated room as starting point or end destination | | 1.0 | US-64: Select classroom as start or end destination - As a user, I would like to be able to select a classroom as a start or end destination.
This story is created from a requirement extracted from #31 to reduce the scope of that story.
**Associated epic:** #20
**Acceptance criteria:**
Step # | Execution Procedure or Input | Expected Results/Outputs | Passed/Failed
--- | --- | --- | ---
1 | Select a room | Should provide option to select designated room as starting point or end destination | | priority | us select classroom as start or end destination as a user i would like to be able to select a classroom as a start or end destination this story is created from a requirement extracted from to reduce the scope of that story associated epic acceptance criteria step execution procedure or input expected results outputs passed failed select a room should provide option to select designated room as starting point or end destination | 1 |
422,945 | 12,288,633,361 | IssuesEvent | 2020-05-09 17:37:10 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Username formatting breaks markdown for links in posts | help wanted priority: medium section: Guilds section: Tavern Chat | 
> "Something I already suspected but just confirmed: Putting a mention in a link markdown breaks the link the person is trying to add (I've had to use a different "@" glyph in the first code demonstration to avoid triggering it expanding the actual mention's code there...)
>
> Typing the following:
>
> [@citrusella really likes Habitica!](https://habitica.com)
>
> produces
>
> [@citrusella really likes Habitica!](https://habitica.com)
>
> which the website likely sees in the following way:
>
> [[@citrusella](/profile/2d6ef231-50b4-4a22-90e7-45eb97147a2c) really likes Habitica!](https://habitica.com)"
(Image included since obviously one can't demonstrate in the same way on github! In other words, including @username in markdown links breaks it because the @mention formatting takes precedence. | 1.0 | Username formatting breaks markdown for links in posts - 
> "Something I already suspected but just confirmed: Putting a mention in a link markdown breaks the link the person is trying to add (I've had to use a different "@" glyph in the first code demonstration to avoid triggering it expanding the actual mention's code there...)
>
> Typing the following:
>
> [@citrusella really likes Habitica!](https://habitica.com)
>
> produces
>
> [@citrusella really likes Habitica!](https://habitica.com)
>
> which the website likely sees in the following way:
>
> [[@citrusella](/profile/2d6ef231-50b4-4a22-90e7-45eb97147a2c) really likes Habitica!](https://habitica.com)"
(Image included since obviously one can't demonstrate in the same way on github! In other words, including @username in markdown links breaks it because the @mention formatting takes precedence. | priority | username formatting breaks markdown for links in posts something i already suspected but just confirmed putting a mention in a link markdown breaks the link the person is trying to add i ve had to use a different glyph in the first code demonstration to avoid triggering it expanding the actual mention s code there typing the following produces which the website likely sees in the following way profile really likes habitica image included since obviously one can t demonstrate in the same way on github in other words including username in markdown links breaks it because the mention formatting takes precedence | 1 |
776,099 | 27,246,883,097 | IssuesEvent | 2023-02-22 03:16:28 | ansible-collections/azure | https://api.github.com/repos/ansible-collections/azure | closed | azure_rm_virtualmachine fails in AzureChinaCloud | has_pr medium_priority | ##### SUMMARY
something called by azure_rm_virtualmachine does not obey AZURE_CLOUD_ENVIRONMENT and attempts to connect to AzureCloud when operating on AzureChinaCloud hosts
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_virtualmachine
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible-playbook [core 2.12.5.post0]
config file = /ansible/ansible.cfg
configured module search path = ['/home/runner/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.8/site-packages/ansible
ansible collection location = /home/runner/.ansible/collections:/ansible/collections
executable location = /usr/local/bin/ansible-playbook
python version = 3.8.13 (default, Jun 24 2022, 15:27:57) [GCC 8.5.0 20210514 (Red Hat 8.5.0-13)]
jinja version = 3.1.2
libyaml = True
```
##### COLLECTION VERSION
<!--- Paste verbatim output from "ansible-galaxy collection list <namespace>.<collection>" between the quotes
for example: ansible-galaxy collection list community.general
-->
```paste below
- name: azure.azcollection
version: '==1.14.0'
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
ANY_ERRORS_FATAL(/ansible/ansible.cfg) = False
COLLECTIONS_PATHS(/ansible/ansible.cfg) = ['/home/runner/.ansible/collections', '/ansible/collections']
DEFAULT_FORKS(/ansible/ansible.cfg) = 50
DEFAULT_GATHER_SUBSET(/ansible/ansible.cfg) = ['all']
DEFAULT_GATHER_TIMEOUT(/ansible/ansible.cfg) = 20
DEFAULT_HASH_BEHAVIOUR(/ansible/ansible.cfg) = merge
DEFAULT_TIMEOUT(/ansible/ansible.cfg) = 20
DEFAULT_TRANSPORT(/ansible/ansible.cfg) = smart
HOST_KEY_CHECKING(/ansible/ansible.cfg) = False
INVENTORY_ENABLED(/ansible/ansible.cfg) = ['vue.azure.azure_rm', 'script', 'yaml', 'ini']
PERSISTENT_COMMAND_TIMEOUT(/ansible/ansible.cfg) = 20
```
##### OS / ENVIRONMENT
running on a recent awx-ee image that has had additional packages/modules installed
##### STEPS TO REPRODUCE
example one: setting vm tags
```yaml
- name: Set patch_version tags
azure.azcollection.azure_rm_virtualmachine:
resource_group: "{{ resource_group }}"
name: "{{ name }}"
tags:
patch_version: "{{ ansible_date_time.iso8601 }}"
zones: "{{ availability_zone }}"
delegate_to: localhost
when:
- "'patch_version' in tags"
- tags['patch_version'] == 'none'
```
example two: rebooting a vm via azure_rm_virtualmachine:
```
- name: Tell Azure to reboot this VM (so UAC setting change takes effect)
azure.azcollection.azure_rm_virtualmachine:
resource_group: "{{ resource_group }}"
name: "{{ name }}"
restarted: yes
zones: "{{ availability_zone }}"
delegate_to: localhost
connection: local
```
##### EXPECTED RESULTS
VM tags set, or VM rebooted, no fatal errors
##### ACTUAL RESULTS
```paste below
TASK [windows : Set patch_version tags] ***************************************************************************************************************
task path: /ansible/roles/windows/tasks/patching_apply.yml:32
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: root
<localhost> EXEC /bin/sh -c 'echo ~root && sleep 0'
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" && echo ansible-tmp-1671211081.1372511-135-112305740467344="` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" ) && sleep 0'
Using module file /home/runner/.ansible/collections/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py
<localhost> PUT /home/runner/.ansible/tmp/ansible-local-64whuibyec/tmpa3z1tq85 TO /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py
<localhost> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0'
<localhost> EXEC /bin/sh -c '/usr/bin/python3 /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0'
<localhost> EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ > /dev/null 2>&1 && sleep 0'
fatal: [************-vm0 -> localhost]: FAILED! => {
"changed": false,
"module_stderr": "ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: ****\r\nTimestamp: 2022-12-16 17:18:11Z\nTraceback (most recent call last):\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module>\n _ansiballz_main()\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main\n invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module\n runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path),\n File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module\n return _run_module_code(code, init_globals, run_name, mod_spec)\n File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code\n _run_code(code, mod_globals, init_globals,\n File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code\n exec(code, run_globals)\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module>\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module\n File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer\n return func(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run\n return first_node.send(pipeline_request) # type: ignore\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n [Previous line repeated 2 more times]\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send\n self.on_request(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request\n self._token = self._credential.get_token(*self._scopes)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token\n token = self._request_token(*scopes, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper\n return fn(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token\n raise ClientAuthenticationError(message=message)\nazure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: *****\r\nTimestamp: 2022-12-16 17:18:11Z\n",
"module_stdout": "",
"msg": "MODULE FAILURE\nSee stdout/stderr for the exact error",
"rc": 1
}
```
decoded:
```
ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.
Trace ID: ****
Correlation ID: ****
Timestamp: 2022-12-16 16:08:27Z
Traceback (most recent call last):
File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module>
_ansiballz_main()
File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main
invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)
File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module
runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path),
File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module
return _run_module_code(code, init_globals, run_name, mod_spec)
File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code
_run_code(code, mod_globals, init_globals,
File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code
exec(code, run_globals)
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module>
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module
File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer
return func(*args, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get
pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run
return first_node.send(pipeline_request) # type: ignore
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send
response = self.next.send(request)
[Previous line repeated 2 more times]
File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send
self.on_request(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request
self._token = self._credential.get_token(*self._scopes)
File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token
token = self._request_token(*scopes, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper
return fn(*args, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token
raise ClientAuthenticationError(message=message)
azure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.
```
| 1.0 | azure_rm_virtualmachine fails in AzureChinaCloud - ##### SUMMARY
something called by azure_rm_virtualmachine does not obey AZURE_CLOUD_ENVIRONMENT and attempts to connect to AzureCloud when operating on AzureChinaCloud hosts
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
azure_rm_virtualmachine
##### ANSIBLE VERSION
<!--- Paste verbatim output from "ansible --version" between quotes -->
```paste below
ansible-playbook [core 2.12.5.post0]
config file = /ansible/ansible.cfg
configured module search path = ['/home/runner/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.8/site-packages/ansible
ansible collection location = /home/runner/.ansible/collections:/ansible/collections
executable location = /usr/local/bin/ansible-playbook
python version = 3.8.13 (default, Jun 24 2022, 15:27:57) [GCC 8.5.0 20210514 (Red Hat 8.5.0-13)]
jinja version = 3.1.2
libyaml = True
```
##### COLLECTION VERSION
<!--- Paste verbatim output from "ansible-galaxy collection list <namespace>.<collection>" between the quotes
for example: ansible-galaxy collection list community.general
-->
```paste below
- name: azure.azcollection
version: '==1.14.0'
```
##### CONFIGURATION
<!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes -->
```paste below
ANY_ERRORS_FATAL(/ansible/ansible.cfg) = False
COLLECTIONS_PATHS(/ansible/ansible.cfg) = ['/home/runner/.ansible/collections', '/ansible/collections']
DEFAULT_FORKS(/ansible/ansible.cfg) = 50
DEFAULT_GATHER_SUBSET(/ansible/ansible.cfg) = ['all']
DEFAULT_GATHER_TIMEOUT(/ansible/ansible.cfg) = 20
DEFAULT_HASH_BEHAVIOUR(/ansible/ansible.cfg) = merge
DEFAULT_TIMEOUT(/ansible/ansible.cfg) = 20
DEFAULT_TRANSPORT(/ansible/ansible.cfg) = smart
HOST_KEY_CHECKING(/ansible/ansible.cfg) = False
INVENTORY_ENABLED(/ansible/ansible.cfg) = ['vue.azure.azure_rm', 'script', 'yaml', 'ini']
PERSISTENT_COMMAND_TIMEOUT(/ansible/ansible.cfg) = 20
```
##### OS / ENVIRONMENT
running on a recent awx-ee image that has had additional packages/modules installed
##### STEPS TO REPRODUCE
example one: setting vm tags
```yaml
- name: Set patch_version tags
azure.azcollection.azure_rm_virtualmachine:
resource_group: "{{ resource_group }}"
name: "{{ name }}"
tags:
patch_version: "{{ ansible_date_time.iso8601 }}"
zones: "{{ availability_zone }}"
delegate_to: localhost
when:
- "'patch_version' in tags"
- tags['patch_version'] == 'none'
```
example two: rebooting a vm via azure_rm_virtualmachine:
```
- name: Tell Azure to reboot this VM (so UAC setting change takes effect)
azure.azcollection.azure_rm_virtualmachine:
resource_group: "{{ resource_group }}"
name: "{{ name }}"
restarted: yes
zones: "{{ availability_zone }}"
delegate_to: localhost
connection: local
```
##### EXPECTED RESULTS
VM tags set, or VM rebooted, no fatal errors
##### ACTUAL RESULTS
```paste below
TASK [windows : Set patch_version tags] ***************************************************************************************************************
task path: /ansible/roles/windows/tasks/patching_apply.yml:32
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: root
<localhost> EXEC /bin/sh -c 'echo ~root && sleep 0'
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /root/.ansible/tmp `"&& mkdir "` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" && echo ansible-tmp-1671211081.1372511-135-112305740467344="` echo /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344 `" ) && sleep 0'
Using module file /home/runner/.ansible/collections/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py
<localhost> PUT /home/runner/.ansible/tmp/ansible-local-64whuibyec/tmpa3z1tq85 TO /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py
<localhost> EXEC /bin/sh -c 'chmod u+x /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0'
<localhost> EXEC /bin/sh -c '/usr/bin/python3 /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py && sleep 0'
<localhost> EXEC /bin/sh -c 'rm -f -r /root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/ > /dev/null 2>&1 && sleep 0'
fatal: [************-vm0 -> localhost]: FAILED! => {
"changed": false,
"module_stderr": "ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: ****\r\nTimestamp: 2022-12-16 17:18:11Z\nTraceback (most recent call last):\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module>\n _ansiballz_main()\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main\n invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)\n File \"/root/.ansible/tmp/ansible-tmp-1671211081.1372511-135-112305740467344/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module\n runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path),\n File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module\n return _run_module_code(code, init_globals, run_name, mod_spec)\n File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code\n _run_code(code, mod_globals, init_globals,\n File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code\n exec(code, run_globals)\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module>\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__\n File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_4tp4ffhc/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module\n File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer\n return func(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get\n pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run\n return first_node.send(pipeline_request) # type: ignore\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send\n response = self.next.send(request)\n [Previous line repeated 2 more times]\n File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send\n response = self.next.send(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send\n self.on_request(request)\n File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request\n self._token = self._credential.get_token(*self._scopes)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token\n token = self._request_token(*scopes, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper\n return fn(*args, **kwargs)\n File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token\n raise ClientAuthenticationError(message=message)\nazure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.\r\nTrace ID: ****\r\nCorrelation ID: *****\r\nTimestamp: 2022-12-16 17:18:11Z\n",
"module_stdout": "",
"msg": "MODULE FAILURE\nSee stdout/stderr for the exact error",
"rc": 1
}
```
decoded:
```
ClientSecretCredential.get_token failed: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.
Trace ID: ****
Correlation ID: ****
Timestamp: 2022-12-16 16:08:27Z
Traceback (most recent call last):
File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 107, in <module>
_ansiballz_main()
File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 99, in _ansiballz_main
invoke_module(zipped_mod, temp_path, ANSIBALLZ_PARAMS)
File \"/root/.ansible/tmp/ansible-tmp-1671206898.300529-135-138589534907732/AnsiballZ_azure_rm_virtualmachine.py\", line 47, in invoke_module
runpy.run_module(mod_name='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', init_globals=dict(_module_fqn='ansible_collections.azure.azcollection.plugins.modules.azure_rm_virtualmachine', _modlib_path=modlib_path),
File \"/usr/lib64/python3.8/runpy.py\", line 207, in run_module
return _run_module_code(code, init_globals, run_name, mod_spec)
File \"/usr/lib64/python3.8/runpy.py\", line 97, in _run_module_code
_run_code(code, mod_globals, init_globals,
File \"/usr/lib64/python3.8/runpy.py\", line 87, in _run_code
exec(code, run_globals)
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2344, in <module>
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 2340, in main
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 963, in __init__
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/module_utils/azure_rm_common.py\", line 469, in __init__
File \"/tmp/ansible_azure.azcollection.azure_rm_virtualmachine_payload_jbwif2z3/ansible_azure.azcollection.azure_rm_virtualmachine_payload.zip/ansible_collections/azure/azcollection/plugins/modules/azure_rm_virtualmachine.py\", line 1114, in exec_module
File \"/usr/local/lib/python3.8/site-packages/azure/core/tracing/decorator.py\", line 78, in wrapper_use_tracer
return func(*args, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/compute/v2021_04_01/operations/_virtual_machines_operations.py\", line 1502, in get
pipeline_response = self._client._pipeline.run(request, stream=False, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 211, in run
return first_node.send(pipeline_request) # type: ignore
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/_base.py\", line 71, in send
response = self.next.send(request)
[Previous line repeated 2 more times]
File \"/usr/local/lib/python3.8/site-packages/azure/mgmt/core/policies/_base.py\", line 47, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_redirect.py\", line 158, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_retry.py\", line 446, in send
response = self.next.send(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 116, in send
self.on_request(request)
File \"/usr/local/lib/python3.8/site-packages/azure/core/pipeline/policies/_authentication.py\", line 93, in on_request
self._token = self._credential.get_token(*self._scopes)
File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/get_token_mixin.py\", line 76, in get_token
token = self._request_token(*scopes, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/decorators.py\", line 56, in wrapper
return fn(*args, **kwargs)
File \"/usr/local/lib/python3.8/site-packages/azure/identity/_internal/client_credential_base.py\", line 40, in _request_token
raise ClientAuthenticationError(message=message)
azure.core.exceptions.ClientAuthenticationError: Authentication failed: AADSTS500011: The resource principal named https://management.azure.com was not found in the tenant named PVUECN. This can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant. You might have sent your authentication request to the wrong tenant.
```
| priority | azure rm virtualmachine fails in azurechinacloud summary something called by azure rm virtualmachine does not obey azure cloud environment and attempts to connect to azurecloud when operating on azurechinacloud hosts issue type bug report component name azure rm virtualmachine ansible version paste below ansible playbook config file ansible ansible cfg configured module search path ansible python module location usr local lib site packages ansible ansible collection location home runner ansible collections ansible collections executable location usr local bin ansible playbook python version default jun jinja version libyaml true collection version between the quotes for example ansible galaxy collection list community general paste below name azure azcollection version configuration paste below any errors fatal ansible ansible cfg false collections paths ansible ansible cfg default forks ansible ansible cfg default gather subset ansible ansible cfg default gather timeout ansible ansible cfg default hash behaviour ansible ansible cfg merge default timeout ansible ansible cfg default transport ansible ansible cfg smart host key checking ansible ansible cfg false inventory enabled ansible ansible cfg persistent command timeout ansible ansible cfg os environment running on a recent awx ee image that has had additional packages modules installed steps to reproduce example one setting vm tags yaml name set patch version tags azure azcollection azure rm virtualmachine resource group resource group name name tags patch version ansible date time zones availability zone delegate to localhost when patch version in tags tags none example two rebooting a vm via azure rm virtualmachine name tell azure to reboot this vm so uac setting change takes effect azure azcollection azure rm virtualmachine resource group resource group name name restarted yes zones availability zone delegate to localhost connection local expected results vm tags set or vm rebooted no fatal errors actual results paste below task task path ansible roles windows tasks patching apply yml establish local connection for user root exec bin sh c echo root sleep exec bin sh c umask mkdir p echo root ansible tmp mkdir echo root ansible tmp ansible tmp echo ansible tmp echo root ansible tmp ansible tmp sleep using module file home runner ansible collections ansible collections azure azcollection plugins modules azure rm virtualmachine py put home runner ansible tmp ansible local to root ansible tmp ansible tmp ansiballz azure rm virtualmachine py exec bin sh c chmod u x root ansible tmp ansible tmp root ansible tmp ansible tmp ansiballz azure rm virtualmachine py sleep exec bin sh c usr bin root ansible tmp ansible tmp ansiballz azure rm virtualmachine py sleep exec bin sh c rm f r root ansible tmp ansible tmp dev null sleep fatal failed changed false module stderr clientsecretcredential get token failed authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant r ntrace id r ncorrelation id r ntimestamp ntraceback most recent call last n file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in n ansiballz main n file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in ansiballz main n invoke module zipped mod temp path ansiballz params n file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in invoke module n runpy run module mod name ansible collections azure azcollection plugins modules azure rm virtualmachine init globals dict module fqn ansible collections azure azcollection plugins modules azure rm virtualmachine modlib path modlib path n file usr runpy py line in run module n return run module code code init globals run name mod spec n file usr runpy py line in run module code n run code code mod globals init globals n file usr runpy py line in run code n exec code run globals n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in main n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in init n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins module utils azure rm common py line in init n file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in exec module n file usr local lib site packages azure core tracing decorator py line in wrapper use tracer n return func args kwargs n file usr local lib site packages azure mgmt compute operations virtual machines operations py line in get n pipeline response self client pipeline run request stream false kwargs n file usr local lib site packages azure core pipeline base py line in run n return first node send pipeline request type ignore n file usr local lib site packages azure core pipeline base py line in send n response self next send request n file usr local lib site packages azure core pipeline base py line in send n response self next send request n file usr local lib site packages azure core pipeline base py line in send n response self next send request n n file usr local lib site packages azure mgmt core policies base py line in send n response self next send request n file usr local lib site packages azure core pipeline policies redirect py line in send n response self next send request n file usr local lib site packages azure core pipeline policies retry py line in send n response self next send request n file usr local lib site packages azure core pipeline policies authentication py line in send n self on request request n file usr local lib site packages azure core pipeline policies authentication py line in on request n self token self credential get token self scopes n file usr local lib site packages azure identity internal get token mixin py line in get token n token self request token scopes kwargs n file usr local lib site packages azure identity internal decorators py line in wrapper n return fn args kwargs n file usr local lib site packages azure identity internal client credential base py line in request token n raise clientauthenticationerror message message nazure core exceptions clientauthenticationerror authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant r ntrace id r ncorrelation id r ntimestamp n module stdout msg module failure nsee stdout stderr for the exact error rc decoded clientsecretcredential get token failed authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant trace id correlation id timestamp traceback most recent call last file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in ansiballz main file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in ansiballz main invoke module zipped mod temp path ansiballz params file root ansible tmp ansible tmp ansiballz azure rm virtualmachine py line in invoke module runpy run module mod name ansible collections azure azcollection plugins modules azure rm virtualmachine init globals dict module fqn ansible collections azure azcollection plugins modules azure rm virtualmachine modlib path modlib path file usr runpy py line in run module return run module code code init globals run name mod spec file usr runpy py line in run module code run code code mod globals init globals file usr runpy py line in run code exec code run globals file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in main file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in init file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins module utils azure rm common py line in init file tmp ansible azure azcollection azure rm virtualmachine payload ansible azure azcollection azure rm virtualmachine payload zip ansible collections azure azcollection plugins modules azure rm virtualmachine py line in exec module file usr local lib site packages azure core tracing decorator py line in wrapper use tracer return func args kwargs file usr local lib site packages azure mgmt compute operations virtual machines operations py line in get pipeline response self client pipeline run request stream false kwargs file usr local lib site packages azure core pipeline base py line in run return first node send pipeline request type ignore file usr local lib site packages azure core pipeline base py line in send response self next send request file usr local lib site packages azure core pipeline base py line in send response self next send request file usr local lib site packages azure core pipeline base py line in send response self next send request file usr local lib site packages azure mgmt core policies base py line in send response self next send request file usr local lib site packages azure core pipeline policies redirect py line in send response self next send request file usr local lib site packages azure core pipeline policies retry py line in send response self next send request file usr local lib site packages azure core pipeline policies authentication py line in send self on request request file usr local lib site packages azure core pipeline policies authentication py line in on request self token self credential get token self scopes file usr local lib site packages azure identity internal get token mixin py line in get token token self request token scopes kwargs file usr local lib site packages azure identity internal decorators py line in wrapper return fn args kwargs file usr local lib site packages azure identity internal client credential base py line in request token raise clientauthenticationerror message message azure core exceptions clientauthenticationerror authentication failed the resource principal named was not found in the tenant named pvuecn this can happen if the application has not been installed by the administrator of the tenant or consented to by any user in the tenant you might have sent your authentication request to the wrong tenant | 1 |
812,561 | 30,342,069,154 | IssuesEvent | 2023-07-11 13:20:27 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | [Coverity CID: 318645] Out-of-bounds access in subsys/bluetooth/controller/ll_sw/ull_adv_aux.c | bug priority: medium area: Bluetooth Coverity area: Bluetooth Controller |
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c#L842
Category: Memory - corruptions
Function: `ll_adv_aux_sr_data_set`
Component: Bluetooth
CID: [318645](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=318645)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c
```
828
829 if (op == BT_HCI_LE_EXT_ADV_OP_INTERM_FRAG ||
830 op == BT_HCI_LE_EXT_ADV_OP_LAST_FRAG) {
831 /* Append fragment to existing data */
832 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA |
833 ULL_ADV_PDU_HDR_FIELD_AD_DATA_APPEND;
>>> CID 318645: (OVERRUN)
>>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20.
834 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu,
835 hdr_add_fields,
836 0,
837 hdr_data);
838 } else {
839 /* Add AD Data and remove any prior presence of Aux Ptr */
806 */
807 *val_ptr++ = len;
808 (void)memcpy(val_ptr, &data, sizeof(data));
809 }
810
811 /* Trigger DID update */
>>> CID 318645: (OVERRUN)
>>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20.
812 err = ull_adv_aux_hdr_set_clear(adv, hdr_add_fields, 0U,
813 hdr_data, &pri_idx, &sec_idx);
814 if (err) {
815 return err;
816 }
817
836 0,
837 hdr_data);
838 } else {
839 /* Add AD Data and remove any prior presence of Aux Ptr */
840 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA |
841 ULL_ADV_PDU_HDR_FIELD_AD_DATA;
>>> CID 318645: (OVERRUN)
>>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20.
842 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu,
843 hdr_add_fields,
844 ULL_ADV_PDU_HDR_FIELD_AUX_PTR,
845 hdr_data);
846 }
847 #if defined(CONFIG_BT_CTLR_ADV_AUX_PDU_LINK)
```
For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_OVERRUN). ([CWE-119](http://cwe.mitre.org/data/definitions/119.html))
Please fix or provide comments in coverity using the link:
https://scan9.scan.coverity.com/reports.htm#v29271/p12996
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| 1.0 | [Coverity CID: 318645] Out-of-bounds access in subsys/bluetooth/controller/ll_sw/ull_adv_aux.c -
Static code scan issues found in file:
https://github.com/zephyrproject-rtos/zephyr/tree/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c#L842
Category: Memory - corruptions
Function: `ll_adv_aux_sr_data_set`
Component: Bluetooth
CID: [318645](https://scan9.scan.coverity.com/reports.htm#v29726/p12996/mergedDefectId=318645)
Details:
https://github.com/zephyrproject-rtos/zephyr/blob/7b2034aaecc4cb2a261973b10b2fa608b29d398c/subsys/bluetooth/controller/ll_sw/ull_adv_aux.c
```
828
829 if (op == BT_HCI_LE_EXT_ADV_OP_INTERM_FRAG ||
830 op == BT_HCI_LE_EXT_ADV_OP_LAST_FRAG) {
831 /* Append fragment to existing data */
832 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA |
833 ULL_ADV_PDU_HDR_FIELD_AD_DATA_APPEND;
>>> CID 318645: (OVERRUN)
>>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20.
834 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu,
835 hdr_add_fields,
836 0,
837 hdr_data);
838 } else {
839 /* Add AD Data and remove any prior presence of Aux Ptr */
806 */
807 *val_ptr++ = len;
808 (void)memcpy(val_ptr, &data, sizeof(data));
809 }
810
811 /* Trigger DID update */
>>> CID 318645: (OVERRUN)
>>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20.
812 err = ull_adv_aux_hdr_set_clear(adv, hdr_add_fields, 0U,
813 hdr_data, &pri_idx, &sec_idx);
814 if (err) {
815 return err;
816 }
817
836 0,
837 hdr_data);
838 } else {
839 /* Add AD Data and remove any prior presence of Aux Ptr */
840 hdr_add_fields |= ULL_ADV_PDU_HDR_FIELD_ADVA |
841 ULL_ADV_PDU_HDR_FIELD_AD_DATA;
>>> CID 318645: (OVERRUN)
>>> Overrunning array "hdr_data" of 16 bytes by passing it to a function which accesses it at byte offset 20.
842 err = ull_adv_aux_pdu_set_clear(adv, sr_pdu_prev, sr_pdu,
843 hdr_add_fields,
844 ULL_ADV_PDU_HDR_FIELD_AUX_PTR,
845 hdr_data);
846 }
847 #if defined(CONFIG_BT_CTLR_ADV_AUX_PDU_LINK)
```
For more information about the violation, check the [Coverity Reference](https://scan9.scan.coverity.com/doc/en/cov_checker_ref.html#static_checker_OVERRUN). ([CWE-119](http://cwe.mitre.org/data/definitions/119.html))
Please fix or provide comments in coverity using the link:
https://scan9.scan.coverity.com/reports.htm#v29271/p12996
Note: This issue was created automatically. Priority was set based on classification
of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
| priority | out of bounds access in subsys bluetooth controller ll sw ull adv aux c static code scan issues found in file category memory corruptions function ll adv aux sr data set component bluetooth cid details if op bt hci le ext adv op interm frag op bt hci le ext adv op last frag append fragment to existing data hdr add fields ull adv pdu hdr field adva ull adv pdu hdr field ad data append cid overrun overrunning array hdr data of bytes by passing it to a function which accesses it at byte offset err ull adv aux pdu set clear adv sr pdu prev sr pdu hdr add fields hdr data else add ad data and remove any prior presence of aux ptr val ptr len void memcpy val ptr data sizeof data trigger did update cid overrun overrunning array hdr data of bytes by passing it to a function which accesses it at byte offset err ull adv aux hdr set clear adv hdr add fields hdr data pri idx sec idx if err return err hdr data else add ad data and remove any prior presence of aux ptr hdr add fields ull adv pdu hdr field adva ull adv pdu hdr field ad data cid overrun overrunning array hdr data of bytes by passing it to a function which accesses it at byte offset err ull adv aux pdu set clear adv sr pdu prev sr pdu hdr add fields ull adv pdu hdr field aux ptr hdr data if defined config bt ctlr adv aux pdu link for more information about the violation check the please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file | 1 |
85,953 | 3,700,889,459 | IssuesEvent | 2016-02-29 10:38:45 | OCHA-DAP/hdx-ckan | https://api.github.com/repos/OCHA-DAP/hdx-ckan | closed | New contribute flow: Horizontal scrollbar showing | bug New Contribute Flow Priority-Medium | I tested in Linux - Firefox and Chrome in FullHD, and in Windows in Firefox:

Notice the scrollbar at the bottom | 1.0 | New contribute flow: Horizontal scrollbar showing - I tested in Linux - Firefox and Chrome in FullHD, and in Windows in Firefox:

Notice the scrollbar at the bottom | priority | new contribute flow horizontal scrollbar showing i tested in linux firefox and chrome in fullhd and in windows in firefox notice the scrollbar at the bottom | 1 |
424,564 | 12,313,098,666 | IssuesEvent | 2020-05-12 14:50:56 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | Come up with a way to handle reviews for the feature replacement workflow | Category: Core Priority: Medium Status: New/Undefined Type: Feature | Two types of reviews could be useful during feature replacement. The first would be the regular conflate review and another could be flagging reviews for linear features snapped post conflation (not yet added; could group reviews by connected ways to reduce the number of them). Since our output is a changeset and the command works outside of the UI review workflow, there is currently no way to handle reviews and they are dropped completely. | 1.0 | Come up with a way to handle reviews for the feature replacement workflow - Two types of reviews could be useful during feature replacement. The first would be the regular conflate review and another could be flagging reviews for linear features snapped post conflation (not yet added; could group reviews by connected ways to reduce the number of them). Since our output is a changeset and the command works outside of the UI review workflow, there is currently no way to handle reviews and they are dropped completely. | priority | come up with a way to handle reviews for the feature replacement workflow two types of reviews could be useful during feature replacement the first would be the regular conflate review and another could be flagging reviews for linear features snapped post conflation not yet added could group reviews by connected ways to reduce the number of them since our output is a changeset and the command works outside of the ui review workflow there is currently no way to handle reviews and they are dropped completely | 1 |
679,463 | 23,233,245,396 | IssuesEvent | 2022-08-03 09:28:25 | owncloud/web | https://api.github.com/repos/owncloud/web | closed | Selected item glues on bottom & scrolls for ⬆️ key up | Type:Bug Priority:p3-medium GA-Blocker | ### Steps to reproduce
1. Login to https://ocis.ocis-web.latest.owncloud.works/
2. upload ~100 files into a single folder
3. select one of the 100 files
4. press key down ⬇️ until scolling starts
5. press key ⬆️ until scrolling starts
6. Selected item glues on bottom and scrolls on every keystroke
https://user-images.githubusercontent.com/26610733/180446452-a27ec8b3-ac8d-45aa-9635-4e21cd9132b4.mp4
### Expected behaviour
selection should "walk" to the top without scrolling, then start scrolling
https://user-images.githubusercontent.com/26610733/180445840-e7deb84f-da62-41c6-b625-156e48980383.mp4
### Actual behaviour
Selected item glues on bottom and scrolls on every keystroke
| 1.0 | Selected item glues on bottom & scrolls for ⬆️ key up - ### Steps to reproduce
1. Login to https://ocis.ocis-web.latest.owncloud.works/
2. upload ~100 files into a single folder
3. select one of the 100 files
4. press key down ⬇️ until scolling starts
5. press key ⬆️ until scrolling starts
6. Selected item glues on bottom and scrolls on every keystroke
https://user-images.githubusercontent.com/26610733/180446452-a27ec8b3-ac8d-45aa-9635-4e21cd9132b4.mp4
### Expected behaviour
selection should "walk" to the top without scrolling, then start scrolling
https://user-images.githubusercontent.com/26610733/180445840-e7deb84f-da62-41c6-b625-156e48980383.mp4
### Actual behaviour
Selected item glues on bottom and scrolls on every keystroke
| priority | selected item glues on bottom scrolls for ⬆️ key up steps to reproduce login to upload files into a single folder select one of the files press key down ⬇️ until scolling starts press key ⬆️ until scrolling starts selected item glues on bottom and scrolls on every keystroke expected behaviour selection should walk to the top without scrolling then start scrolling actual behaviour selected item glues on bottom and scrolls on every keystroke | 1 |
205,052 | 7,093,594,910 | IssuesEvent | 2018-01-12 21:12:42 | certificate-helper/TLS-Inspector | https://api.github.com/repos/certificate-helper/TLS-Inspector | closed | Limit the number of redirects TLS Inspector will follow | CertificateKit bug easy medium priority merged | **Affected Version:**
Since 1.6.0
**Is this a Test Flight version or the App Store version?**
App Store
**Device and iOS Version:**
All
**What steps will reproduce the problem?**
1. Navigate to a web page that triggers a redirect loop
**What is the expected output?**
A specific warning about a redirect loop
**What do you see instead?**
Long loading then eventually timeout
**Please provide any additional information below.**
| 1.0 | Limit the number of redirects TLS Inspector will follow - **Affected Version:**
Since 1.6.0
**Is this a Test Flight version or the App Store version?**
App Store
**Device and iOS Version:**
All
**What steps will reproduce the problem?**
1. Navigate to a web page that triggers a redirect loop
**What is the expected output?**
A specific warning about a redirect loop
**What do you see instead?**
Long loading then eventually timeout
**Please provide any additional information below.**
| priority | limit the number of redirects tls inspector will follow affected version since is this a test flight version or the app store version app store device and ios version all what steps will reproduce the problem navigate to a web page that triggers a redirect loop what is the expected output a specific warning about a redirect loop what do you see instead long loading then eventually timeout please provide any additional information below | 1 |
134,710 | 5,232,910,807 | IssuesEvent | 2017-01-30 11:03:23 | openworm/behavioral_syntax | https://api.github.com/repos/openworm/behavioral_syntax | closed | consider alternatives to compression(MDL) | medium priority | Following Andre's presentation at the OpenWorm journal club today it might be a good idea to look into using:
1) Hidden Markov Models
2) Statistical Analysis: use p-values to see whether there's any information gain from using n grams vs (n-1) grams
3) Use Hierarchical Markov Models
| 1.0 | consider alternatives to compression(MDL) - Following Andre's presentation at the OpenWorm journal club today it might be a good idea to look into using:
1) Hidden Markov Models
2) Statistical Analysis: use p-values to see whether there's any information gain from using n grams vs (n-1) grams
3) Use Hierarchical Markov Models
| priority | consider alternatives to compression mdl following andre s presentation at the openworm journal club today it might be a good idea to look into using hidden markov models statistical analysis use p values to see whether there s any information gain from using n grams vs n grams use hierarchical markov models | 1 |
623,214 | 19,663,333,600 | IssuesEvent | 2022-01-10 19:25:46 | ScottUK/ladojrp-issues | https://api.github.com/repos/ScottUK/ladojrp-issues | closed | PA System For Police | Class: enhancement Priority: medium Scope: scripts | **Describe the feature you'd like implemented**
There should be a PA system for police to use in their vehicles so if they need to yell for a vehicle to pull over or pull closer to the side of the road they can do so by using that implemented PA system.
| 1.0 | PA System For Police - **Describe the feature you'd like implemented**
There should be a PA system for police to use in their vehicles so if they need to yell for a vehicle to pull over or pull closer to the side of the road they can do so by using that implemented PA system.
| priority | pa system for police describe the feature you d like implemented there should be a pa system for police to use in their vehicles so if they need to yell for a vehicle to pull over or pull closer to the side of the road they can do so by using that implemented pa system | 1 |
437,075 | 12,559,884,624 | IssuesEvent | 2020-06-07 20:20:57 | DuckBoss/JJMumbleBot | https://api.github.com/repos/DuckBoss/JJMumbleBot | closed | Better documentation | Medium Priority Next Release Wiki | Hey,
I really like your project! For a better and quicker understanding of your projects and the installation process if would be really helpful to add some sample files (not a image) and improve the quickstart with more details and in depth examples. | 1.0 | Better documentation - Hey,
I really like your project! For a better and quicker understanding of your projects and the installation process if would be really helpful to add some sample files (not a image) and improve the quickstart with more details and in depth examples. | priority | better documentation hey i really like your project for a better and quicker understanding of your projects and the installation process if would be really helpful to add some sample files not a image and improve the quickstart with more details and in depth examples | 1 |
211,567 | 7,202,254,780 | IssuesEvent | 2018-02-06 02:49:34 | Albert481/myTrolley | https://api.github.com/repos/Albert481/myTrolley | closed | Report trolley condition | Medium NTUC Shoppers Priority Normal | report condition of trolley, provide comment box to specify which part if necessary | 1.0 | Report trolley condition - report condition of trolley, provide comment box to specify which part if necessary | priority | report trolley condition report condition of trolley provide comment box to specify which part if necessary | 1 |
218,543 | 7,331,716,185 | IssuesEvent | 2018-03-05 14:23:41 | Mapita/shorter | https://api.github.com/repos/Mapita/shorter | opened | Use Travis CI to automatically run unit tests before approving PRs | effort: low meta: blocked meta: next release priority: medium type: integration | Travis CI should be used to run tests for the master branch and before merging PRs.
Blocked by https://github.com/Mapita/shorter/issues/4 | 1.0 | Use Travis CI to automatically run unit tests before approving PRs - Travis CI should be used to run tests for the master branch and before merging PRs.
Blocked by https://github.com/Mapita/shorter/issues/4 | priority | use travis ci to automatically run unit tests before approving prs travis ci should be used to run tests for the master branch and before merging prs blocked by | 1 |
382,258 | 11,303,041,845 | IssuesEvent | 2020-01-17 19:07:34 | chicagopython/chipy.org | https://api.github.com/repos/chicagopython/chipy.org | closed | Remove Maybe option from RSVP selection | priority: medium status: revision needed type: enhancement | As we will be using the website going forward, let's remove the ambiguity of Maybe from RSVP options.
I'm gonna take this one. Will be a good way to dig into the codebase. | 1.0 | Remove Maybe option from RSVP selection - As we will be using the website going forward, let's remove the ambiguity of Maybe from RSVP options.
I'm gonna take this one. Will be a good way to dig into the codebase. | priority | remove maybe option from rsvp selection as we will be using the website going forward let s remove the ambiguity of maybe from rsvp options i m gonna take this one will be a good way to dig into the codebase | 1 |
696,676 | 23,911,107,177 | IssuesEvent | 2022-09-09 08:17:45 | Chatterino/chatterino2 | https://api.github.com/repos/Chatterino/chatterino2 | closed | Deprecation of IRC commands | enhancement Platform: Twitch Priority: Medium Deprecation: Twitch IRC Commands | On or around Feb 18th, 2023, Twitch will deprecate all commands currently executed through IRC, `excluding /disconnect & /me`
More information on this can be found [here.](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486)
This will mean the following commands will need to be transferred from IRC to Helix, in Chatterino:
<details>
<summary>Command List</summary>
| Completed | Command | Documentation |
| ----------- | ----------- | ----------- |
| ❌ | `/announce` | https://dev.twitch.tv/docs/api/reference#send-chat-announcement |
| ❌ | `/ban` | https://dev.twitch.tv/docs/api/reference#ban-user ³|
| ❌ | `/unban` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴|
| ❌ | `/clear` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹|
| ❌ | `/color` | https://dev.twitch.tv/docs/api/reference#update-user-chat-color |
| ❌ | `/commercial` | https://dev.twitch.tv/docs/api/reference#start-commercial |
| ❌ | `/delete` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹|
| ❌ | `/emoteonly` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/emoteonlyoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/followers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/followersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ✅ | `/marker` | https://dev.twitch.tv/docs/api/reference#create-stream-marker |
| ❌ | `/mod` | https://dev.twitch.tv/docs/api/reference#add-channel-moderator |
| ❌ | `/mods` | https://dev.twitch.tv/docs/api/reference#get-moderators ⁵|
| ❌ | `/unmod` | https://dev.twitch.tv/docs/api/reference#remove-channel-moderator |
| ❌ | `/raid` | https://dev.twitch.tv/docs/api/reference#start-a-raid ⁵|
| ❌ | `/unraid` | https://dev.twitch.tv/docs/api/reference#cancel-a-raid ⁵|
| ❌ | `/slow` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/slowoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/subscribers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/subscribersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/timeout` | https://dev.twitch.tv/docs/api/reference#ban-user ³|
| ❌ | `/untimeout` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴|
| ❌ | `/uniquechat` | https://dev.twitch.tv/docs/api/reference#update-chat-settings |
| ❌ | `/uniquechatoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings |
| ❌ | `/vip` | https://dev.twitch.tv/docs/api/reference#add-channel-vip |
| ❌ | `/vips` | https://dev.twitch.tv/docs/api/reference#get-vips ⁵|
| ❌ | `/unvip` | https://dev.twitch.tv/docs/api/reference#remove-channel-vip |
| ❌ | `/w` | https://dev.twitch.tv/docs/api/reference#send-whisper ⁶|
1. `/clear` & `/delete` use the same API call
2. `/emoteonly`, `/emoteonlyoff`, `/followers`, `/followersoff`, `/slow`, `/slowoff`, `/subscribers`, `/subscribersoff`, `/uniquechat`, `/uniquechatoff` all use the same API call
3. `/ban` & `/timeout` use the same API call
4. `/unban` & `/untimeout` use the same API call
5. `/mods`, `/vips`, `/raid`, and `/unraid` are now broadcaster only commands
6. `/w` now requires Twitch phone number verification `This is not the same as 2FA`
<!--- here are the emojis you'll need to update this: ✅❌ --->
</details>
The newly required scopes have already been handled in Chatterino/website#77
As noted from the list above, the following commands will need custom error handling
- `/mods` - Will need a custom error for non-broadcasters stating they are no longer able use this command.
- `/vips` - ""
- `/raid` - ""
- `/unraid` - ""
- `/w` - Will require custom error handling when receiving back an error code 401, which will indicate the user has not verified their phone number with Twitch.
- Likely will point to the whisper section of the wiki, which will need to be updated showing where to add said phone number. `settings -> security` | 1.0 | Deprecation of IRC commands - On or around Feb 18th, 2023, Twitch will deprecate all commands currently executed through IRC, `excluding /disconnect & /me`
More information on this can be found [here.](https://discuss.dev.twitch.tv/t/deprecation-of-chat-commands-through-irc/40486)
This will mean the following commands will need to be transferred from IRC to Helix, in Chatterino:
<details>
<summary>Command List</summary>
| Completed | Command | Documentation |
| ----------- | ----------- | ----------- |
| ❌ | `/announce` | https://dev.twitch.tv/docs/api/reference#send-chat-announcement |
| ❌ | `/ban` | https://dev.twitch.tv/docs/api/reference#ban-user ³|
| ❌ | `/unban` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴|
| ❌ | `/clear` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹|
| ❌ | `/color` | https://dev.twitch.tv/docs/api/reference#update-user-chat-color |
| ❌ | `/commercial` | https://dev.twitch.tv/docs/api/reference#start-commercial |
| ❌ | `/delete` | https://dev.twitch.tv/docs/api/reference#delete-chat-messages ¹|
| ❌ | `/emoteonly` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/emoteonlyoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/followers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/followersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ✅ | `/marker` | https://dev.twitch.tv/docs/api/reference#create-stream-marker |
| ❌ | `/mod` | https://dev.twitch.tv/docs/api/reference#add-channel-moderator |
| ❌ | `/mods` | https://dev.twitch.tv/docs/api/reference#get-moderators ⁵|
| ❌ | `/unmod` | https://dev.twitch.tv/docs/api/reference#remove-channel-moderator |
| ❌ | `/raid` | https://dev.twitch.tv/docs/api/reference#start-a-raid ⁵|
| ❌ | `/unraid` | https://dev.twitch.tv/docs/api/reference#cancel-a-raid ⁵|
| ❌ | `/slow` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/slowoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/subscribers` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/subscribersoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings ²|
| ❌ | `/timeout` | https://dev.twitch.tv/docs/api/reference#ban-user ³|
| ❌ | `/untimeout` | https://dev.twitch.tv/docs/api/reference#unban-user ⁴|
| ❌ | `/uniquechat` | https://dev.twitch.tv/docs/api/reference#update-chat-settings |
| ❌ | `/uniquechatoff` | https://dev.twitch.tv/docs/api/reference#update-chat-settings |
| ❌ | `/vip` | https://dev.twitch.tv/docs/api/reference#add-channel-vip |
| ❌ | `/vips` | https://dev.twitch.tv/docs/api/reference#get-vips ⁵|
| ❌ | `/unvip` | https://dev.twitch.tv/docs/api/reference#remove-channel-vip |
| ❌ | `/w` | https://dev.twitch.tv/docs/api/reference#send-whisper ⁶|
1. `/clear` & `/delete` use the same API call
2. `/emoteonly`, `/emoteonlyoff`, `/followers`, `/followersoff`, `/slow`, `/slowoff`, `/subscribers`, `/subscribersoff`, `/uniquechat`, `/uniquechatoff` all use the same API call
3. `/ban` & `/timeout` use the same API call
4. `/unban` & `/untimeout` use the same API call
5. `/mods`, `/vips`, `/raid`, and `/unraid` are now broadcaster only commands
6. `/w` now requires Twitch phone number verification `This is not the same as 2FA`
<!--- here are the emojis you'll need to update this: ✅❌ --->
</details>
The newly required scopes have already been handled in Chatterino/website#77
As noted from the list above, the following commands will need custom error handling
- `/mods` - Will need a custom error for non-broadcasters stating they are no longer able use this command.
- `/vips` - ""
- `/raid` - ""
- `/unraid` - ""
- `/w` - Will require custom error handling when receiving back an error code 401, which will indicate the user has not verified their phone number with Twitch.
- Likely will point to the whisper section of the wiki, which will need to be updated showing where to add said phone number. `settings -> security` | priority | deprecation of irc commands on or around feb twitch will deprecate all commands currently executed through irc excluding disconnect me more information on this can be found this will mean the following commands will need to be transferred from irc to helix in chatterino command list completed command documentation ❌ announce ❌ ban ³ ❌ unban ⁴ ❌ clear ¹ ❌ color ❌ commercial ❌ delete ¹ ❌ emoteonly ² ❌ emoteonlyoff ² ❌ followers ² ❌ followersoff ² ✅ marker ❌ mod ❌ mods ⁵ ❌ unmod ❌ raid ⁵ ❌ unraid ⁵ ❌ slow ² ❌ slowoff ² ❌ subscribers ² ❌ subscribersoff ² ❌ timeout ³ ❌ untimeout ⁴ ❌ uniquechat ❌ uniquechatoff ❌ vip ❌ vips ⁵ ❌ unvip ❌ w ⁶ clear delete use the same api call emoteonly emoteonlyoff followers followersoff slow slowoff subscribers subscribersoff uniquechat uniquechatoff all use the same api call ban timeout use the same api call unban untimeout use the same api call mods vips raid and unraid are now broadcaster only commands w now requires twitch phone number verification this is not the same as the newly required scopes have already been handled in chatterino website as noted from the list above the following commands will need custom error handling mods will need a custom error for non broadcasters stating they are no longer able use this command vips raid unraid w will require custom error handling when receiving back an error code which will indicate the user has not verified their phone number with twitch likely will point to the whisper section of the wiki which will need to be updated showing where to add said phone number settings security | 1 |
781,477 | 27,439,114,080 | IssuesEvent | 2023-03-02 09:45:30 | vaticle/typedb-behaviour | https://api.github.com/repos/vaticle/typedb-behaviour | opened | Improve user management features | type: feature priority: medium | ## Problem to Solve
Our user management feature is fairly bare-bones right now, with TypeDB Cluster carrying the bulk of the weight for testing user management.
## Current Workaround
We test much of user management with TypeDB Cluster integration tests.
## Proposed Solution
We should expand the test suite and the available steps that can be executed during a test. These will be more stateful and akin to our tests which open and close sessions and transactions.
| 1.0 | Improve user management features - ## Problem to Solve
Our user management feature is fairly bare-bones right now, with TypeDB Cluster carrying the bulk of the weight for testing user management.
## Current Workaround
We test much of user management with TypeDB Cluster integration tests.
## Proposed Solution
We should expand the test suite and the available steps that can be executed during a test. These will be more stateful and akin to our tests which open and close sessions and transactions.
| priority | improve user management features problem to solve our user management feature is fairly bare bones right now with typedb cluster carrying the bulk of the weight for testing user management current workaround we test much of user management with typedb cluster integration tests proposed solution we should expand the test suite and the available steps that can be executed during a test these will be more stateful and akin to our tests which open and close sessions and transactions | 1 |
60,919 | 3,135,692,769 | IssuesEvent | 2015-09-10 16:19:32 | PeerSay/Atlas | https://api.github.com/repos/PeerSay/Atlas | closed | When all topic's requirements are undefined (i.e. '?') - the overall topic grade should be undefined | Priority: Medium | If all topic grades are '?' then the overall of the topic should show '?' and not 0. | 1.0 | When all topic's requirements are undefined (i.e. '?') - the overall topic grade should be undefined - If all topic grades are '?' then the overall of the topic should show '?' and not 0. | priority | when all topic s requirements are undefined i e the overall topic grade should be undefined if all topic grades are then the overall of the topic should show and not | 1 |
748,700 | 26,133,662,416 | IssuesEvent | 2022-12-29 09:16:21 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | [Quest] Victory At Death's Breach! | NPC Quest - Cataclysm (1-60) Priority: Medium Status: Confirmed Bug Report from Discord | Unintendet
OP
— 12/10/2022 6:54 PM
Quest = Victory At Death's Breach! ... Maybe I was to stupid for it but the Teleport didn't port me Up, or the Fly-guy to Fly Up is missing. | 1.0 | [Quest] Victory At Death's Breach! - Unintendet
OP
— 12/10/2022 6:54 PM
Quest = Victory At Death's Breach! ... Maybe I was to stupid for it but the Teleport didn't port me Up, or the Fly-guy to Fly Up is missing. | priority | victory at death s breach unintendet op — pm quest victory at death s breach maybe i was to stupid for it but the teleport didn t port me up or the fly guy to fly up is missing | 1 |
351,954 | 10,525,703,995 | IssuesEvent | 2019-09-30 15:33:47 | forceworkbench/forceworkbench | https://api.github.com/repos/forceworkbench/forceworkbench | closed | Show realtionship name instead of Type for relationship query headers | Component-Query Priority-Medium Scheduled-Backlog bug imported | _Original author: ryan.bra...@gmail.com (February 01, 2009 11:11:47)_
For a query like [SELECT Opportunity.LastModifiedBy.Name FROM
OpportunityContactRole], don't show Opportunity.User.Name.
Not sure if this is going to be possible because of the way the SOAP client
parses the tags
_Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=146_
| 1.0 | Show realtionship name instead of Type for relationship query headers - _Original author: ryan.bra...@gmail.com (February 01, 2009 11:11:47)_
For a query like [SELECT Opportunity.LastModifiedBy.Name FROM
OpportunityContactRole], don't show Opportunity.User.Name.
Not sure if this is going to be possible because of the way the SOAP client
parses the tags
_Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=146_
| priority | show realtionship name instead of type for relationship query headers original author ryan bra gmail com february for a query like select opportunity lastmodifiedby name from opportunitycontactrole don t show opportunity user name not sure if this is going to be possible because of the way the soap client parses the tags original issue | 1 |
351,921 | 10,525,701,180 | IssuesEvent | 2019-09-30 15:33:26 | forceworkbench/forceworkbench | https://api.github.com/repos/forceworkbench/forceworkbench | closed | add SOQL Pagination support to query builder | Component-Query Priority-Medium Scheduled-Backlog enhancement imported | _Original author: ryan.bra...@gmail.com (August 05, 2012 00:22:33)_
SOQL Pagination—Generally Available
SOQL Pagination enables you to specify the starting row of the result set your SOQL query returns. You use SOQL Pagination
via the SOQL OFFSET clause. Using SOQL OFFSET is helpful for paging through large result sets when you need to
quickly jump to a particular subset of the entire results. As the offset calculation is done on the server and only the result subset
is returned, using OFFSET is more efficient than retrieving the full result set and then filtering the results locally. For example,
this SOQL query returns a result set that skips the first 10 rows:
SELECT Name
FROM Merchandise__c
WHERE Price__c > 5.0
ORDER BY Name
LIMIT 100
OFFSET 10
The maximum offset is 2,000 rows. Requesting an offset greater than 2,000 will result in a NUMBER_OUTSIDE_VALID_RANGE
error.
_Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=584_
| 1.0 | add SOQL Pagination support to query builder - _Original author: ryan.bra...@gmail.com (August 05, 2012 00:22:33)_
SOQL Pagination—Generally Available
SOQL Pagination enables you to specify the starting row of the result set your SOQL query returns. You use SOQL Pagination
via the SOQL OFFSET clause. Using SOQL OFFSET is helpful for paging through large result sets when you need to
quickly jump to a particular subset of the entire results. As the offset calculation is done on the server and only the result subset
is returned, using OFFSET is more efficient than retrieving the full result set and then filtering the results locally. For example,
this SOQL query returns a result set that skips the first 10 rows:
SELECT Name
FROM Merchandise__c
WHERE Price__c > 5.0
ORDER BY Name
LIMIT 100
OFFSET 10
The maximum offset is 2,000 rows. Requesting an offset greater than 2,000 will result in a NUMBER_OUTSIDE_VALID_RANGE
error.
_Original issue: http://code.google.com/p/forceworkbench/issues/detail?id=584_
| priority | add soql pagination support to query builder original author ryan bra gmail com august soql pagination—generally available soql pagination enables you to specify the starting row of the result set your soql query returns you use soql pagination via the soql offset clause using soql offset is helpful for paging through large result sets when you need to quickly jump to a particular subset of the entire results as the offset calculation is done on the server and only the result subset is returned using offset is more efficient than retrieving the full result set and then filtering the results locally for example this soql query returns a result set that skips the first rows select name from merchandise c where price c gt order by name limit offset the maximum offset is rows requesting an offset greater than will result in a number outside valid range error original issue | 1 |
397,247 | 11,725,717,475 | IssuesEvent | 2020-03-10 13:26:32 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio-ui] Add a link from Studio to acknowledgements page in docs | priority: medium task | **Is your feature request related to a problem? Please describe.**
Add a link from Studio to acknowledgements page in docs
**Describe the solution you'd like**
Please add a link from the `Help` -> `About` screen in Studio with the link pointing to https://docs.craftercms.org/en/3.1/acknowledgements/index.html
(Something like the screen below, the message is not centered, but that's the idea =) )

`Crafter CMS is made possible by these other_open source software projects_ `
where `_open source software projects_` points to the acknowledgements in the docs
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | [studio-ui] Add a link from Studio to acknowledgements page in docs - **Is your feature request related to a problem? Please describe.**
Add a link from Studio to acknowledgements page in docs
**Describe the solution you'd like**
Please add a link from the `Help` -> `About` screen in Studio with the link pointing to https://docs.craftercms.org/en/3.1/acknowledgements/index.html
(Something like the screen below, the message is not centered, but that's the idea =) )

`Crafter CMS is made possible by these other_open source software projects_ `
where `_open source software projects_` points to the acknowledgements in the docs
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | add a link from studio to acknowledgements page in docs is your feature request related to a problem please describe add a link from studio to acknowledgements page in docs describe the solution you d like please add a link from the help about screen in studio with the link pointing to something like the screen below the message is not centered but that s the idea crafter cms is made possible by these other open source software projects where open source software projects points to the acknowledgements in the docs describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here | 1 |
758,289 | 26,548,974,730 | IssuesEvent | 2023-01-20 05:01:38 | apimatic/go-core-runtime | https://api.github.com/repos/apimatic/go-core-runtime | opened | [FEATURE REQUEST] - Add support for Retries and Backoff | enhancement priority-medium | Go's native HTTP client does not support Retrying.
Retrying allows an application to recover from transient failures while trying to reach out to a resource. So adding support for Retries and Backoff will only improve the API calls made. | 1.0 | [FEATURE REQUEST] - Add support for Retries and Backoff - Go's native HTTP client does not support Retrying.
Retrying allows an application to recover from transient failures while trying to reach out to a resource. So adding support for Retries and Backoff will only improve the API calls made. | priority | add support for retries and backoff go s native http client does not support retrying retrying allows an application to recover from transient failures while trying to reach out to a resource so adding support for retries and backoff will only improve the api calls made | 1 |
435,337 | 12,534,423,650 | IssuesEvent | 2020-06-04 19:24:24 | Energy-Innovation/eps-us | https://api.github.com/repos/Energy-Innovation/eps-us | opened | Allow carbon tax rate to be set separately for district heat & hydrogen sector, instead of inheriting rate from industry sector | medium priority | The district heat and hydrogen sector uses the same fuels as the industry sector. When the carbon tax lever was designed, it was assumed the tax was levied upstream on these fuels, and so the rate seen by industrial facilities and district heat/hydrogen facilities would be the same. But some countries consider district heat (at least, maybe also hydrogen supply) to be part of the energy sector and may exempt them from carbon taxes or tax them differently than the industry sector. Change the carbon tax lever so it is set separately for the district heat and hydrogen supply sector, rather than inheriting the industry sector's carbon tax rate. | 1.0 | Allow carbon tax rate to be set separately for district heat & hydrogen sector, instead of inheriting rate from industry sector - The district heat and hydrogen sector uses the same fuels as the industry sector. When the carbon tax lever was designed, it was assumed the tax was levied upstream on these fuels, and so the rate seen by industrial facilities and district heat/hydrogen facilities would be the same. But some countries consider district heat (at least, maybe also hydrogen supply) to be part of the energy sector and may exempt them from carbon taxes or tax them differently than the industry sector. Change the carbon tax lever so it is set separately for the district heat and hydrogen supply sector, rather than inheriting the industry sector's carbon tax rate. | priority | allow carbon tax rate to be set separately for district heat hydrogen sector instead of inheriting rate from industry sector the district heat and hydrogen sector uses the same fuels as the industry sector when the carbon tax lever was designed it was assumed the tax was levied upstream on these fuels and so the rate seen by industrial facilities and district heat hydrogen facilities would be the same but some countries consider district heat at least maybe also hydrogen supply to be part of the energy sector and may exempt them from carbon taxes or tax them differently than the industry sector change the carbon tax lever so it is set separately for the district heat and hydrogen supply sector rather than inheriting the industry sector s carbon tax rate | 1 |
280,066 | 8,677,745,211 | IssuesEvent | 2018-11-30 17:40:38 | Killeroo/PowerPing | https://api.github.com/repos/Killeroo/PowerPing | closed | Fix Control C event handler | Bug Medium Priority | Event C handler currently not working properly, works with delay/lag
Responsiveness tested in:
- _Sending Normally_ (requires more testing)
- Graph
- _Flood_ (requires more testing)
- Scan | 1.0 | Fix Control C event handler - Event C handler currently not working properly, works with delay/lag
Responsiveness tested in:
- _Sending Normally_ (requires more testing)
- Graph
- _Flood_ (requires more testing)
- Scan | priority | fix control c event handler event c handler currently not working properly works with delay lag responsiveness tested in sending normally requires more testing graph flood requires more testing scan | 1 |
416,585 | 12,148,927,724 | IssuesEvent | 2020-04-24 15:20:09 | scality/metalk8s | https://api.github.com/repos/scality/metalk8s | opened | Add solution get in python CLI | complexity:easy complexity:medium kind:enhancement priority:low topic:solutions | **Component**:
'salt', 'cli', 'solution'
**Why this is needed**:
Centralized CLI
**What should be done**:
MetalK8s solution get command should be part of `metalk8sctl` python CLI
```
metalk8sctl solution get <name>
```
This command will just show information about an available solution.
Name argument is not mandatory, if no name then all available and activated solutions are listed.
NOTE: We may want to handle regexp matching (not mandatory for first iteration)
**Implementation proposal** (strongly recommended):
This command will just retrieve all solutions information according to the CLI arguments.
Check design documentation about Centralized CLI for more detail. | 1.0 | Add solution get in python CLI - **Component**:
'salt', 'cli', 'solution'
**Why this is needed**:
Centralized CLI
**What should be done**:
MetalK8s solution get command should be part of `metalk8sctl` python CLI
```
metalk8sctl solution get <name>
```
This command will just show information about an available solution.
Name argument is not mandatory, if no name then all available and activated solutions are listed.
NOTE: We may want to handle regexp matching (not mandatory for first iteration)
**Implementation proposal** (strongly recommended):
This command will just retrieve all solutions information according to the CLI arguments.
Check design documentation about Centralized CLI for more detail. | priority | add solution get in python cli component salt cli solution why this is needed centralized cli what should be done solution get command should be part of python cli solution get this command will just show information about an available solution name argument is not mandatory if no name then all available and activated solutions are listed note we may want to handle regexp matching not mandatory for first iteration implementation proposal strongly recommended this command will just retrieve all solutions information according to the cli arguments check design documentation about centralized cli for more detail | 1 |
463,542 | 13,283,460,503 | IssuesEvent | 2020-08-24 03:17:51 | mandarineorg/mandarinets | https://api.github.com/repos/mandarineorg/mandarinets | closed | Add Support for Middleware targetting methods instead of controllers | Medium Priority | **[Context]**
Right now, in order to use Mandarine's middleware, you need to create a `@Middleware` component targetting a RegExp of a route. What we want to do now is to add a feature for the developer to add middleware to the method. This can be located under the options of the Http method ex: `@GET(route, options)` where options would have a middleware field, or we could also create a decorator `UseMiddleware`.
| 1.0 | Add Support for Middleware targetting methods instead of controllers - **[Context]**
Right now, in order to use Mandarine's middleware, you need to create a `@Middleware` component targetting a RegExp of a route. What we want to do now is to add a feature for the developer to add middleware to the method. This can be located under the options of the Http method ex: `@GET(route, options)` where options would have a middleware field, or we could also create a decorator `UseMiddleware`.
| priority | add support for middleware targetting methods instead of controllers right now in order to use mandarine s middleware you need to create a middleware component targetting a regexp of a route what we want to do now is to add a feature for the developer to add middleware to the method this can be located under the options of the http method ex get route options where options would have a middleware field or we could also create a decorator usemiddleware | 1 |
541,200 | 15,823,075,090 | IssuesEvent | 2021-04-05 23:48:22 | dtcenter/METplus | https://api.github.com/repos/dtcenter/METplus | reopened | Apply TCPairs, TCStat, CyclonePlotter, GridStat to EMC 2020 Data | component: use case configuration component: use case wrapper priority: medium requestor: NCAR requestor: NOAA/other type: enhancement type: task | *Replace italics below with details for this issue.*
## Describe the Task ##
*Learn how to use TCPairs, TCStat, CyclonePlotter, and GridStat in the context of cyclones, using existing sample data from GitHub*
### Time Estimate ###
*Estimate the amount of work required here.*
*3 calendar days of work spread out over December.*
*Future work will either extend this task or create another one*
### Sub-Issues ###
Consider breaking the task down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
* John O will provide more data probably in January.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required
- [x] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [ ] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Task Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | Apply TCPairs, TCStat, CyclonePlotter, GridStat to EMC 2020 Data - *Replace italics below with details for this issue.*
## Describe the Task ##
*Learn how to use TCPairs, TCStat, CyclonePlotter, and GridStat in the context of cyclones, using existing sample data from GitHub*
### Time Estimate ###
*Estimate the amount of work required here.*
*3 calendar days of work spread out over December.*
*Future work will either extend this task or create another one*
### Sub-Issues ###
Consider breaking the task down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
* John O will provide more data probably in January.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required
- [x] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Review **projects** and select relevant **Repository** and **Organization** ones or add "alert:NEED PROJECT ASSIGNMENT" label
- [ ] Select **milestone** to next major version milestone or "Future Versions"
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Task Checklist ##
See the [METplus Workflow](https://dtcenter.github.io/METplus/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)**, **Project(s)**, **Milestone**, and **Linked issues**
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | apply tcpairs tcstat cycloneplotter gridstat to emc data replace italics below with details for this issue describe the task learn how to use tcpairs tcstat cycloneplotter and gridstat in the context of cyclones using existing sample data from github time estimate estimate the amount of work required here calendar days of work spread out over december future work will either extend this task or create another one sub issues consider breaking the task down into sub issues add a checkbox for each sub issue here relevant deadlines john o will provide more data probably in january funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone review projects and select relevant repository and organization ones or add alert need project assignment label select milestone to next major version milestone or future versions define related issue s consider the impact to the other metplus components task checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s project s milestone and linked issues iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
33,380 | 2,764,522,937 | IssuesEvent | 2015-04-29 15:50:26 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | closed | Ability to Publish Dataverse when Trying to Publish Dataset (if user has permission to do so) | Priority: Medium Status: QA Type: Feature | ---
Author Name: **Elizabeth Quigley** (@eaquigley)
Original Redmine Issue: 4065, https://redmine.hmdc.harvard.edu/issues/4065
Original Date: 2014-06-03
---
Need to add in functionality so that a user can publish a dataverse and a dataset at the same time if they are trying to publish a dataset in an unpublished dataverse. This will relate to permissions as the system will be able to identify that a user has access to publish the dataverse. If not, then we need messaging that states why they can't publish the dataset.
| 1.0 | Ability to Publish Dataverse when Trying to Publish Dataset (if user has permission to do so) - ---
Author Name: **Elizabeth Quigley** (@eaquigley)
Original Redmine Issue: 4065, https://redmine.hmdc.harvard.edu/issues/4065
Original Date: 2014-06-03
---
Need to add in functionality so that a user can publish a dataverse and a dataset at the same time if they are trying to publish a dataset in an unpublished dataverse. This will relate to permissions as the system will be able to identify that a user has access to publish the dataverse. If not, then we need messaging that states why they can't publish the dataset.
| priority | ability to publish dataverse when trying to publish dataset if user has permission to do so author name elizabeth quigley eaquigley original redmine issue original date need to add in functionality so that a user can publish a dataverse and a dataset at the same time if they are trying to publish a dataset in an unpublished dataverse this will relate to permissions as the system will be able to identify that a user has access to publish the dataverse if not then we need messaging that states why they can t publish the dataset | 1 |
423,766 | 12,301,890,926 | IssuesEvent | 2020-05-11 16:05:04 | wri/gfw-mapbuilder | https://api.github.com/repos/wri/gfw-mapbuilder | closed | Loading state for analysis tab | 4.x Upgrade medium priority | Clicking on poly on the map and going to analysis tab should have a better loading state. Maybe a spinner would be good? Right now it is just saying "geostore is registering".
See the spinner which shows when we click "Run analsysis" for the vega chart. | 1.0 | Loading state for analysis tab - Clicking on poly on the map and going to analysis tab should have a better loading state. Maybe a spinner would be good? Right now it is just saying "geostore is registering".
See the spinner which shows when we click "Run analsysis" for the vega chart. | priority | loading state for analysis tab clicking on poly on the map and going to analysis tab should have a better loading state maybe a spinner would be good right now it is just saying geostore is registering see the spinner which shows when we click run analsysis for the vega chart | 1 |
818,328 | 30,683,813,084 | IssuesEvent | 2023-07-26 10:56:03 | netdata/netdata-cloud | https://api.github.com/repos/netdata/netdata-cloud | closed | [Bug]: overview search bar looks misalligned | bug priority/medium cloud-frontend | ### Bug description

### Expected behavior
like it was before
### Steps to reproduce
1. go [here](https://app.netdata.cloud/spaces/netdata-demo/rooms/all-nodes/overview#metrics_correlation=false&after=-900&before=0&utc=Europe%2FLondon&offset=%2B1&timezoneName=Edinburgh%2C%20London&modal=&modalTab=&d8a4e0c5-7c79-4145-900e-83a9f06fcb6a--chartName=menu_system)
### Screenshots
_No response_
### Error Logs
_No response_
### Desktop
OS: [e.g. iOS]
Browser [e.g. chrome, safari]
Browser Version [e.g. 22]
### Additional context
_No response_ | 1.0 | [Bug]: overview search bar looks misalligned - ### Bug description

### Expected behavior
like it was before
### Steps to reproduce
1. go [here](https://app.netdata.cloud/spaces/netdata-demo/rooms/all-nodes/overview#metrics_correlation=false&after=-900&before=0&utc=Europe%2FLondon&offset=%2B1&timezoneName=Edinburgh%2C%20London&modal=&modalTab=&d8a4e0c5-7c79-4145-900e-83a9f06fcb6a--chartName=menu_system)
### Screenshots
_No response_
### Error Logs
_No response_
### Desktop
OS: [e.g. iOS]
Browser [e.g. chrome, safari]
Browser Version [e.g. 22]
### Additional context
_No response_ | priority | overview search bar looks misalligned bug description expected behavior like it was before steps to reproduce go screenshots no response error logs no response desktop os browser browser version additional context no response | 1 |
346,966 | 10,422,231,303 | IssuesEvent | 2019-09-16 08:32:26 | zdnscloud/singlecloud | https://api.github.com/repos/zdnscloud/singlecloud | opened | lvm存储被删除后, 依然会出现在storageclass列表里 | bug priority: Medium | 分支: add-cluster-scversion
1. 创建一个名称为lvmstorage的lvm存储,
2. 删除这个存储,
3. 创建一个名称为lvm的lvm存储,
4. 删除这个存储,
结果:
get storageclass 时, 被删除的lvm和lvmstorage会出现在列表里.



| 1.0 | lvm存储被删除后, 依然会出现在storageclass列表里 - 分支: add-cluster-scversion
1. 创建一个名称为lvmstorage的lvm存储,
2. 删除这个存储,
3. 创建一个名称为lvm的lvm存储,
4. 删除这个存储,
结果:
get storageclass 时, 被删除的lvm和lvmstorage会出现在列表里.



| priority | lvm存储被删除后 依然会出现在storageclass列表里 分支 add cluster scversion 创建一个名称为lvmstorage的lvm存储 删除这个存储 创建一个名称为lvm的lvm存储 删除这个存储 结果 get storageclass 时 被删除的lvm和lvmstorage会出现在列表里 | 1 |
46,456 | 2,957,660,074 | IssuesEvent | 2015-07-08 17:27:05 | chef/chef | https://api.github.com/repos/chef/chef | closed | MacOSX Service provider - Running services are not restarted | Bug Chef Core Medium Priority | ### Version:
12.2.1
### Environment:
MacOSX
Any cookbook that uses the provider service/macosx.rb
source: https://github.com/chef/chef/blob/master/lib/chef/provider/service/macosx.rb
### Scenario:
Restart services when they are running or stopped.
It just restart the service when it is stopped. However when it is running just stop it.
### Steps to Reproduce:
Use the provider service in some cookbook and run it in a MacOSX.
Example:
service "SOME-SERVICE" do
action :restart
end
### Expected Result:
Expect to the service be restarted.
### Actual Result:
When the service is running, it is stopped but not started. Debug message:
[2015-04-01T14:41:25+01:00] DEBUG: service[thoughtworks.go-agent-1] already running, not starting
[2015-04-01T14:41:25+01:00] INFO: service[thoughtworks.go-agent-1] restarted
When the service is stopped, it is started correctly.
### Problem in source:
When start_service and stop_service methods are called it is checked if the attribute 'running' is 'true' or 'false' respectively. However this attribute is defined at the beginning of the service instance definition and not updated if it the service is stopped or started.
When you restart the service when it is running, the flag is defined as "true" and the stop_service stops the service. However the start_service doesn't do nothing because the 'running' attribute is still 'true' (not updated).
### Proposed solution:
Remove in the start_service method (line 81 source code) the conditional:
if @current_resource.running
Chef::Log.debug("#{@new_resource} already running, not starting")
And to be consistent remove also in the stop_service method (line 93 source code) the conditional:
unless @current_resource.running
Chef::Log.debug("#{@new_resource} not running, not stopping")
| 1.0 | MacOSX Service provider - Running services are not restarted - ### Version:
12.2.1
### Environment:
MacOSX
Any cookbook that uses the provider service/macosx.rb
source: https://github.com/chef/chef/blob/master/lib/chef/provider/service/macosx.rb
### Scenario:
Restart services when they are running or stopped.
It just restart the service when it is stopped. However when it is running just stop it.
### Steps to Reproduce:
Use the provider service in some cookbook and run it in a MacOSX.
Example:
service "SOME-SERVICE" do
action :restart
end
### Expected Result:
Expect to the service be restarted.
### Actual Result:
When the service is running, it is stopped but not started. Debug message:
[2015-04-01T14:41:25+01:00] DEBUG: service[thoughtworks.go-agent-1] already running, not starting
[2015-04-01T14:41:25+01:00] INFO: service[thoughtworks.go-agent-1] restarted
When the service is stopped, it is started correctly.
### Problem in source:
When start_service and stop_service methods are called it is checked if the attribute 'running' is 'true' or 'false' respectively. However this attribute is defined at the beginning of the service instance definition and not updated if it the service is stopped or started.
When you restart the service when it is running, the flag is defined as "true" and the stop_service stops the service. However the start_service doesn't do nothing because the 'running' attribute is still 'true' (not updated).
### Proposed solution:
Remove in the start_service method (line 81 source code) the conditional:
if @current_resource.running
Chef::Log.debug("#{@new_resource} already running, not starting")
And to be consistent remove also in the stop_service method (line 93 source code) the conditional:
unless @current_resource.running
Chef::Log.debug("#{@new_resource} not running, not stopping")
| priority | macosx service provider running services are not restarted version environment macosx any cookbook that uses the provider service macosx rb source scenario restart services when they are running or stopped it just restart the service when it is stopped however when it is running just stop it steps to reproduce use the provider service in some cookbook and run it in a macosx example service some service do action restart end expected result expect to the service be restarted actual result when the service is running it is stopped but not started debug message debug service already running not starting info service restarted when the service is stopped it is started correctly problem in source when start service and stop service methods are called it is checked if the attribute running is true or false respectively however this attribute is defined at the beginning of the service instance definition and not updated if it the service is stopped or started when you restart the service when it is running the flag is defined as true and the stop service stops the service however the start service doesn t do nothing because the running attribute is still true not updated proposed solution remove in the start service method line source code the conditional if current resource running chef log debug new resource already running not starting and to be consistent remove also in the stop service method line source code the conditional unless current resource running chef log debug new resource not running not stopping | 1 |
502,031 | 14,538,545,208 | IssuesEvent | 2020-12-15 10:37:03 | bdecon/econ_data | https://api.github.com/repos/bdecon/econ_data | opened | bd CPS: New variable DISCTYPE | new variable priority: medium | The BLS definition for discouraged workers covers only those who want a job, looked for one in the last 12 months, are available to take one, but don't think there is one available for them. In the NILFREASON variable I've defined discouraged workers as anyone who wants a job but isn't in the labor force.
To make the two match, I want to create a new variable called DISCTYPE that is defined as:
`Discouraged` where `PRDISC == 1`, `Marginally attached` where `PRDISC == 2`, `No recent search` where `PRJOBSEA == 5`, and `Unavailable` where `PRDISC == 3`. It will be defined only for those with `NILFREASON == "Discouraged"`.
I can then remove some of the other variables, MRGNATT and PRDISC, for example. | 1.0 | bd CPS: New variable DISCTYPE - The BLS definition for discouraged workers covers only those who want a job, looked for one in the last 12 months, are available to take one, but don't think there is one available for them. In the NILFREASON variable I've defined discouraged workers as anyone who wants a job but isn't in the labor force.
To make the two match, I want to create a new variable called DISCTYPE that is defined as:
`Discouraged` where `PRDISC == 1`, `Marginally attached` where `PRDISC == 2`, `No recent search` where `PRJOBSEA == 5`, and `Unavailable` where `PRDISC == 3`. It will be defined only for those with `NILFREASON == "Discouraged"`.
I can then remove some of the other variables, MRGNATT and PRDISC, for example. | priority | bd cps new variable disctype the bls definition for discouraged workers covers only those who want a job looked for one in the last months are available to take one but don t think there is one available for them in the nilfreason variable i ve defined discouraged workers as anyone who wants a job but isn t in the labor force to make the two match i want to create a new variable called disctype that is defined as discouraged where prdisc marginally attached where prdisc no recent search where prjobsea and unavailable where prdisc it will be defined only for those with nilfreason discouraged i can then remove some of the other variables mrgnatt and prdisc for example | 1 |
251,724 | 8,026,010,231 | IssuesEvent | 2018-07-27 01:13:40 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | USER ISSUE: To many Skillpoints | Medium Priority | **Version:** 0.7.2.3 beta
**Steps to Reproduce:**
I got 11.914 Skillpoint in ine day... i didn't change the server options...
**Expected behavior:**
I should have about 30 skillpoints....
**Actual behavior:**
| 1.0 | USER ISSUE: To many Skillpoints - **Version:** 0.7.2.3 beta
**Steps to Reproduce:**
I got 11.914 Skillpoint in ine day... i didn't change the server options...
**Expected behavior:**
I should have about 30 skillpoints....
**Actual behavior:**
| priority | user issue to many skillpoints version beta steps to reproduce i got skillpoint in ine day i didn t change the server options expected behavior i should have about skillpoints actual behavior | 1 |
711,788 | 24,475,399,382 | IssuesEvent | 2022-10-08 05:06:21 | roq-trading/roq-issues | https://api.github.com/repos/roq-trading/roq-issues | closed | [roq-server] Now using --cache_dir instead of --auth_cache_dir | chore medium priority | Authentication tokens will be written to `<auth_cache_dir>/<name>/auth/`.
This change was done to align with changes needed to cache configuration history, see
* #287 | 1.0 | [roq-server] Now using --cache_dir instead of --auth_cache_dir - Authentication tokens will be written to `<auth_cache_dir>/<name>/auth/`.
This change was done to align with changes needed to cache configuration history, see
* #287 | priority | now using cache dir instead of auth cache dir authentication tokens will be written to auth this change was done to align with changes needed to cache configuration history see | 1 |
167,215 | 6,334,526,072 | IssuesEvent | 2017-07-26 16:50:59 | ACEmulator/ACE | https://api.github.com/repos/ACEmulator/ACE | closed | Background maintenance thread | priority:medium | We need some sort of low priority thread to run and periodically perform routine maintenance tasks such as removing characters marked for deletion. | 1.0 | Background maintenance thread - We need some sort of low priority thread to run and periodically perform routine maintenance tasks such as removing characters marked for deletion. | priority | background maintenance thread we need some sort of low priority thread to run and periodically perform routine maintenance tasks such as removing characters marked for deletion | 1 |
722,263 | 24,856,607,169 | IssuesEvent | 2022-10-27 03:15:15 | JasonBock/Rocks | https://api.github.com/repos/JasonBock/Rocks | opened | Projected Types With Open Generics Are Not Handled Correctly | bug Medium Priority | To reproduce:
```csharp
using Rocks;
using System;
public interface IPixel { }
public interface IPixel<TSelf> : IPixel, IEquatable<TSelf>
where TSelf : unmanaged, IPixel<TSelf>
{ }
public interface IUseSpanWithOpenGeneric
{
void From<TSourcePixel>(
ReadOnlySpan<TSourcePixel> sourcePixels)
where TSourcePixel : unmanaged, IPixel<TSourcePixel>;
}
public static class Test
{
public static void Go()
{
var expectations = Rock.Create<IUseSpanWithOpenGeneric>();
}
}
```
This will lead to the following errors:
```
```
This was found on ImageSharp's `PixelOperation` type.
| 1.0 | Projected Types With Open Generics Are Not Handled Correctly - To reproduce:
```csharp
using Rocks;
using System;
public interface IPixel { }
public interface IPixel<TSelf> : IPixel, IEquatable<TSelf>
where TSelf : unmanaged, IPixel<TSelf>
{ }
public interface IUseSpanWithOpenGeneric
{
void From<TSourcePixel>(
ReadOnlySpan<TSourcePixel> sourcePixels)
where TSourcePixel : unmanaged, IPixel<TSourcePixel>;
}
public static class Test
{
public static void Go()
{
var expectations = Rock.Create<IUseSpanWithOpenGeneric>();
}
}
```
This will lead to the following errors:
```
```
This was found on ImageSharp's `PixelOperation` type.
| priority | projected types with open generics are not handled correctly to reproduce csharp using rocks using system public interface ipixel public interface ipixel ipixel iequatable where tself unmanaged ipixel public interface iusespanwithopengeneric void from readonlyspan sourcepixels where tsourcepixel unmanaged ipixel public static class test public static void go var expectations rock create this will lead to the following errors this was found on imagesharp s pixeloperation type | 1 |
674,516 | 23,054,367,233 | IssuesEvent | 2022-07-25 02:05:43 | FEeasy404/GameUs | https://api.github.com/repos/FEeasy404/GameUs | closed | 게시글 작성 페이지 및 게시물 업로드 구현 | ✨Feat 🖐Priority: Medium | ## 추가 기능 설명
3.1.11 게시글 작성 페이지
- 게시글을 작성할 수 있는 페이지로, 하단 메뉴바에서 게시글 작성 을 클릭하면 표시됩니다.
- 글이 입력되거나 사진이 업로드 되면 업로드 버튼이 활성화되고, 버튼을 누르면 게시글이 업로드됩니다.
## 할 일
- [x] 게시글 작성 페이지 UI
- [x] 하단 메뉴바 클릭시 게시글 작성 페이지 뜨도록 구현
- [x] 글 입력시 업로드 버튼 활성화 구현
- [x] 버튼 클릭시 게시글 업로드 구현
## ETC
| 1.0 | 게시글 작성 페이지 및 게시물 업로드 구현 - ## 추가 기능 설명
3.1.11 게시글 작성 페이지
- 게시글을 작성할 수 있는 페이지로, 하단 메뉴바에서 게시글 작성 을 클릭하면 표시됩니다.
- 글이 입력되거나 사진이 업로드 되면 업로드 버튼이 활성화되고, 버튼을 누르면 게시글이 업로드됩니다.
## 할 일
- [x] 게시글 작성 페이지 UI
- [x] 하단 메뉴바 클릭시 게시글 작성 페이지 뜨도록 구현
- [x] 글 입력시 업로드 버튼 활성화 구현
- [x] 버튼 클릭시 게시글 업로드 구현
## ETC
| priority | 게시글 작성 페이지 및 게시물 업로드 구현 추가 기능 설명 게시글 작성 페이지 게시글을 작성할 수 있는 페이지로 하단 메뉴바에서 게시글 작성 을 클릭하면 표시됩니다 글이 입력되거나 사진이 업로드 되면 업로드 버튼이 활성화되고 버튼을 누르면 게시글이 업로드됩니다 할 일 게시글 작성 페이지 ui 하단 메뉴바 클릭시 게시글 작성 페이지 뜨도록 구현 글 입력시 업로드 버튼 활성화 구현 버튼 클릭시 게시글 업로드 구현 etc | 1 |
802,699 | 29,044,455,652 | IssuesEvent | 2023-05-13 11:25:15 | darktable-org/darktable | https://api.github.com/repos/darktable-org/darktable | closed | Framing (borders.c): can't enter custom ratio in dropdown | priority: medium reproduce: confirmed scope: UI bug: pending release notes: pending | ### Describe the bug
The tooltip for the Aspect dropdown (not the Aspect ratio slider) says one can enter a custom ratio in the form w:h. This is not actually possible. While one *can* enter a text string instead of selecting a predefined entry, the text is not actually evaluated - it is displayed for the dropdown but the aspect ratio slider is not updated.
### Steps to reproduce
1. Open the Framing module on an image
2. Click (or right-click) on "aspect"
3. Type a value - decimal number, w:h, or w/h
4. See the typed text displayed when the dropdown collapses
5. Note that the aspect ratio slider has not changed
### Expected behavior
The behavior should match the tooltip, or custom text should be disabled for the dropdown and the tooltip for the aspect ratio slider updated. A one-line change in common/calculator.c would permit the bauhaus to accept "7:6" as well as "7/6" for ratios.
### Logfile | Screenshot | Screencast
_No response_
### Commit
_No response_
### Where did you install darktable from?
self compiled
### darktable version
master 4.3.0+2056
### What OS are you using?
Linux
### What is the version of your OS?
Mageia 8.1
### Describe your system?
_No response_
### Are you using OpenCL GPU in darktable?
None
### If yes, what is the GPU card and driver?
_No response_
### Please provide additional context if applicable. You can attach files too, but might need to rename to .txt or .zip
_No response_ | 1.0 | Framing (borders.c): can't enter custom ratio in dropdown - ### Describe the bug
The tooltip for the Aspect dropdown (not the Aspect ratio slider) says one can enter a custom ratio in the form w:h. This is not actually possible. While one *can* enter a text string instead of selecting a predefined entry, the text is not actually evaluated - it is displayed for the dropdown but the aspect ratio slider is not updated.
### Steps to reproduce
1. Open the Framing module on an image
2. Click (or right-click) on "aspect"
3. Type a value - decimal number, w:h, or w/h
4. See the typed text displayed when the dropdown collapses
5. Note that the aspect ratio slider has not changed
### Expected behavior
The behavior should match the tooltip, or custom text should be disabled for the dropdown and the tooltip for the aspect ratio slider updated. A one-line change in common/calculator.c would permit the bauhaus to accept "7:6" as well as "7/6" for ratios.
### Logfile | Screenshot | Screencast
_No response_
### Commit
_No response_
### Where did you install darktable from?
self compiled
### darktable version
master 4.3.0+2056
### What OS are you using?
Linux
### What is the version of your OS?
Mageia 8.1
### Describe your system?
_No response_
### Are you using OpenCL GPU in darktable?
None
### If yes, what is the GPU card and driver?
_No response_
### Please provide additional context if applicable. You can attach files too, but might need to rename to .txt or .zip
_No response_ | priority | framing borders c can t enter custom ratio in dropdown describe the bug the tooltip for the aspect dropdown not the aspect ratio slider says one can enter a custom ratio in the form w h this is not actually possible while one can enter a text string instead of selecting a predefined entry the text is not actually evaluated it is displayed for the dropdown but the aspect ratio slider is not updated steps to reproduce open the framing module on an image click or right click on aspect type a value decimal number w h or w h see the typed text displayed when the dropdown collapses note that the aspect ratio slider has not changed expected behavior the behavior should match the tooltip or custom text should be disabled for the dropdown and the tooltip for the aspect ratio slider updated a one line change in common calculator c would permit the bauhaus to accept as well as for ratios logfile screenshot screencast no response commit no response where did you install darktable from self compiled darktable version master what os are you using linux what is the version of your os mageia describe your system no response are you using opencl gpu in darktable none if yes what is the gpu card and driver no response please provide additional context if applicable you can attach files too but might need to rename to txt or zip no response | 1 |
649,629 | 21,316,757,672 | IssuesEvent | 2022-04-16 12:16:26 | dita-ot/dita-ot | https://api.github.com/repos/dita-ot/dita-ot | closed | Extra error reported in console when converting topic to HTML5 with args.filter | bug priority/medium plugin/html5 stale | Publishing any DITA topic to HTML5 using an args.filter pointing to some ditaval file results in an extra error in the console output:
[xslt] I/O error reported by XML parser processing file:/${dita.input.valfile.url}: \${dita.input.valfile.url} (The system cannot find the file specified)
This seems to be the case because this ANT target:
<target name="html5.topic.init" unless="noMap">
which populates the "dita.input.valfile.url" param is executed only when DITA Maps are published, and not individual topics. | 1.0 | Extra error reported in console when converting topic to HTML5 with args.filter - Publishing any DITA topic to HTML5 using an args.filter pointing to some ditaval file results in an extra error in the console output:
[xslt] I/O error reported by XML parser processing file:/${dita.input.valfile.url}: \${dita.input.valfile.url} (The system cannot find the file specified)
This seems to be the case because this ANT target:
<target name="html5.topic.init" unless="noMap">
which populates the "dita.input.valfile.url" param is executed only when DITA Maps are published, and not individual topics. | priority | extra error reported in console when converting topic to with args filter publishing any dita topic to using an args filter pointing to some ditaval file results in an extra error in the console output i o error reported by xml parser processing file dita input valfile url dita input valfile url the system cannot find the file specified this seems to be the case because this ant target which populates the dita input valfile url param is executed only when dita maps are published and not individual topics | 1 |
624,024 | 19,684,775,856 | IssuesEvent | 2022-01-11 20:44:44 | GameFreedomGG/Sindragosa | https://api.github.com/repos/GameFreedomGG/Sindragosa | closed | [Profession][Item] Mining - Thorium Nodes & Arcane Crystals | Item Priority: Medium Professions Status: Confirmed Bug (Awaiting Development) | Decription: Player reported that drops for Thorium are messed up. Currently nearly all the time Arcane Crystal's drop, but Large Opals, Blue Sapphires etc should drop too.
How it works: Only Arcane Crystal's are dropping.
How it should work: Arcane 10% drop rate. Blue Sapphire 3%. Huge Emerald 3%. Large Opal 3%. Azerothian Diamond 3%.
Source (you should point out proofs of your report, please give us some source):
https://wotlk-twinhead.twinstar.cz/?object=175404 | 1.0 | [Profession][Item] Mining - Thorium Nodes & Arcane Crystals - Decription: Player reported that drops for Thorium are messed up. Currently nearly all the time Arcane Crystal's drop, but Large Opals, Blue Sapphires etc should drop too.
How it works: Only Arcane Crystal's are dropping.
How it should work: Arcane 10% drop rate. Blue Sapphire 3%. Huge Emerald 3%. Large Opal 3%. Azerothian Diamond 3%.
Source (you should point out proofs of your report, please give us some source):
https://wotlk-twinhead.twinstar.cz/?object=175404 | priority | mining thorium nodes arcane crystals decription player reported that drops for thorium are messed up currently nearly all the time arcane crystal s drop but large opals blue sapphires etc should drop too how it works only arcane crystal s are dropping how it should work arcane drop rate blue sapphire huge emerald large opal azerothian diamond source you should point out proofs of your report please give us some source | 1 |
798,636 | 28,291,784,807 | IssuesEvent | 2023-04-09 09:52:21 | KDT3-final-project-team2/backend | https://api.github.com/repos/KDT3-final-project-team2/backend | closed | [Fix] 기업회원 지원자 목록에 지원서id 추가 | For: API Priority: Medium Status: In Progress | ## 🔨개발 할 기능
지원자 목록 출력 DTO에 지원서 id 추가하기
## 🧩 세부 기능
- [ ] DTO에 applicationID 추가하기
## 📖 참고 사항
| 1.0 | [Fix] 기업회원 지원자 목록에 지원서id 추가 - ## 🔨개발 할 기능
지원자 목록 출력 DTO에 지원서 id 추가하기
## 🧩 세부 기능
- [ ] DTO에 applicationID 추가하기
## 📖 참고 사항
| priority | 기업회원 지원자 목록에 지원서id 추가 🔨개발 할 기능 지원자 목록 출력 dto에 지원서 id 추가하기 🧩 세부 기능 dto에 applicationid 추가하기 📖 참고 사항 | 1 |
708,063 | 24,328,922,804 | IssuesEvent | 2022-09-30 17:23:51 | kleros/kleros-v2 | https://api.github.com/repos/kleros/kleros-v2 | closed | Minor bug: after changing a court's minStake, jurors with a smaller stake are still drawn | Priority: Medium Type: Bug :bug: Package: Contracts | The issue also exists in v1. | 1.0 | Minor bug: after changing a court's minStake, jurors with a smaller stake are still drawn - The issue also exists in v1. | priority | minor bug after changing a court s minstake jurors with a smaller stake are still drawn the issue also exists in | 1 |
190,983 | 6,824,470,437 | IssuesEvent | 2017-11-08 06:23:41 | certificate-helper/TLS-Inspector | https://api.github.com/repos/certificate-helper/TLS-Inspector | closed | App URI protocol broken | bug medium priority merged | **Affected Version:**
Current
**Is this a Test Flight version or the App Store version?**
App Store
**Device and iOS Version:**
All
**What steps will reproduce the problem?**
1. Use test app to open `certinspector://inspect/google.com`
**What is the expected output?**
Google.com shows
**What do you see instead?**
Nothing
**Please provide any additional information below.**
Should remove feature. | 1.0 | App URI protocol broken - **Affected Version:**
Current
**Is this a Test Flight version or the App Store version?**
App Store
**Device and iOS Version:**
All
**What steps will reproduce the problem?**
1. Use test app to open `certinspector://inspect/google.com`
**What is the expected output?**
Google.com shows
**What do you see instead?**
Nothing
**Please provide any additional information below.**
Should remove feature. | priority | app uri protocol broken affected version current is this a test flight version or the app store version app store device and ios version all what steps will reproduce the problem use test app to open certinspector inspect google com what is the expected output google com shows what do you see instead nothing please provide any additional information below should remove feature | 1 |
548,712 | 16,074,198,331 | IssuesEvent | 2021-04-25 02:55:58 | rich-iannone/pointblank | https://api.github.com/repos/rich-iannone/pointblank | opened | Provide styled console output when using `yaml_exec()` | Difficulty: [2] Intermediate Effort: [2] Medium Priority: [3] High Type: ★ Enhancement | When using `yaml_exec()` to process YAML agents and informants *en masse*, it would be nice to be notified in the console about what happened (during interactive sessions). We can use {cli}-formatted messages like elsewhere in the package.
Also, the function should invisibly return *something* about what was written. Right now, it always returns `NULL`.
| 1.0 | Provide styled console output when using `yaml_exec()` - When using `yaml_exec()` to process YAML agents and informants *en masse*, it would be nice to be notified in the console about what happened (during interactive sessions). We can use {cli}-formatted messages like elsewhere in the package.
Also, the function should invisibly return *something* about what was written. Right now, it always returns `NULL`.
| priority | provide styled console output when using yaml exec when using yaml exec to process yaml agents and informants en masse it would be nice to be notified in the console about what happened during interactive sessions we can use cli formatted messages like elsewhere in the package also the function should invisibly return something about what was written right now it always returns null | 1 |
188,843 | 6,782,488,138 | IssuesEvent | 2017-10-30 08:20:55 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] Includes more files than it should | 1. Type: Bug Priority: Medium Status: Completed Time: ~1 hour | <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
This is my tsconfig:
```
{
"compilerOptions": {
"target": "es5",
"module": "commonjs",
"moduleResolution": "node",
"sourceMap": true,
"emitDecoratorMetadata": true,
"experimentalDecorators": true,
"lib": [ "es2015", "dom" ],
"rootDir": ".",
"outDir": "dist"
},
"include": [
"app/**/*.ts"
],
"exclude": [
"app/main.prod.ts"
]
}
```
But compodoc still includes files from dist/ (the outDir) and other directories in the root directory, which leads to all my modules, components, etc. being included two times in the docs.
The expected behavior is that compodoc only includes the files specified in "files" or "include" if either of those are specified. That's the way the typescript compiler handles it.
##### **Operating System, Node.js, npm, compodoc version(s)**
Ubuntu 17.04, Node v8.5.0, npm v5.3.0, compodoc v1.0.1
##### **Angular configuration, a `package.json` file in the root folder**
Here is my project: https://github.com/tradity/tradity-client/
##### **Compodoc installed globally or locally ?**
compodoc is installed locally | 1.0 | [BUG] Includes more files than it should - <!--
> Please follow the issue template below for bug reports and queries.
> For issue, start the label of the title with [BUG]
> For feature requests, start the label of the title with [FEATURE] and explain your use case and ideas clearly below, you can remove sections which are not relevant.
-->
##### **Overview of the issue**
This is my tsconfig:
```
{
"compilerOptions": {
"target": "es5",
"module": "commonjs",
"moduleResolution": "node",
"sourceMap": true,
"emitDecoratorMetadata": true,
"experimentalDecorators": true,
"lib": [ "es2015", "dom" ],
"rootDir": ".",
"outDir": "dist"
},
"include": [
"app/**/*.ts"
],
"exclude": [
"app/main.prod.ts"
]
}
```
But compodoc still includes files from dist/ (the outDir) and other directories in the root directory, which leads to all my modules, components, etc. being included two times in the docs.
The expected behavior is that compodoc only includes the files specified in "files" or "include" if either of those are specified. That's the way the typescript compiler handles it.
##### **Operating System, Node.js, npm, compodoc version(s)**
Ubuntu 17.04, Node v8.5.0, npm v5.3.0, compodoc v1.0.1
##### **Angular configuration, a `package.json` file in the root folder**
Here is my project: https://github.com/tradity/tradity-client/
##### **Compodoc installed globally or locally ?**
compodoc is installed locally | priority | includes more files than it should please follow the issue template below for bug reports and queries for issue start the label of the title with for feature requests start the label of the title with and explain your use case and ideas clearly below you can remove sections which are not relevant overview of the issue this is my tsconfig compileroptions target module commonjs moduleresolution node sourcemap true emitdecoratormetadata true experimentaldecorators true lib rootdir outdir dist include app ts exclude app main prod ts but compodoc still includes files from dist the outdir and other directories in the root directory which leads to all my modules components etc being included two times in the docs the expected behavior is that compodoc only includes the files specified in files or include if either of those are specified that s the way the typescript compiler handles it operating system node js npm compodoc version s ubuntu node npm compodoc angular configuration a package json file in the root folder here is my project compodoc installed globally or locally compodoc is installed locally | 1 |
4,047 | 2,544,727,167 | IssuesEvent | 2015-01-29 12:25:13 | pychess/pychess | https://api.github.com/repos/pychess/pychess | closed | Chessrules | Component-Docs imported Milestone-Release1.0 Priority-Medium Type-Task Usability | _From [lobais](https://code.google.com/u/lobais/) on September 06, 2006 16:42:39_
We should have a help item.
Other than simple uihelp, it should also contain the rules and a small
toturial. http://en.wikipedia.org/wiki/Chess has a nice table of the most simple
answer to "what is chess"
_Original issue: http://code.google.com/p/pychess/issues/detail?id=37_ | 1.0 | Chessrules - _From [lobais](https://code.google.com/u/lobais/) on September 06, 2006 16:42:39_
We should have a help item.
Other than simple uihelp, it should also contain the rules and a small
toturial. http://en.wikipedia.org/wiki/Chess has a nice table of the most simple
answer to "what is chess"
_Original issue: http://code.google.com/p/pychess/issues/detail?id=37_ | priority | chessrules from on september we should have a help item other than simple uihelp it should also contain the rules and a small toturial has a nice table of the most simple answer to what is chess original issue | 1 |
236,174 | 7,747,213,742 | IssuesEvent | 2018-05-30 01:57:13 | medic/medic-webapp | https://api.github.com/repos/medic/medic-webapp | closed | Disabled transitions still run | Priority: 2 - Medium Status: 1 - Triaged Type: Bug | **Steps to reproduce**:
- disable the `default_response` transition
- submit an unstructured SMS
**What should happen**:
- the `default_responses` transition should not run on the doc
**What actually happens**:
- the `default_responses` transition does run on the doc
See example [here](https://standard-release.dev.medicmobile.org/medic/410ca5bf6c30c2cfa17f4e8e0ad8ab5c): `"transitions":{"default_responses":{"last_rev":2,"seq":674080,"ok":true}}`
```
2018-05-29T20:29:17.063Z - info: info: Reloading configuration
2018-05-29T20:29:17.852Z - info: info: Loading transition "maintain_info_document"
2018-05-29T20:29:17.854Z - info: info: Loading transition "update_clinics"
2018-05-29T20:29:17.855Z - info: info: Loading transition "registration"
2018-05-29T20:29:17.861Z - info: info: Loading transition "accept_patient_reports"
2018-05-29T20:29:17.861Z - info: info: Loading transition "generate_patient_id_on_people"
2018-05-29T20:29:17.862Z - info: warn: Disabled transition "default_responses"
2018-05-29T20:29:17.862Z - info: warn: Disabled transition "update_sent_by"
2018-05-29T20:29:17.863Z - info: warn: Disabled transition "update_sent_forms"
2018-05-29T20:29:17.863Z - info: warn: Disabled transition "conditional_alerts"
2018-05-29T20:29:17.863Z - info: warn: Disabled transition "multi_report_alerts"
2018-05-29T20:29:17.864Z - info: info: Loading transition "update_notifications"
2018-05-29T20:29:17.864Z - info: warn: Disabled transition "update_scheduled_reports"
2018-05-29T20:29:17.865Z - info: warn: Disabled transition "resolve_pending"
2018-05-29T20:29:45.394Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad88dad seq 674064
2018-05-29T20:29:49.822Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89af9 seq 674068
2018-05-29T20:29:53.152Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89d49 seq 674072
2018-05-29T20:30:21.240Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8aace seq 674076
2018-05-29T20:32:00.006Z - info: info: checking schedule again in 5 minutes
2018-05-29T20:32:08.386Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8ab5c seq 674080
```
This may help towards #4539.
Seen on `standard-release.dev` running `2.14.3-beta.2` | 1.0 | Disabled transitions still run - **Steps to reproduce**:
- disable the `default_response` transition
- submit an unstructured SMS
**What should happen**:
- the `default_responses` transition should not run on the doc
**What actually happens**:
- the `default_responses` transition does run on the doc
See example [here](https://standard-release.dev.medicmobile.org/medic/410ca5bf6c30c2cfa17f4e8e0ad8ab5c): `"transitions":{"default_responses":{"last_rev":2,"seq":674080,"ok":true}}`
```
2018-05-29T20:29:17.063Z - info: info: Reloading configuration
2018-05-29T20:29:17.852Z - info: info: Loading transition "maintain_info_document"
2018-05-29T20:29:17.854Z - info: info: Loading transition "update_clinics"
2018-05-29T20:29:17.855Z - info: info: Loading transition "registration"
2018-05-29T20:29:17.861Z - info: info: Loading transition "accept_patient_reports"
2018-05-29T20:29:17.861Z - info: info: Loading transition "generate_patient_id_on_people"
2018-05-29T20:29:17.862Z - info: warn: Disabled transition "default_responses"
2018-05-29T20:29:17.862Z - info: warn: Disabled transition "update_sent_by"
2018-05-29T20:29:17.863Z - info: warn: Disabled transition "update_sent_forms"
2018-05-29T20:29:17.863Z - info: warn: Disabled transition "conditional_alerts"
2018-05-29T20:29:17.863Z - info: warn: Disabled transition "multi_report_alerts"
2018-05-29T20:29:17.864Z - info: info: Loading transition "update_notifications"
2018-05-29T20:29:17.864Z - info: warn: Disabled transition "update_scheduled_reports"
2018-05-29T20:29:17.865Z - info: warn: Disabled transition "resolve_pending"
2018-05-29T20:29:45.394Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad88dad seq 674064
2018-05-29T20:29:49.822Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89af9 seq 674068
2018-05-29T20:29:53.152Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad89d49 seq 674072
2018-05-29T20:30:21.240Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8aace seq 674076
2018-05-29T20:32:00.006Z - info: info: checking schedule again in 5 minutes
2018-05-29T20:32:08.386Z - info: info: saved changes on doc 410ca5bf6c30c2cfa17f4e8e0ad8ab5c seq 674080
```
This may help towards #4539.
Seen on `standard-release.dev` running `2.14.3-beta.2` | priority | disabled transitions still run steps to reproduce disable the default response transition submit an unstructured sms what should happen the default responses transition should not run on the doc what actually happens the default responses transition does run on the doc see example transitions default responses last rev seq ok true info info reloading configuration info info loading transition maintain info document info info loading transition update clinics info info loading transition registration info info loading transition accept patient reports info info loading transition generate patient id on people info warn disabled transition default responses info warn disabled transition update sent by info warn disabled transition update sent forms info warn disabled transition conditional alerts info warn disabled transition multi report alerts info info loading transition update notifications info warn disabled transition update scheduled reports info warn disabled transition resolve pending info info saved changes on doc seq info info saved changes on doc seq info info saved changes on doc seq info info saved changes on doc seq info info checking schedule again in minutes info info saved changes on doc seq this may help towards seen on standard release dev running beta | 1 |
528,223 | 15,362,148,046 | IssuesEvent | 2021-03-01 19:05:08 | thehyve/delphyne | https://api.github.com/repos/thehyve/delphyne | opened | Cleanup old source_vocabulary_ids in STCM_version | STCM medium priority | If I change the vocabulary ids in `stcm_versions.tsv` (say the custom vocabularies I was referencing have been deleted and I want to use different ones), the old vocabularies are not automatically removed from the `SourceToConceptMapVersion` table in the database. Once the constraints are re-applied, I get a FK constraints error because the deleted vocabularies are still being referenced.
Solution: implement stcm version cleanup. | 1.0 | Cleanup old source_vocabulary_ids in STCM_version - If I change the vocabulary ids in `stcm_versions.tsv` (say the custom vocabularies I was referencing have been deleted and I want to use different ones), the old vocabularies are not automatically removed from the `SourceToConceptMapVersion` table in the database. Once the constraints are re-applied, I get a FK constraints error because the deleted vocabularies are still being referenced.
Solution: implement stcm version cleanup. | priority | cleanup old source vocabulary ids in stcm version if i change the vocabulary ids in stcm versions tsv say the custom vocabularies i was referencing have been deleted and i want to use different ones the old vocabularies are not automatically removed from the sourcetoconceptmapversion table in the database once the constraints are re applied i get a fk constraints error because the deleted vocabularies are still being referenced solution implement stcm version cleanup | 1 |
643,190 | 20,925,544,918 | IssuesEvent | 2022-03-24 22:19:50 | cds-snc/resources-ressources | https://api.github.com/repos/cds-snc/resources-ressources | closed | Prototype Version 1 | Medium Priority | Priorité moyenne | # Summary
[PR for prototype](https://github.com/cds-snc/resources-ressources/pull/6
)
First pass at a coded prototype, based on figma designs.
**Done when**:
The prototype can be visited through a link | 1.0 | Prototype Version 1 - # Summary
[PR for prototype](https://github.com/cds-snc/resources-ressources/pull/6
)
First pass at a coded prototype, based on figma designs.
**Done when**:
The prototype can be visited through a link | priority | prototype version summary first pass at a coded prototype based on figma designs done when the prototype can be visited through a link | 1 |
683,326 | 23,377,282,952 | IssuesEvent | 2022-08-11 05:29:55 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Cannot update versions in multi-line `replace` block in go.mod | type:bug priority-3-medium manager:gomod status:ready reproduction:confirmed | ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
31.81.3, also 32.26.2
### Please select which platform you are using if self-hosting.
github.com
### If you're self-hosting Renovate, tell us what version of the platform you run.
ubuntu-latest
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
renovate does not update versions in a multi-line `replace` block in a go.mod file. It only works for the single line replace.
I created an example repo [here](https://github.com/avorima/renovate-gomod-example).
Not updated: https://github.com/avorima/renovate-gomod-example/runs/5413565230?check_suite_focus=true
```
replace (
k8s.io/client-go => k8s.io/client-go v0.21.9
sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7
)
```
Updated: https://github.com/avorima/renovate-gomod-example/runs/5413625323?check_suite_focus=true
```
replace k8s.io/client-go => k8s.io/client-go v0.21.9
replace sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7
```
### Relevant debug logs
_No response_
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | 1.0 | Cannot update versions in multi-line `replace` block in go.mod - ### How are you running Renovate?
Self-hosted
### If you're self-hosting Renovate, tell us what version of Renovate you run.
31.81.3, also 32.26.2
### Please select which platform you are using if self-hosting.
github.com
### If you're self-hosting Renovate, tell us what version of the platform you run.
ubuntu-latest
### Was this something which used to work for you, and then stopped?
I never saw this working
### Describe the bug
renovate does not update versions in a multi-line `replace` block in a go.mod file. It only works for the single line replace.
I created an example repo [here](https://github.com/avorima/renovate-gomod-example).
Not updated: https://github.com/avorima/renovate-gomod-example/runs/5413565230?check_suite_focus=true
```
replace (
k8s.io/client-go => k8s.io/client-go v0.21.9
sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7
)
```
Updated: https://github.com/avorima/renovate-gomod-example/runs/5413625323?check_suite_focus=true
```
replace k8s.io/client-go => k8s.io/client-go v0.21.9
replace sigs.k8s.io/controller-runtime => sigs.k8s.io/controller-runtime v0.9.7
```
### Relevant debug logs
_No response_
### Have you created a minimal reproduction repository?
I have linked to a minimal reproduction repository in the bug description | priority | cannot update versions in multi line replace block in go mod how are you running renovate self hosted if you re self hosting renovate tell us what version of renovate you run also please select which platform you are using if self hosting github com if you re self hosting renovate tell us what version of the platform you run ubuntu latest was this something which used to work for you and then stopped i never saw this working describe the bug renovate does not update versions in a multi line replace block in a go mod file it only works for the single line replace i created an example repo not updated replace io client go io client go sigs io controller runtime sigs io controller runtime updated replace io client go io client go replace sigs io controller runtime sigs io controller runtime relevant debug logs no response have you created a minimal reproduction repository i have linked to a minimal reproduction repository in the bug description | 1 |
333,421 | 10,121,981,188 | IssuesEvent | 2019-07-31 16:49:55 | salesagility/SuiteCRM | https://api.github.com/repos/salesagility/SuiteCRM | closed | Cookie path is not respected if globally set | Fix Proposed Medium Priority Resolved: Next Release bug | <!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
<!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. --->
#### Issue
When logging in, the ```ck_login_id_20``` and ```ck_login_language_20``` cookies are set using the ```setCookie``` wrapper in SugarApplication.php, if the path has been changed using ```session.cookie_path``` then it is ignored and defaults to '/'
#### Expected Behavior
It should respect any settings made in PHP config and set to the correct value
#### Actual Behavior
The cookie path is set to '/' because of the function signature
#### Possible Fix
Change the sig from:
```php
public static function setCookie(
$name, $value, $expire = 0, $path = '/', $domain = null, $secure = false, $httponly = true
) {
```
To
```php
public static function setCookie(
$name, $value, $expire = 0, $path = null, $domain = null, $secure = false, $httponly = true
) {
```
#### Steps to Reproduce
1. Install suitecrm inside a folder (/alpha)
2. Add the following to .htaccess
```
php_value session.cookie_path /alpha
```
3. Login and check path for cookies
#### Context
I run multiple CRM's in differnet folders under a single domain (crm.domain/aplha, crm.domain/beta etc) and other cookies are set correctly, but these ones are not.
#### Your Environment
* SuiteCRM Version used: Version 7.11.1 / Sugar Version 6.5.25 (Build 344)
* Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Firefox
* Environment name and version (e.g. MySQL, PHP 7): PHP 7.2.14, MySQL 8
* Operating System and version (e.g Ubuntu 16.04): CentOS 7
| 1.0 | Cookie path is not respected if globally set - <!--- Provide a general summary of the issue in the **Title** above -->
<!--- Before you open an issue, please check if a similar issue already exists or has been closed before. --->
<!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. --->
#### Issue
When logging in, the ```ck_login_id_20``` and ```ck_login_language_20``` cookies are set using the ```setCookie``` wrapper in SugarApplication.php, if the path has been changed using ```session.cookie_path``` then it is ignored and defaults to '/'
#### Expected Behavior
It should respect any settings made in PHP config and set to the correct value
#### Actual Behavior
The cookie path is set to '/' because of the function signature
#### Possible Fix
Change the sig from:
```php
public static function setCookie(
$name, $value, $expire = 0, $path = '/', $domain = null, $secure = false, $httponly = true
) {
```
To
```php
public static function setCookie(
$name, $value, $expire = 0, $path = null, $domain = null, $secure = false, $httponly = true
) {
```
#### Steps to Reproduce
1. Install suitecrm inside a folder (/alpha)
2. Add the following to .htaccess
```
php_value session.cookie_path /alpha
```
3. Login and check path for cookies
#### Context
I run multiple CRM's in differnet folders under a single domain (crm.domain/aplha, crm.domain/beta etc) and other cookies are set correctly, but these ones are not.
#### Your Environment
* SuiteCRM Version used: Version 7.11.1 / Sugar Version 6.5.25 (Build 344)
* Browser name and version (e.g. Chrome Version 51.0.2704.63 (64-bit)): Firefox
* Environment name and version (e.g. MySQL, PHP 7): PHP 7.2.14, MySQL 8
* Operating System and version (e.g Ubuntu 16.04): CentOS 7
| priority | cookie path is not respected if globally set issue when logging in the ck login id and ck login language cookies are set using the setcookie wrapper in sugarapplication php if the path has been changed using session cookie path then it is ignored and defaults to expected behavior it should respect any settings made in php config and set to the correct value actual behavior the cookie path is set to because of the function signature possible fix change the sig from php public static function setcookie name value expire path domain null secure false httponly true to php public static function setcookie name value expire path null domain null secure false httponly true steps to reproduce install suitecrm inside a folder alpha add the following to htaccess php value session cookie path alpha login and check path for cookies context i run multiple crm s in differnet folders under a single domain crm domain aplha crm domain beta etc and other cookies are set correctly but these ones are not your environment suitecrm version used version sugar version build browser name and version e g chrome version bit firefox environment name and version e g mysql php php mysql operating system and version e g ubuntu centos | 1 |
534,913 | 15,651,867,434 | IssuesEvent | 2021-03-23 10:41:55 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | REST API - In POST method on the same endpoint is not working | bug priority: medium | **Describe the bug**
We can perform a GET on /wp-json/buddyboss/v1/topics as well as many other endpoints that respond with a GET request.
This validates that the authentication is working. But when we attempt to prefer a POST or Patch on the same endpoint it is NOT successful.
**To Reproduce**
When we attempt to prefer a POST or Patch on the same endpoint it is NOT successful.
https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503.
**Expected behavior**
when I attempt to prefer a POST or Patch on the same endpoint it is need work.
**Screenshots**
https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503.
**Support ticket links**
https://secure.helpscout.net/conversation/1453624113/130633?folderId=4265312
| 1.0 | REST API - In POST method on the same endpoint is not working - **Describe the bug**
We can perform a GET on /wp-json/buddyboss/v1/topics as well as many other endpoints that respond with a GET request.
This validates that the authentication is working. But when we attempt to prefer a POST or Patch on the same endpoint it is NOT successful.
**To Reproduce**
When we attempt to prefer a POST or Patch on the same endpoint it is NOT successful.
https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503.
**Expected behavior**
when I attempt to prefer a POST or Patch on the same endpoint it is need work.
**Screenshots**
https://demos.buddyboss.com/online-communities/wp-json/buddyboss/v1/topics/5503.
**Support ticket links**
https://secure.helpscout.net/conversation/1453624113/130633?folderId=4265312
| priority | rest api in post method on the same endpoint is not working describe the bug we can perform a get on wp json buddyboss topics as well as many other endpoints that respond with a get request this validates that the authentication is working but when we attempt to prefer a post or patch on the same endpoint it is not successful to reproduce when we attempt to prefer a post or patch on the same endpoint it is not successful expected behavior when i attempt to prefer a post or patch on the same endpoint it is need work screenshots support ticket links | 1 |
77,288 | 3,506,336,821 | IssuesEvent | 2016-01-08 05:52:12 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Arenas and battlegrounds not even working (BB #336) | Category: Instances migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** henhouse
**Original Date:** 01.11.2010 12:27:31 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/336
<hr>
You can join arenas but the doors will never open.
Battlegrounds you can join but you cannot see anyone in them, the doors open there. | 1.0 | Arenas and battlegrounds not even working (BB #336) - This issue was migrated from bitbucket.
**Original Reporter:** henhouse
**Original Date:** 01.11.2010 12:27:31 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/336
<hr>
You can join arenas but the doors will never open.
Battlegrounds you can join but you cannot see anyone in them, the doors open there. | priority | arenas and battlegrounds not even working bb this issue was migrated from bitbucket original reporter henhouse original date gmt original priority major original type bug original state resolved direct link you can join arenas but the doors will never open battlegrounds you can join but you cannot see anyone in them the doors open there | 1 |
498,792 | 14,433,225,147 | IssuesEvent | 2020-12-07 04:15:00 | SD2E/experimental-intent-parser | https://api.github.com/repos/SD2E/experimental-intent-parser | closed | “Failed to access SynBioHub” when running “Suggest Additions by Spelling” | bug medium priority | When running “Add to spellchecker dictionary” from“Suggest Additions by Spelling”, IP will report “Failed to access SynBioHub”. What should happen here is each term that a users indicates “add to spellcheck dictionary,” is expected to appear in the dictionary. IP is then expected to store those english words so that it knows which term has already been added to the dictionary. | 1.0 | “Failed to access SynBioHub” when running “Suggest Additions by Spelling” - When running “Add to spellchecker dictionary” from“Suggest Additions by Spelling”, IP will report “Failed to access SynBioHub”. What should happen here is each term that a users indicates “add to spellcheck dictionary,” is expected to appear in the dictionary. IP is then expected to store those english words so that it knows which term has already been added to the dictionary. | priority | “failed to access synbiohub” when running “suggest additions by spelling” when running “add to spellchecker dictionary” from“suggest additions by spelling” ip will report “failed to access synbiohub” what should happen here is each term that a users indicates “add to spellcheck dictionary ” is expected to appear in the dictionary ip is then expected to store those english words so that it knows which term has already been added to the dictionary | 1 |
472,013 | 13,614,694,339 | IssuesEvent | 2020-09-23 13:34:28 | craftercms/craftercms | https://api.github.com/repos/craftercms/craftercms | closed | [studio] Ability to Unpublish the Published Content | CI new feature priority: medium wontfix | As a content author, I should be able to unpublish the published content. At the moment if we want to do this, we have to delete the content item/items and republish again.
The solution I like would be an option to unpublish a published content item/items So we can still keep the same item in authoring.
| 1.0 | [studio] Ability to Unpublish the Published Content - As a content author, I should be able to unpublish the published content. At the moment if we want to do this, we have to delete the content item/items and republish again.
The solution I like would be an option to unpublish a published content item/items So we can still keep the same item in authoring.
| priority | ability to unpublish the published content as a content author i should be able to unpublish the published content at the moment if we want to do this we have to delete the content item items and republish again the solution i like would be an option to unpublish a published content item items so we can still keep the same item in authoring | 1 |
741,216 | 25,784,460,021 | IssuesEvent | 2022-12-09 18:57:19 | EESSI/eessi-bot-software-layer | https://api.github.com/repos/EESSI/eessi-bot-software-layer | opened | Improve tarball handling | enhancement difficulty:medium priority:medium | Currently the script `EESSI-pilot-install-software.sh` of the software-layer repository creates a tarball with the name `eessi-VERSION-COMPONENT-OS-ARCH-TIMESTAMP.tar.gz` and the script `eessi-upload-to-staging` of the eessi-bot-software-layer repository creates a metadata file for a tarball that is uploaded to an S3 bucket. The format of the filename is used a several places. Any change to this may break functionality at different places (in the bot here and elsewhere, e.g., the autoingestion script running on the Stratum 0).
We should come up with a more robust handling of the information needed. See discussion at https://github.com/EESSI/eessi-bot-software-layer/pull/62#discussion_r1023955186 | 1.0 | Improve tarball handling - Currently the script `EESSI-pilot-install-software.sh` of the software-layer repository creates a tarball with the name `eessi-VERSION-COMPONENT-OS-ARCH-TIMESTAMP.tar.gz` and the script `eessi-upload-to-staging` of the eessi-bot-software-layer repository creates a metadata file for a tarball that is uploaded to an S3 bucket. The format of the filename is used a several places. Any change to this may break functionality at different places (in the bot here and elsewhere, e.g., the autoingestion script running on the Stratum 0).
We should come up with a more robust handling of the information needed. See discussion at https://github.com/EESSI/eessi-bot-software-layer/pull/62#discussion_r1023955186 | priority | improve tarball handling currently the script eessi pilot install software sh of the software layer repository creates a tarball with the name eessi version component os arch timestamp tar gz and the script eessi upload to staging of the eessi bot software layer repository creates a metadata file for a tarball that is uploaded to an bucket the format of the filename is used a several places any change to this may break functionality at different places in the bot here and elsewhere e g the autoingestion script running on the stratum we should come up with a more robust handling of the information needed see discussion at | 1 |
618,895 | 19,490,531,951 | IssuesEvent | 2021-12-27 05:07:54 | itsmeow/betteranimalsplus | https://api.github.com/repos/itsmeow/betteranimalsplus | opened | Fabric crash when loading into a server with fresh client near reindeer | crash priority:medium side:client mc:1.16 platform:fabric mc:1.17 mc:1.18 | ## Describe what you were doing before the crash
If a player connects to a Fabric dedicated server on a fresh client that has not joined a singleplayer world or opened the Mods menu, and loads in next to reindeer, it causes a crash, because the client configuration has not initialized.
**Workaround**: Load into a singleplayer world, or open the Mods menu and click BAP's config (w/ Mod Menu and Cloth Config installed)
## To Reproduce
Steps to reproduce the behavior (if possible):
1. Delete `betteranimalsplus-client.json5`
2. Start game
3. Click Multiplayer
4. Connect to a server w/ BAP while nearby reindeer
5. See crash
## Crash Log (REQUIRED)
https://pastebin.com/T1c7Wk93
## Versions
- Minecraft Version: 1.16.5-1.18.1
- Architectury API Version: N/A
- Platform (Forge/Fabric): Fabric
- Forge Version: N/A
- Fabric Loader Version: 0.12.12
- Fabric API Version: N/A
- Mod Version: 11.0.0/11.0.1
| 1.0 | Fabric crash when loading into a server with fresh client near reindeer - ## Describe what you were doing before the crash
If a player connects to a Fabric dedicated server on a fresh client that has not joined a singleplayer world or opened the Mods menu, and loads in next to reindeer, it causes a crash, because the client configuration has not initialized.
**Workaround**: Load into a singleplayer world, or open the Mods menu and click BAP's config (w/ Mod Menu and Cloth Config installed)
## To Reproduce
Steps to reproduce the behavior (if possible):
1. Delete `betteranimalsplus-client.json5`
2. Start game
3. Click Multiplayer
4. Connect to a server w/ BAP while nearby reindeer
5. See crash
## Crash Log (REQUIRED)
https://pastebin.com/T1c7Wk93
## Versions
- Minecraft Version: 1.16.5-1.18.1
- Architectury API Version: N/A
- Platform (Forge/Fabric): Fabric
- Forge Version: N/A
- Fabric Loader Version: 0.12.12
- Fabric API Version: N/A
- Mod Version: 11.0.0/11.0.1
| priority | fabric crash when loading into a server with fresh client near reindeer describe what you were doing before the crash if a player connects to a fabric dedicated server on a fresh client that has not joined a singleplayer world or opened the mods menu and loads in next to reindeer it causes a crash because the client configuration has not initialized workaround load into a singleplayer world or open the mods menu and click bap s config w mod menu and cloth config installed to reproduce steps to reproduce the behavior if possible delete betteranimalsplus client start game click multiplayer connect to a server w bap while nearby reindeer see crash crash log required versions minecraft version architectury api version n a platform forge fabric fabric forge version n a fabric loader version fabric api version n a mod version | 1 |
355,653 | 10,583,239,116 | IssuesEvent | 2019-10-08 13:20:57 | AbsaOSS/enceladus | https://api.github.com/repos/AbsaOSS/enceladus | closed | Add more options to control resources to helper scripts | feature priority: medium run scripts | ## Background
Running Standardization and Conformance on big files revealed options that are essential to run these jobs efficiently.
## Feature
Need to expend `spark-submit` resource request options. Add support for the ollowing:
- `--executor-cores`
- `--conf spark.executor.memoryOverhead=2048`
- `--conf spark.memory.fraction=0.1`
| 1.0 | Add more options to control resources to helper scripts - ## Background
Running Standardization and Conformance on big files revealed options that are essential to run these jobs efficiently.
## Feature
Need to expend `spark-submit` resource request options. Add support for the ollowing:
- `--executor-cores`
- `--conf spark.executor.memoryOverhead=2048`
- `--conf spark.memory.fraction=0.1`
| priority | add more options to control resources to helper scripts background running standardization and conformance on big files revealed options that are essential to run these jobs efficiently feature need to expend spark submit resource request options add support for the ollowing executor cores conf spark executor memoryoverhead conf spark memory fraction | 1 |
57,215 | 3,081,249,057 | IssuesEvent | 2015-08-22 14:40:06 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Variable size testitems | 020 bug duplicate imported Priority-Medium | _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on June 01, 2014 05:37:26_
Do we want to add variable-size testItems? If so... do it!
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=437_ | 1.0 | Variable size testitems - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on June 01, 2014 05:37:26_
Do we want to add variable-size testItems? If so... do it!
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=437_ | priority | variable size testitems from on june do we want to add variable size testitems if so do it original issue | 1 |
59,366 | 3,105,870,595 | IssuesEvent | 2015-08-31 23:32:44 | urbit/urbit | https://api.github.com/repos/urbit/urbit | opened | New FP crashes with SIGABRT on some platforms | bug difficulty medium platform specific priority medium | @ohAitch and @galenwp report:
```
~zod:dojo> (add:rd .~2 .~2)address 0xfffffffffffffffd out of loom!
bail: oops
Abort trap: 6
```
More portability stuff. Looks OSX-specific, so I can't debug it because I have no access to OSX. | 1.0 | New FP crashes with SIGABRT on some platforms - @ohAitch and @galenwp report:
```
~zod:dojo> (add:rd .~2 .~2)address 0xfffffffffffffffd out of loom!
bail: oops
Abort trap: 6
```
More portability stuff. Looks OSX-specific, so I can't debug it because I have no access to OSX. | priority | new fp crashes with sigabrt on some platforms ohaitch and galenwp report zod dojo add rd address out of loom bail oops abort trap more portability stuff looks osx specific so i can t debug it because i have no access to osx | 1 |
797,808 | 28,180,807,085 | IssuesEvent | 2023-04-04 02:09:03 | nimblehq/ic-flutter-taher-toby | https://api.github.com/repos/nimblehq/ic-flutter-taher-toby | closed | Set up CD for iOS (TestFlight) | type : chore priority : medium @0.1.0 @0.2.0 | ## Why
In order to have convenient testing and debugging, a proper CD must be added.
**Note**: This has already been added in `ios_deploy_to_testflight.yml` (staging), but the one for production is still missing.
## Acceptance Criteria
- Create a new project on App Store Connect
- Add necessary environment variables to secrets
- Every time we merge to `develop`, an iOS build has to be deployed to TestFlight (Staging)
- Every time we merge to `main`, an iOS build has to be deployed to TestFlight (Production)
- Rename `ios_deploy_to_testflight.yml` to `ios_deploy_staging.yml`
- Remove `ios_deploy_to_app_store.yml`
## Design
N/A
## Resources
https://developer.apple.com/app-store-connect/ | 1.0 | Set up CD for iOS (TestFlight) - ## Why
In order to have convenient testing and debugging, a proper CD must be added.
**Note**: This has already been added in `ios_deploy_to_testflight.yml` (staging), but the one for production is still missing.
## Acceptance Criteria
- Create a new project on App Store Connect
- Add necessary environment variables to secrets
- Every time we merge to `develop`, an iOS build has to be deployed to TestFlight (Staging)
- Every time we merge to `main`, an iOS build has to be deployed to TestFlight (Production)
- Rename `ios_deploy_to_testflight.yml` to `ios_deploy_staging.yml`
- Remove `ios_deploy_to_app_store.yml`
## Design
N/A
## Resources
https://developer.apple.com/app-store-connect/ | priority | set up cd for ios testflight why in order to have convenient testing and debugging a proper cd must be added note this has already been added in ios deploy to testflight yml staging but the one for production is still missing acceptance criteria create a new project on app store connect add necessary environment variables to secrets every time we merge to develop an ios build has to be deployed to testflight staging every time we merge to main an ios build has to be deployed to testflight production rename ios deploy to testflight yml to ios deploy staging yml remove ios deploy to app store yml design n a resources | 1 |
108,323 | 4,331,105,486 | IssuesEvent | 2016-07-26 22:16:05 | SpeedCurve-Metrics/SpeedCurve | https://api.github.com/repos/SpeedCurve-Metrics/SpeedCurve | closed | checkbox to preserve UA string | priority medium status new type enhancement | Similar to public WebPageTest, have a checkbox to preserve original UA string (ie, do NOT add "PTST"). Some services (such as ads) serve different content if PTST is present.
Note that this same behavior can be achieved by using scripting, although that scripted URL could only be run on one browser (since the UA string in the script is specific to a single browser). | 1.0 | checkbox to preserve UA string - Similar to public WebPageTest, have a checkbox to preserve original UA string (ie, do NOT add "PTST"). Some services (such as ads) serve different content if PTST is present.
Note that this same behavior can be achieved by using scripting, although that scripted URL could only be run on one browser (since the UA string in the script is specific to a single browser). | priority | checkbox to preserve ua string similar to public webpagetest have a checkbox to preserve original ua string ie do not add ptst some services such as ads serve different content if ptst is present note that this same behavior can be achieved by using scripting although that scripted url could only be run on one browser since the ua string in the script is specific to a single browser | 1 |
642,418 | 20,887,420,754 | IssuesEvent | 2022-03-23 07:25:23 | giancarlopernudisegura/cmput404 | https://api.github.com/repos/giancarlopernudisegura/cmput404 | closed | [User Story]: Remove Nodes | Priority Level: Medium Implementation Difficulty: Medium User Story | ### User Story
As a server admin, I want to be able to remove nodes and stop sharing with them.
### Objects
Inbox | 1.0 | [User Story]: Remove Nodes - ### User Story
As a server admin, I want to be able to remove nodes and stop sharing with them.
### Objects
Inbox | priority | remove nodes user story as a server admin i want to be able to remove nodes and stop sharing with them objects inbox | 1 |
25,926 | 2,684,046,392 | IssuesEvent | 2015-03-28 16:10:22 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | не передаются нажатия кнопок в GUI приложения во вкладках ConEmu | 1 star bug imported Priority-Medium | _From [SAnTA...@gmail.com](https://code.google.com/u/105830449921090467193/) on February 09, 2012 01:34:28_
Required information! OS version: Win7 SP1 x64 ConEmu version: build 120205 (7z version)
Far version: Far Manager, version 3.0 (build 2434) x64
PuTTy version: 2012-02-09 (x86-32)
Давно пользуюсь ConEmu и все терзала мысль о том чт оне хватает в ней PuTTy, а тут вдруг выясняю что работы по внедрению ведутся полным ходом. Решил попробовать. Запускается волшебно вот только на нажатия стрелок и на F1-F12 клавиши PuTTy не реагирует.
Комбинации Ctrl+N и Ctrl+P в качестве замены стрелкам вверх и вниз, соответственно, работают, но как то с ними не очень удобно.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=490_ | 1.0 | не передаются нажатия кнопок в GUI приложения во вкладках ConEmu - _From [SAnTA...@gmail.com](https://code.google.com/u/105830449921090467193/) on February 09, 2012 01:34:28_
Required information! OS version: Win7 SP1 x64 ConEmu version: build 120205 (7z version)
Far version: Far Manager, version 3.0 (build 2434) x64
PuTTy version: 2012-02-09 (x86-32)
Давно пользуюсь ConEmu и все терзала мысль о том чт оне хватает в ней PuTTy, а тут вдруг выясняю что работы по внедрению ведутся полным ходом. Решил попробовать. Запускается волшебно вот только на нажатия стрелок и на F1-F12 клавиши PuTTy не реагирует.
Комбинации Ctrl+N и Ctrl+P в качестве замены стрелкам вверх и вниз, соответственно, работают, но как то с ними не очень удобно.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=490_ | priority | не передаются нажатия кнопок в gui приложения во вкладках conemu from on february required information os version conemu version build version far version far manager version build putty version давно пользуюсь conemu и все терзала мысль о том чт оне хватает в ней putty а тут вдруг выясняю что работы по внедрению ведутся полным ходом решил попробовать запускается волшебно вот только на нажатия стрелок и на клавиши putty не реагирует комбинации ctrl n и ctrl p в качестве замены стрелкам вверх и вниз соответственно работают но как то с ними не очень удобно original issue | 1 |
378,567 | 11,204,620,145 | IssuesEvent | 2020-01-05 07:39:57 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | opened | Tradingview.com templates need time zone details. | Priority: Medium V2 Audit | TradingView.com supports user configurable time zone. The market description for these templates needs to specify that the user should use UTC time zone when determining the open/close price. | 1.0 | Tradingview.com templates need time zone details. - TradingView.com supports user configurable time zone. The market description for these templates needs to specify that the user should use UTC time zone when determining the open/close price. | priority | tradingview com templates need time zone details tradingview com supports user configurable time zone the market description for these templates needs to specify that the user should use utc time zone when determining the open close price | 1 |
434,514 | 12,519,519,902 | IssuesEvent | 2020-06-03 14:33:33 | Twin-Cities-Mutual-Aid/twin-cities-aid-distribution-locations | https://api.github.com/repos/Twin-Cities-Mutual-Aid/twin-cities-aid-distribution-locations | opened | Security concerns | Priority: Medium Type: Discussion | Wanted to open an issue so we can have an ongoing discussion for how to address security related to this project. A first step is captured in #27.
My main concern in raising this is the possibility that this map could be used to target donation centers for harassment, but that's also just a risk inherent to the project. I'm sure there are other concerns as well.
This doesn't necessarily need to be limited to technical solutions but anything that will help keep our volunteers safe and prevent any malicious attacks on the site. | 1.0 | Security concerns - Wanted to open an issue so we can have an ongoing discussion for how to address security related to this project. A first step is captured in #27.
My main concern in raising this is the possibility that this map could be used to target donation centers for harassment, but that's also just a risk inherent to the project. I'm sure there are other concerns as well.
This doesn't necessarily need to be limited to technical solutions but anything that will help keep our volunteers safe and prevent any malicious attacks on the site. | priority | security concerns wanted to open an issue so we can have an ongoing discussion for how to address security related to this project a first step is captured in my main concern in raising this is the possibility that this map could be used to target donation centers for harassment but that s also just a risk inherent to the project i m sure there are other concerns as well this doesn t necessarily need to be limited to technical solutions but anything that will help keep our volunteers safe and prevent any malicious attacks on the site | 1 |
886 | 2,504,525,266 | IssuesEvent | 2015-01-10 09:15:30 | Araq/Nim | https://api.github.com/repos/Araq/Nim | opened | Add file handle procedures to async modules | Medium Priority Stdlib | Aside from file handles, the async modules should also be able to deal with file handles/descriptors, such as reading and writing. | 1.0 | Add file handle procedures to async modules - Aside from file handles, the async modules should also be able to deal with file handles/descriptors, such as reading and writing. | priority | add file handle procedures to async modules aside from file handles the async modules should also be able to deal with file handles descriptors such as reading and writing | 1 |
478,379 | 13,778,655,432 | IssuesEvent | 2020-10-08 12:46:47 | level73/membernet | https://api.github.com/repos/level73/membernet | opened | backend: long load time confusing users as when manage entity is clicked, it opens new form instead of list | Platform: Membernet Priority: Medium Type: Enhancement | i am not sure if you can do anything about this, but this is causing confusion for new users that are not aware that htey need to let the page load before selecting the specific management entity for viewing/editing the list- as otherwise it opens a new entry... can we do something about this? at least a line of text at the top? | 1.0 | backend: long load time confusing users as when manage entity is clicked, it opens new form instead of list - i am not sure if you can do anything about this, but this is causing confusion for new users that are not aware that htey need to let the page load before selecting the specific management entity for viewing/editing the list- as otherwise it opens a new entry... can we do something about this? at least a line of text at the top? | priority | backend long load time confusing users as when manage entity is clicked it opens new form instead of list i am not sure if you can do anything about this but this is causing confusion for new users that are not aware that htey need to let the page load before selecting the specific management entity for viewing editing the list as otherwise it opens a new entry can we do something about this at least a line of text at the top | 1 |
706,998 | 24,290,975,493 | IssuesEvent | 2022-09-29 05:56:10 | awslabs/aws-saas-boost | https://api.github.com/repos/awslabs/aws-saas-boost | closed | Updating AppConfig deletes SaaS Boost created HostedZone | bug priority-medium application-management tenant-onboarding | If you have a SaaS Boost environment that automatically created a HostedZone because it was configured with a DomainName, updating the AppConfig afterwards (even if just updating something like a service description) will result in the HostedZone getting deleted.
### Reproduction Steps
<!--
minimal amount of code that causes the bug (if possible) or a reference:
-->
1. Stand up a SaaS Boost environment
2. Configure it with a Domain Name and 1 service
3. See that a HostedZone is created in Route53 with the provided Domain Name as its name and the description `"${DomainName} Public DNS Zone"`
4. Update the service description
5. See that Settings service updates the config and fires an "Application Configuration Changed" event in the `sb-${env}-settings-update-config` CloudWatch LogGroup
6. See that the Onboarding service finds the existing HostedZone and updates the base stack in `sb-${env}-onboarding-events`
7. See that the HostedZone is eventually deleted.
### What did you expect to happen?
I expect that the HostedZone would stick around.
### What actually happened?
The HostedZone was deleted.
### Environment
- **AWS Region :** us-west-2
- **AWS SaaS Boost Version :** `main` at https://github.com/awslabs/aws-saas-boost/commit/176187a405f9ce574544745ed8ec01e0db7172dd
- **Workload OS (Linux or Windows) :** N/A
### Other
<!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, slack, etc -->
This is because the condition on which the HostedZone is created is based on whether a HostedZone is passed (the intention here was to only create a HostedZone if one does not exist): [source](https://github.com/awslabs/aws-saas-boost/blob/main/resources/saas-boost-core.yaml#L64-L67)
```
# in resources/saas-boost-core.yaml
Conditions:
CreateHostedZone: !And
- !Not [!Equals [!Ref DomainName, '']]
- !Equals [!Ref HostedZone, '']
```
However, the Onboarding Service sets the `HostedZone` parameter when updating the stack to whatever exists in Route53, which will be the Boost-created HostedZone if none already existed: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L1771-L1801)
```
String hostedZone = getExistingHostedZone(domainName);
// If there's an existing hosted zone, we need to tell the AppConfig about it
// Otherwise, if there's a domain name, CloudFormation will create a hosted zone
// and the stack listener will tell AppConfig about the newly created one.
if (Utils.isNotBlank(hostedZone)) {
LOGGER.info("Publishing appConfig update event for Route53 hosted zone {}", hostedZone);
Utils.publishEvent(eventBridge, SAAS_BOOST_EVENT_BUS, EVENT_SOURCE,
"Application Configuration Resource Changed",
Map.of("hostedZone", hostedZone));
}
...
Parameter.builder().parameterKey("HostedZone").parameterValue(hostedZone).build(),
```
Using the `getExistingHostedZone` function: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L2188-L2226). It's worth noting that this is also where we only reuse existing HostedZones if SaaS Boost already created them, an issue brought up by users in https://github.com/awslabs/aws-saas-boost/discussions/299 and https://github.com/awslabs/aws-saas-boost/issues/257.
```
protected String getExistingHostedZone(String domainName) {
String existingHostedZone = "";
if (Utils.isNotEmpty(domainName)) {
String nextDnsName = null;
String nextHostedZone = null;
ListHostedZonesByNameResponse response;
do {
response = route53.listHostedZonesByName(ListHostedZonesByNameRequest.builder()
.dnsName(nextDnsName)
.hostedZoneId(nextHostedZone)
.maxItems("100")
.build()
);
nextDnsName = response.nextDNSName();
nextHostedZone = response.nextHostedZoneId();
if (response.hasHostedZones()) {
for (HostedZone hostedZone : response.hostedZones()) {
// If there are multiple hosted zones for a given domain name, what should we do?
// We could sort the response by "CallerReference" which appears to be a timestamp.
// In the documentation, we can just tell people if they're suffering from
// https://github.com/awslabs/aws-saas-boost/issues/74 to go clean things up manually first?
if (hostedZone.name().startsWith(domainName)
&& hostedZone.config() != null
&& Boolean.FALSE.equals(hostedZone.config().privateZone())) {
// Created by SaaS Boost CloudFormation?
// TODO do we do this check? seems safest for now.
if ((domainName + " Public DNS zone").equals(hostedZone.config().comment())) {
LOGGER.info("Found existing hosted zone {} for domain {}", hostedZone, domainName);
// Hosted zone id will be prefixed with /hostedzone/
existingHostedZone = hostedZone.id().replace("/hostedzone/", "");
break;
}
}
}
}
} while (response.isTruncated());
}
return existingHostedZone;
}
```
When the Onboarding Service updates the stack passing in the created HostedZone id, the condition evaluates to `false` and CloudFormation deletes the HostedZone.
One possible way to fix this is to view the existing stack resources for the core stack before passing template parameters. Pseudo-code follows:
```
if appConfig.domainName:
stackParameters.domainName = appConfig.domainName
stackResources = cfn.getStackResources(sb-env-core)
if "hostedZone" not in stackResources && route53.hostedZoneExists(name=appConfig.domainName):
// there exists a hostedZone in the user's account that works for this domain name
// but SaaS Boost didn't create it. so add it in as a parameter: meaning the CloudFormation
// condition will evaluate to false and no hostedZone will be created (and none will be deleted,
// since it didn't already exist owned by this template)
stackParameters.hostedZone = route53.hostedZone(name=appConfig.domainName).id
else:
// either hostedZone was created by SaaS Boost or there is no hosted zone in the user account
// that works for this domain name. in either case we want the "CreateHostedZone" condition
// to evaluate to true so either:
// - the SaaS Boost created hosted zone won't be deleted
// or
// - a hosted zone will be created by SaaS Boost, since none exist
// we do this by providing no value for the hostedZone parameter
stackParameters.hostedZone = "" // the same as having no else in production code
```
---
This is :bug: Bug Report
| 1.0 | Updating AppConfig deletes SaaS Boost created HostedZone - If you have a SaaS Boost environment that automatically created a HostedZone because it was configured with a DomainName, updating the AppConfig afterwards (even if just updating something like a service description) will result in the HostedZone getting deleted.
### Reproduction Steps
<!--
minimal amount of code that causes the bug (if possible) or a reference:
-->
1. Stand up a SaaS Boost environment
2. Configure it with a Domain Name and 1 service
3. See that a HostedZone is created in Route53 with the provided Domain Name as its name and the description `"${DomainName} Public DNS Zone"`
4. Update the service description
5. See that Settings service updates the config and fires an "Application Configuration Changed" event in the `sb-${env}-settings-update-config` CloudWatch LogGroup
6. See that the Onboarding service finds the existing HostedZone and updates the base stack in `sb-${env}-onboarding-events`
7. See that the HostedZone is eventually deleted.
### What did you expect to happen?
I expect that the HostedZone would stick around.
### What actually happened?
The HostedZone was deleted.
### Environment
- **AWS Region :** us-west-2
- **AWS SaaS Boost Version :** `main` at https://github.com/awslabs/aws-saas-boost/commit/176187a405f9ce574544745ed8ec01e0db7172dd
- **Workload OS (Linux or Windows) :** N/A
### Other
<!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, slack, etc -->
This is because the condition on which the HostedZone is created is based on whether a HostedZone is passed (the intention here was to only create a HostedZone if one does not exist): [source](https://github.com/awslabs/aws-saas-boost/blob/main/resources/saas-boost-core.yaml#L64-L67)
```
# in resources/saas-boost-core.yaml
Conditions:
CreateHostedZone: !And
- !Not [!Equals [!Ref DomainName, '']]
- !Equals [!Ref HostedZone, '']
```
However, the Onboarding Service sets the `HostedZone` parameter when updating the stack to whatever exists in Route53, which will be the Boost-created HostedZone if none already existed: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L1771-L1801)
```
String hostedZone = getExistingHostedZone(domainName);
// If there's an existing hosted zone, we need to tell the AppConfig about it
// Otherwise, if there's a domain name, CloudFormation will create a hosted zone
// and the stack listener will tell AppConfig about the newly created one.
if (Utils.isNotBlank(hostedZone)) {
LOGGER.info("Publishing appConfig update event for Route53 hosted zone {}", hostedZone);
Utils.publishEvent(eventBridge, SAAS_BOOST_EVENT_BUS, EVENT_SOURCE,
"Application Configuration Resource Changed",
Map.of("hostedZone", hostedZone));
}
...
Parameter.builder().parameterKey("HostedZone").parameterValue(hostedZone).build(),
```
Using the `getExistingHostedZone` function: [source](https://github.com/awslabs/aws-saas-boost/blob/main/services/onboarding-service/src/main/java/com/amazon/aws/partners/saasfactory/saasboost/OnboardingService.java#L2188-L2226). It's worth noting that this is also where we only reuse existing HostedZones if SaaS Boost already created them, an issue brought up by users in https://github.com/awslabs/aws-saas-boost/discussions/299 and https://github.com/awslabs/aws-saas-boost/issues/257.
```
protected String getExistingHostedZone(String domainName) {
String existingHostedZone = "";
if (Utils.isNotEmpty(domainName)) {
String nextDnsName = null;
String nextHostedZone = null;
ListHostedZonesByNameResponse response;
do {
response = route53.listHostedZonesByName(ListHostedZonesByNameRequest.builder()
.dnsName(nextDnsName)
.hostedZoneId(nextHostedZone)
.maxItems("100")
.build()
);
nextDnsName = response.nextDNSName();
nextHostedZone = response.nextHostedZoneId();
if (response.hasHostedZones()) {
for (HostedZone hostedZone : response.hostedZones()) {
// If there are multiple hosted zones for a given domain name, what should we do?
// We could sort the response by "CallerReference" which appears to be a timestamp.
// In the documentation, we can just tell people if they're suffering from
// https://github.com/awslabs/aws-saas-boost/issues/74 to go clean things up manually first?
if (hostedZone.name().startsWith(domainName)
&& hostedZone.config() != null
&& Boolean.FALSE.equals(hostedZone.config().privateZone())) {
// Created by SaaS Boost CloudFormation?
// TODO do we do this check? seems safest for now.
if ((domainName + " Public DNS zone").equals(hostedZone.config().comment())) {
LOGGER.info("Found existing hosted zone {} for domain {}", hostedZone, domainName);
// Hosted zone id will be prefixed with /hostedzone/
existingHostedZone = hostedZone.id().replace("/hostedzone/", "");
break;
}
}
}
}
} while (response.isTruncated());
}
return existingHostedZone;
}
```
When the Onboarding Service updates the stack passing in the created HostedZone id, the condition evaluates to `false` and CloudFormation deletes the HostedZone.
One possible way to fix this is to view the existing stack resources for the core stack before passing template parameters. Pseudo-code follows:
```
if appConfig.domainName:
stackParameters.domainName = appConfig.domainName
stackResources = cfn.getStackResources(sb-env-core)
if "hostedZone" not in stackResources && route53.hostedZoneExists(name=appConfig.domainName):
// there exists a hostedZone in the user's account that works for this domain name
// but SaaS Boost didn't create it. so add it in as a parameter: meaning the CloudFormation
// condition will evaluate to false and no hostedZone will be created (and none will be deleted,
// since it didn't already exist owned by this template)
stackParameters.hostedZone = route53.hostedZone(name=appConfig.domainName).id
else:
// either hostedZone was created by SaaS Boost or there is no hosted zone in the user account
// that works for this domain name. in either case we want the "CreateHostedZone" condition
// to evaluate to true so either:
// - the SaaS Boost created hosted zone won't be deleted
// or
// - a hosted zone will be created by SaaS Boost, since none exist
// we do this by providing no value for the hostedZone parameter
stackParameters.hostedZone = "" // the same as having no else in production code
```
---
This is :bug: Bug Report
| priority | updating appconfig deletes saas boost created hostedzone if you have a saas boost environment that automatically created a hostedzone because it was configured with a domainname updating the appconfig afterwards even if just updating something like a service description will result in the hostedzone getting deleted reproduction steps minimal amount of code that causes the bug if possible or a reference stand up a saas boost environment configure it with a domain name and service see that a hostedzone is created in with the provided domain name as its name and the description domainname public dns zone update the service description see that settings service updates the config and fires an application configuration changed event in the sb env settings update config cloudwatch loggroup see that the onboarding service finds the existing hostedzone and updates the base stack in sb env onboarding events see that the hostedzone is eventually deleted what did you expect to happen i expect that the hostedzone would stick around what actually happened the hostedzone was deleted environment aws region us west aws saas boost version main at workload os linux or windows n a other this is because the condition on which the hostedzone is created is based on whether a hostedzone is passed the intention here was to only create a hostedzone if one does not exist in resources saas boost core yaml conditions createhostedzone and not equals however the onboarding service sets the hostedzone parameter when updating the stack to whatever exists in which will be the boost created hostedzone if none already existed string hostedzone getexistinghostedzone domainname if there s an existing hosted zone we need to tell the appconfig about it otherwise if there s a domain name cloudformation will create a hosted zone and the stack listener will tell appconfig about the newly created one if utils isnotblank hostedzone logger info publishing appconfig update event for hosted zone hostedzone utils publishevent eventbridge saas boost event bus event source application configuration resource changed map of hostedzone hostedzone parameter builder parameterkey hostedzone parametervalue hostedzone build using the getexistinghostedzone function it s worth noting that this is also where we only reuse existing hostedzones if saas boost already created them an issue brought up by users in and protected string getexistinghostedzone string domainname string existinghostedzone if utils isnotempty domainname string nextdnsname null string nexthostedzone null listhostedzonesbynameresponse response do response listhostedzonesbyname listhostedzonesbynamerequest builder dnsname nextdnsname hostedzoneid nexthostedzone maxitems build nextdnsname response nextdnsname nexthostedzone response nexthostedzoneid if response hashostedzones for hostedzone hostedzone response hostedzones if there are multiple hosted zones for a given domain name what should we do we could sort the response by callerreference which appears to be a timestamp in the documentation we can just tell people if they re suffering from to go clean things up manually first if hostedzone name startswith domainname hostedzone config null boolean false equals hostedzone config privatezone created by saas boost cloudformation todo do we do this check seems safest for now if domainname public dns zone equals hostedzone config comment logger info found existing hosted zone for domain hostedzone domainname hosted zone id will be prefixed with hostedzone existinghostedzone hostedzone id replace hostedzone break while response istruncated return existinghostedzone when the onboarding service updates the stack passing in the created hostedzone id the condition evaluates to false and cloudformation deletes the hostedzone one possible way to fix this is to view the existing stack resources for the core stack before passing template parameters pseudo code follows if appconfig domainname stackparameters domainname appconfig domainname stackresources cfn getstackresources sb env core if hostedzone not in stackresources hostedzoneexists name appconfig domainname there exists a hostedzone in the user s account that works for this domain name but saas boost didn t create it so add it in as a parameter meaning the cloudformation condition will evaluate to false and no hostedzone will be created and none will be deleted since it didn t already exist owned by this template stackparameters hostedzone hostedzone name appconfig domainname id else either hostedzone was created by saas boost or there is no hosted zone in the user account that works for this domain name in either case we want the createhostedzone condition to evaluate to true so either the saas boost created hosted zone won t be deleted or a hosted zone will be created by saas boost since none exist we do this by providing no value for the hostedzone parameter stackparameters hostedzone the same as having no else in production code this is bug bug report | 1 |
485,662 | 13,996,699,275 | IssuesEvent | 2020-10-28 06:30:16 | AY2021S1-CS2113-T14-3/tp | https://api.github.com/repos/AY2021S1-CS2113-T14-3/tp | closed | Automatically create file directories and folders for modules | priority.Medium type.Story type.Task | As a student, I want to be able to automatically create file directories and folders on my computer (i.e. 2021S1 → CS2113 and this folder contains ‘Tutorial’ and ‘Lecture Notes’ folders) for the modules that I am taking this semester, so that I do not have to manually create them. | 1.0 | Automatically create file directories and folders for modules - As a student, I want to be able to automatically create file directories and folders on my computer (i.e. 2021S1 → CS2113 and this folder contains ‘Tutorial’ and ‘Lecture Notes’ folders) for the modules that I am taking this semester, so that I do not have to manually create them. | priority | automatically create file directories and folders for modules as a student i want to be able to automatically create file directories and folders on my computer i e → and this folder contains ‘tutorial’ and ‘lecture notes’ folders for the modules that i am taking this semester so that i do not have to manually create them | 1 |
26,316 | 2,684,309,894 | IssuesEvent | 2015-03-28 21:16:31 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | opened | ConEmuHk screws up output of long lines | 1 star bug imported Priority-Medium | _From [bahamutzero8825](https://code.google.com/u/bahamutzero8825/) on March 04, 2013 22:04:57_
OS version: Windows 8 x64 ConEmu version: 130304 x64
Far version (if you are using Far Manager): not using Far
I've been developing a program in Python where multiple threads and children process can write to the same terminal at the same time. This has generally not been an issue, but I have a couple logging calls that output long lines and ConEmuHk seems to mess up the Python interpreter (CPython x64 3.3.0 if that matters) and I get garbled output (separate lines get interleaved). If ConEmuHk is not injected or I am viewing a remote session over SSH, the output is fine. I noticed this in 130223, and it seems to be better in 130304 (the former mangled several lines, the latter only 2-3). *Steps to reproduction* 1. Enable ConEmuHk injection
2. Write more than one thing to the terminal at once.
3. Get garbled output.
If you cannot reproduce this on your own, I will write a short Python script to reproduce it.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=973_ | 1.0 | ConEmuHk screws up output of long lines - _From [bahamutzero8825](https://code.google.com/u/bahamutzero8825/) on March 04, 2013 22:04:57_
OS version: Windows 8 x64 ConEmu version: 130304 x64
Far version (if you are using Far Manager): not using Far
I've been developing a program in Python where multiple threads and children process can write to the same terminal at the same time. This has generally not been an issue, but I have a couple logging calls that output long lines and ConEmuHk seems to mess up the Python interpreter (CPython x64 3.3.0 if that matters) and I get garbled output (separate lines get interleaved). If ConEmuHk is not injected or I am viewing a remote session over SSH, the output is fine. I noticed this in 130223, and it seems to be better in 130304 (the former mangled several lines, the latter only 2-3). *Steps to reproduction* 1. Enable ConEmuHk injection
2. Write more than one thing to the terminal at once.
3. Get garbled output.
If you cannot reproduce this on your own, I will write a short Python script to reproduce it.
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=973_ | priority | conemuhk screws up output of long lines from on march os version windows conemu version far version if you are using far manager not using far i ve been developing a program in python where multiple threads and children process can write to the same terminal at the same time this has generally not been an issue but i have a couple logging calls that output long lines and conemuhk seems to mess up the python interpreter cpython if that matters and i get garbled output separate lines get interleaved if conemuhk is not injected or i am viewing a remote session over ssh the output is fine i noticed this in and it seems to be better in the former mangled several lines the latter only steps to reproduction enable conemuhk injection write more than one thing to the terminal at once get garbled output if you cannot reproduce this on your own i will write a short python script to reproduce it original issue | 1 |
309,294 | 9,466,471,626 | IssuesEvent | 2019-04-18 04:43:17 | wso2/product-is | https://api.github.com/repos/wso2/product-is | reopened | When reCaptcha is enabled, multi option login steps are not shown for given number of failed attempts. | Affected/5.8.0-Alpha2 Complexity/Medium Component/Adaptive Auth Priority/High Severity/Critical Type/Bug | - Add a Service Provider (OIDC/SSO samples)
- Enable reCaptcha (after 1 failed login)
- Enable multi-option login (eg: basic auth as step 1 and Email OTP as step 2)
- Add a new claim to store the failed attempts before login.
- Use script based adaptive authentication and added the following code.
```
// This variable is used to define the number of invalid attempts allowed before prompting the second factor
var invalidAttemptsToStepup = 2;
var failedLoginAttemptsBeforeSuccessClaim= 'http://wso2.org/claims/identity/failedLoginAttemptsBeforeSuccess';
function onLoginRequest(context) {
doLogin(context);
}
function doLogin(context) {
executeStep(1, {
onSuccess : function(context){
var user = context.steps[1].subject;
if (isExceedInvalidAttempts(user)) {
executeStep(2);
}
},
onFail : function(context) {
// Retry the login..
doLogin(context);
}
});
}
function isExceedInvalidAttempts(user) {
if (user.localClaims[failedLoginAttemptsBeforeSuccessClaim] >= invalidAttemptsToStepup) {
return true;
} else {
return false;
}
}
```
When reCaptcha is set to come up after 1 failed login attempt and Email OTP to come up after 2 failed login attempts, to get E-mail OTP (which is set as the 2nd step) it takes 5 or more invalid login attempts. | 1.0 | When reCaptcha is enabled, multi option login steps are not shown for given number of failed attempts. - - Add a Service Provider (OIDC/SSO samples)
- Enable reCaptcha (after 1 failed login)
- Enable multi-option login (eg: basic auth as step 1 and Email OTP as step 2)
- Add a new claim to store the failed attempts before login.
- Use script based adaptive authentication and added the following code.
```
// This variable is used to define the number of invalid attempts allowed before prompting the second factor
var invalidAttemptsToStepup = 2;
var failedLoginAttemptsBeforeSuccessClaim= 'http://wso2.org/claims/identity/failedLoginAttemptsBeforeSuccess';
function onLoginRequest(context) {
doLogin(context);
}
function doLogin(context) {
executeStep(1, {
onSuccess : function(context){
var user = context.steps[1].subject;
if (isExceedInvalidAttempts(user)) {
executeStep(2);
}
},
onFail : function(context) {
// Retry the login..
doLogin(context);
}
});
}
function isExceedInvalidAttempts(user) {
if (user.localClaims[failedLoginAttemptsBeforeSuccessClaim] >= invalidAttemptsToStepup) {
return true;
} else {
return false;
}
}
```
When reCaptcha is set to come up after 1 failed login attempt and Email OTP to come up after 2 failed login attempts, to get E-mail OTP (which is set as the 2nd step) it takes 5 or more invalid login attempts. | priority | when recaptcha is enabled multi option login steps are not shown for given number of failed attempts add a service provider oidc sso samples enable recaptcha after failed login enable multi option login eg basic auth as step and email otp as step add a new claim to store the failed attempts before login use script based adaptive authentication and added the following code this variable is used to define the number of invalid attempts allowed before prompting the second factor var invalidattemptstostepup var failedloginattemptsbeforesuccessclaim function onloginrequest context dologin context function dologin context executestep onsuccess function context var user context steps subject if isexceedinvalidattempts user executestep onfail function context retry the login dologin context function isexceedinvalidattempts user if user localclaims invalidattemptstostepup return true else return false when recaptcha is set to come up after failed login attempt and email otp to come up after failed login attempts to get e mail otp which is set as the step it takes or more invalid login attempts | 1 |
112,240 | 4,513,777,752 | IssuesEvent | 2016-09-04 13:48:30 | thommoboy/There-are-no-brakes | https://api.github.com/repos/thommoboy/There-are-no-brakes | closed | Industry level: players and box pass through the wall if elevator shove them | bug Industrialists Priority Medium | if player stand someplace that elevator will come, elevator will shove and let player pass through the wall(even out of map) | 1.0 | Industry level: players and box pass through the wall if elevator shove them - if player stand someplace that elevator will come, elevator will shove and let player pass through the wall(even out of map) | priority | industry level players and box pass through the wall if elevator shove them if player stand someplace that elevator will come elevator will shove and let player pass through the wall even out of map | 1 |
644,376 | 20,976,063,121 | IssuesEvent | 2022-03-28 15:18:48 | msavastio/ZohoCreatorWidgetReact | https://api.github.com/repos/msavastio/ZohoCreatorWidgetReact | opened | Rework project audit UI | medium priority business priority | - [ ] Remove float and total w/ float from "As Engineered"
- [ ] Add in Proposal Amount field when doing Opportunity => Closed Won
- [ ] As Contracted now becomes as Engineered with Project Amount and Float - these are static values
- [ ] As Engineered with Changes remains unchanged
- [ ] Equipment budget will be original quotes +/- any change orders internal or external | 2.0 | Rework project audit UI - - [ ] Remove float and total w/ float from "As Engineered"
- [ ] Add in Proposal Amount field when doing Opportunity => Closed Won
- [ ] As Contracted now becomes as Engineered with Project Amount and Float - these are static values
- [ ] As Engineered with Changes remains unchanged
- [ ] Equipment budget will be original quotes +/- any change orders internal or external | priority | rework project audit ui remove float and total w float from as engineered add in proposal amount field when doing opportunity closed won as contracted now becomes as engineered with project amount and float these are static values as engineered with changes remains unchanged equipment budget will be original quotes any change orders internal or external | 1 |
31,369 | 2,732,897,037 | IssuesEvent | 2015-04-17 10:04:07 | tiku01/oryx-editor | https://api.github.com/repos/tiku01/oryx-editor | closed | I want to tag models as I save them | auto-migrated Priority-Medium Type-Enhancement | ```
What steps will reproduce the problem?
1. Creating many models that belong to one project
2. when save the model
3. go to the repository everytime, refresh view, select model, save
What is the expected output?
I'd like to tag the models directly in the savescreen dialog
What do you see instead?
Otherwise it takes much longer and I need 4 steps to do it.
Please provide any additional information below.
```
Original issue reported on code.google.com by `alexande...@googlemail.com` on 1 Dec 2008 at 11:50 | 1.0 | I want to tag models as I save them - ```
What steps will reproduce the problem?
1. Creating many models that belong to one project
2. when save the model
3. go to the repository everytime, refresh view, select model, save
What is the expected output?
I'd like to tag the models directly in the savescreen dialog
What do you see instead?
Otherwise it takes much longer and I need 4 steps to do it.
Please provide any additional information below.
```
Original issue reported on code.google.com by `alexande...@googlemail.com` on 1 Dec 2008 at 11:50 | priority | i want to tag models as i save them what steps will reproduce the problem creating many models that belong to one project when save the model go to the repository everytime refresh view select model save what is the expected output i d like to tag the models directly in the savescreen dialog what do you see instead otherwise it takes much longer and i need steps to do it please provide any additional information below original issue reported on code google com by alexande googlemail com on dec at | 1 |
487,357 | 14,045,141,402 | IssuesEvent | 2020-11-02 00:22:50 | codidact/qpixel | https://api.github.com/repos/codidact/qpixel | opened | Markdown tries to apply block-level styling to comments, losing content | area: frontend complexity: unassessed priority: medium type: bug | https://meta.codidact.com/questions/278611 is a bug report about comments that begin with "1. " -- the number and period disappear. In comments, @luap42 says it's a more general problem with Markdown that looks like it's for block elements:
> Furthermore, it appears that the issue isn't specifically with numbers followed by a dot but by everything that can cause Markdown to produce something (lists, headings, quotes, ...). The HTML is sanitized differently for comments, but Markdown doesn't know that and hence produces a list/header/comment. The sanitizer then removes the <blockquote> tag and hence it looks like the number (or dash or #) was removed. I think the right solution would be to make the Markdown generator know that we are in "inline Markdown" mode and hence not generate block-level stuff. — luap42 is a ghost 👻 14 days ago | 1.0 | Markdown tries to apply block-level styling to comments, losing content - https://meta.codidact.com/questions/278611 is a bug report about comments that begin with "1. " -- the number and period disappear. In comments, @luap42 says it's a more general problem with Markdown that looks like it's for block elements:
> Furthermore, it appears that the issue isn't specifically with numbers followed by a dot but by everything that can cause Markdown to produce something (lists, headings, quotes, ...). The HTML is sanitized differently for comments, but Markdown doesn't know that and hence produces a list/header/comment. The sanitizer then removes the <blockquote> tag and hence it looks like the number (or dash or #) was removed. I think the right solution would be to make the Markdown generator know that we are in "inline Markdown" mode and hence not generate block-level stuff. — luap42 is a ghost 👻 14 days ago | priority | markdown tries to apply block level styling to comments losing content is a bug report about comments that begin with the number and period disappear in comments says it s a more general problem with markdown that looks like it s for block elements furthermore it appears that the issue isn t specifically with numbers followed by a dot but by everything that can cause markdown to produce something lists headings quotes the html is sanitized differently for comments but markdown doesn t know that and hence produces a list header comment the sanitizer then removes the tag and hence it looks like the number or dash or was removed i think the right solution would be to make the markdown generator know that we are in inline markdown mode and hence not generate block level stuff — is a ghost 👻 days ago | 1 |
178,726 | 6,617,919,427 | IssuesEvent | 2017-09-21 05:19:02 | za419/VersionControl | https://api.github.com/repos/za419/VersionControl | closed | Rename the project | enhancement help wanted Medium Priority | Proposed names:
- Odin (by @kenellorando)
- Seshat (the Egyptian goddess of wisdom, knowledge, and writing; a record keeper)
- Hero[dotus] (the father of history) | 1.0 | Rename the project - Proposed names:
- Odin (by @kenellorando)
- Seshat (the Egyptian goddess of wisdom, knowledge, and writing; a record keeper)
- Hero[dotus] (the father of history) | priority | rename the project proposed names odin by kenellorando seshat the egyptian goddess of wisdom knowledge and writing a record keeper hero the father of history | 1 |
767,820 | 26,941,435,311 | IssuesEvent | 2023-02-08 02:46:16 | Benjamin-Loison/YouTube-operational-API | https://api.github.com/repos/Benjamin-Loison/YouTube-operational-API | closed | Propose a Docker version? | enhancement help wanted discussion low priority medium | As found in [the forks of this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/network/members), could inspire myself of [this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/compare/main...ThomasMargnac:YouTube-operational-API:main) to add Docker support.
**UPDATE:** this fork has been removed.
Related to #132. | 1.0 | Propose a Docker version? - As found in [the forks of this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/network/members), could inspire myself of [this repository](https://github.com/Benjamin-Loison/YouTube-operational-API/compare/main...ThomasMargnac:YouTube-operational-API:main) to add Docker support.
**UPDATE:** this fork has been removed.
Related to #132. | priority | propose a docker version as found in could inspire myself of to add docker support update this fork has been removed related to | 1 |
182,261 | 6,668,425,077 | IssuesEvent | 2017-10-03 15:45:16 | dmusican/Elegit | https://api.github.com/repos/dmusican/Elegit | opened | Elegit uses a lot of memory. Is this necessary? | performance priority medium | Elegit uses a lot of memory, which slows down itself and the computers it runs on. We should profile this to see where all that memory is going and if it can be reduced. | 1.0 | Elegit uses a lot of memory. Is this necessary? - Elegit uses a lot of memory, which slows down itself and the computers it runs on. We should profile this to see where all that memory is going and if it can be reduced. | priority | elegit uses a lot of memory is this necessary elegit uses a lot of memory which slows down itself and the computers it runs on we should profile this to see where all that memory is going and if it can be reduced | 1 |
311,899 | 9,540,182,675 | IssuesEvent | 2019-04-30 18:51:02 | ansible/awx | https://api.github.com/repos/ansible/awx | opened | Investigate SCIM support | component:api priority:medium state:needs_devel type:enhancement | ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
SCIM allows social-auth users (SAML, etc.) to automatically be provisioned *and deprovisioned* when necessary, allowing for more secure usage of social auth.
django lib here: https://github.com/15five/django-scim2
| 1.0 | Investigate SCIM support - ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
SCIM allows social-auth users (SAML, etc.) to automatically be provisioned *and deprovisioned* when necessary, allowing for more secure usage of social auth.
django lib here: https://github.com/15five/django-scim2
| priority | investigate scim support issue type feature idea summary scim allows social auth users saml etc to automatically be provisioned and deprovisioned when necessary allowing for more secure usage of social auth django lib here | 1 |
726,677 | 25,007,475,778 | IssuesEvent | 2022-11-03 13:00:35 | bounswe/bounswe2022group2 | https://api.github.com/repos/bounswe/bounswe2022group2 | opened | Milestone 1: Review the Milestone | priority-medium status-new milestone | ### Issue Description
With the first customer presentation coming to an end, we have developed, deployed, and released the 0.1.0-alpha of our application, Learnify. As part of the **Milestone Group Review** part of the Milestone 1, we have planned to provide the **Review the Milestone** part as a collaborative work of three people: @xltvy, @egemenatikk, @surmelienes1
This issue is the main issue of the Review the Milestone part. Each responsible member will open their own sub-issues regarding the parts, they will work on and then link their sub-issues under this issue.
The Milestone 1 can be reached from [here](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md).
### Step Details
Steps that will be performed:
- [ ] Every responsible member will create their sub-issues.
- [ ] Every responsible will complete their parts.
- [ ] All material will be added to the [report](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md).
### Final Actions
After every sub-issue is resolved and our material is added to the milestone report, this issue will be closed.
### Responsible People
- [ ] Bahrican Yesil
- [X] Egemen Atik
- [X] Altay Acar
- [ ] Ezgi Aysel Batı
- [X] Muhammed Enes Sürmeli
- [ ] Onur Kömürcü
- [ ] Mehmet Batuhan Çelik
- [ ] Hasan Can Erol
- [ ] Ecenur Sezer
- [ ] Koray Tekin
- [ ] Mehmet Gökay Yıldız
### Deadline of the Issue
03.11.2022 - Thursday - 23.59
### Reviewer
Egemen Atik
### Deadline for the Review
04.11.2022 - Friday - 16.00
### Final To-Dos
- [ ] Every responsible shared an info about the sections/parts s/he will perform.
- [ ] Every responsible mentioned the sub-issue contains the details of his/her work, if a sub-issue is created.
- [ ] Every responsible mentioned this issue in the description of his/her sub-issue, if a sub-issue is created. | 1.0 | Milestone 1: Review the Milestone - ### Issue Description
With the first customer presentation coming to an end, we have developed, deployed, and released the 0.1.0-alpha of our application, Learnify. As part of the **Milestone Group Review** part of the Milestone 1, we have planned to provide the **Review the Milestone** part as a collaborative work of three people: @xltvy, @egemenatikk, @surmelienes1
This issue is the main issue of the Review the Milestone part. Each responsible member will open their own sub-issues regarding the parts, they will work on and then link their sub-issues under this issue.
The Milestone 1 can be reached from [here](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md).
### Step Details
Steps that will be performed:
- [ ] Every responsible member will create their sub-issues.
- [ ] Every responsible will complete their parts.
- [ ] All material will be added to the [report](https://github.com/bounswe/bounswe2022group2/blob/master/deliverables/CMPE451_Customer_Presentation_Milestone_1/deliverables.md).
### Final Actions
After every sub-issue is resolved and our material is added to the milestone report, this issue will be closed.
### Responsible People
- [ ] Bahrican Yesil
- [X] Egemen Atik
- [X] Altay Acar
- [ ] Ezgi Aysel Batı
- [X] Muhammed Enes Sürmeli
- [ ] Onur Kömürcü
- [ ] Mehmet Batuhan Çelik
- [ ] Hasan Can Erol
- [ ] Ecenur Sezer
- [ ] Koray Tekin
- [ ] Mehmet Gökay Yıldız
### Deadline of the Issue
03.11.2022 - Thursday - 23.59
### Reviewer
Egemen Atik
### Deadline for the Review
04.11.2022 - Friday - 16.00
### Final To-Dos
- [ ] Every responsible shared an info about the sections/parts s/he will perform.
- [ ] Every responsible mentioned the sub-issue contains the details of his/her work, if a sub-issue is created.
- [ ] Every responsible mentioned this issue in the description of his/her sub-issue, if a sub-issue is created. | priority | milestone review the milestone issue description with the first customer presentation coming to an end we have developed deployed and released the alpha of our application learnify as part of the milestone group review part of the milestone we have planned to provide the review the milestone part as a collaborative work of three people xltvy egemenatikk this issue is the main issue of the review the milestone part each responsible member will open their own sub issues regarding the parts they will work on and then link their sub issues under this issue the milestone can be reached from step details steps that will be performed every responsible member will create their sub issues every responsible will complete their parts all material will be added to the final actions after every sub issue is resolved and our material is added to the milestone report this issue will be closed responsible people bahrican yesil egemen atik altay acar ezgi aysel batı muhammed enes sürmeli onur kömürcü mehmet batuhan çelik hasan can erol ecenur sezer koray tekin mehmet gökay yıldız deadline of the issue thursday reviewer egemen atik deadline for the review friday final to dos every responsible shared an info about the sections parts s he will perform every responsible mentioned the sub issue contains the details of his her work if a sub issue is created every responsible mentioned this issue in the description of his her sub issue if a sub issue is created | 1 |
739,465 | 25,598,310,902 | IssuesEvent | 2022-12-01 17:54:12 | zowe/imperative | https://api.github.com/repos/zowe/imperative | closed | Unexpected prompting error: 'prepared.args[option.name][0].toUpperCase is not a function' | bug priority-medium | Prompting code causes an error when a parameter is numeric by trying to use the string 'toUpperCase function. Even if the parameter is entered as a string yargs converts it to numeric. | 1.0 | Unexpected prompting error: 'prepared.args[option.name][0].toUpperCase is not a function' - Prompting code causes an error when a parameter is numeric by trying to use the string 'toUpperCase function. Even if the parameter is entered as a string yargs converts it to numeric. | priority | unexpected prompting error prepared args touppercase is not a function prompting code causes an error when a parameter is numeric by trying to use the string touppercase function even if the parameter is entered as a string yargs converts it to numeric | 1 |
731,220 | 25,206,488,716 | IssuesEvent | 2022-11-13 18:48:34 | Cheos137/ArmorpointsPlusplus | https://api.github.com/repos/Cheos137/ArmorpointsPlusplus | closed | [Feature Request, INTERNAL]: defer config caching to forge | feature request minecraft: 1.16 future update priority: medium minecraft: 1.18 minecraft: 1.19 forge | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
remove caching config implementations on forge versions that support caching by themselves
### Solution
defer caching logic to forge to clean up code
### Alternatives
_No response_
### Additional Context
_No response_ | 1.0 | [Feature Request, INTERNAL]: defer config caching to forge - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
remove caching config implementations on forge versions that support caching by themselves
### Solution
defer caching logic to forge to clean up code
### Alternatives
_No response_
### Additional Context
_No response_ | priority | defer config caching to forge is there an existing issue for this i have searched the existing issues description remove caching config implementations on forge versions that support caching by themselves solution defer caching logic to forge to clean up code alternatives no response additional context no response | 1 |
100,404 | 4,087,156,659 | IssuesEvent | 2016-06-01 09:01:12 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | Passing method to macro causes seg fault | Medium Priority Semcheck | works when you only define the method:
```nimrod
import macros
type A = ref object
macro testMacro(a: expr): stmt =
result = newNimNode(nnkStmtList)
let methDef = a.findChild(it.kind == nnkMethodDef)
result.add methDef
method test(x: A) {.testMacro.} =
echo "hello"
test(A())
```
doesn't work if you don't define method:
```nimrod
import macros
type A = ref object
macro testMacro(a: expr): stmt =
discard
method test(x: A) {.testMacro.} =
echo "hello"
```
doesn't work if you add multiple statements to returned statement list:
```nimrod
import macros
type A = ref object
macro testMacro(a: expr): stmt =
result = newNimNode(nnkStmtList)
let methDef = a.findChild(it.kind == nnkMethodDef)
let methIdent = methDef.findChild(it.kind == nnkIdent)
result.add methDef
let procTyp = newNimNode(nnkProcTy).add(methDef.params).add(newEmptyNode())
let thisMethSym = genSym(nskLet, "thisMethod")
let identDefs = newIdentDefs(thisMethSym, procTyp, methIdent)
let letSection = newNimNode(nnkLetSection).add(identDefs)
result.add letSection
method test(x: A) {.testMacro.} =
echo "hello"
``` | 1.0 | Passing method to macro causes seg fault - works when you only define the method:
```nimrod
import macros
type A = ref object
macro testMacro(a: expr): stmt =
result = newNimNode(nnkStmtList)
let methDef = a.findChild(it.kind == nnkMethodDef)
result.add methDef
method test(x: A) {.testMacro.} =
echo "hello"
test(A())
```
doesn't work if you don't define method:
```nimrod
import macros
type A = ref object
macro testMacro(a: expr): stmt =
discard
method test(x: A) {.testMacro.} =
echo "hello"
```
doesn't work if you add multiple statements to returned statement list:
```nimrod
import macros
type A = ref object
macro testMacro(a: expr): stmt =
result = newNimNode(nnkStmtList)
let methDef = a.findChild(it.kind == nnkMethodDef)
let methIdent = methDef.findChild(it.kind == nnkIdent)
result.add methDef
let procTyp = newNimNode(nnkProcTy).add(methDef.params).add(newEmptyNode())
let thisMethSym = genSym(nskLet, "thisMethod")
let identDefs = newIdentDefs(thisMethSym, procTyp, methIdent)
let letSection = newNimNode(nnkLetSection).add(identDefs)
result.add letSection
method test(x: A) {.testMacro.} =
echo "hello"
``` | priority | passing method to macro causes seg fault works when you only define the method nimrod import macros type a ref object macro testmacro a expr stmt result newnimnode nnkstmtlist let methdef a findchild it kind nnkmethoddef result add methdef method test x a testmacro echo hello test a doesn t work if you don t define method nimrod import macros type a ref object macro testmacro a expr stmt discard method test x a testmacro echo hello doesn t work if you add multiple statements to returned statement list nimrod import macros type a ref object macro testmacro a expr stmt result newnimnode nnkstmtlist let methdef a findchild it kind nnkmethoddef let methident methdef findchild it kind nnkident result add methdef let proctyp newnimnode nnkprocty add methdef params add newemptynode let thismethsym gensym nsklet thismethod let identdefs newidentdefs thismethsym proctyp methident let letsection newnimnode nnkletsection add identdefs result add letsection method test x a testmacro echo hello | 1 |
247,512 | 7,919,426,332 | IssuesEvent | 2018-07-04 16:46:56 | Xceptance/neodymium-library | https://api.github.com/repos/Xceptance/neodymium-library | closed | Enhance NeodymiumCucumberRunner with @Browser annotation | Medium Priority doneInDevelop feature recheckedInDevelop | It would be nice to be able to add a @Browser annotation. This way it would reduce the duplication of test data tables in feature files. | 1.0 | Enhance NeodymiumCucumberRunner with @Browser annotation - It would be nice to be able to add a @Browser annotation. This way it would reduce the duplication of test data tables in feature files. | priority | enhance neodymiumcucumberrunner with browser annotation it would be nice to be able to add a browser annotation this way it would reduce the duplication of test data tables in feature files | 1 |
770,570 | 27,045,629,794 | IssuesEvent | 2023-02-13 09:33:57 | robotframework/robotframework | https://api.github.com/repos/robotframework/robotframework | closed | New `robot:flatten` tag for "flattening" keyword structures | enhancement priority: high effort: medium | Introduction
------------
With nested keyword structures, especially with recursive keyword calls and with WHILE and FOR loops, the log file can get hard do understand with many different nesting levels. Such nested structures also increase output.xml size, because even a simple keyword like
```robotframework
*** Keywords ***
Keyword
Log Robot
Log Framework
```
creates this much content:
```xml
<kw name="Keyword">
<kw name="Log" library="BuiltIn">
<arg>Robot</arg>
<doc>Logs the given message with the given level.</doc>
<msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.663"/>
</kw>
<kw name="Log" library="BuiltIn">
<arg>Framework</arg>
<doc>Logs the given message with the given level.</doc>
<msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/>
</kw>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/>
</kw>
```
We have had `--flattenkeywords` option for "flattening" such structures since RF 2.8.2 (#1551) and it works great. When a keyword is flattened, its child keywords and control structures are removed otherwise, but all their messages are preserved. It doesn't affect output.xml generated during execution, but flattening happens when output.xml files are parsed and can save huge amounts of memory. When `--flattenkeywords` is used with Rebot, it is possible to create a new flattened output.xml. For example, the above structure is converted into this if `Keyowrd` is flattened:
```
<kw name="Keyword">
<doc>_*Content flattened.*_</doc>
<msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg>
<msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/>
</kw>
```
Proposal
--------
Flattening works based on keyword names and based on tags, but it needs to be activated separately from the command line. This issue proposes adding new built-in tag `robot:flatten` that activates this behavior automatically. Removing top level keywords from tests and leaving only their messages doesn't make sense, so `robot:flatten` should be usable only as a keyword tag.
This functionality should work already during execution so that flattened keywords and control structures are never written to output.xml file. This avoid output.xml file growing big and is likely to also enhance the performance a bit.
Open questions
---------------
There are some open questions related to the design still:
- [ ] Should `start/end_keyword` listener methods be called with flattened keywords? I believe not, but I don't feel too strongly about this.
- [ ] Should we add *Content flattened* to keyword documentation like we do with `--flattenkeywords`? I believe not. There's the `robot:flatten` tag to indicate that anyway.
- [ ] Should `--flattenkeywords` be changed to work during execution as well? I believe yes, but that requires a separate issue.
- [ ] Should automatic TRACE level logging of arguments and return values of flattened keywords be disabled? I believe yes, but this isn't high priority.
Possible future enhancements
------------------------------
`--flattenkeywords` allows flattening WHILE or FOR loops or all loop iterations. Something like that would be convenient with built-in tags as well. We could consider something like `robot:flatten:while` and `robot:flatten:iteration` to support that, but I believe that's something that can wait until future versions.
Another alternative would be allowing tags with control structures as shown in the example below. This would require parser and model changes but could also have other use cases. That's certainly out of the scope of RF 6.1, though.
```robotframework
*** Keywords ***
Keyword
WHILE True
[Tags] robot:flatten
Nested
END
```
| 1.0 | New `robot:flatten` tag for "flattening" keyword structures - Introduction
------------
With nested keyword structures, especially with recursive keyword calls and with WHILE and FOR loops, the log file can get hard do understand with many different nesting levels. Such nested structures also increase output.xml size, because even a simple keyword like
```robotframework
*** Keywords ***
Keyword
Log Robot
Log Framework
```
creates this much content:
```xml
<kw name="Keyword">
<kw name="Log" library="BuiltIn">
<arg>Robot</arg>
<doc>Logs the given message with the given level.</doc>
<msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.663"/>
</kw>
<kw name="Log" library="BuiltIn">
<arg>Framework</arg>
<doc>Logs the given message with the given level.</doc>
<msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/>
</kw>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/>
</kw>
```
We have had `--flattenkeywords` option for "flattening" such structures since RF 2.8.2 (#1551) and it works great. When a keyword is flattened, its child keywords and control structures are removed otherwise, but all their messages are preserved. It doesn't affect output.xml generated during execution, but flattening happens when output.xml files are parsed and can save huge amounts of memory. When `--flattenkeywords` is used with Rebot, it is possible to create a new flattened output.xml. For example, the above structure is converted into this if `Keyowrd` is flattened:
```
<kw name="Keyword">
<doc>_*Content flattened.*_</doc>
<msg timestamp="20230103 20:06:36.663" level="INFO">Robot</msg>
<msg timestamp="20230103 20:06:36.663" level="INFO">Framework</msg>
<status status="PASS" starttime="20230103 20:06:36.663" endtime="20230103 20:06:36.664"/>
</kw>
```
Proposal
--------
Flattening works based on keyword names and based on tags, but it needs to be activated separately from the command line. This issue proposes adding new built-in tag `robot:flatten` that activates this behavior automatically. Removing top level keywords from tests and leaving only their messages doesn't make sense, so `robot:flatten` should be usable only as a keyword tag.
This functionality should work already during execution so that flattened keywords and control structures are never written to output.xml file. This avoid output.xml file growing big and is likely to also enhance the performance a bit.
Open questions
---------------
There are some open questions related to the design still:
- [ ] Should `start/end_keyword` listener methods be called with flattened keywords? I believe not, but I don't feel too strongly about this.
- [ ] Should we add *Content flattened* to keyword documentation like we do with `--flattenkeywords`? I believe not. There's the `robot:flatten` tag to indicate that anyway.
- [ ] Should `--flattenkeywords` be changed to work during execution as well? I believe yes, but that requires a separate issue.
- [ ] Should automatic TRACE level logging of arguments and return values of flattened keywords be disabled? I believe yes, but this isn't high priority.
Possible future enhancements
------------------------------
`--flattenkeywords` allows flattening WHILE or FOR loops or all loop iterations. Something like that would be convenient with built-in tags as well. We could consider something like `robot:flatten:while` and `robot:flatten:iteration` to support that, but I believe that's something that can wait until future versions.
Another alternative would be allowing tags with control structures as shown in the example below. This would require parser and model changes but could also have other use cases. That's certainly out of the scope of RF 6.1, though.
```robotframework
*** Keywords ***
Keyword
WHILE True
[Tags] robot:flatten
Nested
END
```
| priority | new robot flatten tag for flattening keyword structures introduction with nested keyword structures especially with recursive keyword calls and with while and for loops the log file can get hard do understand with many different nesting levels such nested structures also increase output xml size because even a simple keyword like robotframework keywords keyword log robot log framework creates this much content xml robot logs the given message with the given level robot framework logs the given message with the given level framework we have had flattenkeywords option for flattening such structures since rf and it works great when a keyword is flattened its child keywords and control structures are removed otherwise but all their messages are preserved it doesn t affect output xml generated during execution but flattening happens when output xml files are parsed and can save huge amounts of memory when flattenkeywords is used with rebot it is possible to create a new flattened output xml for example the above structure is converted into this if keyowrd is flattened content flattened robot framework proposal flattening works based on keyword names and based on tags but it needs to be activated separately from the command line this issue proposes adding new built in tag robot flatten that activates this behavior automatically removing top level keywords from tests and leaving only their messages doesn t make sense so robot flatten should be usable only as a keyword tag this functionality should work already during execution so that flattened keywords and control structures are never written to output xml file this avoid output xml file growing big and is likely to also enhance the performance a bit open questions there are some open questions related to the design still should start end keyword listener methods be called with flattened keywords i believe not but i don t feel too strongly about this should we add content flattened to keyword documentation like we do with flattenkeywords i believe not there s the robot flatten tag to indicate that anyway should flattenkeywords be changed to work during execution as well i believe yes but that requires a separate issue should automatic trace level logging of arguments and return values of flattened keywords be disabled i believe yes but this isn t high priority possible future enhancements flattenkeywords allows flattening while or for loops or all loop iterations something like that would be convenient with built in tags as well we could consider something like robot flatten while and robot flatten iteration to support that but i believe that s something that can wait until future versions another alternative would be allowing tags with control structures as shown in the example below this would require parser and model changes but could also have other use cases that s certainly out of the scope of rf though robotframework keywords keyword while true robot flatten nested end | 1 |
759,956 | 26,620,080,245 | IssuesEvent | 2023-01-24 10:32:46 | EddieHubCommunity/LinkFree | https://api.github.com/repos/EddieHubCommunity/LinkFree | closed | [DOCS] Wrong links in the FAQ | 🏁 status: ready for dev 🟨 priority: medium 🛠 goal: fix good first issue 📄 aspect: text | ### Description
There are various broken links on the FAQ page (such as the single-user mode or Hacktoberfest pages) which should be fixed immediately.
_Steps to contribute_
1. Fork the repository [here](https://github.com/EddieHubCommunity/LinkFree/fork).
2. Create a new branch named `fix_faq_links`.
3. Open the `faqs.mdx` file, which is located at `pages/docs/`.
4. In line 9, change the single-user mode link to `advanced/single-user-mode`. We do not want to include `docs` as otherwise, it would lead to `docs/docs/...` which is invalid.
5. In line 15, change the Hacktoberfest link to `contributing/hacktoberfest`.
6. The Discord invitation link seems to be broken as well. Change it to `http://discord.eddiehub.org` (located in line 21).
7. Commit and push your changes.
8. Create a [pull request](https://github.com/EddieHubCommunity/LinkFree/compare).
9. You're done now! 🎉 You just have to wait until the maintainers check your PR.
If you need any help, let us know about it and we'll gladly help you. 🙂
### Screenshots
_No response_
### Additional information
This is a good first issue. Please leave it open to first-time contributors and beginners.
| 1.0 | [DOCS] Wrong links in the FAQ - ### Description
There are various broken links on the FAQ page (such as the single-user mode or Hacktoberfest pages) which should be fixed immediately.
_Steps to contribute_
1. Fork the repository [here](https://github.com/EddieHubCommunity/LinkFree/fork).
2. Create a new branch named `fix_faq_links`.
3. Open the `faqs.mdx` file, which is located at `pages/docs/`.
4. In line 9, change the single-user mode link to `advanced/single-user-mode`. We do not want to include `docs` as otherwise, it would lead to `docs/docs/...` which is invalid.
5. In line 15, change the Hacktoberfest link to `contributing/hacktoberfest`.
6. The Discord invitation link seems to be broken as well. Change it to `http://discord.eddiehub.org` (located in line 21).
7. Commit and push your changes.
8. Create a [pull request](https://github.com/EddieHubCommunity/LinkFree/compare).
9. You're done now! 🎉 You just have to wait until the maintainers check your PR.
If you need any help, let us know about it and we'll gladly help you. 🙂
### Screenshots
_No response_
### Additional information
This is a good first issue. Please leave it open to first-time contributors and beginners.
| priority | wrong links in the faq description there are various broken links on the faq page such as the single user mode or hacktoberfest pages which should be fixed immediately steps to contribute fork the repository create a new branch named fix faq links open the faqs mdx file which is located at pages docs in line change the single user mode link to advanced single user mode we do not want to include docs as otherwise it would lead to docs docs which is invalid in line change the hacktoberfest link to contributing hacktoberfest the discord invitation link seems to be broken as well change it to located in line commit and push your changes create a you re done now 🎉 you just have to wait until the maintainers check your pr if you need any help let us know about it and we ll gladly help you 🙂 screenshots no response additional information this is a good first issue please leave it open to first time contributors and beginners | 1 |
533,377 | 15,589,876,073 | IssuesEvent | 2021-03-18 08:39:14 | sopra-fs21-group-02/server | https://api.github.com/repos/sopra-fs21-group-02/server | opened | Implement filtering for the map view | area:map priority:medium task | Make the map view consider the currently configured filters and only display elements that are wanted according to the filters.
## Estimate
6h
## User Story
This task belongs to user story #3 | 1.0 | Implement filtering for the map view - Make the map view consider the currently configured filters and only display elements that are wanted according to the filters.
## Estimate
6h
## User Story
This task belongs to user story #3 | priority | implement filtering for the map view make the map view consider the currently configured filters and only display elements that are wanted according to the filters estimate user story this task belongs to user story | 1 |
253,215 | 8,052,926,704 | IssuesEvent | 2018-08-01 20:56:20 | medic/medic-webapp | https://api.github.com/repos/medic/medic-webapp | opened | Update the on/off handling to mark an individual as muted | Needs Triage Priority: 2 - Medium Type: Feature | This request is a continuation of the request in ticket #4767. It is not an immediate need, but if it is not too much extra work beyond 4767, it probably makes sense to do at the same time.
Full details and additional screenshots can be found in the design spec: https://docs.google.com/document/d/1we_Qu1B4x8PuAi9pxiIu4MeXtdH1kY9W29kMidHThsc/edit?usp=sharing
**Summary of individual muting**
- We would like to make sure that the family muting is done in a way that can be logically and technically extended to individuals, even if that isn't actually implemented now.
- Individuals can already be muted from their profiles using the features built for the “OFF” form, but we need to make an extension of that to also mark the patient muted and make the UI changes to their profile to reflect the “Muted” status
- Generally speaking, there may be muted individuals in an un-muted household. However, there may NOT be un-muted individuals in a muted household.
- If a user wants to unmute an individual in a muted household, the whole household (including all family members) will need to be unmuted. The user may then individually mute any other household members as needed.
- Muting schedules attached to a CHW as a person in the app (in the case that they are fired, quit, or retire) will be technically available but considering the programmatic implications of this (re-assigning, performance metrics) is outside the scope of this MVP.
**UI changes to the main People page list: none**
- If an individual is muted from an individual profile form, then the greater family remains unmuted, and there is no change to that family’s appearance on the main People list

**UI changes to the family profile**
- If an individual is muted from an individual on/off form, then the greater family remains unmuted, and there is no change to that family’s appearance (family icons stays pink, no “Muted” text)
- There should be a change made to the “Family Members” condition card. Any individuals that are muted within an unmuted family should be specially styled in this card so that they stand out from active unmuted family members.
- Suggestion: The person’s name becomes grey colored, and instead of (or in addition to) the age on the second line, we also display the word “Muted”. See example above where Jon Bass is muted.

**UI changes to profiles**
- Very similar to profile styling for death
- The person/place icon becomes grey (#A7A9AC)
- And we display the text “Muted” on the second line below the person/place name
- Any forms that were previously available on the profile remain available
- Instead of the “Mute person/place” form, the user now sees “Unmute person/place” | 1.0 | Update the on/off handling to mark an individual as muted - This request is a continuation of the request in ticket #4767. It is not an immediate need, but if it is not too much extra work beyond 4767, it probably makes sense to do at the same time.
Full details and additional screenshots can be found in the design spec: https://docs.google.com/document/d/1we_Qu1B4x8PuAi9pxiIu4MeXtdH1kY9W29kMidHThsc/edit?usp=sharing
**Summary of individual muting**
- We would like to make sure that the family muting is done in a way that can be logically and technically extended to individuals, even if that isn't actually implemented now.
- Individuals can already be muted from their profiles using the features built for the “OFF” form, but we need to make an extension of that to also mark the patient muted and make the UI changes to their profile to reflect the “Muted” status
- Generally speaking, there may be muted individuals in an un-muted household. However, there may NOT be un-muted individuals in a muted household.
- If a user wants to unmute an individual in a muted household, the whole household (including all family members) will need to be unmuted. The user may then individually mute any other household members as needed.
- Muting schedules attached to a CHW as a person in the app (in the case that they are fired, quit, or retire) will be technically available but considering the programmatic implications of this (re-assigning, performance metrics) is outside the scope of this MVP.
**UI changes to the main People page list: none**
- If an individual is muted from an individual profile form, then the greater family remains unmuted, and there is no change to that family’s appearance on the main People list

**UI changes to the family profile**
- If an individual is muted from an individual on/off form, then the greater family remains unmuted, and there is no change to that family’s appearance (family icons stays pink, no “Muted” text)
- There should be a change made to the “Family Members” condition card. Any individuals that are muted within an unmuted family should be specially styled in this card so that they stand out from active unmuted family members.
- Suggestion: The person’s name becomes grey colored, and instead of (or in addition to) the age on the second line, we also display the word “Muted”. See example above where Jon Bass is muted.

**UI changes to profiles**
- Very similar to profile styling for death
- The person/place icon becomes grey (#A7A9AC)
- And we display the text “Muted” on the second line below the person/place name
- Any forms that were previously available on the profile remain available
- Instead of the “Mute person/place” form, the user now sees “Unmute person/place” | priority | update the on off handling to mark an individual as muted this request is a continuation of the request in ticket it is not an immediate need but if it is not too much extra work beyond it probably makes sense to do at the same time full details and additional screenshots can be found in the design spec summary of individual muting we would like to make sure that the family muting is done in a way that can be logically and technically extended to individuals even if that isn t actually implemented now individuals can already be muted from their profiles using the features built for the “off” form but we need to make an extension of that to also mark the patient muted and make the ui changes to their profile to reflect the “muted” status generally speaking there may be muted individuals in an un muted household however there may not be un muted individuals in a muted household if a user wants to unmute an individual in a muted household the whole household including all family members will need to be unmuted the user may then individually mute any other household members as needed muting schedules attached to a chw as a person in the app in the case that they are fired quit or retire will be technically available but considering the programmatic implications of this re assigning performance metrics is outside the scope of this mvp ui changes to the main people page list none if an individual is muted from an individual profile form then the greater family remains unmuted and there is no change to that family’s appearance on the main people list ui changes to the family profile if an individual is muted from an individual on off form then the greater family remains unmuted and there is no change to that family’s appearance family icons stays pink no “muted” text there should be a change made to the “family members” condition card any individuals that are muted within an unmuted family should be specially styled in this card so that they stand out from active unmuted family members suggestion the person’s name becomes grey colored and instead of or in addition to the age on the second line we also display the word “muted” see example above where jon bass is muted ui changes to profiles very similar to profile styling for death the person place icon becomes grey and we display the text “muted” on the second line below the person place name any forms that were previously available on the profile remain available instead of the “mute person place” form the user now sees “unmute person place” | 1 |
16,749 | 2,615,122,965 | IssuesEvent | 2015-03-01 05:50:24 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | opened | Need an example for google shopping API....Please | auto-migrated Priority-Medium Type-Sample | ```
Which Google API and version (e.g. Google Calendar Data API version 2)?1.5
What format (e.g. JSON, Atom)?Json
What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?ClientLogin
Java environment (e.g. Java 6, Android 2.3, App Engine)?android 2.3
External references, such as API reference guide?GoogleShopping API
Please provide any additional information below.Need an example for searching
an API using android ..Please help me out
```
Original issue reported on code.google.com by `saikiran...@gmail.com` on 13 Oct 2011 at 11:05 | 1.0 | Need an example for google shopping API....Please - ```
Which Google API and version (e.g. Google Calendar Data API version 2)?1.5
What format (e.g. JSON, Atom)?Json
What Authentation (e.g. OAuth, OAuth 2, ClientLogin)?ClientLogin
Java environment (e.g. Java 6, Android 2.3, App Engine)?android 2.3
External references, such as API reference guide?GoogleShopping API
Please provide any additional information below.Need an example for searching
an API using android ..Please help me out
```
Original issue reported on code.google.com by `saikiran...@gmail.com` on 13 Oct 2011 at 11:05 | priority | need an example for google shopping api please which google api and version e g google calendar data api version what format e g json atom json what authentation e g oauth oauth clientlogin clientlogin java environment e g java android app engine android external references such as api reference guide googleshopping api please provide any additional information below need an example for searching an api using android please help me out original issue reported on code google com by saikiran gmail com on oct at | 1 |
358,687 | 10,631,312,269 | IssuesEvent | 2019-10-15 08:05:57 | facebookresearch/nevergrad | https://api.github.com/repos/facebookresearch/nevergrad | opened | Consistent population-based algorithms | Difficulty: High Priority: Medium Type: Enhancement | Population based algorithms are written in a variety of ways while some common code could be extracted.
Also, for now they don't allow asking more than the population, which can be cumbersome (especially if some evaluation died without providing a value: in this case an individual of the population will always be ignored) | 1.0 | Consistent population-based algorithms - Population based algorithms are written in a variety of ways while some common code could be extracted.
Also, for now they don't allow asking more than the population, which can be cumbersome (especially if some evaluation died without providing a value: in this case an individual of the population will always be ignored) | priority | consistent population based algorithms population based algorithms are written in a variety of ways while some common code could be extracted also for now they don t allow asking more than the population which can be cumbersome especially if some evaluation died without providing a value in this case an individual of the population will always be ignored | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.