Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
28,246 | 2,700,668,000 | IssuesEvent | 2015-04-04 12:34:05 | cs2103jan2015-f13-1j/main | https://api.github.com/repos/cs2103jan2015-f13-1j/main | opened | Implement simple functionality for right sidebar | priority.medium type.enhancement | Depending on toggle status of right sidebar, display floating tasks, deadline tasks and/or timed tasks (using search functionality for the time being).
Depending on the status of the right radio button, display tasks due today, tomorrow, this week or all time (floating tasks are always displayed) | 1.0 | Implement simple functionality for right sidebar - Depending on toggle status of right sidebar, display floating tasks, deadline tasks and/or timed tasks (using search functionality for the time being).
Depending on the status of the right radio button, display tasks due today, tomorrow, this week or all time (floating tasks are always displayed) | priority | implement simple functionality for right sidebar depending on toggle status of right sidebar display floating tasks deadline tasks and or timed tasks using search functionality for the time being depending on the status of the right radio button display tasks due today tomorrow this week or all time floating tasks are always displayed | 1 |
464,908 | 13,348,000,794 | IssuesEvent | 2020-08-29 16:20:37 | debops/debops | https://api.github.com/repos/debops/debops | closed | [debops.users] My user isn't added to the admins group by ansible when running a customized playbook | meta priority: medium | Preface: I'm new to Ansible/debops, and I'm doing a bit of a loopy setup, so my setup is confusing. Any help you can offer is much appreciated!
I am setting up a custom ansible playbook that I can use for my personal server infrastructure. I am re-using most of debops as the basis for my playbooks to try to leverage the existing conventions defined to avoid re-inventing the wheel.
That said, I'm not using debops as designed and I'm also using ansible-pull, which is changing things up a bit for me.
I have a mostly working playbook ready to serve as my basic setup that I can then build applications on top of. However, I have one major issue:
When I run my playbook, the user account it creates ("devin") is not being added to the "admins" group. If I run `adduser devin admins` after running ansible-pull, everything works fine. But I can't figure out how to make it work.
I don't observe any difference when running on a fresh host vs running the playbook again on the same host - should I expect a difference?
Any advice on a) what might be going wrong or b) best strategies for debugging? As I said, I'm new to ansible so any tips would help. I got as far as editing /root/.ansible/collections/ansible_collections/debops/debops/roles/users/tasks/main.yml to see if I could figure it out but didn't make any progress after 30 minutes of trying things blindly.
Code I'm using https://github.com/devvmh/ansible-playbook-core/tree/devel
Specifically, this code is not functioning as expected:
```
- role: debops.debops.system_groups
- role: debops.debops.users
vars:
- users__accounts:
- name: 'devin'
- groups: ['admins']
```
Output I get:
```
TASK [debops.debops.users : Manage additional UNIX groups for UNIX accounts] *********************************************************
skipping: [localhost] => (item={u'state': u'present', u'gecos': u'', u'name': u'devin', u'groups': u''})
```
If I remove the when block from this step in tasks/main.yml, I get this output, but it still doesn't work:
```
TASK [debops.debops.users : Manage additional UNIX groups for UNIX accounts] *********************************************************
ok: [localhost] => (item={u'state': u'present', u'gecos': u'', u'name': u'devin', u'groups': u''})
```
Thanks for any help or tips you can provide! | 1.0 | [debops.users] My user isn't added to the admins group by ansible when running a customized playbook - Preface: I'm new to Ansible/debops, and I'm doing a bit of a loopy setup, so my setup is confusing. Any help you can offer is much appreciated!
I am setting up a custom ansible playbook that I can use for my personal server infrastructure. I am re-using most of debops as the basis for my playbooks to try to leverage the existing conventions defined to avoid re-inventing the wheel.
That said, I'm not using debops as designed and I'm also using ansible-pull, which is changing things up a bit for me.
I have a mostly working playbook ready to serve as my basic setup that I can then build applications on top of. However, I have one major issue:
When I run my playbook, the user account it creates ("devin") is not being added to the "admins" group. If I run `adduser devin admins` after running ansible-pull, everything works fine. But I can't figure out how to make it work.
I don't observe any difference when running on a fresh host vs running the playbook again on the same host - should I expect a difference?
Any advice on a) what might be going wrong or b) best strategies for debugging? As I said, I'm new to ansible so any tips would help. I got as far as editing /root/.ansible/collections/ansible_collections/debops/debops/roles/users/tasks/main.yml to see if I could figure it out but didn't make any progress after 30 minutes of trying things blindly.
Code I'm using https://github.com/devvmh/ansible-playbook-core/tree/devel
Specifically, this code is not functioning as expected:
```
- role: debops.debops.system_groups
- role: debops.debops.users
vars:
- users__accounts:
- name: 'devin'
- groups: ['admins']
```
Output I get:
```
TASK [debops.debops.users : Manage additional UNIX groups for UNIX accounts] *********************************************************
skipping: [localhost] => (item={u'state': u'present', u'gecos': u'', u'name': u'devin', u'groups': u''})
```
If I remove the when block from this step in tasks/main.yml, I get this output, but it still doesn't work:
```
TASK [debops.debops.users : Manage additional UNIX groups for UNIX accounts] *********************************************************
ok: [localhost] => (item={u'state': u'present', u'gecos': u'', u'name': u'devin', u'groups': u''})
```
Thanks for any help or tips you can provide! | priority | my user isn t added to the admins group by ansible when running a customized playbook preface i m new to ansible debops and i m doing a bit of a loopy setup so my setup is confusing any help you can offer is much appreciated i am setting up a custom ansible playbook that i can use for my personal server infrastructure i am re using most of debops as the basis for my playbooks to try to leverage the existing conventions defined to avoid re inventing the wheel that said i m not using debops as designed and i m also using ansible pull which is changing things up a bit for me i have a mostly working playbook ready to serve as my basic setup that i can then build applications on top of however i have one major issue when i run my playbook the user account it creates devin is not being added to the admins group if i run adduser devin admins after running ansible pull everything works fine but i can t figure out how to make it work i don t observe any difference when running on a fresh host vs running the playbook again on the same host should i expect a difference any advice on a what might be going wrong or b best strategies for debugging as i said i m new to ansible so any tips would help i got as far as editing root ansible collections ansible collections debops debops roles users tasks main yml to see if i could figure it out but didn t make any progress after minutes of trying things blindly code i m using specifically this code is not functioning as expected role debops debops system groups role debops debops users vars users accounts name devin groups output i get task skipping item u state u present u gecos u u name u devin u groups u if i remove the when block from this step in tasks main yml i get this output but it still doesn t work task ok item u state u present u gecos u u name u devin u groups u thanks for any help or tips you can provide | 1 |
3,806 | 2,540,574,898 | IssuesEvent | 2015-01-27 22:43:55 | SiCKRAGETV/sickrage-issues | https://api.github.com/repos/SiCKRAGETV/sickrage-issues | closed | POSTPROCESSER stops running | 1: Bug / issue 2: Medium Priority 3: Unconfirmed branch: master | Branch: Master
Commit Hash: (401cb666016e45b42fe675bafdf7908ad6c2b9bb)
OS: Win 7 pro 64bit
Python Version: 2.7.3 (default, Apr 10 2012, 23:24:47) [MSC v.1500 64 bit (AMD64)
I'm having a problem with automatic post processing. When I first start sickrage the post processing will work fine, but at some point it stops. Usually a day or so after starting sickrage. During the time when this is happening there are no "POSTPROCESSER" entries in the log. If I restart sickrage, post processing will resume and the backlog of completed downloads will process. Generally it will also process new downloads for about a day, then it stops again.
I'm not sure when this started but everything was running fine for months before I had a problem. I have auto updates turned on and the problem may have coincided with an update.
I have the post processing set to run every 10 minutes. Below is a log for an hour period with no POSTPROCESSER entries. I'm not sure if it will be of any help but it at least proves that post processing isn't running.
http://pastebin.com/eHHN2RbV | 1.0 | POSTPROCESSER stops running - Branch: Master
Commit Hash: (401cb666016e45b42fe675bafdf7908ad6c2b9bb)
OS: Win 7 pro 64bit
Python Version: 2.7.3 (default, Apr 10 2012, 23:24:47) [MSC v.1500 64 bit (AMD64)
I'm having a problem with automatic post processing. When I first start sickrage the post processing will work fine, but at some point it stops. Usually a day or so after starting sickrage. During the time when this is happening there are no "POSTPROCESSER" entries in the log. If I restart sickrage, post processing will resume and the backlog of completed downloads will process. Generally it will also process new downloads for about a day, then it stops again.
I'm not sure when this started but everything was running fine for months before I had a problem. I have auto updates turned on and the problem may have coincided with an update.
I have the post processing set to run every 10 minutes. Below is a log for an hour period with no POSTPROCESSER entries. I'm not sure if it will be of any help but it at least proves that post processing isn't running.
http://pastebin.com/eHHN2RbV | priority | postprocesser stops running branch master commit hash os win pro python version default apr msc v bit i m having a problem with automatic post processing when i first start sickrage the post processing will work fine but at some point it stops usually a day or so after starting sickrage during the time when this is happening there are no postprocesser entries in the log if i restart sickrage post processing will resume and the backlog of completed downloads will process generally it will also process new downloads for about a day then it stops again i m not sure when this started but everything was running fine for months before i had a problem i have auto updates turned on and the problem may have coincided with an update i have the post processing set to run every minutes below is a log for an hour period with no postprocesser entries i m not sure if it will be of any help but it at least proves that post processing isn t running | 1 |
788,431 | 27,752,853,566 | IssuesEvent | 2023-03-15 22:27:05 | unstructuredstudio/zubhub | https://api.github.com/repos/unstructuredstudio/zubhub | closed | Clicking on a username in the comment box takes you to the current user's profile page | bug good first issue medium priority | **Describe the bug**
If you click on a username in the comment box, it will take you to the current user's profile page. Ideally, it should take you to the user's profile whose avatar or username is clicked.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit the following project: https://zubhub.unstructured.studio/projects/61045c12-6eae-4678-abda-4f62161a2a8a
2. Scroll to the comments section
3. Click on the username or avatar in a comment box
4. Notice that you are redirected to your own profile page
**Expected behavior**
Ideally, you were redirected to the clicked user's profile page.

| 1.0 | Clicking on a username in the comment box takes you to the current user's profile page - **Describe the bug**
If you click on a username in the comment box, it will take you to the current user's profile page. Ideally, it should take you to the user's profile whose avatar or username is clicked.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit the following project: https://zubhub.unstructured.studio/projects/61045c12-6eae-4678-abda-4f62161a2a8a
2. Scroll to the comments section
3. Click on the username or avatar in a comment box
4. Notice that you are redirected to your own profile page
**Expected behavior**
Ideally, you were redirected to the clicked user's profile page.

| priority | clicking on a username in the comment box takes you to the current user s profile page describe the bug if you click on a username in the comment box it will take you to the current user s profile page ideally it should take you to the user s profile whose avatar or username is clicked to reproduce steps to reproduce the behavior visit the following project scroll to the comments section click on the username or avatar in a comment box notice that you are redirected to your own profile page expected behavior ideally you were redirected to the clicked user s profile page | 1 |
173,566 | 6,527,887,711 | IssuesEvent | 2017-08-30 03:57:53 | orange-alliance/the-orange-alliance | https://api.github.com/repos/orange-alliance/the-orange-alliance | closed | Add year to Season on Team Page | bug Medium Priority | To have constancy the page should show season as XXXX/YYYY (Example: 2016/2017) | 1.0 | Add year to Season on Team Page - To have constancy the page should show season as XXXX/YYYY (Example: 2016/2017) | priority | add year to season on team page to have constancy the page should show season as xxxx yyyy example | 1 |
118,427 | 4,744,923,981 | IssuesEvent | 2016-10-21 04:05:27 | CovertJaguar/Railcraft | https://api.github.com/repos/CovertJaguar/Railcraft | closed | Redstone Condition Incorrect on RF Loader | bug priority-medium | Description:
The "complete" redstone condition for the RF loader should be "when cart is full", but instead it's "Process until cart is empty".
Tested With:
RailCraft: 1.10.2-10.0.0-beta-3
Forge: 1.10.2-12.18.2.2099 | 1.0 | Redstone Condition Incorrect on RF Loader - Description:
The "complete" redstone condition for the RF loader should be "when cart is full", but instead it's "Process until cart is empty".
Tested With:
RailCraft: 1.10.2-10.0.0-beta-3
Forge: 1.10.2-12.18.2.2099 | priority | redstone condition incorrect on rf loader description the complete redstone condition for the rf loader should be when cart is full but instead it s process until cart is empty tested with railcraft beta forge | 1 |
395,966 | 11,699,295,092 | IssuesEvent | 2020-03-06 15:23:41 | luna/enso | https://api.github.com/repos/luna/enso | closed | Desugar Operators to Functions | Category: Compiler Category: Core Change: Breaking Difficulty: Core Contributor Priority: Medium Type: Enhancement | ### Summary
Operators in Enso are just syntactic sugar for functions, but analysis passes don't need to know about that.
### Value
Analysis passes don't want to have to deal with syntax sugar, and at this stage in the compiler implementation this is the only sugar we support.
### Specification
- [x] Implement a pass that transforms all uses of operators into standard prefix function applications.
- [x] The output of this pass is not analysis, but just a graph without operator sugar in it.
- [x] Implement operators as standard builting functions in the interpreter to support this.
- [ ] Implement an extensible optimisation pass that treats known, fully-saturated functions with known arity specially such that they can be compiled to specific nodes. This should contain the node constructor function in the metadata from the pass, and map from arity to function name to the necessary data.
### Acceptance Criteria & Test Cases
- Operators can be successfully desugared to functions. | 1.0 | Desugar Operators to Functions - ### Summary
Operators in Enso are just syntactic sugar for functions, but analysis passes don't need to know about that.
### Value
Analysis passes don't want to have to deal with syntax sugar, and at this stage in the compiler implementation this is the only sugar we support.
### Specification
- [x] Implement a pass that transforms all uses of operators into standard prefix function applications.
- [x] The output of this pass is not analysis, but just a graph without operator sugar in it.
- [x] Implement operators as standard builting functions in the interpreter to support this.
- [ ] Implement an extensible optimisation pass that treats known, fully-saturated functions with known arity specially such that they can be compiled to specific nodes. This should contain the node constructor function in the metadata from the pass, and map from arity to function name to the necessary data.
### Acceptance Criteria & Test Cases
- Operators can be successfully desugared to functions. | priority | desugar operators to functions summary operators in enso are just syntactic sugar for functions but analysis passes don t need to know about that value analysis passes don t want to have to deal with syntax sugar and at this stage in the compiler implementation this is the only sugar we support specification implement a pass that transforms all uses of operators into standard prefix function applications the output of this pass is not analysis but just a graph without operator sugar in it implement operators as standard builting functions in the interpreter to support this implement an extensible optimisation pass that treats known fully saturated functions with known arity specially such that they can be compiled to specific nodes this should contain the node constructor function in the metadata from the pass and map from arity to function name to the necessary data acceptance criteria test cases operators can be successfully desugared to functions | 1 |
642,790 | 20,913,520,985 | IssuesEvent | 2022-03-24 11:23:52 | netdata/netdata-cloud | https://api.github.com/repos/netdata/netdata-cloud | closed | [BUG] chart families not showing properly in overview screen | bug priority/medium visualizations-team | <!---
If you are a member of the Netdata organization, add the label 'internal submit'.
-->
**Describe the bug**
I have a room with one node. This node about 20 anomalies collector jobs running on it and so has lots of anomalies contexts in the menu on the right.
When i am in the node itself i see correct menus:

But for some reason when i go to the overview for the room i lose those menu's and it has for some reason sort of flattened out all the "System Overview" sections into their own menus for some reason.

**To Reproduce**
Create a agent with a lot of non standard collectors and/or multiple jobs per collector. Chanage the priority of those jobs to be 80-90 and so appear at top of menu list.
Create a space and room with just that node. Look at the differences between the overview and the node view of the dashboard itself.
**Expected behavior**
I could see all the menu sections in the overview screen just like i do on the node screen.
**Screenshots**
As above
**Additional context**
I'm happy to invite anyone to my room if easier to debug that way as might be a tricky one to recreate.
| 1.0 | [BUG] chart families not showing properly in overview screen - <!---
If you are a member of the Netdata organization, add the label 'internal submit'.
-->
**Describe the bug**
I have a room with one node. This node about 20 anomalies collector jobs running on it and so has lots of anomalies contexts in the menu on the right.
When i am in the node itself i see correct menus:

But for some reason when i go to the overview for the room i lose those menu's and it has for some reason sort of flattened out all the "System Overview" sections into their own menus for some reason.

**To Reproduce**
Create a agent with a lot of non standard collectors and/or multiple jobs per collector. Chanage the priority of those jobs to be 80-90 and so appear at top of menu list.
Create a space and room with just that node. Look at the differences between the overview and the node view of the dashboard itself.
**Expected behavior**
I could see all the menu sections in the overview screen just like i do on the node screen.
**Screenshots**
As above
**Additional context**
I'm happy to invite anyone to my room if easier to debug that way as might be a tricky one to recreate.
| priority | chart families not showing properly in overview screen if you are a member of the netdata organization add the label internal submit describe the bug i have a room with one node this node about anomalies collector jobs running on it and so has lots of anomalies contexts in the menu on the right when i am in the node itself i see correct menus but for some reason when i go to the overview for the room i lose those menu s and it has for some reason sort of flattened out all the system overview sections into their own menus for some reason to reproduce create a agent with a lot of non standard collectors and or multiple jobs per collector chanage the priority of those jobs to be and so appear at top of menu list create a space and room with just that node look at the differences between the overview and the node view of the dashboard itself expected behavior i could see all the menu sections in the overview screen just like i do on the node screen screenshots as above additional context i m happy to invite anyone to my room if easier to debug that way as might be a tricky one to recreate | 1 |
705,287 | 24,229,601,869 | IssuesEvent | 2022-09-26 17:02:39 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Import schema qualification fix | kind/bug area/ysql priority/medium 2.12 Backport Required 2.14 Backport Required | Jira Link: [DB-3617](https://yugabyte.atlassian.net/browse/DB-3617)
### Description
Import schema qualification fix | 1.0 | [YSQL] Import schema qualification fix - Jira Link: [DB-3617](https://yugabyte.atlassian.net/browse/DB-3617)
### Description
Import schema qualification fix | priority | import schema qualification fix jira link description import schema qualification fix | 1 |
431,302 | 12,476,988,360 | IssuesEvent | 2020-05-29 14:22:24 | ansible/ansible-lint | https://api.github.com/repos/ansible/ansible-lint | closed | auto-detect modules with collection layouts | help wanted needs_implementation priority/medium status/new type/enhancement type/proposal | Ansible-lint ability to just work without any extra tuning can be improved by making it automatically define `ANSIBLE_LIBRARY=plugins/modules` when the variable is not already defined.
This would follow the [official collection repository layout](https://docs.ansible.com/ansible/latest/dev_guide/developing_collections.html) and make ansible-lint more likely work without extra configuration.
This feature should be enabled only with auto-detection mode because it needs to know what is the repository root location. That is because the tool could be called from any subdirectory and we do expect to give the same kind of results.
When implement this feature should allow people to remove extra code added to files like `tox.ini` or `.pre-commit-config.yaml` that define ANSIBLE_LIBRARY in order to be able perform the linting. | 1.0 | auto-detect modules with collection layouts - Ansible-lint ability to just work without any extra tuning can be improved by making it automatically define `ANSIBLE_LIBRARY=plugins/modules` when the variable is not already defined.
This would follow the [official collection repository layout](https://docs.ansible.com/ansible/latest/dev_guide/developing_collections.html) and make ansible-lint more likely work without extra configuration.
This feature should be enabled only with auto-detection mode because it needs to know what is the repository root location. That is because the tool could be called from any subdirectory and we do expect to give the same kind of results.
When implement this feature should allow people to remove extra code added to files like `tox.ini` or `.pre-commit-config.yaml` that define ANSIBLE_LIBRARY in order to be able perform the linting. | priority | auto detect modules with collection layouts ansible lint ability to just work without any extra tuning can be improved by making it automatically define ansible library plugins modules when the variable is not already defined this would follow the and make ansible lint more likely work without extra configuration this feature should be enabled only with auto detection mode because it needs to know what is the repository root location that is because the tool could be called from any subdirectory and we do expect to give the same kind of results when implement this feature should allow people to remove extra code added to files like tox ini or pre commit config yaml that define ansible library in order to be able perform the linting | 1 |
262,874 | 8,272,568,746 | IssuesEvent | 2018-09-16 21:35:13 | minio/minio | https://api.github.com/repos/minio/minio | closed | [Distributed] Disks are never formatted if instances mixed with binary and docker hosts | priority: medium won't fix | I have set up 4 minio servers in 4 different physical locations. 3 Are running the native linux binary and one is running in a docker container.
**All instances are using this exact version:**
- Version: 2018-08-21T00:37:20Z
- Release-Tag: RELEASE.2018-08-21T00-37-20Z
- Commit-ID: 2d84b02bc429ad059a060001bccd712491e6e6a3
I am able to ping and curl all instances from all locations and even from inside the container. The answer from CURL is the expected XML with the message "XMinioServerNotInitialized" from all 4 nodes
I have all 4 sites open via my browser so I know they have the same auth info and that they are running correctly.
But it seems they never format the disks.
Here is a video of the servers in action
[Video of all servers waiting for format ](https://pictshare.net/raw/m4szy91lqp.mp4)
However if I replace the docker container with a binary as well, everything works without a problem.
## Expected Behavior
Format should happen and servers should be initialized
## Current Behavior
forever hanging in "Waiting for the first server to format the disks."
## Possible Solution
Replace docker container with linux binary, then it works.
Also it might be noted in the docs that mixing docker and binary will cause problems even at the same version
## Steps to Reproduce (for bugs)
1. Start 3 machines (different networks, exposed port 9000) with `minio server http://server1:9000/data http://server2:9000/data http://server3:9000/data http://server4:9000/data`
2. Start one docker container `docker run -it --rm -e MINIO_ACCESS_KEY=mykey -e MINIO_SECRET_KEY=mysecret -p 9000:9000 -v /mnt:/data minio/minio server http://server1:9000/data http://server2:9000/data http://server3:9000/data http://server4:9000/data`
3. Watch them never sync
## Context
I tried to set up a distributed system on 4 physical locations with portforwarding and domains enabled. Didn't work unless I let them all run form the linux binary instead of docker
## Your Environment
* Version used (`minio version`): 2018-08-21T00:37:20Z
* Environment name and version: Debian 8 and 9 and one digitalocean docker instance
* Server type and version: mixed native hardare, vhosts and cloud instances
| 1.0 | [Distributed] Disks are never formatted if instances mixed with binary and docker hosts - I have set up 4 minio servers in 4 different physical locations. 3 Are running the native linux binary and one is running in a docker container.
**All instances are using this exact version:**
- Version: 2018-08-21T00:37:20Z
- Release-Tag: RELEASE.2018-08-21T00-37-20Z
- Commit-ID: 2d84b02bc429ad059a060001bccd712491e6e6a3
I am able to ping and curl all instances from all locations and even from inside the container. The answer from CURL is the expected XML with the message "XMinioServerNotInitialized" from all 4 nodes
I have all 4 sites open via my browser so I know they have the same auth info and that they are running correctly.
But it seems they never format the disks.
Here is a video of the servers in action
[Video of all servers waiting for format ](https://pictshare.net/raw/m4szy91lqp.mp4)
However if I replace the docker container with a binary as well, everything works without a problem.
## Expected Behavior
Format should happen and servers should be initialized
## Current Behavior
forever hanging in "Waiting for the first server to format the disks."
## Possible Solution
Replace docker container with linux binary, then it works.
Also it might be noted in the docs that mixing docker and binary will cause problems even at the same version
## Steps to Reproduce (for bugs)
1. Start 3 machines (different networks, exposed port 9000) with `minio server http://server1:9000/data http://server2:9000/data http://server3:9000/data http://server4:9000/data`
2. Start one docker container `docker run -it --rm -e MINIO_ACCESS_KEY=mykey -e MINIO_SECRET_KEY=mysecret -p 9000:9000 -v /mnt:/data minio/minio server http://server1:9000/data http://server2:9000/data http://server3:9000/data http://server4:9000/data`
3. Watch them never sync
## Context
I tried to set up a distributed system on 4 physical locations with portforwarding and domains enabled. Didn't work unless I let them all run form the linux binary instead of docker
## Your Environment
* Version used (`minio version`): 2018-08-21T00:37:20Z
* Environment name and version: Debian 8 and 9 and one digitalocean docker instance
* Server type and version: mixed native hardare, vhosts and cloud instances
| priority | disks are never formatted if instances mixed with binary and docker hosts i have set up minio servers in different physical locations are running the native linux binary and one is running in a docker container all instances are using this exact version version release tag release commit id i am able to ping and curl all instances from all locations and even from inside the container the answer from curl is the expected xml with the message xminioservernotinitialized from all nodes i have all sites open via my browser so i know they have the same auth info and that they are running correctly but it seems they never format the disks here is a video of the servers in action however if i replace the docker container with a binary as well everything works without a problem expected behavior format should happen and servers should be initialized current behavior forever hanging in waiting for the first server to format the disks possible solution replace docker container with linux binary then it works also it might be noted in the docs that mixing docker and binary will cause problems even at the same version steps to reproduce for bugs start machines different networks exposed port with minio server start one docker container docker run it rm e minio access key mykey e minio secret key mysecret p v mnt data minio minio server watch them never sync context i tried to set up a distributed system on physical locations with portforwarding and domains enabled didn t work unless i let them all run form the linux binary instead of docker your environment version used minio version environment name and version debian and and one digitalocean docker instance server type and version mixed native hardare vhosts and cloud instances | 1 |
272,860 | 8,518,511,161 | IssuesEvent | 2018-11-01 11:53:37 | GoldenSoftwareLtd/gedemin | https://api.github.com/repos/GoldenSoftwareLtd/gedemin | closed | Закрытие счёта с 0 суммой | Priority-Medium Type-Enhancement check | Originally reported on Google Code with ID 3219
```
Чисто теоретически можно сделать заказ с 0 суммой. К примеру, скидка 100% или ранее
принятый аванс покрыл всю сумму заказа.
Нужно предусмотреть закрытие такого заказ.
1. Печатать предчек.
2. Дать возможность вызвать окно оплаты
3. В окне оплаты печатается счёт но по фискальному регистратору если сумма 0 то не
печатать чек (хотя если аванс был принят то как бы надо закрыть такой заказ)
```
Reported by `stasgm` on 2013-08-22 14:57:12
| 1.0 | Закрытие счёта с 0 суммой - Originally reported on Google Code with ID 3219
```
Чисто теоретически можно сделать заказ с 0 суммой. К примеру, скидка 100% или ранее
принятый аванс покрыл всю сумму заказа.
Нужно предусмотреть закрытие такого заказ.
1. Печатать предчек.
2. Дать возможность вызвать окно оплаты
3. В окне оплаты печатается счёт но по фискальному регистратору если сумма 0 то не
печатать чек (хотя если аванс был принят то как бы надо закрыть такой заказ)
```
Reported by `stasgm` on 2013-08-22 14:57:12
| priority | закрытие счёта с суммой originally reported on google code with id чисто теоретически можно сделать заказ с суммой к примеру скидка или ранее принятый аванс покрыл всю сумму заказа нужно предусмотреть закрытие такого заказ печатать предчек дать возможность вызвать окно оплаты в окне оплаты печатается счёт но по фискальному регистратору если сумма то не печатать чек хотя если аванс был принят то как бы надо закрыть такой заказ reported by stasgm on | 1 |
445,654 | 12,834,617,005 | IssuesEvent | 2020-07-07 11:23:48 | NgyAnthony/APOS | https://api.github.com/repos/NgyAnthony/APOS | opened | Can't have an empty textbox for barcode | bug medium priority | Trying to make the textbox empty won't work because it tries to convert an empty string into a long value which obviously doesn't work. | 1.0 | Can't have an empty textbox for barcode - Trying to make the textbox empty won't work because it tries to convert an empty string into a long value which obviously doesn't work. | priority | can t have an empty textbox for barcode trying to make the textbox empty won t work because it tries to convert an empty string into a long value which obviously doesn t work | 1 |
793,393 | 27,994,302,698 | IssuesEvent | 2023-03-27 07:20:11 | AY2223S2-CS2113-T13-3/tp | https://api.github.com/repos/AY2223S2-CS2113-T13-3/tp | closed | Commands interface | Medium priority | An interface to allow the Parser class to call commands directly. Currently, the stopgap measure is to pass in the whole object into the parser class when instantiated so that the commands can be accessed. | 1.0 | Commands interface - An interface to allow the Parser class to call commands directly. Currently, the stopgap measure is to pass in the whole object into the parser class when instantiated so that the commands can be accessed. | priority | commands interface an interface to allow the parser class to call commands directly currently the stopgap measure is to pass in the whole object into the parser class when instantiated so that the commands can be accessed | 1 |
119,382 | 4,769,225,353 | IssuesEvent | 2016-10-26 11:50:55 | Cadasta/cadasta-platform | https://api.github.com/repos/Cadasta/cadasta-platform | closed | Uploading a XLSForm with empty labels throws IntegrityError | bug medium priority | ### Steps to reproduce the error
- Create a new project
- In the wizard, upload an XLSForm with empty labels
- Finish the remaining steps in the wizards and save the project.
### Actual behavior
Saving the project throws an IntegrityError. The field `label_xlat` in `QuestionOption` models can be an empty `dict` but not `None`. [This line](https://github.com/Cadasta/cadasta-platform/blob/master/cadasta/questionnaires/managers.py#L64) reads empty values to `None`, which leads to the exception. The line should read `label_xlat=o.get('label_xlat', o.get('label', {}))` instead.
### Expected behavior
The form should be processed without throwing an exception.
| 1.0 | Uploading a XLSForm with empty labels throws IntegrityError - ### Steps to reproduce the error
- Create a new project
- In the wizard, upload an XLSForm with empty labels
- Finish the remaining steps in the wizards and save the project.
### Actual behavior
Saving the project throws an IntegrityError. The field `label_xlat` in `QuestionOption` models can be an empty `dict` but not `None`. [This line](https://github.com/Cadasta/cadasta-platform/blob/master/cadasta/questionnaires/managers.py#L64) reads empty values to `None`, which leads to the exception. The line should read `label_xlat=o.get('label_xlat', o.get('label', {}))` instead.
### Expected behavior
The form should be processed without throwing an exception.
| priority | uploading a xlsform with empty labels throws integrityerror steps to reproduce the error create a new project in the wizard upload an xlsform with empty labels finish the remaining steps in the wizards and save the project actual behavior saving the project throws an integrityerror the field label xlat in questionoption models can be an empty dict but not none reads empty values to none which leads to the exception the line should read label xlat o get label xlat o get label instead expected behavior the form should be processed without throwing an exception | 1 |
279,102 | 8,657,530,072 | IssuesEvent | 2018-11-27 21:36:00 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Setting x._backward_hooks to None doesn't error, doesn't clear hooks | bootcamp medium priority | ## 🐛 Bug
Steps to reproduce the behavior:
```
import torch
y = torch.ones(5, 5, requires_grad=True)
counter = [0]
def bw_hook(grad):
counter[0] += 1
z = y + y
test = z.register_hook(bw_hook)
print(z._backward_hooks)
z._backward_hooks = None
print(z._backward_hooks)
z.backward(torch.ones(5, 5), retain_graph=True)
print(counter)
```
this prints:
```
$ python hooky.py
OrderedDict([(0, <function bw_hook at 0x7f2c9ab69e18>)])
None
[1]
```
## Expected behavior
I expect to either get an error, saying you can't set `_backward_hooks` to None, or for hooks to be cleared if I do this. | 1.0 | Setting x._backward_hooks to None doesn't error, doesn't clear hooks - ## 🐛 Bug
Steps to reproduce the behavior:
```
import torch
y = torch.ones(5, 5, requires_grad=True)
counter = [0]
def bw_hook(grad):
counter[0] += 1
z = y + y
test = z.register_hook(bw_hook)
print(z._backward_hooks)
z._backward_hooks = None
print(z._backward_hooks)
z.backward(torch.ones(5, 5), retain_graph=True)
print(counter)
```
this prints:
```
$ python hooky.py
OrderedDict([(0, <function bw_hook at 0x7f2c9ab69e18>)])
None
[1]
```
## Expected behavior
I expect to either get an error, saying you can't set `_backward_hooks` to None, or for hooks to be cleared if I do this. | priority | setting x backward hooks to none doesn t error doesn t clear hooks 🐛 bug steps to reproduce the behavior import torch y torch ones requires grad true counter def bw hook grad counter z y y test z register hook bw hook print z backward hooks z backward hooks none print z backward hooks z backward torch ones retain graph true print counter this prints python hooky py ordereddict none expected behavior i expect to either get an error saying you can t set backward hooks to none or for hooks to be cleared if i do this | 1 |
48,832 | 3,000,286,755 | IssuesEvent | 2015-07-24 00:06:59 | opendatakit/opendatakit | https://api.github.com/repos/opendatakit/opendatakit | closed | cannot use livereload feature of grunt | Priority-Medium Survey Type-Other | Originally reported on Google Code with ID 1058
```
Please add a comment to this issue if you encounter this problem:
Running:
grunt
Causes the webpage to never display (it is always spinning trying to load the page
and never times out).
Running:
grunt --verbose connect:livereload:keepalive
Displays it.
This starts grunt, but disables the file-change detection mechanisms that automatically
reload an HTML page when it or any javascript file it uses has been modified.
```
Reported by `mitchellsundt` on 2014-09-05 16:15:33
| 1.0 | cannot use livereload feature of grunt - Originally reported on Google Code with ID 1058
```
Please add a comment to this issue if you encounter this problem:
Running:
grunt
Causes the webpage to never display (it is always spinning trying to load the page
and never times out).
Running:
grunt --verbose connect:livereload:keepalive
Displays it.
This starts grunt, but disables the file-change detection mechanisms that automatically
reload an HTML page when it or any javascript file it uses has been modified.
```
Reported by `mitchellsundt` on 2014-09-05 16:15:33
| priority | cannot use livereload feature of grunt originally reported on google code with id please add a comment to this issue if you encounter this problem running grunt causes the webpage to never display it is always spinning trying to load the page and never times out running grunt verbose connect livereload keepalive displays it this starts grunt but disables the file change detection mechanisms that automatically reload an html page when it or any javascript file it uses has been modified reported by mitchellsundt on | 1 |
458,921 | 13,184,163,965 | IssuesEvent | 2020-08-12 18:52:43 | LBL-EESA/TECA | https://api.github.com/repos/LBL-EESA/TECA | opened | use pre-built docker image for travis-ci | 2_medium_priority | **Is your feature request related to a problem? Please describe.**
Travis-CI runs are slow, each run wastes time by building and installing all of the dependencies from scratch every time.
**Describe the solution you'd like**
We will publish docker images containing all the dependencies on docker hub and fetch them from the tests.
https://docs.travis-ci.com/user/docker/#using-a-docker-image-from-a-repository-in-a-build
**Describe alternatives you've considered**
**Additional context**
More of an issue now that we are running each docker flavor multiple times
| 1.0 | use pre-built docker image for travis-ci - **Is your feature request related to a problem? Please describe.**
Travis-CI runs are slow, each run wastes time by building and installing all of the dependencies from scratch every time.
**Describe the solution you'd like**
We will publish docker images containing all the dependencies on docker hub and fetch them from the tests.
https://docs.travis-ci.com/user/docker/#using-a-docker-image-from-a-repository-in-a-build
**Describe alternatives you've considered**
**Additional context**
More of an issue now that we are running each docker flavor multiple times
| priority | use pre built docker image for travis ci is your feature request related to a problem please describe travis ci runs are slow each run wastes time by building and installing all of the dependencies from scratch every time describe the solution you d like we will publish docker images containing all the dependencies on docker hub and fetch them from the tests describe alternatives you ve considered additional context more of an issue now that we are running each docker flavor multiple times | 1 |
31,995 | 2,742,582,080 | IssuesEvent | 2015-04-21 17:09:06 | boxkite/ckanext-donneesqctheme | https://api.github.com/repos/boxkite/ckanext-donneesqctheme | closed | Users to be member of all groups by default | Medium Priority | I don't know if this is new with 2.3, but it seems that now a user has to be member of a group to put datasts in that group.
In Données Québec, all users should be able to link datasets with all groups. Would it be possible to have the users linked with all the groups when created? | 1.0 | Users to be member of all groups by default - I don't know if this is new with 2.3, but it seems that now a user has to be member of a group to put datasts in that group.
In Données Québec, all users should be able to link datasets with all groups. Would it be possible to have the users linked with all the groups when created? | priority | users to be member of all groups by default i don t know if this is new with but it seems that now a user has to be member of a group to put datasts in that group in données québec all users should be able to link datasets with all groups would it be possible to have the users linked with all the groups when created | 1 |
84,985 | 3,683,134,972 | IssuesEvent | 2016-02-24 12:50:49 | PlanHubMe/PlanHub | https://api.github.com/repos/PlanHubMe/PlanHub | closed | A way of overriding a user's planner. | enhancement medium priority | Here is how it can go:
1. Go to admin panel
2. Click user data override
3. Give unique link, (expires in 3 hrs) for user to sign in.
4. User signs in, goes to page verifying giving temporary (24 hr) account access to administrator who is helping you.
5. Administrator has various options, but **cannot edit events directly**.
6. Administrator selects "Finished" button.
7. Session is closed, and cannot be reopened without going back to step 1
Admin options:
* Erase all data
* Change user's name
* Export data (JSON)
* Sends email to user with data
* Import data (JSON)
* Sends email to user with unique line (expires in 3hrs) for user to paste data into.
This is a lot of programming work, but it can be useful. | 1.0 | A way of overriding a user's planner. - Here is how it can go:
1. Go to admin panel
2. Click user data override
3. Give unique link, (expires in 3 hrs) for user to sign in.
4. User signs in, goes to page verifying giving temporary (24 hr) account access to administrator who is helping you.
5. Administrator has various options, but **cannot edit events directly**.
6. Administrator selects "Finished" button.
7. Session is closed, and cannot be reopened without going back to step 1
Admin options:
* Erase all data
* Change user's name
* Export data (JSON)
* Sends email to user with data
* Import data (JSON)
* Sends email to user with unique line (expires in 3hrs) for user to paste data into.
This is a lot of programming work, but it can be useful. | priority | a way of overriding a user s planner here is how it can go go to admin panel click user data override give unique link expires in hrs for user to sign in user signs in goes to page verifying giving temporary hr account access to administrator who is helping you administrator has various options but cannot edit events directly administrator selects finished button session is closed and cannot be reopened without going back to step admin options erase all data change user s name export data json sends email to user with data import data json sends email to user with unique line expires in for user to paste data into this is a lot of programming work but it can be useful | 1 |
398,296 | 11,739,455,330 | IssuesEvent | 2020-03-11 17:44:33 | thaliawww/ThaliApp | https://api.github.com/repos/thaliawww/ThaliApp | closed | User cannot deregister after deregistration deadline | bug priority: medium | In GitLab by @pingiun on Nov 11, 2019, 13:18
### One-sentence description
User cannot deregister after deregistration deadline
### Current behaviour / Reproducing the bug
<!-- Please write what is happening and how we could reproduce it, if relevant -->
1. Register for an event
2. Wait for deregistration deadline
3. Try to deregister, knowing that you will be fined
4. There is no deregistration button
### Expected behaviour
Just like on the website, you should be able to deregister and get a fine.
<!-- Please write how what happened did not meet your expectations --> | 1.0 | User cannot deregister after deregistration deadline - In GitLab by @pingiun on Nov 11, 2019, 13:18
### One-sentence description
User cannot deregister after deregistration deadline
### Current behaviour / Reproducing the bug
<!-- Please write what is happening and how we could reproduce it, if relevant -->
1. Register for an event
2. Wait for deregistration deadline
3. Try to deregister, knowing that you will be fined
4. There is no deregistration button
### Expected behaviour
Just like on the website, you should be able to deregister and get a fine.
<!-- Please write how what happened did not meet your expectations --> | priority | user cannot deregister after deregistration deadline in gitlab by pingiun on nov one sentence description user cannot deregister after deregistration deadline current behaviour reproducing the bug register for an event wait for deregistration deadline try to deregister knowing that you will be fined there is no deregistration button expected behaviour just like on the website you should be able to deregister and get a fine | 1 |
1,930 | 2,521,824,496 | IssuesEvent | 2015-01-19 17:09:40 | oculusinfo/aperture-tiles | https://api.github.com/repos/oculusinfo/aperture-tiles | closed | Map and Layer files expect the pyramid with different labels | bug client invalid P2 - Medium Priority refactor | The map configuration expects its pyramid configuration to be labelled "PyramidConfig.
The layer configuration expects its pyramid configuration to be labelled "pyramid".
The labels should be the same in both cases.
We should go through other similar cases and make sure we're consistent across the board. | 1.0 | Map and Layer files expect the pyramid with different labels - The map configuration expects its pyramid configuration to be labelled "PyramidConfig.
The layer configuration expects its pyramid configuration to be labelled "pyramid".
The labels should be the same in both cases.
We should go through other similar cases and make sure we're consistent across the board. | priority | map and layer files expect the pyramid with different labels the map configuration expects its pyramid configuration to be labelled pyramidconfig the layer configuration expects its pyramid configuration to be labelled pyramid the labels should be the same in both cases we should go through other similar cases and make sure we re consistent across the board | 1 |
358,904 | 10,651,690,966 | IssuesEvent | 2019-10-17 10:57:09 | dotkom/onlineweb4 | https://api.github.com/repos/dotkom/onlineweb4 | closed | Receipts are sent multiple times for individual transactions | Easy Package: Payment Priority: Medium Status: Available Type: Bug | **Describe the bug**
Receipts are mailed multiple times, even though only one payment occurs. The receipts have differing IDs.
**To Reproduce**
- Purchase something in the webshop, and observe that two receipt mails are received.
- Add money to your wallet for the same behavior.
Probably also occurs when paying for events.
**Additional context**
My guess, I haven't checked that this is precisely the case: Receipts are sent when PaymentRelation objects are saved. Following the recent changes to payments using Stripe, payments get saved multiple times. Perhaps getting asked for 3D-Secure authentication results in a third mail? | 1.0 | Receipts are sent multiple times for individual transactions - **Describe the bug**
Receipts are mailed multiple times, even though only one payment occurs. The receipts have differing IDs.
**To Reproduce**
- Purchase something in the webshop, and observe that two receipt mails are received.
- Add money to your wallet for the same behavior.
Probably also occurs when paying for events.
**Additional context**
My guess, I haven't checked that this is precisely the case: Receipts are sent when PaymentRelation objects are saved. Following the recent changes to payments using Stripe, payments get saved multiple times. Perhaps getting asked for 3D-Secure authentication results in a third mail? | priority | receipts are sent multiple times for individual transactions describe the bug receipts are mailed multiple times even though only one payment occurs the receipts have differing ids to reproduce purchase something in the webshop and observe that two receipt mails are received add money to your wallet for the same behavior probably also occurs when paying for events additional context my guess i haven t checked that this is precisely the case receipts are sent when paymentrelation objects are saved following the recent changes to payments using stripe payments get saved multiple times perhaps getting asked for secure authentication results in a third mail | 1 |
244,952 | 7,880,709,382 | IssuesEvent | 2018-06-26 16:42:14 | aowen87/FOO | https://api.github.com/repos/aowen87/FOO | closed | add support for external ghosts to SAMRAI plugin | Expected Use: 3 - Occasional Impact: 3 - Medium OS: All Priority: Normal Support Group: Any Target Version: 2.13.0 feature version: 2.12.3 | cyrus has an example dataset from noah elliot | 1.0 | add support for external ghosts to SAMRAI plugin - cyrus has an example dataset from noah elliot | priority | add support for external ghosts to samrai plugin cyrus has an example dataset from noah elliot | 1 |
608,011 | 18,795,933,322 | IssuesEvent | 2021-11-08 22:19:45 | mosaicml/yahp | https://api.github.com/repos/mosaicml/yahp | opened | Expose dot notation helpers conversion between dot notation/nested syntax | enhancement Medium Priority | It would be extremely useful to be able to convert between dot notation and nested syntax in order to support new features like parameter sweeping and other configuration tools. | 1.0 | Expose dot notation helpers conversion between dot notation/nested syntax - It would be extremely useful to be able to convert between dot notation and nested syntax in order to support new features like parameter sweeping and other configuration tools. | priority | expose dot notation helpers conversion between dot notation nested syntax it would be extremely useful to be able to convert between dot notation and nested syntax in order to support new features like parameter sweeping and other configuration tools | 1 |
462,627 | 13,250,445,670 | IssuesEvent | 2020-08-19 23:03:51 | onicagroup/runway | https://api.github.com/repos/onicagroup/runway | closed | [TODO] investigate implementing a code formatter | maintenance priority:medium | # Viable Options
- [autopep](https://github.com/hhatto/autopep8)
- [black](https://github.com/psf/black)
- [yapf](https://github.com/google/yapf)
# Considerations
- Should we implement one?
- Which one should we implement?
- How will it impact the code base?
- How will it impact the developers/maintainers? | 1.0 | [TODO] investigate implementing a code formatter - # Viable Options
- [autopep](https://github.com/hhatto/autopep8)
- [black](https://github.com/psf/black)
- [yapf](https://github.com/google/yapf)
# Considerations
- Should we implement one?
- Which one should we implement?
- How will it impact the code base?
- How will it impact the developers/maintainers? | priority | investigate implementing a code formatter viable options considerations should we implement one which one should we implement how will it impact the code base how will it impact the developers maintainers | 1 |
582,414 | 17,360,987,911 | IssuesEvent | 2021-07-29 20:35:55 | kleros/court | https://api.github.com/repos/kleros/court | closed | Email Notifications | Priority: Medium Status: Available Type: Enhancement :sparkles: | We should implement the new "Uber style" email notification that was designed by Plinio. | 1.0 | Email Notifications - We should implement the new "Uber style" email notification that was designed by Plinio. | priority | email notifications we should implement the new uber style email notification that was designed by plinio | 1 |
170,803 | 6,471,810,643 | IssuesEvent | 2017-08-17 12:37:27 | semperfiwebdesign/semperpluginstheme | https://api.github.com/repos/semperfiwebdesign/semperpluginstheme | opened | Reorganize documentation | Priority | Medium | https://semperplugins.com/documentation/
Not sure if this is a theme or plugin issue, but the documentation can be difficult to look at.
For example:
1) If there there are more items for one category than another, both are pushed down (check out the space under file editor module)
2) Maybe there should be a way to easily distinguish guides from settings.
3) I wonder if API should be on it's own page. | 1.0 | Reorganize documentation - https://semperplugins.com/documentation/
Not sure if this is a theme or plugin issue, but the documentation can be difficult to look at.
For example:
1) If there there are more items for one category than another, both are pushed down (check out the space under file editor module)
2) Maybe there should be a way to easily distinguish guides from settings.
3) I wonder if API should be on it's own page. | priority | reorganize documentation not sure if this is a theme or plugin issue but the documentation can be difficult to look at for example if there there are more items for one category than another both are pushed down check out the space under file editor module maybe there should be a way to easily distinguish guides from settings i wonder if api should be on it s own page | 1 |
344,103 | 10,340,005,168 | IssuesEvent | 2019-09-03 20:46:00 | acl-services/paprika | https://api.github.com/repos/acl-services/paprika | closed | Make Paprika storybook available publicly | Medium Priority → Task WIP | **Estimation: ~1 day**
Making paprika storybook public:
- [ ] domain for this?
- [ ] coordinate Jedi team to help with this
- [ ] script the release process for doing this. | 1.0 | Make Paprika storybook available publicly - **Estimation: ~1 day**
Making paprika storybook public:
- [ ] domain for this?
- [ ] coordinate Jedi team to help with this
- [ ] script the release process for doing this. | priority | make paprika storybook available publicly estimation day making paprika storybook public domain for this coordinate jedi team to help with this script the release process for doing this | 1 |
655,417 | 21,689,745,309 | IssuesEvent | 2022-05-09 14:25:02 | zeyneplervesarp/swe574-javagang | https://api.github.com/repos/zeyneplervesarp/swe574-javagang | closed | Change minutes to hours for time credits | enhancement backend high priority difficulty-medium | The balance of the users have been implemented with minutes, they should be hours.
Refactor the domain, dto, service classes and the faker objects in load database class. | 1.0 | Change minutes to hours for time credits - The balance of the users have been implemented with minutes, they should be hours.
Refactor the domain, dto, service classes and the faker objects in load database class. | priority | change minutes to hours for time credits the balance of the users have been implemented with minutes they should be hours refactor the domain dto service classes and the faker objects in load database class | 1 |
370,154 | 10,926,022,054 | IssuesEvent | 2019-11-22 13:53:16 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Inaccessible titles after deleting other titles | Medium Priority QA Staging | Version: 0.8.1.4 beta
In the Registrar GUI, deleting a title higher up in the list of titles causes all titles below the deleted title to disappear from the GUI and become inaccessible. These titles still exist as titles in /titlelist, law propositions, etc but do not appear in the registrar. | 1.0 | Inaccessible titles after deleting other titles - Version: 0.8.1.4 beta
In the Registrar GUI, deleting a title higher up in the list of titles causes all titles below the deleted title to disappear from the GUI and become inaccessible. These titles still exist as titles in /titlelist, law propositions, etc but do not appear in the registrar. | priority | inaccessible titles after deleting other titles version beta in the registrar gui deleting a title higher up in the list of titles causes all titles below the deleted title to disappear from the gui and become inaccessible these titles still exist as titles in titlelist law propositions etc but do not appear in the registrar | 1 |
215,486 | 7,294,496,033 | IssuesEvent | 2018-02-26 00:11:39 | minio/minio | https://api.github.com/repos/minio/minio | closed | Unable to heal objects | priority: medium | Unable to heal objects in distributed minio. even if the cluster have always ben running except for the upgrade i did before the heal command to latest version.
I am not able to send the sosreport since its time consuming to clear sensitive data. But I am willing to send you reports of specific parts if you require it. Just let me know what you need.
## Expected Behavior
Running `mc admin heal` should heal objects.
## Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1.
2.
3.
4.
## Context
ls -l for the different minio instances
minio01:
```
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
```
minio02:
```
drwxr-xr-x 2 minio minio 35 24 apr 13.26 00ad64d8-b767-4f6e-b7f6-5282d0e9f058.png
drwxr-xr-x 2 minio minio 35 24 apr 13.28 05dce802-9eee-446b-8c1a-39fe80bf2e92.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 2086f4a1-e06e-40c0-9d95-9046cb04a80e.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 2c88345a-1218-49ea-8be9-92e572fab683.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 3234d90e-be31-44f8-9ba1-cb3a3a97d51e.zip
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.28 361594f7-d42d-4e6e-92a8-957799697c36.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.27 41ef03e5-e7f1-48fd-aa2f-a86f232fc9ca.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 455a68bf-10b1-4092-b3f6-170fdac06d18.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 48f2b7c6-3325-41ad-959b-1f45881bbd8e.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 550e4bb7-b0f0-4aac-862b-679db72058ca.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 5587c75b-b0a6-4479-b810-a705c1015c45.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 58a3afd6-8f6c-4460-837b-0b182b6e6435.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 5de5ca3e-443f-4305-8010-95a1cf6994ee.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 7586fba1-142d-4833-b3b6-f8a30f94c23b.jpg
drwxr-xr-x 2 minio minio 35 22 sep 07.25 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 881df61c-933b-42ca-9a7a-6701665b735a.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 88d450ef-50e7-4d58-bbbc-978631ac08dc.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.27 925a3bdc-5c93-4e27-955a-038a22f57185.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 9b2c5946-ccfb-4763-af36-1ccf2683baff.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 a05d07d6-3e35-4771-90f0-84801e4a9654.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 a905a220-07c1-4979-9452-1462c590b96a.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 b6cc1d50-d215-477c-b7b6-0a63c3358fd1.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 c241949c-8b76-448e-826d-75453d62e6fe.pdf
drwxr-xr-x 2 minio minio 35 22 sep 07.25 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 dde3c849-7891-48a5-8c60-e547c48b8201.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 e1ea8118-f8ee-495f-8d8a-435139907993.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 e2c8ded6-524e-485f-997c-33d19e1d762b.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 e452af3e-9a91-4fce-b028-fba91144a521.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 f619c4d9-b185-4e79-a88d-f2957aec8a30.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 ff3921d4-6531-4dd6-ac0e-fd00296ca690.pdf
```
minio03:
```
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
```
minio04:
```
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
```
## Your Environment
* minio disk/node count: 4
* Version used (`minio version`): 2017-08-05T00:00:53Z
* Environment name and version (e.g. nginx 1.9.1): HAproxy 1.5
* Server type and version: VMware VSphere 6.5
* Operating System and version (`uname -a`): CentOS 7, Linux 3.10.0-514.10.2.el7.x86_64
* Storage: SSD SAN connected with fibrechannel
* Link to your project:
| 1.0 | Unable to heal objects - Unable to heal objects in distributed minio. even if the cluster have always ben running except for the upgrade i did before the heal command to latest version.
I am not able to send the sosreport since its time consuming to clear sensitive data. But I am willing to send you reports of specific parts if you require it. Just let me know what you need.
## Expected Behavior
Running `mc admin heal` should heal objects.
## Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1.
2.
3.
4.
## Context
ls -l for the different minio instances
minio01:
```
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
```
minio02:
```
drwxr-xr-x 2 minio minio 35 24 apr 13.26 00ad64d8-b767-4f6e-b7f6-5282d0e9f058.png
drwxr-xr-x 2 minio minio 35 24 apr 13.28 05dce802-9eee-446b-8c1a-39fe80bf2e92.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 2086f4a1-e06e-40c0-9d95-9046cb04a80e.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 2c88345a-1218-49ea-8be9-92e572fab683.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 3234d90e-be31-44f8-9ba1-cb3a3a97d51e.zip
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.28 361594f7-d42d-4e6e-92a8-957799697c36.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.27 41ef03e5-e7f1-48fd-aa2f-a86f232fc9ca.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 455a68bf-10b1-4092-b3f6-170fdac06d18.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 48f2b7c6-3325-41ad-959b-1f45881bbd8e.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 550e4bb7-b0f0-4aac-862b-679db72058ca.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 5587c75b-b0a6-4479-b810-a705c1015c45.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 58a3afd6-8f6c-4460-837b-0b182b6e6435.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 5de5ca3e-443f-4305-8010-95a1cf6994ee.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 7586fba1-142d-4833-b3b6-f8a30f94c23b.jpg
drwxr-xr-x 2 minio minio 35 22 sep 07.25 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 881df61c-933b-42ca-9a7a-6701665b735a.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 88d450ef-50e7-4d58-bbbc-978631ac08dc.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.27 925a3bdc-5c93-4e27-955a-038a22f57185.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 9b2c5946-ccfb-4763-af36-1ccf2683baff.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 a05d07d6-3e35-4771-90f0-84801e4a9654.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 a905a220-07c1-4979-9452-1462c590b96a.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 b6cc1d50-d215-477c-b7b6-0a63c3358fd1.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 c241949c-8b76-448e-826d-75453d62e6fe.pdf
drwxr-xr-x 2 minio minio 35 22 sep 07.25 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 dde3c849-7891-48a5-8c60-e547c48b8201.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 e1ea8118-f8ee-495f-8d8a-435139907993.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 e2c8ded6-524e-485f-997c-33d19e1d762b.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 e452af3e-9a91-4fce-b028-fba91144a521.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.28 f619c4d9-b185-4e79-a88d-f2957aec8a30.jpg
drwxr-xr-x 2 minio minio 35 24 apr 13.26 ff3921d4-6531-4dd6-ac0e-fd00296ca690.pdf
```
minio03:
```
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
```
minio04:
```
drwxr-xr-x 2 minio minio 35 24 apr 09.07 3546c1e2-b191-4b61-bdde-2411225c0625.pdf
drwxr-xr-x 2 minio minio 35 24 apr 13.26 83b17557-edea-4662-9d79-f54faeb4259c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 b85be709-aa98-4e73-9b19-4a04b4861a1c.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.57 c8885a40-d5df-4f90-9872-0af7c46026c0.jpg
drwxr-xr-x 2 minio minio 35 24 apr 12.58 d8ab0052-3057-4baa-aa9a-e082271a2cc8.jpg
```
## Your Environment
* minio disk/node count: 4
* Version used (`minio version`): 2017-08-05T00:00:53Z
* Environment name and version (e.g. nginx 1.9.1): HAproxy 1.5
* Server type and version: VMware VSphere 6.5
* Operating System and version (`uname -a`): CentOS 7, Linux 3.10.0-514.10.2.el7.x86_64
* Storage: SSD SAN connected with fibrechannel
* Link to your project:
| priority | unable to heal objects unable to heal objects in distributed minio even if the cluster have always ben running except for the upgrade i did before the heal command to latest version i am not able to send the sosreport since its time consuming to clear sensitive data but i am willing to send you reports of specific parts if you require it just let me know what you need expected behavior running mc admin heal should heal objects current behavior possible solution steps to reproduce for bugs context ls l for the different minio instances drwxr xr x minio minio apr bdde pdf drwxr xr x minio minio apr edea jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr png drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr pdf drwxr xr x minio minio apr zip drwxr xr x minio minio apr bdde pdf drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr pdf drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio sep edea jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr bbbc pdf drwxr xr x minio minio apr jpg drwxr xr x minio minio apr ccfb jpg drwxr xr x minio minio apr pdf drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr pdf drwxr xr x minio minio sep jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr pdf drwxr xr x minio minio apr bdde pdf drwxr xr x minio minio apr edea jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr bdde pdf drwxr xr x minio minio apr edea jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg drwxr xr x minio minio apr jpg your environment minio disk node count version used minio version environment name and version e g nginx haproxy server type and version vmware vsphere operating system and version uname a centos linux storage ssd san connected with fibrechannel link to your project | 1 |
144,732 | 5,544,867,667 | IssuesEvent | 2017-03-22 20:12:31 | astropy/astropy | https://api.github.com/repos/astropy/astropy | closed | Make a pretty printer for OrderedDict | Effort-medium Feature Request Package-novice Priority-Low | `OrderedDict` is used in many places within astropy, but printing an OrderedDict object gives an ugly and difficult-to-read output. `pprint` is no better. Some quick googling led to some stackoverflow answers but nothing really helpful.
Having a readable representation of these objects would be useful. One idea is to customize `pprint.pformat` to work with OrderedDict, then use that as `__repr__` or `__str__` in astropy the OrderDict class.
If someone is bored maybe this would be a good little project...
| 1.0 | Make a pretty printer for OrderedDict - `OrderedDict` is used in many places within astropy, but printing an OrderedDict object gives an ugly and difficult-to-read output. `pprint` is no better. Some quick googling led to some stackoverflow answers but nothing really helpful.
Having a readable representation of these objects would be useful. One idea is to customize `pprint.pformat` to work with OrderedDict, then use that as `__repr__` or `__str__` in astropy the OrderDict class.
If someone is bored maybe this would be a good little project...
| priority | make a pretty printer for ordereddict ordereddict is used in many places within astropy but printing an ordereddict object gives an ugly and difficult to read output pprint is no better some quick googling led to some stackoverflow answers but nothing really helpful having a readable representation of these objects would be useful one idea is to customize pprint pformat to work with ordereddict then use that as repr or str in astropy the orderdict class if someone is bored maybe this would be a good little project | 1 |
757,923 | 26,535,828,457 | IssuesEvent | 2023-01-19 15:35:00 | notofonts/latin-greek-cyrillic | https://api.github.com/repos/notofonts/latin-greek-cyrillic | closed | Some fonts might need update for USE and default mark zeroing behavior change in HB | in-evaluation Android Priority-Medium | In HarfBuzz 1.2.0 I change Universal Shaping Engine mark zeroing behavior to match (undocumented) Microsoft behavior:
https://github.com/behdad/harfbuzz/commit/da41e48f0a1a6af6d44ef25185d2421a29bd4166
We should go over those of noto-fonts that go through USE and double-check that they behave correctly after this change. If the fonts were tested against Windows, they should work better with HB now. If, on the other hand, they were built against HB, we might need to hotpatch them to add the width of marks back in 'dist' GPOS feature.
| 1.0 | Some fonts might need update for USE and default mark zeroing behavior change in HB - In HarfBuzz 1.2.0 I change Universal Shaping Engine mark zeroing behavior to match (undocumented) Microsoft behavior:
https://github.com/behdad/harfbuzz/commit/da41e48f0a1a6af6d44ef25185d2421a29bd4166
We should go over those of noto-fonts that go through USE and double-check that they behave correctly after this change. If the fonts were tested against Windows, they should work better with HB now. If, on the other hand, they were built against HB, we might need to hotpatch them to add the width of marks back in 'dist' GPOS feature.
| priority | some fonts might need update for use and default mark zeroing behavior change in hb in harfbuzz i change universal shaping engine mark zeroing behavior to match undocumented microsoft behavior we should go over those of noto fonts that go through use and double check that they behave correctly after this change if the fonts were tested against windows they should work better with hb now if on the other hand they were built against hb we might need to hotpatch them to add the width of marks back in dist gpos feature | 1 |
418,123 | 12,193,620,079 | IssuesEvent | 2020-04-29 14:39:46 | graknlabs/client-java | https://api.github.com/repos/graknlabs/client-java | closed | Implement missing BDD test added for explanation tests | priority: medium | In this PR: https://github.com/graknlabs/verification/pull/33, we added 2 new BDD steps. However, we only implemented them in Grakn Core, and we had to create empty "placeholders" for them in this PR: https://github.com/graknlabs/client-java/pull/94.
We now need to implement these 2 BDD steps in Client Java. | 1.0 | Implement missing BDD test added for explanation tests - In this PR: https://github.com/graknlabs/verification/pull/33, we added 2 new BDD steps. However, we only implemented them in Grakn Core, and we had to create empty "placeholders" for them in this PR: https://github.com/graknlabs/client-java/pull/94.
We now need to implement these 2 BDD steps in Client Java. | priority | implement missing bdd test added for explanation tests in this pr we added new bdd steps however we only implemented them in grakn core and we had to create empty placeholders for them in this pr we now need to implement these bdd steps in client java | 1 |
238,669 | 7,781,806,970 | IssuesEvent | 2018-06-06 02:35:34 | PathwayCommons/factoid | https://api.github.com/repos/PathwayCommons/factoid | closed | Reach text highlighting | 2-medium-priority feature | - Re-engineer current text input ( here: http://factoid.baderlab.org/new-document/fill )
- Get answer from reach
- Highlight text in text input with Reach annotations | 1.0 | Reach text highlighting - - Re-engineer current text input ( here: http://factoid.baderlab.org/new-document/fill )
- Get answer from reach
- Highlight text in text input with Reach annotations | priority | reach text highlighting re engineer current text input here get answer from reach highlight text in text input with reach annotations | 1 |
777,818 | 27,295,272,065 | IssuesEvent | 2023-02-23 19:47:03 | dvrpc/TrackingProgress | https://api.github.com/repos/dvrpc/TrackingProgress | closed | Remove "bug" notes | medium priority | @hachadorian, chart 3 for Greenhouse Gas Emissions and charts 2 and 3 for Job Growth have a note below them ("Note: Due to a current bug in the software, percentages display in the mouse-over as decimals.") that now can be removed now that the bug has been fixed. | 1.0 | Remove "bug" notes - @hachadorian, chart 3 for Greenhouse Gas Emissions and charts 2 and 3 for Job Growth have a note below them ("Note: Due to a current bug in the software, percentages display in the mouse-over as decimals.") that now can be removed now that the bug has been fixed. | priority | remove bug notes hachadorian chart for greenhouse gas emissions and charts and for job growth have a note below them note due to a current bug in the software percentages display in the mouse over as decimals that now can be removed now that the bug has been fixed | 1 |
374,214 | 11,082,028,110 | IssuesEvent | 2019-12-13 11:06:03 | DigitalCampus/oppia-mobile-android | https://api.github.com/repos/DigitalCampus/oppia-mobile-android | closed | Initial mockups for interface for user to see their quiz attempts (and responses) | Medium priority enhancement est-8-hours | At the moment the user can't look back at all their quiz attempts/responses - so need to add an interface for this in the app.
Firstly we need to make some mock ups for how this will display (and where it will be linked from), before finalising the tech dev.
The basic rationale for this is so that when supervisors visit users, they can review on the phone what the user has been doing. | 1.0 | Initial mockups for interface for user to see their quiz attempts (and responses) - At the moment the user can't look back at all their quiz attempts/responses - so need to add an interface for this in the app.
Firstly we need to make some mock ups for how this will display (and where it will be linked from), before finalising the tech dev.
The basic rationale for this is so that when supervisors visit users, they can review on the phone what the user has been doing. | priority | initial mockups for interface for user to see their quiz attempts and responses at the moment the user can t look back at all their quiz attempts responses so need to add an interface for this in the app firstly we need to make some mock ups for how this will display and where it will be linked from before finalising the tech dev the basic rationale for this is so that when supervisors visit users they can review on the phone what the user has been doing | 1 |
433,872 | 12,511,735,158 | IssuesEvent | 2020-06-02 21:09:08 | ChainSafe/forest | https://api.github.com/repos/ChainSafe/forest | closed | Connecting state transition to chain system | Blockchain Priority: 3 - Medium VM | **Task summary**
We have implemented the logic for making the state transition via `apply_blocks` however we still need to connect the VM subsystem with the blockchain subsystem. This will likely be in the `state_manager` component.
- [Apply Blocks - StateManager](https://github.com/ChainSafe/forest/blob/master/blockchain/state_manager/src/lib.rs#L96)
This task may be dependent on the implementation of graph sync #379 as the state transition will be called during the syncing process.
**Specification reference**
<!-- Provide a reference to the specification as to what is being implemented. -->
-
**Other information and links**
<!-- Add any other context, existing implementation reference or screenshots about the task here. -->
The below link shows the logic that is being used in lotus to update state from the blockchain subsystem, this should be used as a reference and not necessarily a 1:1 blueprint.
- [`TipSetState`: Lotus connecting logic](https://github.com/filecoin-project/lotus/blob/master/chain/stmgr/stmgr.go#L54)
<!-- Thank you 💪 --> | 1.0 | Connecting state transition to chain system - **Task summary**
We have implemented the logic for making the state transition via `apply_blocks` however we still need to connect the VM subsystem with the blockchain subsystem. This will likely be in the `state_manager` component.
- [Apply Blocks - StateManager](https://github.com/ChainSafe/forest/blob/master/blockchain/state_manager/src/lib.rs#L96)
This task may be dependent on the implementation of graph sync #379 as the state transition will be called during the syncing process.
**Specification reference**
<!-- Provide a reference to the specification as to what is being implemented. -->
-
**Other information and links**
<!-- Add any other context, existing implementation reference or screenshots about the task here. -->
The below link shows the logic that is being used in lotus to update state from the blockchain subsystem, this should be used as a reference and not necessarily a 1:1 blueprint.
- [`TipSetState`: Lotus connecting logic](https://github.com/filecoin-project/lotus/blob/master/chain/stmgr/stmgr.go#L54)
<!-- Thank you 💪 --> | priority | connecting state transition to chain system task summary we have implemented the logic for making the state transition via apply blocks however we still need to connect the vm subsystem with the blockchain subsystem this will likely be in the state manager component this task may be dependent on the implementation of graph sync as the state transition will be called during the syncing process specification reference other information and links the below link shows the logic that is being used in lotus to update state from the blockchain subsystem this should be used as a reference and not necessarily a blueprint | 1 |
430,042 | 12,440,466,791 | IssuesEvent | 2020-05-26 12:03:24 | RichardFav/AnalysisGUI | https://api.github.com/repos/RichardFav/AnalysisGUI | closed | two functions under the Rotation Analysis crash when using fixed-free mcomp files | MEDIUM priority bug | test experiment: 6-Expt-Filter-calc.mcomp
general filter: exclude all except RSPd and RSPg
The following functions under Rotation Analysis crash:
1-Phase Spiking Rate Comparison (Individual Cell)
2-Phase Spiking Rate Comparison (Whole Experiment)
Note that while the other functions run (i.e. Spiking Rate Heatmap and Kinematic Spiking Frequency), I can't tell whether they are including the correct number of cells. I can only confirm that the counts are correct on the "Overall Direction Bias" function because I can see the total number of cells there, and once filtering by AHV cells that match with what is in the freely moving analysis.
| 1.0 | two functions under the Rotation Analysis crash when using fixed-free mcomp files - test experiment: 6-Expt-Filter-calc.mcomp
general filter: exclude all except RSPd and RSPg
The following functions under Rotation Analysis crash:
1-Phase Spiking Rate Comparison (Individual Cell)
2-Phase Spiking Rate Comparison (Whole Experiment)
Note that while the other functions run (i.e. Spiking Rate Heatmap and Kinematic Spiking Frequency), I can't tell whether they are including the correct number of cells. I can only confirm that the counts are correct on the "Overall Direction Bias" function because I can see the total number of cells there, and once filtering by AHV cells that match with what is in the freely moving analysis.
| priority | two functions under the rotation analysis crash when using fixed free mcomp files test experiment expt filter calc mcomp general filter exclude all except rspd and rspg the following functions under rotation analysis crash phase spiking rate comparison individual cell phase spiking rate comparison whole experiment note that while the other functions run i e spiking rate heatmap and kinematic spiking frequency i can t tell whether they are including the correct number of cells i can only confirm that the counts are correct on the overall direction bias function because i can see the total number of cells there and once filtering by ahv cells that match with what is in the freely moving analysis | 1 |
747,519 | 26,089,378,303 | IssuesEvent | 2022-12-26 09:03:35 | EddieHubCommunity/LinkFree | https://api.github.com/repos/EddieHubCommunity/LinkFree | closed | [DOCS] Markdown formatting issue in `README.md` | 🏁 status: ready for dev 🟨 priority: medium 🛠 goal: fix good first issue 📄 aspect: text | ### Description
There's an issue that prevents proper Markdown formatting in the `README.md` file.
_Steps to contribute_
1. Fork the repository [**here**](https://github.com/EddieHubCommunity/LinkFree/).
2. Create a new branch in your fork named `fix_markdown`.
3. Open the `README.md` file.
4. In line 17, remove the space before the `**` at the end. This will fix the bold text being formatted wrongly. To do it easier, replace the current content with the following content:
```markdown
**Optional fields: `milestones`, `type(personal | community)`, `socials`, `testimonials` and `events`**\
```
5. Commit and push your changes.
6. Create a [**pull request**](https://github.com/EddieHubCommunity/LinkFree/compare).
7. You're done now! 🎉 You just have to wait until the maintainers review your PR.
If you need any help, let us know! 😃
### Screenshots
_No response_
### Additional information
This is a good first issue. Please leave it open to first-time contributors and beginners. 🙂
| 1.0 | [DOCS] Markdown formatting issue in `README.md` - ### Description
There's an issue that prevents proper Markdown formatting in the `README.md` file.
_Steps to contribute_
1. Fork the repository [**here**](https://github.com/EddieHubCommunity/LinkFree/).
2. Create a new branch in your fork named `fix_markdown`.
3. Open the `README.md` file.
4. In line 17, remove the space before the `**` at the end. This will fix the bold text being formatted wrongly. To do it easier, replace the current content with the following content:
```markdown
**Optional fields: `milestones`, `type(personal | community)`, `socials`, `testimonials` and `events`**\
```
5. Commit and push your changes.
6. Create a [**pull request**](https://github.com/EddieHubCommunity/LinkFree/compare).
7. You're done now! 🎉 You just have to wait until the maintainers review your PR.
If you need any help, let us know! 😃
### Screenshots
_No response_
### Additional information
This is a good first issue. Please leave it open to first-time contributors and beginners. 🙂
| priority | markdown formatting issue in readme md description there s an issue that prevents proper markdown formatting in the readme md file steps to contribute fork the repository create a new branch in your fork named fix markdown open the readme md file in line remove the space before the at the end this will fix the bold text being formatted wrongly to do it easier replace the current content with the following content markdown optional fields milestones type personal community socials testimonials and events commit and push your changes create a you re done now 🎉 you just have to wait until the maintainers review your pr if you need any help let us know 😃 screenshots no response additional information this is a good first issue please leave it open to first time contributors and beginners 🙂 | 1 |
771,675 | 27,088,660,727 | IssuesEvent | 2023-02-14 19:02:56 | cs-utulsa/Encrypted-Chat-Service | https://api.github.com/repos/cs-utulsa/Encrypted-Chat-Service | closed | Coding: fInput Class | coding Priority 1 Medium Effort | Please read the file transfer architecture before starting this story. Link [here](https://www.w3schools.com/python/python_file_handling.asp) for more details on files in Python
The goal of this story is create a class for reading in and placing files. When a user wants to send a file, they will use a button on the GUI to start the process. This will then use Tkinter to have them select the path of the file. At this point, the fInput class will be called with the parameter being the path (In a string) the user selected, the method should be name fReadFile(path). The fInput class should then read in the file, save the file name as a string and return the file object. There should also be a get method to get the file name.
When a file is received, it will be decrypted and then the fInput class will be called again with two parameters. The first is the file itself and the second is the path of where the file should be placed. This method should be called fCreateFile(file, path). This method should return a true or false if it was successful.
- [x] Create fInput Class
- [x] Define Method headers (Including a main method)
- [x] Create field for the file and path (String) objects
- [x] Create fReadFile method
- [x] Create getFileName method
- [x] Create fCreateFile Method
- [x] Code should be functional for any file type.
- [x] Code should be clean and well commented | 1.0 | Coding: fInput Class - Please read the file transfer architecture before starting this story. Link [here](https://www.w3schools.com/python/python_file_handling.asp) for more details on files in Python
The goal of this story is create a class for reading in and placing files. When a user wants to send a file, they will use a button on the GUI to start the process. This will then use Tkinter to have them select the path of the file. At this point, the fInput class will be called with the parameter being the path (In a string) the user selected, the method should be name fReadFile(path). The fInput class should then read in the file, save the file name as a string and return the file object. There should also be a get method to get the file name.
When a file is received, it will be decrypted and then the fInput class will be called again with two parameters. The first is the file itself and the second is the path of where the file should be placed. This method should be called fCreateFile(file, path). This method should return a true or false if it was successful.
- [x] Create fInput Class
- [x] Define Method headers (Including a main method)
- [x] Create field for the file and path (String) objects
- [x] Create fReadFile method
- [x] Create getFileName method
- [x] Create fCreateFile Method
- [x] Code should be functional for any file type.
- [x] Code should be clean and well commented | priority | coding finput class please read the file transfer architecture before starting this story link for more details on files in python the goal of this story is create a class for reading in and placing files when a user wants to send a file they will use a button on the gui to start the process this will then use tkinter to have them select the path of the file at this point the finput class will be called with the parameter being the path in a string the user selected the method should be name freadfile path the finput class should then read in the file save the file name as a string and return the file object there should also be a get method to get the file name when a file is received it will be decrypted and then the finput class will be called again with two parameters the first is the file itself and the second is the path of where the file should be placed this method should be called fcreatefile file path this method should return a true or false if it was successful create finput class define method headers including a main method create field for the file and path string objects create freadfile method create getfilename method create fcreatefile method code should be functional for any file type code should be clean and well commented | 1 |
159,452 | 6,046,549,899 | IssuesEvent | 2017-06-12 12:26:11 | libarchive/libarchive | https://api.github.com/repos/libarchive/libarchive | closed | Add a python wrapper | OpSys-All Priority-Medium Type-Enhancement | Original [issue 28](https://code.google.com/p/libarchive/issues/detail?id=28) created by Google Code user `pombredanne` on 2009-05-13T00:43:22.000Z:
```
Having a python wrapper for libarchive would be awesome!
Either a ctypes or swig wrapper
```
| 1.0 | Add a python wrapper - Original [issue 28](https://code.google.com/p/libarchive/issues/detail?id=28) created by Google Code user `pombredanne` on 2009-05-13T00:43:22.000Z:
```
Having a python wrapper for libarchive would be awesome!
Either a ctypes or swig wrapper
```
| priority | add a python wrapper original created by google code user pombredanne on having a python wrapper for libarchive would be awesome either a ctypes or swig wrapper | 1 |
25,622 | 2,683,866,518 | IssuesEvent | 2015-03-28 12:04:07 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | ConEmu.Maximus5.100212: EMenu popup after mouse move | 2–5 stars bug imported Priority-Medium | _From [pult....@gmail.com](https://code.google.com/u/108086555794593142651/) on February 12, 2010 18:27:41_
Версия ОС: Win7 32bit
Версия FAR: 2.0 build 1393 x86 Описание бага... Используется плагин из поставки far-a: EMenu.
Выпадение системного Shell-овского меню по нажатию правой мыши происходит
после последующего перемещения курсора мыши :(
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=183_ | 1.0 | ConEmu.Maximus5.100212: EMenu popup after mouse move - _From [pult....@gmail.com](https://code.google.com/u/108086555794593142651/) on February 12, 2010 18:27:41_
Версия ОС: Win7 32bit
Версия FAR: 2.0 build 1393 x86 Описание бага... Используется плагин из поставки far-a: EMenu.
Выпадение системного Shell-овского меню по нажатию правой мыши происходит
после последующего перемещения курсора мыши :(
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=183_ | priority | conemu emenu popup after mouse move from on february версия ос версия far build описание бага используется плагин из поставки far a emenu выпадение системного shell овского меню по нажатию правой мыши происходит после последующего перемещения курсора мыши original issue | 1 |
818,265 | 30,681,077,839 | IssuesEvent | 2023-07-26 09:12:14 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | reopened | [YSQL] Flakiness in org.yb.pgsql.TestPgAnalyze.testUniformRandomSampling | kind/bug area/ysql priority/medium | Jira Link: [DB-4911](https://yugabyte.atlassian.net/browse/DB-4911)
### Description
Flakiness can be seen in centos-clang15-debug builds for 2 tests -
1. org.yb.pgsql.TestPgAnalyze.testUniformRandomSamplingWithPaging
2. org.yb.pgsql.TestPgAnalyze.testUniformRandomSampling
[DB-4911]: https://yugabyte.atlassian.net/browse/DB-4911?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | 1.0 | [YSQL] Flakiness in org.yb.pgsql.TestPgAnalyze.testUniformRandomSampling - Jira Link: [DB-4911](https://yugabyte.atlassian.net/browse/DB-4911)
### Description
Flakiness can be seen in centos-clang15-debug builds for 2 tests -
1. org.yb.pgsql.TestPgAnalyze.testUniformRandomSamplingWithPaging
2. org.yb.pgsql.TestPgAnalyze.testUniformRandomSampling
[DB-4911]: https://yugabyte.atlassian.net/browse/DB-4911?atlOrigin=eyJpIjoiNWRkNTljNzYxNjVmNDY3MDlhMDU5Y2ZhYzA5YTRkZjUiLCJwIjoiZ2l0aHViLWNvbS1KU1cifQ | priority | flakiness in org yb pgsql testpganalyze testuniformrandomsampling jira link description flakiness can be seen in centos debug builds for tests org yb pgsql testpganalyze testuniformrandomsamplingwithpaging org yb pgsql testpganalyze testuniformrandomsampling | 1 |
372,029 | 11,008,137,583 | IssuesEvent | 2019-12-04 09:56:38 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Update available configs for hosted worlds | Fixed Medium Priority | We recently hid a bunch of config options from hosted worlds:
https://github.com/StrangeLoopGames/Eco/pull/5020
Need to unhide one of those. The Pause plugin, needs to display option for UnpauseAtFirstLogin
Make sure your working on branch that has the new config updates (they dont seem to be in staging yet, but should be) | 1.0 | Update available configs for hosted worlds - We recently hid a bunch of config options from hosted worlds:
https://github.com/StrangeLoopGames/Eco/pull/5020
Need to unhide one of those. The Pause plugin, needs to display option for UnpauseAtFirstLogin
Make sure your working on branch that has the new config updates (they dont seem to be in staging yet, but should be) | priority | update available configs for hosted worlds we recently hid a bunch of config options from hosted worlds need to unhide one of those the pause plugin needs to display option for unpauseatfirstlogin make sure your working on branch that has the new config updates they dont seem to be in staging yet but should be | 1 |
209,853 | 7,180,470,076 | IssuesEvent | 2018-01-31 23:26:40 | AZMAG/map-RecreationViewer | https://api.github.com/repos/AZMAG/map-RecreationViewer | closed | Bike layers stay on after unchecking group checkbox | Issue: Bug Priority: Medium | Steps to reproduce:
1. Click the bikeways checkbox
2. Zoom into a fairly close extent
3. Uncheck the "Bicycle Infrastructure" group checkbox
4. Observe that the bikeways layers are still visible
| 1.0 | Bike layers stay on after unchecking group checkbox - Steps to reproduce:
1. Click the bikeways checkbox
2. Zoom into a fairly close extent
3. Uncheck the "Bicycle Infrastructure" group checkbox
4. Observe that the bikeways layers are still visible
| priority | bike layers stay on after unchecking group checkbox steps to reproduce click the bikeways checkbox zoom into a fairly close extent uncheck the bicycle infrastructure group checkbox observe that the bikeways layers are still visible | 1 |
327,050 | 9,965,529,368 | IssuesEvent | 2019-07-08 08:59:12 | code4romania/monitorizare-vot-votanti-api | https://api.github.com/repos/code4romania/monitorizare-vot-votanti-api | closed | Edit the type of incidents available in the application | api enhancement laravel may-release medium priority php | Enhance the IncidentType controller to accept a POST method for adding incident types from the frontend.
The datamodel of the request should be the same as the datamodel persisted:
- Label
- Code
- Name
Please check IncidentTypeController.php and the api_routes.php to create the new route.
Enhance the IncidentType controller to accept a POST method for adding a list of incident types from the front end. The datamodel of the objects sent from frontend is the same as above:
- Label
- Code
- Name
Please check IncidentTypeController.php and the api_routes.php to create the new route. | 1.0 | Edit the type of incidents available in the application - Enhance the IncidentType controller to accept a POST method for adding incident types from the frontend.
The datamodel of the request should be the same as the datamodel persisted:
- Label
- Code
- Name
Please check IncidentTypeController.php and the api_routes.php to create the new route.
Enhance the IncidentType controller to accept a POST method for adding a list of incident types from the front end. The datamodel of the objects sent from frontend is the same as above:
- Label
- Code
- Name
Please check IncidentTypeController.php and the api_routes.php to create the new route. | priority | edit the type of incidents available in the application enhance the incidenttype controller to accept a post method for adding incident types from the frontend the datamodel of the request should be the same as the datamodel persisted label code name please check incidenttypecontroller php and the api routes php to create the new route enhance the incidenttype controller to accept a post method for adding a list of incident types from the front end the datamodel of the objects sent from frontend is the same as above label code name please check incidenttypecontroller php and the api routes php to create the new route | 1 |
404,544 | 11,859,001,250 | IssuesEvent | 2020-03-25 12:34:53 | AugurProject/augur | https://api.github.com/repos/AugurProject/augur | opened | add 'recently depleted liquidity' to the sort drop down in the my created markets section of portfolio | Add post v2 launch Priority: Medium | To allow user to bring all markets to the top with depleted liquidity in order of most recently depleted | 1.0 | add 'recently depleted liquidity' to the sort drop down in the my created markets section of portfolio - To allow user to bring all markets to the top with depleted liquidity in order of most recently depleted | priority | add recently depleted liquidity to the sort drop down in the my created markets section of portfolio to allow user to bring all markets to the top with depleted liquidity in order of most recently depleted | 1 |
404,847 | 11,863,700,011 | IssuesEvent | 2020-03-25 20:13:41 | noobaa/noobaa-core | https://api.github.com/repos/noobaa/noobaa-core | closed | Export CSV quietly fails to work | Priority 3 Medium Severity 3 Supportability UI implementation UX-feedback | ### Environment info
- Version: **1.11.4-fd0d23b**
- Deployment: **AZURE**
- Customer: **demo**
### Actual behavior
1.Audit Log - click on export as CSV. nothing happens
### Expected behavior
1. show progress or in process
2. download audit file
### Steps to reproduce
1. simply use the demo system
### Screenshots or Logs or other output that would be helpful
(If large, please upload as attachment)
1. the client sends request to the BE
`ACTION DISPATHCED loadAuditEntries with {categories: Array(8), count: 25}
api.js:15 RPC REQUEST events_api.read_activity_log {event: "(^node.)|(^obj.)|(^bucket.)|(^account.)|(^resource.)|(^dbg.)|(^cluster.)|(^conf.)", limit: 25} ==> {logs: Array(25)}
actions.js:36 ACTION DISPATHCED exportAuditEnteries with {categories: Array(8)}
api.js:15 RPC REQUEST events_api.export_activity_log {event: "(^node.)|(^obj.)|(^bucket.)|(^account.)|(^resource.)|(^dbg.)|(^cluster.)|(^conf.)"} ==> /public/audit.csv
`
2. The BE is writing the log data for a min - doesn't make sense.
3. nothing happens from that point on (no download)
4. audit file size is 23 MB - can be found [here](http://swiggity.eastus.cloudapp.azure.com:8080/fe/systems/demo/buckets/customers/objects/audit_3568.zip)
5. log attached [here](http://swiggity.eastus.cloudapp.azure.com:8080/fe/systems/demo/buckets/customers/objects/noobaa_issue_3568.log.zip)
look for export_activity_log and "writing data" - debug messages that I added to mark the export function in the BE.

| 1.0 | Export CSV quietly fails to work - ### Environment info
- Version: **1.11.4-fd0d23b**
- Deployment: **AZURE**
- Customer: **demo**
### Actual behavior
1.Audit Log - click on export as CSV. nothing happens
### Expected behavior
1. show progress or in process
2. download audit file
### Steps to reproduce
1. simply use the demo system
### Screenshots or Logs or other output that would be helpful
(If large, please upload as attachment)
1. the client sends request to the BE
`ACTION DISPATHCED loadAuditEntries with {categories: Array(8), count: 25}
api.js:15 RPC REQUEST events_api.read_activity_log {event: "(^node.)|(^obj.)|(^bucket.)|(^account.)|(^resource.)|(^dbg.)|(^cluster.)|(^conf.)", limit: 25} ==> {logs: Array(25)}
actions.js:36 ACTION DISPATHCED exportAuditEnteries with {categories: Array(8)}
api.js:15 RPC REQUEST events_api.export_activity_log {event: "(^node.)|(^obj.)|(^bucket.)|(^account.)|(^resource.)|(^dbg.)|(^cluster.)|(^conf.)"} ==> /public/audit.csv
`
2. The BE is writing the log data for a min - doesn't make sense.
3. nothing happens from that point on (no download)
4. audit file size is 23 MB - can be found [here](http://swiggity.eastus.cloudapp.azure.com:8080/fe/systems/demo/buckets/customers/objects/audit_3568.zip)
5. log attached [here](http://swiggity.eastus.cloudapp.azure.com:8080/fe/systems/demo/buckets/customers/objects/noobaa_issue_3568.log.zip)
look for export_activity_log and "writing data" - debug messages that I added to mark the export function in the BE.

| priority | export csv quietly fails to work environment info version deployment azure customer demo actual behavior audit log click on export as csv nothing happens expected behavior show progress or in process download audit file steps to reproduce simply use the demo system screenshots or logs or other output that would be helpful if large please upload as attachment the client sends request to the be action dispathced loadauditentries with categories array count api js rpc request events api read activity log event node obj bucket account resource dbg cluster conf limit logs array actions js action dispathced exportauditenteries with categories array api js rpc request events api export activity log event node obj bucket account resource dbg cluster conf public audit csv the be is writing the log data for a min doesn t make sense nothing happens from that point on no download audit file size is mb can be found log attached look for export activity log and writing data debug messages that i added to mark the export function in the be | 1 |
102,749 | 4,159,994,523 | IssuesEvent | 2016-06-17 11:25:03 | CanonicalLtd/maas-docs | https://api.github.com/repos/CanonicalLtd/maas-docs | opened | Replace API references with CLI commands | medium priority | The API references within the documentation are confusing and difficult to understand, especially as the API is already documented. When an example can't be given in the GUI, I propose we provide CLI examples instead. For instance:
replace:
POST /api/2.0/ipaddresses/ op=reserve
with:
maas admin ipaddresses reserve ip_address=192.168.122.148 | 1.0 | Replace API references with CLI commands - The API references within the documentation are confusing and difficult to understand, especially as the API is already documented. When an example can't be given in the GUI, I propose we provide CLI examples instead. For instance:
replace:
POST /api/2.0/ipaddresses/ op=reserve
with:
maas admin ipaddresses reserve ip_address=192.168.122.148 | priority | replace api references with cli commands the api references within the documentation are confusing and difficult to understand especially as the api is already documented when an example can t be given in the gui i propose we provide cli examples instead for instance replace post api ipaddresses op reserve with maas admin ipaddresses reserve ip address | 1 |
552,237 | 16,235,641,715 | IssuesEvent | 2021-05-07 00:08:04 | iaebots/iae | https://api.github.com/repos/iaebots/iae | closed | Developer's avatar on navbar. | Medium priority | Add developer's avatar on navbar as a button to a dropdown menu, with options to go to profile and log-out. | 1.0 | Developer's avatar on navbar. - Add developer's avatar on navbar as a button to a dropdown menu, with options to go to profile and log-out. | priority | developer s avatar on navbar add developer s avatar on navbar as a button to a dropdown menu with options to go to profile and log out | 1 |
614,627 | 19,187,265,928 | IssuesEvent | 2021-12-05 12:17:23 | ut-issl/c2a-core | https://api.github.com/repos/ut-issl/c2a-core | opened | c89のgcc Wextra Warningに対応すると,あるマイコンIDEにてWarningが出てしまう | bug priority::medium | ## 概要
c89のgcc Wextra Warningに対応すると,あるマイコンIDEにてWarningが出てしまう
## 詳細
- https://github.com/ut-issl/c2a-core/pull/73 にて, `const uint32_t` などの返り値を `uint32_t` などにした
- `warning: type qualifiers ignored on function return type [-Wignored-qualifiers]` が発生
- 基本型をconstつけて返しても,値渡しなため意味がないから
- 一方で,以下などでは,const構造体のメンバをconstなしで返している,として,あるマイコンIDEでは `C0008 (W) Conversion in return` を出してしまう.
- https://github.com/ut-issl/c2a-core/blob/1280836a5c6a6a5fc7a4fc8a5a9211134345b5b5/Drivers/Super/driver_super.c#L1513-L1516
## close条件
どちらの環境でもWarningがでなくなったら
## 備考
どうしようか.とりあえず,上の例だと,意味はないんだけど `(uint32_t)` とキャストさせる?
| 1.0 | c89のgcc Wextra Warningに対応すると,あるマイコンIDEにてWarningが出てしまう - ## 概要
c89のgcc Wextra Warningに対応すると,あるマイコンIDEにてWarningが出てしまう
## 詳細
- https://github.com/ut-issl/c2a-core/pull/73 にて, `const uint32_t` などの返り値を `uint32_t` などにした
- `warning: type qualifiers ignored on function return type [-Wignored-qualifiers]` が発生
- 基本型をconstつけて返しても,値渡しなため意味がないから
- 一方で,以下などでは,const構造体のメンバをconstなしで返している,として,あるマイコンIDEでは `C0008 (W) Conversion in return` を出してしまう.
- https://github.com/ut-issl/c2a-core/blob/1280836a5c6a6a5fc7a4fc8a5a9211134345b5b5/Drivers/Super/driver_super.c#L1513-L1516
## close条件
どちらの環境でもWarningがでなくなったら
## 備考
どうしようか.とりあえず,上の例だと,意味はないんだけど `(uint32_t)` とキャストさせる?
| priority | wextra warningに対応すると,あるマイコンideにてwarningが出てしまう 概要 wextra warningに対応すると,あるマイコンideにてwarningが出てしまう 詳細 にて, const t などの返り値を t などにした warning type qualifiers ignored on function return type が発生 基本型をconstつけて返しても,値渡しなため意味がないから 一方で,以下などでは,const構造体のメンバをconstなしで返している,として,あるマイコンideでは w conversion in return を出してしまう. close条件 どちらの環境でもwarningがでなくなったら 備考 どうしようか.とりあえず,上の例だと,意味はないんだけど t とキャストさせる? | 1 |
697,562 | 23,943,879,277 | IssuesEvent | 2022-09-12 04:35:45 | nokotan/siv3d-studio | https://api.github.com/repos/nokotan/siv3d-studio | closed | 自前のアセットファイル追加対応(画像やテキストファイル) | priority: medium | - priority: medium
最終的に Siv3D 公式チュートリアルの各種サンプルを動かせるようにしたい。
任意のファイルも簡単に追加できるようにしたい。
技術的に難しくない部分から段階的な実装で良い。 | 1.0 | 自前のアセットファイル追加対応(画像やテキストファイル) - - priority: medium
最終的に Siv3D 公式チュートリアルの各種サンプルを動かせるようにしたい。
任意のファイルも簡単に追加できるようにしたい。
技術的に難しくない部分から段階的な実装で良い。 | priority | 自前のアセットファイル追加対応(画像やテキストファイル) priority medium 最終的に 公式チュートリアルの各種サンプルを動かせるようにしたい。 任意のファイルも簡単に追加できるようにしたい。 技術的に難しくない部分から段階的な実装で良い。 | 1 |
368,484 | 10,879,462,130 | IssuesEvent | 2019-11-17 01:53:22 | rainlab/blog-plugin | https://api.github.com/repos/rainlab/blog-plugin | closed | Error after upgrade to 1.3.3 version | Priority: Medium Type: Bug | Hi, After upgrading to 1.3.3 we get this error:
"SQLSTATE[42S22]: Column not found: 1054 Unknown column 'rainlab_blog_posts.page' in 'where clause' (SQL: select rainlab_blog_posts.* from rainlab_blog_posts left join rainlab_translate_indexes on rainlab_blog_posts.id = rainlab_translate_indexes.model_id and rainlab_translate_indexes.model_type = RainLab\Blog\Models\Post and rainlab_translate_indexes.locale = en where (rainlab_blog_posts.page = 0 or (rainlab_translate_indexes.item = page and rainlab_translate_indexes.value = 0)) limit 1)" on line 664 of /home/cnpcd/public_html/vendor/laravel/framework/src/Illuminate/Database/Connection.php
For the specific change of language with the Translate plugin functional before upgrade.
The website for tst is this: http://www.cnpcd.ro/noutati and change languages from the bottom romana|english
All worked as a charm but after the update that errors ocur and can't find a fix for that. The error in the backecd is:
PDOException: SQLSTATE[42S22]: Column not found: 1054 Unknown column 'rainlab_blog_posts.page' in 'where clause' in /home/cnpcd/public_html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOConnection.php:77
The transalte plugins instaled are:
1.
Translate | Enables multi-lingual websites. | 1.6.0 | Alexey Bobkov, Samuel Georges
-- | -- | -- | --
2.
Translate Extended | Adds browser's language detection and language route prefixes to the Rainlab's Translate plugin. | 1.0.7 | Excodus
-- | -- | -- | --
Thank you. | 1.0 | Error after upgrade to 1.3.3 version - Hi, After upgrading to 1.3.3 we get this error:
"SQLSTATE[42S22]: Column not found: 1054 Unknown column 'rainlab_blog_posts.page' in 'where clause' (SQL: select rainlab_blog_posts.* from rainlab_blog_posts left join rainlab_translate_indexes on rainlab_blog_posts.id = rainlab_translate_indexes.model_id and rainlab_translate_indexes.model_type = RainLab\Blog\Models\Post and rainlab_translate_indexes.locale = en where (rainlab_blog_posts.page = 0 or (rainlab_translate_indexes.item = page and rainlab_translate_indexes.value = 0)) limit 1)" on line 664 of /home/cnpcd/public_html/vendor/laravel/framework/src/Illuminate/Database/Connection.php
For the specific change of language with the Translate plugin functional before upgrade.
The website for tst is this: http://www.cnpcd.ro/noutati and change languages from the bottom romana|english
All worked as a charm but after the update that errors ocur and can't find a fix for that. The error in the backecd is:
PDOException: SQLSTATE[42S22]: Column not found: 1054 Unknown column 'rainlab_blog_posts.page' in 'where clause' in /home/cnpcd/public_html/vendor/doctrine/dbal/lib/Doctrine/DBAL/Driver/PDOConnection.php:77
The transalte plugins instaled are:
1.
Translate | Enables multi-lingual websites. | 1.6.0 | Alexey Bobkov, Samuel Georges
-- | -- | -- | --
2.
Translate Extended | Adds browser's language detection and language route prefixes to the Rainlab's Translate plugin. | 1.0.7 | Excodus
-- | -- | -- | --
Thank you. | priority | error after upgrade to version hi after upgrading to we get this error sqlstate column not found unknown column rainlab blog posts page in where clause sql select rainlab blog posts from rainlab blog posts left join rainlab translate indexes on rainlab blog posts id rainlab translate indexes model id and rainlab translate indexes model type rainlab blog models post and rainlab translate indexes locale en where rainlab blog posts page or rainlab translate indexes item page and rainlab translate indexes value limit on line of home cnpcd public html vendor laravel framework src illuminate database connection php for the specific change of language with the translate plugin functional before upgrade the website for tst is this and change languages from the bottom romana english all worked as a charm but after the update that errors ocur and can t find a fix for that the error in the backecd is pdoexception sqlstate column not found unknown column rainlab blog posts page in where clause in home cnpcd public html vendor doctrine dbal lib doctrine dbal driver pdoconnection php the transalte plugins instaled are translate enables multi lingual websites alexey bobkov samuel georges translate extended adds browser s language detection and language route prefixes to the rainlab s translate plugin excodus thank you | 1 |
688,779 | 23,596,251,696 | IssuesEvent | 2022-08-23 19:32:42 | episphere/connectApp | https://api.github.com/repos/episphere/connectApp | closed | Site-specific My Samples page content | Medium Priority MVP | Configure My Samples page with content for the five sites for which content is linked below. Please note that once these are mocked up in dev, we’ll need to review internally and also share the final drafts with the sites so they can confirm that all information is accurate.
HealthPartners: https://nih.box.com/s/5f96oni440zklaxt20i4qxg4fvqfuwim
Sanford: https://nih.box.com/s/e5xonzmm27rvd77crt1mnrbalit51a7v
Marshfield: https://nih.box.com/s/6wwbw9bkvevmpk0v9d053h8mokb7jm0d
Henry Ford: https://nih.box.com/s/ppp1rp3ooljceidv17lao3n0tpw9dpom
UChicago: https://nih.box.com/s/mkipmbpsdz36hgrjuk8iev564gkpr2v0
For each of the KPs, implement a message placeholder on the My Samples page until further notice. Language here:
**We will begin collecting samples later this year. We will send you an email with instructions and next steps when it is time to donate samples. Thank you for being part of Connect!**
| 1.0 | Site-specific My Samples page content - Configure My Samples page with content for the five sites for which content is linked below. Please note that once these are mocked up in dev, we’ll need to review internally and also share the final drafts with the sites so they can confirm that all information is accurate.
HealthPartners: https://nih.box.com/s/5f96oni440zklaxt20i4qxg4fvqfuwim
Sanford: https://nih.box.com/s/e5xonzmm27rvd77crt1mnrbalit51a7v
Marshfield: https://nih.box.com/s/6wwbw9bkvevmpk0v9d053h8mokb7jm0d
Henry Ford: https://nih.box.com/s/ppp1rp3ooljceidv17lao3n0tpw9dpom
UChicago: https://nih.box.com/s/mkipmbpsdz36hgrjuk8iev564gkpr2v0
For each of the KPs, implement a message placeholder on the My Samples page until further notice. Language here:
**We will begin collecting samples later this year. We will send you an email with instructions and next steps when it is time to donate samples. Thank you for being part of Connect!**
| priority | site specific my samples page content configure my samples page with content for the five sites for which content is linked below please note that once these are mocked up in dev we’ll need to review internally and also share the final drafts with the sites so they can confirm that all information is accurate healthpartners sanford marshfield henry ford uchicago for each of the kps implement a message placeholder on the my samples page until further notice language here we will begin collecting samples later this year we will send you an email with instructions and next steps when it is time to donate samples thank you for being part of connect | 1 |
244,730 | 7,879,359,153 | IssuesEvent | 2018-06-26 13:12:49 | weglot/translate-wordpress | https://api.github.com/repos/weglot/translate-wordpress | closed | Add iframe and "congrats" message on first save | priority: medium status: confirmed type: enhancement | Like in V1, when the user first save settings, a box with a "Congrats" message should appear.
This box will contain an iframe to generate the first API call.
Also, we should add the button "Edit my translations" that links to Weglot. | 1.0 | Add iframe and "congrats" message on first save - Like in V1, when the user first save settings, a box with a "Congrats" message should appear.
This box will contain an iframe to generate the first API call.
Also, we should add the button "Edit my translations" that links to Weglot. | priority | add iframe and congrats message on first save like in when the user first save settings a box with a congrats message should appear this box will contain an iframe to generate the first api call also we should add the button edit my translations that links to weglot | 1 |
311,841 | 9,539,586,706 | IssuesEvent | 2019-04-30 17:19:57 | ansible/awx | https://api.github.com/repos/ansible/awx | opened | [WIP] Error Modal displayed when Copy Credential doesn't display | component:ui flag:🎱 priority:medium state:needs_info type:bug | ##### ISSUE TYPE
- Bug Report
##### SUMMARY
Copy Credential doesn't work if viewing more than 20 credentials per page
##### ENVIRONMENT
* AWX version: 4.0.0
* AWX install method: openshift, minishift, docker on linux, docker for mac, boot2docker
* Ansible version: X.Y.Z
* Operating System: macOS
* Web Browser: Firefox 60ESR
##### STEPS TO REPRODUCE
1. Create more than 20 Credentials
1. On Credentials page, view 50/100 Credentials
1. Copy one of the credentials
##### EXPECTED RESULTS
Credential copied as if viewing 20 credentials
##### ACTUAL RESULTS
Error modal.
##### ADDITIONAL INFORMATION

| 1.0 | [WIP] Error Modal displayed when Copy Credential doesn't display - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
Copy Credential doesn't work if viewing more than 20 credentials per page
##### ENVIRONMENT
* AWX version: 4.0.0
* AWX install method: openshift, minishift, docker on linux, docker for mac, boot2docker
* Ansible version: X.Y.Z
* Operating System: macOS
* Web Browser: Firefox 60ESR
##### STEPS TO REPRODUCE
1. Create more than 20 Credentials
1. On Credentials page, view 50/100 Credentials
1. Copy one of the credentials
##### EXPECTED RESULTS
Credential copied as if viewing 20 credentials
##### ACTUAL RESULTS
Error modal.
##### ADDITIONAL INFORMATION

| priority | error modal displayed when copy credential doesn t display issue type bug report summary copy credential doesn t work if viewing more than credentials per page environment awx version awx install method openshift minishift docker on linux docker for mac ansible version x y z operating system macos web browser firefox steps to reproduce create more than credentials on credentials page view credentials copy one of the credentials expected results credential copied as if viewing credentials actual results error modal additional information | 1 |
316,886 | 9,658,210,668 | IssuesEvent | 2019-05-20 10:24:46 | medic/medic | https://api.github.com/repos/medic/medic | closed | Improve Sentinel performance? | Priority: 2 - Medium Sentinel Type: Performance | **Zerothly**, we should benchmark sentinel, see if we consider it fast enough, and then look at progressively more complex solutions to increase its throughput.
Even without benchmarking (though do this first, maybe it's Good Enough™ already!), Sentinel has some clear places for optimisation.
**Firstly**, it has an intentional 50ms delay in it between processing each document. According to the git log and #907, this was done because sentinel going at [ludicrous speed](https://www.youtube.com/watch?v=ygE01sOhzz0&feature=youtu.be&t=60) caused CouchDB to freeze and become unstable. We should investigate how true this is, and why. Is this still true with CouchDB 2.0? Is this just because of lots of individual writes? Or is this a combination of writes and view regenerations in between those writes? If a delay really is required, is there a small value that we could use (it used to be 500ms!)? Is there another metric entirely other than "arbitrary sleep time" that we could use to maintain stability? Answering these questions will help determine some larger solutions.
**Secondly**, there are a couple of places: `deleteInfoDoc` and `deleteReadDocs`; which could be converted into asynchronous queues[1] instead of their actions blocking more processing. These may not be run that much, because they only run when a document is deleted, so it might not be worth the code change effort.
**Thirdly**, there are a couple of different easy strategies for improving performance that don't require re-architecting much. Both strategies I'm about to describe would require we add some kind of `partition` function (other name ideas: `index`, `groupBy`) to the transition api, which each transition would implement. `partition` would be a function that would return, given a document, a value which is used to describe how to group / categorise the action the transition performs on that document. This can then be used to try to increase performance:
1. Increase throughput by increasing the number of queues we have, from one to as many queues as we have unique values from `partition`.
2. Reduce writes by putting writes into a queue, and not writing until we're going to process a document for which there is already the same `partition` value in the queue.
The current thought about the implementation of `partition` is that for most things it would be the patient id, or perhaps the chw's / contact's uuid. However, for some transitions it would be a constant value (eg the transition's name), to force concurrency inside that transition. The clearest example being multi report alerts (after you see N things send a message: this cannot be parallelised without re-writing the transition and probably the transition processing flow).
So in the first approach we will be writing document from two CHPs concurrently, while still maintaining ordering inside those CHPs. And in the second example, we would delay writing until we were trying to process a document for a CHP for which we already have a queued write. The first approach will smash CouchDB harder, the second would smash CouchDB less. Which approach is best (or even deciding to do both) is dependent on the modern reasoning behind the 50ms delay.
**Fourthly**, and similarly to the previous point, we could look at relationships between one transition and all others, and itself. We could do this to categorise transitions and work out which ones can be parallelised without trouble, and which rely on others (can't accept patient reports before the registration is processed) or itself (multi-report-alerts, as described above). It's not clear if this individual action is more useful than just mostly doing it in the third point though.
**Lastly**, it's time to give up the idea that you can just re-write bits of sentinel, and we have look at a whole new architecture. Yikes.
[1] With all queues that is we talk about in this doc, we'd expect that: they are disk-backed; pick up queued items on re-boot, and where relevant flush after N seconds regardless. Once we add queues we'd also need to make sure the "seq we're up to" flow works properly, as that would be disrupted by this kind of thing. | 1.0 | Improve Sentinel performance? - **Zerothly**, we should benchmark sentinel, see if we consider it fast enough, and then look at progressively more complex solutions to increase its throughput.
Even without benchmarking (though do this first, maybe it's Good Enough™ already!), Sentinel has some clear places for optimisation.
**Firstly**, it has an intentional 50ms delay in it between processing each document. According to the git log and #907, this was done because sentinel going at [ludicrous speed](https://www.youtube.com/watch?v=ygE01sOhzz0&feature=youtu.be&t=60) caused CouchDB to freeze and become unstable. We should investigate how true this is, and why. Is this still true with CouchDB 2.0? Is this just because of lots of individual writes? Or is this a combination of writes and view regenerations in between those writes? If a delay really is required, is there a small value that we could use (it used to be 500ms!)? Is there another metric entirely other than "arbitrary sleep time" that we could use to maintain stability? Answering these questions will help determine some larger solutions.
**Secondly**, there are a couple of places: `deleteInfoDoc` and `deleteReadDocs`; which could be converted into asynchronous queues[1] instead of their actions blocking more processing. These may not be run that much, because they only run when a document is deleted, so it might not be worth the code change effort.
**Thirdly**, there are a couple of different easy strategies for improving performance that don't require re-architecting much. Both strategies I'm about to describe would require we add some kind of `partition` function (other name ideas: `index`, `groupBy`) to the transition api, which each transition would implement. `partition` would be a function that would return, given a document, a value which is used to describe how to group / categorise the action the transition performs on that document. This can then be used to try to increase performance:
1. Increase throughput by increasing the number of queues we have, from one to as many queues as we have unique values from `partition`.
2. Reduce writes by putting writes into a queue, and not writing until we're going to process a document for which there is already the same `partition` value in the queue.
The current thought about the implementation of `partition` is that for most things it would be the patient id, or perhaps the chw's / contact's uuid. However, for some transitions it would be a constant value (eg the transition's name), to force concurrency inside that transition. The clearest example being multi report alerts (after you see N things send a message: this cannot be parallelised without re-writing the transition and probably the transition processing flow).
So in the first approach we will be writing document from two CHPs concurrently, while still maintaining ordering inside those CHPs. And in the second example, we would delay writing until we were trying to process a document for a CHP for which we already have a queued write. The first approach will smash CouchDB harder, the second would smash CouchDB less. Which approach is best (or even deciding to do both) is dependent on the modern reasoning behind the 50ms delay.
**Fourthly**, and similarly to the previous point, we could look at relationships between one transition and all others, and itself. We could do this to categorise transitions and work out which ones can be parallelised without trouble, and which rely on others (can't accept patient reports before the registration is processed) or itself (multi-report-alerts, as described above). It's not clear if this individual action is more useful than just mostly doing it in the third point though.
**Lastly**, it's time to give up the idea that you can just re-write bits of sentinel, and we have look at a whole new architecture. Yikes.
[1] With all queues that is we talk about in this doc, we'd expect that: they are disk-backed; pick up queued items on re-boot, and where relevant flush after N seconds regardless. Once we add queues we'd also need to make sure the "seq we're up to" flow works properly, as that would be disrupted by this kind of thing. | priority | improve sentinel performance zerothly we should benchmark sentinel see if we consider it fast enough and then look at progressively more complex solutions to increase its throughput even without benchmarking though do this first maybe it s good enough™ already sentinel has some clear places for optimisation firstly it has an intentional delay in it between processing each document according to the git log and this was done because sentinel going at caused couchdb to freeze and become unstable we should investigate how true this is and why is this still true with couchdb is this just because of lots of individual writes or is this a combination of writes and view regenerations in between those writes if a delay really is required is there a small value that we could use it used to be is there another metric entirely other than arbitrary sleep time that we could use to maintain stability answering these questions will help determine some larger solutions secondly there are a couple of places deleteinfodoc and deletereaddocs which could be converted into asynchronous queues instead of their actions blocking more processing these may not be run that much because they only run when a document is deleted so it might not be worth the code change effort thirdly there are a couple of different easy strategies for improving performance that don t require re architecting much both strategies i m about to describe would require we add some kind of partition function other name ideas index groupby to the transition api which each transition would implement partition would be a function that would return given a document a value which is used to describe how to group categorise the action the transition performs on that document this can then be used to try to increase performance increase throughput by increasing the number of queues we have from one to as many queues as we have unique values from partition reduce writes by putting writes into a queue and not writing until we re going to process a document for which there is already the same partition value in the queue the current thought about the implementation of partition is that for most things it would be the patient id or perhaps the chw s contact s uuid however for some transitions it would be a constant value eg the transition s name to force concurrency inside that transition the clearest example being multi report alerts after you see n things send a message this cannot be parallelised without re writing the transition and probably the transition processing flow so in the first approach we will be writing document from two chps concurrently while still maintaining ordering inside those chps and in the second example we would delay writing until we were trying to process a document for a chp for which we already have a queued write the first approach will smash couchdb harder the second would smash couchdb less which approach is best or even deciding to do both is dependent on the modern reasoning behind the delay fourthly and similarly to the previous point we could look at relationships between one transition and all others and itself we could do this to categorise transitions and work out which ones can be parallelised without trouble and which rely on others can t accept patient reports before the registration is processed or itself multi report alerts as described above it s not clear if this individual action is more useful than just mostly doing it in the third point though lastly it s time to give up the idea that you can just re write bits of sentinel and we have look at a whole new architecture yikes with all queues that is we talk about in this doc we d expect that they are disk backed pick up queued items on re boot and where relevant flush after n seconds regardless once we add queues we d also need to make sure the seq we re up to flow works properly as that would be disrupted by this kind of thing | 1 |
710,307 | 24,414,186,384 | IssuesEvent | 2022-10-05 14:39:34 | assemblee-virtuelle/semapps | https://api.github.com/repos/assemblee-virtuelle/semapps | closed | Arrêt de Vercel - Utilisation de docker pour le frontend | medium difficulty 6. Déploiements medium priority | **Problématique**
Vercel, qui nous servait à publier le frontend devient payant à partir du 12 mai 2022. Nous allons donc passer progressivement tous les frontends (Archipel, Meta, etc...) sur docker (comme le middleware).
**Composants concernés**
Configuration Docker | 1.0 | Arrêt de Vercel - Utilisation de docker pour le frontend - **Problématique**
Vercel, qui nous servait à publier le frontend devient payant à partir du 12 mai 2022. Nous allons donc passer progressivement tous les frontends (Archipel, Meta, etc...) sur docker (comme le middleware).
**Composants concernés**
Configuration Docker | priority | arrêt de vercel utilisation de docker pour le frontend problématique vercel qui nous servait à publier le frontend devient payant à partir du mai nous allons donc passer progressivement tous les frontends archipel meta etc sur docker comme le middleware composants concernés configuration docker | 1 |
611,508 | 18,957,210,527 | IssuesEvent | 2021-11-18 21:52:13 | bounswe/2021SpringGroup7 | https://api.github.com/repos/bounswe/2021SpringGroup7 | closed | CM-3 Add Login Page | Type: Enhancement Status: In Progress Priority: Medium Mobile | Is your proposal related to a problem?
--------------------------------------
<!--
Provide a clear and concise description of what the problem is.
For example, "I'm always frustrated when..."
-->
No, it is a new and first page of our application. Users can enter the application with authenticated.
Describe the solution you'd like
--------------------------------
<!--
Provide a clear and concise description of what you want to happen.
-->
I have to add two inputs with native base to user enter their username or email address and password. It also has a register page link to sign up if user doesn't has an account in our app.
| 1.0 | CM-3 Add Login Page - Is your proposal related to a problem?
--------------------------------------
<!--
Provide a clear and concise description of what the problem is.
For example, "I'm always frustrated when..."
-->
No, it is a new and first page of our application. Users can enter the application with authenticated.
Describe the solution you'd like
--------------------------------
<!--
Provide a clear and concise description of what you want to happen.
-->
I have to add two inputs with native base to user enter their username or email address and password. It also has a register page link to sign up if user doesn't has an account in our app.
| priority | cm add login page is your proposal related to a problem provide a clear and concise description of what the problem is for example i m always frustrated when no it is a new and first page of our application users can enter the application with authenticated describe the solution you d like provide a clear and concise description of what you want to happen i have to add two inputs with native base to user enter their username or email address and password it also has a register page link to sign up if user doesn t has an account in our app | 1 |
663,025 | 22,160,017,155 | IssuesEvent | 2022-06-04 11:09:22 | Aerodymier/roblox-endorsed-weapons-fork | https://api.github.com/repos/Aerodymier/roblox-endorsed-weapons-fork | opened | Add an optional fire button for mobile devices | enhancement medium-priority easy | Currently guns fire with tapping on the screen on mobile devices which might not be ideal for guns like sniper rifles. | 1.0 | Add an optional fire button for mobile devices - Currently guns fire with tapping on the screen on mobile devices which might not be ideal for guns like sniper rifles. | priority | add an optional fire button for mobile devices currently guns fire with tapping on the screen on mobile devices which might not be ideal for guns like sniper rifles | 1 |
615,470 | 19,255,713,029 | IssuesEvent | 2021-12-09 11:02:30 | amosproj/amos2021ws05-fin-prod-port-quick-check | https://api.github.com/repos/amosproj/amos2021ws05-fin-prod-port-quick-check | opened | Code Style in frontend | priority: medium frontend | ## User story
1. As a frontend develper
2. I want a new code style
3. So that it is easier to work on it afterwards
## Acceptance criteria
* code style is implemented
## Definition of done
* Approved by product owner
* Tests have been written (e.g. Unit test, integration test etc..)
* Code has been peer reviewed and approved
* No syntax or runtime errors emerged
* Code has to be included in the release candidate
| 1.0 | Code Style in frontend - ## User story
1. As a frontend develper
2. I want a new code style
3. So that it is easier to work on it afterwards
## Acceptance criteria
* code style is implemented
## Definition of done
* Approved by product owner
* Tests have been written (e.g. Unit test, integration test etc..)
* Code has been peer reviewed and approved
* No syntax or runtime errors emerged
* Code has to be included in the release candidate
| priority | code style in frontend user story as a frontend develper i want a new code style so that it is easier to work on it afterwards acceptance criteria code style is implemented definition of done approved by product owner tests have been written e g unit test integration test etc code has been peer reviewed and approved no syntax or runtime errors emerged code has to be included in the release candidate | 1 |
743,379 | 25,896,472,389 | IssuesEvent | 2022-12-14 23:05:14 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | [Glyph] Glyph of the Monkey & Glyph of the Penguin | Class: Hunter Item Priority: Medium Status: Needs Confirmation Bug Report from Discord | Jimmy17
OP
— Yesterday at 1:12 AM
if you have both those glyphs on. it should be random , sometimes into monkey sometimes into penguin. atm it's only penguin. | 1.0 | [Glyph] Glyph of the Monkey & Glyph of the Penguin - Jimmy17
OP
— Yesterday at 1:12 AM
if you have both those glyphs on. it should be random , sometimes into monkey sometimes into penguin. atm it's only penguin. | priority | glyph of the monkey glyph of the penguin op — yesterday at am if you have both those glyphs on it should be random sometimes into monkey sometimes into penguin atm it s only penguin | 1 |
448,878 | 12,959,480,279 | IssuesEvent | 2020-07-20 13:04:07 | kubesphere/console | https://api.github.com/repos/kubesphere/console | closed | Network policy shows in project when disabled | area/console kind/bug kind/need-to-verify priority/medium | **Describe the bug**
Network policy is disabled, but still shows when creating project
<img width="962" alt="Screen Shot 2020-07-20 at 12 02 17 PM" src="https://user-images.githubusercontent.com/28859385/87898531-f4359b80-ca80-11ea-8442-ecfdd6944a51.png">
**Versions used(KubeSphere/Kubernetes)**
KubeSphere: 3.0.0-dev
Kubernetes: 1.15.12
**Environment**
allinone with network policy disabled
**Expected behavior**
Network policy not shown | 1.0 | Network policy shows in project when disabled - **Describe the bug**
Network policy is disabled, but still shows when creating project
<img width="962" alt="Screen Shot 2020-07-20 at 12 02 17 PM" src="https://user-images.githubusercontent.com/28859385/87898531-f4359b80-ca80-11ea-8442-ecfdd6944a51.png">
**Versions used(KubeSphere/Kubernetes)**
KubeSphere: 3.0.0-dev
Kubernetes: 1.15.12
**Environment**
allinone with network policy disabled
**Expected behavior**
Network policy not shown | priority | network policy shows in project when disabled describe the bug network policy is disabled but still shows when creating project img width alt screen shot at pm src versions used kubesphere kubernetes kubesphere dev kubernetes environment allinone with network policy disabled expected behavior network policy not shown | 1 |
640,238 | 20,777,493,124 | IssuesEvent | 2022-03-16 11:54:11 | supercrafter333/BetterBan | https://api.github.com/repos/supercrafter333/BetterBan | closed | [BUG:] kick-message-with-time: show {reason} at the end of the phrase | bug Priority: Medium WIP | # Bug: [kick-message-with-time: show {reason} at the end of the phrase]

```yml
kick-message-with-time: "§cBaneado por un administrador{line}§dMotivo: §7{reason}{line}§bTermina el dia: §a{time}{line}§6Apelaciones: §5discord.gg/x3a3wjsdTT"
```
### Informations
BetterBan Version: 1.4.1
Server-OS: Ubuntu
PHP Version: 8.0
PocketMine-MP Version: 4.2.3
### Error
X
### readjustment instructions
[I've done...]
[1. Ban a player with a random time (1 minute, 3 hours, whatever you prefer)]
[2. Set a random reason such as AAAAAAAA]
[3. When the player has been banned, the reason will appear at the end of the ban even if you have not put it there.]
### fix suggestions [optional]
| 1.0 | [BUG:] kick-message-with-time: show {reason} at the end of the phrase - # Bug: [kick-message-with-time: show {reason} at the end of the phrase]

```yml
kick-message-with-time: "§cBaneado por un administrador{line}§dMotivo: §7{reason}{line}§bTermina el dia: §a{time}{line}§6Apelaciones: §5discord.gg/x3a3wjsdTT"
```
### Informations
BetterBan Version: 1.4.1
Server-OS: Ubuntu
PHP Version: 8.0
PocketMine-MP Version: 4.2.3
### Error
X
### readjustment instructions
[I've done...]
[1. Ban a player with a random time (1 minute, 3 hours, whatever you prefer)]
[2. Set a random reason such as AAAAAAAA]
[3. When the player has been banned, the reason will appear at the end of the ban even if you have not put it there.]
### fix suggestions [optional]
| priority | kick message with time show reason at the end of the phrase bug yml kick message with time §cbaneado por un administrador line §dmotivo § reason line §btermina el dia §a time line § § gg informations betterban version server os ubuntu php version pocketmine mp version error x readjustment instructions fix suggestions | 1 |
77,103 | 3,506,261,089 | IssuesEvent | 2016-01-08 05:03:56 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Karazan script crash (BB #164) | Category: Crash migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** xaqq
**Original Date:** 06.06.2010 14:45:02 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/164
<hr>
Hello, here is a core crash;
Go to Karazan, mod your hp and the go in the room before Moroes'one. Pull all the mobs and then, the core will crash. | 1.0 | Karazan script crash (BB #164) - This issue was migrated from bitbucket.
**Original Reporter:** xaqq
**Original Date:** 06.06.2010 14:45:02 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/164
<hr>
Hello, here is a core crash;
Go to Karazan, mod your hp and the go in the room before Moroes'one. Pull all the mobs and then, the core will crash. | priority | karazan script crash bb this issue was migrated from bitbucket original reporter xaqq original date gmt original priority major original type bug original state resolved direct link hello here is a core crash go to karazan mod your hp and the go in the room before moroes one pull all the mobs and then the core will crash | 1 |
285,916 | 8,780,993,918 | IssuesEvent | 2018-12-19 19:02:46 | gravityview/GravityView | https://api.github.com/repos/gravityview/GravityView | opened | Add height / width setting to File Upload field settings | Core: Administration Core: Fields Difficulty: Low Enhancement Priority: Medium | When embedding an image on Multiple Entries, we set the size to be 250px (because of back-compat). Single entries have no size limit.
This requires lots of customization for what should be a simple thing: showing an image. | 1.0 | Add height / width setting to File Upload field settings - When embedding an image on Multiple Entries, we set the size to be 250px (because of back-compat). Single entries have no size limit.
This requires lots of customization for what should be a simple thing: showing an image. | priority | add height width setting to file upload field settings when embedding an image on multiple entries we set the size to be because of back compat single entries have no size limit this requires lots of customization for what should be a simple thing showing an image | 1 |
239,019 | 7,785,801,461 | IssuesEvent | 2018-06-06 16:54:25 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | closed | Changes for power line data exploration | Category: Core Priority: Medium Status: Ready For Review Type: Feature Type: Maintenance | These changes are for ease of use to aid in power utility data exploration:
* attribute-count --> tag-info
* attribute-comparison --> tag-comparison
* accuracy-distribution --> tag-accuracy-distribution
* change tag-info so it can read formats other than OGR
* change tag limit on tag-info from config option to command arg
* allow for listing unique tag keys only with tag-info
* make TagCriterion support multiple tags
* SetTagVisitor --> SetTagValueVisitor | 1.0 | Changes for power line data exploration - These changes are for ease of use to aid in power utility data exploration:
* attribute-count --> tag-info
* attribute-comparison --> tag-comparison
* accuracy-distribution --> tag-accuracy-distribution
* change tag-info so it can read formats other than OGR
* change tag limit on tag-info from config option to command arg
* allow for listing unique tag keys only with tag-info
* make TagCriterion support multiple tags
* SetTagVisitor --> SetTagValueVisitor | priority | changes for power line data exploration these changes are for ease of use to aid in power utility data exploration attribute count tag info attribute comparison tag comparison accuracy distribution tag accuracy distribution change tag info so it can read formats other than ogr change tag limit on tag info from config option to command arg allow for listing unique tag keys only with tag info make tagcriterion support multiple tags settagvisitor settagvaluevisitor | 1 |
712,606 | 24,500,708,472 | IssuesEvent | 2022-10-10 12:35:03 | simonvbrae/flordi-web | https://api.github.com/repos/simonvbrae/flordi-web | opened | Internationalisation | enhancement medium priority | The website should be available in english and dutch
- Implement internationalisation
- Add translations of context | 1.0 | Internationalisation - The website should be available in english and dutch
- Implement internationalisation
- Add translations of context | priority | internationalisation the website should be available in english and dutch implement internationalisation add translations of context | 1 |
147,998 | 5,657,140,205 | IssuesEvent | 2017-04-10 05:44:38 | k0shk0sh/FastHub | https://api.github.com/repos/k0shk0sh/FastHub | closed | Add an mark as read action to notifications in the notification panel | Priority: Medium Status: Accepted Type: Enhancement Type: Feature Request | There are times where I don't want to have to open the notification. A mark As read action would be helpful. | 1.0 | Add an mark as read action to notifications in the notification panel - There are times where I don't want to have to open the notification. A mark As read action would be helpful. | priority | add an mark as read action to notifications in the notification panel there are times where i don t want to have to open the notification a mark as read action would be helpful | 1 |
30,354 | 2,723,580,046 | IssuesEvent | 2015-04-14 13:30:07 | CruxFramework/crux-widgets | https://api.github.com/repos/CruxFramework/crux-widgets | closed | Sometimes Crux Compiler process does not die | bug CruxCompiler imported Milestone-2.2.1 Priority-Medium | _From [brenola...@gmail.com](https://code.google.com/u/110659575854244841723/) on May 06, 2010 18:25:51_
What steps will reproduce the problem? 1. Runs the script compiler's task classname =
"br.com.sysmap.crux.tools.compile.CruxModulesCompilerTask"
The java process does not die after compiling the script
Workaround:
Kill the java process by task manager in windows or kill java process in linux;
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=97_ | 1.0 | Sometimes Crux Compiler process does not die - _From [brenola...@gmail.com](https://code.google.com/u/110659575854244841723/) on May 06, 2010 18:25:51_
What steps will reproduce the problem? 1. Runs the script compiler's task classname =
"br.com.sysmap.crux.tools.compile.CruxModulesCompilerTask"
The java process does not die after compiling the script
Workaround:
Kill the java process by task manager in windows or kill java process in linux;
_Original issue: http://code.google.com/p/crux-framework/issues/detail?id=97_ | priority | sometimes crux compiler process does not die from on may what steps will reproduce the problem runs the script compiler s task classname br com sysmap crux tools compile cruxmodulescompilertask the java process does not die after compiling the script workaround kill the java process by task manager in windows or kill java process in linux original issue | 1 |
769,962 | 27,024,172,617 | IssuesEvent | 2023-02-11 11:34:16 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | closed | Refactor Bazel to use autoReplace + artifacts | priority-3-medium type:refactor manager:bazel status:ready | I think we should refactor our Bazel handling to use getArtifacts instead of trying to update them all in one go using updateDependency().
Whenever URLs, digests or checksums are dependent on the version - and not independently updatable - it should usually mean we use artifacts. | 1.0 | Refactor Bazel to use autoReplace + artifacts - I think we should refactor our Bazel handling to use getArtifacts instead of trying to update them all in one go using updateDependency().
Whenever URLs, digests or checksums are dependent on the version - and not independently updatable - it should usually mean we use artifacts. | priority | refactor bazel to use autoreplace artifacts i think we should refactor our bazel handling to use getartifacts instead of trying to update them all in one go using updatedependency whenever urls digests or checksums are dependent on the version and not independently updatable it should usually mean we use artifacts | 1 |
368,403 | 10,878,339,065 | IssuesEvent | 2019-11-16 16:58:38 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Orb of Rebirth remains non-free an extra day (creates misleading message on website) | good first issue priority: medium section: other status: issue: in progress | When the Orb of Rebirth stopped being free consistently from level 100, it was announced in the Aspiring Socialites guild with this message (emphasis is mine):
> "Hey Socialites! We deployed a tweak with this Gala that power users might run into. Previously, there was a sort of exploit where folks could use Fix Character Values + the Orb of Rebirth to evade having to pay Gems to change classes. Now, the freebie level 100+ Rebirth is only **usable once every 45 days**. Users who are affected by the limitation will see a small orange banner when they open the Orb from the Market saying how many days are left in this countdown."
However with the current code, it actually becomes free again after 46 days, not 45. I realise this is pretty minor but it would be easy to fix, and it's causing a misleading display message on the website (described below).
The current API code is this: https://github.com/HabitRPG/habitica/blob/2bfc41b1c3174db869209d6e4d74950f6920d222/website/common/script/libs/isFreeRebirth.js#L4-L14
It could be changed to match the official announcement by changing `> 45` to `>= 45`
---
The website's behaviour matches the API (i.e., it shows the Orb as free or not free according to the decision made by the API), but because the API makes you wait one extra day, the website's message is confusing on the final day (the day when it should already be free again) - on that day the website says "0 days until FREE Orb of Rebirth", as reported by @Evil_Overmom (0396e6e1-0edc-46f5-9449-24891ba16698):
> "Hi -- I'm having a problem with the orb of rebirth. It tells me I have 0 days til I get a free orb, but no option to get the free orb.

If the API's code is fixed as described above, then the Orb will become available a day earlier and you'll never see that misleading message.
FYI this is the website's code but it does not need to be changed if the API change above is made:
https://github.com/HabitRPG/habitica/blob/2bfc41b1c3174db869209d6e4d74950f6920d222/website/client/components/shops/buyModal.vue#L382-L384
https://github.com/HabitRPG/habitica/blob/2bfc41b1c3174db869209d6e4d74950f6920d222/website/client/components/shops/buyModal.vue#L82-L85
| 1.0 | Orb of Rebirth remains non-free an extra day (creates misleading message on website) - When the Orb of Rebirth stopped being free consistently from level 100, it was announced in the Aspiring Socialites guild with this message (emphasis is mine):
> "Hey Socialites! We deployed a tweak with this Gala that power users might run into. Previously, there was a sort of exploit where folks could use Fix Character Values + the Orb of Rebirth to evade having to pay Gems to change classes. Now, the freebie level 100+ Rebirth is only **usable once every 45 days**. Users who are affected by the limitation will see a small orange banner when they open the Orb from the Market saying how many days are left in this countdown."
However with the current code, it actually becomes free again after 46 days, not 45. I realise this is pretty minor but it would be easy to fix, and it's causing a misleading display message on the website (described below).
The current API code is this: https://github.com/HabitRPG/habitica/blob/2bfc41b1c3174db869209d6e4d74950f6920d222/website/common/script/libs/isFreeRebirth.js#L4-L14
It could be changed to match the official announcement by changing `> 45` to `>= 45`
---
The website's behaviour matches the API (i.e., it shows the Orb as free or not free according to the decision made by the API), but because the API makes you wait one extra day, the website's message is confusing on the final day (the day when it should already be free again) - on that day the website says "0 days until FREE Orb of Rebirth", as reported by @Evil_Overmom (0396e6e1-0edc-46f5-9449-24891ba16698):
> "Hi -- I'm having a problem with the orb of rebirth. It tells me I have 0 days til I get a free orb, but no option to get the free orb.

If the API's code is fixed as described above, then the Orb will become available a day earlier and you'll never see that misleading message.
FYI this is the website's code but it does not need to be changed if the API change above is made:
https://github.com/HabitRPG/habitica/blob/2bfc41b1c3174db869209d6e4d74950f6920d222/website/client/components/shops/buyModal.vue#L382-L384
https://github.com/HabitRPG/habitica/blob/2bfc41b1c3174db869209d6e4d74950f6920d222/website/client/components/shops/buyModal.vue#L82-L85
| priority | orb of rebirth remains non free an extra day creates misleading message on website when the orb of rebirth stopped being free consistently from level it was announced in the aspiring socialites guild with this message emphasis is mine hey socialites we deployed a tweak with this gala that power users might run into previously there was a sort of exploit where folks could use fix character values the orb of rebirth to evade having to pay gems to change classes now the freebie level rebirth is only usable once every days users who are affected by the limitation will see a small orange banner when they open the orb from the market saying how many days are left in this countdown however with the current code it actually becomes free again after days not i realise this is pretty minor but it would be easy to fix and it s causing a misleading display message on the website described below the current api code is this it could be changed to match the official announcement by changing to the website s behaviour matches the api i e it shows the orb as free or not free according to the decision made by the api but because the api makes you wait one extra day the website s message is confusing on the final day the day when it should already be free again on that day the website says days until free orb of rebirth as reported by evil overmom hi i m having a problem with the orb of rebirth it tells me i have days til i get a free orb but no option to get the free orb if the api s code is fixed as described above then the orb will become available a day earlier and you ll never see that misleading message fyi this is the website s code but it does not need to be changed if the api change above is made | 1 |
592,034 | 17,868,596,458 | IssuesEvent | 2021-09-06 12:40:39 | TencentBlueKing/bk-iam-saas | https://api.github.com/repos/TencentBlueKing/bk-iam-saas | closed | [Engine] 上云版本处理 | Type: Enhancement Layer: Backend Priority: Medium Size: S | 1. [x] 接入日志采集
1. [x] 接入sentry
1. [x] metric告警
- Pod存活告警
- 增量同步失败
- 快照失败
- 操过5分钟没有进行增量同步
1. [ ] healthz监控
1. [ ] 接入普罗米修斯
- dashboard api/task | 1.0 | [Engine] 上云版本处理 - 1. [x] 接入日志采集
1. [x] 接入sentry
1. [x] metric告警
- Pod存活告警
- 增量同步失败
- 快照失败
- 操过5分钟没有进行增量同步
1. [ ] healthz监控
1. [ ] 接入普罗米修斯
- dashboard api/task | priority | 上云版本处理 接入日志采集 接入sentry metric告警 pod存活告警 增量同步失败 快照失败 healthz监控 接入普罗米修斯 dashboard api task | 1 |
308,677 | 9,441,849,308 | IssuesEvent | 2019-04-15 03:49:56 | cilium/cilium | https://api.github.com/repos/cilium/cilium | closed | Default IPv6 prefix allocated from reserved address space | good-first-issue kind/bug priority/medium stale | As far as I'm able to determine, the default IPv6 cluster prefix of f00d::/48 is part of a [reserved block of addresses](https://www.iana.org/assignments/ipv6-address-space/ipv6-address-space.xhtml). Wouldn't a ula prefix be more appropriate as a default? Perhaps either a fixed fdxx prefix or one generated at cluster build time? | 1.0 | Default IPv6 prefix allocated from reserved address space - As far as I'm able to determine, the default IPv6 cluster prefix of f00d::/48 is part of a [reserved block of addresses](https://www.iana.org/assignments/ipv6-address-space/ipv6-address-space.xhtml). Wouldn't a ula prefix be more appropriate as a default? Perhaps either a fixed fdxx prefix or one generated at cluster build time? | priority | default prefix allocated from reserved address space as far as i m able to determine the default cluster prefix of is part of a wouldn t a ula prefix be more appropriate as a default perhaps either a fixed fdxx prefix or one generated at cluster build time | 1 |
140,431 | 5,408,936,902 | IssuesEvent | 2017-03-01 01:51:20 | ubc/compair | https://api.github.com/repos/ubc/compair | opened | Improvements to assignment setup instructions for instructors | enhancement front end instructor request medium priority | - [ ] Note that instructors/TAs can affect scores with their comparisons
- [ ] Note what are replies for
- [ ] Basically, all of the checkboxes need more explanation (under comparisons)
- [ ] Note that grades are participation only | 1.0 | Improvements to assignment setup instructions for instructors - - [ ] Note that instructors/TAs can affect scores with their comparisons
- [ ] Note what are replies for
- [ ] Basically, all of the checkboxes need more explanation (under comparisons)
- [ ] Note that grades are participation only | priority | improvements to assignment setup instructions for instructors note that instructors tas can affect scores with their comparisons note what are replies for basically all of the checkboxes need more explanation under comparisons note that grades are participation only | 1 |
757,039 | 26,494,428,759 | IssuesEvent | 2023-01-18 03:22:55 | NucciTheBoss/cleantest | https://api.github.com/repos/NucciTheBoss/cleantest | closed | Cleantest blows up `/tmp` with files | Priority: Medium Type: Cleanup | Cleantest blows up the `/tmp` directory on Linux systems with the various pickles, scriptlets, and tarballs that it needs to upload into the test environment providers. `/tmp` is emptied out after every reboot, but that does not do you much good if you almost never reboot your system.
As such, rather than just getting the name of the temporary directory on the host, instead use a named temporary file created with the `tempfile` module. This should cut down on the amount of space being consumed by cleantest on `/tmp`. Lastly, this will allow for better organization of files within the test environment. | 1.0 | Cleantest blows up `/tmp` with files - Cleantest blows up the `/tmp` directory on Linux systems with the various pickles, scriptlets, and tarballs that it needs to upload into the test environment providers. `/tmp` is emptied out after every reboot, but that does not do you much good if you almost never reboot your system.
As such, rather than just getting the name of the temporary directory on the host, instead use a named temporary file created with the `tempfile` module. This should cut down on the amount of space being consumed by cleantest on `/tmp`. Lastly, this will allow for better organization of files within the test environment. | priority | cleantest blows up tmp with files cleantest blows up the tmp directory on linux systems with the various pickles scriptlets and tarballs that it needs to upload into the test environment providers tmp is emptied out after every reboot but that does not do you much good if you almost never reboot your system as such rather than just getting the name of the temporary directory on the host instead use a named temporary file created with the tempfile module this should cut down on the amount of space being consumed by cleantest on tmp lastly this will allow for better organization of files within the test environment | 1 |
522,694 | 15,165,771,551 | IssuesEvent | 2021-02-12 15:29:28 | NeuraLegion/nexploit-cli | https://api.github.com/repos/NeuraLegion/nexploit-cli | closed | Add support on repeater part to Web Sockets | Priority: medium Type: enhancement | The engine supports scanning Web Sockets, but the repeater was built without this in mind, hence:
```
Error executing request: "GET wss://example.com/"
Cause: Invalid protocol: wss:
```
We need to be sure to add support for Web Sockets to the repeater by implementing a new `RequestExecutor`. You can use [ws](https://www.npmjs.com/package/ws) or any other libraries. Maybe, it requires some changes from the engine side. | 1.0 | Add support on repeater part to Web Sockets - The engine supports scanning Web Sockets, but the repeater was built without this in mind, hence:
```
Error executing request: "GET wss://example.com/"
Cause: Invalid protocol: wss:
```
We need to be sure to add support for Web Sockets to the repeater by implementing a new `RequestExecutor`. You can use [ws](https://www.npmjs.com/package/ws) or any other libraries. Maybe, it requires some changes from the engine side. | priority | add support on repeater part to web sockets the engine supports scanning web sockets but the repeater was built without this in mind hence error executing request get wss example com cause invalid protocol wss we need to be sure to add support for web sockets to the repeater by implementing a new requestexecutor you can use or any other libraries maybe it requires some changes from the engine side | 1 |
623,539 | 19,671,560,613 | IssuesEvent | 2022-01-11 07:57:40 | haydarsinan/SWE573-2021-Fall | https://api.github.com/repos/haydarsinan/SWE573-2021-Fall | closed | Attendees and Locations Update for Create Service and Event | Bug Medium Priority | On creation forms of services and events, location and attendees should be in details, not in creation pages. | 1.0 | Attendees and Locations Update for Create Service and Event - On creation forms of services and events, location and attendees should be in details, not in creation pages. | priority | attendees and locations update for create service and event on creation forms of services and events location and attendees should be in details not in creation pages | 1 |
434,381 | 12,517,401,761 | IssuesEvent | 2020-06-03 11:04:58 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | API: Reevaluate access require NODES_CREATE administrative role | Priority: Medium Type: Bug | **Describe the bug**
If you don't have NODES_CREATE administrative role but NODES_READ and NODES_UPDATE, you can't reevaluate access of a node.
**To Reproduce**
Steps to reproduce the behavior:
1. Create following admin access:
```ini
[tech]
actions=AUDITING_READ,RADIUS_LOG_READ,NODES_READ,NODES_UPDATE,SECURITY_EVENTS_READ,DHCP_OPTION_82_READ,USERS_READ
allowed_roles=
allowed_node_roles=
description=Technicien niveau 1
allowed_access_levels=
allowed_actions=
```
2. Create a user with this admin access
3. Try to reevaluate access of a node
=> message: "Unauthorized access, lacking the NODES_CREATE administrative role"
**Expected behavior**
NODES_UPDATE administrative role should be enough
| 1.0 | API: Reevaluate access require NODES_CREATE administrative role - **Describe the bug**
If you don't have NODES_CREATE administrative role but NODES_READ and NODES_UPDATE, you can't reevaluate access of a node.
**To Reproduce**
Steps to reproduce the behavior:
1. Create following admin access:
```ini
[tech]
actions=AUDITING_READ,RADIUS_LOG_READ,NODES_READ,NODES_UPDATE,SECURITY_EVENTS_READ,DHCP_OPTION_82_READ,USERS_READ
allowed_roles=
allowed_node_roles=
description=Technicien niveau 1
allowed_access_levels=
allowed_actions=
```
2. Create a user with this admin access
3. Try to reevaluate access of a node
=> message: "Unauthorized access, lacking the NODES_CREATE administrative role"
**Expected behavior**
NODES_UPDATE administrative role should be enough
| priority | api reevaluate access require nodes create administrative role describe the bug if you don t have nodes create administrative role but nodes read and nodes update you can t reevaluate access of a node to reproduce steps to reproduce the behavior create following admin access ini actions auditing read radius log read nodes read nodes update security events read dhcp option read users read allowed roles allowed node roles description technicien niveau allowed access levels allowed actions create a user with this admin access try to reevaluate access of a node message unauthorized access lacking the nodes create administrative role expected behavior nodes update administrative role should be enough | 1 |
616,646 | 19,308,811,771 | IssuesEvent | 2021-12-13 14:20:14 | lokka30/Treasury | https://api.github.com/repos/lokka30/Treasury | closed | Add method to get all accounts a player is member of/has X permission on | enhancement priority: medium developer thoughts wanted approved | Right now if I want to get all the accounts a player has, for example, `withdraw` permission on then I have to go thruough all accounts and check, not really optimal. The `EconomyProvider` could have a
`void retriveAllAccountsPlayerIsMemberOf(UUID player, EconomySubscriber<Collection<? extends Acount>>)`
`void retriveAllAccountsIdPlayerIsMemberOf(UUID player, EconomySubscriber<Collection<UUID>>)`
`void retriveAllAccountsPlayerHasPermission(UUID player, BankAccountPermission[] permissions, EconomySubscriber<Collection<? extends BankAccount>>)`
`void retriveAllAccountsIdPlayerHasPermission(UUID player, BankAccountPermission[] permissions, EconomySubscriber<Collection<UUID>>)`
| 1.0 | Add method to get all accounts a player is member of/has X permission on - Right now if I want to get all the accounts a player has, for example, `withdraw` permission on then I have to go thruough all accounts and check, not really optimal. The `EconomyProvider` could have a
`void retriveAllAccountsPlayerIsMemberOf(UUID player, EconomySubscriber<Collection<? extends Acount>>)`
`void retriveAllAccountsIdPlayerIsMemberOf(UUID player, EconomySubscriber<Collection<UUID>>)`
`void retriveAllAccountsPlayerHasPermission(UUID player, BankAccountPermission[] permissions, EconomySubscriber<Collection<? extends BankAccount>>)`
`void retriveAllAccountsIdPlayerHasPermission(UUID player, BankAccountPermission[] permissions, EconomySubscriber<Collection<UUID>>)`
| priority | add method to get all accounts a player is member of has x permission on right now if i want to get all the accounts a player has for example withdraw permission on then i have to go thruough all accounts and check not really optimal the economyprovider could have a void retriveallaccountsplayerismemberof uuid player economysubscriber void retriveallaccountsidplayerismemberof uuid player economysubscriber void retriveallaccountsplayerhaspermission uuid player bankaccountpermission permissions economysubscriber void retriveallaccountsidplayerhaspermission uuid player bankaccountpermission permissions economysubscriber | 1 |
650,404 | 21,390,822,545 | IssuesEvent | 2022-04-21 06:54:06 | eclipse/dirigible | https://api.github.com/repos/eclipse/dirigible | opened | [EDM] Put all the generated artefacts under a 'gen' folder | web-ide usability priority-high efforts-medium component-template | Put all the generated artefacts under a 'gen' folder. It will be easy and definitive, when regeneration is needed what to be cleaned. | 1.0 | [EDM] Put all the generated artefacts under a 'gen' folder - Put all the generated artefacts under a 'gen' folder. It will be easy and definitive, when regeneration is needed what to be cleaned. | priority | put all the generated artefacts under a gen folder put all the generated artefacts under a gen folder it will be easy and definitive when regeneration is needed what to be cleaned | 1 |
204,631 | 7,089,566,631 | IssuesEvent | 2018-01-12 03:34:41 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Tools to recover from a host failure | Medium Priority WMAgent | Assuming total loss of a host, we need a tool that can identify what we lost (WorkQueue elements running in the agent), reset them and invalidate the already produced output from DBS. There is no requirement to recover the work already done with finer granularity.
Putting this in medium priority as an Ops request, no milestone or deadline but should be soon.
| 1.0 | Tools to recover from a host failure - Assuming total loss of a host, we need a tool that can identify what we lost (WorkQueue elements running in the agent), reset them and invalidate the already produced output from DBS. There is no requirement to recover the work already done with finer granularity.
Putting this in medium priority as an Ops request, no milestone or deadline but should be soon.
| priority | tools to recover from a host failure assuming total loss of a host we need a tool that can identify what we lost workqueue elements running in the agent reset them and invalidate the already produced output from dbs there is no requirement to recover the work already done with finer granularity putting this in medium priority as an ops request no milestone or deadline but should be soon | 1 |
706,024 | 24,257,691,519 | IssuesEvent | 2022-09-27 19:19:36 | CDCgov/prime-reportstream | https://api.github.com/repos/CDCgov/prime-reportstream | closed | Emerge Medstaffing - IL State Dept received 2021 data | onboarding-ops support Medium Priority | ## Problem statement
Nicholas McGee from DoIT Illinois recently recieved files from 2021 COVID results. This is hindering our ability to finish 2021 year end close outs.
## What you need to know
ORC|RE|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|||||||||^Watson^Dr. Richard||(312)206-1064^WPN^PH^^1^312^2061064|20211223180000-0000||||||Emerge Medstaffing LLC^L|3330 W 177th St^^Hazel Crest^IL^60429^^^^17031|(888)855-2995^WPN^PH^^1^888^8552995|^^^^60429
OBR|1|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|97097-0^SARS-CoV-2 (COVID-19) Ag [Presence] in Upper respiratory specimen by Rapid immunoassay^LN^^^^2.70|||20211223174500-0000|20211223174500-0000||||||||^Watson^Dr. Richard|(312)206-1064^WPN^PH^^1^312^2061064|||||20220808175431-0000|||F
OBX|1|CWE|97097-0^SARS-CoV-2 (COVID-19) Ag [Presence] in Upper respiratory specimen by Rapid immunoassay^LN^^^^2.70||260373001^Detected^SCT|||A^Abnormal (applies to non-numeric results)^HL70078^^^^2.7|||F|||20211223174500-0000|14D2185111^Emerge Medstaffing LLC^CLIA||INDICAID COVID-19 Rapid Antigen Test_PHASE Scientific International, Ltd._EUA^INDICAID COVID-19 Rapid Antigen Test^99ELR^^^^2.70^^INDICAID COVID-19 Rapid Antigen Test_PHASE Scientific International, Ltd._EUA||20211223180000-0000||||Emerge Medstaffing LLC^^^^^CLIA&2.16.840.1.113883.4.7&ISO^XX^^^14D2185111|3330 W 177th St^^Hazel Crest^IL^60429^^^^17031
SPM|1|a480a48c-3ef2-4a95-a87a-d28a220e145a&Emerge Medstaffing L&14D2185111&CLIA^a480a48c-3ef2-4a95-a87a-d28a220e145a&Emerge Medstaffing L&14D2185111&CLIA||697989009^Anterior nares swab^SCT^^^^2.67||||53342003^Internal nose structure (body structure)^SCT^^^^2020-09-01|||||||||20211223174500-0000|20211223174506-0000
## Acceptance criteria
- Please confirm that this data is valid and the reason for the delayed reporting.
- In the future, please provide a notice or send out a mass email to the users submitting data and let them know that they should not be sending COVID data from 2021 as we are working on year end close outs?
## To do
- Assign to Engineer
| 1.0 | Emerge Medstaffing - IL State Dept received 2021 data - ## Problem statement
Nicholas McGee from DoIT Illinois recently recieved files from 2021 COVID results. This is hindering our ability to finish 2021 year end close outs.
## What you need to know
ORC|RE|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|||||||||^Watson^Dr. Richard||(312)206-1064^WPN^PH^^1^312^2061064|20211223180000-0000||||||Emerge Medstaffing LLC^L|3330 W 177th St^^Hazel Crest^IL^60429^^^^17031|(888)855-2995^WPN^PH^^1^888^8552995|^^^^60429
OBR|1|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|a480a48c-3ef2-4a95-a87a-d28a220e145a^Emerge Medstaffing L^14D2185111^CLIA|97097-0^SARS-CoV-2 (COVID-19) Ag [Presence] in Upper respiratory specimen by Rapid immunoassay^LN^^^^2.70|||20211223174500-0000|20211223174500-0000||||||||^Watson^Dr. Richard|(312)206-1064^WPN^PH^^1^312^2061064|||||20220808175431-0000|||F
OBX|1|CWE|97097-0^SARS-CoV-2 (COVID-19) Ag [Presence] in Upper respiratory specimen by Rapid immunoassay^LN^^^^2.70||260373001^Detected^SCT|||A^Abnormal (applies to non-numeric results)^HL70078^^^^2.7|||F|||20211223174500-0000|14D2185111^Emerge Medstaffing LLC^CLIA||INDICAID COVID-19 Rapid Antigen Test_PHASE Scientific International, Ltd._EUA^INDICAID COVID-19 Rapid Antigen Test^99ELR^^^^2.70^^INDICAID COVID-19 Rapid Antigen Test_PHASE Scientific International, Ltd._EUA||20211223180000-0000||||Emerge Medstaffing LLC^^^^^CLIA&2.16.840.1.113883.4.7&ISO^XX^^^14D2185111|3330 W 177th St^^Hazel Crest^IL^60429^^^^17031
SPM|1|a480a48c-3ef2-4a95-a87a-d28a220e145a&Emerge Medstaffing L&14D2185111&CLIA^a480a48c-3ef2-4a95-a87a-d28a220e145a&Emerge Medstaffing L&14D2185111&CLIA||697989009^Anterior nares swab^SCT^^^^2.67||||53342003^Internal nose structure (body structure)^SCT^^^^2020-09-01|||||||||20211223174500-0000|20211223174506-0000
## Acceptance criteria
- Please confirm that this data is valid and the reason for the delayed reporting.
- In the future, please provide a notice or send out a mass email to the users submitting data and let them know that they should not be sending COVID data from 2021 as we are working on year end close outs?
## To do
- Assign to Engineer
| priority | emerge medstaffing il state dept received data problem statement nicholas mcgee from doit illinois recently recieved files from covid results this is hindering our ability to finish year end close outs what you need to know orc re emerge medstaffing l clia emerge medstaffing l clia watson dr richard wpn ph emerge medstaffing llc l w st hazel crest il wpn ph obr emerge medstaffing l clia emerge medstaffing l clia sars cov covid ag in upper respiratory specimen by rapid immunoassay ln watson dr richard wpn ph f obx cwe sars cov covid ag in upper respiratory specimen by rapid immunoassay ln detected sct a abnormal applies to non numeric results f emerge medstaffing llc clia indicaid covid rapid antigen test phase scientific international ltd eua indicaid covid rapid antigen test indicaid covid rapid antigen test phase scientific international ltd eua emerge medstaffing llc clia iso xx w st hazel crest il spm emerge medstaffing l clia emerge medstaffing l clia anterior nares swab sct internal nose structure body structure sct acceptance criteria please confirm that this data is valid and the reason for the delayed reporting in the future please provide a notice or send out a mass email to the users submitting data and let them know that they should not be sending covid data from as we are working on year end close outs to do assign to engineer | 1 |
398,983 | 11,742,573,698 | IssuesEvent | 2020-03-12 01:17:47 | thaliawww/concrexit | https://api.github.com/repos/thaliawww/concrexit | closed | Calendar listview should be 'all future events' | bug priority: medium | In GitLab by @se-bastiaan on Dec 4, 2018, 12:52
### One-sentence description
Calendar listview should be 'all future events'
### Current behaviour / Reproducing the bug
Shows events in one week
<!--
1. Step 1
2. Step 2
3. ???
4. Breakage
-->
### Expected behaviour
Shows all future events
### Why?
Multiple people have expressed their discontent with the new workings of the listview on the website, in their eyes this is a bug. | 1.0 | Calendar listview should be 'all future events' - In GitLab by @se-bastiaan on Dec 4, 2018, 12:52
### One-sentence description
Calendar listview should be 'all future events'
### Current behaviour / Reproducing the bug
Shows events in one week
<!--
1. Step 1
2. Step 2
3. ???
4. Breakage
-->
### Expected behaviour
Shows all future events
### Why?
Multiple people have expressed their discontent with the new workings of the listview on the website, in their eyes this is a bug. | priority | calendar listview should be all future events in gitlab by se bastiaan on dec one sentence description calendar listview should be all future events current behaviour reproducing the bug shows events in one week step step breakage expected behaviour shows all future events why multiple people have expressed their discontent with the new workings of the listview on the website in their eyes this is a bug | 1 |
438,065 | 12,610,603,399 | IssuesEvent | 2020-06-12 05:31:05 | AuDigitalHealth/ci-fhir-r4 | https://api.github.com/repos/AuDigitalHealth/ci-fhir-r4 | closed | My Health Record Pathology Report | feature/change-profile size/medium status/in-progress tag/priority | <!-- Make sure to NOT include any sensitive or personally identifiable information. -->
# Prerequisites
- [x] I have searched [open](https://github.com/ci-fhir-r4/Issues/issues) and [closed](https://github.com/ci-fhir-r4/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested or reported
- [x] I have written a descriptive issue title
# The feature
## Change description
The purpose of this profile is to define a short form representation providing a rich text representation of the report as issued by the diagnostic service provider with supporting key elements of clinical relevance for the electronic exchange of pathology reports between healthcare providers and the My Health Record system infrastructure in Australia.
## What it actually enables people to do
<!-- A description of what the change / new feature will allow people to do, e.g. behaviour change, software change, validation etc. -->
## Mockups
If applicable, add screenshots or mockups to help explain the issue / feature.
## How awesome would it be?
<!-- A description of the value of this request to help us understand its value in comparison to the other awesome things we are doing. -->
## Workarounds
<!-- Are there any viable workarounds in case this cannot be addressed quickly?-->
## Additional context
Add any other context or links about the issue or feature here, e.g. replicate this page over here in this context. | 1.0 | My Health Record Pathology Report - <!-- Make sure to NOT include any sensitive or personally identifiable information. -->
# Prerequisites
- [x] I have searched [open](https://github.com/ci-fhir-r4/Issues/issues) and [closed](https://github.com/ci-fhir-r4/Issues/issues?utf8=%E2%9C%93&q=is%3Aissue+is%3Aclosed) issues to make sure it isn't already requested or reported
- [x] I have written a descriptive issue title
# The feature
## Change description
The purpose of this profile is to define a short form representation providing a rich text representation of the report as issued by the diagnostic service provider with supporting key elements of clinical relevance for the electronic exchange of pathology reports between healthcare providers and the My Health Record system infrastructure in Australia.
## What it actually enables people to do
<!-- A description of what the change / new feature will allow people to do, e.g. behaviour change, software change, validation etc. -->
## Mockups
If applicable, add screenshots or mockups to help explain the issue / feature.
## How awesome would it be?
<!-- A description of the value of this request to help us understand its value in comparison to the other awesome things we are doing. -->
## Workarounds
<!-- Are there any viable workarounds in case this cannot be addressed quickly?-->
## Additional context
Add any other context or links about the issue or feature here, e.g. replicate this page over here in this context. | priority | my health record pathology report prerequisites i have searched and issues to make sure it isn t already requested or reported i have written a descriptive issue title the feature change description the purpose of this profile is to define a short form representation providing a rich text representation of the report as issued by the diagnostic service provider with supporting key elements of clinical relevance for the electronic exchange of pathology reports between healthcare providers and the my health record system infrastructure in australia what it actually enables people to do mockups if applicable add screenshots or mockups to help explain the issue feature how awesome would it be workarounds additional context add any other context or links about the issue or feature here e g replicate this page over here in this context | 1 |
483,045 | 13,917,827,254 | IssuesEvent | 2020-10-21 06:42:22 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | Blog post's video Embed is not working in activity feed. | bug priority: medium | **Describe the bug**
When posting blog posts containing a video. The video doesn't embed on the activity feed. Videos are now displaying as URL instead of the video.
This all started happening after the latest update v1.5.3
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Admin Dashboard >> Posts >> Add New' and add a post with a Youtube video
2. Navigate to the Activity Feed page on the frontend.
3. See Error.
**Expected behavior**
It should embed the video in the activity feed.
**Screenshots**
Backend:
https://prnt.sc/v3kzmw
Front End Activity Feed:
https://prnt.sc/v3kz3t
**Support ticket links**
https://secure.helpscout.net/conversation/1312707010/103861/
| 1.0 | Blog post's video Embed is not working in activity feed. - **Describe the bug**
When posting blog posts containing a video. The video doesn't embed on the activity feed. Videos are now displaying as URL instead of the video.
This all started happening after the latest update v1.5.3
**To Reproduce**
Steps to reproduce the behavior:
1. Go to 'Admin Dashboard >> Posts >> Add New' and add a post with a Youtube video
2. Navigate to the Activity Feed page on the frontend.
3. See Error.
**Expected behavior**
It should embed the video in the activity feed.
**Screenshots**
Backend:
https://prnt.sc/v3kzmw
Front End Activity Feed:
https://prnt.sc/v3kz3t
**Support ticket links**
https://secure.helpscout.net/conversation/1312707010/103861/
| priority | blog post s video embed is not working in activity feed describe the bug when posting blog posts containing a video the video doesn t embed on the activity feed videos are now displaying as url instead of the video this all started happening after the latest update to reproduce steps to reproduce the behavior go to admin dashboard posts add new and add a post with a youtube video navigate to the activity feed page on the frontend see error expected behavior it should embed the video in the activity feed screenshots backend front end activity feed support ticket links | 1 |
651,185 | 21,468,621,888 | IssuesEvent | 2022-04-26 07:27:14 | LuanRT/YouTube.js | https://api.github.com/repos/LuanRT/YouTube.js | closed | Add support for removing videos from a playlist | enhancement good first issue priority: medium | ## Expected Behavior
The library should be able to remove multiple videos from a given playlist.
## Current Behavior
Can only add videos.
## Sample Code
The Innertube API seems to require a parameter called setVideoId to remove a video from a playlist. It is not a regular video id but rather a special id that can be retrieved by requesting playlist info using the browse endpoint.
After getting the setVideoId, the following payload should be sent to the `browse/edit_playlist` endpoint:
```js
{
action: 'ACTION_REMOVE_VIDEO',
setVideoId: string
}
``` | 1.0 | Add support for removing videos from a playlist - ## Expected Behavior
The library should be able to remove multiple videos from a given playlist.
## Current Behavior
Can only add videos.
## Sample Code
The Innertube API seems to require a parameter called setVideoId to remove a video from a playlist. It is not a regular video id but rather a special id that can be retrieved by requesting playlist info using the browse endpoint.
After getting the setVideoId, the following payload should be sent to the `browse/edit_playlist` endpoint:
```js
{
action: 'ACTION_REMOVE_VIDEO',
setVideoId: string
}
``` | priority | add support for removing videos from a playlist expected behavior the library should be able to remove multiple videos from a given playlist current behavior can only add videos sample code the innertube api seems to require a parameter called setvideoid to remove a video from a playlist it is not a regular video id but rather a special id that can be retrieved by requesting playlist info using the browse endpoint after getting the setvideoid the following payload should be sent to the browse edit playlist endpoint js action action remove video setvideoid string | 1 |
404,079 | 11,851,917,699 | IssuesEvent | 2020-03-24 18:57:09 | department-of-veterans-affairs/caseflow | https://api.github.com/repos/department-of-veterans-affairs/caseflow | closed | Feature Testing by Roles | Priority: Medium Product: caseflow-hearings Stakeholder: BVA Team: Tango 💃 Type: Investigation Type: Tech-Improvement | We've found that it's often difficult to understand exactly how a type of use interacts with Caseflow. For the hearing apps, we have Hearing Coordinators, VLJs, Regional Office staff, and VSOs that all interact with the same sets of pages slightly differently.
We'd like to experiment with reorganizing our feature tests around user roles rather than simply pages. We could either change the `/spec/feature/hearings` directory to either `<role_name>/<page>` or `<page>/<role_name>`.
We think this would give folks who have not worked on hearings a better sense of how the app is used across roles while also allowing us to test specific use cases more thoroughly. | 1.0 | Feature Testing by Roles - We've found that it's often difficult to understand exactly how a type of use interacts with Caseflow. For the hearing apps, we have Hearing Coordinators, VLJs, Regional Office staff, and VSOs that all interact with the same sets of pages slightly differently.
We'd like to experiment with reorganizing our feature tests around user roles rather than simply pages. We could either change the `/spec/feature/hearings` directory to either `<role_name>/<page>` or `<page>/<role_name>`.
We think this would give folks who have not worked on hearings a better sense of how the app is used across roles while also allowing us to test specific use cases more thoroughly. | priority | feature testing by roles we ve found that it s often difficult to understand exactly how a type of use interacts with caseflow for the hearing apps we have hearing coordinators vljs regional office staff and vsos that all interact with the same sets of pages slightly differently we d like to experiment with reorganizing our feature tests around user roles rather than simply pages we could either change the spec feature hearings directory to either or we think this would give folks who have not worked on hearings a better sense of how the app is used across roles while also allowing us to test specific use cases more thoroughly | 1 |
795,918 | 28,092,013,457 | IssuesEvent | 2023-03-30 13:37:54 | Photon-Health/client | https://api.github.com/repos/Photon-Health/client | closed | Re-add DataDog to App | 🛠 Medium Priority App | Datadog seems to have been crashing our App in certain situations. We should try re-adding with different settings and monitor for adverse performance on our application.
Could be in these settings:
``` JS
datadogRum.init({
applicationId: process.env.REACT_APP_DATADOG_RUM_APPLICATION_ID as string,
clientToken: process.env.REACT_APP_DATADOG_RUM_CLIENT_TOKEN as string,
site: 'datadoghq.com',
service: pkg.name,
env: process.env.REACT_APP_ENV_NAME,
version: pkg.version,
// HERE AND DOWN
sampleRate: 100,
sessionReplaySampleRate: 100,
trackInteractions: true,
trackResources: true,
trackLongTasks: true
});
```
also grab latest version
thx @jknapp25 for helping think through this
| 1.0 | Re-add DataDog to App - Datadog seems to have been crashing our App in certain situations. We should try re-adding with different settings and monitor for adverse performance on our application.
Could be in these settings:
``` JS
datadogRum.init({
applicationId: process.env.REACT_APP_DATADOG_RUM_APPLICATION_ID as string,
clientToken: process.env.REACT_APP_DATADOG_RUM_CLIENT_TOKEN as string,
site: 'datadoghq.com',
service: pkg.name,
env: process.env.REACT_APP_ENV_NAME,
version: pkg.version,
// HERE AND DOWN
sampleRate: 100,
sessionReplaySampleRate: 100,
trackInteractions: true,
trackResources: true,
trackLongTasks: true
});
```
also grab latest version
thx @jknapp25 for helping think through this
| priority | re add datadog to app datadog seems to have been crashing our app in certain situations we should try re adding with different settings and monitor for adverse performance on our application could be in these settings js datadogrum init applicationid process env react app datadog rum application id as string clienttoken process env react app datadog rum client token as string site datadoghq com service pkg name env process env react app env name version pkg version here and down samplerate sessionreplaysamplerate trackinteractions true trackresources true tracklongtasks true also grab latest version thx for helping think through this | 1 |
101,264 | 4,111,667,163 | IssuesEvent | 2016-06-07 07:25:26 | BugBusterSWE/documentation | https://api.github.com/repos/BugBusterSWE/documentation | closed | Inserire misurazioni BV e SV | priority:medium | *Documento in cui si trova il problema*:
Piano di Qualifica
Activity #511
*Descrizione del problema*:
Inserire misurazioni BV e SV
Link task: [https://bugbusters.teamwork.com/tasks/6938649](https://bugbusters.teamwork.com/tasks/6938649) | 1.0 | Inserire misurazioni BV e SV - *Documento in cui si trova il problema*:
Piano di Qualifica
Activity #511
*Descrizione del problema*:
Inserire misurazioni BV e SV
Link task: [https://bugbusters.teamwork.com/tasks/6938649](https://bugbusters.teamwork.com/tasks/6938649) | priority | inserire misurazioni bv e sv documento in cui si trova il problema piano di qualifica activity descrizione del problema inserire misurazioni bv e sv link task | 1 |
782,582 | 27,500,459,607 | IssuesEvent | 2023-03-05 16:35:30 | teambit/bit | https://api.github.com/repos/teambit/bit | closed | Yarn doesn't use .yarnrc registry | type/bug type/feature priority/medium | ### Describe the bug
Yarn doesn't use .yarnrc registry, but it uses .npmrc at least.
### Steps to Reproduce
1. Define private registry in .yarnrc
1. Use `"packageManager": "teambit.dependencies/yarn"`
1. Run `bbit install`
1. See 404 errors for private packages from registry defined in .yarnrc
### Expected Behavior
1. No http errors, e.g. 404.
1. All packages available in private registry are installed.
### Screenshots, exceptions and logs
```
YN0001: │ HTTPError: @company/graphql-api@npm:1.5.0: Response code 404 (Not Found)
at Request.<anonymous> (/Users/user/.bvm/versions/0.0.331/bit-0.0.331/node_modules/got/dist/source/as-promise/index.js:117:42)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:97:5)
```
### Specifications
* Bit version: 0.0.331
* Node version: v12.16.2
* npm / yarn version: 6.14.4 / 1.22.5
* Platform: MacOS Darwin Kernel Version 19.6.0
| 1.0 | Yarn doesn't use .yarnrc registry - ### Describe the bug
Yarn doesn't use .yarnrc registry, but it uses .npmrc at least.
### Steps to Reproduce
1. Define private registry in .yarnrc
1. Use `"packageManager": "teambit.dependencies/yarn"`
1. Run `bbit install`
1. See 404 errors for private packages from registry defined in .yarnrc
### Expected Behavior
1. No http errors, e.g. 404.
1. All packages available in private registry are installed.
### Screenshots, exceptions and logs
```
YN0001: │ HTTPError: @company/graphql-api@npm:1.5.0: Response code 404 (Not Found)
at Request.<anonymous> (/Users/user/.bvm/versions/0.0.331/bit-0.0.331/node_modules/got/dist/source/as-promise/index.js:117:42)
at runMicrotasks (<anonymous>)
at processTicksAndRejections (internal/process/task_queues.js:97:5)
```
### Specifications
* Bit version: 0.0.331
* Node version: v12.16.2
* npm / yarn version: 6.14.4 / 1.22.5
* Platform: MacOS Darwin Kernel Version 19.6.0
| priority | yarn doesn t use yarnrc registry describe the bug yarn doesn t use yarnrc registry but it uses npmrc at least steps to reproduce define private registry in yarnrc use packagemanager teambit dependencies yarn run bbit install see errors for private packages from registry defined in yarnrc expected behavior no http errors e g all packages available in private registry are installed screenshots exceptions and logs │ httperror company graphql api npm response code not found at request users user bvm versions bit node modules got dist source as promise index js at runmicrotasks at processticksandrejections internal process task queues js specifications bit version node version npm yarn version platform macos darwin kernel version | 1 |
343,918 | 10,338,314,829 | IssuesEvent | 2019-09-03 16:37:48 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | k_busy_wait not working when using 32KHz timer driver | bug has-pr priority: medium | k_busy_wait results in inaccurate delays causing MEC1501 I2C driver to always report timeout.
This breaks I2C tests in this board.
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=mec15xxevb_assy6853
3. make
**Expected behavior**
k_busy_wait should result in expected delays.
**Impact**
This affects behavior from drivers relaying in k_busy_wait.
Additionally usage of I2C driver cause most tests to fail
**Additional context**
| 1.0 | k_busy_wait not working when using 32KHz timer driver - k_busy_wait results in inaccurate delays causing MEC1501 I2C driver to always report timeout.
This breaks I2C tests in this board.
**To Reproduce**
Steps to reproduce the behavior:
1. mkdir build; cd build
2. cmake -DBOARD=mec15xxevb_assy6853
3. make
**Expected behavior**
k_busy_wait should result in expected delays.
**Impact**
This affects behavior from drivers relaying in k_busy_wait.
Additionally usage of I2C driver cause most tests to fail
**Additional context**
| priority | k busy wait not working when using timer driver k busy wait results in inaccurate delays causing driver to always report timeout this breaks tests in this board to reproduce steps to reproduce the behavior mkdir build cd build cmake dboard make expected behavior k busy wait should result in expected delays impact this affects behavior from drivers relaying in k busy wait additionally usage of driver cause most tests to fail additional context | 1 |
212,739 | 7,242,273,315 | IssuesEvent | 2018-02-14 06:44:49 | ODIQueensland/data-curator | https://api.github.com/repos/ODIQueensland/data-curator | closed | TSV file in exported Data Package should have correct format and CSV Dialect | env:MacOS env:Windows est:Moderate fn:Export priority:Medium problem:Bug | ### Expected Behaviour (for problems)
TSV file in an exported Data Package should have the correct [format](http://frictionlessdata.io/specs/data-resource/#optional-properties) in the data resource and delimiter in [CSV Dialect](http://frictionlessdata.io/specs/csv-dialect/#specification)
### Current Behaviour
datapackage.json has:
- format: csv
- no csv dialect delimiter: \t
### Steps to Reproduce
1. create data
2. save as TSV
3. export data package
4. inspect datapackage.json
### Your Environment
* Data Curator version 0.5.0
* Operating System and version Windows 7
| 1.0 | TSV file in exported Data Package should have correct format and CSV Dialect - ### Expected Behaviour (for problems)
TSV file in an exported Data Package should have the correct [format](http://frictionlessdata.io/specs/data-resource/#optional-properties) in the data resource and delimiter in [CSV Dialect](http://frictionlessdata.io/specs/csv-dialect/#specification)
### Current Behaviour
datapackage.json has:
- format: csv
- no csv dialect delimiter: \t
### Steps to Reproduce
1. create data
2. save as TSV
3. export data package
4. inspect datapackage.json
### Your Environment
* Data Curator version 0.5.0
* Operating System and version Windows 7
| priority | tsv file in exported data package should have correct format and csv dialect expected behaviour for problems tsv file in an exported data package should have the correct in the data resource and delimiter in current behaviour datapackage json has format csv no csv dialect delimiter t steps to reproduce create data save as tsv export data package inspect datapackage json your environment data curator version operating system and version windows | 1 |
487,524 | 14,047,830,253 | IssuesEvent | 2020-11-02 07:52:33 | Puzzlepart/prosjektportalen | https://api.github.com/repos/Puzzlepart/prosjektportalen | opened | Possiblity to sort "Om prosjektet" fields | Complexity: medium Priority: low enhancement | **Is your feature request related to a problem? Please describe.**
Three assist customers has asked how to sort the fields under "Om prosjektet". This is no longer working if it worked before, and is therefore a requested feature. (Assist-2926) Before there was a contentType "Prosjektforside", if I remember correctly this is where the field where ordered before, but this Ct has been deprecated a long time ago.
**Describe the solution you'd like**
Sort the fields under "Om prosjektet". Either by the order in a contenttype (Prosjektegenskaper) or from one of the fields lists.
**Where in the Project Portal is this relevant?**
ProjectHome SitePage

| 1.0 | Possiblity to sort "Om prosjektet" fields - **Is your feature request related to a problem? Please describe.**
Three assist customers has asked how to sort the fields under "Om prosjektet". This is no longer working if it worked before, and is therefore a requested feature. (Assist-2926) Before there was a contentType "Prosjektforside", if I remember correctly this is where the field where ordered before, but this Ct has been deprecated a long time ago.
**Describe the solution you'd like**
Sort the fields under "Om prosjektet". Either by the order in a contenttype (Prosjektegenskaper) or from one of the fields lists.
**Where in the Project Portal is this relevant?**
ProjectHome SitePage

| priority | possiblity to sort om prosjektet fields is your feature request related to a problem please describe three assist customers has asked how to sort the fields under om prosjektet this is no longer working if it worked before and is therefore a requested feature assist before there was a contenttype prosjektforside if i remember correctly this is where the field where ordered before but this ct has been deprecated a long time ago describe the solution you d like sort the fields under om prosjektet either by the order in a contenttype prosjektegenskaper or from one of the fields lists where in the project portal is this relevant projecthome sitepage | 1 |
340,144 | 10,267,301,399 | IssuesEvent | 2019-08-23 00:57:24 | minio/minio | https://api.github.com/repos/minio/minio | closed | Lifecycle delete not working | priority: medium triage | <!--- Provide a general summary of the issue in the Title above -->
Followed examples in https://github.com/minio/minio/tree/master/docs/lifecycle, bucket contents are not deleted after expiry. This is working for others so assuming something incorrect in my environment.
## Expected Behavior
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
Objects in bucket should be deleted after expiry date is reached
## Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
Objects remain in bucket
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
Unsure
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. Create bucket 'temporary-uploads' and upload objects to it
```
mc ls min/temporary-uploads/
[2019-08-15 16:26:40 PDT] 26B issue.1
[2019-08-15 16:26:40 PDT] 26B issue.10
[2019-08-15 16:26:40 PDT] 26B issue.11
[2019-08-15 16:26:40 PDT] 26B issue.12
[2019-08-15 16:26:40 PDT] 26B issue.13
[2019-08-15 16:26:40 PDT] 26B issue.14
[2019-08-15 16:26:40 PDT] 26B issue.15
[2019-08-15 16:26:40 PDT] 26B issue.16
[2019-08-15 16:26:40 PDT] 26B issue.17
[2019-08-15 16:26:40 PDT] 26B issue.18
[2019-08-15 16:26:40 PDT] 26B issue.19
[2019-08-15 16:26:40 PDT] 26B issue.2
[2019-08-15 16:26:40 PDT] 26B issue.20
[2019-08-15 16:26:40 PDT] 26B issue.3
[2019-08-15 16:26:40 PDT] 26B issue.4
[2019-08-15 16:26:40 PDT] 26B issue.5
[2019-08-15 16:26:40 PDT] 26B issue.6
[2019-08-15 16:26:40 PDT] 26B issue.7
[2019-08-15 16:26:40 PDT] 26B issue.8
[2019-08-15 16:26:40 PDT] 26B issue.9
```
2. Create json file with policy
```
cat ~/bucket-lifecycle.json
{
"Rules": [
{
"Expiration": {
"Date": "2019-08-16T00:00:00.000Z"
},
"ID": "Delete very old messenger pictures",
"Filter": {
"Prefix": "uploads/2015/"
},
"Status": "Enabled"
},
{
"Expiration": {
"Date": "2019-08-17T00:00:00.000Z"
},
"ID": "Delete temporary uploads",
"Filter": {
"Prefix": "temporary-uploads/"
},
"Status": "Enabled"
}
]
}
```
3. Apply policy
```
export AWS_ACCESS_KEY_ID="your-access-key"
export AWS_SECRET_ACCESS_KEY="your-secret-key"
aws s3api put-bucket-lifecycle-configuration --bucket temporary-uploads --endpoint-url http://localhost:9000 --lifecycle-configuration file://bucket-lifecycle.json
```
4. Wait for expiration, objects are not deleted
## Context
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## Regression
<!-- Is this issue a regression? (Yes / No) -->
<!-- If Yes, optionally please include minio version or commit id or PR# that caused this regression, if you have these details. -->
No
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio version`):
Version: 2019-08-14T20:37:41Z
* Operating System and version (`uname -a`):
Ubuntu 18.04
| 1.0 | Lifecycle delete not working - <!--- Provide a general summary of the issue in the Title above -->
Followed examples in https://github.com/minio/minio/tree/master/docs/lifecycle, bucket contents are not deleted after expiry. This is working for others so assuming something incorrect in my environment.
## Expected Behavior
<!--- If you're describing a bug, tell us what should happen -->
<!--- If you're suggesting a change/improvement, tell us how it should work -->
Objects in bucket should be deleted after expiry date is reached
## Current Behavior
<!--- If describing a bug, tell us what happens instead of the expected behavior -->
<!--- If suggesting a change/improvement, explain the difference from current behavior -->
Objects remain in bucket
## Possible Solution
<!--- Not obligatory, but suggest a fix/reason for the bug, -->
<!--- or ideas how to implement the addition or change -->
Unsure
## Steps to Reproduce (for bugs)
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
1. Create bucket 'temporary-uploads' and upload objects to it
```
mc ls min/temporary-uploads/
[2019-08-15 16:26:40 PDT] 26B issue.1
[2019-08-15 16:26:40 PDT] 26B issue.10
[2019-08-15 16:26:40 PDT] 26B issue.11
[2019-08-15 16:26:40 PDT] 26B issue.12
[2019-08-15 16:26:40 PDT] 26B issue.13
[2019-08-15 16:26:40 PDT] 26B issue.14
[2019-08-15 16:26:40 PDT] 26B issue.15
[2019-08-15 16:26:40 PDT] 26B issue.16
[2019-08-15 16:26:40 PDT] 26B issue.17
[2019-08-15 16:26:40 PDT] 26B issue.18
[2019-08-15 16:26:40 PDT] 26B issue.19
[2019-08-15 16:26:40 PDT] 26B issue.2
[2019-08-15 16:26:40 PDT] 26B issue.20
[2019-08-15 16:26:40 PDT] 26B issue.3
[2019-08-15 16:26:40 PDT] 26B issue.4
[2019-08-15 16:26:40 PDT] 26B issue.5
[2019-08-15 16:26:40 PDT] 26B issue.6
[2019-08-15 16:26:40 PDT] 26B issue.7
[2019-08-15 16:26:40 PDT] 26B issue.8
[2019-08-15 16:26:40 PDT] 26B issue.9
```
2. Create json file with policy
```
cat ~/bucket-lifecycle.json
{
"Rules": [
{
"Expiration": {
"Date": "2019-08-16T00:00:00.000Z"
},
"ID": "Delete very old messenger pictures",
"Filter": {
"Prefix": "uploads/2015/"
},
"Status": "Enabled"
},
{
"Expiration": {
"Date": "2019-08-17T00:00:00.000Z"
},
"ID": "Delete temporary uploads",
"Filter": {
"Prefix": "temporary-uploads/"
},
"Status": "Enabled"
}
]
}
```
3. Apply policy
```
export AWS_ACCESS_KEY_ID="your-access-key"
export AWS_SECRET_ACCESS_KEY="your-secret-key"
aws s3api put-bucket-lifecycle-configuration --bucket temporary-uploads --endpoint-url http://localhost:9000 --lifecycle-configuration file://bucket-lifecycle.json
```
4. Wait for expiration, objects are not deleted
## Context
<!--- How has this issue affected you? What are you trying to accomplish? -->
<!--- Providing context helps us come up with a solution that is most useful in the real world -->
## Regression
<!-- Is this issue a regression? (Yes / No) -->
<!-- If Yes, optionally please include minio version or commit id or PR# that caused this regression, if you have these details. -->
No
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Version used (`minio version`):
Version: 2019-08-14T20:37:41Z
* Operating System and version (`uname -a`):
Ubuntu 18.04
| priority | lifecycle delete not working followed examples in bucket contents are not deleted after expiry this is working for others so assuming something incorrect in my environment expected behavior objects in bucket should be deleted after expiry date is reached current behavior objects remain in bucket possible solution unsure steps to reproduce for bugs create bucket temporary uploads and upload objects to it mc ls min temporary uploads issue issue issue issue issue issue issue issue issue issue issue issue issue issue issue issue issue issue issue issue create json file with policy cat bucket lifecycle json rules expiration date id delete very old messenger pictures filter prefix uploads status enabled expiration date id delete temporary uploads filter prefix temporary uploads status enabled apply policy export aws access key id your access key export aws secret access key your secret key aws put bucket lifecycle configuration bucket temporary uploads endpoint url lifecycle configuration file bucket lifecycle json wait for expiration objects are not deleted context regression no your environment version used minio version version operating system and version uname a ubuntu | 1 |
253,587 | 8,058,128,442 | IssuesEvent | 2018-08-02 17:28:38 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | set_stats fails in callback plugin on isolated node | component:api flag:🎱 priority:medium state:needs_devel type:bug | ##### SUMMARY
When using set_stats, callback is not written. An error is generated, but playbook run successfully. Subsequent playbook in a workflow template does not receive set_stats on execution.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
awx/lib/awx_display_callback/module.py
##### ANSIBLE VERSION
```
ansible-playbook 2.6.1
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/var/lib/awx/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible-playbook
python version = 2.7.5 (default, Feb 20 2018, 09:19:12) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)]
Using /etc/ansible/ansible.cfg as config file
```
##### OS / ENVIRONMENT
Red Hat Enterprise Linux Server release 7.4 (Maipo)
Ansible Tower 3.2.5
Ansible 2.6.1
##### STEPS TO REPRODUCE
```yaml
---
- hosts: tower
become: no
gather_facts: no
tasks:
- set_stats:
data:
extra_vars:
some_fact: 'foo'
some_fact: 'foo'
```
##### EXPECTED RESULTS
```
CUSTOM STATS: ******************************************************************
RUN: { "extra_vars": { "some_fact": "foo" }, "some_fact": "foo"}
```
##### ACTUAL RESULTS
```
[WARNING]: Failure using method (v2_playbook_on_stats) in callback plugin
(<awx_display_callback.module.AWXDefaultCallbackModule object at
0x7fa9b563bf10>): [Errno 17] File exists:
'/var/lib/awx/.ansible/tmp/awx_3094_b7FWcd/artifacts'
Callback Exception:
File "/usr/lib/python2.7/site-packages/ansible/executor/task_queue_manager.py", line 374, in send_callback
method(*new_args, **kwargs)
File "/lib/python2.7/site-packages/awx/lib/awx_display_callback/module.py", line 323, in v2_playbook_on_stats
File "/usr/lib64/python2.7/os.py", line 157, in makedirs
mkdir(name, mode)
```
##### CODE TO BLAME
https://github.com/ansible/awx/blob/devel/awx/lib/awx_display_callback/module.py#L311
( last line in block below)
```
if custom_artifact_data:
# create the directory for custom stats artifacts to live in (if it doesn't exist)
custom_artifacts_dir = os.path.join(os.getenv('AWX_PRIVATE_DATA_DIR'), 'artifacts')
os.makedirs(custom_artifacts_dir, mode=stat.S_IXUSR + stat.S_IWUSR + stat.S_IRUSR)
```
It appears that on an isolated node, the custom_artifact_dir is automatically created. Add if condition to os.makedirs fixes the observed behavior:
```
if custom_artifact_data:
# create the directory for custom stats artifacts to live in (if it doesn't exist)
custom_artifacts_dir = os.path.join(os.getenv('AWX_PRIVATE_DATA_DIR'), 'artifacts')
if not os.path.isdir(custom_artifacts_dir):
os.makedirs(custom_artifacts_dir, mode=stat.S_IXUSR + stat.S_IWUSR + stat.S_IRUSR)
```
| 1.0 | set_stats fails in callback plugin on isolated node - ##### SUMMARY
When using set_stats, callback is not written. An error is generated, but playbook run successfully. Subsequent playbook in a workflow template does not receive set_stats on execution.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
awx/lib/awx_display_callback/module.py
##### ANSIBLE VERSION
```
ansible-playbook 2.6.1
config file = /etc/ansible/ansible.cfg
configured module search path = [u'/var/lib/awx/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules']
ansible python module location = /usr/lib/python2.7/site-packages/ansible
executable location = /usr/bin/ansible-playbook
python version = 2.7.5 (default, Feb 20 2018, 09:19:12) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)]
Using /etc/ansible/ansible.cfg as config file
```
##### OS / ENVIRONMENT
Red Hat Enterprise Linux Server release 7.4 (Maipo)
Ansible Tower 3.2.5
Ansible 2.6.1
##### STEPS TO REPRODUCE
```yaml
---
- hosts: tower
become: no
gather_facts: no
tasks:
- set_stats:
data:
extra_vars:
some_fact: 'foo'
some_fact: 'foo'
```
##### EXPECTED RESULTS
```
CUSTOM STATS: ******************************************************************
RUN: { "extra_vars": { "some_fact": "foo" }, "some_fact": "foo"}
```
##### ACTUAL RESULTS
```
[WARNING]: Failure using method (v2_playbook_on_stats) in callback plugin
(<awx_display_callback.module.AWXDefaultCallbackModule object at
0x7fa9b563bf10>): [Errno 17] File exists:
'/var/lib/awx/.ansible/tmp/awx_3094_b7FWcd/artifacts'
Callback Exception:
File "/usr/lib/python2.7/site-packages/ansible/executor/task_queue_manager.py", line 374, in send_callback
method(*new_args, **kwargs)
File "/lib/python2.7/site-packages/awx/lib/awx_display_callback/module.py", line 323, in v2_playbook_on_stats
File "/usr/lib64/python2.7/os.py", line 157, in makedirs
mkdir(name, mode)
```
##### CODE TO BLAME
https://github.com/ansible/awx/blob/devel/awx/lib/awx_display_callback/module.py#L311
( last line in block below)
```
if custom_artifact_data:
# create the directory for custom stats artifacts to live in (if it doesn't exist)
custom_artifacts_dir = os.path.join(os.getenv('AWX_PRIVATE_DATA_DIR'), 'artifacts')
os.makedirs(custom_artifacts_dir, mode=stat.S_IXUSR + stat.S_IWUSR + stat.S_IRUSR)
```
It appears that on an isolated node, the custom_artifact_dir is automatically created. Add if condition to os.makedirs fixes the observed behavior:
```
if custom_artifact_data:
# create the directory for custom stats artifacts to live in (if it doesn't exist)
custom_artifacts_dir = os.path.join(os.getenv('AWX_PRIVATE_DATA_DIR'), 'artifacts')
if not os.path.isdir(custom_artifacts_dir):
os.makedirs(custom_artifacts_dir, mode=stat.S_IXUSR + stat.S_IWUSR + stat.S_IRUSR)
```
| priority | set stats fails in callback plugin on isolated node summary when using set stats callback is not written an error is generated but playbook run successfully subsequent playbook in a workflow template does not receive set stats on execution issue type bug report component name awx lib awx display callback module py ansible version ansible playbook config file etc ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location usr bin ansible playbook python version default feb using etc ansible ansible cfg as config file os environment red hat enterprise linux server release maipo ansible tower ansible steps to reproduce yaml hosts tower become no gather facts no tasks set stats data extra vars some fact foo some fact foo expected results custom stats run extra vars some fact foo some fact foo actual results failure using method playbook on stats in callback plugin awx display callback module awxdefaultcallbackmodule object at file exists var lib awx ansible tmp awx artifacts callback exception file usr lib site packages ansible executor task queue manager py line in send callback method new args kwargs file lib site packages awx lib awx display callback module py line in playbook on stats file usr os py line in makedirs mkdir name mode code to blame last line in block below if custom artifact data create the directory for custom stats artifacts to live in if it doesn t exist custom artifacts dir os path join os getenv awx private data dir artifacts os makedirs custom artifacts dir mode stat s ixusr stat s iwusr stat s irusr it appears that on an isolated node the custom artifact dir is automatically created add if condition to os makedirs fixes the observed behavior if custom artifact data create the directory for custom stats artifacts to live in if it doesn t exist custom artifacts dir os path join os getenv awx private data dir artifacts if not os path isdir custom artifacts dir os makedirs custom artifacts dir mode stat s ixusr stat s iwusr stat s irusr | 1 |
209,253 | 7,166,823,721 | IssuesEvent | 2018-01-29 18:29:18 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | ram_report not working for qemu targets | bug priority: medium | commit 4bd904ccaec5ebcddfb8d9be6290facbd38beec7 introduced a bug where ram_reports are not working anymore with qemu_x86:
add CONFIG_BUILD_OUTPUT_BIN=y to the prj.conf
when building with ram_report for qemu_x86, you get:
```
Traceback (most recent call last):
File "/work/zephyr/github/zephyr/scripts/footprint/size_report", line 420, in <module>
base + '/', None)
File "/work/zephyr/github/zephyr/scripts/footprint/size_report", line 194, in generate_target_memory_section
symbols_paths = dict(load_symbols_and_paths(bin_nm, elf_file_abs, source_dir))
File "/work/zephyr/github/zephyr/scripts/footprint/size_report", line 57, in load_symbols_and_paths
processed_path = Path(path).relative_to(Path(path_to_strip))
File "/usr/lib64/python3.6/pathlib.py", line 870, in relative_to
.format(str(self), str(formatted)))
ValueError: '/usr/src/debug/libgcc/6.2.0-r0/gcc-6.2.0/build.i586-zephyr-elf.i586-zephyr-elf/libgcc/../../../../../../../work-shared/gcc-6.2.0-r0/gcc-6.2.0/libgcc/libgcc2.c' does not start with '/work/zephyr/github/zephyr'
make[4]: *** [zephyr/cmake/reports/CMakeFiles/ram_report.dir/build.make:57: zephyr/cmake/reports/CMakeFiles/ram_report] Error 1
make[3]: *** [CMakeFiles/Makefile2:1951: zephyr/cmake/reports/CMakeFiles/ram_report.dir/all] Error 2
make[2]: *** [CMakeFiles/Makefile2:1958: zephyr/cmake/reports/CMakeFiles/ram_report.dir/rule] Error 2
make[1]: *** [Makefile:495: ram_report] Error 2
make: *** [/home/nashif/bin/zb:23: ram_report] Error 2
``` | 1.0 | ram_report not working for qemu targets - commit 4bd904ccaec5ebcddfb8d9be6290facbd38beec7 introduced a bug where ram_reports are not working anymore with qemu_x86:
add CONFIG_BUILD_OUTPUT_BIN=y to the prj.conf
when building with ram_report for qemu_x86, you get:
```
Traceback (most recent call last):
File "/work/zephyr/github/zephyr/scripts/footprint/size_report", line 420, in <module>
base + '/', None)
File "/work/zephyr/github/zephyr/scripts/footprint/size_report", line 194, in generate_target_memory_section
symbols_paths = dict(load_symbols_and_paths(bin_nm, elf_file_abs, source_dir))
File "/work/zephyr/github/zephyr/scripts/footprint/size_report", line 57, in load_symbols_and_paths
processed_path = Path(path).relative_to(Path(path_to_strip))
File "/usr/lib64/python3.6/pathlib.py", line 870, in relative_to
.format(str(self), str(formatted)))
ValueError: '/usr/src/debug/libgcc/6.2.0-r0/gcc-6.2.0/build.i586-zephyr-elf.i586-zephyr-elf/libgcc/../../../../../../../work-shared/gcc-6.2.0-r0/gcc-6.2.0/libgcc/libgcc2.c' does not start with '/work/zephyr/github/zephyr'
make[4]: *** [zephyr/cmake/reports/CMakeFiles/ram_report.dir/build.make:57: zephyr/cmake/reports/CMakeFiles/ram_report] Error 1
make[3]: *** [CMakeFiles/Makefile2:1951: zephyr/cmake/reports/CMakeFiles/ram_report.dir/all] Error 2
make[2]: *** [CMakeFiles/Makefile2:1958: zephyr/cmake/reports/CMakeFiles/ram_report.dir/rule] Error 2
make[1]: *** [Makefile:495: ram_report] Error 2
make: *** [/home/nashif/bin/zb:23: ram_report] Error 2
``` | priority | ram report not working for qemu targets commit introduced a bug where ram reports are not working anymore with qemu add config build output bin y to the prj conf when building with ram report for qemu you get traceback most recent call last file work zephyr github zephyr scripts footprint size report line in base none file work zephyr github zephyr scripts footprint size report line in generate target memory section symbols paths dict load symbols and paths bin nm elf file abs source dir file work zephyr github zephyr scripts footprint size report line in load symbols and paths processed path path path relative to path path to strip file usr pathlib py line in relative to format str self str formatted valueerror usr src debug libgcc gcc build zephyr elf zephyr elf libgcc work shared gcc gcc libgcc c does not start with work zephyr github zephyr make error make error make error make error make error | 1 |
136,581 | 5,285,278,739 | IssuesEvent | 2017-02-08 04:02:19 | dhowe/AdNauseam | https://api.github.com/repos/dhowe/AdNauseam | closed | Reactive button in lists doesn't reactive the list in the background | Bug PRIORITY: Medium | To recreate:
1.disable easylist -> apply changes
2.reactive the list through reactive button in notification.
3.go to nytimes.com, ad parsing/hiding seem to not working anymore because easylist is still disabled
4.Refresh the background console (third party filters are reloaded) -> ad parsing/hiding works again. | 1.0 | Reactive button in lists doesn't reactive the list in the background - To recreate:
1.disable easylist -> apply changes
2.reactive the list through reactive button in notification.
3.go to nytimes.com, ad parsing/hiding seem to not working anymore because easylist is still disabled
4.Refresh the background console (third party filters are reloaded) -> ad parsing/hiding works again. | priority | reactive button in lists doesn t reactive the list in the background to recreate disable easylist apply changes reactive the list through reactive button in notification go to nytimes com ad parsing hiding seem to not working anymore because easylist is still disabled refresh the background console third party filters are reloaded ad parsing hiding works again | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.