Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 957 | labels stringlengths 4 795 | body stringlengths 1 259k | index stringclasses 12
values | text_combine stringlengths 96 259k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
617,764 | 19,404,167,081 | IssuesEvent | 2021-12-19 18:06:24 | codidact/qpixel | https://api.github.com/repos/codidact/qpixel | closed | Preview renders unsupported HTML tags, but final post (correctly) doesn't | area: html/css/js type: bug priority: medium complexity: unassessed | https://meta.codidact.com/posts/284503
https://meta.codidact.com/posts/284505
When composing a post, an unsupported HTML tag (`div` and `kbd` in these reports) was accepted and rendered. However, after submission the post doesn't render it, which is correct because we don't support the tag. This is confusing to users who haven't memorized (or don't look up) which HTML tags we do/don't support.
I know there's at least one other issue about differences between preview and final rendering, though I couldn't find it. We use different libraries in the two cases so differences aren't surprising. If we can't use the same library (or logic) to render the Markdown in both cases, is there anything we can do to provide some feedback when editing? Can we "lint" the post body and indicate if we found something? Maybe, as with missing alt text, that could be something we do when the user clicks the "post" button, so it doesn't have to be a performance drain.
| 1.0 | Preview renders unsupported HTML tags, but final post (correctly) doesn't - https://meta.codidact.com/posts/284503
https://meta.codidact.com/posts/284505
When composing a post, an unsupported HTML tag (`div` and `kbd` in these reports) was accepted and rendered. However, after submission the post doesn't render it, which is correct because we don't support the tag. This is confusing to users who haven't memorized (or don't look up) which HTML tags we do/don't support.
I know there's at least one other issue about differences between preview and final rendering, though I couldn't find it. We use different libraries in the two cases so differences aren't surprising. If we can't use the same library (or logic) to render the Markdown in both cases, is there anything we can do to provide some feedback when editing? Can we "lint" the post body and indicate if we found something? Maybe, as with missing alt text, that could be something we do when the user clicks the "post" button, so it doesn't have to be a performance drain.
| priority | preview renders unsupported html tags but final post correctly doesn t when composing a post an unsupported html tag div and kbd in these reports was accepted and rendered however after submission the post doesn t render it which is correct because we don t support the tag this is confusing to users who haven t memorized or don t look up which html tags we do don t support i know there s at least one other issue about differences between preview and final rendering though i couldn t find it we use different libraries in the two cases so differences aren t surprising if we can t use the same library or logic to render the markdown in both cases is there anything we can do to provide some feedback when editing can we lint the post body and indicate if we found something maybe as with missing alt text that could be something we do when the user clicks the post button so it doesn t have to be a performance drain | 1 |
567,309 | 16,854,480,669 | IssuesEvent | 2021-06-21 03:26:56 | adirh3/Fluent-Search | https://api.github.com/repos/adirh3/Fluent-Search | closed | Change size in MB to look cleaner | Medium Priority UI/UX bug | This is no-brainer and standard interface
kilobyte, megabyte, gigabyte, terabyte instead of "Size of MB 0.03891485883838". This looks very hard to read for all users, novice and expert alike.

| 1.0 | Change size in MB to look cleaner - This is no-brainer and standard interface
kilobyte, megabyte, gigabyte, terabyte instead of "Size of MB 0.03891485883838". This looks very hard to read for all users, novice and expert alike.

| priority | change size in mb to look cleaner this is no brainer and standard interface kilobyte megabyte gigabyte terabyte instead of size of mb this looks very hard to read for all users novice and expert alike | 1 |
627,412 | 19,904,337,359 | IssuesEvent | 2022-01-25 11:10:40 | docker-mailserver/docker-mailserver | https://api.github.com/repos/docker-mailserver/docker-mailserver | opened | Can't create email accounts while enabling LDAP | kind/bug meta/needs triage priority/medium | ### Miscellaneous first checks
- [X] I checked that all ports are open and not blocked by my ISP / hosting provider.
- [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct.
### Affected Component(s)
Mail creation, deletion and listing
### What happened and when does this occur?
```Markdown
While trying to create the email account using the setup.sh; I'm encountering the error below.
Waiting for dovecot to create /var/mail/deol.com/deoltito...
Waiting for dovecot to create /var/mail/deol.com/deoltito...
Waiting for dovecot to create /var/mail/deol.com/deoltito...
Waiting for dovecot to create /var/mail/deol.com/deoltito...
This message goes on until I stops it. However the email account won't be created.
When I try to list the available email accounts, this particular error is seen under it.
==========
# ./setup.sh email list
Fatal: Unknown command 'quota', but plugin quota exists. Try to set mail_plugins=quota
/usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ")
/usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ")
* new@deol.com ( / ) [%]
==========
```
### What did you expect to happen?
```Markdown
I believe it has something to with the LDAP integration. When I disable LDAP in the compose file, the email accounts can be created, listed and deleted without any issues.
```
### How do we replicate the issue?
```Markdown
1. Try to create a compose file with LDAP enabled and integrated in it
2. Try creating an email account after that
3. Try to list or delete the email accounts too
...
```
### DMS version
v10.4.0
### What operating system is DMS running on?
Linux
### What instruction set architecture is DMS running on?
x86_64 / AMD64
### What container orchestration tool are you using?
Docker Compose
### docker-compose.yml
```yaml
version: '3.8'
services:
mailserver:
image: docker.io/mailserver/docker-mailserver:latest
container_name: mailserver
hostname: mail
domainname: deol.com
ports:
- "25:25"
- "143:143"
- "587:587"
- "993:993"
volumes:
- ./docker-data/dms/mail-data/:/var/mail/
- ./docker-data/dms/mail-state/:/var/mail-state/
- ./docker-data/dms/mail-logs/:/var/log/mail/
- ./docker-data/dms/config/:/tmp/docker-mailserver/
- /etc/localtime:/etc/localtime:ro
environment:
- ENABLE_SPAMASSASSIN=1
- SPAMASSASSIN_SPAM_TO_INBOX=1
- ENABLE_CLAMAV=1
- ENABLE_FAIL2BAN=1
- ENABLE_POSTGREY=1
- ENABLE_SASLAUTHD=1
- ONE_DIR=1
- DMS_DEBUG=1
- ENABLE_LDAP=1
- LDAP_SERVER_HOST=LDAPNEW # your ldap container/IP/ServerName
- LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com
- LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com
- LDAP_BIND_PW=
- ENABLE_SASLAUTHD=1
- SASLAUTHD_MECHANISMS=ldap
- SASLAUTHD_LDAP_SERVER=LDAPNEW
- SASLAUTHD_LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com
- SASLAUTHD_LDAP_PASSWORD=
- SASLAUTHD_LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com
- SASLAUTHD_LDAP_FILTER=(&(objectClass=PostfixBookMailAccount)(uniqueIdentifier=%U))
- POSTMASTER_ADDRESS=postmaster@deol.com
- POSTFIX_MESSAGE_SIZE_LIMIT=100000000
cap_add:
- NET_ADMIN
- SYS_PTRACE
```
### Relevant log output
_No response_
### Other relevant information
_No response_
### What level of experience do you have with Docker and mail servers?
- [ ] I am inexperienced with docker
- [ ] I am inexperienced with mail servers
- [ ] I am uncomfortable with the CLI
### Code of conduct
- [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree
- [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution
### Improvements to this form?
_No response_ | 1.0 | Can't create email accounts while enabling LDAP - ### Miscellaneous first checks
- [X] I checked that all ports are open and not blocked by my ISP / hosting provider.
- [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct.
### Affected Component(s)
Mail creation, deletion and listing
### What happened and when does this occur?
```Markdown
While trying to create the email account using the setup.sh; I'm encountering the error below.
Waiting for dovecot to create /var/mail/deol.com/deoltito...
Waiting for dovecot to create /var/mail/deol.com/deoltito...
Waiting for dovecot to create /var/mail/deol.com/deoltito...
Waiting for dovecot to create /var/mail/deol.com/deoltito...
This message goes on until I stops it. However the email account won't be created.
When I try to list the available email accounts, this particular error is seen under it.
==========
# ./setup.sh email list
Fatal: Unknown command 'quota', but plugin quota exists. Try to set mail_plugins=quota
/usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ")
/usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ")
* new@deol.com ( / ) [%]
==========
```
### What did you expect to happen?
```Markdown
I believe it has something to with the LDAP integration. When I disable LDAP in the compose file, the email accounts can be created, listed and deleted without any issues.
```
### How do we replicate the issue?
```Markdown
1. Try to create a compose file with LDAP enabled and integrated in it
2. Try creating an email account after that
3. Try to list or delete the email accounts too
...
```
### DMS version
v10.4.0
### What operating system is DMS running on?
Linux
### What instruction set architecture is DMS running on?
x86_64 / AMD64
### What container orchestration tool are you using?
Docker Compose
### docker-compose.yml
```yaml
version: '3.8'
services:
mailserver:
image: docker.io/mailserver/docker-mailserver:latest
container_name: mailserver
hostname: mail
domainname: deol.com
ports:
- "25:25"
- "143:143"
- "587:587"
- "993:993"
volumes:
- ./docker-data/dms/mail-data/:/var/mail/
- ./docker-data/dms/mail-state/:/var/mail-state/
- ./docker-data/dms/mail-logs/:/var/log/mail/
- ./docker-data/dms/config/:/tmp/docker-mailserver/
- /etc/localtime:/etc/localtime:ro
environment:
- ENABLE_SPAMASSASSIN=1
- SPAMASSASSIN_SPAM_TO_INBOX=1
- ENABLE_CLAMAV=1
- ENABLE_FAIL2BAN=1
- ENABLE_POSTGREY=1
- ENABLE_SASLAUTHD=1
- ONE_DIR=1
- DMS_DEBUG=1
- ENABLE_LDAP=1
- LDAP_SERVER_HOST=LDAPNEW # your ldap container/IP/ServerName
- LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com
- LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com
- LDAP_BIND_PW=
- ENABLE_SASLAUTHD=1
- SASLAUTHD_MECHANISMS=ldap
- SASLAUTHD_LDAP_SERVER=LDAPNEW
- SASLAUTHD_LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com
- SASLAUTHD_LDAP_PASSWORD=
- SASLAUTHD_LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com
- SASLAUTHD_LDAP_FILTER=(&(objectClass=PostfixBookMailAccount)(uniqueIdentifier=%U))
- POSTMASTER_ADDRESS=postmaster@deol.com
- POSTFIX_MESSAGE_SIZE_LIMIT=100000000
cap_add:
- NET_ADMIN
- SYS_PTRACE
```
### Relevant log output
_No response_
### Other relevant information
_No response_
### What level of experience do you have with Docker and mail servers?
- [ ] I am inexperienced with docker
- [ ] I am inexperienced with mail servers
- [ ] I am uncomfortable with the CLI
### Code of conduct
- [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree
- [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution
### Improvements to this form?
_No response_ | priority | can t create email accounts while enabling ldap miscellaneous first checks i checked that all ports are open and not blocked by my isp hosting provider i know that ssl errors are likely the result of a wrong setup on the user side and not caused by dms itself i m confident my setup is correct affected component s mail creation deletion and listing what happened and when does this occur markdown while trying to create the email account using the setup sh i m encountering the error below waiting for dovecot to create var mail deol com deoltito waiting for dovecot to create var mail deol com deoltito waiting for dovecot to create var mail deol com deoltito waiting for dovecot to create var mail deol com deoltito this message goes on until i stops it however the email account won t be created when i try to list the available email accounts this particular error is seen under it setup sh email list fatal unknown command quota but plugin quota exists try to set mail plugins quota usr local bin listmailuser line syntax error operand expected error token is usr local bin listmailuser line syntax error operand expected error token is new deol com what did you expect to happen markdown i believe it has something to with the ldap integration when i disable ldap in the compose file the email accounts can be created listed and deleted without any issues how do we replicate the issue markdown try to create a compose file with ldap enabled and integrated in it try creating an email account after that try to list or delete the email accounts too dms version what operating system is dms running on linux what instruction set architecture is dms running on what container orchestration tool are you using docker compose docker compose yml yaml version services mailserver image docker io mailserver docker mailserver latest container name mailserver hostname mail domainname deol com ports volumes docker data dms mail data var mail docker data dms mail state var mail state docker data dms mail logs var log mail docker data dms config tmp docker mailserver etc localtime etc localtime ro environment enable spamassassin spamassassin spam to inbox enable clamav enable enable postgrey enable saslauthd one dir dms debug enable ldap ldap server host ldapnew your ldap container ip servername ldap search base ou people dc ds dc domain dc com ldap bind dn cn admin dc ds dc domain dc com ldap bind pw enable saslauthd saslauthd mechanisms ldap saslauthd ldap server ldapnew saslauthd ldap bind dn cn admin dc ds dc domain dc com saslauthd ldap password saslauthd ldap search base ou people dc ds dc domain dc com saslauthd ldap filter objectclass postfixbookmailaccount uniqueidentifier u postmaster address postmaster deol com postfix message size limit cap add net admin sys ptrace relevant log output no response other relevant information no response what level of experience do you have with docker and mail servers i am inexperienced with docker i am inexperienced with mail servers i am uncomfortable with the cli code of conduct i have read this project s and i agree i have read the and the and i searched the but could not find a solution improvements to this form no response | 1 |
666,300 | 22,349,529,677 | IssuesEvent | 2022-06-15 10:45:34 | SAP/xsk | https://api.github.com/repos/SAP/xsk | closed | [IDE] CSV Editor - Total Records Count | enhancement priority-low effort-medium usability tooling shadow | It would be great if the CSV editor is displaying the total records count somewhere:
**Sample rows count:**

**CSV Editor:**

| 1.0 | [IDE] CSV Editor - Total Records Count - It would be great if the CSV editor is displaying the total records count somewhere:
**Sample rows count:**

**CSV Editor:**

| priority | csv editor total records count it would be great if the csv editor is displaying the total records count somewhere sample rows count csv editor | 1 |
92,675 | 3,872,900,201 | IssuesEvent | 2016-04-11 15:15:54 | jcgregorio/httplib2 | https://api.github.com/repos/jcgregorio/httplib2 | closed | bdist_rpm fails | bug imported Priority-Medium | _From [brian.la...@gmail.com](https://code.google.com/u/108532649133345963591/) on December 29, 2009 16:24:09_
What steps will reproduce the problem? 1. python setup.py bdist_rpm What is the expected output? What do you see instead? EXPECTED::
+ python setup.py build
running build
running build_py
creating build
creating build/lib
creating build/lib/httplib2
copying httplib2/__init__.py -> build/lib/httplib2
copying httplib2/iri2uri.py -> build/lib/httplib2
+ exit 0
ACTUAL RESULT::
+ python setup.py build
running build
running build_py
error: package directory 'python2/httplib2' does not exist
error: Bad exit status from /var/tmp/rpm-tmp.50855 (%build) What version of the product are you using? On what operating system? Latest check-in on any RPM based distro (Fedora/RHEL/Centos etc.) Please provide any additional information below. I believe that the RPM isn't being generated correctly with the new python2
and python3 subdirectories. I'm not a huge user of python distutils...but
I'll try to look and see if I can fix.
_Original issue: http://code.google.com/p/httplib2/issues/detail?id=85_ | 1.0 | bdist_rpm fails - _From [brian.la...@gmail.com](https://code.google.com/u/108532649133345963591/) on December 29, 2009 16:24:09_
What steps will reproduce the problem? 1. python setup.py bdist_rpm What is the expected output? What do you see instead? EXPECTED::
+ python setup.py build
running build
running build_py
creating build
creating build/lib
creating build/lib/httplib2
copying httplib2/__init__.py -> build/lib/httplib2
copying httplib2/iri2uri.py -> build/lib/httplib2
+ exit 0
ACTUAL RESULT::
+ python setup.py build
running build
running build_py
error: package directory 'python2/httplib2' does not exist
error: Bad exit status from /var/tmp/rpm-tmp.50855 (%build) What version of the product are you using? On what operating system? Latest check-in on any RPM based distro (Fedora/RHEL/Centos etc.) Please provide any additional information below. I believe that the RPM isn't being generated correctly with the new python2
and python3 subdirectories. I'm not a huge user of python distutils...but
I'll try to look and see if I can fix.
_Original issue: http://code.google.com/p/httplib2/issues/detail?id=85_ | priority | bdist rpm fails from on december what steps will reproduce the problem python setup py bdist rpm what is the expected output what do you see instead expected python setup py build running build running build py creating build creating build lib creating build lib copying init py build lib copying py build lib exit actual result python setup py build running build running build py error package directory does not exist error bad exit status from var tmp rpm tmp build what version of the product are you using on what operating system latest check in on any rpm based distro fedora rhel centos etc please provide any additional information below i believe that the rpm isn t being generated correctly with the new and subdirectories i m not a huge user of python distutils but i ll try to look and see if i can fix original issue | 1 |
171,596 | 6,491,685,465 | IssuesEvent | 2017-08-21 10:41:07 | softdevteam/krun | https://api.github.com/repos/softdevteam/krun | opened | Move masking of the core cycle counter outside the time section. | enhancement medium priority (a clear improvement but not a blocker for publication) | This loop:
https://github.com/softdevteam/krun/blob/9c38f01a7e6292669d7c9f83780d44e2abae8279/libkrun/libkruntime.c#L401
Can be moved into the getter functions, thus moving the masking operations out of the timed section. | 1.0 | Move masking of the core cycle counter outside the time section. - This loop:
https://github.com/softdevteam/krun/blob/9c38f01a7e6292669d7c9f83780d44e2abae8279/libkrun/libkruntime.c#L401
Can be moved into the getter functions, thus moving the masking operations out of the timed section. | priority | move masking of the core cycle counter outside the time section this loop can be moved into the getter functions thus moving the masking operations out of the timed section | 1 |
11,563 | 2,610,142,072 | IssuesEvent | 2015-02-26 18:44:46 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | seed line should NOT be visible right away | auto-migrated Priority-Medium Type-Enhancement | ```
What steps will reproduce the problem?
1. latest revision offers a neat functionality, the possibility of editing the
seed line
2. this functionality is useful for testing and has been requested by users
3. however despite being in the game configuration page, it should not be
visible right away for not always the user needs to change that line and has
the high likelihood of scaring the user with too much technicality
What is the expected output? What do you see instead?
in my view there should be a nice button that when pushed either shows a popup
or simply makes visible that line
```
-----
Original issue reported on code.google.com by `vittorio...@gmail.com` on 20 Dec 2010 at 12:02
* Blocking: #115 | 1.0 | seed line should NOT be visible right away - ```
What steps will reproduce the problem?
1. latest revision offers a neat functionality, the possibility of editing the
seed line
2. this functionality is useful for testing and has been requested by users
3. however despite being in the game configuration page, it should not be
visible right away for not always the user needs to change that line and has
the high likelihood of scaring the user with too much technicality
What is the expected output? What do you see instead?
in my view there should be a nice button that when pushed either shows a popup
or simply makes visible that line
```
-----
Original issue reported on code.google.com by `vittorio...@gmail.com` on 20 Dec 2010 at 12:02
* Blocking: #115 | priority | seed line should not be visible right away what steps will reproduce the problem latest revision offers a neat functionality the possibility of editing the seed line this functionality is useful for testing and has been requested by users however despite being in the game configuration page it should not be visible right away for not always the user needs to change that line and has the high likelihood of scaring the user with too much technicality what is the expected output what do you see instead in my view there should be a nice button that when pushed either shows a popup or simply makes visible that line original issue reported on code google com by vittorio gmail com on dec at blocking | 1 |
70,631 | 3,332,949,470 | IssuesEvent | 2015-11-11 22:28:04 | angular/material | https://api.github.com/repos/angular/material | closed | menuContents has zero items | priority: medium | In menuDirective.js, the link function call mdMenuCtrl.init.
if the md-menu-item is hard coded, like in the case of demoBasicUsage, then in menuController.js this will return items : 'menuContainer[0].querySelectorAll('md-menu-item')'.
if the md-menu-item is not hard coded, like in the case of demoMenuPositionModes and demoMenuWidth, then in menuController.js this will return zero items: 'menuContainer[0].querySelectorAll('md-menu-item')'.
```javascript
function link(scope, element, attrs, ctrls) {
var mdMenuCtrl = ctrls[0];
var isInMenuBar = ctrls[1] != undefined;
// Move everything into a md-menu-container and pass it to the controller
var menuContainer = angular.element(
'<div class="md-open-menu-container md-whiteframe-z2"></div>'
);
var menuContents = element.children()[1];
menuContainer.append(menuContents);
if (isInMenuBar) {
element.append(menuContainer);
menuContainer[0].style.display = 'none';
}
mdMenuCtrl.init(menuContainer, { isInMenuBar: isInMenuBar });
scope.$on('$destroy', function() {
mdMenuCtrl
.destroy()
.finally(function(){
menuContainer.remove();
});
});
}
``` | 1.0 | menuContents has zero items - In menuDirective.js, the link function call mdMenuCtrl.init.
if the md-menu-item is hard coded, like in the case of demoBasicUsage, then in menuController.js this will return items : 'menuContainer[0].querySelectorAll('md-menu-item')'.
if the md-menu-item is not hard coded, like in the case of demoMenuPositionModes and demoMenuWidth, then in menuController.js this will return zero items: 'menuContainer[0].querySelectorAll('md-menu-item')'.
```javascript
function link(scope, element, attrs, ctrls) {
var mdMenuCtrl = ctrls[0];
var isInMenuBar = ctrls[1] != undefined;
// Move everything into a md-menu-container and pass it to the controller
var menuContainer = angular.element(
'<div class="md-open-menu-container md-whiteframe-z2"></div>'
);
var menuContents = element.children()[1];
menuContainer.append(menuContents);
if (isInMenuBar) {
element.append(menuContainer);
menuContainer[0].style.display = 'none';
}
mdMenuCtrl.init(menuContainer, { isInMenuBar: isInMenuBar });
scope.$on('$destroy', function() {
mdMenuCtrl
.destroy()
.finally(function(){
menuContainer.remove();
});
});
}
``` | priority | menucontents has zero items in menudirective js the link function call mdmenuctrl init if the md menu item is hard coded like in the case of demobasicusage then in menucontroller js this will return items menucontainer queryselectorall md menu item if the md menu item is not hard coded like in the case of demomenupositionmodes and demomenuwidth then in menucontroller js this will return zero items menucontainer queryselectorall md menu item javascript function link scope element attrs ctrls var mdmenuctrl ctrls var isinmenubar ctrls undefined move everything into a md menu container and pass it to the controller var menucontainer angular element var menucontents element children menucontainer append menucontents if isinmenubar element append menucontainer menucontainer style display none mdmenuctrl init menucontainer isinmenubar isinmenubar scope on destroy function mdmenuctrl destroy finally function menucontainer remove | 1 |
462,654 | 13,250,971,352 | IssuesEvent | 2020-08-20 00:41:55 | TB-Modeling/modeltb.org | https://api.github.com/repos/TB-Modeling/modeltb.org | opened | Professional downloadable briefs | Priority 2/3: Medium new feature new idea | ## Description
We can host professional-looking briefs on TB and other topics. | 1.0 | Professional downloadable briefs - ## Description
We can host professional-looking briefs on TB and other topics. | priority | professional downloadable briefs description we can host professional looking briefs on tb and other topics | 1 |
273,401 | 8,530,257,133 | IssuesEvent | 2018-11-03 20:34:54 | minio/minio | https://api.github.com/repos/minio/minio | closed | Crashing with Out of Memory Errors (refs 6164) | community priority: medium | I hate to be the bearer of bad news, @harshavardhana , but the fix has not been working for me. I've upgraded to all the most recent Minio.exe builds since the last post in this issue, but with no success. I've delayed responding here in case the fix was in a build later than I thought, and because I found an issue in my environment and wondered if it was related. I've determined it is not but will outline what I've found in the past month.
I noticed as time went on that Minio started reporting read errors on my V: drive, but not any of the others. Separately, I discovered while trying to copy a set of large files from my V: drive (completely unrelated to Minio) that there was at least one corrupted file in the mix. This led me to believe my V drive was experiencing data corruption, and I wondered if it was the culprit of problems all along. I replaced both my V and X drives. Because they're identical and were purchased at the same time, I replaced both just in case the X drive was soon to follow. Two brand-new 6 TB drives are in their place.
This allowed me a chance to test Minio's healing capabilities, which so far are working OK. However, I've not been able to fully test it (I've only healed about half my data set so far) because Minio continues to crash with out of memory errors after 6-7 hours of run time. I brainstormed that if this out of memory error cannot be fixed I could heal in two stages to work-around the bug. First the V and X drives, which should heal their data in this "new" format. Then delete the Minio data on W and Y and heal back to those drives. Thus getting all my data into the new format in a very round-about (and dangerous) fashion.
But like I said, because Minio continues to crash with out of memory exceptions during the heal, I've yet to accomplish this in full. My previous screenshot shows my Minio data at 294 GB, and currently I've only healed about 154 GB. The healed amount tends to change unpredictably. Sometimes it will run for hours and appear to make 0 progress (the V and X data directories are not larger), and other times they will grow by several GB. It's almost as if the process is starting over every single time and it only makes progress if it happens to run longer than last time. I'm not sure, though.
Further, even without the heal running, the W and Y drives continue to be hammered while Minio consumes more and more memory for several hours before crashing. This is puzzling to me, because with the number of times I've run Minio it has to have a total run time of at least 50 hours, maybe much more. Assuming a read/write speed of just 10 MB a second, which is very slow for a SATA drive, even platter-based, it would have been able to read/write 1.8 TB of data from each drive. Even if speed were half that at 5 MB/s, that's still 900 GB. Half that, 2.5 MB/s would be 450 GB in that time period. Given my total data set is ~300 GB, I am extremely puzzled as to why it's not done yet unless it is restarting progress every single time. Which isn't logical given that no software would try to complete 300 GB worth of work before writing out progress. That wouldn't make sense.
Can you shed any light into what's going on here? My last resort is to simply delete everything and restart with no data. That's not the end of the world because my most important backup copy is off-site. Having stuff locally is for a more convenient full-copy.
I've attached CPU and MEM logs that I gathered yesterday, letting it run for a few hours and _not_ healing, just doing it's behind-the-scenes work. Again, it slammed usage of my W and Y drives, as well as the CPU. I stopped it before a crash to get the log output. The CPU log is shorter than the mem one due to my schedule, but it still has a couple hours.
Thank you.
[mem_20180918.txt](https://github.com/minio/minio/files/2397728/mem_20180918.txt)
[cpu_20180918.txt](https://github.com/minio/minio/files/2397729/cpu_20180918.txt)
_Originally posted by @tylerforsythe in https://github.com/minio/minio/issue_comments#issuecomment-422851849_ | 1.0 | Crashing with Out of Memory Errors (refs 6164) - I hate to be the bearer of bad news, @harshavardhana , but the fix has not been working for me. I've upgraded to all the most recent Minio.exe builds since the last post in this issue, but with no success. I've delayed responding here in case the fix was in a build later than I thought, and because I found an issue in my environment and wondered if it was related. I've determined it is not but will outline what I've found in the past month.
I noticed as time went on that Minio started reporting read errors on my V: drive, but not any of the others. Separately, I discovered while trying to copy a set of large files from my V: drive (completely unrelated to Minio) that there was at least one corrupted file in the mix. This led me to believe my V drive was experiencing data corruption, and I wondered if it was the culprit of problems all along. I replaced both my V and X drives. Because they're identical and were purchased at the same time, I replaced both just in case the X drive was soon to follow. Two brand-new 6 TB drives are in their place.
This allowed me a chance to test Minio's healing capabilities, which so far are working OK. However, I've not been able to fully test it (I've only healed about half my data set so far) because Minio continues to crash with out of memory errors after 6-7 hours of run time. I brainstormed that if this out of memory error cannot be fixed I could heal in two stages to work-around the bug. First the V and X drives, which should heal their data in this "new" format. Then delete the Minio data on W and Y and heal back to those drives. Thus getting all my data into the new format in a very round-about (and dangerous) fashion.
But like I said, because Minio continues to crash with out of memory exceptions during the heal, I've yet to accomplish this in full. My previous screenshot shows my Minio data at 294 GB, and currently I've only healed about 154 GB. The healed amount tends to change unpredictably. Sometimes it will run for hours and appear to make 0 progress (the V and X data directories are not larger), and other times they will grow by several GB. It's almost as if the process is starting over every single time and it only makes progress if it happens to run longer than last time. I'm not sure, though.
Further, even without the heal running, the W and Y drives continue to be hammered while Minio consumes more and more memory for several hours before crashing. This is puzzling to me, because with the number of times I've run Minio it has to have a total run time of at least 50 hours, maybe much more. Assuming a read/write speed of just 10 MB a second, which is very slow for a SATA drive, even platter-based, it would have been able to read/write 1.8 TB of data from each drive. Even if speed were half that at 5 MB/s, that's still 900 GB. Half that, 2.5 MB/s would be 450 GB in that time period. Given my total data set is ~300 GB, I am extremely puzzled as to why it's not done yet unless it is restarting progress every single time. Which isn't logical given that no software would try to complete 300 GB worth of work before writing out progress. That wouldn't make sense.
Can you shed any light into what's going on here? My last resort is to simply delete everything and restart with no data. That's not the end of the world because my most important backup copy is off-site. Having stuff locally is for a more convenient full-copy.
I've attached CPU and MEM logs that I gathered yesterday, letting it run for a few hours and _not_ healing, just doing it's behind-the-scenes work. Again, it slammed usage of my W and Y drives, as well as the CPU. I stopped it before a crash to get the log output. The CPU log is shorter than the mem one due to my schedule, but it still has a couple hours.
Thank you.
[mem_20180918.txt](https://github.com/minio/minio/files/2397728/mem_20180918.txt)
[cpu_20180918.txt](https://github.com/minio/minio/files/2397729/cpu_20180918.txt)
_Originally posted by @tylerforsythe in https://github.com/minio/minio/issue_comments#issuecomment-422851849_ | priority | crashing with out of memory errors refs i hate to be the bearer of bad news harshavardhana but the fix has not been working for me i ve upgraded to all the most recent minio exe builds since the last post in this issue but with no success i ve delayed responding here in case the fix was in a build later than i thought and because i found an issue in my environment and wondered if it was related i ve determined it is not but will outline what i ve found in the past month i noticed as time went on that minio started reporting read errors on my v drive but not any of the others separately i discovered while trying to copy a set of large files from my v drive completely unrelated to minio that there was at least one corrupted file in the mix this led me to believe my v drive was experiencing data corruption and i wondered if it was the culprit of problems all along i replaced both my v and x drives because they re identical and were purchased at the same time i replaced both just in case the x drive was soon to follow two brand new tb drives are in their place this allowed me a chance to test minio s healing capabilities which so far are working ok however i ve not been able to fully test it i ve only healed about half my data set so far because minio continues to crash with out of memory errors after hours of run time i brainstormed that if this out of memory error cannot be fixed i could heal in two stages to work around the bug first the v and x drives which should heal their data in this new format then delete the minio data on w and y and heal back to those drives thus getting all my data into the new format in a very round about and dangerous fashion but like i said because minio continues to crash with out of memory exceptions during the heal i ve yet to accomplish this in full my previous screenshot shows my minio data at gb and currently i ve only healed about gb the healed amount tends to change unpredictably sometimes it will run for hours and appear to make progress the v and x data directories are not larger and other times they will grow by several gb it s almost as if the process is starting over every single time and it only makes progress if it happens to run longer than last time i m not sure though further even without the heal running the w and y drives continue to be hammered while minio consumes more and more memory for several hours before crashing this is puzzling to me because with the number of times i ve run minio it has to have a total run time of at least hours maybe much more assuming a read write speed of just mb a second which is very slow for a sata drive even platter based it would have been able to read write tb of data from each drive even if speed were half that at mb s that s still gb half that mb s would be gb in that time period given my total data set is gb i am extremely puzzled as to why it s not done yet unless it is restarting progress every single time which isn t logical given that no software would try to complete gb worth of work before writing out progress that wouldn t make sense can you shed any light into what s going on here my last resort is to simply delete everything and restart with no data that s not the end of the world because my most important backup copy is off site having stuff locally is for a more convenient full copy i ve attached cpu and mem logs that i gathered yesterday letting it run for a few hours and not healing just doing it s behind the scenes work again it slammed usage of my w and y drives as well as the cpu i stopped it before a crash to get the log output the cpu log is shorter than the mem one due to my schedule but it still has a couple hours thank you originally posted by tylerforsythe in | 1 |
604,574 | 18,714,856,193 | IssuesEvent | 2021-11-03 02:12:14 | AY2122S1-CS2103T-T12-3/tp | https://api.github.com/repos/AY2122S1-CS2103T-T12-3/tp | closed | [PE-D] UG TOC | priority.Medium | 
I think for your sectioning of TOC,
The `feature` should be an overall header for all the subsections.
Something like below
- Features
- Contact Features
- Appointment Features
- Other Features
<!--session: 1635494624855-6c4169c8-8f98-434c-bb03-2bfcaad4f18c-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: Timothyoung97/ped#2 | 1.0 | [PE-D] UG TOC - 
I think for your sectioning of TOC,
The `feature` should be an overall header for all the subsections.
Something like below
- Features
- Contact Features
- Appointment Features
- Other Features
<!--session: 1635494624855-6c4169c8-8f98-434c-bb03-2bfcaad4f18c-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: Timothyoung97/ped#2 | priority | ug toc i think for your sectioning of toc the feature should be an overall header for all the subsections something like below features contact features appointment features other features labels severity low type documentationbug original ped | 1 |
727,944 | 25,060,654,219 | IssuesEvent | 2022-11-07 00:59:44 | AY2223S1-CS2103T-W08-2/tp | https://api.github.com/repos/AY2223S1-CS2103T-W08-2/tp | closed | [PE-D][Tester B] find command duplicates person | priority.High priority.Medium | Using the default data provided, running 'find Irfan` displays Irfan's details correclty.
But then running `find Irfan Ibrahim` causes does nothing.
Then run `find Irfan`, Irfan is duplicated in the list.
To recover, the app must be reopened.
There might be a deeper issue here where clients are randomly getting duplicated in the model.
<!--session: 1666944033620-ae8585f7-7738-461c-9ea8-0b2de73b6950--><!--Version: Web v3.4.4-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: Thing1Thing2/ped#10 | 2.0 | [PE-D][Tester B] find command duplicates person - Using the default data provided, running 'find Irfan` displays Irfan's details correclty.
But then running `find Irfan Ibrahim` causes does nothing.
Then run `find Irfan`, Irfan is duplicated in the list.
To recover, the app must be reopened.
There might be a deeper issue here where clients are randomly getting duplicated in the model.
<!--session: 1666944033620-ae8585f7-7738-461c-9ea8-0b2de73b6950--><!--Version: Web v3.4.4-->
-------------
Labels: `severity.Medium` `type.FunctionalityBug`
original: Thing1Thing2/ped#10 | priority | find command duplicates person using the default data provided running find irfan displays irfan s details correclty but then running find irfan ibrahim causes does nothing then run find irfan irfan is duplicated in the list to recover the app must be reopened there might be a deeper issue here where clients are randomly getting duplicated in the model labels severity medium type functionalitybug original ped | 1 |
738,045 | 25,542,863,056 | IssuesEvent | 2022-11-29 16:28:50 | envoyproxy/gateway | https://api.github.com/repos/envoyproxy/gateway | closed | Wildcard in HttpRoute hostnames generates exact match | bug area/translator priority/medium | *Description*:
When `listeners` spec in the Gateway does not specify hosts and HttpRoute spec specifies _wildcard_ host envoy-gateway generates Exact match for the `:authority` that results in matching failures and hence 404.
E.g. for HttpRoute host name `*.example.com` it generates
```
HeaderMatches:
- Exact: '*.example.com'
Name: :authority
Prefix: null
SafeRegex: null
```
The problem seems to be in [this code](https://github.com/envoyproxy/gateway/blob/cede77ebbecf25fbe8994cfd7643e8295bf75712/internal/gatewayapi/translator.go#L1166) that compares host to `*` while it probably needs to consider `*.example.com` cases
**Expected:**
Regex match is generated instead of Exact match and request passes the filter
*Repro steps*:
Apply this Gateway manifest
```
apiVersion: gateway.networking.k8s.io/v1beta1
kind: Gateway
metadata:
name: eg
spec:
gatewayClassName: eg
listeners:
- name: http
protocol: HTTP
port: 80
```
Apply this HttpRoute manifest:
```
apiVersion: gateway.networking.k8s.io/v1beta1
kind: HTTPRoute
metadata:
name: backend
spec:
parentRefs:
- name: eg
hostnames:
- "*.example.com"
rules:
- backendRefs:
- name: paveliak-playground-1
port: 8080
matches:
- path:
type: PathPrefix
value: /
```
*Environment*:
- Minikube 1.27.1
- K8s 1.25.2
- envoy-gateway (tried both `v0.2.0` and `main`)
*Logs*:
```
HTTP:
- Address: 0.0.0.0
Hostnames:
- '*'
Name: default-eg-http
Port: 10080
Routes:
- AddRequestHeaders: null
BackendWeights:
Invalid: 0
Valid: 0
Destinations:
- Host: 10.100.121.40
Port: 8080
Weight: 1
DirectResponse: null
HeaderMatches:
- Exact: '*.example.com'
Name: :authority
Prefix: null
SafeRegex: null
Name: default-backend-rule-0-match-0-*.example.com
PathMatch:
Exact: null
Name: ""
Prefix: /
SafeRegex: null
QueryParamMatches: null
Redirect: null
RemoveRequestHeaders: null
TLS: null
TCP: null
{"runner": "gateway-api", "output": "xds-ir"}
```
As a result this works `curl http://10.110.191.160 -H "Host: *.example.com"`
But this request returns 404 `curl http://10.110.191.160 -H "Host: www.example.com"`
| 1.0 | Wildcard in HttpRoute hostnames generates exact match - *Description*:
When `listeners` spec in the Gateway does not specify hosts and HttpRoute spec specifies _wildcard_ host envoy-gateway generates Exact match for the `:authority` that results in matching failures and hence 404.
E.g. for HttpRoute host name `*.example.com` it generates
```
HeaderMatches:
- Exact: '*.example.com'
Name: :authority
Prefix: null
SafeRegex: null
```
The problem seems to be in [this code](https://github.com/envoyproxy/gateway/blob/cede77ebbecf25fbe8994cfd7643e8295bf75712/internal/gatewayapi/translator.go#L1166) that compares host to `*` while it probably needs to consider `*.example.com` cases
**Expected:**
Regex match is generated instead of Exact match and request passes the filter
*Repro steps*:
Apply this Gateway manifest
```
apiVersion: gateway.networking.k8s.io/v1beta1
kind: Gateway
metadata:
name: eg
spec:
gatewayClassName: eg
listeners:
- name: http
protocol: HTTP
port: 80
```
Apply this HttpRoute manifest:
```
apiVersion: gateway.networking.k8s.io/v1beta1
kind: HTTPRoute
metadata:
name: backend
spec:
parentRefs:
- name: eg
hostnames:
- "*.example.com"
rules:
- backendRefs:
- name: paveliak-playground-1
port: 8080
matches:
- path:
type: PathPrefix
value: /
```
*Environment*:
- Minikube 1.27.1
- K8s 1.25.2
- envoy-gateway (tried both `v0.2.0` and `main`)
*Logs*:
```
HTTP:
- Address: 0.0.0.0
Hostnames:
- '*'
Name: default-eg-http
Port: 10080
Routes:
- AddRequestHeaders: null
BackendWeights:
Invalid: 0
Valid: 0
Destinations:
- Host: 10.100.121.40
Port: 8080
Weight: 1
DirectResponse: null
HeaderMatches:
- Exact: '*.example.com'
Name: :authority
Prefix: null
SafeRegex: null
Name: default-backend-rule-0-match-0-*.example.com
PathMatch:
Exact: null
Name: ""
Prefix: /
SafeRegex: null
QueryParamMatches: null
Redirect: null
RemoveRequestHeaders: null
TLS: null
TCP: null
{"runner": "gateway-api", "output": "xds-ir"}
```
As a result this works `curl http://10.110.191.160 -H "Host: *.example.com"`
But this request returns 404 `curl http://10.110.191.160 -H "Host: www.example.com"`
| priority | wildcard in httproute hostnames generates exact match description when listeners spec in the gateway does not specify hosts and httproute spec specifies wildcard host envoy gateway generates exact match for the authority that results in matching failures and hence e g for httproute host name example com it generates headermatches exact example com name authority prefix null saferegex null the problem seems to be in that compares host to while it probably needs to consider example com cases expected regex match is generated instead of exact match and request passes the filter repro steps apply this gateway manifest apiversion gateway networking io kind gateway metadata name eg spec gatewayclassname eg listeners name http protocol http port apply this httproute manifest apiversion gateway networking io kind httproute metadata name backend spec parentrefs name eg hostnames example com rules backendrefs name paveliak playground port matches path type pathprefix value environment minikube envoy gateway tried both and main logs http address hostnames name default eg http port routes addrequestheaders null backendweights invalid valid destinations host port weight directresponse null headermatches exact example com name authority prefix null saferegex null name default backend rule match example com pathmatch exact null name prefix saferegex null queryparammatches null redirect null removerequestheaders null tls null tcp null runner gateway api output xds ir as a result this works curl h host example com but this request returns curl h host | 1 |
302,963 | 9,300,893,328 | IssuesEvent | 2019-03-23 17:22:54 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Username searches in the Hall of Heroes should be case-insensitive | good first issue priority: medium section: Achievements/Popups/Notifications section: other status: issue: in progress | As described in issue https://github.com/HabitRPG/habitica/issues/10972 and its pull request https://github.com/HabitRPG/habitica/pull/10980 , the Hall of Heroes allows moderators to search for users by the Username. Currently that search is case-sensitive. It should be case-insensitive. I.e., searching for "examplename" would let you find a user called "ExampleName".
Any contributor who wants to work on this should read the top post at https://github.com/HabitRPG/habitica/issues/10972 to learn about how the moderator search feature works and how you can access it on your local install. https://github.com/HabitRPG/habitica/pull/10980/files will show you how the current case-sensitive search is implemented.
_(NB We're not considering this to be a bug in https://github.com/HabitRPG/habitica/pull/10980. We neglected to say that the search should be case-insensitive and so that PR did implement the search feature as intended at the time!)_ | 1.0 | Username searches in the Hall of Heroes should be case-insensitive - As described in issue https://github.com/HabitRPG/habitica/issues/10972 and its pull request https://github.com/HabitRPG/habitica/pull/10980 , the Hall of Heroes allows moderators to search for users by the Username. Currently that search is case-sensitive. It should be case-insensitive. I.e., searching for "examplename" would let you find a user called "ExampleName".
Any contributor who wants to work on this should read the top post at https://github.com/HabitRPG/habitica/issues/10972 to learn about how the moderator search feature works and how you can access it on your local install. https://github.com/HabitRPG/habitica/pull/10980/files will show you how the current case-sensitive search is implemented.
_(NB We're not considering this to be a bug in https://github.com/HabitRPG/habitica/pull/10980. We neglected to say that the search should be case-insensitive and so that PR did implement the search feature as intended at the time!)_ | priority | username searches in the hall of heroes should be case insensitive as described in issue and its pull request the hall of heroes allows moderators to search for users by the username currently that search is case sensitive it should be case insensitive i e searching for examplename would let you find a user called examplename any contributor who wants to work on this should read the top post at to learn about how the moderator search feature works and how you can access it on your local install will show you how the current case sensitive search is implemented nb we re not considering this to be a bug in we neglected to say that the search should be case insensitive and so that pr did implement the search feature as intended at the time | 1 |
596,841 | 18,145,008,004 | IssuesEvent | 2021-09-25 09:06:36 | google/mozc | https://api.github.com/repos/google/mozc | closed | Missing "こと***" entries | Priority-Medium auto-migrated OpSys-All Type-Conversion | ```
I've got them automatically and fixed them manually.
Maybe there are some mistakes and you need to calculate the scores.
Check "tmp.koto" please.
e.g.
======================================================
すごいこと 2665 2235 5782 凄いこと
そういうこと 3006 2235 4853 そういうこと
======================================================
```
Original issue reported on code.google.com by `heathros...@gmail.com` on 9 Nov 2010 at 7:26
Attachments:
- [tmp.koto](https://storage.googleapis.com/google-code-attachments/mozc/issue-67/comment-0/tmp.koto)
| 1.0 | Missing "こと***" entries - ```
I've got them automatically and fixed them manually.
Maybe there are some mistakes and you need to calculate the scores.
Check "tmp.koto" please.
e.g.
======================================================
すごいこと 2665 2235 5782 凄いこと
そういうこと 3006 2235 4853 そういうこと
======================================================
```
Original issue reported on code.google.com by `heathros...@gmail.com` on 9 Nov 2010 at 7:26
Attachments:
- [tmp.koto](https://storage.googleapis.com/google-code-attachments/mozc/issue-67/comment-0/tmp.koto)
| priority | missing こと entries i ve got them automatically and fixed them manually maybe there are some mistakes and you need to calculate the scores check tmp koto please e g すごいこと 凄いこと そういうこと そういうこと original issue reported on code google com by heathros gmail com on nov at attachments | 1 |
442,168 | 12,740,990,356 | IssuesEvent | 2020-06-26 04:36:10 | openmsupply/mobile | https://api.github.com/repos/openmsupply/mobile | opened | Merged Prescriber from Desktop store doesn't sync to Mobile Store | 5.0.4 Docs: not needed Effort: medium Priority: immediate | ## Describe the bug
Merged Prescriber from Desktop store doesn't sync to Mobile Store
### To reproduce
Steps to reproduce the behavior:
1. In mobile store, create atleast two prescriber .
2. Create Prescription from one of the created Prescriber and sync
3. Merge these prescriber in Desktop Store such that Keep Prescriber with no Prescription and Merge Prescriber with Prescription.
4. Check Mobile Store
5. See Error. Prescribers is not sync



### Expected behaviour
Sync should work as expected and Merged prescriber should be shown in Mobile as well
### Proposed Solution
Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase.
### Version and device info
- App version: v5.0.4
- Tablet model: API 21
- Desktop version: v412RC04
### Additional context
Add any other context about the problem here.
| 1.0 | Merged Prescriber from Desktop store doesn't sync to Mobile Store - ## Describe the bug
Merged Prescriber from Desktop store doesn't sync to Mobile Store
### To reproduce
Steps to reproduce the behavior:
1. In mobile store, create atleast two prescriber .
2. Create Prescription from one of the created Prescriber and sync
3. Merge these prescriber in Desktop Store such that Keep Prescriber with no Prescription and Merge Prescriber with Prescription.
4. Check Mobile Store
5. See Error. Prescribers is not sync



### Expected behaviour
Sync should work as expected and Merged prescriber should be shown in Mobile as well
### Proposed Solution
Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase.
### Version and device info
- App version: v5.0.4
- Tablet model: API 21
- Desktop version: v412RC04
### Additional context
Add any other context about the problem here.
| priority | merged prescriber from desktop store doesn t sync to mobile store describe the bug merged prescriber from desktop store doesn t sync to mobile store to reproduce steps to reproduce the behavior in mobile store create atleast two prescriber create prescription from one of the created prescriber and sync merge these prescriber in desktop store such that keep prescriber with no prescription and merge prescriber with prescription check mobile store see error prescribers is not sync expected behaviour sync should work as expected and merged prescriber should be shown in mobile as well proposed solution leave if you don t know how to fix implement edit this issue description and explain here if you know the best path of implementing the fix within the codebase version and device info app version tablet model api desktop version additional context add any other context about the problem here | 1 |
118,917 | 4,757,600,475 | IssuesEvent | 2016-10-24 17:02:50 | geosolutions-it/geotools | https://api.github.com/repos/geosolutions-it/geotools | opened | Communitty interaction | C009-2016-MONGODB Priority: Medium Task | Since we are preserving the current MongoDB behavior and only users that want to use complex feature will need to do something, we only need to convince the community about how we intend to support the complex features. Another point that needs to be discussed is the dependency on app-schema (if we don't manage to avoid it). | 1.0 | Communitty interaction - Since we are preserving the current MongoDB behavior and only users that want to use complex feature will need to do something, we only need to convince the community about how we intend to support the complex features. Another point that needs to be discussed is the dependency on app-schema (if we don't manage to avoid it). | priority | communitty interaction since we are preserving the current mongodb behavior and only users that want to use complex feature will need to do something we only need to convince the community about how we intend to support the complex features another point that needs to be discussed is the dependency on app schema if we don t manage to avoid it | 1 |
22,105 | 2,645,590,069 | IssuesEvent | 2015-03-13 00:06:08 | pedromorgan/flightgear-issues-test | https://api.github.com/repos/pedromorgan/flightgear-issues-test | closed | MPmap is out off sync | apt.dat bug imported mpmap pigeon Priority-Medium | _From [pedromor...@gmail.com](https://code.google.com/u/112081649209717344658/) on February 03, 2010 07:45:16_
What steps will reproduce the problem? * the nav.dat is updated in cvs, but changes are not reflected on
mmapservers
* mpmap needs to be in sync with cvs
_Original issue: http://code.google.com/p/flightgear-bugs/issues/detail?id=28_ | 1.0 | MPmap is out off sync - _From [pedromor...@gmail.com](https://code.google.com/u/112081649209717344658/) on February 03, 2010 07:45:16_
What steps will reproduce the problem? * the nav.dat is updated in cvs, but changes are not reflected on
mmapservers
* mpmap needs to be in sync with cvs
_Original issue: http://code.google.com/p/flightgear-bugs/issues/detail?id=28_ | priority | mpmap is out off sync from on february what steps will reproduce the problem the nav dat is updated in cvs but changes are not reflected on mmapservers mpmap needs to be in sync with cvs original issue | 1 |
271,612 | 8,485,749,511 | IssuesEvent | 2018-10-26 08:48:57 | cms-gem-daq-project/cmsgemos | https://api.github.com/repos/cms-gem-daq-project/cmsgemos | closed | Exception in amc_info_uhal.py if -r option used at cmd line | Priority: Medium Status: Help Wanted Type: Bug | Based on feedback from DAQ expert tried to issue a reset command to the CTP7 with command line option -r.
Generated the exception shown: [amc_info_resetExcept.txt](https://github.com/cms-gem-daq-project/cmsgemos/files/1089047/amc_info_resetExcept.txt)
| 1.0 | Exception in amc_info_uhal.py if -r option used at cmd line - Based on feedback from DAQ expert tried to issue a reset command to the CTP7 with command line option -r.
Generated the exception shown: [amc_info_resetExcept.txt](https://github.com/cms-gem-daq-project/cmsgemos/files/1089047/amc_info_resetExcept.txt)
| priority | exception in amc info uhal py if r option used at cmd line based on feedback from daq expert tried to issue a reset command to the with command line option r generated the exception shown | 1 |
465,173 | 13,358,091,790 | IssuesEvent | 2020-08-31 11:02:33 | ooni/probe | https://api.github.com/repos/ooni/probe | closed | Make the OONI Probe desktop app rely entirely on the golang engine | enhancement epic funder/otf20 ooni/probe-desktop priority/medium | We are in the process of consolidating all the code related to running network experiments inside of a new golang based engine called probe-engine, which will replace measurement-kit.
As part of this activity, we will make the OONI Probe desktop (Windows and macOS) app rely entirely on the golang engine.
We are not aiming to port the OONI Probe desktop apps to a new codebase. We are also not aiming to port the OONI CLI to a new codebase. The work we are proposing is rather to reduce the amount of C++ dependencies used by the OONI Probe CLI and the desktop apps. Below we explain in detail which components we are planning to modify, why we are doing that, and why we have chosen to start integrating these changes in the desktop apps, rather than in the mobile apps.
The current OONI Probe desktop apps for Windows and macOS are Electron-based apps implemented at github.com/ooni/probe-desktop. These apps use the CLI interface written in golang to run OONI experiments and perform other functions.
The codebase of the OONI Probe CLI interface is available at github.com/ooni/probe-cli. More in detail, we have specified how the Desktop app will exec out the CLI to perform specific tasks and we have also specified the data format emitted by the CLI when running tests.
The CLI, in turn, is based on a lower-level “engine” library written in golang and available at github.com/ooni/probe-engine. This library aims to include a unified desktop and mobile implementation of all experiments. However, inside this library, we are still linking to our old C++ engine, Measurement Kit, to implement most OONI network experiments. (Measurement Kit is composed of several GitHub repositories available at github.com/measurement-kit.)
The work described in this activity is about gradually phasing out Measurement Kit, so that in the end all experiments are written in Go rather than in C++. This will mostly happen inside the probe-engine and will mostly concern changes in the implementation of tests (i.e. rewriting from C++ to Go). We believe that the API exposed by probe-engine will not change, except for minor changes in the options supported by the new experiments (for example, we may drop an option that does not make sense in the Go context or we may add an option that did not make sense for a C++ implementation).
This should explain why we believe that writing more OONI experiments in Go as part of probe-engine does not overlap with our previous work on creating the OONI Probe desktop apps (supported by our last OTF contract). In our view, this proposed work is simply the natural continuation of what we have been working on over the past year.
The advantages of rewriting in Go are the following:
* It will reduce the code size, because we can drop the C dependencies (whereas we cannot drop the dependency from Go since we want to ship a Psiphon experiment).
* It allows us to write experiments that can easily and safely be interrupted, because Go natively support this functionality, which is instead rather tricky to get right in C++.
* It allows us to easily wrap standard library functionality, hence allowing for easier low level measurements.
* It allows us to quickly cross compile for the operating systems we care about (Windows, macOS, Linux, Android, iOS); full builds will complete in minutes rather than hours, thus enabling faster development cycles.
* It will make OONI Probe safer, because Go is memory safe and has built-in support for concurrency.
Regarding the choice of focusing on the OONI Probe desktop app first (and on the mobile apps next) the reason is simple. Since probe-engine is already being used by the OONI Probe desktop apps via probe-cli, we have estimated that the effort (and cost) for doing this for the desktop apps is less than doing it on mobile, where we haven’t shipped golang code in production yet. In fact, we can update the desktop apps by recompiling probe-cli to use a new version of probe-engine that reimplements all tests in Go, with none or minimal API changes.
| 1.0 | Make the OONI Probe desktop app rely entirely on the golang engine - We are in the process of consolidating all the code related to running network experiments inside of a new golang based engine called probe-engine, which will replace measurement-kit.
As part of this activity, we will make the OONI Probe desktop (Windows and macOS) app rely entirely on the golang engine.
We are not aiming to port the OONI Probe desktop apps to a new codebase. We are also not aiming to port the OONI CLI to a new codebase. The work we are proposing is rather to reduce the amount of C++ dependencies used by the OONI Probe CLI and the desktop apps. Below we explain in detail which components we are planning to modify, why we are doing that, and why we have chosen to start integrating these changes in the desktop apps, rather than in the mobile apps.
The current OONI Probe desktop apps for Windows and macOS are Electron-based apps implemented at github.com/ooni/probe-desktop. These apps use the CLI interface written in golang to run OONI experiments and perform other functions.
The codebase of the OONI Probe CLI interface is available at github.com/ooni/probe-cli. More in detail, we have specified how the Desktop app will exec out the CLI to perform specific tasks and we have also specified the data format emitted by the CLI when running tests.
The CLI, in turn, is based on a lower-level “engine” library written in golang and available at github.com/ooni/probe-engine. This library aims to include a unified desktop and mobile implementation of all experiments. However, inside this library, we are still linking to our old C++ engine, Measurement Kit, to implement most OONI network experiments. (Measurement Kit is composed of several GitHub repositories available at github.com/measurement-kit.)
The work described in this activity is about gradually phasing out Measurement Kit, so that in the end all experiments are written in Go rather than in C++. This will mostly happen inside the probe-engine and will mostly concern changes in the implementation of tests (i.e. rewriting from C++ to Go). We believe that the API exposed by probe-engine will not change, except for minor changes in the options supported by the new experiments (for example, we may drop an option that does not make sense in the Go context or we may add an option that did not make sense for a C++ implementation).
This should explain why we believe that writing more OONI experiments in Go as part of probe-engine does not overlap with our previous work on creating the OONI Probe desktop apps (supported by our last OTF contract). In our view, this proposed work is simply the natural continuation of what we have been working on over the past year.
The advantages of rewriting in Go are the following:
* It will reduce the code size, because we can drop the C dependencies (whereas we cannot drop the dependency from Go since we want to ship a Psiphon experiment).
* It allows us to write experiments that can easily and safely be interrupted, because Go natively support this functionality, which is instead rather tricky to get right in C++.
* It allows us to easily wrap standard library functionality, hence allowing for easier low level measurements.
* It allows us to quickly cross compile for the operating systems we care about (Windows, macOS, Linux, Android, iOS); full builds will complete in minutes rather than hours, thus enabling faster development cycles.
* It will make OONI Probe safer, because Go is memory safe and has built-in support for concurrency.
Regarding the choice of focusing on the OONI Probe desktop app first (and on the mobile apps next) the reason is simple. Since probe-engine is already being used by the OONI Probe desktop apps via probe-cli, we have estimated that the effort (and cost) for doing this for the desktop apps is less than doing it on mobile, where we haven’t shipped golang code in production yet. In fact, we can update the desktop apps by recompiling probe-cli to use a new version of probe-engine that reimplements all tests in Go, with none or minimal API changes.
| priority | make the ooni probe desktop app rely entirely on the golang engine we are in the process of consolidating all the code related to running network experiments inside of a new golang based engine called probe engine which will replace measurement kit as part of this activity we will make the ooni probe desktop windows and macos app rely entirely on the golang engine we are not aiming to port the ooni probe desktop apps to a new codebase we are also not aiming to port the ooni cli to a new codebase the work we are proposing is rather to reduce the amount of c dependencies used by the ooni probe cli and the desktop apps below we explain in detail which components we are planning to modify why we are doing that and why we have chosen to start integrating these changes in the desktop apps rather than in the mobile apps the current ooni probe desktop apps for windows and macos are electron based apps implemented at github com ooni probe desktop these apps use the cli interface written in golang to run ooni experiments and perform other functions the codebase of the ooni probe cli interface is available at github com ooni probe cli more in detail we have specified how the desktop app will exec out the cli to perform specific tasks and we have also specified the data format emitted by the cli when running tests the cli in turn is based on a lower level “engine” library written in golang and available at github com ooni probe engine this library aims to include a unified desktop and mobile implementation of all experiments however inside this library we are still linking to our old c engine measurement kit to implement most ooni network experiments measurement kit is composed of several github repositories available at github com measurement kit the work described in this activity is about gradually phasing out measurement kit so that in the end all experiments are written in go rather than in c this will mostly happen inside the probe engine and will mostly concern changes in the implementation of tests i e rewriting from c to go we believe that the api exposed by probe engine will not change except for minor changes in the options supported by the new experiments for example we may drop an option that does not make sense in the go context or we may add an option that did not make sense for a c implementation this should explain why we believe that writing more ooni experiments in go as part of probe engine does not overlap with our previous work on creating the ooni probe desktop apps supported by our last otf contract in our view this proposed work is simply the natural continuation of what we have been working on over the past year the advantages of rewriting in go are the following it will reduce the code size because we can drop the c dependencies whereas we cannot drop the dependency from go since we want to ship a psiphon experiment it allows us to write experiments that can easily and safely be interrupted because go natively support this functionality which is instead rather tricky to get right in c it allows us to easily wrap standard library functionality hence allowing for easier low level measurements it allows us to quickly cross compile for the operating systems we care about windows macos linux android ios full builds will complete in minutes rather than hours thus enabling faster development cycles it will make ooni probe safer because go is memory safe and has built in support for concurrency regarding the choice of focusing on the ooni probe desktop app first and on the mobile apps next the reason is simple since probe engine is already being used by the ooni probe desktop apps via probe cli we have estimated that the effort and cost for doing this for the desktop apps is less than doing it on mobile where we haven’t shipped golang code in production yet in fact we can update the desktop apps by recompiling probe cli to use a new version of probe engine that reimplements all tests in go with none or minimal api changes | 1 |
272,107 | 8,499,185,828 | IssuesEvent | 2018-10-29 16:32:48 | spacetelescope/specviz | https://api.github.com/repos/spacetelescope/specviz | closed | Smoothing fails when trying it a second time | bug gui medium-priority | Smoothing a spectrum the first time, works fine. But when I try to smooth it a second time (click again on the spectrum on the left and the hit the smoothing button) the dialog window looks like this and it won't let me click anything on it.

That happens even if I delete the first smoothing spectrum. | 1.0 | Smoothing fails when trying it a second time - Smoothing a spectrum the first time, works fine. But when I try to smooth it a second time (click again on the spectrum on the left and the hit the smoothing button) the dialog window looks like this and it won't let me click anything on it.

That happens even if I delete the first smoothing spectrum. | priority | smoothing fails when trying it a second time smoothing a spectrum the first time works fine but when i try to smooth it a second time click again on the spectrum on the left and the hit the smoothing button the dialog window looks like this and it won t let me click anything on it that happens even if i delete the first smoothing spectrum | 1 |
357,821 | 10,618,352,164 | IssuesEvent | 2019-10-13 03:38:10 | carbon-design-system/ibm-dotcom-library | https://api.github.com/repos/carbon-design-system/ibm-dotcom-library | opened | Masthead | Search autosuggest not working in Firefox | bug dev dotcom migrate priority: medium to be triaged | _Jeff-Chew created the following on Aug 16:_
### Detailed description
@Kenny-Lam noticed this while testing the masthead locally. Autosuggest currently isn't working in Firefox.
### Steps to reproduce the issue
1. Go to https://ibmdotcom-react.netlify.com in Firefox
2. Activate search and type in a query (minimum 3 characters)
3. Search results not appearing
### Additional information
- Currently the preview isn't there as it's not merged to the main test environment. Will update the ticket once we can reproduce there.
_Original issue: https://github.ibm.com/webstandards/digital-design/issues/1473_ | 1.0 | Masthead | Search autosuggest not working in Firefox - _Jeff-Chew created the following on Aug 16:_
### Detailed description
@Kenny-Lam noticed this while testing the masthead locally. Autosuggest currently isn't working in Firefox.
### Steps to reproduce the issue
1. Go to https://ibmdotcom-react.netlify.com in Firefox
2. Activate search and type in a query (minimum 3 characters)
3. Search results not appearing
### Additional information
- Currently the preview isn't there as it's not merged to the main test environment. Will update the ticket once we can reproduce there.
_Original issue: https://github.ibm.com/webstandards/digital-design/issues/1473_ | priority | masthead search autosuggest not working in firefox jeff chew created the following on aug detailed description kenny lam noticed this while testing the masthead locally autosuggest currently isn t working in firefox steps to reproduce the issue go to in firefox activate search and type in a query minimum characters search results not appearing additional information currently the preview isn t there as it s not merged to the main test environment will update the ticket once we can reproduce there original issue | 1 |
332,364 | 10,092,004,950 | IssuesEvent | 2019-07-26 15:32:51 | wherebyus/general-tasks | https://api.github.com/repos/wherebyus/general-tasks | closed | When I try to type in the photo credit field on WP, it takes FOR-E-VER and sometimes just freezes and doesn't work. | Priority: Medium Product: Newsletters Type: Not A Bug UX: Not Validated | ## Feature or problem
See above ^
## UX Validation
Not Validated
### Suggested priority
Medium
### Stakeholders
*Submitted:* caitlin
### Definition of done
How will we know when this feature is complete?
### Subtasks
A detailed list of changes that need to be made or subtasks. One checkbox per.
- [ ] Brew the coffee
## Developer estimate
To help the team accurately estimate the complexity of this task,
take a moment to walk through this list and estimate each item. At the end, you can total
the estimates and round to the nearest prime number.
If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking
this issue into multiple smaller issues.
- [ ] Changes to the database ()
- [ ] Changes to the API ()
- [ ] Testing Changes to the API ()
- [ ] Changes to Application Code ()
- [ ] Adding or updating unit tests ()
- [ ] Local developer testing ()
### Total developer estimate: 0
## Additional estimate
- [ ] Code review ()
- [ ] QA Testing ()
- [ ] Stakeholder Sign-off ()
- [ ] Deploy to Production ()
### Total additional estimate:
## QA Notes
Detailed instructions for testing, one checkbox per test to be completed.
### Contextual tests
- [ ] Accessibility check
- [ ] Cross-browser check (Edge, Chrome, Firefox)
- [ ] Responsive check
| 1.0 | When I try to type in the photo credit field on WP, it takes FOR-E-VER and sometimes just freezes and doesn't work. - ## Feature or problem
See above ^
## UX Validation
Not Validated
### Suggested priority
Medium
### Stakeholders
*Submitted:* caitlin
### Definition of done
How will we know when this feature is complete?
### Subtasks
A detailed list of changes that need to be made or subtasks. One checkbox per.
- [ ] Brew the coffee
## Developer estimate
To help the team accurately estimate the complexity of this task,
take a moment to walk through this list and estimate each item. At the end, you can total
the estimates and round to the nearest prime number.
If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking
this issue into multiple smaller issues.
- [ ] Changes to the database ()
- [ ] Changes to the API ()
- [ ] Testing Changes to the API ()
- [ ] Changes to Application Code ()
- [ ] Adding or updating unit tests ()
- [ ] Local developer testing ()
### Total developer estimate: 0
## Additional estimate
- [ ] Code review ()
- [ ] QA Testing ()
- [ ] Stakeholder Sign-off ()
- [ ] Deploy to Production ()
### Total additional estimate:
## QA Notes
Detailed instructions for testing, one checkbox per test to be completed.
### Contextual tests
- [ ] Accessibility check
- [ ] Cross-browser check (Edge, Chrome, Firefox)
- [ ] Responsive check
| priority | when i try to type in the photo credit field on wp it takes for e ver and sometimes just freezes and doesn t work feature or problem see above ux validation not validated suggested priority medium stakeholders submitted caitlin definition of done how will we know when this feature is complete subtasks a detailed list of changes that need to be made or subtasks one checkbox per brew the coffee developer estimate to help the team accurately estimate the complexity of this task take a moment to walk through this list and estimate each item at the end you can total the estimates and round to the nearest prime number if any of these are at a or higher or if the total is above a consider breaking this issue into multiple smaller issues changes to the database changes to the api testing changes to the api changes to application code adding or updating unit tests local developer testing total developer estimate additional estimate code review qa testing stakeholder sign off deploy to production total additional estimate qa notes detailed instructions for testing one checkbox per test to be completed contextual tests accessibility check cross browser check edge chrome firefox responsive check | 1 |
651,058 | 21,448,829,960 | IssuesEvent | 2022-04-25 09:16:26 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | opened | [Редагувати профіль] After changing the role from "Керівник" to "Відвідувач", the user cannot return the role "Керівник" | bug Priority: Medium | **Environment:** Windows 10 Professional, Chrome 96.0.4664.110
**Reproducible:** always
**Build found:** https://speak-ukrainian.org.ua/dev/
**Preconditions**
1. Log in as a "Керівник"on https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль"
2. Change user role to "Відвідувач" and save changes
3. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль"
4. Try to change the user role back to the role "Керівник"
5. Pay attention to the result of this action
**Actual result**
The user with the "Керівник" role can lose its role without the possibility of going back
**Expected result**
The user with the "Керівник" role cannot change his role to the "Відвідувач" role


| 1.0 | [Редагувати профіль] After changing the role from "Керівник" to "Відвідувач", the user cannot return the role "Керівник" - **Environment:** Windows 10 Professional, Chrome 96.0.4664.110
**Reproducible:** always
**Build found:** https://speak-ukrainian.org.ua/dev/
**Preconditions**
1. Log in as a "Керівник"on https://speak-ukrainian.org.ua/dev/
**Steps to reproduce**
1. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль"
2. Change user role to "Відвідувач" and save changes
3. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль"
4. Try to change the user role back to the role "Керівник"
5. Pay attention to the result of this action
**Actual result**
The user with the "Керівник" role can lose its role without the possibility of going back
**Expected result**
The user with the "Керівник" role cannot change his role to the "Відвідувач" role


| priority | after changing the role from керівник to відвідувач the user cannot return the role керівник environment windows professional chrome reproducible always build found preconditions log in as a керівник on steps to reproduce go to the user cabinet at the top right corner мій профіль редагувати профіль change user role to відвідувач and save changes go to the user cabinet at the top right corner мій профіль редагувати профіль try to change the user role back to the role керівник pay attention to the result of this action actual result the user with the керівник role can lose its role without the possibility of going back expected result the user with the керівник role cannot change his role to the відвідувач role | 1 |
824,111 | 31,141,621,856 | IssuesEvent | 2023-08-16 00:49:36 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | Wrong hp amount in a query for npc with id=46645 | NPC Priority: Medium Zone: TolBarad | [//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
A mistake with this npc HP modifier, probably it has extra 000 or something in query responsive for HP
**How to reproduce:**
--- just check the hp
**How it should work:**
-- shouldn't be 30 mln hp here
**Database links:**
https://www.wowhead.com/npc=46645/cursed-stallion
| 1.0 | Wrong hp amount in a query for npc with id=46645 - [//]: # (REMBEMBER! Add links to things related to the bug using for example:)
[//]: # (http://wowhead.com/)
[//]: # (cata-twinhead.twinstar.cz)
**Description:**
A mistake with this npc HP modifier, probably it has extra 000 or something in query responsive for HP
**How to reproduce:**
--- just check the hp
**How it should work:**
-- shouldn't be 30 mln hp here
**Database links:**
https://www.wowhead.com/npc=46645/cursed-stallion
| priority | wrong hp amount in a query for npc with id rembember add links to things related to the bug using for example cata twinhead twinstar cz description a mistake with this npc hp modifier probably it has extra or something in query responsive for hp how to reproduce just check the hp how it should work shouldn t be mln hp here database links | 1 |
117,448 | 4,716,402,809 | IssuesEvent | 2016-10-16 01:42:47 | lzlw/HBONowHelper | https://api.github.com/repos/lzlw/HBONowHelper | opened | Get info via PROD id | enhancement in progress medium priority | http://content.hbonow.com/content/id/v1/PROD592264/detail/web-v1.json
This may possibly ensure that all videos will have valid durations in popup info boxes. | 1.0 | Get info via PROD id - http://content.hbonow.com/content/id/v1/PROD592264/detail/web-v1.json
This may possibly ensure that all videos will have valid durations in popup info boxes. | priority | get info via prod id this may possibly ensure that all videos will have valid durations in popup info boxes | 1 |
519,239 | 15,048,143,208 | IssuesEvent | 2021-02-03 09:49:46 | vmware/clarity | https://api.github.com/repos/vmware/clarity | closed | Modal is also dismissed when pressing the ESC key to close a combobox's dropdown panel. | component: combobox component: dropdown component: modal priority: 4 medium-high status: backlog type: bug v4 v5 | ## Describe the bug
I have a combobox inside a modal but when you try to press the ESC key to close the it's dropdown panel, the modal is also get dismissed.

## How to reproduce
https://stackblitz.com/edit/clarity-v4-light-theme-jz6if2?file=src/app/app.component.ts
Steps to reproduce the behavior:
1. Toggle the dropdown panel of the combo box.
2. Press ESC key.
## Expected behavior
It should only dismiss/close the items panel/dropdown panel of the combo box and not the modal.
## Versions
4.0.2
**App**
- Angular: 10
- Node: 12
- Clarity: 4.0.2
**Device:**
- Type: [e.g. MacBook]
- OS: [e.g. iOS]
- Browser: Chrome/Edge Chromium
- Version [e.g. 22]
## Additional notes
Also, you will notice that there's a console error when pressing the ESC key.
| 1.0 | Modal is also dismissed when pressing the ESC key to close a combobox's dropdown panel. - ## Describe the bug
I have a combobox inside a modal but when you try to press the ESC key to close the it's dropdown panel, the modal is also get dismissed.

## How to reproduce
https://stackblitz.com/edit/clarity-v4-light-theme-jz6if2?file=src/app/app.component.ts
Steps to reproduce the behavior:
1. Toggle the dropdown panel of the combo box.
2. Press ESC key.
## Expected behavior
It should only dismiss/close the items panel/dropdown panel of the combo box and not the modal.
## Versions
4.0.2
**App**
- Angular: 10
- Node: 12
- Clarity: 4.0.2
**Device:**
- Type: [e.g. MacBook]
- OS: [e.g. iOS]
- Browser: Chrome/Edge Chromium
- Version [e.g. 22]
## Additional notes
Also, you will notice that there's a console error when pressing the ESC key.
| priority | modal is also dismissed when pressing the esc key to close a combobox s dropdown panel describe the bug i have a combobox inside a modal but when you try to press the esc key to close the it s dropdown panel the modal is also get dismissed how to reproduce steps to reproduce the behavior toggle the dropdown panel of the combo box press esc key expected behavior it should only dismiss close the items panel dropdown panel of the combo box and not the modal versions app angular node clarity device type os browser chrome edge chromium version additional notes also you will notice that there s a console error when pressing the esc key | 1 |
198,951 | 6,979,423,108 | IssuesEvent | 2017-12-12 20:59:15 | compodoc/compodoc | https://api.github.com/repos/compodoc/compodoc | closed | [BUG] Partial documentation is generated | Maximum call stack size exceeded | 1. Type: Bug Priority: Medium Status: Accepted Time: ~3 hours | ##### **Overview of the issue**
When generating the documentation i get this error:
```javascript
RangeError: Maximum call stack size exceeded
```
The documentation is partial with missing components, directives and services.
I do have a couple of base64 images in the code, but there is a related issue stating this bug was fixed.
##### **Operating System, Node.js, npm, compodoc version(s)**
node: 8.5.0
npm: 5.3.0
os: win32 x64
compodoc: 1.0.3
##### **Angular configuration, a `package.json` file in the root folder**
```json
"devDependencies": {
"angular/animations": "4.4.4",
"angular/cli": "1.4.5",
"angular/common": "4.4.4",
"angular/compiler": "4.4.4",
"angular/compiler-cli": "4.4.4",
"angular/core": "4.4.4",
"angular/forms": "4.4.4",
"angular/http": "4.4.4",
"angular/language-service": "4.4.4",
"angular/platform-browser": "4.4.4",
"angular/platform-browser-dynamic": "4.4.4",
"angular/router": "4.4.4",
"compodoc/compodoc": "1.0.3",
"types/chai-as-promised": "^7.1.0",
"types/jasmine": "~2.5.54",
"types/jasminewd2": "2.0.3",
"types/node": "6.0.89",
"bootstrap": "4.0.0-beta",
"chai": "4.1.2",
"chai-as-promised": "7.1.1",
"codelyzer": "~3.1.1",
"core-js": "2.5.1",
"cucumber": "2.2.0",
"cucumber-tsflow": "2.2.0",
"font-awesome": "4.7.0",
"jasmine-core": "~2.6.2",
"jasmine-spec-reporter": "~4.1.0",
"jquery": "3.2.1",
"karma": "1.7.1",
"karma-chrome-launcher": "~2.1.1",
"karma-cli": "~1.0.1",
"karma-coverage-istanbul-reporter": "1.2.1",
"karma-jasmine": "~1.1.0",
"karma-jasmine-html-reporter": "0.2.2",
"popper.js": "1.12.5",
"primeng": "4.2.2",
"protractor": "~4.0.0",
"protractor-cucumber-framework": "4.0.8",
"rxjs": "5.4.3",
"ts-node": "~3.2.0",
"tslint": "~5.3.2",
"typescript": "~2.3.4",
"zone.js": "0.8.18"
},
```
##### **Related issues**
An issues was closed but still the problem persist.
[#269](https://github.com/compodoc/compodoc/issues/269)
| 1.0 | [BUG] Partial documentation is generated | Maximum call stack size exceeded - ##### **Overview of the issue**
When generating the documentation i get this error:
```javascript
RangeError: Maximum call stack size exceeded
```
The documentation is partial with missing components, directives and services.
I do have a couple of base64 images in the code, but there is a related issue stating this bug was fixed.
##### **Operating System, Node.js, npm, compodoc version(s)**
node: 8.5.0
npm: 5.3.0
os: win32 x64
compodoc: 1.0.3
##### **Angular configuration, a `package.json` file in the root folder**
```json
"devDependencies": {
"angular/animations": "4.4.4",
"angular/cli": "1.4.5",
"angular/common": "4.4.4",
"angular/compiler": "4.4.4",
"angular/compiler-cli": "4.4.4",
"angular/core": "4.4.4",
"angular/forms": "4.4.4",
"angular/http": "4.4.4",
"angular/language-service": "4.4.4",
"angular/platform-browser": "4.4.4",
"angular/platform-browser-dynamic": "4.4.4",
"angular/router": "4.4.4",
"compodoc/compodoc": "1.0.3",
"types/chai-as-promised": "^7.1.0",
"types/jasmine": "~2.5.54",
"types/jasminewd2": "2.0.3",
"types/node": "6.0.89",
"bootstrap": "4.0.0-beta",
"chai": "4.1.2",
"chai-as-promised": "7.1.1",
"codelyzer": "~3.1.1",
"core-js": "2.5.1",
"cucumber": "2.2.0",
"cucumber-tsflow": "2.2.0",
"font-awesome": "4.7.0",
"jasmine-core": "~2.6.2",
"jasmine-spec-reporter": "~4.1.0",
"jquery": "3.2.1",
"karma": "1.7.1",
"karma-chrome-launcher": "~2.1.1",
"karma-cli": "~1.0.1",
"karma-coverage-istanbul-reporter": "1.2.1",
"karma-jasmine": "~1.1.0",
"karma-jasmine-html-reporter": "0.2.2",
"popper.js": "1.12.5",
"primeng": "4.2.2",
"protractor": "~4.0.0",
"protractor-cucumber-framework": "4.0.8",
"rxjs": "5.4.3",
"ts-node": "~3.2.0",
"tslint": "~5.3.2",
"typescript": "~2.3.4",
"zone.js": "0.8.18"
},
```
##### **Related issues**
An issues was closed but still the problem persist.
[#269](https://github.com/compodoc/compodoc/issues/269)
| priority | partial documentation is generated maximum call stack size exceeded overview of the issue when generating the documentation i get this error javascript rangeerror maximum call stack size exceeded the documentation is partial with missing components directives and services i do have a couple of images in the code but there is a related issue stating this bug was fixed operating system node js npm compodoc version s node npm os compodoc angular configuration a package json file in the root folder json devdependencies angular animations angular cli angular common angular compiler angular compiler cli angular core angular forms angular http angular language service angular platform browser angular platform browser dynamic angular router compodoc compodoc types chai as promised types jasmine types types node bootstrap beta chai chai as promised codelyzer core js cucumber cucumber tsflow font awesome jasmine core jasmine spec reporter jquery karma karma chrome launcher karma cli karma coverage istanbul reporter karma jasmine karma jasmine html reporter popper js primeng protractor protractor cucumber framework rxjs ts node tslint typescript zone js related issues an issues was closed but still the problem persist | 1 |
400,709 | 11,779,593,671 | IssuesEvent | 2020-03-16 18:19:31 | buddyboss/buddyboss-platform | https://api.github.com/repos/buddyboss/buddyboss-platform | opened | Email view - HTML and Email Tokens are stripped down from Email Invites when used with MailGun | bug component: invites priority: medium | **Describe the bug**
On Email view, HTML and Email Tokens are stripped down from Email Invites when used with MailGun
**To Reproduce**
Steps to reproduce the behavior:
1. Set up a MailGun account (free version)
2. Install MailGun Plugin https://wordpress.org/plugins/mailgun/
3. Setup credentials in MailGun plguin from created mailgun account
4. On frontend, try to invite a user on Email Invites on your profile
5. Check invited users email, all html and email tokens are stripped
**Expected behavior**
Email Invites should have a layout
**Screenshots**
https://www.loom.com/share/816b15a82b124e77be38e979f2a4b7c0


**Support ticket links**
https://buddyboss.zendesk.com/agent/tickets/63879
| 1.0 | Email view - HTML and Email Tokens are stripped down from Email Invites when used with MailGun - **Describe the bug**
On Email view, HTML and Email Tokens are stripped down from Email Invites when used with MailGun
**To Reproduce**
Steps to reproduce the behavior:
1. Set up a MailGun account (free version)
2. Install MailGun Plugin https://wordpress.org/plugins/mailgun/
3. Setup credentials in MailGun plguin from created mailgun account
4. On frontend, try to invite a user on Email Invites on your profile
5. Check invited users email, all html and email tokens are stripped
**Expected behavior**
Email Invites should have a layout
**Screenshots**
https://www.loom.com/share/816b15a82b124e77be38e979f2a4b7c0


**Support ticket links**
https://buddyboss.zendesk.com/agent/tickets/63879
| priority | email view html and email tokens are stripped down from email invites when used with mailgun describe the bug on email view html and email tokens are stripped down from email invites when used with mailgun to reproduce steps to reproduce the behavior set up a mailgun account free version install mailgun plugin setup credentials in mailgun plguin from created mailgun account on frontend try to invite a user on email invites on your profile check invited users email all html and email tokens are stripped expected behavior email invites should have a layout screenshots support ticket links | 1 |
197,369 | 6,954,842,441 | IssuesEvent | 2017-12-07 03:51:47 | ODIQueensland/data-curator | https://api.github.com/repos/ODIQueensland/data-curator | closed | Provide "pick list" of tables and fields for Foreign and Primary Keys | f:Feature-request fn:Set-Properties priority:Medium status:In-progress | ### Desired Behaviour
Allow the user to select from a list of valid table or field names to specify the [Foreign ](http://specs.frictionlessdata.io/table-schema/#foreign-keys)and [Primary Keys](http://specs.frictionlessdata.io/table-schema/#primary-key).
Support the selection of multiple fields for keys by adding like the current sources function
| 1.0 | Provide "pick list" of tables and fields for Foreign and Primary Keys - ### Desired Behaviour
Allow the user to select from a list of valid table or field names to specify the [Foreign ](http://specs.frictionlessdata.io/table-schema/#foreign-keys)and [Primary Keys](http://specs.frictionlessdata.io/table-schema/#primary-key).
Support the selection of multiple fields for keys by adding like the current sources function
| priority | provide pick list of tables and fields for foreign and primary keys desired behaviour allow the user to select from a list of valid table or field names to specify the support the selection of multiple fields for keys by adding like the current sources function | 1 |
530,595 | 15,434,473,661 | IssuesEvent | 2021-03-07 03:33:00 | timmo001/system-bridge | https://api.github.com/repos/timmo001/system-bridge | opened | Set Windows Desktop and Start Menu Shortcut | enhancement priority-medium | # Description
Icons are currently set to `system-bridge` for Windows. Linux is as it should be.
## Suggested Actions

to

| 1.0 | Set Windows Desktop and Start Menu Shortcut - # Description
Icons are currently set to `system-bridge` for Windows. Linux is as it should be.
## Suggested Actions

to

| priority | set windows desktop and start menu shortcut description icons are currently set to system bridge for windows linux is as it should be suggested actions to | 1 |
16,570 | 2,615,119,082 | IssuesEvent | 2015-03-01 05:44:56 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | opened | Spreadsheet (List?) access (specifically to iterate through all the rows on a given sheet) | auto-migrated Priority-Medium Type-Sample | ```
Which API and version (e.g. Google Calendar Data API version 2)?
google-api-java-client v1.3.1 (or greater if no one gets around to this before
the next drop)
What format (e.g. JSON, Atom)?
JSON
What Authentication (e.g. OAuth, OAuth 2, Android, ClientLogin)?
Whatever is recommended when using Google credentials. It's all new to me.
Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.2)?
Java 6, Windows 7 x64, Netbeans 6.9.1
External references, such as API reference guide?
Please provide any additional information below.
All existing samples are for the legacy GDATA API as far as I can tell.
```
Original issue reported on code.google.com by `Predato...@gmail.com` on 3 Mar 2011 at 4:15 | 1.0 | Spreadsheet (List?) access (specifically to iterate through all the rows on a given sheet) - ```
Which API and version (e.g. Google Calendar Data API version 2)?
google-api-java-client v1.3.1 (or greater if no one gets around to this before
the next drop)
What format (e.g. JSON, Atom)?
JSON
What Authentication (e.g. OAuth, OAuth 2, Android, ClientLogin)?
Whatever is recommended when using Google credentials. It's all new to me.
Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.2)?
Java 6, Windows 7 x64, Netbeans 6.9.1
External references, such as API reference guide?
Please provide any additional information below.
All existing samples are for the legacy GDATA API as far as I can tell.
```
Original issue reported on code.google.com by `Predato...@gmail.com` on 3 Mar 2011 at 4:15 | priority | spreadsheet list access specifically to iterate through all the rows on a given sheet which api and version e g google calendar data api version google api java client or greater if no one gets around to this before the next drop what format e g json atom json what authentication e g oauth oauth android clientlogin whatever is recommended when using google credentials it s all new to me java environment e g java android app engine java windows netbeans external references such as api reference guide please provide any additional information below all existing samples are for the legacy gdata api as far as i can tell original issue reported on code google com by predato gmail com on mar at | 1 |
412,090 | 12,035,187,763 | IssuesEvent | 2020-04-13 17:25:51 | ChainSafe/forest | https://api.github.com/repos/ChainSafe/forest | opened | Implement Randomness | Priority: 3 - Medium VM | In the Lotus implementation, they use a RAND library within the VM context.
| 1.0 | Implement Randomness - In the Lotus implementation, they use a RAND library within the VM context.
| priority | implement randomness in the lotus implementation they use a rand library within the vm context | 1 |
202,796 | 7,055,241,966 | IssuesEvent | 2018-01-04 07:03:58 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Hazelcast : Set Map Configuration to Cluster Node without Creating Instance of Hazelcast locally | Priority: Medium Source: Community Team: Core Type: Enhancement | Hi
Below are my cluster members and I wan to pass some Map configuration to those node so once that Map create all node use that configuration.
Members [3] {
Member [169.185.209.99]:5701
Member [165.203.252.170]:5701
Member [169.185.209.169]:5701
}
How can i pass Map config without creating hazlecast Instance on (169.164.232.143) my local system.
this is really urgent..
Thanks
Vipul
| 1.0 | Hazelcast : Set Map Configuration to Cluster Node without Creating Instance of Hazelcast locally - Hi
Below are my cluster members and I wan to pass some Map configuration to those node so once that Map create all node use that configuration.
Members [3] {
Member [169.185.209.99]:5701
Member [165.203.252.170]:5701
Member [169.185.209.169]:5701
}
How can i pass Map config without creating hazlecast Instance on (169.164.232.143) my local system.
this is really urgent..
Thanks
Vipul
| priority | hazelcast set map configuration to cluster node without creating instance of hazelcast locally hi below are my cluster members and i wan to pass some map configuration to those node so once that map create all node use that configuration members member member member how can i pass map config without creating hazlecast instance on my local system this is really urgent thanks vipul | 1 |
28,492 | 2,704,467,267 | IssuesEvent | 2015-04-07 00:49:56 | kevinchandler/test | https://api.github.com/repos/kevinchandler/test | closed | dfsafsdfsdfds | bug-report-cs Priority: Medium | Description: dfsafsdfsdfds
Priority: Medium
Date Reported: 04-23-2015
User Email: kevin@onradpad.com
Profile: https://onradpad.com/admin/landlords/837750
Platform: Mac
Able to Replicate: No
Reported By: Kevin C. - kevin@onradpad.com | 1.0 | dfsafsdfsdfds - Description: dfsafsdfsdfds
Priority: Medium
Date Reported: 04-23-2015
User Email: kevin@onradpad.com
Profile: https://onradpad.com/admin/landlords/837750
Platform: Mac
Able to Replicate: No
Reported By: Kevin C. - kevin@onradpad.com | priority | dfsafsdfsdfds description dfsafsdfsdfds priority medium date reported user email kevin onradpad com profile platform mac able to replicate no reported by kevin c kevin onradpad com | 1 |
452,899 | 13,061,011,366 | IssuesEvent | 2020-07-30 13:19:35 | clastix/capsule | https://api.github.com/repos/clastix/capsule | closed | Make the logs level configurable | enhancement medium-priority | It would be nice to have a configurable logs level for the capsule controller, e.g. info, warning, debug, trace, .. This will help during troubleshooting. | 1.0 | Make the logs level configurable - It would be nice to have a configurable logs level for the capsule controller, e.g. info, warning, debug, trace, .. This will help during troubleshooting. | priority | make the logs level configurable it would be nice to have a configurable logs level for the capsule controller e g info warning debug trace this will help during troubleshooting | 1 |
399,744 | 11,759,908,903 | IssuesEvent | 2020-03-13 18:16:07 | ChainSafe/ChainBridgeV2 | https://api.github.com/repos/ChainSafe/ChainBridgeV2 | closed | Convert JS scripts to CLI | Ethereum Priority: 3 - Medium | The current JS scripts are a hacked together CLI. They should be switched to a command based approach.
The commands should be:`deploy` & `transfer`, with the respective flags. | 1.0 | Convert JS scripts to CLI - The current JS scripts are a hacked together CLI. They should be switched to a command based approach.
The commands should be:`deploy` & `transfer`, with the respective flags. | priority | convert js scripts to cli the current js scripts are a hacked together cli they should be switched to a command based approach the commands should be deploy transfer with the respective flags | 1 |
326,906 | 9,962,356,286 | IssuesEvent | 2019-07-07 13:55:04 | PMEAL/OpenPNM | https://api.github.com/repos/PMEAL/OpenPNM | closed | Networks with 1 pore and 0 throats break stuff | Next Minor Release Priority - Medium bug | Not that this would ever happen, but it might be worth throwing an error or warning? | 1.0 | Networks with 1 pore and 0 throats break stuff - Not that this would ever happen, but it might be worth throwing an error or warning? | priority | networks with pore and throats break stuff not that this would ever happen but it might be worth throwing an error or warning | 1 |
43,989 | 2,894,941,497 | IssuesEvent | 2015-06-16 05:05:00 | wchill/steamSummerMinigame | https://api.github.com/repos/wchill/steamSummerMinigame | closed | [Suggestion] Auto Use 'Treasure' On Bosses | 2 - Medium Priority | This would make sense as is also does a free metal detector too, ideally for a Boss who has more than 60% health as per 'Raining Gold.' | 1.0 | [Suggestion] Auto Use 'Treasure' On Bosses - This would make sense as is also does a free metal detector too, ideally for a Boss who has more than 60% health as per 'Raining Gold.' | priority | auto use treasure on bosses this would make sense as is also does a free metal detector too ideally for a boss who has more than health as per raining gold | 1 |
77,297 | 3,506,339,225 | IssuesEvent | 2016-01-08 05:53:38 | OregonCore/OregonCore | https://api.github.com/repos/OregonCore/OregonCore | closed | Full rage when arena begins (BB #344) | migrated Priority: Medium Type: Bug | This issue was migrated from bitbucket.
**Original Reporter:** henhouse
**Original Date:** 07.11.2010 18:33:33 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/344
<hr>
It's possible to attack an enemy and get full rage then entering an arena with full rage. It doesn't reset to 0 at the start of an arena match. | 1.0 | Full rage when arena begins (BB #344) - This issue was migrated from bitbucket.
**Original Reporter:** henhouse
**Original Date:** 07.11.2010 18:33:33 GMT+0000
**Original Priority:** major
**Original Type:** bug
**Original State:** resolved
**Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/344
<hr>
It's possible to attack an enemy and get full rage then entering an arena with full rage. It doesn't reset to 0 at the start of an arena match. | priority | full rage when arena begins bb this issue was migrated from bitbucket original reporter henhouse original date gmt original priority major original type bug original state resolved direct link it s possible to attack an enemy and get full rage then entering an arena with full rage it doesn t reset to at the start of an arena match | 1 |
738,271 | 25,550,823,828 | IssuesEvent | 2022-11-29 23:31:11 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [xCluster] Race condition with SPLIT_OP processing and CdcConsumer shutdown | kind/bug area/docdb priority/medium | Jira Link: [DB-836](https://yugabyte.atlassian.net/browse/DB-836)
### Description
Noticed when working on `XClusterAutomaticTabletSplitITest.AutomaticTabletSplitting`.
There is a race occurring during the shutdown path, if we shutdown immediately after sending a DeleteUniverseReplication request and before waiting for all the CDCConsumers to be properly cleared.
From my investigations, we seem to be getting stuck in `rpcs_->Abort({&write_handle_});` during `~TwoDCOutputClient()`, but unclear as to what its waiting on..
Currently working around in tests by waiting for pollers to be deleted by DeleteUniverseReplication before continuing with the shutdown. | 1.0 | [xCluster] Race condition with SPLIT_OP processing and CdcConsumer shutdown - Jira Link: [DB-836](https://yugabyte.atlassian.net/browse/DB-836)
### Description
Noticed when working on `XClusterAutomaticTabletSplitITest.AutomaticTabletSplitting`.
There is a race occurring during the shutdown path, if we shutdown immediately after sending a DeleteUniverseReplication request and before waiting for all the CDCConsumers to be properly cleared.
From my investigations, we seem to be getting stuck in `rpcs_->Abort({&write_handle_});` during `~TwoDCOutputClient()`, but unclear as to what its waiting on..
Currently working around in tests by waiting for pollers to be deleted by DeleteUniverseReplication before continuing with the shutdown. | priority | race condition with split op processing and cdcconsumer shutdown jira link description noticed when working on xclusterautomatictabletsplititest automatictabletsplitting there is a race occurring during the shutdown path if we shutdown immediately after sending a deleteuniversereplication request and before waiting for all the cdcconsumers to be properly cleared from my investigations we seem to be getting stuck in rpcs abort write handle during twodcoutputclient but unclear as to what its waiting on currently working around in tests by waiting for pollers to be deleted by deleteuniversereplication before continuing with the shutdown | 1 |
728,827 | 25,094,734,667 | IssuesEvent | 2022-11-08 09:21:54 | Zenika/skillZ | https://api.github.com/repos/Zenika/skillZ | closed | add admin emails | Priority: Medium | I've just added Benjamin Adrien and me for the tests. The core team and Pierre should be added in the clever env var | 1.0 | add admin emails - I've just added Benjamin Adrien and me for the tests. The core team and Pierre should be added in the clever env var | priority | add admin emails i ve just added benjamin adrien and me for the tests the core team and pierre should be added in the clever env var | 1 |
41,132 | 2,868,977,969 | IssuesEvent | 2015-06-05 22:20:10 | dart-lang/sdk | https://api.github.com/repos/dart-lang/sdk | closed | Make pub run use an isolate instead of a process for the spawned program | Area-Pub Priority-Medium Triaged Type-Enhancement | Right now, it runs the app in a separate process. We should use an isolate because it will play nicer with pipes and detecting the stdioType in the spawned app. | 1.0 | Make pub run use an isolate instead of a process for the spawned program - Right now, it runs the app in a separate process. We should use an isolate because it will play nicer with pipes and detecting the stdioType in the spawned app. | priority | make pub run use an isolate instead of a process for the spawned program right now it runs the app in a separate process we should use an isolate because it will play nicer with pipes and detecting the stdiotype in the spawned app | 1 |
465,077 | 13,351,240,303 | IssuesEvent | 2020-08-30 12:35:19 | ncssar/sartopo_address | https://api.github.com/repos/ncssar/sartopo_address | opened | inconsistent defafult-folder-per-custom-marker behavior | Priority: medium bug | clean up this behavior:
- folder name should revert to evac- folder name, not to 'Folder', when evac- symbol is selected and folder combo box is released
- custom-default-folder name doesn't always appear in folder combo box; see video https://www.youtube.com/watch?v=1U9VOujH7Es | 1.0 | inconsistent defafult-folder-per-custom-marker behavior - clean up this behavior:
- folder name should revert to evac- folder name, not to 'Folder', when evac- symbol is selected and folder combo box is released
- custom-default-folder name doesn't always appear in folder combo box; see video https://www.youtube.com/watch?v=1U9VOujH7Es | priority | inconsistent defafult folder per custom marker behavior clean up this behavior folder name should revert to evac folder name not to folder when evac symbol is selected and folder combo box is released custom default folder name doesn t always appear in folder combo box see video | 1 |
282,644 | 8,709,073,472 | IssuesEvent | 2018-12-06 12:55:14 | 7552-2C-2018/App-Server | https://api.github.com/repos/7552-2C-2018/App-Server | closed | Get Questions | Priority-Medium feature | Devuelve la preguntas y respuestas de una publicación.
Se envia facebookID, token, id de publicacion{facebookID y pubDate}.
Devuelve JSONArray de {"question": string, "answer": string}
answer puede no estar | 1.0 | Get Questions - Devuelve la preguntas y respuestas de una publicación.
Se envia facebookID, token, id de publicacion{facebookID y pubDate}.
Devuelve JSONArray de {"question": string, "answer": string}
answer puede no estar | priority | get questions devuelve la preguntas y respuestas de una publicación se envia facebookid token id de publicacion facebookid y pubdate devuelve jsonarray de question string answer string answer puede no estar | 1 |
714,379 | 24,559,915,275 | IssuesEvent | 2022-10-12 19:14:28 | asastats/channel | https://api.github.com/repos/asastats/channel | closed | One of the Alchemon addresses fails to be added to the related collection | bug medium priority addressed | DragMZ in [Discord](https://discord.com/channels/906917846754418770/908054330265960478/1029809964669218857):
> OJGTHEJ2O5NXN7FVXDZZEEJTUEQHHCIYIE5MWY6BEFVVLZ2KANJODBOKGA Alchemons collection isn't recognized.
| 1.0 | One of the Alchemon addresses fails to be added to the related collection - DragMZ in [Discord](https://discord.com/channels/906917846754418770/908054330265960478/1029809964669218857):
> OJGTHEJ2O5NXN7FVXDZZEEJTUEQHHCIYIE5MWY6BEFVVLZ2KANJODBOKGA Alchemons collection isn't recognized.
| priority | one of the alchemon addresses fails to be added to the related collection dragmz in alchemons collection isn t recognized | 1 |
247,040 | 7,896,142,210 | IssuesEvent | 2018-06-29 07:29:01 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Provide an OpenCLI() cli method which launches a new cli instance | Expected Use: 3 - Occasional Feature Impact: 3 - Medium Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 2 - Minor Irritation Support Group: DOE/ASC version: 2.0.0 | rm-id: 38
rm-submitter: Cyrus Harrison
rm-submit-date: 05/14/2010 03:05 pm
When using the visitmodule, Bob Corey expected visit.AddArgument("-cli") to pop up a new cli window when visit.Launch() was called.
What he actually wants is the ability to launch a new cli instance from the VisIt python module, similar to OpenGUI(). So we should provide an OpenCLI() method to enable this.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Cyrus Harrison
Original creation: 06/21/2010 07:14 pm
Original update: 07/14/2010 03:01 pm
Ticket number: 23 | 1.0 | Provide an OpenCLI() cli method which launches a new cli instance - rm-id: 38
rm-submitter: Cyrus Harrison
rm-submit-date: 05/14/2010 03:05 pm
When using the visitmodule, Bob Corey expected visit.AddArgument("-cli") to pop up a new cli window when visit.Launch() was called.
What he actually wants is the ability to launch a new cli instance from the VisIt python module, similar to OpenGUI(). So we should provide an OpenCLI() method to enable this.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Cyrus Harrison
Original creation: 06/21/2010 07:14 pm
Original update: 07/14/2010 03:01 pm
Ticket number: 23 | priority | provide an opencli cli method which launches a new cli instance rm id rm submitter cyrus harrison rm submit date pm when using the visitmodule bob corey expected visit addargument cli to pop up a new cli window when visit launch was called what he actually wants is the ability to launch a new cli instance from the visit python module similar to opengui so we should provide an opencli method to enable this redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author cyrus harrison original creation pm original update pm ticket number | 1 |
3,160 | 2,537,393,613 | IssuesEvent | 2015-01-26 20:16:38 | web2py/web2py | https://api.github.com/repos/web2py/web2py | opened | Querying on 1:M Relationships in Grid and Smartgrid | 1 star enhancement imported Priority-Medium | _From [mike.leo...@gmail.com](https://code.google.com/u/109626380233121956961/) on September 05, 2012 22:07:38_
Hello,
We're building a web2py application that leverages SQLFORM.smartgrid heavily. Out of the box, this does not support filtering on 1:M relationships, or reference fields on the model. We understand the concerns about this: It could potentially be a huge performance hit, especially if the related table has a large number of rows.
On the other hand, this functionality is very valuable. Being able to filter on related tables is a very common need in applications. Often, users need to filter on "lookup" tables that have a relatively small number of rows, and the performance hit is negligible. This is a very common feature in Django, and in Rails admin interfaces like ActiveScaffold and RailsAdmin.
Example:
Given a model like this:
db.define_table('book',
Field('user_id', 'reference auth_user'),
Field('title'),
Field('genre'))
And a controller like this:
def index():
grid = SQLFORM.smartgrid(db.book)
return dict(grid=grid)
The query interface will look like this by default: http://i49.tinypic.com/2ugjt02.jpg Note how this includes title and genre, but not the auth_user reference.
We patched web2py to allow to for 1:M reference fields in the query interface. The resulting interface and generated query look something like this: http://i46.tinypic.com/28taq74.png Here is our simple patch to allow searching on reference fields like this (web2py version 1.99.7): https://gist.github.com/ff543a5c6d3bc14b9079 This is obviously NOT ready for production; it's a simple proof of concept. Note the potentially very expensive "ALL" query. Because of the performance risks, something like this only makes sense if developers can opt in, i.e. explicitly say so if they want to include a given reference field in the query interface.
1. Would the web2py maintiners be interested in a feature like this if it is opt-in? I'm willing to devote some time to submit a more robust patch if I could get some guidance.
2. I also have a simple implementation for querying on list:reference fields that I can share as well.
Finally, I just want to thank the web2py team. I've really enjoyed using the framework and hope to help out as much as I can.
Thanks,
Mike Leone
Principal, Engineering
Panoptic Development, Inc.
\-----------------------
On Friday, August 31, 2012 5:48:17 PM UTC-4, Massimo Di Pierro wrote:
Hello Mike,
thanks for your patch? Would you please attach it to a google code issue so we do not forget. We cannot add new feaures until all 2.0.x issues have been resolved.
_Original issue: http://code.google.com/p/web2py/issues/detail?id=979_ | 1.0 | Querying on 1:M Relationships in Grid and Smartgrid - _From [mike.leo...@gmail.com](https://code.google.com/u/109626380233121956961/) on September 05, 2012 22:07:38_
Hello,
We're building a web2py application that leverages SQLFORM.smartgrid heavily. Out of the box, this does not support filtering on 1:M relationships, or reference fields on the model. We understand the concerns about this: It could potentially be a huge performance hit, especially if the related table has a large number of rows.
On the other hand, this functionality is very valuable. Being able to filter on related tables is a very common need in applications. Often, users need to filter on "lookup" tables that have a relatively small number of rows, and the performance hit is negligible. This is a very common feature in Django, and in Rails admin interfaces like ActiveScaffold and RailsAdmin.
Example:
Given a model like this:
db.define_table('book',
Field('user_id', 'reference auth_user'),
Field('title'),
Field('genre'))
And a controller like this:
def index():
grid = SQLFORM.smartgrid(db.book)
return dict(grid=grid)
The query interface will look like this by default: http://i49.tinypic.com/2ugjt02.jpg Note how this includes title and genre, but not the auth_user reference.
We patched web2py to allow to for 1:M reference fields in the query interface. The resulting interface and generated query look something like this: http://i46.tinypic.com/28taq74.png Here is our simple patch to allow searching on reference fields like this (web2py version 1.99.7): https://gist.github.com/ff543a5c6d3bc14b9079 This is obviously NOT ready for production; it's a simple proof of concept. Note the potentially very expensive "ALL" query. Because of the performance risks, something like this only makes sense if developers can opt in, i.e. explicitly say so if they want to include a given reference field in the query interface.
1. Would the web2py maintiners be interested in a feature like this if it is opt-in? I'm willing to devote some time to submit a more robust patch if I could get some guidance.
2. I also have a simple implementation for querying on list:reference fields that I can share as well.
Finally, I just want to thank the web2py team. I've really enjoyed using the framework and hope to help out as much as I can.
Thanks,
Mike Leone
Principal, Engineering
Panoptic Development, Inc.
\-----------------------
On Friday, August 31, 2012 5:48:17 PM UTC-4, Massimo Di Pierro wrote:
Hello Mike,
thanks for your patch? Would you please attach it to a google code issue so we do not forget. We cannot add new feaures until all 2.0.x issues have been resolved.
_Original issue: http://code.google.com/p/web2py/issues/detail?id=979_ | priority | querying on m relationships in grid and smartgrid from on september hello we re building a application that leverages sqlform smartgrid heavily out of the box this does not support filtering on m relationships or reference fields on the model we understand the concerns about this it could potentially be a huge performance hit especially if the related table has a large number of rows on the other hand this functionality is very valuable being able to filter on related tables is a very common need in applications often users need to filter on lookup tables that have a relatively small number of rows and the performance hit is negligible this is a very common feature in django and in rails admin interfaces like activescaffold and railsadmin example given a model like this db define table book field user id reference auth user field title field genre and a controller like this def index grid sqlform smartgrid db book return dict grid grid the query interface will look like this by default note how this includes title and genre but not the auth user reference we patched to allow to for m reference fields in the query interface the resulting interface and generated query look something like this here is our simple patch to allow searching on reference fields like this version this is obviously not ready for production it s a simple proof of concept note the potentially very expensive all query because of the performance risks something like this only makes sense if developers can opt in i e explicitly say so if they want to include a given reference field in the query interface would the maintiners be interested in a feature like this if it is opt in i m willing to devote some time to submit a more robust patch if i could get some guidance i also have a simple implementation for querying on list reference fields that i can share as well finally i just want to thank the team i ve really enjoyed using the framework and hope to help out as much as i can thanks mike leone principal engineering panoptic development inc on friday august pm utc massimo di pierro wrote hello mike thanks for your patch would you please attach it to a google code issue so we do not forget we cannot add new feaures until all x issues have been resolved original issue | 1 |
543,970 | 15,888,076,453 | IssuesEvent | 2021-04-10 05:30:48 | AY2021S2-CS2103T-T12-4/tp | https://api.github.com/repos/AY2021S2-CS2103T-T12-4/tp | closed | DG: Implementation - Send & Run feature | priority.High severity.Medium | - refactor description to focus on run instead as the request feature talks about send extensively
- update to the UML diagram
- be ready by tonight or saturday morning | 1.0 | DG: Implementation - Send & Run feature - - refactor description to focus on run instead as the request feature talks about send extensively
- update to the UML diagram
- be ready by tonight or saturday morning | priority | dg implementation send run feature refactor description to focus on run instead as the request feature talks about send extensively update to the uml diagram be ready by tonight or saturday morning | 1 |
196,931 | 6,950,957,130 | IssuesEvent | 2017-12-06 12:49:50 | elementary/music | https://api.github.com/repos/elementary/music | closed | Album Art not populated until a song from any given album is played first (art embedded in ID3 tag) | Priority: Medium | Like it says on the tin.
I have ~85G of music. After waiting the 30-60 minutes for it to scan and populate the noise database, I noticed two things:
1. Noise was frozen, and taking up 100% of one of the CPU cores
2. Not a single album cover, from any album or single, was visible
After restarting the product, I played a song. About 4 seconds into the song, the ablum art showed up, and remains visible to this day.
I'm not sure if this is planned behavior or not, but since noise already scans all of the files' metadata to populate the DB, would it not make sense to also scan for the cover and populate the DB (or .local/cache/noise directory) with the covers in 1 shot? I think I would have been OK with waiting slightly longer and seeing all of my album art when the scan was done as opposed to how it works right now, where I've been playing songs on shuffle and only have the albums who's songs have played be populated with art.
Launchpad Details: [#LP1215558](https://bugs.launchpad.net/bugs/1215558) GonzO - 2013-08-22 17:28:04 +0000 | 1.0 | Album Art not populated until a song from any given album is played first (art embedded in ID3 tag) - Like it says on the tin.
I have ~85G of music. After waiting the 30-60 minutes for it to scan and populate the noise database, I noticed two things:
1. Noise was frozen, and taking up 100% of one of the CPU cores
2. Not a single album cover, from any album or single, was visible
After restarting the product, I played a song. About 4 seconds into the song, the ablum art showed up, and remains visible to this day.
I'm not sure if this is planned behavior or not, but since noise already scans all of the files' metadata to populate the DB, would it not make sense to also scan for the cover and populate the DB (or .local/cache/noise directory) with the covers in 1 shot? I think I would have been OK with waiting slightly longer and seeing all of my album art when the scan was done as opposed to how it works right now, where I've been playing songs on shuffle and only have the albums who's songs have played be populated with art.
Launchpad Details: [#LP1215558](https://bugs.launchpad.net/bugs/1215558) GonzO - 2013-08-22 17:28:04 +0000 | priority | album art not populated until a song from any given album is played first art embedded in tag like it says on the tin i have of music after waiting the minutes for it to scan and populate the noise database i noticed two things noise was frozen and taking up of one of the cpu cores not a single album cover from any album or single was visible after restarting the product i played a song about seconds into the song the ablum art showed up and remains visible to this day i m not sure if this is planned behavior or not but since noise already scans all of the files metadata to populate the db would it not make sense to also scan for the cover and populate the db or local cache noise directory with the covers in shot i think i would have been ok with waiting slightly longer and seeing all of my album art when the scan was done as opposed to how it works right now where i ve been playing songs on shuffle and only have the albums who s songs have played be populated with art launchpad details gonzo | 1 |
641,717 | 20,832,832,181 | IssuesEvent | 2022-03-19 18:29:02 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | LoRaWAN: Unable to correctly join networks of any version on LTS | bug priority: medium LTS area: LoRa | **Describe the bug**
The LoRaWAN subsystem is unable to correctly join networks of any LoRaWAN protocol version on the LTS branch.
LoRaWAN fork: https://github.com/zephyrproject-rtos/loramac-node
The relevant versions for this issue:
1. v1.0.3 (Previous version)
2. v1.0.4 (Current version targeted by the Zephyr fork)
3. v1.1.x (Not supported at all by the release that the Zephyr fork is based upon)
The PR https://github.com/zephyrproject-rtos/loramac-node/pull/9 updated the release the fork was based upon from some older version (v1.0.3?) to v1.0.4. Part of the specification changes from v1.0.3 to v1.0.4 was a change in requirements around the device nonce value, sent in the join request as a way of avoiding replay attacks.
*LoRaWAN v1.0.3 Specification (6.2.4)*
> DevNonce is a random value.1 For each end-device, the network server keeps track of a certain number of DevNonce values used by the end-device in the past, and ignores join requests with any of these DevNonce values from that end-device.
*LoRaWAN v1.0.4 Specification (6.2.5)*
> DevNonce is a counter starting at 0 when the end-device is initially powered up and incremented with every Join-Request. A DevNonce value SHALL never be reused for a given JoinEUI value. If the end-device can be power-cycled, then DevNonce SHALL be persistent (e.g., stored in a non-volatile memory). Resetting DevNonce without changing JoinEUI will cause the Join Server to discard the Join-Requests of the end-device. For each end-device, the Join Server keeps track of the last DevNonce value used by the end device and ignores Join-Requests if DevNonce is not incremented.
With the changes in https://github.com/zephyrproject-rtos/loramac-node/pull/9, the DevNonce is no longer a random value, but instead restarts from 0 on each reboot. This results in network servers expecting either v1.0.3 or v1.04 to reject join requests due to duplicate DevNonce values. For a v1.0.3 we need to wait for the network servers internal cache to be evicted. For v1.0.4 we need to send an additional join request each time the device reboots. The Nth reboot will require at least N join requests before a join has a chance of succeeding. In both cases, for a device compliant with the join request backoff recommendations, this can quickly take hours.
At the same time, the LoRaWAN subsystem does not expose a mechanism for storing and restoring the NVM context required for correct operation under v1.0.4. This is currently a work in progress in #40193
This issue was picked up on in #36571, but was not followed through to the logical conclusion that we are now unable to properly join any network server compliant with the spec.
**To Reproduce**
Build the `zephyr/samples/subsys/lorawan/class_a` sample, and attempt to join a LoRaWAN network multiple times without clearing anything on the network server.
**Expected behavior**
The LoRaWAN subsystem should be able to correctly join network servers, ideally either v1.0.3 or v1.0.4 versions.
**Impact**
LoRaWAN functionality is significantly degraded when devices reboot for any reason (Faults, OTA upgrades etc).
The degradation increases with each reboot.
**Proposed Solution**
Merging #40193 into the LTS branch is likely too large a change to be accepted. Providing a mechanism to query and restore the loramac-node crypto NVS state is sufficient to fix this issue for both network types. This can be done by hooking into the `NvmDataChange` callback without additional changes to loramac-node.
**Environment (please complete the following information):**
- v2.7.1 LTS | 1.0 | LoRaWAN: Unable to correctly join networks of any version on LTS - **Describe the bug**
The LoRaWAN subsystem is unable to correctly join networks of any LoRaWAN protocol version on the LTS branch.
LoRaWAN fork: https://github.com/zephyrproject-rtos/loramac-node
The relevant versions for this issue:
1. v1.0.3 (Previous version)
2. v1.0.4 (Current version targeted by the Zephyr fork)
3. v1.1.x (Not supported at all by the release that the Zephyr fork is based upon)
The PR https://github.com/zephyrproject-rtos/loramac-node/pull/9 updated the release the fork was based upon from some older version (v1.0.3?) to v1.0.4. Part of the specification changes from v1.0.3 to v1.0.4 was a change in requirements around the device nonce value, sent in the join request as a way of avoiding replay attacks.
*LoRaWAN v1.0.3 Specification (6.2.4)*
> DevNonce is a random value.1 For each end-device, the network server keeps track of a certain number of DevNonce values used by the end-device in the past, and ignores join requests with any of these DevNonce values from that end-device.
*LoRaWAN v1.0.4 Specification (6.2.5)*
> DevNonce is a counter starting at 0 when the end-device is initially powered up and incremented with every Join-Request. A DevNonce value SHALL never be reused for a given JoinEUI value. If the end-device can be power-cycled, then DevNonce SHALL be persistent (e.g., stored in a non-volatile memory). Resetting DevNonce without changing JoinEUI will cause the Join Server to discard the Join-Requests of the end-device. For each end-device, the Join Server keeps track of the last DevNonce value used by the end device and ignores Join-Requests if DevNonce is not incremented.
With the changes in https://github.com/zephyrproject-rtos/loramac-node/pull/9, the DevNonce is no longer a random value, but instead restarts from 0 on each reboot. This results in network servers expecting either v1.0.3 or v1.04 to reject join requests due to duplicate DevNonce values. For a v1.0.3 we need to wait for the network servers internal cache to be evicted. For v1.0.4 we need to send an additional join request each time the device reboots. The Nth reboot will require at least N join requests before a join has a chance of succeeding. In both cases, for a device compliant with the join request backoff recommendations, this can quickly take hours.
At the same time, the LoRaWAN subsystem does not expose a mechanism for storing and restoring the NVM context required for correct operation under v1.0.4. This is currently a work in progress in #40193
This issue was picked up on in #36571, but was not followed through to the logical conclusion that we are now unable to properly join any network server compliant with the spec.
**To Reproduce**
Build the `zephyr/samples/subsys/lorawan/class_a` sample, and attempt to join a LoRaWAN network multiple times without clearing anything on the network server.
**Expected behavior**
The LoRaWAN subsystem should be able to correctly join network servers, ideally either v1.0.3 or v1.0.4 versions.
**Impact**
LoRaWAN functionality is significantly degraded when devices reboot for any reason (Faults, OTA upgrades etc).
The degradation increases with each reboot.
**Proposed Solution**
Merging #40193 into the LTS branch is likely too large a change to be accepted. Providing a mechanism to query and restore the loramac-node crypto NVS state is sufficient to fix this issue for both network types. This can be done by hooking into the `NvmDataChange` callback without additional changes to loramac-node.
**Environment (please complete the following information):**
- v2.7.1 LTS | priority | lorawan unable to correctly join networks of any version on lts describe the bug the lorawan subsystem is unable to correctly join networks of any lorawan protocol version on the lts branch lorawan fork the relevant versions for this issue previous version current version targeted by the zephyr fork x not supported at all by the release that the zephyr fork is based upon the pr updated the release the fork was based upon from some older version to part of the specification changes from to was a change in requirements around the device nonce value sent in the join request as a way of avoiding replay attacks lorawan specification devnonce is a random value for each end device the network server keeps track of a certain number of devnonce values used by the end device in the past and ignores join requests with any of these devnonce values from that end device lorawan specification devnonce is a counter starting at when the end device is initially powered up and incremented with every join request a devnonce value shall never be reused for a given joineui value if the end device can be power cycled then devnonce shall be persistent e g stored in a non volatile memory resetting devnonce without changing joineui will cause the join server to discard the join requests of the end device for each end device the join server keeps track of the last devnonce value used by the end device and ignores join requests if devnonce is not incremented with the changes in the devnonce is no longer a random value but instead restarts from on each reboot this results in network servers expecting either or to reject join requests due to duplicate devnonce values for a we need to wait for the network servers internal cache to be evicted for we need to send an additional join request each time the device reboots the nth reboot will require at least n join requests before a join has a chance of succeeding in both cases for a device compliant with the join request backoff recommendations this can quickly take hours at the same time the lorawan subsystem does not expose a mechanism for storing and restoring the nvm context required for correct operation under this is currently a work in progress in this issue was picked up on in but was not followed through to the logical conclusion that we are now unable to properly join any network server compliant with the spec to reproduce build the zephyr samples subsys lorawan class a sample and attempt to join a lorawan network multiple times without clearing anything on the network server expected behavior the lorawan subsystem should be able to correctly join network servers ideally either or versions impact lorawan functionality is significantly degraded when devices reboot for any reason faults ota upgrades etc the degradation increases with each reboot proposed solution merging into the lts branch is likely too large a change to be accepted providing a mechanism to query and restore the loramac node crypto nvs state is sufficient to fix this issue for both network types this can be done by hooking into the nvmdatachange callback without additional changes to loramac node environment please complete the following information lts | 1 |
266,231 | 8,364,461,107 | IssuesEvent | 2018-10-03 23:09:46 | nelsonnyland/WorldwideMovieDatabase | https://api.github.com/repos/nelsonnyland/WorldwideMovieDatabase | opened | Create an example mockup of the home page. | good first issue medium priority | We should design an example sketch or diagram for what we want the home page to look like.
We could use Visio or just a normal drawing.
We could then upload it to Github and add it to our project in a resource folder.
We should make sure to label the diagram so we know what each piece and widget does.
| 1.0 | Create an example mockup of the home page. - We should design an example sketch or diagram for what we want the home page to look like.
We could use Visio or just a normal drawing.
We could then upload it to Github and add it to our project in a resource folder.
We should make sure to label the diagram so we know what each piece and widget does.
| priority | create an example mockup of the home page we should design an example sketch or diagram for what we want the home page to look like we could use visio or just a normal drawing we could then upload it to github and add it to our project in a resource folder we should make sure to label the diagram so we know what each piece and widget does | 1 |
634,424 | 20,361,023,107 | IssuesEvent | 2022-02-20 17:35:01 | moonstar-x/discord-tts-bot | https://api.github.com/repos/moonstar-x/discord-tts-bot | opened | Add support for TTS Tool - Microsoft and Amazon voices. | Priority: Medium Type: Feature Request | #### :zap: Describe the New Feature
The site https://ttstool.com/ exposes a public API to use with some voices from Microsoft and Amazon engines. I think maybe the Amazon ones would be duplicate as specified in #71, however this tool supports SSML and some options for volume, pitch, and rate.
To request this API, send a POST to:
```text
https://support.readaloud.app/ttstool/createParts
```
With a body shaped like this:
```text
[{
"voiceId": "Amazon Polish (Ewa)",
"ssml": "<speak version=\"1.0\" xml:lang=\"pl-PL\"><prosody volume='x-loud' rate='x-fast' pitch='x-high'>yeah</prosody></speak>"
}]
```
This returns a an JSON array with the IDs of the voices. You can technically pass in multiple items to the previous array to get multiple TTS streams in the same request. The response looks like this:
```text
[
"efh7l3ll"
]
```
You can then use this ID in a GET request like this:
```text
https://support.readaloud.app/ttstool/getParts?q=efh7l3ll
```
This returns an `audio/mpeg` stream. | 1.0 | Add support for TTS Tool - Microsoft and Amazon voices. - #### :zap: Describe the New Feature
The site https://ttstool.com/ exposes a public API to use with some voices from Microsoft and Amazon engines. I think maybe the Amazon ones would be duplicate as specified in #71, however this tool supports SSML and some options for volume, pitch, and rate.
To request this API, send a POST to:
```text
https://support.readaloud.app/ttstool/createParts
```
With a body shaped like this:
```text
[{
"voiceId": "Amazon Polish (Ewa)",
"ssml": "<speak version=\"1.0\" xml:lang=\"pl-PL\"><prosody volume='x-loud' rate='x-fast' pitch='x-high'>yeah</prosody></speak>"
}]
```
This returns a an JSON array with the IDs of the voices. You can technically pass in multiple items to the previous array to get multiple TTS streams in the same request. The response looks like this:
```text
[
"efh7l3ll"
]
```
You can then use this ID in a GET request like this:
```text
https://support.readaloud.app/ttstool/getParts?q=efh7l3ll
```
This returns an `audio/mpeg` stream. | priority | add support for tts tool microsoft and amazon voices zap describe the new feature the site exposes a public api to use with some voices from microsoft and amazon engines i think maybe the amazon ones would be duplicate as specified in however this tool supports ssml and some options for volume pitch and rate to request this api send a post to text with a body shaped like this text voiceid amazon polish ewa ssml yeah this returns a an json array with the ids of the voices you can technically pass in multiple items to the previous array to get multiple tts streams in the same request the response looks like this text you can then use this id in a get request like this text this returns an audio mpeg stream | 1 |
91,942 | 3,863,517,812 | IssuesEvent | 2016-04-08 09:45:57 | iamxavier/elmah | https://api.github.com/repos/iamxavier/elmah | closed | Support Logging to Azure Table Storage | auto-migrated Priority-Medium Type-Enhancement | ```
What new or enhanced feature are you proposing?
Add support for logging to Azure Table Storage
What goal would this enhancement help you achieve?
Ability to use ELMAH with Windows Azure + Table Storage.
```
Original issue reported on code.google.com by `pbsbac...@gmail.com` on 27 May 2009 at 1:30 | 1.0 | Support Logging to Azure Table Storage - ```
What new or enhanced feature are you proposing?
Add support for logging to Azure Table Storage
What goal would this enhancement help you achieve?
Ability to use ELMAH with Windows Azure + Table Storage.
```
Original issue reported on code.google.com by `pbsbac...@gmail.com` on 27 May 2009 at 1:30 | priority | support logging to azure table storage what new or enhanced feature are you proposing add support for logging to azure table storage what goal would this enhancement help you achieve ability to use elmah with windows azure table storage original issue reported on code google com by pbsbac gmail com on may at | 1 |
38,662 | 2,849,832,411 | IssuesEvent | 2015-05-31 01:49:36 | GiraffaFS/giraffa | https://api.github.com/repos/GiraffaFS/giraffa | closed | Release Giraffa 0.2.0 | Priority-Medium Type-Task | Release Steps:
* Move remaining issues from Milestone 0.2.0 to Milestone 0.3.0
* Update poms to 0.2.0
* Create release tag, named release_0.2.0
* Update poms to 0..3.0-SNAPSHOT
* Update Wiki documentation to include new release. Include Release notes generated from git messages
* Run Jenkins RELEASE build for new release option.
| 1.0 | Release Giraffa 0.2.0 - Release Steps:
* Move remaining issues from Milestone 0.2.0 to Milestone 0.3.0
* Update poms to 0.2.0
* Create release tag, named release_0.2.0
* Update poms to 0..3.0-SNAPSHOT
* Update Wiki documentation to include new release. Include Release notes generated from git messages
* Run Jenkins RELEASE build for new release option.
| priority | release giraffa release steps move remaining issues from milestone to milestone update poms to create release tag named release update poms to snapshot update wiki documentation to include new release include release notes generated from git messages run jenkins release build for new release option | 1 |
599,578 | 18,277,556,473 | IssuesEvent | 2021-10-04 20:49:22 | poissonconsulting/ypr | https://api.github.com/repos/poissonconsulting/ypr | opened | ypr_tabulate_yield add population name column to table | Effort: 2 Medium Priority: 4 Low Type: Enhancement Difficulty: 1 Simple | ## Your Idea
The table that is outputted by ypr_tabulate_yield.ypr_populations could have a column that adds the names of each population to the output of the table so you can easily tell which is which | 1.0 | ypr_tabulate_yield add population name column to table - ## Your Idea
The table that is outputted by ypr_tabulate_yield.ypr_populations could have a column that adds the names of each population to the output of the table so you can easily tell which is which | priority | ypr tabulate yield add population name column to table your idea the table that is outputted by ypr tabulate yield ypr populations could have a column that adds the names of each population to the output of the table so you can easily tell which is which | 1 |
805,570 | 29,577,045,728 | IssuesEvent | 2023-06-07 00:16:51 | TimeWalkOrg/Manhattan | https://api.github.com/repos/TimeWalkOrg/Manhattan | closed | Creek origins should be more natural | 2 Medium Priority | The ends of the creeks (where they appear to originate) look unnatural. The water is springing out of the rock like a powerful spring, though this is a small island.
Can we add a few smaller grooves uphill of the "origin" that are dry that appear to "feed" the creeks? Slowing down the water (per Issue #3 ) will help. Can we maybe cover the origins of the creeks with dead leaf and small rock cover of some kind?

| 1.0 | Creek origins should be more natural - The ends of the creeks (where they appear to originate) look unnatural. The water is springing out of the rock like a powerful spring, though this is a small island.
Can we add a few smaller grooves uphill of the "origin" that are dry that appear to "feed" the creeks? Slowing down the water (per Issue #3 ) will help. Can we maybe cover the origins of the creeks with dead leaf and small rock cover of some kind?

| priority | creek origins should be more natural the ends of the creeks where they appear to originate look unnatural the water is springing out of the rock like a powerful spring though this is a small island can we add a few smaller grooves uphill of the origin that are dry that appear to feed the creeks slowing down the water per issue will help can we maybe cover the origins of the creeks with dead leaf and small rock cover of some kind | 1 |
248,665 | 7,934,747,908 | IssuesEvent | 2018-07-08 23:00:17 | SETI/pds-opus | https://api.github.com/repos/SETI/pds-opus | opened | Sort on multiple fields | A-Enhancement B-OPUS Django B-OPUS JS Effort 2 Medium Priority TBD | Currently it's possible to sort on a single field by using the "view table" view and clicking on a column header. However, there are times you want to sort on multiple columns, since columns like "planet" or "mission" can have a large number of entries that are all identical. For example, you might want to sort on "Target Body" in ascending order, then on "Filter" in ascending order, then on "Observation Duration" in descending order.
The JPL Imaging Node "Image Atlas" allows this kind of searching. We should too.
| 1.0 | Sort on multiple fields - Currently it's possible to sort on a single field by using the "view table" view and clicking on a column header. However, there are times you want to sort on multiple columns, since columns like "planet" or "mission" can have a large number of entries that are all identical. For example, you might want to sort on "Target Body" in ascending order, then on "Filter" in ascending order, then on "Observation Duration" in descending order.
The JPL Imaging Node "Image Atlas" allows this kind of searching. We should too.
| priority | sort on multiple fields currently it s possible to sort on a single field by using the view table view and clicking on a column header however there are times you want to sort on multiple columns since columns like planet or mission can have a large number of entries that are all identical for example you might want to sort on target body in ascending order then on filter in ascending order then on observation duration in descending order the jpl imaging node image atlas allows this kind of searching we should too | 1 |
417,369 | 12,158,999,347 | IssuesEvent | 2020-04-26 07:04:25 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | opened | Check for repeating error events from pods in openshift-storage namespace | Medium Priority | This check can be done as a single test case or at the end of relevant test runs.
Relevant test runs means those runs which will not introduce any errors which will in turn trigger error events in pods. eg: tier1 runs.
This will also cover bugs like this - [1777591](https://bugzilla.redhat.com/show_bug.cgi?id=1777591) | 1.0 | Check for repeating error events from pods in openshift-storage namespace - This check can be done as a single test case or at the end of relevant test runs.
Relevant test runs means those runs which will not introduce any errors which will in turn trigger error events in pods. eg: tier1 runs.
This will also cover bugs like this - [1777591](https://bugzilla.redhat.com/show_bug.cgi?id=1777591) | priority | check for repeating error events from pods in openshift storage namespace this check can be done as a single test case or at the end of relevant test runs relevant test runs means those runs which will not introduce any errors which will in turn trigger error events in pods eg runs this will also cover bugs like this | 1 |
500,904 | 14,516,983,689 | IssuesEvent | 2020-12-13 17:53:14 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Some Installations of ui_next do not render anything in UI on MS Edge Classic and IE11 | component:ui_next priority:medium state:needs_devel type:bug | ##### ISSUE TYPE
- Bug Report
##### SUMMARY
In the bug bash, we set up a dev server to point to a remote tower api server. When a user of IE11 or Edge Classic attempts to load the page, they see nothing. No errors in the network or console logs.
This has a self signed cert pointed to a yolo'd tower with it's own cert.
This page loads in the chromium-based browsers and firefox.
##### ENVIRONMENT
* AWX version: 9.2.0
* AWX install method: ui_next local dev install on a remote ec2 target
* Operating System: Windows 10
* Web Browser: IE11 and Edge Classic (18)
##### STEPS TO REPRODUCE
1. Log in to web UI on IE11 or Edge Classic
##### EXPECTED RESULTS
Functional UI
##### ACTUAL RESULTS
Blank page. No console or network errors
| 1.0 | Some Installations of ui_next do not render anything in UI on MS Edge Classic and IE11 - ##### ISSUE TYPE
- Bug Report
##### SUMMARY
In the bug bash, we set up a dev server to point to a remote tower api server. When a user of IE11 or Edge Classic attempts to load the page, they see nothing. No errors in the network or console logs.
This has a self signed cert pointed to a yolo'd tower with it's own cert.
This page loads in the chromium-based browsers and firefox.
##### ENVIRONMENT
* AWX version: 9.2.0
* AWX install method: ui_next local dev install on a remote ec2 target
* Operating System: Windows 10
* Web Browser: IE11 and Edge Classic (18)
##### STEPS TO REPRODUCE
1. Log in to web UI on IE11 or Edge Classic
##### EXPECTED RESULTS
Functional UI
##### ACTUAL RESULTS
Blank page. No console or network errors
| priority | some installations of ui next do not render anything in ui on ms edge classic and issue type bug report summary in the bug bash we set up a dev server to point to a remote tower api server when a user of or edge classic attempts to load the page they see nothing no errors in the network or console logs this has a self signed cert pointed to a yolo d tower with it s own cert this page loads in the chromium based browsers and firefox environment awx version awx install method ui next local dev install on a remote target operating system windows web browser and edge classic steps to reproduce log in to web ui on or edge classic expected results functional ui actual results blank page no console or network errors | 1 |
652,791 | 21,561,960,306 | IssuesEvent | 2022-05-01 09:47:45 | xournalpp/xournalpp | https://api.github.com/repos/xournalpp/xournalpp | closed | Make colors configurable with only stylus input and option to add configured color bins to toolbar | enhancement priority::medium | I think additional configurable color slots, that keep the colour, after you have selected a standard color, would be nice. A colour wheel would also be interesting | 1.0 | Make colors configurable with only stylus input and option to add configured color bins to toolbar - I think additional configurable color slots, that keep the colour, after you have selected a standard color, would be nice. A colour wheel would also be interesting | priority | make colors configurable with only stylus input and option to add configured color bins to toolbar i think additional configurable color slots that keep the colour after you have selected a standard color would be nice a colour wheel would also be interesting | 1 |
230,706 | 7,613,018,389 | IssuesEvent | 2018-05-01 19:40:15 | fgpv-vpgf/fgpv-vpgf | https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf | closed | Legend toggling and data table viewing issue with attributes that contain apostrophes (single quotes) | bug-type: regression priority: medium problem: bug | Layers displayed with queries on stings that contain apostrophes (single quotes) throw an error.

`http://geoappext.nrcan.gc.ca/arcgis/rest/services/MMS/IMA_F/MapServer/0/query?f=json&where=(TYPE_ENTENTE_NOM%20%3D%20'Protocole%20d'entente')&returnGeometry=true&spatialRel=esriSpatialRelIntersects&geometry=%7B%22xmin%22%3A-4074491.517018034%2C%22ymin%22%3A-192876.4845059691%2C%22xmax%22%3A-10483.389001777861%2C%22ymax%22%3A3871131.643510287%2C%22spatialReference%22%3A%7B%22wkid%22%3A3978%7D%7D&geometryType=esriGeometryEnvelope&inSR=3978&outFields=OBJECTID%2CTYPE_ENTENTE_NOM%2CPROJECT_NAME_FR&outSR=3978`
`{"error":{"code":400,"message":"Unable to complete operation.","details":[]}}`
| 1.0 | Legend toggling and data table viewing issue with attributes that contain apostrophes (single quotes) - Layers displayed with queries on stings that contain apostrophes (single quotes) throw an error.

`http://geoappext.nrcan.gc.ca/arcgis/rest/services/MMS/IMA_F/MapServer/0/query?f=json&where=(TYPE_ENTENTE_NOM%20%3D%20'Protocole%20d'entente')&returnGeometry=true&spatialRel=esriSpatialRelIntersects&geometry=%7B%22xmin%22%3A-4074491.517018034%2C%22ymin%22%3A-192876.4845059691%2C%22xmax%22%3A-10483.389001777861%2C%22ymax%22%3A3871131.643510287%2C%22spatialReference%22%3A%7B%22wkid%22%3A3978%7D%7D&geometryType=esriGeometryEnvelope&inSR=3978&outFields=OBJECTID%2CTYPE_ENTENTE_NOM%2CPROJECT_NAME_FR&outSR=3978`
`{"error":{"code":400,"message":"Unable to complete operation.","details":[]}}`
| priority | legend toggling and data table viewing issue with attributes that contain apostrophes single quotes layers displayed with queries on stings that contain apostrophes single quotes throw an error error code message unable to complete operation details | 1 |
807,897 | 30,023,443,319 | IssuesEvent | 2023-06-27 02:49:15 | AnOpenSauceDev/Methane-mod | https://api.github.com/repos/AnOpenSauceDev/Methane-mod | closed | [Issue] 1.20 is broken | enhancement compat Methane-Next priority:medium | ## Description of what happened
methane worked fine till 23w14a (by editing fabric.mod.json), but it broke for 23w16a
## Mods used
methane 1.7
cloth config 10.0.96
fabric api for 23w16a
## possible ways to replicate this bug
join a server; then it crashes
maybe u already want to update methane to 1.20
might be an issue with MatrixStack screen. not sure tho
it seems to be a fundamental change, because all 1.19.4 mods that worked fine for 23w14a broke for 23w16a (game crashes while launching/joining server)
## crashreport
---- Minecraft Crash Report ----
// Shall we play a game?
Time: 2023-04-21 01:03:34
Description: Unexpected error
java.lang.AbstractMethodError: Receiver class me.wolfie.methane.client.HudRenderListener does not define or inherit an implementation of the resolved method 'abstract void onHudRender(net.minecraft.class_332, float)' of interface net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.
at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27)
at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393)
at net.minecraft.class_329.method_1753(class_329.java:371)
at net.minecraft.class_757.method_3192(class_757.java:918)
at net.minecraft.class_310.method_1523(class_310.java:1203)
at net.minecraft.class_310.method_1514(class_310.java:786)
at net.minecraft.client.main.Main.main(Main.java:240)
at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462)
at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74)
at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23)
A detailed walkthrough of the error, its code path and all known details is as follows:
---------------------------------------------------------------------------------------
-- Head --
Thread: Render thread
Stacktrace:
at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27)
at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393)
at net.minecraft.class_329.method_1753(class_329.java:371)
-- Affected level --
Details:
All players: 1 total; [class_746['Cyclopropinon'/2857, l='ClientLevel', x=192.87, y=133.00, z=-246.02]]
Chunk stats: 729, 248
Level dimension: minecraft:the_nether
Level spawn location: World: (0,73,0), Section: (at 0,9,0 in 0,4,0; chunk contains blocks 0,0,0 to 15,255,15), Region: (0,0; contains chunks 0,0 to 31,31, blocks 0,0,0 to 511,255,511)
Level time: 119784655 game time, 129434541 day time
Server brand: fabric
Server type: Non-integrated multiplayer server
Stacktrace:
at net.minecraft.class_638.method_8538(class_638.java:455)
at net.minecraft.class_310.method_1587(class_310.java:2394)
at net.minecraft.class_310.method_1514(class_310.java:810)
at net.minecraft.client.main.Main.main(Main.java:240)
at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462)
at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74)
at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23)
-- Last reload --
Details:
Reload number: 1
Reload reason: initial
Finished: Yes
Packs: vanilla, Fabric Mods
-- System Details --
Details:
Minecraft Version: 23w16a
Minecraft Version ID: 23w16a
Operating System: Windows 10 (amd64) version 10.0
Java Version: 17.0.3, Microsoft
Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Microsoft
Memory: 359664912 bytes (343 MiB) / 671088640 bytes (640 MiB) up to 2147483648 bytes (2048 MiB)
CPUs: 4
Processor Vendor: GenuineIntel
Processor Name: Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz
Identifier: Intel64 Family 6 Model 37 Stepping 5
Microarchitecture: Westmere (Client)
Frequency (GHz): 2.66
Number of physical packages: 1
Number of physical CPUs: 2
Number of logical CPUs: 4
Graphics card #0 name: NVIDIA NVS 3100M
Graphics card #0 vendor: NVIDIA (0x10de)
Graphics card #0 VRAM (MB): 512.00
Graphics card #0 deviceId: 0x0a6c
Graphics card #0 versionInfo: DriverVersion=21.21.13.4201
Memory slot #0 capacity (MB): 4096.00
Memory slot #0 clockSpeed (GHz): 1.33
Memory slot #0 type: DDR3
Virtual memory max (MB): 15219.67
Virtual memory used (MB): 10379.44
Swap memory total (MB): 11264.00
Swap memory used (MB): 2109.98
JVM Flags: 9 total; -XX:HeapDumpPath=MojangTricksIntelDriversForPerformance_javaw.exe_minecraft.exe.heapdump -Xss1M -Xmx2G -XX:+UnlockExperimentalVMOptions -XX:+UseG1GC -XX:G1NewSizePercent=20 -XX:G1ReservePercent=20 -XX:MaxGCPauseMillis=50 -XX:G1HeapRegionSize=32M
Fabric Mods:
cloth-config: Cloth Config v10 10.0.96
cloth-basic-math: cloth-basic-math 0.6.1
fabric-api: Fabric API 0.78.1+1.20
fabric-api-base: Fabric API Base 0.4.23+9ff28bce67
fabric-api-lookup-api-v1: Fabric API Lookup API (v1) 1.6.26+eff2638667
fabric-biome-api-v1: Fabric Biome API (v1) 13.0.6+348a9c6467
fabric-block-api-v1: Fabric Block API (v1) 1.0.5+e022e5d167
fabric-blockrenderlayer-v1: Fabric BlockRenderLayer Registration (v1) 1.1.33+c2e6f67467
fabric-client-tags-api-v1: Fabric Client Tags 1.0.14+1134c5b867
fabric-command-api-v1: Fabric Command API (v1) 1.2.26+f71b366f67
fabric-command-api-v2: Fabric Command API (v2) 2.2.5+df5b2a9d67
fabric-commands-v0: Fabric Commands (v0) 0.2.43+df3654b367
fabric-containers-v0: Fabric Containers (v0) 0.1.55+df3654b367
fabric-content-registries-v0: Fabric Content Registries (v0) 4.0.0+eff2638667
fabric-convention-tags-v1: Fabric Convention Tags 1.4.0+9a7c5daa67
fabric-crash-report-info-v1: Fabric Crash Report Info (v1) 0.2.14+aeb40ebe67
fabric-data-generation-api-v1: Fabric Data Generation API (v1) 12.0.0+eff2638667
fabric-dimensions-v1: Fabric Dimensions API (v1) 2.1.44+7f87f8fa67
fabric-entity-events-v1: Fabric Entity Events (v1) 1.5.14+eff2638667
fabric-events-interaction-v0: Fabric Events Interaction (v0) 0.4.42+a1ccd7bf67
fabric-events-lifecycle-v0: Fabric Events Lifecycle (v0) 0.2.53+df3654b367
fabric-game-rule-api-v1: Fabric Game Rule API (v1) 1.0.33+eff2638667
fabric-item-api-v1: Fabric Item API (v1) 2.1.18+eff2638667
fabric-item-group-api-v1: Fabric Item Group API (v1) 4.0.0+eff2638667
fabric-key-binding-api-v1: Fabric Key Binding API (v1) 1.0.32+c477957e67
fabric-keybindings-v0: Fabric Key Bindings (v0) 0.2.30+df3654b367
fabric-lifecycle-events-v1: Fabric Lifecycle Events (v1) 2.2.14+5da15ca167
fabric-loot-api-v2: Fabric Loot API (v2) 1.1.29+eff2638667
fabric-loot-tables-v1: Fabric Loot Tables (v1) 1.1.33+9e7660c667
fabric-message-api-v1: Fabric Message API (v1) 5.1.0+1ee8be4067
fabric-mining-level-api-v1: Fabric Mining Level API (v1) 2.1.39+eff2638667
fabric-models-v0: Fabric Models (v0) 0.3.29+11ba9c3b67
fabric-networking-api-v1: Fabric Networking API (v1) 1.3.2+eff2638667
fabric-networking-v0: Fabric Networking (v0) 0.3.42+df3654b367
fabric-object-builder-api-v1: Fabric Object Builder API (v1) 10.0.0+eff2638667
fabric-particles-v1: Fabric Particles (v1) 1.0.22+f1e4495b67
fabric-recipe-api-v1: Fabric Recipe API (v1) 1.0.9+a1ccd7bf67
fabric-registry-sync-v0: Fabric Registry Sync (v0) 2.1.5+eff2638667
fabric-renderer-api-v1: Fabric Renderer API (v1) 2.2.5+eff2638667
fabric-renderer-indigo: Fabric Renderer - Indigo 1.1.1+81e8c57667
fabric-renderer-registries-v1: Fabric Renderer Registries (v1) 3.2.38+df3654b367
fabric-rendering-data-attachment-v1: Fabric Rendering Data Attachment (v1) 0.3.27+afca2f3e67
fabric-rendering-fluids-v1: Fabric Rendering Fluids (v1) 3.0.20+f1e4495b67
fabric-rendering-v0: Fabric Rendering (v0) 1.1.41+df3654b367
fabric-rendering-v1: Fabric Rendering (v1) 3.0.0+eff2638667
fabric-resource-conditions-api-v1: Fabric Resource Conditions API (v1) 2.3.0+e6c7d4ee67
fabric-resource-loader-v0: Fabric Resource Loader (v0) 0.11.1+03ffe37867
fabric-screen-api-v1: Fabric Screen API (v1) 2.0.0+eff2638667
fabric-screen-handler-api-v1: Fabric Screen Handler API (v1) 1.3.21+eff2638667
fabric-sound-api-v1: Fabric Sound API (v1) 1.0.8+75e9821167
fabric-transfer-api-v1: Fabric Transfer API (v1) 3.1.1+eff2638667
fabric-transitive-access-wideners-v1: Fabric Transitive Access Wideners (v1) 4.0.1+848ffaab67
fabricloader: Fabric Loader 0.14.19
java: OpenJDK 64-Bit Server VM 17
methane: Methane 1.7
minecraft: Minecraft 1.20-alpha.23.16.a
Launched Version: fabric-loader-0.14.19-23w16a
Backend library: LWJGL version 3.3.1 SNAPSHOT
Backend API: NVS 3100M/PCIe/SSE2 GL version 3.2.0, NVIDIA Corporation
Window size: 1440x837
GL Caps: Using framebuffer using OpenGL 3.2
GL debug messages:
Using VBOs: Yes
Is Modded: Definitely; Client brand changed to 'fabric'
Type: Client (map_client.txt)
Graphics mode: fancy
Resource Packs: fabric
Current Language: en_us
CPU: 4x Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz | 1.0 | [Issue] 1.20 is broken - ## Description of what happened
methane worked fine till 23w14a (by editing fabric.mod.json), but it broke for 23w16a
## Mods used
methane 1.7
cloth config 10.0.96
fabric api for 23w16a
## possible ways to replicate this bug
join a server; then it crashes
maybe u already want to update methane to 1.20
might be an issue with MatrixStack screen. not sure tho
it seems to be a fundamental change, because all 1.19.4 mods that worked fine for 23w14a broke for 23w16a (game crashes while launching/joining server)
## crashreport
---- Minecraft Crash Report ----
// Shall we play a game?
Time: 2023-04-21 01:03:34
Description: Unexpected error
java.lang.AbstractMethodError: Receiver class me.wolfie.methane.client.HudRenderListener does not define or inherit an implementation of the resolved method 'abstract void onHudRender(net.minecraft.class_332, float)' of interface net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.
at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27)
at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393)
at net.minecraft.class_329.method_1753(class_329.java:371)
at net.minecraft.class_757.method_3192(class_757.java:918)
at net.minecraft.class_310.method_1523(class_310.java:1203)
at net.minecraft.class_310.method_1514(class_310.java:786)
at net.minecraft.client.main.Main.main(Main.java:240)
at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462)
at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74)
at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23)
A detailed walkthrough of the error, its code path and all known details is as follows:
---------------------------------------------------------------------------------------
-- Head --
Thread: Render thread
Stacktrace:
at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27)
at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393)
at net.minecraft.class_329.method_1753(class_329.java:371)
-- Affected level --
Details:
All players: 1 total; [class_746['Cyclopropinon'/2857, l='ClientLevel', x=192.87, y=133.00, z=-246.02]]
Chunk stats: 729, 248
Level dimension: minecraft:the_nether
Level spawn location: World: (0,73,0), Section: (at 0,9,0 in 0,4,0; chunk contains blocks 0,0,0 to 15,255,15), Region: (0,0; contains chunks 0,0 to 31,31, blocks 0,0,0 to 511,255,511)
Level time: 119784655 game time, 129434541 day time
Server brand: fabric
Server type: Non-integrated multiplayer server
Stacktrace:
at net.minecraft.class_638.method_8538(class_638.java:455)
at net.minecraft.class_310.method_1587(class_310.java:2394)
at net.minecraft.class_310.method_1514(class_310.java:810)
at net.minecraft.client.main.Main.main(Main.java:240)
at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462)
at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74)
at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23)
-- Last reload --
Details:
Reload number: 1
Reload reason: initial
Finished: Yes
Packs: vanilla, Fabric Mods
-- System Details --
Details:
Minecraft Version: 23w16a
Minecraft Version ID: 23w16a
Operating System: Windows 10 (amd64) version 10.0
Java Version: 17.0.3, Microsoft
Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Microsoft
Memory: 359664912 bytes (343 MiB) / 671088640 bytes (640 MiB) up to 2147483648 bytes (2048 MiB)
CPUs: 4
Processor Vendor: GenuineIntel
Processor Name: Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz
Identifier: Intel64 Family 6 Model 37 Stepping 5
Microarchitecture: Westmere (Client)
Frequency (GHz): 2.66
Number of physical packages: 1
Number of physical CPUs: 2
Number of logical CPUs: 4
Graphics card #0 name: NVIDIA NVS 3100M
Graphics card #0 vendor: NVIDIA (0x10de)
Graphics card #0 VRAM (MB): 512.00
Graphics card #0 deviceId: 0x0a6c
Graphics card #0 versionInfo: DriverVersion=21.21.13.4201
Memory slot #0 capacity (MB): 4096.00
Memory slot #0 clockSpeed (GHz): 1.33
Memory slot #0 type: DDR3
Virtual memory max (MB): 15219.67
Virtual memory used (MB): 10379.44
Swap memory total (MB): 11264.00
Swap memory used (MB): 2109.98
JVM Flags: 9 total; -XX:HeapDumpPath=MojangTricksIntelDriversForPerformance_javaw.exe_minecraft.exe.heapdump -Xss1M -Xmx2G -XX:+UnlockExperimentalVMOptions -XX:+UseG1GC -XX:G1NewSizePercent=20 -XX:G1ReservePercent=20 -XX:MaxGCPauseMillis=50 -XX:G1HeapRegionSize=32M
Fabric Mods:
cloth-config: Cloth Config v10 10.0.96
cloth-basic-math: cloth-basic-math 0.6.1
fabric-api: Fabric API 0.78.1+1.20
fabric-api-base: Fabric API Base 0.4.23+9ff28bce67
fabric-api-lookup-api-v1: Fabric API Lookup API (v1) 1.6.26+eff2638667
fabric-biome-api-v1: Fabric Biome API (v1) 13.0.6+348a9c6467
fabric-block-api-v1: Fabric Block API (v1) 1.0.5+e022e5d167
fabric-blockrenderlayer-v1: Fabric BlockRenderLayer Registration (v1) 1.1.33+c2e6f67467
fabric-client-tags-api-v1: Fabric Client Tags 1.0.14+1134c5b867
fabric-command-api-v1: Fabric Command API (v1) 1.2.26+f71b366f67
fabric-command-api-v2: Fabric Command API (v2) 2.2.5+df5b2a9d67
fabric-commands-v0: Fabric Commands (v0) 0.2.43+df3654b367
fabric-containers-v0: Fabric Containers (v0) 0.1.55+df3654b367
fabric-content-registries-v0: Fabric Content Registries (v0) 4.0.0+eff2638667
fabric-convention-tags-v1: Fabric Convention Tags 1.4.0+9a7c5daa67
fabric-crash-report-info-v1: Fabric Crash Report Info (v1) 0.2.14+aeb40ebe67
fabric-data-generation-api-v1: Fabric Data Generation API (v1) 12.0.0+eff2638667
fabric-dimensions-v1: Fabric Dimensions API (v1) 2.1.44+7f87f8fa67
fabric-entity-events-v1: Fabric Entity Events (v1) 1.5.14+eff2638667
fabric-events-interaction-v0: Fabric Events Interaction (v0) 0.4.42+a1ccd7bf67
fabric-events-lifecycle-v0: Fabric Events Lifecycle (v0) 0.2.53+df3654b367
fabric-game-rule-api-v1: Fabric Game Rule API (v1) 1.0.33+eff2638667
fabric-item-api-v1: Fabric Item API (v1) 2.1.18+eff2638667
fabric-item-group-api-v1: Fabric Item Group API (v1) 4.0.0+eff2638667
fabric-key-binding-api-v1: Fabric Key Binding API (v1) 1.0.32+c477957e67
fabric-keybindings-v0: Fabric Key Bindings (v0) 0.2.30+df3654b367
fabric-lifecycle-events-v1: Fabric Lifecycle Events (v1) 2.2.14+5da15ca167
fabric-loot-api-v2: Fabric Loot API (v2) 1.1.29+eff2638667
fabric-loot-tables-v1: Fabric Loot Tables (v1) 1.1.33+9e7660c667
fabric-message-api-v1: Fabric Message API (v1) 5.1.0+1ee8be4067
fabric-mining-level-api-v1: Fabric Mining Level API (v1) 2.1.39+eff2638667
fabric-models-v0: Fabric Models (v0) 0.3.29+11ba9c3b67
fabric-networking-api-v1: Fabric Networking API (v1) 1.3.2+eff2638667
fabric-networking-v0: Fabric Networking (v0) 0.3.42+df3654b367
fabric-object-builder-api-v1: Fabric Object Builder API (v1) 10.0.0+eff2638667
fabric-particles-v1: Fabric Particles (v1) 1.0.22+f1e4495b67
fabric-recipe-api-v1: Fabric Recipe API (v1) 1.0.9+a1ccd7bf67
fabric-registry-sync-v0: Fabric Registry Sync (v0) 2.1.5+eff2638667
fabric-renderer-api-v1: Fabric Renderer API (v1) 2.2.5+eff2638667
fabric-renderer-indigo: Fabric Renderer - Indigo 1.1.1+81e8c57667
fabric-renderer-registries-v1: Fabric Renderer Registries (v1) 3.2.38+df3654b367
fabric-rendering-data-attachment-v1: Fabric Rendering Data Attachment (v1) 0.3.27+afca2f3e67
fabric-rendering-fluids-v1: Fabric Rendering Fluids (v1) 3.0.20+f1e4495b67
fabric-rendering-v0: Fabric Rendering (v0) 1.1.41+df3654b367
fabric-rendering-v1: Fabric Rendering (v1) 3.0.0+eff2638667
fabric-resource-conditions-api-v1: Fabric Resource Conditions API (v1) 2.3.0+e6c7d4ee67
fabric-resource-loader-v0: Fabric Resource Loader (v0) 0.11.1+03ffe37867
fabric-screen-api-v1: Fabric Screen API (v1) 2.0.0+eff2638667
fabric-screen-handler-api-v1: Fabric Screen Handler API (v1) 1.3.21+eff2638667
fabric-sound-api-v1: Fabric Sound API (v1) 1.0.8+75e9821167
fabric-transfer-api-v1: Fabric Transfer API (v1) 3.1.1+eff2638667
fabric-transitive-access-wideners-v1: Fabric Transitive Access Wideners (v1) 4.0.1+848ffaab67
fabricloader: Fabric Loader 0.14.19
java: OpenJDK 64-Bit Server VM 17
methane: Methane 1.7
minecraft: Minecraft 1.20-alpha.23.16.a
Launched Version: fabric-loader-0.14.19-23w16a
Backend library: LWJGL version 3.3.1 SNAPSHOT
Backend API: NVS 3100M/PCIe/SSE2 GL version 3.2.0, NVIDIA Corporation
Window size: 1440x837
GL Caps: Using framebuffer using OpenGL 3.2
GL debug messages:
Using VBOs: Yes
Is Modded: Definitely; Client brand changed to 'fabric'
Type: Client (map_client.txt)
Graphics mode: fancy
Resource Packs: fabric
Current Language: en_us
CPU: 4x Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz | priority | is broken description of what happened methane worked fine till by editing fabric mod json but it broke for mods used methane cloth config fabric api for possible ways to replicate this bug join a server then it crashes maybe u already want to update methane to might be an issue with matrixstack screen not sure tho it seems to be a fundamental change because all mods that worked fine for broke for game crashes while launching joining server crashreport minecraft crash report shall we play a game time description unexpected error java lang abstractmethoderror receiver class me wolfie methane client hudrenderlistener does not define or inherit an implementation of the resolved method abstract void onhudrender net minecraft class float of interface net fabricmc fabric api client rendering hudrendercallback at net fabricmc fabric api client rendering hudrendercallback lambda static hudrendercallback java at net minecraft class handler fabric rendering render class java at net minecraft class method class java at net minecraft class method class java at net minecraft class method class java at net minecraft class method class java at net minecraft client main main main main java at net fabricmc loader impl game minecraft minecraftgameprovider launch minecraftgameprovider java at net fabricmc loader impl launch knot knot launch knot java at net fabricmc loader impl launch knot knotclient main knotclient java a detailed walkthrough of the error its code path and all known details is as follows head thread render thread stacktrace at net fabricmc fabric api client rendering hudrendercallback lambda static hudrendercallback java at net minecraft class handler fabric rendering render class java at net minecraft class method class java affected level details all players total chunk stats level dimension minecraft the nether level spawn location world section at in chunk contains blocks to region contains chunks to blocks to level time game time day time server brand fabric server type non integrated multiplayer server stacktrace at net minecraft class method class java at net minecraft class method class java at net minecraft class method class java at net minecraft client main main main main java at net fabricmc loader impl game minecraft minecraftgameprovider launch minecraftgameprovider java at net fabricmc loader impl launch knot knot launch knot java at net fabricmc loader impl launch knot knotclient main knotclient java last reload details reload number reload reason initial finished yes packs vanilla fabric mods system details details minecraft version minecraft version id operating system windows version java version microsoft java vm version openjdk bit server vm mixed mode microsoft memory bytes mib bytes mib up to bytes mib cpus processor vendor genuineintel processor name intel r core tm cpu m identifier family model stepping microarchitecture westmere client frequency ghz number of physical packages number of physical cpus number of logical cpus graphics card name nvidia nvs graphics card vendor nvidia graphics card vram mb graphics card deviceid graphics card versioninfo driverversion memory slot capacity mb memory slot clockspeed ghz memory slot type virtual memory max mb virtual memory used mb swap memory total mb swap memory used mb jvm flags total xx heapdumppath mojangtricksinteldriversforperformance javaw exe minecraft exe heapdump xx unlockexperimentalvmoptions xx xx xx xx maxgcpausemillis xx fabric mods cloth config cloth config cloth basic math cloth basic math fabric api fabric api fabric api base fabric api base fabric api lookup api fabric api lookup api fabric biome api fabric biome api fabric block api fabric block api fabric blockrenderlayer fabric blockrenderlayer registration fabric client tags api fabric client tags fabric command api fabric command api fabric command api fabric command api fabric commands fabric commands fabric containers fabric containers fabric content registries fabric content registries fabric convention tags fabric convention tags fabric crash report info fabric crash report info fabric data generation api fabric data generation api fabric dimensions fabric dimensions api fabric entity events fabric entity events fabric events interaction fabric events interaction fabric events lifecycle fabric events lifecycle fabric game rule api fabric game rule api fabric item api fabric item api fabric item group api fabric item group api fabric key binding api fabric key binding api fabric keybindings fabric key bindings fabric lifecycle events fabric lifecycle events fabric loot api fabric loot api fabric loot tables fabric loot tables fabric message api fabric message api fabric mining level api fabric mining level api fabric models fabric models fabric networking api fabric networking api fabric networking fabric networking fabric object builder api fabric object builder api fabric particles fabric particles fabric recipe api fabric recipe api fabric registry sync fabric registry sync fabric renderer api fabric renderer api fabric renderer indigo fabric renderer indigo fabric renderer registries fabric renderer registries fabric rendering data attachment fabric rendering data attachment fabric rendering fluids fabric rendering fluids fabric rendering fabric rendering fabric rendering fabric rendering fabric resource conditions api fabric resource conditions api fabric resource loader fabric resource loader fabric screen api fabric screen api fabric screen handler api fabric screen handler api fabric sound api fabric sound api fabric transfer api fabric transfer api fabric transitive access wideners fabric transitive access wideners fabricloader fabric loader java openjdk bit server vm methane methane minecraft minecraft alpha a launched version fabric loader backend library lwjgl version snapshot backend api nvs pcie gl version nvidia corporation window size gl caps using framebuffer using opengl gl debug messages using vbos yes is modded definitely client brand changed to fabric type client map client txt graphics mode fancy resource packs fabric current language en us cpu intel r core tm cpu m | 1 |
247,842 | 7,924,251,041 | IssuesEvent | 2018-07-05 16:19:01 | uracreative/task-management | https://api.github.com/repos/uracreative/task-management | closed | OTF report June 2018 | Internal: Operations Internal: Overdue Priority: Medium | Documentation for monthly report of activities of Ura Design as part of OTF. Please add here on bullet points every time we have done tasks related to OTF.
Deadline: 26.06.2018. | 1.0 | OTF report June 2018 - Documentation for monthly report of activities of Ura Design as part of OTF. Please add here on bullet points every time we have done tasks related to OTF.
Deadline: 26.06.2018. | priority | otf report june documentation for monthly report of activities of ura design as part of otf please add here on bullet points every time we have done tasks related to otf deadline | 1 |
501,248 | 14,524,178,703 | IssuesEvent | 2020-12-14 11:05:05 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.2 staging-1872] Industrial elecator collider problem | Category: Gameplay Priority: Medium | This part doesn't have collider:

Therefore, if the block under the elevator does not have a square edge, you will fall by car:


It wiill be harder but you can fall with square edgge too:

I think that at least visually we have a gap, but it should be removed:

| 1.0 | [0.9.2 staging-1872] Industrial elecator collider problem - This part doesn't have collider:

Therefore, if the block under the elevator does not have a square edge, you will fall by car:


It wiill be harder but you can fall with square edgge too:

I think that at least visually we have a gap, but it should be removed:

| priority | industrial elecator collider problem this part doesn t have collider therefore if the block under the elevator does not have a square edge you will fall by car it wiill be harder but you can fall with square edgge too i think that at least visually we have a gap but it should be removed | 1 |
54,999 | 3,071,798,917 | IssuesEvent | 2015-08-19 14:05:04 | RobotiumTech/robotium | https://api.github.com/repos/RobotiumTech/robotium | closed | ClickonButton fails on buttons only with background images | bug imported Priority-Medium | _From [yair.oha...@gmail.com](https://code.google.com/u/103769070358968325894/) on February 09, 2011 13:58:24_
What steps will reproduce the problem? 1.open dialog with buttons which have no text and backround specifics.
2.execute solo.clickonbutton(index) 3. What is the expected output? What do you see instead? I expect to see a click - nothing happens. What version of the product are you using? On what operating system? version 2.1 on Win 7 ,64 bit Please provide any additional information below.
**Attachment:** [servings_edit_dialog.xml](http://code.google.com/p/robotium/issues/detail?id=66)
_Original issue: http://code.google.com/p/robotium/issues/detail?id=66_ | 1.0 | ClickonButton fails on buttons only with background images - _From [yair.oha...@gmail.com](https://code.google.com/u/103769070358968325894/) on February 09, 2011 13:58:24_
What steps will reproduce the problem? 1.open dialog with buttons which have no text and backround specifics.
2.execute solo.clickonbutton(index) 3. What is the expected output? What do you see instead? I expect to see a click - nothing happens. What version of the product are you using? On what operating system? version 2.1 on Win 7 ,64 bit Please provide any additional information below.
**Attachment:** [servings_edit_dialog.xml](http://code.google.com/p/robotium/issues/detail?id=66)
_Original issue: http://code.google.com/p/robotium/issues/detail?id=66_ | priority | clickonbutton fails on buttons only with background images from on february what steps will reproduce the problem open dialog with buttons which have no text and backround specifics execute solo clickonbutton index what is the expected output what do you see instead i expect to see a click nothing happens what version of the product are you using on what operating system version on win bit please provide any additional information below attachment original issue | 1 |
684,886 | 23,436,680,018 | IssuesEvent | 2022-08-15 10:38:50 | conan-io/conan | https://api.github.com/repos/conan-io/conan | closed | [feature] Access to license and homepage for dependencies in conanfile.py | type: feature priority: medium complex: low | For some automated license report we would need to have access to the attributes `description`, `license` and `homepage` which can be set in the conan recipe.
As self.deps_cpp_info, self.deps_env_info or self.deps_user_info will be removed in conan 2.0, this should probably be added to [self.dependencies](https://docs.conan.io/en/latest/reference/conanfile/dependencies.html#dependencies-interface).
> Do not use self.deps_cpp_info, self.deps_env_info or self.deps_user_info, these have been removed in 2.0.
During this change it would be great to just have access to all attributes for the dependencies
| 1.0 | [feature] Access to license and homepage for dependencies in conanfile.py - For some automated license report we would need to have access to the attributes `description`, `license` and `homepage` which can be set in the conan recipe.
As self.deps_cpp_info, self.deps_env_info or self.deps_user_info will be removed in conan 2.0, this should probably be added to [self.dependencies](https://docs.conan.io/en/latest/reference/conanfile/dependencies.html#dependencies-interface).
> Do not use self.deps_cpp_info, self.deps_env_info or self.deps_user_info, these have been removed in 2.0.
During this change it would be great to just have access to all attributes for the dependencies
| priority | access to license and homepage for dependencies in conanfile py for some automated license report we would need to have access to the attributes description license and homepage which can be set in the conan recipe as self deps cpp info self deps env info or self deps user info will be removed in conan this should probably be added to do not use self deps cpp info self deps env info or self deps user info these have been removed in during this change it would be great to just have access to all attributes for the dependencies | 1 |
682,857 | 23,359,948,951 | IssuesEvent | 2022-08-10 10:46:25 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Bluetooth controller extended advertisement crashes in lll layer | bug priority: medium area: Bluetooth area: Bluetooth Controller | Zephyr version:
NCS 2.0.0 - Zephyr tag v3.0.99-ncs1
I'm running and Bluetooth host and controller combination of nRF9160(Host) and nRF52833(Controller) using the BT_LL_SW_SPLIT variant. The application is very complicated so i cannot provide a reproducible sample. I have managed to capture an stack trace using Memfault. The problem occurs rather random once in a few hours.

[lll.c](https://github.com/nrfconnect/sdk-zephyr/blob/v3.0.99-ncs1/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c)
Note that the trace is capture with disabled CONFIG_BT_ASSERT=n on the controller.
We had issues with advertisement raising radio tx not ready. In production we disabled the assert to remove unnecessary restarts with cost of some advertisements not getting transmitted.
I can try to catch some debug outputs using the BT_ASSERT, however the assert string is usually lost due to LOG_DEFFER. To catch the issue with memfault is also possibility with the BT_ASSERT but requires more work.
Afterwards i tried to optimize IRQ times by these settings(replaced Zero Latency Interrupts) and haven't tested the assert variant ever since:
```
CONFIG_BT_CTLR_ULL_HIGH_PRIO=1
CONFIG_BT_CTLR_ULL_LOW_PRIO=2
```
Basic description of the application.
It broadcasts a large amount of advertisements sets
6 sets of advertisements using BT legacy(1 connectable).
6 sets of advertisements using BT long range (Coded phy).
Interval of all advertisements are 4HZ.
Application performs regular ADV_DATA updates for the advertisements with interval 1Hz-4Hz.
Communication between Controller and Host is implemented by custom layer. I can provide HCI traces using RTT BT debug if necessary. However, we are using same layer to implement other communication between the chips and we had no problems with it. It is basically H5(Confirmed messages, and retransmission) over multiplexed UART with Flow Control. The controller and host does not seem to generate any error logs.
Configuration of the host(BT part):
```
# Enable Bluetooth stack and libraries
CONFIG_BT=y
CONFIG_BT_PERIPHERAL=y
CONFIG_BT_NO_DRIVER=y
CONFIG_BT_H4=n
CONFIG_BT_SMP=y
CONFIG_BT_KEYS_OVERWRITE_OLDEST=y
CONFIG_BT_HCI_VS=y
CONFIG_BT_HCI_VS_EXT=y
CONFIG_BT_WAIT_NOP=n
CONFIG_BT_DEVICE_NAME="Mini"
CONFIG_BT_DEVICE_APPEARANCE=666
CONFIG_BT_EXT_ADV=y
CONFIG_BT_EXT_ADV_MAX_ADV_SET=15
CONFIG_BT_USER_PHY_UPDATE=y
CONFIG_BT_MAX_CONN=4
# Allow for large Bluetooth data packets.
CONFIG_BT_BUF_ACL_RX_SIZE=255
CONFIG_BT_BUF_CMD_TX_SIZE=255
CONFIG_BT_L2CAP_TX_MTU=247
CONFIG_BT_L2CAP_TX_BUF_COUNT=10
CONFIG_BT_L2CAP_DYNAMIC_CHANNEL=y
CONFIG_BT_ATT_PREPARE_COUNT=2
```
Configuration of the controller(BT part):
```
#BT Driver
CONFIG_BT=y
CONFIG_BT_CTLR=y
CONFIG_BT_CTLR_TX_PWR_PLUS_8=y
CONFIG_NET_BUF=y
CONFIG_BT_HCI_RAW=y
CONFIG_BT_CENTRAL=n
CONFIG_BT_EXT_ADV=y
CONFIG_BT_CTLR_ADV_EXT=y
CONFIG_BT_CTLR_ADV_SET=15
CONFIG_BT_CTLR_ADV_DATA_LEN_MAX=100
CONFIG_BT_CTLR_LE_ENC=y
CONFIG_BT_CTLR_ADVANCED_FEATURES=y
CONFIG_BT_CTLR_OPTIMIZE_FOR_SPEED=y
CONFIG_BT_CTLR_SCHED_ADVANCED=y
CONFIG_BT_MAX_CONN=4
CONFIG_BT_CTLR_TX_PWR_DYNAMIC_CONTROL=y
CONFIG_BT_CTLR_RX_BUFFERS=16
CONFIG_BT_CTLR_PHY_CODED=y
CONFIG_BT_CTLR_ULL_HIGH_PRIO=1
CONFIG_BT_CTLR_ULL_LOW_PRIO=2
CONFIG_BT_CTLR_LOW_LAT=n
CONFIG_BT_ASSERT=n
CONFIG_BT_LL_SOFTDEVICE=n
CONFIG_BT_LL_SW_SPLIT=y
CONFIG_MPSL=n
#BT Minimizing
CONFIG_BT_CTLR_PHY_2M=n
CONFIG_BT_CTLR_PRIVACY=n
CONFIG_BT_CTLR_LE_PING=n
CONFIG_BT_CTLR_LE_ENC=n
CONFIG_BT_CTLR_CRYPTO=n
```
What other outputs might be helpful to identify the origin of the problem ? | 1.0 | Bluetooth controller extended advertisement crashes in lll layer - Zephyr version:
NCS 2.0.0 - Zephyr tag v3.0.99-ncs1
I'm running and Bluetooth host and controller combination of nRF9160(Host) and nRF52833(Controller) using the BT_LL_SW_SPLIT variant. The application is very complicated so i cannot provide a reproducible sample. I have managed to capture an stack trace using Memfault. The problem occurs rather random once in a few hours.

[lll.c](https://github.com/nrfconnect/sdk-zephyr/blob/v3.0.99-ncs1/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c)
Note that the trace is capture with disabled CONFIG_BT_ASSERT=n on the controller.
We had issues with advertisement raising radio tx not ready. In production we disabled the assert to remove unnecessary restarts with cost of some advertisements not getting transmitted.
I can try to catch some debug outputs using the BT_ASSERT, however the assert string is usually lost due to LOG_DEFFER. To catch the issue with memfault is also possibility with the BT_ASSERT but requires more work.
Afterwards i tried to optimize IRQ times by these settings(replaced Zero Latency Interrupts) and haven't tested the assert variant ever since:
```
CONFIG_BT_CTLR_ULL_HIGH_PRIO=1
CONFIG_BT_CTLR_ULL_LOW_PRIO=2
```
Basic description of the application.
It broadcasts a large amount of advertisements sets
6 sets of advertisements using BT legacy(1 connectable).
6 sets of advertisements using BT long range (Coded phy).
Interval of all advertisements are 4HZ.
Application performs regular ADV_DATA updates for the advertisements with interval 1Hz-4Hz.
Communication between Controller and Host is implemented by custom layer. I can provide HCI traces using RTT BT debug if necessary. However, we are using same layer to implement other communication between the chips and we had no problems with it. It is basically H5(Confirmed messages, and retransmission) over multiplexed UART with Flow Control. The controller and host does not seem to generate any error logs.
Configuration of the host(BT part):
```
# Enable Bluetooth stack and libraries
CONFIG_BT=y
CONFIG_BT_PERIPHERAL=y
CONFIG_BT_NO_DRIVER=y
CONFIG_BT_H4=n
CONFIG_BT_SMP=y
CONFIG_BT_KEYS_OVERWRITE_OLDEST=y
CONFIG_BT_HCI_VS=y
CONFIG_BT_HCI_VS_EXT=y
CONFIG_BT_WAIT_NOP=n
CONFIG_BT_DEVICE_NAME="Mini"
CONFIG_BT_DEVICE_APPEARANCE=666
CONFIG_BT_EXT_ADV=y
CONFIG_BT_EXT_ADV_MAX_ADV_SET=15
CONFIG_BT_USER_PHY_UPDATE=y
CONFIG_BT_MAX_CONN=4
# Allow for large Bluetooth data packets.
CONFIG_BT_BUF_ACL_RX_SIZE=255
CONFIG_BT_BUF_CMD_TX_SIZE=255
CONFIG_BT_L2CAP_TX_MTU=247
CONFIG_BT_L2CAP_TX_BUF_COUNT=10
CONFIG_BT_L2CAP_DYNAMIC_CHANNEL=y
CONFIG_BT_ATT_PREPARE_COUNT=2
```
Configuration of the controller(BT part):
```
#BT Driver
CONFIG_BT=y
CONFIG_BT_CTLR=y
CONFIG_BT_CTLR_TX_PWR_PLUS_8=y
CONFIG_NET_BUF=y
CONFIG_BT_HCI_RAW=y
CONFIG_BT_CENTRAL=n
CONFIG_BT_EXT_ADV=y
CONFIG_BT_CTLR_ADV_EXT=y
CONFIG_BT_CTLR_ADV_SET=15
CONFIG_BT_CTLR_ADV_DATA_LEN_MAX=100
CONFIG_BT_CTLR_LE_ENC=y
CONFIG_BT_CTLR_ADVANCED_FEATURES=y
CONFIG_BT_CTLR_OPTIMIZE_FOR_SPEED=y
CONFIG_BT_CTLR_SCHED_ADVANCED=y
CONFIG_BT_MAX_CONN=4
CONFIG_BT_CTLR_TX_PWR_DYNAMIC_CONTROL=y
CONFIG_BT_CTLR_RX_BUFFERS=16
CONFIG_BT_CTLR_PHY_CODED=y
CONFIG_BT_CTLR_ULL_HIGH_PRIO=1
CONFIG_BT_CTLR_ULL_LOW_PRIO=2
CONFIG_BT_CTLR_LOW_LAT=n
CONFIG_BT_ASSERT=n
CONFIG_BT_LL_SOFTDEVICE=n
CONFIG_BT_LL_SW_SPLIT=y
CONFIG_MPSL=n
#BT Minimizing
CONFIG_BT_CTLR_PHY_2M=n
CONFIG_BT_CTLR_PRIVACY=n
CONFIG_BT_CTLR_LE_PING=n
CONFIG_BT_CTLR_LE_ENC=n
CONFIG_BT_CTLR_CRYPTO=n
```
What other outputs might be helpful to identify the origin of the problem ? | priority | bluetooth controller extended advertisement crashes in lll layer zephyr version ncs zephyr tag i m running and bluetooth host and controller combination of host and controller using the bt ll sw split variant the application is very complicated so i cannot provide a reproducible sample i have managed to capture an stack trace using memfault the problem occurs rather random once in a few hours note that the trace is capture with disabled config bt assert n on the controller we had issues with advertisement raising radio tx not ready in production we disabled the assert to remove unnecessary restarts with cost of some advertisements not getting transmitted i can try to catch some debug outputs using the bt assert however the assert string is usually lost due to log deffer to catch the issue with memfault is also possibility with the bt assert but requires more work afterwards i tried to optimize irq times by these settings replaced zero latency interrupts and haven t tested the assert variant ever since config bt ctlr ull high prio config bt ctlr ull low prio basic description of the application it broadcasts a large amount of advertisements sets sets of advertisements using bt legacy connectable sets of advertisements using bt long range coded phy interval of all advertisements are application performs regular adv data updates for the advertisements with interval communication between controller and host is implemented by custom layer i can provide hci traces using rtt bt debug if necessary however we are using same layer to implement other communication between the chips and we had no problems with it it is basically confirmed messages and retransmission over multiplexed uart with flow control the controller and host does not seem to generate any error logs configuration of the host bt part enable bluetooth stack and libraries config bt y config bt peripheral y config bt no driver y config bt n config bt smp y config bt keys overwrite oldest y config bt hci vs y config bt hci vs ext y config bt wait nop n config bt device name mini config bt device appearance config bt ext adv y config bt ext adv max adv set config bt user phy update y config bt max conn allow for large bluetooth data packets config bt buf acl rx size config bt buf cmd tx size config bt tx mtu config bt tx buf count config bt dynamic channel y config bt att prepare count configuration of the controller bt part bt driver config bt y config bt ctlr y config bt ctlr tx pwr plus y config net buf y config bt hci raw y config bt central n config bt ext adv y config bt ctlr adv ext y config bt ctlr adv set config bt ctlr adv data len max config bt ctlr le enc y config bt ctlr advanced features y config bt ctlr optimize for speed y config bt ctlr sched advanced y config bt max conn config bt ctlr tx pwr dynamic control y config bt ctlr rx buffers config bt ctlr phy coded y config bt ctlr ull high prio config bt ctlr ull low prio config bt ctlr low lat n config bt assert n config bt ll softdevice n config bt ll sw split y config mpsl n bt minimizing config bt ctlr phy n config bt ctlr privacy n config bt ctlr le ping n config bt ctlr le enc n config bt ctlr crypto n what other outputs might be helpful to identify the origin of the problem | 1 |
736,842 | 25,490,042,703 | IssuesEvent | 2022-11-26 23:52:53 | Thenlie/Streamability | https://api.github.com/repos/Thenlie/Streamability | closed | 📝Install and configure Tailwind CSS | type: feature priority: medium | Set up tailwind as suggested in the documentation. Make sure this is running painlessly for dev, it should auto update with the site when developing but remove unused styles when bundling for production.
This should be set up initially with the understanding we will need to make some custom classes and possible using the `@apply` directive to add them to the DOM.
References #63 | 1.0 | 📝Install and configure Tailwind CSS - Set up tailwind as suggested in the documentation. Make sure this is running painlessly for dev, it should auto update with the site when developing but remove unused styles when bundling for production.
This should be set up initially with the understanding we will need to make some custom classes and possible using the `@apply` directive to add them to the DOM.
References #63 | priority | 📝install and configure tailwind css set up tailwind as suggested in the documentation make sure this is running painlessly for dev it should auto update with the site when developing but remove unused styles when bundling for production this should be set up initially with the understanding we will need to make some custom classes and possible using the apply directive to add them to the dom references | 1 |
289,026 | 8,854,282,774 | IssuesEvent | 2019-01-09 00:37:36 | visit-dav/issues-test | https://api.github.com/repos/visit-dav/issues-test | closed | Gadget reader failure on re-open | bug crash likelihood medium priority reviewed severity high wrong results | Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1352
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Gadget reader failure on re-open
Assigned to: Kathleen Biagas
Category: -
Target version: 2.6.2
Author: Kathleen Biagas
Start: 02/26/2013
Due date:
% Done: 0%
Estimated time: 2.00 hours
Created: 02/26/2013 06:19 pm
Updated: 02/27/2013 07:10 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.6.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database.
Comments:
Made the reader close its file handles.M databases/Gadget/avtGadgetFileFormat.CM databases/Gadget/avtGadgetFileFormat.h
| 1.0 | Gadget reader failure on re-open - Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1352
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Gadget reader failure on re-open
Assigned to: Kathleen Biagas
Category: -
Target version: 2.6.2
Author: Kathleen Biagas
Start: 02/26/2013
Due date:
% Done: 0%
Estimated time: 2.00 hours
Created: 02/26/2013 06:19 pm
Updated: 02/27/2013 07:10 pm
Likelihood: 3 - Occasional
Severity: 4 - Crash / Wrong Results
Found in version: 2.6.0
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database.
Comments:
Made the reader close its file handles.M databases/Gadget/avtGadgetFileFormat.CM databases/Gadget/avtGadgetFileFormat.h
| priority | gadget reader failure on re open allen harvey reported strangeness with the gadget reader he has a lot of files grouped in a virtual database if he saves settings from a session where he has opened this database then restarts visit and makes a plot clicking reopen will yield in the information window visit could not read from the file home customerdata allenharvey gadget snap database the generated error message was there was an error opening home customerdata allenharvey gadget snap database it may be an invalid file visit tried using the following file format readers to open the file gadget the following error s may be helpful in identifying the problem the selected database type was not correct for the given file the message from the database was cant open file i added code to print out the errno and message from the failed fopen attempt in the gadget reader which is in the constructor by the way and got this error too many open files strangely enough if you delete the config files and retry visit does not complain when reopening the database redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject gadget reader failure on re open assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time hours created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description allen harvey reported strangeness with the gadget reader he has a lot of files grouped in a virtual database if he saves settings from a session where he has opened this database then restarts visit and makes a plot clicking reopen will yield in the information window visit could not read from the file home customerdata allenharvey gadget snap database the generated error message was there was an error opening home customerdata allenharvey gadget snap database it may be an invalid file visit tried using the following file format readers to open the file gadget the following error s may be helpful in identifying the problem the selected database type was not correct for the given file the message from the database was cant open file i added code to print out the errno and message from the failed fopen attempt in the gadget reader which is in the constructor by the way and got this error too many open files strangely enough if you delete the config files and retry visit does not complain when reopening the database comments made the reader close its file handles m databases gadget avtgadgetfileformat cm databases gadget avtgadgetfileformat h | 1 |
150,833 | 5,791,823,112 | IssuesEvent | 2017-05-02 07:34:24 | Caleydo/taggle | https://api.github.com/repos/Caleydo/taggle | opened | Apply column stratification of first categorical column correctly | bug medium priority | * Release number or git hash:
* Web browser version and OS:
* Environment (local or deployed):
### Steps to reproduce
1. Add a matrix (e.g. Number1Hits)
2. Add a categorical column for this matrix (e.g. continent)
3. Split matrix by min/mean/max (type doesn't matter)
### Observed behavior
* The matrix is not splitted into 3 vectors as expected
* The user first has to click the stratify button of the categorical column and then split

### Expected behavior
* The stratification should be applied automatically when adding the first categorical column to a matrix
| 1.0 | Apply column stratification of first categorical column correctly - * Release number or git hash:
* Web browser version and OS:
* Environment (local or deployed):
### Steps to reproduce
1. Add a matrix (e.g. Number1Hits)
2. Add a categorical column for this matrix (e.g. continent)
3. Split matrix by min/mean/max (type doesn't matter)
### Observed behavior
* The matrix is not splitted into 3 vectors as expected
* The user first has to click the stratify button of the categorical column and then split

### Expected behavior
* The stratification should be applied automatically when adding the first categorical column to a matrix
| priority | apply column stratification of first categorical column correctly release number or git hash web browser version and os environment local or deployed steps to reproduce add a matrix e g add a categorical column for this matrix e g continent split matrix by min mean max type doesn t matter observed behavior the matrix is not splitted into vectors as expected the user first has to click the stratify button of the categorical column and then split expected behavior the stratification should be applied automatically when adding the first categorical column to a matrix | 1 |
320,980 | 9,791,862,032 | IssuesEvent | 2019-06-10 15:58:29 | OperationCode/front-end | https://api.github.com/repos/OperationCode/front-end | closed | Prop-type error in Select + Input component | Priority: Medium Type: Developer Experience / Refactoring help wanted | # Bug Report
## What is the current behavior?
I've duct-taped the behavior in https://github.com/OperationCode/front-end/pull/506 , but if you undo that PR and run forms with a Select where `isMulti` is true, you'll see a prop-type error where `form.touched.$fieldName` is an array and it should be a boolean.
**Error Stack**
```
Error: Uncaught [Error: Warning: Failed prop type: Invalid prop `form.touched.programmingLanguages` of type `array` supplied to `Select`, expected `boolean`.
in Select (created by FieldInner)
in FieldInner (created by Context.Consumer)
in FormikConnect(FieldInner) (at Technology.js:71)
in div (at Technology.js:70)
in Technology (at Technology.test.js:39)
in form (at Form.js:5)
in Unknown (created by Context.Consumer)
in FormikConnectedForm (at Technology.test.js:38)
in Formik (created by WrapperComponent)
in WrapperComponent]
```
**Implementation**
```javascript
<Field
className={styles.fullWidth}
name="programmingLanguages"
label="Programming Languages That Interest You"
component={Select}
isMulti
options={[{ label: 'JavaScript', value: 'javascript' }, { label: 'Python', value: 'python' }, /* etc */]}
disabled={isSubmitting}
placeholder="Select at least one..."
/>
```
## What is the expected behavior?
Essentially `form.touched` should be an object where ever form field name is represented in the keys and the values are a boolean on whether or not they've been "touched".
## What steps did you take to get this behavior?
Undo changes in aforementioned duct tape PR, play around inside profile update forms, and see run-time prop-type error.
## Additional Context
I started doubting the source code of Formik, and was able to get what I believe is a troubling result (copied code from `formik/src/utils.ts` into console on lodash's website):


[Here's a Code Sandbox](https://codesandbox.io/embed/710qrr0qjx?fontsize=14) with a `react-select` component inside a Formik context where `isMulti` is true and `touched` is always a boolean.
| 1.0 | Prop-type error in Select + Input component - # Bug Report
## What is the current behavior?
I've duct-taped the behavior in https://github.com/OperationCode/front-end/pull/506 , but if you undo that PR and run forms with a Select where `isMulti` is true, you'll see a prop-type error where `form.touched.$fieldName` is an array and it should be a boolean.
**Error Stack**
```
Error: Uncaught [Error: Warning: Failed prop type: Invalid prop `form.touched.programmingLanguages` of type `array` supplied to `Select`, expected `boolean`.
in Select (created by FieldInner)
in FieldInner (created by Context.Consumer)
in FormikConnect(FieldInner) (at Technology.js:71)
in div (at Technology.js:70)
in Technology (at Technology.test.js:39)
in form (at Form.js:5)
in Unknown (created by Context.Consumer)
in FormikConnectedForm (at Technology.test.js:38)
in Formik (created by WrapperComponent)
in WrapperComponent]
```
**Implementation**
```javascript
<Field
className={styles.fullWidth}
name="programmingLanguages"
label="Programming Languages That Interest You"
component={Select}
isMulti
options={[{ label: 'JavaScript', value: 'javascript' }, { label: 'Python', value: 'python' }, /* etc */]}
disabled={isSubmitting}
placeholder="Select at least one..."
/>
```
## What is the expected behavior?
Essentially `form.touched` should be an object where ever form field name is represented in the keys and the values are a boolean on whether or not they've been "touched".
## What steps did you take to get this behavior?
Undo changes in aforementioned duct tape PR, play around inside profile update forms, and see run-time prop-type error.
## Additional Context
I started doubting the source code of Formik, and was able to get what I believe is a troubling result (copied code from `formik/src/utils.ts` into console on lodash's website):


[Here's a Code Sandbox](https://codesandbox.io/embed/710qrr0qjx?fontsize=14) with a `react-select` component inside a Formik context where `isMulti` is true and `touched` is always a boolean.
| priority | prop type error in select input component bug report what is the current behavior i ve duct taped the behavior in but if you undo that pr and run forms with a select where ismulti is true you ll see a prop type error where form touched fieldname is an array and it should be a boolean error stack error uncaught error warning failed prop type invalid prop form touched programminglanguages of type array supplied to select expected boolean in select created by fieldinner in fieldinner created by context consumer in formikconnect fieldinner at technology js in div at technology js in technology at technology test js in form at form js in unknown created by context consumer in formikconnectedform at technology test js in formik created by wrappercomponent in wrappercomponent implementation javascript field classname styles fullwidth name programminglanguages label programming languages that interest you component select ismulti options disabled issubmitting placeholder select at least one what is the expected behavior essentially form touched should be an object where ever form field name is represented in the keys and the values are a boolean on whether or not they ve been touched what steps did you take to get this behavior undo changes in aforementioned duct tape pr play around inside profile update forms and see run time prop type error additional context i started doubting the source code of formik and was able to get what i believe is a troubling result copied code from formik src utils ts into console on lodash s website with a react select component inside a formik context where ismulti is true and touched is always a boolean | 1 |
730,600 | 25,181,228,053 | IssuesEvent | 2022-11-11 13:52:05 | eclipse/lsp4jakarta | https://api.github.com/repos/eclipse/lsp4jakarta | closed | Update snippets with more relevant names, prefixes and descriptions for Jakarta EE 9+ | 1 medium priority | The Jakarta RESTful Web Services snippets still refer to "JAX-RS" in the description: https://github.com/eclipse/lsp4jakarta/blob/7fb1fe4fdfac031819f426240b6227b06ab712d1/jakarta.ls/src/main/resources/jax-rs.json#L2-L25
Update to use more relevant names, prefixes and descriptions for Jakarta EE 9+
Also consider if any of the other snippets should be updated in the same way:
https://github.com/eclipse/lsp4jakarta/tree/main/jakarta.ls/src/main/resources | 1.0 | Update snippets with more relevant names, prefixes and descriptions for Jakarta EE 9+ - The Jakarta RESTful Web Services snippets still refer to "JAX-RS" in the description: https://github.com/eclipse/lsp4jakarta/blob/7fb1fe4fdfac031819f426240b6227b06ab712d1/jakarta.ls/src/main/resources/jax-rs.json#L2-L25
Update to use more relevant names, prefixes and descriptions for Jakarta EE 9+
Also consider if any of the other snippets should be updated in the same way:
https://github.com/eclipse/lsp4jakarta/tree/main/jakarta.ls/src/main/resources | priority | update snippets with more relevant names prefixes and descriptions for jakarta ee the jakarta restful web services snippets still refer to jax rs in the description update to use more relevant names prefixes and descriptions for jakarta ee also consider if any of the other snippets should be updated in the same way | 1 |
235,138 | 7,734,949,419 | IssuesEvent | 2018-05-27 09:17:30 | Radarr/Radarr | https://api.github.com/repos/Radarr/Radarr | closed | Add AHD_Internal Flag | confirmed enhancement indexer priority:medium | Would it be possible to add an AHD indexer? There is a search api available at
`https://awesome-hd.me/searchapi.php?action=imdbsearch&passkey=%s&imdb=%s&internal=%s`
In this passkey is the users passkey, imdb is the IMDb ID int he form tt1234567 and internal is a boolean that will set the results to only include AHD internals if true or all releases if false or not included.
This will return the following for a search for Bride of Chucky (IMDb ID tt0144120)
```
<?xml version="1.0"?>
<searchresults>
<authkey>AUTHKEY REMOVED</authkey>
<torrent>
<id>18809</id>
<groupid>9687</groupid>
<time>2014-03-18 22:29:33</time>
<userid>33249</userid>
<size>22661245411</size>
<snatched>35</snatched>
<seeders>1</seeders>
<leechers>0</leechers>
<releasegroup>HiFi</releasegroup>
<resolution>1080p</resolution>
<media>Blu-ray</media>
<format>MKV</format>
<encoding>VC-1 Remux</encoding>
<audioformat>DTS-HD MA</audioformat>
<audiobitrate>Variable</audiobitrate>
<audiochannels>6</audiochannels>
<subtitles>English, French, Spanish</subtitles>
<encodestatus>AHDGolddxva</encodestatus>
<freeleech>0.50</freeleech>
<cover>
http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</cover>
<smallcover>
http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</smallcover>
<year>1998</year>
<name>Bride of Chucky</name>
<imdb>tt0144120</imdb>
<type>Movie</type>
<plotoutline>
Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll.
</plotoutline>
</torrent>
<torrent>
<id>17500</id>
<groupid>9687</groupid>
<time>2014-02-02 06:07:05</time>
<userid>37948</userid>
<size>22661245137</size>
<snatched>6</snatched>
<seeders>0</seeders>
<leechers>0</leechers>
<releasegroup>User-Upload</releasegroup>
<resolution>1080p</resolution>
<media>Blu-ray</media>
<format>MKV</format>
<encoding>VC-1 Remux</encoding>
<audioformat>DTS-HD MA</audioformat>
<audiobitrate>Variable</audiobitrate>
<audiochannels>6</audiochannels>
<subtitles>English, Spanish, French</subtitles>
<encodestatus>AHDGolddxva</encodestatus>
<freeleech>0.50</freeleech>
<cover>
http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</cover>
<smallcover>
http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</smallcover>
<year>1998</year>
<name>Bride of Chucky</name>
<imdb>tt0144120</imdb>
<type>Movie</type>
<plotoutline>
Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll.
</plotoutline>
</torrent>
</searchresults>
```
The download link for the above then would be
`https://awesome-hd.me/torrents.php?action=download&id=%s&authkey=%s&torrent_pass=%s'`
the authkey and id are returned in the search results above and the torrent_pass is the passkey as used earlier. | 1.0 | Add AHD_Internal Flag - Would it be possible to add an AHD indexer? There is a search api available at
`https://awesome-hd.me/searchapi.php?action=imdbsearch&passkey=%s&imdb=%s&internal=%s`
In this passkey is the users passkey, imdb is the IMDb ID int he form tt1234567 and internal is a boolean that will set the results to only include AHD internals if true or all releases if false or not included.
This will return the following for a search for Bride of Chucky (IMDb ID tt0144120)
```
<?xml version="1.0"?>
<searchresults>
<authkey>AUTHKEY REMOVED</authkey>
<torrent>
<id>18809</id>
<groupid>9687</groupid>
<time>2014-03-18 22:29:33</time>
<userid>33249</userid>
<size>22661245411</size>
<snatched>35</snatched>
<seeders>1</seeders>
<leechers>0</leechers>
<releasegroup>HiFi</releasegroup>
<resolution>1080p</resolution>
<media>Blu-ray</media>
<format>MKV</format>
<encoding>VC-1 Remux</encoding>
<audioformat>DTS-HD MA</audioformat>
<audiobitrate>Variable</audiobitrate>
<audiochannels>6</audiochannels>
<subtitles>English, French, Spanish</subtitles>
<encodestatus>AHDGolddxva</encodestatus>
<freeleech>0.50</freeleech>
<cover>
http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</cover>
<smallcover>
http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</smallcover>
<year>1998</year>
<name>Bride of Chucky</name>
<imdb>tt0144120</imdb>
<type>Movie</type>
<plotoutline>
Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll.
</plotoutline>
</torrent>
<torrent>
<id>17500</id>
<groupid>9687</groupid>
<time>2014-02-02 06:07:05</time>
<userid>37948</userid>
<size>22661245137</size>
<snatched>6</snatched>
<seeders>0</seeders>
<leechers>0</leechers>
<releasegroup>User-Upload</releasegroup>
<resolution>1080p</resolution>
<media>Blu-ray</media>
<format>MKV</format>
<encoding>VC-1 Remux</encoding>
<audioformat>DTS-HD MA</audioformat>
<audiobitrate>Variable</audiobitrate>
<audiochannels>6</audiochannels>
<subtitles>English, Spanish, French</subtitles>
<encodestatus>AHDGolddxva</encodestatus>
<freeleech>0.50</freeleech>
<cover>
http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</cover>
<smallcover>
http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg
</smallcover>
<year>1998</year>
<name>Bride of Chucky</name>
<imdb>tt0144120</imdb>
<type>Movie</type>
<plotoutline>
Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll.
</plotoutline>
</torrent>
</searchresults>
```
The download link for the above then would be
`https://awesome-hd.me/torrents.php?action=download&id=%s&authkey=%s&torrent_pass=%s'`
the authkey and id are returned in the search results above and the torrent_pass is the passkey as used earlier. | priority | add ahd internal flag would it be possible to add an ahd indexer there is a search api available at in this passkey is the users passkey imdb is the imdb id int he form and internal is a boolean that will set the results to only include ahd internals if true or all releases if false or not included this will return the following for a search for bride of chucky imdb id authkey removed hifi blu ray mkv vc remux dts hd ma variable english french spanish ahdgolddxva bride of chucky movie chucky the doll possessed by a serial killer discovers the perfect mate to kill and revive into the body of another doll user upload blu ray mkv vc remux dts hd ma variable english spanish french ahdgolddxva bride of chucky movie chucky the doll possessed by a serial killer discovers the perfect mate to kill and revive into the body of another doll the download link for the above then would be the authkey and id are returned in the search results above and the torrent pass is the passkey as used earlier | 1 |
26,463 | 2,684,555,866 | IssuesEvent | 2015-03-29 03:30:53 | gtcasl/gpuocelot | https://api.github.com/repos/gtcasl/gpuocelot | closed | Ocelot does not build from source | bug imported Priority-Medium | _From [lokendra...@gmail.com](https://code.google.com/u/102201010740957725244/) on January 10, 2012 01:41:00_
What steps will reproduce the problem? 1. ./build.py --install -p /home/zenome/local --no_llvm What is the expected output? What do you see instead? Expected: compilation successful
Output:
Installing ocelot...
config_h_build([".release_build/configure.h"], ["configure.h.in"])
g++ -o .release_build/ocelot/tools/OcelotConfig.o -c -O2 -Wall -Werror -std=c++0x -DOCELOT_CXXFLAGS="\"-O2 -Wall -Werror -std=c++0x\"" -DPACKAGE="\"ocelot\"" -DVERSION="\"2.1.1624\"" -DOCELOT_PREFIX_PATH="\"/home/zenome/local\"" -DOCELOT_LDFLAGS="\"-locelot\"" -L"\"/home/zenome/local/lib\"" -DOCELOT_INCLUDE_PATH="\"/home/zenome/local/include\"" -DOCELOT_LIB_PATH="\"/home/zenome/local/lib\"" -DOCELOT_BIN_PATH="\"/home/zenome/local/bin\"" -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/OcelotConfig.cpp
g++ -o .release_build/ocelot/tools/CFG.o -c -O2 -Wall -Werror -std=c++0x -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/CFG.cpp
ocelot/tools/CFG.cpp: In function 'void analyze(const char*)':
ocelot/tools/CFG.cpp:56:11: error: 'class ir::PTXKernel' has no member named 'dom_tree'
ocelot/tools/CFG.cpp:61:11: error: 'class ir::PTXKernel' has no member named 'pdom_tree'
ocelot/tools/CFG.cpp:66:11: error: 'class ir::PTXKernel' has no member named 'ctrl_tree' What version of the product are you using? On what operating system? Ocelot: 1649
Ubuntu-11.04
GCC/++ : 4.5.2
Boost: 1.46.0 Please provide any additional information below.
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=61_ | 1.0 | Ocelot does not build from source - _From [lokendra...@gmail.com](https://code.google.com/u/102201010740957725244/) on January 10, 2012 01:41:00_
What steps will reproduce the problem? 1. ./build.py --install -p /home/zenome/local --no_llvm What is the expected output? What do you see instead? Expected: compilation successful
Output:
Installing ocelot...
config_h_build([".release_build/configure.h"], ["configure.h.in"])
g++ -o .release_build/ocelot/tools/OcelotConfig.o -c -O2 -Wall -Werror -std=c++0x -DOCELOT_CXXFLAGS="\"-O2 -Wall -Werror -std=c++0x\"" -DPACKAGE="\"ocelot\"" -DVERSION="\"2.1.1624\"" -DOCELOT_PREFIX_PATH="\"/home/zenome/local\"" -DOCELOT_LDFLAGS="\"-locelot\"" -L"\"/home/zenome/local/lib\"" -DOCELOT_INCLUDE_PATH="\"/home/zenome/local/include\"" -DOCELOT_LIB_PATH="\"/home/zenome/local/lib\"" -DOCELOT_BIN_PATH="\"/home/zenome/local/bin\"" -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/OcelotConfig.cpp
g++ -o .release_build/ocelot/tools/CFG.o -c -O2 -Wall -Werror -std=c++0x -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/CFG.cpp
ocelot/tools/CFG.cpp: In function 'void analyze(const char*)':
ocelot/tools/CFG.cpp:56:11: error: 'class ir::PTXKernel' has no member named 'dom_tree'
ocelot/tools/CFG.cpp:61:11: error: 'class ir::PTXKernel' has no member named 'pdom_tree'
ocelot/tools/CFG.cpp:66:11: error: 'class ir::PTXKernel' has no member named 'ctrl_tree' What version of the product are you using? On what operating system? Ocelot: 1649
Ubuntu-11.04
GCC/++ : 4.5.2
Boost: 1.46.0 Please provide any additional information below.
_Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=61_ | priority | ocelot does not build from source from on january what steps will reproduce the problem build py install p home zenome local no llvm what is the expected output what do you see instead expected compilation successful output installing ocelot config h build g o release build ocelot tools ocelotconfig o c wall werror std c docelot cxxflags wall werror std c dpackage ocelot dversion docelot prefix path home zenome local docelot ldflags locelot l home zenome local lib docelot include path home zenome local include docelot lib path home zenome local lib docelot bin path home zenome local bin i i release build i home zenome local share boost install include i usr include i usr include ocelot tools ocelotconfig cpp g o release build ocelot tools cfg o c wall werror std c i i release build i home zenome local share boost install include i usr include i usr include ocelot tools cfg cpp ocelot tools cfg cpp in function void analyze const char ocelot tools cfg cpp error class ir ptxkernel has no member named dom tree ocelot tools cfg cpp error class ir ptxkernel has no member named pdom tree ocelot tools cfg cpp error class ir ptxkernel has no member named ctrl tree what version of the product are you using on what operating system ocelot ubuntu gcc boost please provide any additional information below original issue | 1 |
22,293 | 2,648,595,179 | IssuesEvent | 2015-03-14 02:15:07 | dhamp/eiskaltdcpp | https://api.github.com/repos/dhamp/eiskaltdcpp | opened | feature to log bot private messages | imported Priority-Medium Type-FeatureRequest | _From [hg2rbw](https://code.google.com/u/hg2rbw/) on May 07, 2011 20:42:37_
feature to log bot private messages like opchat
_Original issue: http://code.google.com/p/eiskaltdc/issues/detail?id=1067_ | 1.0 | feature to log bot private messages - _From [hg2rbw](https://code.google.com/u/hg2rbw/) on May 07, 2011 20:42:37_
feature to log bot private messages like opchat
_Original issue: http://code.google.com/p/eiskaltdc/issues/detail?id=1067_ | priority | feature to log bot private messages from on may feature to log bot private messages like opchat original issue | 1 |
449,328 | 12,967,346,953 | IssuesEvent | 2020-07-21 02:53:46 | ansible/awx | https://api.github.com/repos/ansible/awx | closed | Improve Diff Output When Specifying Show Changes in Job Template | component:ui priority:medium state:needs_devel type:enhancement | ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
<!-- Briefly describe the problem or desired enhancement. -->
Per an user's request, they would like to see an improved output when specifying Show Changes in a job template.
For example, if a job has thousands of lines of output, they would like to be able to filter for just the diff changes (i.e. For reporting purposes). | 1.0 | Improve Diff Output When Specifying Show Changes in Job Template - ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
<!-- Briefly describe the problem or desired enhancement. -->
Per an user's request, they would like to see an improved output when specifying Show Changes in a job template.
For example, if a job has thousands of lines of output, they would like to be able to filter for just the diff changes (i.e. For reporting purposes). | priority | improve diff output when specifying show changes in job template issue type feature idea summary per an user s request they would like to see an improved output when specifying show changes in a job template for example if a job has thousands of lines of output they would like to be able to filter for just the diff changes i e for reporting purposes | 1 |
229,995 | 7,603,084,074 | IssuesEvent | 2018-04-29 10:33:52 | resin-io/resin-supervisor | https://api.github.com/repos/resin-io/resin-supervisor | closed | Trailing slash in WORKDIR causes a service restart loop | Low-hanging fruit Medium Priority flow/in-progress type/bug | If the image for a service has `Config.WorkingDir = "/usr/src/app/"`, the resulting container will have `Config.WorkingDir = "/usr/src/app"` - so the isSameContainer check in src/compose/service.coffee will always be false.
This results in an endless loop of the supervisor killing and restarting the affected service.
Workaround is to define WORKDIR without trailing slashes (i.e. `/usr/src/app` instead of `/usr/src/app/`).
The fix would be to remove trailing slashes from `@workingDir` in the Service constructor.
<img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_bbyz) | 1.0 | Trailing slash in WORKDIR causes a service restart loop - If the image for a service has `Config.WorkingDir = "/usr/src/app/"`, the resulting container will have `Config.WorkingDir = "/usr/src/app"` - so the isSameContainer check in src/compose/service.coffee will always be false.
This results in an endless loop of the supervisor killing and restarting the affected service.
Workaround is to define WORKDIR without trailing slashes (i.e. `/usr/src/app` instead of `/usr/src/app/`).
The fix would be to remove trailing slashes from `@workingDir` in the Service constructor.
<img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_bbyz) | priority | trailing slash in workdir causes a service restart loop if the image for a service has config workingdir usr src app the resulting container will have config workingdir usr src app so the issamecontainer check in src compose service coffee will always be false this results in an endless loop of the supervisor killing and restarting the affected service workaround is to define workdir without trailing slashes i e usr src app instead of usr src app the fix would be to remove trailing slashes from workingdir in the service constructor | 1 |
689,195 | 23,611,361,487 | IssuesEvent | 2022-08-24 12:41:41 | Knowledge-Management-Capstone/knowledge-management-dashboard | https://api.github.com/repos/Knowledge-Management-Capstone/knowledge-management-dashboard | opened | KMND-133 Revamp `DisscussionPage` UI | type:enhancement estimated-sp:2 priority:medium | ## Description
The problem with current UI is hard to display when the message is sent.
I propose to use Slack/Discord like UI to overcome this.
Please refer to [this repository](https://github.com/stuyy/chat-platform-react)
| 1.0 | KMND-133 Revamp `DisscussionPage` UI - ## Description
The problem with current UI is hard to display when the message is sent.
I propose to use Slack/Discord like UI to overcome this.
Please refer to [this repository](https://github.com/stuyy/chat-platform-react)
| priority | kmnd revamp disscussionpage ui description the problem with current ui is hard to display when the message is sent i propose to use slack discord like ui to overcome this please refer to | 1 |
129,428 | 5,096,915,235 | IssuesEvent | 2017-01-03 19:45:26 | dteviot/WebToEpub | https://api.github.com/repos/dteviot/WebToEpub | reopened | [imgur] grab images from imgur | bug medium priority | I'm pretty sure there was talk about this but I can't seem to find it any more so I made a separate issue for sake of visibility.
An example would be "https://crimsonmagic.me/2016/11/16/gifting-10-1/" [FIXED LINK], where there are various image -links- like "http://imgur.com/K4CZyyP.jpg" that don't get grabbed (or shown).
There's also the more complex case of linked imgur galleries like on "http://skythewoodtl.com/fanfic-gifting/" [ADDED LINK], where you have a link to an imgur gallery with all the illustration before the novel ("Images" at the end).
| 1.0 | [imgur] grab images from imgur - I'm pretty sure there was talk about this but I can't seem to find it any more so I made a separate issue for sake of visibility.
An example would be "https://crimsonmagic.me/2016/11/16/gifting-10-1/" [FIXED LINK], where there are various image -links- like "http://imgur.com/K4CZyyP.jpg" that don't get grabbed (or shown).
There's also the more complex case of linked imgur galleries like on "http://skythewoodtl.com/fanfic-gifting/" [ADDED LINK], where you have a link to an imgur gallery with all the illustration before the novel ("Images" at the end).
| priority | grab images from imgur i m pretty sure there was talk about this but i can t seem to find it any more so i made a separate issue for sake of visibility an example would be where there are various image links like that don t get grabbed or shown there s also the more complex case of linked imgur galleries like on where you have a link to an imgur gallery with all the illustration before the novel images at the end | 1 |
249,429 | 7,961,728,018 | IssuesEvent | 2018-07-13 11:55:30 | magicDGS/ReadTools | https://api.github.com/repos/magicDGS/ReadTools | closed | Check usages of GATKRead defensive copy methods and substitute for performance improvement | Priority: Medium Status: Accepted Type: Enhancement | For example, `CheckQualityReadTransformer` makes always a defensive copy for the qualities, when it is not required because there is only an assertion and not change of qualities. This might reduce performance (concretely in the example case, it is performed quite often). | 1.0 | Check usages of GATKRead defensive copy methods and substitute for performance improvement - For example, `CheckQualityReadTransformer` makes always a defensive copy for the qualities, when it is not required because there is only an assertion and not change of qualities. This might reduce performance (concretely in the example case, it is performed quite often). | priority | check usages of gatkread defensive copy methods and substitute for performance improvement for example checkqualityreadtransformer makes always a defensive copy for the qualities when it is not required because there is only an assertion and not change of qualities this might reduce performance concretely in the example case it is performed quite often | 1 |
167,287 | 6,335,898,146 | IssuesEvent | 2017-07-26 19:53:06 | ngageoint/hootenanny | https://api.github.com/repos/ngageoint/hootenanny | opened | Error conflating single Rafah tile with HG workflow | Category: Core Priority: Medium Status: Defined Type: Bug | `scripts/core/ServiceOsmApiDbHootApiDbConflate.sh tmp/rafah.osm tmp/Rafah_UTDS_5_FCODE_noemptyFC_v61.osm "" "unifying" ServiceOsmApiDbHootApiDbRafahConflateTest2 true 1
Error running apply-changeset:
Error executing query: ERROR: duplicate key value violates unique constraint "changesets_pkey"
DETAIL: Key (id)=(6) already exists.
(23505) QPSQL: Unable to create query (INSERT INTO changesets (id, user_id, created_at, closed_at) VALUES (6, 1, (now() at time zone 'utc'), (now() at time zone 'utc'));)`
This actually happens for any random tile selected. | 1.0 | Error conflating single Rafah tile with HG workflow - `scripts/core/ServiceOsmApiDbHootApiDbConflate.sh tmp/rafah.osm tmp/Rafah_UTDS_5_FCODE_noemptyFC_v61.osm "" "unifying" ServiceOsmApiDbHootApiDbRafahConflateTest2 true 1
Error running apply-changeset:
Error executing query: ERROR: duplicate key value violates unique constraint "changesets_pkey"
DETAIL: Key (id)=(6) already exists.
(23505) QPSQL: Unable to create query (INSERT INTO changesets (id, user_id, created_at, closed_at) VALUES (6, 1, (now() at time zone 'utc'), (now() at time zone 'utc'));)`
This actually happens for any random tile selected. | priority | error conflating single rafah tile with hg workflow scripts core serviceosmapidbhootapidbconflate sh tmp rafah osm tmp rafah utds fcode noemptyfc osm unifying true error running apply changeset error executing query error duplicate key value violates unique constraint changesets pkey detail key id already exists qpsql unable to create query insert into changesets id user id created at closed at values now at time zone utc now at time zone utc this actually happens for any random tile selected | 1 |
439,219 | 12,679,487,608 | IssuesEvent | 2020-06-19 11:54:20 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | closed | LT.61006 | :beetle: bug :beetle: :grey_exclamation: priority medium | **Mod Version**
f63a4a0d
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
LT.61006 can make no sense for a lot of characters
**Steps to reproduce the issue:**
Fire LT.61006
**Upload an attachment below: .zip of your save, or screenshots:**
<details>
<summary>Click to expand</summary>

</details> | 1.0 | LT.61006 - **Mod Version**
f63a4a0d
**What expansions do you have installed?**
All
**Please explain your issue in as much detail as possible:**
LT.61006 can make no sense for a lot of characters
**Steps to reproduce the issue:**
Fire LT.61006
**Upload an attachment below: .zip of your save, or screenshots:**
<details>
<summary>Click to expand</summary>

</details> | priority | lt mod version what expansions do you have installed all please explain your issue in as much detail as possible lt can make no sense for a lot of characters steps to reproduce the issue fire lt upload an attachment below zip of your save or screenshots click to expand | 1 |
509,438 | 14,731,027,317 | IssuesEvent | 2021-01-06 14:07:28 | profclems/glab | https://api.github.com/repos/profclems/glab | reopened | Friendlier recovery when issue/mr create fails | help wanted kind: enhancement priority: medium | **Describe the feature or problem you'd like to solve**
When `glab mr create` fails or cancelled, the body content typed using the editor is lost.
Sometimes the user may type a long MR description and just lose it due to a simple mistake, such as not adding any commit to the MR branch yet (e.g., when creating draft MRs).
**Propose a Solution**
To me, the best user experience would be if the person can just resume where they left off so whenever someone authors an MR or issue, we save their title & body in a file scoped to that repository.
If the create operation succeeded, we delete the file. Otherwise, the save file stays on disk and is used to pre-fill the next `issue/mr` create invocation, which is most likely a retry after a server or validation failure.
Or instead of prefilling, we provide `--recover` flag to continue from the last session
**Additional context**
<!-- Add any other context or screenshots about the feature request here. -->
| 1.0 | Friendlier recovery when issue/mr create fails - **Describe the feature or problem you'd like to solve**
When `glab mr create` fails or cancelled, the body content typed using the editor is lost.
Sometimes the user may type a long MR description and just lose it due to a simple mistake, such as not adding any commit to the MR branch yet (e.g., when creating draft MRs).
**Propose a Solution**
To me, the best user experience would be if the person can just resume where they left off so whenever someone authors an MR or issue, we save their title & body in a file scoped to that repository.
If the create operation succeeded, we delete the file. Otherwise, the save file stays on disk and is used to pre-fill the next `issue/mr` create invocation, which is most likely a retry after a server or validation failure.
Or instead of prefilling, we provide `--recover` flag to continue from the last session
**Additional context**
<!-- Add any other context or screenshots about the feature request here. -->
| priority | friendlier recovery when issue mr create fails describe the feature or problem you d like to solve when glab mr create fails or cancelled the body content typed using the editor is lost sometimes the user may type a long mr description and just lose it due to a simple mistake such as not adding any commit to the mr branch yet e g when creating draft mrs propose a solution to me the best user experience would be if the person can just resume where they left off so whenever someone authors an mr or issue we save their title body in a file scoped to that repository if the create operation succeeded we delete the file otherwise the save file stays on disk and is used to pre fill the next issue mr create invocation which is most likely a retry after a server or validation failure or instead of prefilling we provide recover flag to continue from the last session additional context | 1 |
17,084 | 2,615,130,537 | IssuesEvent | 2015-03-01 06:00:25 | chrsmith/google-api-java-client | https://api.github.com/repos/chrsmith/google-api-java-client | opened | Maps Coordinate sample | auto-migrated Priority-Medium Type-Sample | ```
Which Google API and version: coordinate v1
Java environment (e.g. Java 6, Android 2.3, App Engine)? Java6
External references, such as API reference guide?
https://developers.google.com/coordinate/
Please provide any additional information below.
Service accounts are not currently supported on Google Maps Coordinate API, so
the web server flow should demonstrated. Just like in the Fusion Tables API
sample -
http://code.google.com/p/google-api-java-client/wiki/APIs#Fusion_Tables_API
```
Original issue reported on code.google.com by `nicol...@google.com` on 4 Dec 2012 at 7:33 | 1.0 | Maps Coordinate sample - ```
Which Google API and version: coordinate v1
Java environment (e.g. Java 6, Android 2.3, App Engine)? Java6
External references, such as API reference guide?
https://developers.google.com/coordinate/
Please provide any additional information below.
Service accounts are not currently supported on Google Maps Coordinate API, so
the web server flow should demonstrated. Just like in the Fusion Tables API
sample -
http://code.google.com/p/google-api-java-client/wiki/APIs#Fusion_Tables_API
```
Original issue reported on code.google.com by `nicol...@google.com` on 4 Dec 2012 at 7:33 | priority | maps coordinate sample which google api and version coordinate java environment e g java android app engine external references such as api reference guide please provide any additional information below service accounts are not currently supported on google maps coordinate api so the web server flow should demonstrated just like in the fusion tables api sample original issue reported on code google com by nicol google com on dec at | 1 |
766,175 | 26,873,659,755 | IssuesEvent | 2023-02-04 19:33:42 | ita-social-projects/TeachUA | https://api.github.com/repos/ita-social-projects/TeachUA | closed | [API.Task] "Status 500 Internal server error" response on attempt to create task with valid values | bug Priority: Medium API | **Environment:** macOS High Sierra Version 10.13.6
Postman for Mac Version 10.8.7
**Reproducible:** always
**Preconditions**
1. Run Postman
2. Create new request > Choose 'POST' in 'Method' drop-down
3. Enter request URL: https://speak-ukrainian.org.ua/dev/api/challenge/x/task
4. Navigate to 'Authorization' tab, choose 'Bearer Token' in 'Type' drop-down
5. Set 'accessToken' of user with admin rights "eyJhbGciOiJIUzUxMiJ9.eyJzdWIiOiJhZG1pbkBnbWFpbC5jb20iLCJqdGkiOiIxIiwiZXhwIjoxNjc1Mzc0ODczfQ.dyThALXsOb7qrrG2cr2iUEVAUKloiGTei7Z5GCgaVZcpjPL0a3E-RPBb2Q75b8bBPvXb1YURPgqPKqBNRV5NiA"
* You need to replace "x" in URL with the existing challenge ID (you can check the site or DB)
TD1: "{
"name": " namenamename1213#$% ",
"description": " descriptiondescriptiondescriptiondescriptiondescription12345$%%^$# ",
"picture": "/upload/test/test.png",
"startDate": "2021-12-03"
}"
**Steps to reproduce**
1. Choose body>raw>JSON and enter data(TD1) in field
2. Click on 'Send' button
**Actual result**
Server is responded with status 500
**Expected result**
Server is responded with status 200. Task is successfully created.
**User story and test case links**
[Test case](https://jira.softserve.academy/browse/TUA-441)
**Labels to be added**
"Bug", Priority ("pri: Medium "), Severity ("severity: Major"), Type "API"

| 1.0 | [API.Task] "Status 500 Internal server error" response on attempt to create task with valid values - **Environment:** macOS High Sierra Version 10.13.6
Postman for Mac Version 10.8.7
**Reproducible:** always
**Preconditions**
1. Run Postman
2. Create new request > Choose 'POST' in 'Method' drop-down
3. Enter request URL: https://speak-ukrainian.org.ua/dev/api/challenge/x/task
4. Navigate to 'Authorization' tab, choose 'Bearer Token' in 'Type' drop-down
5. Set 'accessToken' of user with admin rights "eyJhbGciOiJIUzUxMiJ9.eyJzdWIiOiJhZG1pbkBnbWFpbC5jb20iLCJqdGkiOiIxIiwiZXhwIjoxNjc1Mzc0ODczfQ.dyThALXsOb7qrrG2cr2iUEVAUKloiGTei7Z5GCgaVZcpjPL0a3E-RPBb2Q75b8bBPvXb1YURPgqPKqBNRV5NiA"
* You need to replace "x" in URL with the existing challenge ID (you can check the site or DB)
TD1: "{
"name": " namenamename1213#$% ",
"description": " descriptiondescriptiondescriptiondescriptiondescription12345$%%^$# ",
"picture": "/upload/test/test.png",
"startDate": "2021-12-03"
}"
**Steps to reproduce**
1. Choose body>raw>JSON and enter data(TD1) in field
2. Click on 'Send' button
**Actual result**
Server is responded with status 500
**Expected result**
Server is responded with status 200. Task is successfully created.
**User story and test case links**
[Test case](https://jira.softserve.academy/browse/TUA-441)
**Labels to be added**
"Bug", Priority ("pri: Medium "), Severity ("severity: Major"), Type "API"

| priority | status internal server error response on attempt to create task with valid values environment macos high sierra version postman for mac version reproducible always preconditions run postman create new request choose post in method drop down enter request url navigate to authorization tab choose bearer token in type drop down set accesstoken of user with admin rights you need to replace x in url with the existing challenge id you can check the site or db name description picture upload test test png startdate steps to reproduce choose body raw json and enter data in field click on send button actual result server is responded with status expected result server is responded with status task is successfully created user story and test case links labels to be added bug priority pri medium severity severity major type api | 1 |
5,725 | 2,578,768,550 | IssuesEvent | 2015-02-13 02:13:13 | phetsims/website | https://api.github.com/repos/phetsims/website | closed | Header Focus Color Change | Medium Priority | Can the outline be a light color (like white) for focus objects in the blue header area? For example, it’s hard to tell when the PhET logos are focused, and impossible to tell when the “search” button is focused (for me at least). | 1.0 | Header Focus Color Change - Can the outline be a light color (like white) for focus objects in the blue header area? For example, it’s hard to tell when the PhET logos are focused, and impossible to tell when the “search” button is focused (for me at least). | priority | header focus color change can the outline be a light color like white for focus objects in the blue header area for example it’s hard to tell when the phet logos are focused and impossible to tell when the “search” button is focused for me at least | 1 |
87,641 | 3,756,436,865 | IssuesEvent | 2016-03-13 10:26:19 | kotcrab/vis-editor | https://api.github.com/repos/kotcrab/vis-editor | opened | Split FileChooser into separate extension | priority: medium ui | It uses AWT imports which creates lint error when doing gradle build, probably would be the best to split it into separate module. | 1.0 | Split FileChooser into separate extension - It uses AWT imports which creates lint error when doing gradle build, probably would be the best to split it into separate module. | priority | split filechooser into separate extension it uses awt imports which creates lint error when doing gradle build probably would be the best to split it into separate module | 1 |
22,250 | 2,648,473,741 | IssuesEvent | 2015-03-14 00:07:22 | prikhi/pencil | https://api.github.com/repos/prikhi/pencil | closed | "Export as PNG" broken in Firefox 36 | 1 star bug imported Priority-Medium | _From [m.overme...@yahoo.ca](https://code.google.com/u/108010164991601365011/) on March 03, 2015 08:33:14_
What steps will reproduce the problem? 1. Update to Firefox 36
2. Open Pencil Sketching Firefox add-on
3. Try to "Export as PNG" What is the expected output? What do you see instead? The file never appears on the file system. What version of the product are you using? On what operating system? I'm using the Firefox addon in Firefox 36, on Windows 7 Professional. Please provide any additional information below. -"Export as PNG" works as expected in Firefox 35.0.1.
-This is probably due to the changes Firefox made to addon compatibility ( https://blog.mozilla.org/addons/2015/01/13/compatibility-for-firefox-36/ )
-The workaround is to use the Windows build of Pencil.
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=668_ | 1.0 | "Export as PNG" broken in Firefox 36 - _From [m.overme...@yahoo.ca](https://code.google.com/u/108010164991601365011/) on March 03, 2015 08:33:14_
What steps will reproduce the problem? 1. Update to Firefox 36
2. Open Pencil Sketching Firefox add-on
3. Try to "Export as PNG" What is the expected output? What do you see instead? The file never appears on the file system. What version of the product are you using? On what operating system? I'm using the Firefox addon in Firefox 36, on Windows 7 Professional. Please provide any additional information below. -"Export as PNG" works as expected in Firefox 35.0.1.
-This is probably due to the changes Firefox made to addon compatibility ( https://blog.mozilla.org/addons/2015/01/13/compatibility-for-firefox-36/ )
-The workaround is to use the Windows build of Pencil.
_Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=668_ | priority | export as png broken in firefox from on march what steps will reproduce the problem update to firefox open pencil sketching firefox add on try to export as png what is the expected output what do you see instead the file never appears on the file system what version of the product are you using on what operating system i m using the firefox addon in firefox on windows professional please provide any additional information below export as png works as expected in firefox this is probably due to the changes firefox made to addon compatibility the workaround is to use the windows build of pencil original issue | 1 |
443,589 | 12,796,649,795 | IssuesEvent | 2020-07-02 10:50:27 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | [0.9.0 staging-1636] Gigot sleeve shirt icon/model issue | Priority: Medium | So, now Gigot sleeve shirt is looking like that?

But the icon looks different with the vest part | 1.0 | [0.9.0 staging-1636] Gigot sleeve shirt icon/model issue - So, now Gigot sleeve shirt is looking like that?

But the icon looks different with the vest part | priority | gigot sleeve shirt icon model issue so now gigot sleeve shirt is looking like that but the icon looks different with the vest part | 1 |
612,092 | 18,990,667,406 | IssuesEvent | 2021-11-22 06:44:47 | turbot/steampipe-plugin-aws | https://api.github.com/repos/turbot/steampipe-plugin-aws | closed | Add table aws_elasticache_redis_metric_cpu_utilization_daily | enhancement new table priority:medium | **References**
Add any related links that will help us understand the resource, including vendor documentation, related GitHub issues, and Go SDK documentation.
Engine level cloudwatch metric of EngineCPUUtilization can be considered
https://docs.aws.amazon.com/AmazonElastiCache/latest/red-ug/CacheMetrics.Redis.html | 1.0 | Add table aws_elasticache_redis_metric_cpu_utilization_daily - **References**
Add any related links that will help us understand the resource, including vendor documentation, related GitHub issues, and Go SDK documentation.
Engine level cloudwatch metric of EngineCPUUtilization can be considered
https://docs.aws.amazon.com/AmazonElastiCache/latest/red-ug/CacheMetrics.Redis.html | priority | add table aws elasticache redis metric cpu utilization daily references add any related links that will help us understand the resource including vendor documentation related github issues and go sdk documentation engine level cloudwatch metric of enginecpuutilization can be considered | 1 |
800,592 | 28,372,154,628 | IssuesEvent | 2023-04-12 17:50:01 | svthalia/Reaxit | https://api.github.com/repos/svthalia/Reaxit | opened | Double-click unlikes is odd | bug priority: medium | ### Describe the bug
Double-click unlikes is odd
### How to reproduce
Steps to reproduce the behavior:
1. Double click a liked photo
2. Observe
### Expected behaviour
It would like the photo (again), but not unlike. This is unusual, look at instagram
### Screenshots
N/A
### Environment:
N/A
### Additional context
N/A | 1.0 | Double-click unlikes is odd - ### Describe the bug
Double-click unlikes is odd
### How to reproduce
Steps to reproduce the behavior:
1. Double click a liked photo
2. Observe
### Expected behaviour
It would like the photo (again), but not unlike. This is unusual, look at instagram
### Screenshots
N/A
### Environment:
N/A
### Additional context
N/A | priority | double click unlikes is odd describe the bug double click unlikes is odd how to reproduce steps to reproduce the behavior double click a liked photo observe expected behaviour it would like the photo again but not unlike this is unusual look at instagram screenshots n a environment n a additional context n a | 1 |
25,235 | 2,678,312,524 | IssuesEvent | 2015-03-26 09:53:11 | hials/HTML5-GroupProject | https://api.github.com/repos/hials/HTML5-GroupProject | closed | Javascript applications for the "JavaScript Fun" page | medium-priority | We need one or more javascript applications for the "JavaScript Fun" page | 1.0 | Javascript applications for the "JavaScript Fun" page - We need one or more javascript applications for the "JavaScript Fun" page | priority | javascript applications for the javascript fun page we need one or more javascript applications for the javascript fun page | 1 |
692,872 | 23,752,056,104 | IssuesEvent | 2022-08-31 21:44:06 | yugabyte/yugabyte-db | https://api.github.com/repos/yugabyte/yugabyte-db | closed | [YSQL] Strict inequality conditions are pushed down as non-strict inequalities to DocDB | kind/bug area/ysql priority/medium | Jira Link: [DB-1170](https://yugabyte.atlassian.net/browse/DB-1170)
### Description
It seems that if a YSQL user requests a filter of the form `c < 3`, the corresponding request sent to DocDB will have `c <= 3`, causing an extra suite of seeks. The extra rows produced are filtered at the YSQL layer. However, this filtering is not reflected in the `EXPLAIN ANALYZE` output.
Consider the following steps to reproduce this issue:
1) `CREATE TABLE rrr(r1 INT, r2 INT, v INT, PRIMARY KEY(r1 ASC, r2 ASC));`
2) `INSERT INTO rrr SELECT i/10, i%10, i FROM generate_series(1, 100) i;`
3) `set yb_debug_log_docdb_requests=true;`
4) `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 < 2 and r2 = 1;`
5) Observe how the RPC sent by this has a `QL_OP_LESS_THAN_EQUAL` operator in the condition specifier instead of `QL_OP_LESS_THAN`
6) Observe, how `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 <= 2 and r2 = 1;` also takes less time than the query in step 4 as no extra filtering has to be done.
In step 4, note that there were no displayed rows filtered by recheck in the output. You can also verify that rows were indeed filtered by putting a breakpoint on the codepath that returns false over [here](https://github.com/yugabyte/yugabyte-db/blob/0307f302e99737ad6cd0411657e593880bc53c10/src/postgres/src/backend/access/yb_access/yb_scan.c#L1303).
The extra seeks resulting from this become more non-negligible as you add more columns and strict inequality conditions.
| 1.0 | [YSQL] Strict inequality conditions are pushed down as non-strict inequalities to DocDB - Jira Link: [DB-1170](https://yugabyte.atlassian.net/browse/DB-1170)
### Description
It seems that if a YSQL user requests a filter of the form `c < 3`, the corresponding request sent to DocDB will have `c <= 3`, causing an extra suite of seeks. The extra rows produced are filtered at the YSQL layer. However, this filtering is not reflected in the `EXPLAIN ANALYZE` output.
Consider the following steps to reproduce this issue:
1) `CREATE TABLE rrr(r1 INT, r2 INT, v INT, PRIMARY KEY(r1 ASC, r2 ASC));`
2) `INSERT INTO rrr SELECT i/10, i%10, i FROM generate_series(1, 100) i;`
3) `set yb_debug_log_docdb_requests=true;`
4) `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 < 2 and r2 = 1;`
5) Observe how the RPC sent by this has a `QL_OP_LESS_THAN_EQUAL` operator in the condition specifier instead of `QL_OP_LESS_THAN`
6) Observe, how `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 <= 2 and r2 = 1;` also takes less time than the query in step 4 as no extra filtering has to be done.
In step 4, note that there were no displayed rows filtered by recheck in the output. You can also verify that rows were indeed filtered by putting a breakpoint on the codepath that returns false over [here](https://github.com/yugabyte/yugabyte-db/blob/0307f302e99737ad6cd0411657e593880bc53c10/src/postgres/src/backend/access/yb_access/yb_scan.c#L1303).
The extra seeks resulting from this become more non-negligible as you add more columns and strict inequality conditions.
| priority | strict inequality conditions are pushed down as non strict inequalities to docdb jira link description it seems that if a ysql user requests a filter of the form c the corresponding request sent to docdb will have c causing an extra suite of seeks the extra rows produced are filtered at the ysql layer however this filtering is not reflected in the explain analyze output consider the following steps to reproduce this issue create table rrr int int v int primary key asc asc insert into rrr select i i i from generate series i set yb debug log docdb requests true explain analyze select from rrr where and observe how the rpc sent by this has a ql op less than equal operator in the condition specifier instead of ql op less than observe how explain analyze select from rrr where and also takes less time than the query in step as no extra filtering has to be done in step note that there were no displayed rows filtered by recheck in the output you can also verify that rows were indeed filtered by putting a breakpoint on the codepath that returns false over the extra seeks resulting from this become more non negligible as you add more columns and strict inequality conditions | 1 |
270,921 | 8,474,674,241 | IssuesEvent | 2018-10-24 16:46:51 | CS2103-AY1819S1-T13-4/main | https://api.github.com/repos/CS2103-AY1819S1-T13-4/main | opened | Implement Map DS in UniqueModuleList for O(1) look up | priority.Medium severity.Low status.Ongoing type.Enhancement | Improve look up efficiency of UniqueModuleList | 1.0 | Implement Map DS in UniqueModuleList for O(1) look up - Improve look up efficiency of UniqueModuleList | priority | implement map ds in uniquemodulelist for o look up improve look up efficiency of uniquemodulelist | 1 |
454,151 | 13,095,675,985 | IssuesEvent | 2020-08-03 14:29:19 | HabitRPG/habitica | https://api.github.com/repos/HabitRPG/habitica | closed | Modals should open without being scrolled | help wanted priority: medium section: Achievements/Popups/Notifications | Often when long modals pop up, they're scrolled to the very bottom by default, like this:
<img width="533" alt="screen shot 2017-10-27 at 1 40 52 pm" src="https://user-images.githubusercontent.com/3494081/32124586-24485edc-bb1d-11e7-8dfe-abf0f8915283.png">
They should instead pop up with the top of the modal displaying, like this:
<img width="574" alt="screen shot 2017-10-27 at 1 40 55 pm" src="https://user-images.githubusercontent.com/3494081/32124559-0ffbd6e8-bb1d-11e7-90ea-1c8d925c322d.png">
| 1.0 | Modals should open without being scrolled - Often when long modals pop up, they're scrolled to the very bottom by default, like this:
<img width="533" alt="screen shot 2017-10-27 at 1 40 52 pm" src="https://user-images.githubusercontent.com/3494081/32124586-24485edc-bb1d-11e7-8dfe-abf0f8915283.png">
They should instead pop up with the top of the modal displaying, like this:
<img width="574" alt="screen shot 2017-10-27 at 1 40 55 pm" src="https://user-images.githubusercontent.com/3494081/32124559-0ffbd6e8-bb1d-11e7-90ea-1c8d925c322d.png">
| priority | modals should open without being scrolled often when long modals pop up they re scrolled to the very bottom by default like this img width alt screen shot at pm src they should instead pop up with the top of the modal displaying like this img width alt screen shot at pm src | 1 |
651,067 | 21,464,249,123 | IssuesEvent | 2022-04-26 00:49:48 | Dragon1320/anime-notifier | https://api.github.com/repos/Dragon1320/anime-notifier | closed | Establish database connection | priority: medium type: feature module: database | We should be able to establish a connection to MongoDB given some parameters. Credentials should be optional since we probably won't need them when testing locally. You can interpret 'configurable' however you like - be it function arguments or reading a config file. If it is the latter though, I can recommend some useful libraries.
The following should be configurable:
- address
- port
- database
Additionally, the following should be optional:
- username
- password | 1.0 | Establish database connection - We should be able to establish a connection to MongoDB given some parameters. Credentials should be optional since we probably won't need them when testing locally. You can interpret 'configurable' however you like - be it function arguments or reading a config file. If it is the latter though, I can recommend some useful libraries.
The following should be configurable:
- address
- port
- database
Additionally, the following should be optional:
- username
- password | priority | establish database connection we should be able to establish a connection to mongodb given some parameters credentials should be optional since we probably won t need them when testing locally you can interpret configurable however you like be it function arguments or reading a config file if it is the latter though i can recommend some useful libraries the following should be configurable address port database additionally the following should be optional username password | 1 |
693,805 | 23,791,209,389 | IssuesEvent | 2022-09-02 14:42:01 | meisnate12/Plex-Meta-Manager | https://api.github.com/repos/meisnate12/Plex-Meta-Manager | closed | Feature Request: I'd like to see a way to publish various collections on the home screen for a set number of days. | enhancement priority:medium | <!---
Please make sure you submit all Pull Requests to the develop branch not the master branch.
--->
**Is your feature request related to a problem? Please describe.**
Similar to how Netflix has lists you can scroll through, I'd like to do this with Plex so users might be presented with an option they hadn't considered before.
**Describe the solution you'd like**
I'd like to be able to either have a random collection assigned to the screen for a set period of time, or a list of collections that could change on a rotating basis
**Describe alternatives you've considered**
I can make it work on a limited basis using a method similar to the holiday collections that are able to be pinned to the home screen. I have a list of collections and gave them all a date range. My biggest issues are how they are ordered since you can only use custom sorting on one collection, and how to have them change out. Currently I have dates ranges attached to each one.
| 1.0 | Feature Request: I'd like to see a way to publish various collections on the home screen for a set number of days. - <!---
Please make sure you submit all Pull Requests to the develop branch not the master branch.
--->
**Is your feature request related to a problem? Please describe.**
Similar to how Netflix has lists you can scroll through, I'd like to do this with Plex so users might be presented with an option they hadn't considered before.
**Describe the solution you'd like**
I'd like to be able to either have a random collection assigned to the screen for a set period of time, or a list of collections that could change on a rotating basis
**Describe alternatives you've considered**
I can make it work on a limited basis using a method similar to the holiday collections that are able to be pinned to the home screen. I have a list of collections and gave them all a date range. My biggest issues are how they are ordered since you can only use custom sorting on one collection, and how to have them change out. Currently I have dates ranges attached to each one.
| priority | feature request i d like to see a way to publish various collections on the home screen for a set number of days please make sure you submit all pull requests to the develop branch not the master branch is your feature request related to a problem please describe similar to how netflix has lists you can scroll through i d like to do this with plex so users might be presented with an option they hadn t considered before describe the solution you d like i d like to be able to either have a random collection assigned to the screen for a set period of time or a list of collections that could change on a rotating basis describe alternatives you ve considered i can make it work on a limited basis using a method similar to the holiday collections that are able to be pinned to the home screen i have a list of collections and gave them all a date range my biggest issues are how they are ordered since you can only use custom sorting on one collection and how to have them change out currently i have dates ranges attached to each one | 1 |
159,231 | 6,042,453,965 | IssuesEvent | 2017-06-11 13:11:05 | psi-plus/main | https://api.github.com/repos/psi-plus/main | closed | Косметический фикс на кликабельное фото из vCard. | OpSys-All Priority-Medium Type-Enhancement Usability | Original [issue 12](https://code.google.com/p/psi-dev/issues/detail?id=12) created by psi-plus on 2008-12-24T09:00:58.000Z:
1. Запрещает изменять размер окна с фото.
2. Открывает единственное окно.
3. Окно типа ToolWindow.
| 1.0 | Косметический фикс на кликабельное фото из vCard. - Original [issue 12](https://code.google.com/p/psi-dev/issues/detail?id=12) created by psi-plus on 2008-12-24T09:00:58.000Z:
1. Запрещает изменять размер окна с фото.
2. Открывает единственное окно.
3. Окно типа ToolWindow.
| priority | косметический фикс на кликабельное фото из vcard original created by psi plus on запрещает изменять размер окна с фото открывает единственное окно окно типа toolwindow | 1 |
462,468 | 13,247,788,889 | IssuesEvent | 2020-08-19 17:51:02 | radical-cybertools/radical.pilot | https://api.github.com/repos/radical-cybertools/radical.pilot | closed | Configurable SMT level | comp:agent:executor comp:pmgr:launcher layer:rp layer:saga priority:medium topic:resource type:enhancement | https://github.com/radical-cybertools/radical.saga/blob/devel/src/radical/saga/adaptors/lsf/lsfjob.py#L38
`SMT` can be adjusted by a user in some way by either resource config, parameters, or ENV variables (e.g. `export RADICAL_SAGA_SMT`).
I was in a conversation and this was raised if a user can specify different SMT level instead of fixed SMT4. | 1.0 | Configurable SMT level - https://github.com/radical-cybertools/radical.saga/blob/devel/src/radical/saga/adaptors/lsf/lsfjob.py#L38
`SMT` can be adjusted by a user in some way by either resource config, parameters, or ENV variables (e.g. `export RADICAL_SAGA_SMT`).
I was in a conversation and this was raised if a user can specify different SMT level instead of fixed SMT4. | priority | configurable smt level smt can be adjusted by a user in some way by either resource config parameters or env variables e g export radical saga smt i was in a conversation and this was raised if a user can specify different smt level instead of fixed | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.