Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
957
labels
stringlengths
4
795
body
stringlengths
1
259k
index
stringclasses
12 values
text_combine
stringlengths
96
259k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
617,764
19,404,167,081
IssuesEvent
2021-12-19 18:06:24
codidact/qpixel
https://api.github.com/repos/codidact/qpixel
closed
Preview renders unsupported HTML tags, but final post (correctly) doesn't
area: html/css/js type: bug priority: medium complexity: unassessed
https://meta.codidact.com/posts/284503 https://meta.codidact.com/posts/284505 When composing a post, an unsupported HTML tag (`div` and `kbd` in these reports) was accepted and rendered. However, after submission the post doesn't render it, which is correct because we don't support the tag. This is confusing to users who haven't memorized (or don't look up) which HTML tags we do/don't support. I know there's at least one other issue about differences between preview and final rendering, though I couldn't find it. We use different libraries in the two cases so differences aren't surprising. If we can't use the same library (or logic) to render the Markdown in both cases, is there anything we can do to provide some feedback when editing? Can we "lint" the post body and indicate if we found something? Maybe, as with missing alt text, that could be something we do when the user clicks the "post" button, so it doesn't have to be a performance drain.
1.0
Preview renders unsupported HTML tags, but final post (correctly) doesn't - https://meta.codidact.com/posts/284503 https://meta.codidact.com/posts/284505 When composing a post, an unsupported HTML tag (`div` and `kbd` in these reports) was accepted and rendered. However, after submission the post doesn't render it, which is correct because we don't support the tag. This is confusing to users who haven't memorized (or don't look up) which HTML tags we do/don't support. I know there's at least one other issue about differences between preview and final rendering, though I couldn't find it. We use different libraries in the two cases so differences aren't surprising. If we can't use the same library (or logic) to render the Markdown in both cases, is there anything we can do to provide some feedback when editing? Can we "lint" the post body and indicate if we found something? Maybe, as with missing alt text, that could be something we do when the user clicks the "post" button, so it doesn't have to be a performance drain.
priority
preview renders unsupported html tags but final post correctly doesn t when composing a post an unsupported html tag div and kbd in these reports was accepted and rendered however after submission the post doesn t render it which is correct because we don t support the tag this is confusing to users who haven t memorized or don t look up which html tags we do don t support i know there s at least one other issue about differences between preview and final rendering though i couldn t find it we use different libraries in the two cases so differences aren t surprising if we can t use the same library or logic to render the markdown in both cases is there anything we can do to provide some feedback when editing can we lint the post body and indicate if we found something maybe as with missing alt text that could be something we do when the user clicks the post button so it doesn t have to be a performance drain
1
567,309
16,854,480,669
IssuesEvent
2021-06-21 03:26:56
adirh3/Fluent-Search
https://api.github.com/repos/adirh3/Fluent-Search
closed
Change size in MB to look cleaner
Medium Priority UI/UX bug
This is no-brainer and standard interface kilobyte, megabyte, gigabyte, terabyte instead of "Size of MB 0.03891485883838". This looks very hard to read for all users, novice and expert alike. ![image](https://user-images.githubusercontent.com/1836362/120909338-ebae6b00-c628-11eb-9dee-44ab14bcd2c6.png)
1.0
Change size in MB to look cleaner - This is no-brainer and standard interface kilobyte, megabyte, gigabyte, terabyte instead of "Size of MB 0.03891485883838". This looks very hard to read for all users, novice and expert alike. ![image](https://user-images.githubusercontent.com/1836362/120909338-ebae6b00-c628-11eb-9dee-44ab14bcd2c6.png)
priority
change size in mb to look cleaner this is no brainer and standard interface kilobyte megabyte gigabyte terabyte instead of size of mb this looks very hard to read for all users novice and expert alike
1
627,412
19,904,337,359
IssuesEvent
2022-01-25 11:10:40
docker-mailserver/docker-mailserver
https://api.github.com/repos/docker-mailserver/docker-mailserver
opened
Can't create email accounts while enabling LDAP
kind/bug meta/needs triage priority/medium
### Miscellaneous first checks - [X] I checked that all ports are open and not blocked by my ISP / hosting provider. - [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct. ### Affected Component(s) Mail creation, deletion and listing ### What happened and when does this occur? ```Markdown While trying to create the email account using the setup.sh; I'm encountering the error below. Waiting for dovecot to create /var/mail/deol.com/deoltito... Waiting for dovecot to create /var/mail/deol.com/deoltito... Waiting for dovecot to create /var/mail/deol.com/deoltito... Waiting for dovecot to create /var/mail/deol.com/deoltito... This message goes on until I stops it. However the email account won't be created. When I try to list the available email accounts, this particular error is seen under it. ========== # ./setup.sh email list Fatal: Unknown command 'quota', but plugin quota exists. Try to set mail_plugins=quota /usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ") /usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ") * new@deol.com ( / ) [%] ========== ``` ### What did you expect to happen? ```Markdown I believe it has something to with the LDAP integration. When I disable LDAP in the compose file, the email accounts can be created, listed and deleted without any issues. ``` ### How do we replicate the issue? ```Markdown 1. Try to create a compose file with LDAP enabled and integrated in it 2. Try creating an email account after that 3. Try to list or delete the email accounts too ... ``` ### DMS version v10.4.0 ### What operating system is DMS running on? Linux ### What instruction set architecture is DMS running on? x86_64 / AMD64 ### What container orchestration tool are you using? Docker Compose ### docker-compose.yml ```yaml version: '3.8' services: mailserver: image: docker.io/mailserver/docker-mailserver:latest container_name: mailserver hostname: mail domainname: deol.com ports: - "25:25" - "143:143" - "587:587" - "993:993" volumes: - ./docker-data/dms/mail-data/:/var/mail/ - ./docker-data/dms/mail-state/:/var/mail-state/ - ./docker-data/dms/mail-logs/:/var/log/mail/ - ./docker-data/dms/config/:/tmp/docker-mailserver/ - /etc/localtime:/etc/localtime:ro environment: - ENABLE_SPAMASSASSIN=1 - SPAMASSASSIN_SPAM_TO_INBOX=1 - ENABLE_CLAMAV=1 - ENABLE_FAIL2BAN=1 - ENABLE_POSTGREY=1 - ENABLE_SASLAUTHD=1 - ONE_DIR=1 - DMS_DEBUG=1 - ENABLE_LDAP=1 - LDAP_SERVER_HOST=LDAPNEW # your ldap container/IP/ServerName - LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com - LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com - LDAP_BIND_PW= - ENABLE_SASLAUTHD=1 - SASLAUTHD_MECHANISMS=ldap - SASLAUTHD_LDAP_SERVER=LDAPNEW - SASLAUTHD_LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com - SASLAUTHD_LDAP_PASSWORD= - SASLAUTHD_LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com - SASLAUTHD_LDAP_FILTER=(&(objectClass=PostfixBookMailAccount)(uniqueIdentifier=%U)) - POSTMASTER_ADDRESS=postmaster@deol.com - POSTFIX_MESSAGE_SIZE_LIMIT=100000000 cap_add: - NET_ADMIN - SYS_PTRACE ``` ### Relevant log output _No response_ ### Other relevant information _No response_ ### What level of experience do you have with Docker and mail servers? - [ ] I am inexperienced with docker - [ ] I am inexperienced with mail servers - [ ] I am uncomfortable with the CLI ### Code of conduct - [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree - [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution ### Improvements to this form? _No response_
1.0
Can't create email accounts while enabling LDAP - ### Miscellaneous first checks - [X] I checked that all ports are open and not blocked by my ISP / hosting provider. - [X] I know that SSL errors are likely the result of a wrong setup on the user side and not caused by DMS itself. I'm confident my setup is correct. ### Affected Component(s) Mail creation, deletion and listing ### What happened and when does this occur? ```Markdown While trying to create the email account using the setup.sh; I'm encountering the error below. Waiting for dovecot to create /var/mail/deol.com/deoltito... Waiting for dovecot to create /var/mail/deol.com/deoltito... Waiting for dovecot to create /var/mail/deol.com/deoltito... Waiting for dovecot to create /var/mail/deol.com/deoltito... This message goes on until I stops it. However the email account won't be created. When I try to list the available email accounts, this particular error is seen under it. ========== # ./setup.sh email list Fatal: Unknown command 'quota', but plugin quota exists. Try to set mail_plugins=quota /usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ") /usr/local/bin/listmailuser: line 15: 1024 * : syntax error: operand expected (error token is "* ") * new@deol.com ( / ) [%] ========== ``` ### What did you expect to happen? ```Markdown I believe it has something to with the LDAP integration. When I disable LDAP in the compose file, the email accounts can be created, listed and deleted without any issues. ``` ### How do we replicate the issue? ```Markdown 1. Try to create a compose file with LDAP enabled and integrated in it 2. Try creating an email account after that 3. Try to list or delete the email accounts too ... ``` ### DMS version v10.4.0 ### What operating system is DMS running on? Linux ### What instruction set architecture is DMS running on? x86_64 / AMD64 ### What container orchestration tool are you using? Docker Compose ### docker-compose.yml ```yaml version: '3.8' services: mailserver: image: docker.io/mailserver/docker-mailserver:latest container_name: mailserver hostname: mail domainname: deol.com ports: - "25:25" - "143:143" - "587:587" - "993:993" volumes: - ./docker-data/dms/mail-data/:/var/mail/ - ./docker-data/dms/mail-state/:/var/mail-state/ - ./docker-data/dms/mail-logs/:/var/log/mail/ - ./docker-data/dms/config/:/tmp/docker-mailserver/ - /etc/localtime:/etc/localtime:ro environment: - ENABLE_SPAMASSASSIN=1 - SPAMASSASSIN_SPAM_TO_INBOX=1 - ENABLE_CLAMAV=1 - ENABLE_FAIL2BAN=1 - ENABLE_POSTGREY=1 - ENABLE_SASLAUTHD=1 - ONE_DIR=1 - DMS_DEBUG=1 - ENABLE_LDAP=1 - LDAP_SERVER_HOST=LDAPNEW # your ldap container/IP/ServerName - LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com - LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com - LDAP_BIND_PW= - ENABLE_SASLAUTHD=1 - SASLAUTHD_MECHANISMS=ldap - SASLAUTHD_LDAP_SERVER=LDAPNEW - SASLAUTHD_LDAP_BIND_DN=cn=admin,dc=ds,dc=domain,dc=com - SASLAUTHD_LDAP_PASSWORD= - SASLAUTHD_LDAP_SEARCH_BASE=ou=people,dc=ds,dc=domain,dc=com - SASLAUTHD_LDAP_FILTER=(&(objectClass=PostfixBookMailAccount)(uniqueIdentifier=%U)) - POSTMASTER_ADDRESS=postmaster@deol.com - POSTFIX_MESSAGE_SIZE_LIMIT=100000000 cap_add: - NET_ADMIN - SYS_PTRACE ``` ### Relevant log output _No response_ ### Other relevant information _No response_ ### What level of experience do you have with Docker and mail servers? - [ ] I am inexperienced with docker - [ ] I am inexperienced with mail servers - [ ] I am uncomfortable with the CLI ### Code of conduct - [X] I have read this project's [Code of Conduct](https://github.com/docker-mailserver/docker-mailserver/blob/master/CODE_OF_CONDUCT.md) and I agree - [X] I have read the [README](https://github.com/docker-mailserver/docker-mailserver/blob/master/README.md) and the [documentation](https://docker-mailserver.github.io/docker-mailserver/edge/) and I searched the [issue tracker](https://github.com/docker-mailserver/docker-mailserver/issues?q=is%3Aissue) but could not find a solution ### Improvements to this form? _No response_
priority
can t create email accounts while enabling ldap miscellaneous first checks i checked that all ports are open and not blocked by my isp hosting provider i know that ssl errors are likely the result of a wrong setup on the user side and not caused by dms itself i m confident my setup is correct affected component s mail creation deletion and listing what happened and when does this occur markdown while trying to create the email account using the setup sh i m encountering the error below waiting for dovecot to create var mail deol com deoltito waiting for dovecot to create var mail deol com deoltito waiting for dovecot to create var mail deol com deoltito waiting for dovecot to create var mail deol com deoltito this message goes on until i stops it however the email account won t be created when i try to list the available email accounts this particular error is seen under it setup sh email list fatal unknown command quota but plugin quota exists try to set mail plugins quota usr local bin listmailuser line syntax error operand expected error token is usr local bin listmailuser line syntax error operand expected error token is new deol com what did you expect to happen markdown i believe it has something to with the ldap integration when i disable ldap in the compose file the email accounts can be created listed and deleted without any issues how do we replicate the issue markdown try to create a compose file with ldap enabled and integrated in it try creating an email account after that try to list or delete the email accounts too dms version what operating system is dms running on linux what instruction set architecture is dms running on what container orchestration tool are you using docker compose docker compose yml yaml version services mailserver image docker io mailserver docker mailserver latest container name mailserver hostname mail domainname deol com ports volumes docker data dms mail data var mail docker data dms mail state var mail state docker data dms mail logs var log mail docker data dms config tmp docker mailserver etc localtime etc localtime ro environment enable spamassassin spamassassin spam to inbox enable clamav enable enable postgrey enable saslauthd one dir dms debug enable ldap ldap server host ldapnew your ldap container ip servername ldap search base ou people dc ds dc domain dc com ldap bind dn cn admin dc ds dc domain dc com ldap bind pw enable saslauthd saslauthd mechanisms ldap saslauthd ldap server ldapnew saslauthd ldap bind dn cn admin dc ds dc domain dc com saslauthd ldap password saslauthd ldap search base ou people dc ds dc domain dc com saslauthd ldap filter objectclass postfixbookmailaccount uniqueidentifier u postmaster address postmaster deol com postfix message size limit cap add net admin sys ptrace relevant log output no response other relevant information no response what level of experience do you have with docker and mail servers i am inexperienced with docker i am inexperienced with mail servers i am uncomfortable with the cli code of conduct i have read this project s and i agree i have read the and the and i searched the but could not find a solution improvements to this form no response
1
666,300
22,349,529,677
IssuesEvent
2022-06-15 10:45:34
SAP/xsk
https://api.github.com/repos/SAP/xsk
closed
[IDE] CSV Editor - Total Records Count
enhancement priority-low effort-medium usability tooling shadow
It would be great if the CSV editor is displaying the total records count somewhere: **Sample rows count:** ![sample-csv-data](https://user-images.githubusercontent.com/4092083/151515240-a714938c-1ea1-4507-b8ce-c62bf64c6eac.png) **CSV Editor:** ![csv-editor](https://user-images.githubusercontent.com/4092083/151515549-2131e8f9-9ddc-4560-9f0c-f2d7e3affa89.png)
1.0
[IDE] CSV Editor - Total Records Count - It would be great if the CSV editor is displaying the total records count somewhere: **Sample rows count:** ![sample-csv-data](https://user-images.githubusercontent.com/4092083/151515240-a714938c-1ea1-4507-b8ce-c62bf64c6eac.png) **CSV Editor:** ![csv-editor](https://user-images.githubusercontent.com/4092083/151515549-2131e8f9-9ddc-4560-9f0c-f2d7e3affa89.png)
priority
csv editor total records count it would be great if the csv editor is displaying the total records count somewhere sample rows count csv editor
1
92,675
3,872,900,201
IssuesEvent
2016-04-11 15:15:54
jcgregorio/httplib2
https://api.github.com/repos/jcgregorio/httplib2
closed
bdist_rpm fails
bug imported Priority-Medium
_From [brian.la...@gmail.com](https://code.google.com/u/108532649133345963591/) on December 29, 2009 16:24:09_ What steps will reproduce the problem? 1. python setup.py bdist_rpm What is the expected output? What do you see instead? EXPECTED:: + python setup.py build running build running build_py creating build creating build/lib creating build/lib/httplib2 copying httplib2/__init__.py -> build/lib/httplib2 copying httplib2/iri2uri.py -> build/lib/httplib2 + exit 0 ACTUAL RESULT:: + python setup.py build running build running build_py error: package directory 'python2/httplib2' does not exist error: Bad exit status from /var/tmp/rpm-tmp.50855 (%build) What version of the product are you using? On what operating system? Latest check-in on any RPM based distro (Fedora/RHEL/Centos etc.) Please provide any additional information below. I believe that the RPM isn't being generated correctly with the new python2 and python3 subdirectories. I'm not a huge user of python distutils...but I'll try to look and see if I can fix. _Original issue: http://code.google.com/p/httplib2/issues/detail?id=85_
1.0
bdist_rpm fails - _From [brian.la...@gmail.com](https://code.google.com/u/108532649133345963591/) on December 29, 2009 16:24:09_ What steps will reproduce the problem? 1. python setup.py bdist_rpm What is the expected output? What do you see instead? EXPECTED:: + python setup.py build running build running build_py creating build creating build/lib creating build/lib/httplib2 copying httplib2/__init__.py -> build/lib/httplib2 copying httplib2/iri2uri.py -> build/lib/httplib2 + exit 0 ACTUAL RESULT:: + python setup.py build running build running build_py error: package directory 'python2/httplib2' does not exist error: Bad exit status from /var/tmp/rpm-tmp.50855 (%build) What version of the product are you using? On what operating system? Latest check-in on any RPM based distro (Fedora/RHEL/Centos etc.) Please provide any additional information below. I believe that the RPM isn't being generated correctly with the new python2 and python3 subdirectories. I'm not a huge user of python distutils...but I'll try to look and see if I can fix. _Original issue: http://code.google.com/p/httplib2/issues/detail?id=85_
priority
bdist rpm fails from on december what steps will reproduce the problem python setup py bdist rpm what is the expected output what do you see instead expected python setup py build running build running build py creating build creating build lib creating build lib copying init py build lib copying py build lib exit actual result python setup py build running build running build py error package directory does not exist error bad exit status from var tmp rpm tmp build what version of the product are you using on what operating system latest check in on any rpm based distro fedora rhel centos etc please provide any additional information below i believe that the rpm isn t being generated correctly with the new and subdirectories i m not a huge user of python distutils but i ll try to look and see if i can fix original issue
1
171,596
6,491,685,465
IssuesEvent
2017-08-21 10:41:07
softdevteam/krun
https://api.github.com/repos/softdevteam/krun
opened
Move masking of the core cycle counter outside the time section.
enhancement medium priority (a clear improvement but not a blocker for publication)
This loop: https://github.com/softdevteam/krun/blob/9c38f01a7e6292669d7c9f83780d44e2abae8279/libkrun/libkruntime.c#L401 Can be moved into the getter functions, thus moving the masking operations out of the timed section.
1.0
Move masking of the core cycle counter outside the time section. - This loop: https://github.com/softdevteam/krun/blob/9c38f01a7e6292669d7c9f83780d44e2abae8279/libkrun/libkruntime.c#L401 Can be moved into the getter functions, thus moving the masking operations out of the timed section.
priority
move masking of the core cycle counter outside the time section this loop can be moved into the getter functions thus moving the masking operations out of the timed section
1
11,563
2,610,142,072
IssuesEvent
2015-02-26 18:44:46
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
seed line should NOT be visible right away
auto-migrated Priority-Medium Type-Enhancement
``` What steps will reproduce the problem? 1. latest revision offers a neat functionality, the possibility of editing the seed line 2. this functionality is useful for testing and has been requested by users 3. however despite being in the game configuration page, it should not be visible right away for not always the user needs to change that line and has the high likelihood of scaring the user with too much technicality What is the expected output? What do you see instead? in my view there should be a nice button that when pushed either shows a popup or simply makes visible that line ``` ----- Original issue reported on code.google.com by `vittorio...@gmail.com` on 20 Dec 2010 at 12:02 * Blocking: #115
1.0
seed line should NOT be visible right away - ``` What steps will reproduce the problem? 1. latest revision offers a neat functionality, the possibility of editing the seed line 2. this functionality is useful for testing and has been requested by users 3. however despite being in the game configuration page, it should not be visible right away for not always the user needs to change that line and has the high likelihood of scaring the user with too much technicality What is the expected output? What do you see instead? in my view there should be a nice button that when pushed either shows a popup or simply makes visible that line ``` ----- Original issue reported on code.google.com by `vittorio...@gmail.com` on 20 Dec 2010 at 12:02 * Blocking: #115
priority
seed line should not be visible right away what steps will reproduce the problem latest revision offers a neat functionality the possibility of editing the seed line this functionality is useful for testing and has been requested by users however despite being in the game configuration page it should not be visible right away for not always the user needs to change that line and has the high likelihood of scaring the user with too much technicality what is the expected output what do you see instead in my view there should be a nice button that when pushed either shows a popup or simply makes visible that line original issue reported on code google com by vittorio gmail com on dec at blocking
1
70,631
3,332,949,470
IssuesEvent
2015-11-11 22:28:04
angular/material
https://api.github.com/repos/angular/material
closed
menuContents has zero items
priority: medium
In menuDirective.js, the link function call mdMenuCtrl.init. if the md-menu-item is hard coded, like in the case of demoBasicUsage, then in menuController.js this will return items : 'menuContainer[0].querySelectorAll('md-menu-item')'. if the md-menu-item is not hard coded, like in the case of demoMenuPositionModes and demoMenuWidth, then in menuController.js this will return zero items: 'menuContainer[0].querySelectorAll('md-menu-item')'. ```javascript function link(scope, element, attrs, ctrls) { var mdMenuCtrl = ctrls[0]; var isInMenuBar = ctrls[1] != undefined; // Move everything into a md-menu-container and pass it to the controller var menuContainer = angular.element( '<div class="md-open-menu-container md-whiteframe-z2"></div>' ); var menuContents = element.children()[1]; menuContainer.append(menuContents); if (isInMenuBar) { element.append(menuContainer); menuContainer[0].style.display = 'none'; } mdMenuCtrl.init(menuContainer, { isInMenuBar: isInMenuBar }); scope.$on('$destroy', function() { mdMenuCtrl .destroy() .finally(function(){ menuContainer.remove(); }); }); } ```
1.0
menuContents has zero items - In menuDirective.js, the link function call mdMenuCtrl.init. if the md-menu-item is hard coded, like in the case of demoBasicUsage, then in menuController.js this will return items : 'menuContainer[0].querySelectorAll('md-menu-item')'. if the md-menu-item is not hard coded, like in the case of demoMenuPositionModes and demoMenuWidth, then in menuController.js this will return zero items: 'menuContainer[0].querySelectorAll('md-menu-item')'. ```javascript function link(scope, element, attrs, ctrls) { var mdMenuCtrl = ctrls[0]; var isInMenuBar = ctrls[1] != undefined; // Move everything into a md-menu-container and pass it to the controller var menuContainer = angular.element( '<div class="md-open-menu-container md-whiteframe-z2"></div>' ); var menuContents = element.children()[1]; menuContainer.append(menuContents); if (isInMenuBar) { element.append(menuContainer); menuContainer[0].style.display = 'none'; } mdMenuCtrl.init(menuContainer, { isInMenuBar: isInMenuBar }); scope.$on('$destroy', function() { mdMenuCtrl .destroy() .finally(function(){ menuContainer.remove(); }); }); } ```
priority
menucontents has zero items in menudirective js the link function call mdmenuctrl init if the md menu item is hard coded like in the case of demobasicusage then in menucontroller js this will return items menucontainer queryselectorall md menu item if the md menu item is not hard coded like in the case of demomenupositionmodes and demomenuwidth then in menucontroller js this will return zero items menucontainer queryselectorall md menu item javascript function link scope element attrs ctrls var mdmenuctrl ctrls var isinmenubar ctrls undefined move everything into a md menu container and pass it to the controller var menucontainer angular element var menucontents element children menucontainer append menucontents if isinmenubar element append menucontainer menucontainer style display none mdmenuctrl init menucontainer isinmenubar isinmenubar scope on destroy function mdmenuctrl destroy finally function menucontainer remove
1
462,654
13,250,971,352
IssuesEvent
2020-08-20 00:41:55
TB-Modeling/modeltb.org
https://api.github.com/repos/TB-Modeling/modeltb.org
opened
Professional downloadable briefs
Priority 2/3: Medium new feature new idea
## Description We can host professional-looking briefs on TB and other topics.
1.0
Professional downloadable briefs - ## Description We can host professional-looking briefs on TB and other topics.
priority
professional downloadable briefs description we can host professional looking briefs on tb and other topics
1
273,401
8,530,257,133
IssuesEvent
2018-11-03 20:34:54
minio/minio
https://api.github.com/repos/minio/minio
closed
Crashing with Out of Memory Errors (refs 6164)
community priority: medium
I hate to be the bearer of bad news, @harshavardhana , but the fix has not been working for me. I've upgraded to all the most recent Minio.exe builds since the last post in this issue, but with no success. I've delayed responding here in case the fix was in a build later than I thought, and because I found an issue in my environment and wondered if it was related. I've determined it is not but will outline what I've found in the past month. I noticed as time went on that Minio started reporting read errors on my V: drive, but not any of the others. Separately, I discovered while trying to copy a set of large files from my V: drive (completely unrelated to Minio) that there was at least one corrupted file in the mix. This led me to believe my V drive was experiencing data corruption, and I wondered if it was the culprit of problems all along. I replaced both my V and X drives. Because they're identical and were purchased at the same time, I replaced both just in case the X drive was soon to follow. Two brand-new 6 TB drives are in their place. This allowed me a chance to test Minio's healing capabilities, which so far are working OK. However, I've not been able to fully test it (I've only healed about half my data set so far) because Minio continues to crash with out of memory errors after 6-7 hours of run time. I brainstormed that if this out of memory error cannot be fixed I could heal in two stages to work-around the bug. First the V and X drives, which should heal their data in this "new" format. Then delete the Minio data on W and Y and heal back to those drives. Thus getting all my data into the new format in a very round-about (and dangerous) fashion. But like I said, because Minio continues to crash with out of memory exceptions during the heal, I've yet to accomplish this in full. My previous screenshot shows my Minio data at 294 GB, and currently I've only healed about 154 GB. The healed amount tends to change unpredictably. Sometimes it will run for hours and appear to make 0 progress (the V and X data directories are not larger), and other times they will grow by several GB. It's almost as if the process is starting over every single time and it only makes progress if it happens to run longer than last time. I'm not sure, though. Further, even without the heal running, the W and Y drives continue to be hammered while Minio consumes more and more memory for several hours before crashing. This is puzzling to me, because with the number of times I've run Minio it has to have a total run time of at least 50 hours, maybe much more. Assuming a read/write speed of just 10 MB a second, which is very slow for a SATA drive, even platter-based, it would have been able to read/write 1.8 TB of data from each drive. Even if speed were half that at 5 MB/s, that's still 900 GB. Half that, 2.5 MB/s would be 450 GB in that time period. Given my total data set is ~300 GB, I am extremely puzzled as to why it's not done yet unless it is restarting progress every single time. Which isn't logical given that no software would try to complete 300 GB worth of work before writing out progress. That wouldn't make sense. Can you shed any light into what's going on here? My last resort is to simply delete everything and restart with no data. That's not the end of the world because my most important backup copy is off-site. Having stuff locally is for a more convenient full-copy. I've attached CPU and MEM logs that I gathered yesterday, letting it run for a few hours and _not_ healing, just doing it's behind-the-scenes work. Again, it slammed usage of my W and Y drives, as well as the CPU. I stopped it before a crash to get the log output. The CPU log is shorter than the mem one due to my schedule, but it still has a couple hours. Thank you. [mem_20180918.txt](https://github.com/minio/minio/files/2397728/mem_20180918.txt) [cpu_20180918.txt](https://github.com/minio/minio/files/2397729/cpu_20180918.txt) _Originally posted by @tylerforsythe in https://github.com/minio/minio/issue_comments#issuecomment-422851849_
1.0
Crashing with Out of Memory Errors (refs 6164) - I hate to be the bearer of bad news, @harshavardhana , but the fix has not been working for me. I've upgraded to all the most recent Minio.exe builds since the last post in this issue, but with no success. I've delayed responding here in case the fix was in a build later than I thought, and because I found an issue in my environment and wondered if it was related. I've determined it is not but will outline what I've found in the past month. I noticed as time went on that Minio started reporting read errors on my V: drive, but not any of the others. Separately, I discovered while trying to copy a set of large files from my V: drive (completely unrelated to Minio) that there was at least one corrupted file in the mix. This led me to believe my V drive was experiencing data corruption, and I wondered if it was the culprit of problems all along. I replaced both my V and X drives. Because they're identical and were purchased at the same time, I replaced both just in case the X drive was soon to follow. Two brand-new 6 TB drives are in their place. This allowed me a chance to test Minio's healing capabilities, which so far are working OK. However, I've not been able to fully test it (I've only healed about half my data set so far) because Minio continues to crash with out of memory errors after 6-7 hours of run time. I brainstormed that if this out of memory error cannot be fixed I could heal in two stages to work-around the bug. First the V and X drives, which should heal their data in this "new" format. Then delete the Minio data on W and Y and heal back to those drives. Thus getting all my data into the new format in a very round-about (and dangerous) fashion. But like I said, because Minio continues to crash with out of memory exceptions during the heal, I've yet to accomplish this in full. My previous screenshot shows my Minio data at 294 GB, and currently I've only healed about 154 GB. The healed amount tends to change unpredictably. Sometimes it will run for hours and appear to make 0 progress (the V and X data directories are not larger), and other times they will grow by several GB. It's almost as if the process is starting over every single time and it only makes progress if it happens to run longer than last time. I'm not sure, though. Further, even without the heal running, the W and Y drives continue to be hammered while Minio consumes more and more memory for several hours before crashing. This is puzzling to me, because with the number of times I've run Minio it has to have a total run time of at least 50 hours, maybe much more. Assuming a read/write speed of just 10 MB a second, which is very slow for a SATA drive, even platter-based, it would have been able to read/write 1.8 TB of data from each drive. Even if speed were half that at 5 MB/s, that's still 900 GB. Half that, 2.5 MB/s would be 450 GB in that time period. Given my total data set is ~300 GB, I am extremely puzzled as to why it's not done yet unless it is restarting progress every single time. Which isn't logical given that no software would try to complete 300 GB worth of work before writing out progress. That wouldn't make sense. Can you shed any light into what's going on here? My last resort is to simply delete everything and restart with no data. That's not the end of the world because my most important backup copy is off-site. Having stuff locally is for a more convenient full-copy. I've attached CPU and MEM logs that I gathered yesterday, letting it run for a few hours and _not_ healing, just doing it's behind-the-scenes work. Again, it slammed usage of my W and Y drives, as well as the CPU. I stopped it before a crash to get the log output. The CPU log is shorter than the mem one due to my schedule, but it still has a couple hours. Thank you. [mem_20180918.txt](https://github.com/minio/minio/files/2397728/mem_20180918.txt) [cpu_20180918.txt](https://github.com/minio/minio/files/2397729/cpu_20180918.txt) _Originally posted by @tylerforsythe in https://github.com/minio/minio/issue_comments#issuecomment-422851849_
priority
crashing with out of memory errors refs i hate to be the bearer of bad news harshavardhana but the fix has not been working for me i ve upgraded to all the most recent minio exe builds since the last post in this issue but with no success i ve delayed responding here in case the fix was in a build later than i thought and because i found an issue in my environment and wondered if it was related i ve determined it is not but will outline what i ve found in the past month i noticed as time went on that minio started reporting read errors on my v drive but not any of the others separately i discovered while trying to copy a set of large files from my v drive completely unrelated to minio that there was at least one corrupted file in the mix this led me to believe my v drive was experiencing data corruption and i wondered if it was the culprit of problems all along i replaced both my v and x drives because they re identical and were purchased at the same time i replaced both just in case the x drive was soon to follow two brand new tb drives are in their place this allowed me a chance to test minio s healing capabilities which so far are working ok however i ve not been able to fully test it i ve only healed about half my data set so far because minio continues to crash with out of memory errors after hours of run time i brainstormed that if this out of memory error cannot be fixed i could heal in two stages to work around the bug first the v and x drives which should heal their data in this new format then delete the minio data on w and y and heal back to those drives thus getting all my data into the new format in a very round about and dangerous fashion but like i said because minio continues to crash with out of memory exceptions during the heal i ve yet to accomplish this in full my previous screenshot shows my minio data at gb and currently i ve only healed about gb the healed amount tends to change unpredictably sometimes it will run for hours and appear to make progress the v and x data directories are not larger and other times they will grow by several gb it s almost as if the process is starting over every single time and it only makes progress if it happens to run longer than last time i m not sure though further even without the heal running the w and y drives continue to be hammered while minio consumes more and more memory for several hours before crashing this is puzzling to me because with the number of times i ve run minio it has to have a total run time of at least hours maybe much more assuming a read write speed of just mb a second which is very slow for a sata drive even platter based it would have been able to read write tb of data from each drive even if speed were half that at mb s that s still gb half that mb s would be gb in that time period given my total data set is gb i am extremely puzzled as to why it s not done yet unless it is restarting progress every single time which isn t logical given that no software would try to complete gb worth of work before writing out progress that wouldn t make sense can you shed any light into what s going on here my last resort is to simply delete everything and restart with no data that s not the end of the world because my most important backup copy is off site having stuff locally is for a more convenient full copy i ve attached cpu and mem logs that i gathered yesterday letting it run for a few hours and not healing just doing it s behind the scenes work again it slammed usage of my w and y drives as well as the cpu i stopped it before a crash to get the log output the cpu log is shorter than the mem one due to my schedule but it still has a couple hours thank you originally posted by tylerforsythe in
1
604,574
18,714,856,193
IssuesEvent
2021-11-03 02:12:14
AY2122S1-CS2103T-T12-3/tp
https://api.github.com/repos/AY2122S1-CS2103T-T12-3/tp
closed
[PE-D] UG TOC
priority.Medium
![image.png](https://raw.githubusercontent.com/Timothyoung97/ped/main/files/fb45b94c-ffbd-4426-9544-f497776e2da8.png) I think for your sectioning of TOC, The `feature` should be an overall header for all the subsections. Something like below - Features - Contact Features - Appointment Features - Other Features <!--session: 1635494624855-6c4169c8-8f98-434c-bb03-2bfcaad4f18c--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: Timothyoung97/ped#2
1.0
[PE-D] UG TOC - ![image.png](https://raw.githubusercontent.com/Timothyoung97/ped/main/files/fb45b94c-ffbd-4426-9544-f497776e2da8.png) I think for your sectioning of TOC, The `feature` should be an overall header for all the subsections. Something like below - Features - Contact Features - Appointment Features - Other Features <!--session: 1635494624855-6c4169c8-8f98-434c-bb03-2bfcaad4f18c--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: Timothyoung97/ped#2
priority
ug toc i think for your sectioning of toc the feature should be an overall header for all the subsections something like below features contact features appointment features other features labels severity low type documentationbug original ped
1
727,944
25,060,654,219
IssuesEvent
2022-11-07 00:59:44
AY2223S1-CS2103T-W08-2/tp
https://api.github.com/repos/AY2223S1-CS2103T-W08-2/tp
closed
[PE-D][Tester B] find command duplicates person
priority.High priority.Medium
Using the default data provided, running 'find Irfan` displays Irfan's details correclty. But then running `find Irfan Ibrahim` causes does nothing. Then run `find Irfan`, Irfan is duplicated in the list. To recover, the app must be reopened. There might be a deeper issue here where clients are randomly getting duplicated in the model. <!--session: 1666944033620-ae8585f7-7738-461c-9ea8-0b2de73b6950--><!--Version: Web v3.4.4--> ------------- Labels: `severity.Medium` `type.FunctionalityBug` original: Thing1Thing2/ped#10
2.0
[PE-D][Tester B] find command duplicates person - Using the default data provided, running 'find Irfan` displays Irfan's details correclty. But then running `find Irfan Ibrahim` causes does nothing. Then run `find Irfan`, Irfan is duplicated in the list. To recover, the app must be reopened. There might be a deeper issue here where clients are randomly getting duplicated in the model. <!--session: 1666944033620-ae8585f7-7738-461c-9ea8-0b2de73b6950--><!--Version: Web v3.4.4--> ------------- Labels: `severity.Medium` `type.FunctionalityBug` original: Thing1Thing2/ped#10
priority
find command duplicates person using the default data provided running find irfan displays irfan s details correclty but then running find irfan ibrahim causes does nothing then run find irfan irfan is duplicated in the list to recover the app must be reopened there might be a deeper issue here where clients are randomly getting duplicated in the model labels severity medium type functionalitybug original ped
1
738,045
25,542,863,056
IssuesEvent
2022-11-29 16:28:50
envoyproxy/gateway
https://api.github.com/repos/envoyproxy/gateway
closed
Wildcard in HttpRoute hostnames generates exact match
bug area/translator priority/medium
*Description*: When `listeners` spec in the Gateway does not specify hosts and HttpRoute spec specifies _wildcard_ host envoy-gateway generates Exact match for the `:authority` that results in matching failures and hence 404. E.g. for HttpRoute host name `*.example.com` it generates ``` HeaderMatches: - Exact: '*.example.com' Name: :authority Prefix: null SafeRegex: null ``` The problem seems to be in [this code](https://github.com/envoyproxy/gateway/blob/cede77ebbecf25fbe8994cfd7643e8295bf75712/internal/gatewayapi/translator.go#L1166) that compares host to `*` while it probably needs to consider `*.example.com` cases **Expected:** Regex match is generated instead of Exact match and request passes the filter *Repro steps*: Apply this Gateway manifest ``` apiVersion: gateway.networking.k8s.io/v1beta1 kind: Gateway metadata: name: eg spec: gatewayClassName: eg listeners: - name: http protocol: HTTP port: 80 ``` Apply this HttpRoute manifest: ``` apiVersion: gateway.networking.k8s.io/v1beta1 kind: HTTPRoute metadata: name: backend spec: parentRefs: - name: eg hostnames: - "*.example.com" rules: - backendRefs: - name: paveliak-playground-1 port: 8080 matches: - path: type: PathPrefix value: / ``` *Environment*: - Minikube 1.27.1 - K8s 1.25.2 - envoy-gateway (tried both `v0.2.0` and `main`) *Logs*: ``` HTTP: - Address: 0.0.0.0 Hostnames: - '*' Name: default-eg-http Port: 10080 Routes: - AddRequestHeaders: null BackendWeights: Invalid: 0 Valid: 0 Destinations: - Host: 10.100.121.40 Port: 8080 Weight: 1 DirectResponse: null HeaderMatches: - Exact: '*.example.com' Name: :authority Prefix: null SafeRegex: null Name: default-backend-rule-0-match-0-*.example.com PathMatch: Exact: null Name: "" Prefix: / SafeRegex: null QueryParamMatches: null Redirect: null RemoveRequestHeaders: null TLS: null TCP: null {"runner": "gateway-api", "output": "xds-ir"} ``` As a result this works `curl http://10.110.191.160 -H "Host: *.example.com"` But this request returns 404 `curl http://10.110.191.160 -H "Host: www.example.com"`
1.0
Wildcard in HttpRoute hostnames generates exact match - *Description*: When `listeners` spec in the Gateway does not specify hosts and HttpRoute spec specifies _wildcard_ host envoy-gateway generates Exact match for the `:authority` that results in matching failures and hence 404. E.g. for HttpRoute host name `*.example.com` it generates ``` HeaderMatches: - Exact: '*.example.com' Name: :authority Prefix: null SafeRegex: null ``` The problem seems to be in [this code](https://github.com/envoyproxy/gateway/blob/cede77ebbecf25fbe8994cfd7643e8295bf75712/internal/gatewayapi/translator.go#L1166) that compares host to `*` while it probably needs to consider `*.example.com` cases **Expected:** Regex match is generated instead of Exact match and request passes the filter *Repro steps*: Apply this Gateway manifest ``` apiVersion: gateway.networking.k8s.io/v1beta1 kind: Gateway metadata: name: eg spec: gatewayClassName: eg listeners: - name: http protocol: HTTP port: 80 ``` Apply this HttpRoute manifest: ``` apiVersion: gateway.networking.k8s.io/v1beta1 kind: HTTPRoute metadata: name: backend spec: parentRefs: - name: eg hostnames: - "*.example.com" rules: - backendRefs: - name: paveliak-playground-1 port: 8080 matches: - path: type: PathPrefix value: / ``` *Environment*: - Minikube 1.27.1 - K8s 1.25.2 - envoy-gateway (tried both `v0.2.0` and `main`) *Logs*: ``` HTTP: - Address: 0.0.0.0 Hostnames: - '*' Name: default-eg-http Port: 10080 Routes: - AddRequestHeaders: null BackendWeights: Invalid: 0 Valid: 0 Destinations: - Host: 10.100.121.40 Port: 8080 Weight: 1 DirectResponse: null HeaderMatches: - Exact: '*.example.com' Name: :authority Prefix: null SafeRegex: null Name: default-backend-rule-0-match-0-*.example.com PathMatch: Exact: null Name: "" Prefix: / SafeRegex: null QueryParamMatches: null Redirect: null RemoveRequestHeaders: null TLS: null TCP: null {"runner": "gateway-api", "output": "xds-ir"} ``` As a result this works `curl http://10.110.191.160 -H "Host: *.example.com"` But this request returns 404 `curl http://10.110.191.160 -H "Host: www.example.com"`
priority
wildcard in httproute hostnames generates exact match description when listeners spec in the gateway does not specify hosts and httproute spec specifies wildcard host envoy gateway generates exact match for the authority that results in matching failures and hence e g for httproute host name example com it generates headermatches exact example com name authority prefix null saferegex null the problem seems to be in that compares host to while it probably needs to consider example com cases expected regex match is generated instead of exact match and request passes the filter repro steps apply this gateway manifest apiversion gateway networking io kind gateway metadata name eg spec gatewayclassname eg listeners name http protocol http port apply this httproute manifest apiversion gateway networking io kind httproute metadata name backend spec parentrefs name eg hostnames example com rules backendrefs name paveliak playground port matches path type pathprefix value environment minikube envoy gateway tried both and main logs http address hostnames name default eg http port routes addrequestheaders null backendweights invalid valid destinations host port weight directresponse null headermatches exact example com name authority prefix null saferegex null name default backend rule match example com pathmatch exact null name prefix saferegex null queryparammatches null redirect null removerequestheaders null tls null tcp null runner gateway api output xds ir as a result this works curl h host example com but this request returns curl h host
1
302,963
9,300,893,328
IssuesEvent
2019-03-23 17:22:54
HabitRPG/habitica
https://api.github.com/repos/HabitRPG/habitica
closed
Username searches in the Hall of Heroes should be case-insensitive
good first issue priority: medium section: Achievements/Popups/Notifications section: other status: issue: in progress
As described in issue https://github.com/HabitRPG/habitica/issues/10972 and its pull request https://github.com/HabitRPG/habitica/pull/10980 , the Hall of Heroes allows moderators to search for users by the Username. Currently that search is case-sensitive. It should be case-insensitive. I.e., searching for "examplename" would let you find a user called "ExampleName". Any contributor who wants to work on this should read the top post at https://github.com/HabitRPG/habitica/issues/10972 to learn about how the moderator search feature works and how you can access it on your local install. https://github.com/HabitRPG/habitica/pull/10980/files will show you how the current case-sensitive search is implemented. _(NB We're not considering this to be a bug in https://github.com/HabitRPG/habitica/pull/10980. We neglected to say that the search should be case-insensitive and so that PR did implement the search feature as intended at the time!)_
1.0
Username searches in the Hall of Heroes should be case-insensitive - As described in issue https://github.com/HabitRPG/habitica/issues/10972 and its pull request https://github.com/HabitRPG/habitica/pull/10980 , the Hall of Heroes allows moderators to search for users by the Username. Currently that search is case-sensitive. It should be case-insensitive. I.e., searching for "examplename" would let you find a user called "ExampleName". Any contributor who wants to work on this should read the top post at https://github.com/HabitRPG/habitica/issues/10972 to learn about how the moderator search feature works and how you can access it on your local install. https://github.com/HabitRPG/habitica/pull/10980/files will show you how the current case-sensitive search is implemented. _(NB We're not considering this to be a bug in https://github.com/HabitRPG/habitica/pull/10980. We neglected to say that the search should be case-insensitive and so that PR did implement the search feature as intended at the time!)_
priority
username searches in the hall of heroes should be case insensitive as described in issue and its pull request the hall of heroes allows moderators to search for users by the username currently that search is case sensitive it should be case insensitive i e searching for examplename would let you find a user called examplename any contributor who wants to work on this should read the top post at to learn about how the moderator search feature works and how you can access it on your local install will show you how the current case sensitive search is implemented nb we re not considering this to be a bug in we neglected to say that the search should be case insensitive and so that pr did implement the search feature as intended at the time
1
596,841
18,145,008,004
IssuesEvent
2021-09-25 09:06:36
google/mozc
https://api.github.com/repos/google/mozc
closed
Missing "こと***" entries
Priority-Medium auto-migrated OpSys-All Type-Conversion
``` I've got them automatically and fixed them manually. Maybe there are some mistakes and you need to calculate the scores. Check "tmp.koto" please. e.g. ====================================================== すごいこと 2665 2235 5782 凄いこと そういうこと 3006 2235 4853 そういうこと ====================================================== ``` Original issue reported on code.google.com by `heathros...@gmail.com` on 9 Nov 2010 at 7:26 Attachments: - [tmp.koto](https://storage.googleapis.com/google-code-attachments/mozc/issue-67/comment-0/tmp.koto)
1.0
Missing "こと***" entries - ``` I've got them automatically and fixed them manually. Maybe there are some mistakes and you need to calculate the scores. Check "tmp.koto" please. e.g. ====================================================== すごいこと 2665 2235 5782 凄いこと そういうこと 3006 2235 4853 そういうこと ====================================================== ``` Original issue reported on code.google.com by `heathros...@gmail.com` on 9 Nov 2010 at 7:26 Attachments: - [tmp.koto](https://storage.googleapis.com/google-code-attachments/mozc/issue-67/comment-0/tmp.koto)
priority
missing こと entries i ve got them automatically and fixed them manually maybe there are some mistakes and you need to calculate the scores check tmp koto please e g すごいこと 凄いこと そういうこと そういうこと original issue reported on code google com by heathros gmail com on nov at attachments
1
442,168
12,740,990,356
IssuesEvent
2020-06-26 04:36:10
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
opened
Merged Prescriber from Desktop store doesn't sync to Mobile Store
5.0.4 Docs: not needed Effort: medium Priority: immediate
## Describe the bug Merged Prescriber from Desktop store doesn't sync to Mobile Store ### To reproduce Steps to reproduce the behavior: 1. In mobile store, create atleast two prescriber . 2. Create Prescription from one of the created Prescriber and sync 3. Merge these prescriber in Desktop Store such that Keep Prescriber with no Prescription and Merge Prescriber with Prescription. 4. Check Mobile Store 5. See Error. Prescribers is not sync ![Screen Shot 2020-06-26 at 10 16 05 AM](https://user-images.githubusercontent.com/58979436/85820517-1d228380-b796-11ea-9b82-ae55de6bda0e.png) ![Screen Shot 2020-06-26 at 10 15 21 AM](https://user-images.githubusercontent.com/58979436/85820521-1eec4700-b796-11ea-8610-da3b0c24ad5e.png) ![Screen Shot 2020-06-26 at 10 18 00 AM](https://user-images.githubusercontent.com/58979436/85820724-98843500-b796-11ea-8d24-74ca3c8979c9.png) ### Expected behaviour Sync should work as expected and Merged prescriber should be shown in Mobile as well ### Proposed Solution Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase. ### Version and device info - App version: v5.0.4 - Tablet model: API 21 - Desktop version: v412RC04 ### Additional context Add any other context about the problem here.
1.0
Merged Prescriber from Desktop store doesn't sync to Mobile Store - ## Describe the bug Merged Prescriber from Desktop store doesn't sync to Mobile Store ### To reproduce Steps to reproduce the behavior: 1. In mobile store, create atleast two prescriber . 2. Create Prescription from one of the created Prescriber and sync 3. Merge these prescriber in Desktop Store such that Keep Prescriber with no Prescription and Merge Prescriber with Prescription. 4. Check Mobile Store 5. See Error. Prescribers is not sync ![Screen Shot 2020-06-26 at 10 16 05 AM](https://user-images.githubusercontent.com/58979436/85820517-1d228380-b796-11ea-9b82-ae55de6bda0e.png) ![Screen Shot 2020-06-26 at 10 15 21 AM](https://user-images.githubusercontent.com/58979436/85820521-1eec4700-b796-11ea-8610-da3b0c24ad5e.png) ![Screen Shot 2020-06-26 at 10 18 00 AM](https://user-images.githubusercontent.com/58979436/85820724-98843500-b796-11ea-8d24-74ca3c8979c9.png) ### Expected behaviour Sync should work as expected and Merged prescriber should be shown in Mobile as well ### Proposed Solution Leave if you don't know how to fix/implement. Edit this issue description and explain here if you know the best path of implementing the fix within the codebase. ### Version and device info - App version: v5.0.4 - Tablet model: API 21 - Desktop version: v412RC04 ### Additional context Add any other context about the problem here.
priority
merged prescriber from desktop store doesn t sync to mobile store describe the bug merged prescriber from desktop store doesn t sync to mobile store to reproduce steps to reproduce the behavior in mobile store create atleast two prescriber create prescription from one of the created prescriber and sync merge these prescriber in desktop store such that keep prescriber with no prescription and merge prescriber with prescription check mobile store see error prescribers is not sync expected behaviour sync should work as expected and merged prescriber should be shown in mobile as well proposed solution leave if you don t know how to fix implement edit this issue description and explain here if you know the best path of implementing the fix within the codebase version and device info app version tablet model api desktop version additional context add any other context about the problem here
1
118,917
4,757,600,475
IssuesEvent
2016-10-24 17:02:50
geosolutions-it/geotools
https://api.github.com/repos/geosolutions-it/geotools
opened
Communitty interaction
C009-2016-MONGODB Priority: Medium Task
Since we are preserving the current MongoDB behavior and only users that want to use complex feature will need to do something, we only need to convince the community about how we intend to support the complex features. Another point that needs to be discussed is the dependency on app-schema (if we don't manage to avoid it).
1.0
Communitty interaction - Since we are preserving the current MongoDB behavior and only users that want to use complex feature will need to do something, we only need to convince the community about how we intend to support the complex features. Another point that needs to be discussed is the dependency on app-schema (if we don't manage to avoid it).
priority
communitty interaction since we are preserving the current mongodb behavior and only users that want to use complex feature will need to do something we only need to convince the community about how we intend to support the complex features another point that needs to be discussed is the dependency on app schema if we don t manage to avoid it
1
22,105
2,645,590,069
IssuesEvent
2015-03-13 00:06:08
pedromorgan/flightgear-issues-test
https://api.github.com/repos/pedromorgan/flightgear-issues-test
closed
MPmap is out off sync
apt.dat bug imported mpmap pigeon Priority-Medium
_From [pedromor...@gmail.com](https://code.google.com/u/112081649209717344658/) on February 03, 2010 07:45:16_ What steps will reproduce the problem? * the nav.dat is updated in cvs, but changes are not reflected on mmapservers * mpmap needs to be in sync with cvs _Original issue: http://code.google.com/p/flightgear-bugs/issues/detail?id=28_
1.0
MPmap is out off sync - _From [pedromor...@gmail.com](https://code.google.com/u/112081649209717344658/) on February 03, 2010 07:45:16_ What steps will reproduce the problem? * the nav.dat is updated in cvs, but changes are not reflected on mmapservers * mpmap needs to be in sync with cvs _Original issue: http://code.google.com/p/flightgear-bugs/issues/detail?id=28_
priority
mpmap is out off sync from on february what steps will reproduce the problem the nav dat is updated in cvs but changes are not reflected on mmapservers mpmap needs to be in sync with cvs original issue
1
271,612
8,485,749,511
IssuesEvent
2018-10-26 08:48:57
cms-gem-daq-project/cmsgemos
https://api.github.com/repos/cms-gem-daq-project/cmsgemos
closed
Exception in amc_info_uhal.py if -r option used at cmd line
Priority: Medium Status: Help Wanted Type: Bug
Based on feedback from DAQ expert tried to issue a reset command to the CTP7 with command line option -r. Generated the exception shown: [amc_info_resetExcept.txt](https://github.com/cms-gem-daq-project/cmsgemos/files/1089047/amc_info_resetExcept.txt)
1.0
Exception in amc_info_uhal.py if -r option used at cmd line - Based on feedback from DAQ expert tried to issue a reset command to the CTP7 with command line option -r. Generated the exception shown: [amc_info_resetExcept.txt](https://github.com/cms-gem-daq-project/cmsgemos/files/1089047/amc_info_resetExcept.txt)
priority
exception in amc info uhal py if r option used at cmd line based on feedback from daq expert tried to issue a reset command to the with command line option r generated the exception shown
1
465,173
13,358,091,790
IssuesEvent
2020-08-31 11:02:33
ooni/probe
https://api.github.com/repos/ooni/probe
closed
Make the OONI Probe desktop app rely entirely on the golang engine
enhancement epic funder/otf20 ooni/probe-desktop priority/medium
We are in the process of consolidating all the code related to running network experiments inside of a new golang based engine called probe-engine, which will replace measurement-kit. As part of this activity, we will make the OONI Probe desktop (Windows and macOS) app rely entirely on the golang engine. We are not aiming to port the OONI Probe desktop apps to a new codebase. We are also not aiming to port the OONI CLI to a new codebase. The work we are proposing is rather to reduce the amount of C++ dependencies used by the OONI Probe CLI and the desktop apps. Below we explain in detail which components we are planning to modify, why we are doing that, and why we have chosen to start integrating these changes in the desktop apps, rather than in the mobile apps. The current OONI Probe desktop apps for Windows and macOS are Electron-based apps implemented at github.com/ooni/probe-desktop. These apps use the CLI interface written in golang to run OONI experiments and perform other functions. The codebase of the OONI Probe CLI interface is available at github.com/ooni/probe-cli. More in detail, we have specified how the Desktop app will exec out the CLI to perform specific tasks and we have also specified the data format emitted by the CLI when running tests. The CLI, in turn, is based on a lower-level “engine” library written in golang and available at github.com/ooni/probe-engine. This library aims to include a unified desktop and mobile implementation of all experiments. However, inside this library, we are still linking to our old C++ engine, Measurement Kit, to implement most OONI network experiments. (Measurement Kit is composed of several GitHub repositories available at github.com/measurement-kit.) The work described in this activity is about gradually phasing out Measurement Kit, so that in the end all experiments are written in Go rather than in C++. This will mostly happen inside the probe-engine and will mostly concern changes in the implementation of tests (i.e. rewriting from C++ to Go). We believe that the API exposed by probe-engine will not change, except for minor changes in the options supported by the new experiments (for example, we may drop an option that does not make sense in the Go context or we may add an option that did not make sense for a C++ implementation). This should explain why we believe that writing more OONI experiments in Go as part of probe-engine does not overlap with our previous work on creating the OONI Probe desktop apps (supported by our last OTF contract). In our view, this proposed work is simply the natural continuation of what we have been working on over the past year. The advantages of rewriting in Go are the following: * It will reduce the code size, because we can drop the C dependencies (whereas we cannot drop the dependency from Go since we want to ship a Psiphon experiment). * It allows us to write experiments that can easily and safely be interrupted, because Go natively support this functionality, which is instead rather tricky to get right in C++. * It allows us to easily wrap standard library functionality, hence allowing for easier low level measurements. * It allows us to quickly cross compile for the operating systems we care about (Windows, macOS, Linux, Android, iOS); full builds will complete in minutes rather than hours, thus enabling faster development cycles. * It will make OONI Probe safer, because Go is memory safe and has built-in support for concurrency. Regarding the choice of focusing on the OONI Probe desktop app first (and on the mobile apps next) the reason is simple. Since probe-engine is already being used by the OONI Probe desktop apps via probe-cli, we have estimated that the effort (and cost) for doing this for the desktop apps is less than doing it on mobile, where we haven’t shipped golang code in production yet. In fact, we can update the desktop apps by recompiling probe-cli to use a new version of probe-engine that reimplements all tests in Go, with none or minimal API changes.
1.0
Make the OONI Probe desktop app rely entirely on the golang engine - We are in the process of consolidating all the code related to running network experiments inside of a new golang based engine called probe-engine, which will replace measurement-kit. As part of this activity, we will make the OONI Probe desktop (Windows and macOS) app rely entirely on the golang engine. We are not aiming to port the OONI Probe desktop apps to a new codebase. We are also not aiming to port the OONI CLI to a new codebase. The work we are proposing is rather to reduce the amount of C++ dependencies used by the OONI Probe CLI and the desktop apps. Below we explain in detail which components we are planning to modify, why we are doing that, and why we have chosen to start integrating these changes in the desktop apps, rather than in the mobile apps. The current OONI Probe desktop apps for Windows and macOS are Electron-based apps implemented at github.com/ooni/probe-desktop. These apps use the CLI interface written in golang to run OONI experiments and perform other functions. The codebase of the OONI Probe CLI interface is available at github.com/ooni/probe-cli. More in detail, we have specified how the Desktop app will exec out the CLI to perform specific tasks and we have also specified the data format emitted by the CLI when running tests. The CLI, in turn, is based on a lower-level “engine” library written in golang and available at github.com/ooni/probe-engine. This library aims to include a unified desktop and mobile implementation of all experiments. However, inside this library, we are still linking to our old C++ engine, Measurement Kit, to implement most OONI network experiments. (Measurement Kit is composed of several GitHub repositories available at github.com/measurement-kit.) The work described in this activity is about gradually phasing out Measurement Kit, so that in the end all experiments are written in Go rather than in C++. This will mostly happen inside the probe-engine and will mostly concern changes in the implementation of tests (i.e. rewriting from C++ to Go). We believe that the API exposed by probe-engine will not change, except for minor changes in the options supported by the new experiments (for example, we may drop an option that does not make sense in the Go context or we may add an option that did not make sense for a C++ implementation). This should explain why we believe that writing more OONI experiments in Go as part of probe-engine does not overlap with our previous work on creating the OONI Probe desktop apps (supported by our last OTF contract). In our view, this proposed work is simply the natural continuation of what we have been working on over the past year. The advantages of rewriting in Go are the following: * It will reduce the code size, because we can drop the C dependencies (whereas we cannot drop the dependency from Go since we want to ship a Psiphon experiment). * It allows us to write experiments that can easily and safely be interrupted, because Go natively support this functionality, which is instead rather tricky to get right in C++. * It allows us to easily wrap standard library functionality, hence allowing for easier low level measurements. * It allows us to quickly cross compile for the operating systems we care about (Windows, macOS, Linux, Android, iOS); full builds will complete in minutes rather than hours, thus enabling faster development cycles. * It will make OONI Probe safer, because Go is memory safe and has built-in support for concurrency. Regarding the choice of focusing on the OONI Probe desktop app first (and on the mobile apps next) the reason is simple. Since probe-engine is already being used by the OONI Probe desktop apps via probe-cli, we have estimated that the effort (and cost) for doing this for the desktop apps is less than doing it on mobile, where we haven’t shipped golang code in production yet. In fact, we can update the desktop apps by recompiling probe-cli to use a new version of probe-engine that reimplements all tests in Go, with none or minimal API changes.
priority
make the ooni probe desktop app rely entirely on the golang engine we are in the process of consolidating all the code related to running network experiments inside of a new golang based engine called probe engine which will replace measurement kit as part of this activity we will make the ooni probe desktop windows and macos app rely entirely on the golang engine we are not aiming to port the ooni probe desktop apps to a new codebase we are also not aiming to port the ooni cli to a new codebase the work we are proposing is rather to reduce the amount of c dependencies used by the ooni probe cli and the desktop apps below we explain in detail which components we are planning to modify why we are doing that and why we have chosen to start integrating these changes in the desktop apps rather than in the mobile apps the current ooni probe desktop apps for windows and macos are electron based apps implemented at github com ooni probe desktop these apps use the cli interface written in golang to run ooni experiments and perform other functions the codebase of the ooni probe cli interface is available at github com ooni probe cli more in detail we have specified how the desktop app will exec out the cli to perform specific tasks and we have also specified the data format emitted by the cli when running tests the cli in turn is based on a lower level “engine” library written in golang and available at github com ooni probe engine this library aims to include a unified desktop and mobile implementation of all experiments however inside this library we are still linking to our old c engine measurement kit to implement most ooni network experiments measurement kit is composed of several github repositories available at github com measurement kit the work described in this activity is about gradually phasing out measurement kit so that in the end all experiments are written in go rather than in c this will mostly happen inside the probe engine and will mostly concern changes in the implementation of tests i e rewriting from c to go we believe that the api exposed by probe engine will not change except for minor changes in the options supported by the new experiments for example we may drop an option that does not make sense in the go context or we may add an option that did not make sense for a c implementation this should explain why we believe that writing more ooni experiments in go as part of probe engine does not overlap with our previous work on creating the ooni probe desktop apps supported by our last otf contract in our view this proposed work is simply the natural continuation of what we have been working on over the past year the advantages of rewriting in go are the following it will reduce the code size because we can drop the c dependencies whereas we cannot drop the dependency from go since we want to ship a psiphon experiment it allows us to write experiments that can easily and safely be interrupted because go natively support this functionality which is instead rather tricky to get right in c it allows us to easily wrap standard library functionality hence allowing for easier low level measurements it allows us to quickly cross compile for the operating systems we care about windows macos linux android ios full builds will complete in minutes rather than hours thus enabling faster development cycles it will make ooni probe safer because go is memory safe and has built in support for concurrency regarding the choice of focusing on the ooni probe desktop app first and on the mobile apps next the reason is simple since probe engine is already being used by the ooni probe desktop apps via probe cli we have estimated that the effort and cost for doing this for the desktop apps is less than doing it on mobile where we haven’t shipped golang code in production yet in fact we can update the desktop apps by recompiling probe cli to use a new version of probe engine that reimplements all tests in go with none or minimal api changes
1
272,107
8,499,185,828
IssuesEvent
2018-10-29 16:32:48
spacetelescope/specviz
https://api.github.com/repos/spacetelescope/specviz
closed
Smoothing fails when trying it a second time
bug gui medium-priority
Smoothing a spectrum the first time, works fine. But when I try to smooth it a second time (click again on the spectrum on the left and the hit the smoothing button) the dialog window looks like this and it won't let me click anything on it. ![screen shot 2018-10-24 at 11 49 46 am](https://user-images.githubusercontent.com/24398824/47443978-721d2380-d783-11e8-913d-f9a8d1962db2.png) That happens even if I delete the first smoothing spectrum.
1.0
Smoothing fails when trying it a second time - Smoothing a spectrum the first time, works fine. But when I try to smooth it a second time (click again on the spectrum on the left and the hit the smoothing button) the dialog window looks like this and it won't let me click anything on it. ![screen shot 2018-10-24 at 11 49 46 am](https://user-images.githubusercontent.com/24398824/47443978-721d2380-d783-11e8-913d-f9a8d1962db2.png) That happens even if I delete the first smoothing spectrum.
priority
smoothing fails when trying it a second time smoothing a spectrum the first time works fine but when i try to smooth it a second time click again on the spectrum on the left and the hit the smoothing button the dialog window looks like this and it won t let me click anything on it that happens even if i delete the first smoothing spectrum
1
357,821
10,618,352,164
IssuesEvent
2019-10-13 03:38:10
carbon-design-system/ibm-dotcom-library
https://api.github.com/repos/carbon-design-system/ibm-dotcom-library
opened
Masthead | Search autosuggest not working in Firefox
bug dev dotcom migrate priority: medium to be triaged
_Jeff-Chew created the following on Aug 16:_ ### Detailed description @Kenny-Lam noticed this while testing the masthead locally. Autosuggest currently isn't working in Firefox. ### Steps to reproduce the issue 1. Go to https://ibmdotcom-react.netlify.com in Firefox 2. Activate search and type in a query (minimum 3 characters) 3. Search results not appearing ### Additional information - Currently the preview isn't there as it's not merged to the main test environment. Will update the ticket once we can reproduce there. _Original issue: https://github.ibm.com/webstandards/digital-design/issues/1473_
1.0
Masthead | Search autosuggest not working in Firefox - _Jeff-Chew created the following on Aug 16:_ ### Detailed description @Kenny-Lam noticed this while testing the masthead locally. Autosuggest currently isn't working in Firefox. ### Steps to reproduce the issue 1. Go to https://ibmdotcom-react.netlify.com in Firefox 2. Activate search and type in a query (minimum 3 characters) 3. Search results not appearing ### Additional information - Currently the preview isn't there as it's not merged to the main test environment. Will update the ticket once we can reproduce there. _Original issue: https://github.ibm.com/webstandards/digital-design/issues/1473_
priority
masthead search autosuggest not working in firefox jeff chew created the following on aug detailed description kenny lam noticed this while testing the masthead locally autosuggest currently isn t working in firefox steps to reproduce the issue go to in firefox activate search and type in a query minimum characters search results not appearing additional information currently the preview isn t there as it s not merged to the main test environment will update the ticket once we can reproduce there original issue
1
332,364
10,092,004,950
IssuesEvent
2019-07-26 15:32:51
wherebyus/general-tasks
https://api.github.com/repos/wherebyus/general-tasks
closed
When I try to type in the photo credit field on WP, it takes FOR-E-VER and sometimes just freezes and doesn't work.
Priority: Medium Product: Newsletters Type: Not A Bug UX: Not Validated
## Feature or problem See above ^ ## UX Validation Not Validated ### Suggested priority Medium ### Stakeholders *Submitted:* caitlin ### Definition of done How will we know when this feature is complete? ### Subtasks A detailed list of changes that need to be made or subtasks. One checkbox per. - [ ] Brew the coffee ## Developer estimate To help the team accurately estimate the complexity of this task, take a moment to walk through this list and estimate each item. At the end, you can total the estimates and round to the nearest prime number. If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking this issue into multiple smaller issues. - [ ] Changes to the database () - [ ] Changes to the API () - [ ] Testing Changes to the API () - [ ] Changes to Application Code () - [ ] Adding or updating unit tests () - [ ] Local developer testing () ### Total developer estimate: 0 ## Additional estimate - [ ] Code review () - [ ] QA Testing () - [ ] Stakeholder Sign-off () - [ ] Deploy to Production () ### Total additional estimate: ## QA Notes Detailed instructions for testing, one checkbox per test to be completed. ### Contextual tests - [ ] Accessibility check - [ ] Cross-browser check (Edge, Chrome, Firefox) - [ ] Responsive check
1.0
When I try to type in the photo credit field on WP, it takes FOR-E-VER and sometimes just freezes and doesn't work. - ## Feature or problem See above ^ ## UX Validation Not Validated ### Suggested priority Medium ### Stakeholders *Submitted:* caitlin ### Definition of done How will we know when this feature is complete? ### Subtasks A detailed list of changes that need to be made or subtasks. One checkbox per. - [ ] Brew the coffee ## Developer estimate To help the team accurately estimate the complexity of this task, take a moment to walk through this list and estimate each item. At the end, you can total the estimates and round to the nearest prime number. If any of these are at a `5` or higher, or if the total is above a `5`, consider breaking this issue into multiple smaller issues. - [ ] Changes to the database () - [ ] Changes to the API () - [ ] Testing Changes to the API () - [ ] Changes to Application Code () - [ ] Adding or updating unit tests () - [ ] Local developer testing () ### Total developer estimate: 0 ## Additional estimate - [ ] Code review () - [ ] QA Testing () - [ ] Stakeholder Sign-off () - [ ] Deploy to Production () ### Total additional estimate: ## QA Notes Detailed instructions for testing, one checkbox per test to be completed. ### Contextual tests - [ ] Accessibility check - [ ] Cross-browser check (Edge, Chrome, Firefox) - [ ] Responsive check
priority
when i try to type in the photo credit field on wp it takes for e ver and sometimes just freezes and doesn t work feature or problem see above ux validation not validated suggested priority medium stakeholders submitted caitlin definition of done how will we know when this feature is complete subtasks a detailed list of changes that need to be made or subtasks one checkbox per brew the coffee developer estimate to help the team accurately estimate the complexity of this task take a moment to walk through this list and estimate each item at the end you can total the estimates and round to the nearest prime number if any of these are at a or higher or if the total is above a consider breaking this issue into multiple smaller issues changes to the database changes to the api testing changes to the api changes to application code adding or updating unit tests local developer testing total developer estimate additional estimate code review qa testing stakeholder sign off deploy to production total additional estimate qa notes detailed instructions for testing one checkbox per test to be completed contextual tests accessibility check cross browser check edge chrome firefox responsive check
1
651,058
21,448,829,960
IssuesEvent
2022-04-25 09:16:26
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
opened
[Редагувати профіль] After changing the role from "Керівник" to "Відвідувач", the user cannot return the role "Керівник"
bug Priority: Medium
**Environment:** Windows 10 Professional, Chrome 96.0.4664.110 **Reproducible:** always **Build found:** https://speak-ukrainian.org.ua/dev/ **Preconditions** 1. Log in as a "Керівник"on https://speak-ukrainian.org.ua/dev/ **Steps to reproduce** 1. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль" 2. Change user role to "Відвідувач" and save changes 3. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль" 4. Try to change the user role back to the role "Керівник" 5. Pay attention to the result of this action **Actual result** The user with the "Керівник" role can lose its role without the possibility of going back **Expected result** The user with the "Керівник" role cannot change his role to the "Відвідувач" role ![role 1](https://user-images.githubusercontent.com/42620939/165059052-7871c9dd-2d05-4442-8831-15cd3e4e9608.png) ![role 2](https://user-images.githubusercontent.com/42620939/165059055-520e3417-5ebe-4c03-8d4d-51448449de3d.png)
1.0
[Редагувати профіль] After changing the role from "Керівник" to "Відвідувач", the user cannot return the role "Керівник" - **Environment:** Windows 10 Professional, Chrome 96.0.4664.110 **Reproducible:** always **Build found:** https://speak-ukrainian.org.ua/dev/ **Preconditions** 1. Log in as a "Керівник"on https://speak-ukrainian.org.ua/dev/ **Steps to reproduce** 1. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль" 2. Change user role to "Відвідувач" and save changes 3. Go to the user cabinet at the top right corner -"Мій профіль" - "Редагувати профіль" 4. Try to change the user role back to the role "Керівник" 5. Pay attention to the result of this action **Actual result** The user with the "Керівник" role can lose its role without the possibility of going back **Expected result** The user with the "Керівник" role cannot change his role to the "Відвідувач" role ![role 1](https://user-images.githubusercontent.com/42620939/165059052-7871c9dd-2d05-4442-8831-15cd3e4e9608.png) ![role 2](https://user-images.githubusercontent.com/42620939/165059055-520e3417-5ebe-4c03-8d4d-51448449de3d.png)
priority
after changing the role from керівник to відвідувач the user cannot return the role керівник environment windows professional chrome reproducible always build found preconditions log in as a керівник on steps to reproduce go to the user cabinet at the top right corner мій профіль редагувати профіль change user role to відвідувач and save changes go to the user cabinet at the top right corner мій профіль редагувати профіль try to change the user role back to the role керівник pay attention to the result of this action actual result the user with the керівник role can lose its role without the possibility of going back expected result the user with the керівник role cannot change his role to the відвідувач role
1
824,111
31,141,621,856
IssuesEvent
2023-08-16 00:49:36
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
closed
Wrong hp amount in a query for npc with id=46645
NPC Priority: Medium Zone: TolBarad
[//]: # (REMBEMBER! Add links to things related to the bug using for example:) [//]: # (http://wowhead.com/) [//]: # (cata-twinhead.twinstar.cz) **Description:** A mistake with this npc HP modifier, probably it has extra 000 or something in query responsive for HP **How to reproduce:** --- just check the hp **How it should work:** -- shouldn't be 30 mln hp here **Database links:** https://www.wowhead.com/npc=46645/cursed-stallion
1.0
Wrong hp amount in a query for npc with id=46645 - [//]: # (REMBEMBER! Add links to things related to the bug using for example:) [//]: # (http://wowhead.com/) [//]: # (cata-twinhead.twinstar.cz) **Description:** A mistake with this npc HP modifier, probably it has extra 000 or something in query responsive for HP **How to reproduce:** --- just check the hp **How it should work:** -- shouldn't be 30 mln hp here **Database links:** https://www.wowhead.com/npc=46645/cursed-stallion
priority
wrong hp amount in a query for npc with id rembember add links to things related to the bug using for example cata twinhead twinstar cz description a mistake with this npc hp modifier probably it has extra or something in query responsive for hp how to reproduce just check the hp how it should work shouldn t be mln hp here database links
1
117,448
4,716,402,809
IssuesEvent
2016-10-16 01:42:47
lzlw/HBONowHelper
https://api.github.com/repos/lzlw/HBONowHelper
opened
Get info via PROD id
enhancement in progress medium priority
http://content.hbonow.com/content/id/v1/PROD592264/detail/web-v1.json This may possibly ensure that all videos will have valid durations in popup info boxes.
1.0
Get info via PROD id - http://content.hbonow.com/content/id/v1/PROD592264/detail/web-v1.json This may possibly ensure that all videos will have valid durations in popup info boxes.
priority
get info via prod id this may possibly ensure that all videos will have valid durations in popup info boxes
1
519,239
15,048,143,208
IssuesEvent
2021-02-03 09:49:46
vmware/clarity
https://api.github.com/repos/vmware/clarity
closed
Modal is also dismissed when pressing the ESC key to close a combobox's dropdown panel.
component: combobox component: dropdown component: modal priority: 4 medium-high status: backlog type: bug v4 v5
## Describe the bug I have a combobox inside a modal but when you try to press the ESC key to close the it's dropdown panel, the modal is also get dismissed. ![modals](https://user-images.githubusercontent.com/50942719/94930400-ed072100-04f8-11eb-980a-c310693a001e.gif) ## How to reproduce https://stackblitz.com/edit/clarity-v4-light-theme-jz6if2?file=src/app/app.component.ts Steps to reproduce the behavior: 1. Toggle the dropdown panel of the combo box. 2. Press ESC key. ## Expected behavior It should only dismiss/close the items panel/dropdown panel of the combo box and not the modal. ## Versions 4.0.2 **App** - Angular: 10 - Node: 12 - Clarity: 4.0.2 **Device:** - Type: [e.g. MacBook] - OS: [e.g. iOS] - Browser: Chrome/Edge Chromium - Version [e.g. 22] ## Additional notes Also, you will notice that there's a console error when pressing the ESC key.
1.0
Modal is also dismissed when pressing the ESC key to close a combobox's dropdown panel. - ## Describe the bug I have a combobox inside a modal but when you try to press the ESC key to close the it's dropdown panel, the modal is also get dismissed. ![modals](https://user-images.githubusercontent.com/50942719/94930400-ed072100-04f8-11eb-980a-c310693a001e.gif) ## How to reproduce https://stackblitz.com/edit/clarity-v4-light-theme-jz6if2?file=src/app/app.component.ts Steps to reproduce the behavior: 1. Toggle the dropdown panel of the combo box. 2. Press ESC key. ## Expected behavior It should only dismiss/close the items panel/dropdown panel of the combo box and not the modal. ## Versions 4.0.2 **App** - Angular: 10 - Node: 12 - Clarity: 4.0.2 **Device:** - Type: [e.g. MacBook] - OS: [e.g. iOS] - Browser: Chrome/Edge Chromium - Version [e.g. 22] ## Additional notes Also, you will notice that there's a console error when pressing the ESC key.
priority
modal is also dismissed when pressing the esc key to close a combobox s dropdown panel describe the bug i have a combobox inside a modal but when you try to press the esc key to close the it s dropdown panel the modal is also get dismissed how to reproduce steps to reproduce the behavior toggle the dropdown panel of the combo box press esc key expected behavior it should only dismiss close the items panel dropdown panel of the combo box and not the modal versions app angular node clarity device type os browser chrome edge chromium version additional notes also you will notice that there s a console error when pressing the esc key
1
198,951
6,979,423,108
IssuesEvent
2017-12-12 20:59:15
compodoc/compodoc
https://api.github.com/repos/compodoc/compodoc
closed
[BUG] Partial documentation is generated | Maximum call stack size exceeded
1. Type: Bug Priority: Medium Status: Accepted Time: ~3 hours
##### **Overview of the issue** When generating the documentation i get this error: ```javascript RangeError: Maximum call stack size exceeded ``` The documentation is partial with missing components, directives and services. I do have a couple of base64 images in the code, but there is a related issue stating this bug was fixed. ##### **Operating System, Node.js, npm, compodoc version(s)** node: 8.5.0 npm: 5.3.0 os: win32 x64 compodoc: 1.0.3 ##### **Angular configuration, a `package.json` file in the root folder** ```json "devDependencies": { "angular/animations": "4.4.4", "angular/cli": "1.4.5", "angular/common": "4.4.4", "angular/compiler": "4.4.4", "angular/compiler-cli": "4.4.4", "angular/core": "4.4.4", "angular/forms": "4.4.4", "angular/http": "4.4.4", "angular/language-service": "4.4.4", "angular/platform-browser": "4.4.4", "angular/platform-browser-dynamic": "4.4.4", "angular/router": "4.4.4", "compodoc/compodoc": "1.0.3", "types/chai-as-promised": "^7.1.0", "types/jasmine": "~2.5.54", "types/jasminewd2": "2.0.3", "types/node": "6.0.89", "bootstrap": "4.0.0-beta", "chai": "4.1.2", "chai-as-promised": "7.1.1", "codelyzer": "~3.1.1", "core-js": "2.5.1", "cucumber": "2.2.0", "cucumber-tsflow": "2.2.0", "font-awesome": "4.7.0", "jasmine-core": "~2.6.2", "jasmine-spec-reporter": "~4.1.0", "jquery": "3.2.1", "karma": "1.7.1", "karma-chrome-launcher": "~2.1.1", "karma-cli": "~1.0.1", "karma-coverage-istanbul-reporter": "1.2.1", "karma-jasmine": "~1.1.0", "karma-jasmine-html-reporter": "0.2.2", "popper.js": "1.12.5", "primeng": "4.2.2", "protractor": "~4.0.0", "protractor-cucumber-framework": "4.0.8", "rxjs": "5.4.3", "ts-node": "~3.2.0", "tslint": "~5.3.2", "typescript": "~2.3.4", "zone.js": "0.8.18" }, ``` ##### **Related issues** An issues was closed but still the problem persist. [#269](https://github.com/compodoc/compodoc/issues/269)
1.0
[BUG] Partial documentation is generated | Maximum call stack size exceeded - ##### **Overview of the issue** When generating the documentation i get this error: ```javascript RangeError: Maximum call stack size exceeded ``` The documentation is partial with missing components, directives and services. I do have a couple of base64 images in the code, but there is a related issue stating this bug was fixed. ##### **Operating System, Node.js, npm, compodoc version(s)** node: 8.5.0 npm: 5.3.0 os: win32 x64 compodoc: 1.0.3 ##### **Angular configuration, a `package.json` file in the root folder** ```json "devDependencies": { "angular/animations": "4.4.4", "angular/cli": "1.4.5", "angular/common": "4.4.4", "angular/compiler": "4.4.4", "angular/compiler-cli": "4.4.4", "angular/core": "4.4.4", "angular/forms": "4.4.4", "angular/http": "4.4.4", "angular/language-service": "4.4.4", "angular/platform-browser": "4.4.4", "angular/platform-browser-dynamic": "4.4.4", "angular/router": "4.4.4", "compodoc/compodoc": "1.0.3", "types/chai-as-promised": "^7.1.0", "types/jasmine": "~2.5.54", "types/jasminewd2": "2.0.3", "types/node": "6.0.89", "bootstrap": "4.0.0-beta", "chai": "4.1.2", "chai-as-promised": "7.1.1", "codelyzer": "~3.1.1", "core-js": "2.5.1", "cucumber": "2.2.0", "cucumber-tsflow": "2.2.0", "font-awesome": "4.7.0", "jasmine-core": "~2.6.2", "jasmine-spec-reporter": "~4.1.0", "jquery": "3.2.1", "karma": "1.7.1", "karma-chrome-launcher": "~2.1.1", "karma-cli": "~1.0.1", "karma-coverage-istanbul-reporter": "1.2.1", "karma-jasmine": "~1.1.0", "karma-jasmine-html-reporter": "0.2.2", "popper.js": "1.12.5", "primeng": "4.2.2", "protractor": "~4.0.0", "protractor-cucumber-framework": "4.0.8", "rxjs": "5.4.3", "ts-node": "~3.2.0", "tslint": "~5.3.2", "typescript": "~2.3.4", "zone.js": "0.8.18" }, ``` ##### **Related issues** An issues was closed but still the problem persist. [#269](https://github.com/compodoc/compodoc/issues/269)
priority
partial documentation is generated maximum call stack size exceeded overview of the issue when generating the documentation i get this error javascript rangeerror maximum call stack size exceeded the documentation is partial with missing components directives and services i do have a couple of images in the code but there is a related issue stating this bug was fixed operating system node js npm compodoc version s node npm os compodoc angular configuration a package json file in the root folder json devdependencies angular animations angular cli angular common angular compiler angular compiler cli angular core angular forms angular http angular language service angular platform browser angular platform browser dynamic angular router compodoc compodoc types chai as promised types jasmine types types node bootstrap beta chai chai as promised codelyzer core js cucumber cucumber tsflow font awesome jasmine core jasmine spec reporter jquery karma karma chrome launcher karma cli karma coverage istanbul reporter karma jasmine karma jasmine html reporter popper js primeng protractor protractor cucumber framework rxjs ts node tslint typescript zone js related issues an issues was closed but still the problem persist
1
400,709
11,779,593,671
IssuesEvent
2020-03-16 18:19:31
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
opened
Email view - HTML and Email Tokens are stripped down from Email Invites when used with MailGun
bug component: invites priority: medium
**Describe the bug** On Email view, HTML and Email Tokens are stripped down from Email Invites when used with MailGun **To Reproduce** Steps to reproduce the behavior: 1. Set up a MailGun account (free version) 2. Install MailGun Plugin https://wordpress.org/plugins/mailgun/ 3. Setup credentials in MailGun plguin from created mailgun account 4. On frontend, try to invite a user on Email Invites on your profile 5. Check invited users email, all html and email tokens are stripped **Expected behavior** Email Invites should have a layout **Screenshots** https://www.loom.com/share/816b15a82b124e77be38e979f2a4b7c0 ![email-setup](https://user-images.githubusercontent.com/58522224/76788227-7a04ba00-67f5-11ea-8c40-6f63612463b9.JPG) ![email-test2](https://user-images.githubusercontent.com/58522224/76788232-7cffaa80-67f5-11ea-82b5-c460f99053e3.JPG) **Support ticket links** https://buddyboss.zendesk.com/agent/tickets/63879
1.0
Email view - HTML and Email Tokens are stripped down from Email Invites when used with MailGun - **Describe the bug** On Email view, HTML and Email Tokens are stripped down from Email Invites when used with MailGun **To Reproduce** Steps to reproduce the behavior: 1. Set up a MailGun account (free version) 2. Install MailGun Plugin https://wordpress.org/plugins/mailgun/ 3. Setup credentials in MailGun plguin from created mailgun account 4. On frontend, try to invite a user on Email Invites on your profile 5. Check invited users email, all html and email tokens are stripped **Expected behavior** Email Invites should have a layout **Screenshots** https://www.loom.com/share/816b15a82b124e77be38e979f2a4b7c0 ![email-setup](https://user-images.githubusercontent.com/58522224/76788227-7a04ba00-67f5-11ea-8c40-6f63612463b9.JPG) ![email-test2](https://user-images.githubusercontent.com/58522224/76788232-7cffaa80-67f5-11ea-82b5-c460f99053e3.JPG) **Support ticket links** https://buddyboss.zendesk.com/agent/tickets/63879
priority
email view html and email tokens are stripped down from email invites when used with mailgun describe the bug on email view html and email tokens are stripped down from email invites when used with mailgun to reproduce steps to reproduce the behavior set up a mailgun account free version install mailgun plugin setup credentials in mailgun plguin from created mailgun account on frontend try to invite a user on email invites on your profile check invited users email all html and email tokens are stripped expected behavior email invites should have a layout screenshots support ticket links
1
197,369
6,954,842,441
IssuesEvent
2017-12-07 03:51:47
ODIQueensland/data-curator
https://api.github.com/repos/ODIQueensland/data-curator
closed
Provide "pick list" of tables and fields for Foreign and Primary Keys
f:Feature-request fn:Set-Properties priority:Medium status:In-progress
### Desired Behaviour Allow the user to select from a list of valid table or field names to specify the [Foreign ](http://specs.frictionlessdata.io/table-schema/#foreign-keys)and [Primary Keys](http://specs.frictionlessdata.io/table-schema/#primary-key). Support the selection of multiple fields for keys by adding like the current sources function
1.0
Provide "pick list" of tables and fields for Foreign and Primary Keys - ### Desired Behaviour Allow the user to select from a list of valid table or field names to specify the [Foreign ](http://specs.frictionlessdata.io/table-schema/#foreign-keys)and [Primary Keys](http://specs.frictionlessdata.io/table-schema/#primary-key). Support the selection of multiple fields for keys by adding like the current sources function
priority
provide pick list of tables and fields for foreign and primary keys desired behaviour allow the user to select from a list of valid table or field names to specify the support the selection of multiple fields for keys by adding like the current sources function
1
530,595
15,434,473,661
IssuesEvent
2021-03-07 03:33:00
timmo001/system-bridge
https://api.github.com/repos/timmo001/system-bridge
opened
Set Windows Desktop and Start Menu Shortcut
enhancement priority-medium
# Description Icons are currently set to `system-bridge` for Windows. Linux is as it should be. ## Suggested Actions ![image](https://user-images.githubusercontent.com/28114703/110228008-82b63a80-7ef5-11eb-9504-ad185372693d.png) to ![image](https://user-images.githubusercontent.com/28114703/110228017-9661a100-7ef5-11eb-97e0-8426b5afc97b.png)
1.0
Set Windows Desktop and Start Menu Shortcut - # Description Icons are currently set to `system-bridge` for Windows. Linux is as it should be. ## Suggested Actions ![image](https://user-images.githubusercontent.com/28114703/110228008-82b63a80-7ef5-11eb-9504-ad185372693d.png) to ![image](https://user-images.githubusercontent.com/28114703/110228017-9661a100-7ef5-11eb-97e0-8426b5afc97b.png)
priority
set windows desktop and start menu shortcut description icons are currently set to system bridge for windows linux is as it should be suggested actions to
1
16,570
2,615,119,082
IssuesEvent
2015-03-01 05:44:56
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
opened
Spreadsheet (List?) access (specifically to iterate through all the rows on a given sheet)
auto-migrated Priority-Medium Type-Sample
``` Which API and version (e.g. Google Calendar Data API version 2)? google-api-java-client v1.3.1 (or greater if no one gets around to this before the next drop) What format (e.g. JSON, Atom)? JSON What Authentication (e.g. OAuth, OAuth 2, Android, ClientLogin)? Whatever is recommended when using Google credentials. It's all new to me. Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.2)? Java 6, Windows 7 x64, Netbeans 6.9.1 External references, such as API reference guide? Please provide any additional information below. All existing samples are for the legacy GDATA API as far as I can tell. ``` Original issue reported on code.google.com by `Predato...@gmail.com` on 3 Mar 2011 at 4:15
1.0
Spreadsheet (List?) access (specifically to iterate through all the rows on a given sheet) - ``` Which API and version (e.g. Google Calendar Data API version 2)? google-api-java-client v1.3.1 (or greater if no one gets around to this before the next drop) What format (e.g. JSON, Atom)? JSON What Authentication (e.g. OAuth, OAuth 2, Android, ClientLogin)? Whatever is recommended when using Google credentials. It's all new to me. Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.2)? Java 6, Windows 7 x64, Netbeans 6.9.1 External references, such as API reference guide? Please provide any additional information below. All existing samples are for the legacy GDATA API as far as I can tell. ``` Original issue reported on code.google.com by `Predato...@gmail.com` on 3 Mar 2011 at 4:15
priority
spreadsheet list access specifically to iterate through all the rows on a given sheet which api and version e g google calendar data api version google api java client or greater if no one gets around to this before the next drop what format e g json atom json what authentication e g oauth oauth android clientlogin whatever is recommended when using google credentials it s all new to me java environment e g java android app engine java windows netbeans external references such as api reference guide please provide any additional information below all existing samples are for the legacy gdata api as far as i can tell original issue reported on code google com by predato gmail com on mar at
1
412,090
12,035,187,763
IssuesEvent
2020-04-13 17:25:51
ChainSafe/forest
https://api.github.com/repos/ChainSafe/forest
opened
Implement Randomness
Priority: 3 - Medium VM
In the Lotus implementation, they use a RAND library within the VM context.
1.0
Implement Randomness - In the Lotus implementation, they use a RAND library within the VM context.
priority
implement randomness in the lotus implementation they use a rand library within the vm context
1
202,796
7,055,241,966
IssuesEvent
2018-01-04 07:03:58
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Hazelcast : Set Map Configuration to Cluster Node without Creating Instance of Hazelcast locally
Priority: Medium Source: Community Team: Core Type: Enhancement
Hi Below are my cluster members and I wan to pass some Map configuration to those node so once that Map create all node use that configuration. Members [3] { Member [169.185.209.99]:5701 Member [165.203.252.170]:5701 Member [169.185.209.169]:5701 } How can i pass Map config without creating hazlecast Instance on (169.164.232.143) my local system. this is really urgent.. Thanks Vipul
1.0
Hazelcast : Set Map Configuration to Cluster Node without Creating Instance of Hazelcast locally - Hi Below are my cluster members and I wan to pass some Map configuration to those node so once that Map create all node use that configuration. Members [3] { Member [169.185.209.99]:5701 Member [165.203.252.170]:5701 Member [169.185.209.169]:5701 } How can i pass Map config without creating hazlecast Instance on (169.164.232.143) my local system. this is really urgent.. Thanks Vipul
priority
hazelcast set map configuration to cluster node without creating instance of hazelcast locally hi below are my cluster members and i wan to pass some map configuration to those node so once that map create all node use that configuration members member member member how can i pass map config without creating hazlecast instance on my local system this is really urgent thanks vipul
1
28,492
2,704,467,267
IssuesEvent
2015-04-07 00:49:56
kevinchandler/test
https://api.github.com/repos/kevinchandler/test
closed
dfsafsdfsdfds
bug-report-cs Priority: Medium
Description: dfsafsdfsdfds Priority: Medium Date Reported: 04-23-2015 User Email: kevin@onradpad.com Profile: https://onradpad.com/admin/landlords/837750 Platform: Mac Able to Replicate: No Reported By: Kevin C. - kevin@onradpad.com
1.0
dfsafsdfsdfds - Description: dfsafsdfsdfds Priority: Medium Date Reported: 04-23-2015 User Email: kevin@onradpad.com Profile: https://onradpad.com/admin/landlords/837750 Platform: Mac Able to Replicate: No Reported By: Kevin C. - kevin@onradpad.com
priority
dfsafsdfsdfds description dfsafsdfsdfds priority medium date reported user email kevin onradpad com profile platform mac able to replicate no reported by kevin c kevin onradpad com
1
452,899
13,061,011,366
IssuesEvent
2020-07-30 13:19:35
clastix/capsule
https://api.github.com/repos/clastix/capsule
closed
Make the logs level configurable
enhancement medium-priority
It would be nice to have a configurable logs level for the capsule controller, e.g. info, warning, debug, trace, .. This will help during troubleshooting.
1.0
Make the logs level configurable - It would be nice to have a configurable logs level for the capsule controller, e.g. info, warning, debug, trace, .. This will help during troubleshooting.
priority
make the logs level configurable it would be nice to have a configurable logs level for the capsule controller e g info warning debug trace this will help during troubleshooting
1
399,744
11,759,908,903
IssuesEvent
2020-03-13 18:16:07
ChainSafe/ChainBridgeV2
https://api.github.com/repos/ChainSafe/ChainBridgeV2
closed
Convert JS scripts to CLI
Ethereum Priority: 3 - Medium
The current JS scripts are a hacked together CLI. They should be switched to a command based approach. The commands should be:`deploy` & `transfer`, with the respective flags.
1.0
Convert JS scripts to CLI - The current JS scripts are a hacked together CLI. They should be switched to a command based approach. The commands should be:`deploy` & `transfer`, with the respective flags.
priority
convert js scripts to cli the current js scripts are a hacked together cli they should be switched to a command based approach the commands should be deploy transfer with the respective flags
1
326,906
9,962,356,286
IssuesEvent
2019-07-07 13:55:04
PMEAL/OpenPNM
https://api.github.com/repos/PMEAL/OpenPNM
closed
Networks with 1 pore and 0 throats break stuff
Next Minor Release Priority - Medium bug
Not that this would ever happen, but it might be worth throwing an error or warning?
1.0
Networks with 1 pore and 0 throats break stuff - Not that this would ever happen, but it might be worth throwing an error or warning?
priority
networks with pore and throats break stuff not that this would ever happen but it might be worth throwing an error or warning
1
43,989
2,894,941,497
IssuesEvent
2015-06-16 05:05:00
wchill/steamSummerMinigame
https://api.github.com/repos/wchill/steamSummerMinigame
closed
[Suggestion] Auto Use 'Treasure' On Bosses
2 - Medium Priority
This would make sense as is also does a free metal detector too, ideally for a Boss who has more than 60% health as per 'Raining Gold.'
1.0
[Suggestion] Auto Use 'Treasure' On Bosses - This would make sense as is also does a free metal detector too, ideally for a Boss who has more than 60% health as per 'Raining Gold.'
priority
auto use treasure on bosses this would make sense as is also does a free metal detector too ideally for a boss who has more than health as per raining gold
1
77,297
3,506,339,225
IssuesEvent
2016-01-08 05:53:38
OregonCore/OregonCore
https://api.github.com/repos/OregonCore/OregonCore
closed
Full rage when arena begins (BB #344)
migrated Priority: Medium Type: Bug
This issue was migrated from bitbucket. **Original Reporter:** henhouse **Original Date:** 07.11.2010 18:33:33 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/344 <hr> It's possible to attack an enemy and get full rage then entering an arena with full rage. It doesn't reset to 0 at the start of an arena match.
1.0
Full rage when arena begins (BB #344) - This issue was migrated from bitbucket. **Original Reporter:** henhouse **Original Date:** 07.11.2010 18:33:33 GMT+0000 **Original Priority:** major **Original Type:** bug **Original State:** resolved **Direct Link:** https://bitbucket.org/oregon/oregoncore/issues/344 <hr> It's possible to attack an enemy and get full rage then entering an arena with full rage. It doesn't reset to 0 at the start of an arena match.
priority
full rage when arena begins bb this issue was migrated from bitbucket original reporter henhouse original date gmt original priority major original type bug original state resolved direct link it s possible to attack an enemy and get full rage then entering an arena with full rage it doesn t reset to at the start of an arena match
1
738,271
25,550,823,828
IssuesEvent
2022-11-29 23:31:11
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[xCluster] Race condition with SPLIT_OP processing and CdcConsumer shutdown
kind/bug area/docdb priority/medium
Jira Link: [DB-836](https://yugabyte.atlassian.net/browse/DB-836) ### Description Noticed when working on `XClusterAutomaticTabletSplitITest.AutomaticTabletSplitting`. There is a race occurring during the shutdown path, if we shutdown immediately after sending a DeleteUniverseReplication request and before waiting for all the CDCConsumers to be properly cleared. From my investigations, we seem to be getting stuck in `rpcs_->Abort({&write_handle_});` during `~TwoDCOutputClient()`, but unclear as to what its waiting on.. Currently working around in tests by waiting for pollers to be deleted by DeleteUniverseReplication before continuing with the shutdown.
1.0
[xCluster] Race condition with SPLIT_OP processing and CdcConsumer shutdown - Jira Link: [DB-836](https://yugabyte.atlassian.net/browse/DB-836) ### Description Noticed when working on `XClusterAutomaticTabletSplitITest.AutomaticTabletSplitting`. There is a race occurring during the shutdown path, if we shutdown immediately after sending a DeleteUniverseReplication request and before waiting for all the CDCConsumers to be properly cleared. From my investigations, we seem to be getting stuck in `rpcs_->Abort({&write_handle_});` during `~TwoDCOutputClient()`, but unclear as to what its waiting on.. Currently working around in tests by waiting for pollers to be deleted by DeleteUniverseReplication before continuing with the shutdown.
priority
race condition with split op processing and cdcconsumer shutdown jira link description noticed when working on xclusterautomatictabletsplititest automatictabletsplitting there is a race occurring during the shutdown path if we shutdown immediately after sending a deleteuniversereplication request and before waiting for all the cdcconsumers to be properly cleared from my investigations we seem to be getting stuck in rpcs abort write handle during twodcoutputclient but unclear as to what its waiting on currently working around in tests by waiting for pollers to be deleted by deleteuniversereplication before continuing with the shutdown
1
728,827
25,094,734,667
IssuesEvent
2022-11-08 09:21:54
Zenika/skillZ
https://api.github.com/repos/Zenika/skillZ
closed
add admin emails
Priority: Medium
I've just added Benjamin Adrien and me for the tests. The core team and Pierre should be added in the clever env var
1.0
add admin emails - I've just added Benjamin Adrien and me for the tests. The core team and Pierre should be added in the clever env var
priority
add admin emails i ve just added benjamin adrien and me for the tests the core team and pierre should be added in the clever env var
1
41,132
2,868,977,969
IssuesEvent
2015-06-05 22:20:10
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
Make pub run use an isolate instead of a process for the spawned program
Area-Pub Priority-Medium Triaged Type-Enhancement
Right now, it runs the app in a separate process. We should use an isolate because it will play nicer with pipes and detecting the stdioType in the spawned app.
1.0
Make pub run use an isolate instead of a process for the spawned program - Right now, it runs the app in a separate process. We should use an isolate because it will play nicer with pipes and detecting the stdioType in the spawned app.
priority
make pub run use an isolate instead of a process for the spawned program right now it runs the app in a separate process we should use an isolate because it will play nicer with pipes and detecting the stdiotype in the spawned app
1
465,077
13,351,240,303
IssuesEvent
2020-08-30 12:35:19
ncssar/sartopo_address
https://api.github.com/repos/ncssar/sartopo_address
opened
inconsistent defafult-folder-per-custom-marker behavior
Priority: medium bug
clean up this behavior: - folder name should revert to evac- folder name, not to 'Folder', when evac- symbol is selected and folder combo box is released - custom-default-folder name doesn't always appear in folder combo box; see video https://www.youtube.com/watch?v=1U9VOujH7Es
1.0
inconsistent defafult-folder-per-custom-marker behavior - clean up this behavior: - folder name should revert to evac- folder name, not to 'Folder', when evac- symbol is selected and folder combo box is released - custom-default-folder name doesn't always appear in folder combo box; see video https://www.youtube.com/watch?v=1U9VOujH7Es
priority
inconsistent defafult folder per custom marker behavior clean up this behavior folder name should revert to evac folder name not to folder when evac symbol is selected and folder combo box is released custom default folder name doesn t always appear in folder combo box see video
1
282,644
8,709,073,472
IssuesEvent
2018-12-06 12:55:14
7552-2C-2018/App-Server
https://api.github.com/repos/7552-2C-2018/App-Server
closed
Get Questions
Priority-Medium feature
Devuelve la preguntas y respuestas de una publicación. Se envia facebookID, token, id de publicacion{facebookID y pubDate}. Devuelve JSONArray de {"question": string, "answer": string} answer puede no estar
1.0
Get Questions - Devuelve la preguntas y respuestas de una publicación. Se envia facebookID, token, id de publicacion{facebookID y pubDate}. Devuelve JSONArray de {"question": string, "answer": string} answer puede no estar
priority
get questions devuelve la preguntas y respuestas de una publicación se envia facebookid token id de publicacion facebookid y pubdate devuelve jsonarray de question string answer string answer puede no estar
1
714,379
24,559,915,275
IssuesEvent
2022-10-12 19:14:28
asastats/channel
https://api.github.com/repos/asastats/channel
closed
One of the Alchemon addresses fails to be added to the related collection
bug medium priority addressed
DragMZ in [Discord](https://discord.com/channels/906917846754418770/908054330265960478/1029809964669218857): > OJGTHEJ2O5NXN7FVXDZZEEJTUEQHHCIYIE5MWY6BEFVVLZ2KANJODBOKGA Alchemons collection isn't recognized.
1.0
One of the Alchemon addresses fails to be added to the related collection - DragMZ in [Discord](https://discord.com/channels/906917846754418770/908054330265960478/1029809964669218857): > OJGTHEJ2O5NXN7FVXDZZEEJTUEQHHCIYIE5MWY6BEFVVLZ2KANJODBOKGA Alchemons collection isn't recognized.
priority
one of the alchemon addresses fails to be added to the related collection dragmz in alchemons collection isn t recognized
1
247,040
7,896,142,210
IssuesEvent
2018-06-29 07:29:01
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
Provide an OpenCLI() cli method which launches a new cli instance
Expected Use: 3 - Occasional Feature Impact: 3 - Medium Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 2 - Minor Irritation Support Group: DOE/ASC version: 2.0.0
rm-id: 38 rm-submitter: Cyrus Harrison rm-submit-date: 05/14/2010 03:05 pm When using the visitmodule, Bob Corey expected visit.AddArgument("-cli") to pop up a new cli window when visit.Launch() was called. What he actually wants is the ability to launch a new cli instance from the VisIt python module, similar to OpenGUI(). So we should provide an OpenCLI() method to enable this. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Cyrus Harrison Original creation: 06/21/2010 07:14 pm Original update: 07/14/2010 03:01 pm Ticket number: 23
1.0
Provide an OpenCLI() cli method which launches a new cli instance - rm-id: 38 rm-submitter: Cyrus Harrison rm-submit-date: 05/14/2010 03:05 pm When using the visitmodule, Bob Corey expected visit.AddArgument("-cli") to pop up a new cli window when visit.Launch() was called. What he actually wants is the ability to launch a new cli instance from the VisIt python module, similar to OpenGUI(). So we should provide an OpenCLI() method to enable this. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Cyrus Harrison Original creation: 06/21/2010 07:14 pm Original update: 07/14/2010 03:01 pm Ticket number: 23
priority
provide an opencli cli method which launches a new cli instance rm id rm submitter cyrus harrison rm submit date pm when using the visitmodule bob corey expected visit addargument cli to pop up a new cli window when visit launch was called what he actually wants is the ability to launch a new cli instance from the visit python module similar to opengui so we should provide an opencli method to enable this redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author cyrus harrison original creation pm original update pm ticket number
1
3,160
2,537,393,613
IssuesEvent
2015-01-26 20:16:38
web2py/web2py
https://api.github.com/repos/web2py/web2py
opened
Querying on 1:M Relationships in Grid and Smartgrid
1 star enhancement imported Priority-Medium
_From [mike.leo...@gmail.com](https://code.google.com/u/109626380233121956961/) on September 05, 2012 22:07:38_ Hello, We're building a web2py application that leverages SQLFORM.smartgrid heavily. Out of the box, this does not support filtering on 1:M relationships, or reference fields on the model. We understand the concerns about this: It could potentially be a huge performance hit, especially if the related table has a large number of rows. On the other hand, this functionality is very valuable. Being able to filter on related tables is a very common need in applications. Often, users need to filter on "lookup" tables that have a relatively small number of rows, and the performance hit is negligible. This is a very common feature in Django, and in Rails admin interfaces like ActiveScaffold and RailsAdmin. Example: Given a model like this: db.define_table('book', Field('user_id', 'reference auth_user'), Field('title'), Field('genre')) And a controller like this: def index(): grid = SQLFORM.smartgrid(db.book) return dict(grid=grid) The query interface will look like this by default: http://i49.tinypic.com/2ugjt02.jpg Note how this includes title and genre, but not the auth_user reference. We patched web2py to allow to for 1:M reference fields in the query interface. The resulting interface and generated query look something like this: http://i46.tinypic.com/28taq74.png Here is our simple patch to allow searching on reference fields like this (web2py version 1.99.7): https://gist.github.com/ff543a5c6d3bc14b9079 This is obviously NOT ready for production; it's a simple proof of concept. Note the potentially very expensive "ALL" query. Because of the performance risks, something like this only makes sense if developers can opt in, i.e. explicitly say so if they want to include a given reference field in the query interface. 1. Would the web2py maintiners be interested in a feature like this if it is opt-in? I'm willing to devote some time to submit a more robust patch if I could get some guidance. 2. I also have a simple implementation for querying on list:reference fields that I can share as well. Finally, I just want to thank the web2py team. I've really enjoyed using the framework and hope to help out as much as I can. Thanks, Mike Leone Principal, Engineering Panoptic Development, Inc. \----------------------- On Friday, August 31, 2012 5:48:17 PM UTC-4, Massimo Di Pierro wrote: Hello Mike, thanks for your patch? Would you please attach it to a google code issue so we do not forget. We cannot add new feaures until all 2.0.x issues have been resolved. _Original issue: http://code.google.com/p/web2py/issues/detail?id=979_
1.0
Querying on 1:M Relationships in Grid and Smartgrid - _From [mike.leo...@gmail.com](https://code.google.com/u/109626380233121956961/) on September 05, 2012 22:07:38_ Hello, We're building a web2py application that leverages SQLFORM.smartgrid heavily. Out of the box, this does not support filtering on 1:M relationships, or reference fields on the model. We understand the concerns about this: It could potentially be a huge performance hit, especially if the related table has a large number of rows. On the other hand, this functionality is very valuable. Being able to filter on related tables is a very common need in applications. Often, users need to filter on "lookup" tables that have a relatively small number of rows, and the performance hit is negligible. This is a very common feature in Django, and in Rails admin interfaces like ActiveScaffold and RailsAdmin. Example: Given a model like this: db.define_table('book', Field('user_id', 'reference auth_user'), Field('title'), Field('genre')) And a controller like this: def index(): grid = SQLFORM.smartgrid(db.book) return dict(grid=grid) The query interface will look like this by default: http://i49.tinypic.com/2ugjt02.jpg Note how this includes title and genre, but not the auth_user reference. We patched web2py to allow to for 1:M reference fields in the query interface. The resulting interface and generated query look something like this: http://i46.tinypic.com/28taq74.png Here is our simple patch to allow searching on reference fields like this (web2py version 1.99.7): https://gist.github.com/ff543a5c6d3bc14b9079 This is obviously NOT ready for production; it's a simple proof of concept. Note the potentially very expensive "ALL" query. Because of the performance risks, something like this only makes sense if developers can opt in, i.e. explicitly say so if they want to include a given reference field in the query interface. 1. Would the web2py maintiners be interested in a feature like this if it is opt-in? I'm willing to devote some time to submit a more robust patch if I could get some guidance. 2. I also have a simple implementation for querying on list:reference fields that I can share as well. Finally, I just want to thank the web2py team. I've really enjoyed using the framework and hope to help out as much as I can. Thanks, Mike Leone Principal, Engineering Panoptic Development, Inc. \----------------------- On Friday, August 31, 2012 5:48:17 PM UTC-4, Massimo Di Pierro wrote: Hello Mike, thanks for your patch? Would you please attach it to a google code issue so we do not forget. We cannot add new feaures until all 2.0.x issues have been resolved. _Original issue: http://code.google.com/p/web2py/issues/detail?id=979_
priority
querying on m relationships in grid and smartgrid from on september hello we re building a application that leverages sqlform smartgrid heavily out of the box this does not support filtering on m relationships or reference fields on the model we understand the concerns about this it could potentially be a huge performance hit especially if the related table has a large number of rows on the other hand this functionality is very valuable being able to filter on related tables is a very common need in applications often users need to filter on lookup tables that have a relatively small number of rows and the performance hit is negligible this is a very common feature in django and in rails admin interfaces like activescaffold and railsadmin example given a model like this db define table book field user id reference auth user field title field genre and a controller like this def index grid sqlform smartgrid db book return dict grid grid the query interface will look like this by default note how this includes title and genre but not the auth user reference we patched to allow to for m reference fields in the query interface the resulting interface and generated query look something like this here is our simple patch to allow searching on reference fields like this version this is obviously not ready for production it s a simple proof of concept note the potentially very expensive all query because of the performance risks something like this only makes sense if developers can opt in i e explicitly say so if they want to include a given reference field in the query interface would the maintiners be interested in a feature like this if it is opt in i m willing to devote some time to submit a more robust patch if i could get some guidance i also have a simple implementation for querying on list reference fields that i can share as well finally i just want to thank the team i ve really enjoyed using the framework and hope to help out as much as i can thanks mike leone principal engineering panoptic development inc on friday august pm utc massimo di pierro wrote hello mike thanks for your patch would you please attach it to a google code issue so we do not forget we cannot add new feaures until all x issues have been resolved original issue
1
543,970
15,888,076,453
IssuesEvent
2021-04-10 05:30:48
AY2021S2-CS2103T-T12-4/tp
https://api.github.com/repos/AY2021S2-CS2103T-T12-4/tp
closed
DG: Implementation - Send & Run feature
priority.High severity.Medium
- refactor description to focus on run instead as the request feature talks about send extensively - update to the UML diagram - be ready by tonight or saturday morning
1.0
DG: Implementation - Send & Run feature - - refactor description to focus on run instead as the request feature talks about send extensively - update to the UML diagram - be ready by tonight or saturday morning
priority
dg implementation send run feature refactor description to focus on run instead as the request feature talks about send extensively update to the uml diagram be ready by tonight or saturday morning
1
196,931
6,950,957,130
IssuesEvent
2017-12-06 12:49:50
elementary/music
https://api.github.com/repos/elementary/music
closed
Album Art not populated until a song from any given album is played first (art embedded in ID3 tag)
Priority: Medium
Like it says on the tin. I have ~85G of music. After waiting the 30-60 minutes for it to scan and populate the noise database, I noticed two things: 1. Noise was frozen, and taking up 100% of one of the CPU cores 2. Not a single album cover, from any album or single, was visible After restarting the product, I played a song. About 4 seconds into the song, the ablum art showed up, and remains visible to this day. I'm not sure if this is planned behavior or not, but since noise already scans all of the files' metadata to populate the DB, would it not make sense to also scan for the cover and populate the DB (or .local/cache/noise directory) with the covers in 1 shot? I think I would have been OK with waiting slightly longer and seeing all of my album art when the scan was done as opposed to how it works right now, where I've been playing songs on shuffle and only have the albums who's songs have played be populated with art. Launchpad Details: [#LP1215558](https://bugs.launchpad.net/bugs/1215558) GonzO - 2013-08-22 17:28:04 +0000
1.0
Album Art not populated until a song from any given album is played first (art embedded in ID3 tag) - Like it says on the tin. I have ~85G of music. After waiting the 30-60 minutes for it to scan and populate the noise database, I noticed two things: 1. Noise was frozen, and taking up 100% of one of the CPU cores 2. Not a single album cover, from any album or single, was visible After restarting the product, I played a song. About 4 seconds into the song, the ablum art showed up, and remains visible to this day. I'm not sure if this is planned behavior or not, but since noise already scans all of the files' metadata to populate the DB, would it not make sense to also scan for the cover and populate the DB (or .local/cache/noise directory) with the covers in 1 shot? I think I would have been OK with waiting slightly longer and seeing all of my album art when the scan was done as opposed to how it works right now, where I've been playing songs on shuffle and only have the albums who's songs have played be populated with art. Launchpad Details: [#LP1215558](https://bugs.launchpad.net/bugs/1215558) GonzO - 2013-08-22 17:28:04 +0000
priority
album art not populated until a song from any given album is played first art embedded in tag like it says on the tin i have of music after waiting the minutes for it to scan and populate the noise database i noticed two things noise was frozen and taking up of one of the cpu cores not a single album cover from any album or single was visible after restarting the product i played a song about seconds into the song the ablum art showed up and remains visible to this day i m not sure if this is planned behavior or not but since noise already scans all of the files metadata to populate the db would it not make sense to also scan for the cover and populate the db or local cache noise directory with the covers in shot i think i would have been ok with waiting slightly longer and seeing all of my album art when the scan was done as opposed to how it works right now where i ve been playing songs on shuffle and only have the albums who s songs have played be populated with art launchpad details gonzo
1
641,717
20,832,832,181
IssuesEvent
2022-03-19 18:29:02
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
LoRaWAN: Unable to correctly join networks of any version on LTS
bug priority: medium LTS area: LoRa
**Describe the bug** The LoRaWAN subsystem is unable to correctly join networks of any LoRaWAN protocol version on the LTS branch. LoRaWAN fork: https://github.com/zephyrproject-rtos/loramac-node The relevant versions for this issue: 1. v1.0.3 (Previous version) 2. v1.0.4 (Current version targeted by the Zephyr fork) 3. v1.1.x (Not supported at all by the release that the Zephyr fork is based upon) The PR https://github.com/zephyrproject-rtos/loramac-node/pull/9 updated the release the fork was based upon from some older version (v1.0.3?) to v1.0.4. Part of the specification changes from v1.0.3 to v1.0.4 was a change in requirements around the device nonce value, sent in the join request as a way of avoiding replay attacks. *LoRaWAN v1.0.3 Specification (6.2.4)* > DevNonce is a random value.1 For each end-device, the network server keeps track of a certain number of DevNonce values used by the end-device in the past, and ignores join requests with any of these DevNonce values from that end-device. *LoRaWAN v1.0.4 Specification (6.2.5)* > DevNonce is a counter starting at 0 when the end-device is initially powered up and incremented with every Join-Request. A DevNonce value SHALL never be reused for a given JoinEUI value. If the end-device can be power-cycled, then DevNonce SHALL be persistent (e.g., stored in a non-volatile memory). Resetting DevNonce without changing JoinEUI will cause the Join Server to discard the Join-Requests of the end-device. For each end-device, the Join Server keeps track of the last DevNonce value used by the end device and ignores Join-Requests if DevNonce is not incremented. With the changes in https://github.com/zephyrproject-rtos/loramac-node/pull/9, the DevNonce is no longer a random value, but instead restarts from 0 on each reboot. This results in network servers expecting either v1.0.3 or v1.04 to reject join requests due to duplicate DevNonce values. For a v1.0.3 we need to wait for the network servers internal cache to be evicted. For v1.0.4 we need to send an additional join request each time the device reboots. The Nth reboot will require at least N join requests before a join has a chance of succeeding. In both cases, for a device compliant with the join request backoff recommendations, this can quickly take hours. At the same time, the LoRaWAN subsystem does not expose a mechanism for storing and restoring the NVM context required for correct operation under v1.0.4. This is currently a work in progress in #40193 This issue was picked up on in #36571, but was not followed through to the logical conclusion that we are now unable to properly join any network server compliant with the spec. **To Reproduce** Build the `zephyr/samples/subsys/lorawan/class_a` sample, and attempt to join a LoRaWAN network multiple times without clearing anything on the network server. **Expected behavior** The LoRaWAN subsystem should be able to correctly join network servers, ideally either v1.0.3 or v1.0.4 versions. **Impact** LoRaWAN functionality is significantly degraded when devices reboot for any reason (Faults, OTA upgrades etc). The degradation increases with each reboot. **Proposed Solution** Merging #40193 into the LTS branch is likely too large a change to be accepted. Providing a mechanism to query and restore the loramac-node crypto NVS state is sufficient to fix this issue for both network types. This can be done by hooking into the `NvmDataChange` callback without additional changes to loramac-node. **Environment (please complete the following information):** - v2.7.1 LTS
1.0
LoRaWAN: Unable to correctly join networks of any version on LTS - **Describe the bug** The LoRaWAN subsystem is unable to correctly join networks of any LoRaWAN protocol version on the LTS branch. LoRaWAN fork: https://github.com/zephyrproject-rtos/loramac-node The relevant versions for this issue: 1. v1.0.3 (Previous version) 2. v1.0.4 (Current version targeted by the Zephyr fork) 3. v1.1.x (Not supported at all by the release that the Zephyr fork is based upon) The PR https://github.com/zephyrproject-rtos/loramac-node/pull/9 updated the release the fork was based upon from some older version (v1.0.3?) to v1.0.4. Part of the specification changes from v1.0.3 to v1.0.4 was a change in requirements around the device nonce value, sent in the join request as a way of avoiding replay attacks. *LoRaWAN v1.0.3 Specification (6.2.4)* > DevNonce is a random value.1 For each end-device, the network server keeps track of a certain number of DevNonce values used by the end-device in the past, and ignores join requests with any of these DevNonce values from that end-device. *LoRaWAN v1.0.4 Specification (6.2.5)* > DevNonce is a counter starting at 0 when the end-device is initially powered up and incremented with every Join-Request. A DevNonce value SHALL never be reused for a given JoinEUI value. If the end-device can be power-cycled, then DevNonce SHALL be persistent (e.g., stored in a non-volatile memory). Resetting DevNonce without changing JoinEUI will cause the Join Server to discard the Join-Requests of the end-device. For each end-device, the Join Server keeps track of the last DevNonce value used by the end device and ignores Join-Requests if DevNonce is not incremented. With the changes in https://github.com/zephyrproject-rtos/loramac-node/pull/9, the DevNonce is no longer a random value, but instead restarts from 0 on each reboot. This results in network servers expecting either v1.0.3 or v1.04 to reject join requests due to duplicate DevNonce values. For a v1.0.3 we need to wait for the network servers internal cache to be evicted. For v1.0.4 we need to send an additional join request each time the device reboots. The Nth reboot will require at least N join requests before a join has a chance of succeeding. In both cases, for a device compliant with the join request backoff recommendations, this can quickly take hours. At the same time, the LoRaWAN subsystem does not expose a mechanism for storing and restoring the NVM context required for correct operation under v1.0.4. This is currently a work in progress in #40193 This issue was picked up on in #36571, but was not followed through to the logical conclusion that we are now unable to properly join any network server compliant with the spec. **To Reproduce** Build the `zephyr/samples/subsys/lorawan/class_a` sample, and attempt to join a LoRaWAN network multiple times without clearing anything on the network server. **Expected behavior** The LoRaWAN subsystem should be able to correctly join network servers, ideally either v1.0.3 or v1.0.4 versions. **Impact** LoRaWAN functionality is significantly degraded when devices reboot for any reason (Faults, OTA upgrades etc). The degradation increases with each reboot. **Proposed Solution** Merging #40193 into the LTS branch is likely too large a change to be accepted. Providing a mechanism to query and restore the loramac-node crypto NVS state is sufficient to fix this issue for both network types. This can be done by hooking into the `NvmDataChange` callback without additional changes to loramac-node. **Environment (please complete the following information):** - v2.7.1 LTS
priority
lorawan unable to correctly join networks of any version on lts describe the bug the lorawan subsystem is unable to correctly join networks of any lorawan protocol version on the lts branch lorawan fork the relevant versions for this issue previous version current version targeted by the zephyr fork x not supported at all by the release that the zephyr fork is based upon the pr updated the release the fork was based upon from some older version to part of the specification changes from to was a change in requirements around the device nonce value sent in the join request as a way of avoiding replay attacks lorawan specification devnonce is a random value for each end device the network server keeps track of a certain number of devnonce values used by the end device in the past and ignores join requests with any of these devnonce values from that end device lorawan specification devnonce is a counter starting at when the end device is initially powered up and incremented with every join request a devnonce value shall never be reused for a given joineui value if the end device can be power cycled then devnonce shall be persistent e g stored in a non volatile memory resetting devnonce without changing joineui will cause the join server to discard the join requests of the end device for each end device the join server keeps track of the last devnonce value used by the end device and ignores join requests if devnonce is not incremented with the changes in the devnonce is no longer a random value but instead restarts from on each reboot this results in network servers expecting either or to reject join requests due to duplicate devnonce values for a we need to wait for the network servers internal cache to be evicted for we need to send an additional join request each time the device reboots the nth reboot will require at least n join requests before a join has a chance of succeeding in both cases for a device compliant with the join request backoff recommendations this can quickly take hours at the same time the lorawan subsystem does not expose a mechanism for storing and restoring the nvm context required for correct operation under this is currently a work in progress in this issue was picked up on in but was not followed through to the logical conclusion that we are now unable to properly join any network server compliant with the spec to reproduce build the zephyr samples subsys lorawan class a sample and attempt to join a lorawan network multiple times without clearing anything on the network server expected behavior the lorawan subsystem should be able to correctly join network servers ideally either or versions impact lorawan functionality is significantly degraded when devices reboot for any reason faults ota upgrades etc the degradation increases with each reboot proposed solution merging into the lts branch is likely too large a change to be accepted providing a mechanism to query and restore the loramac node crypto nvs state is sufficient to fix this issue for both network types this can be done by hooking into the nvmdatachange callback without additional changes to loramac node environment please complete the following information lts
1
266,231
8,364,461,107
IssuesEvent
2018-10-03 23:09:46
nelsonnyland/WorldwideMovieDatabase
https://api.github.com/repos/nelsonnyland/WorldwideMovieDatabase
opened
Create an example mockup of the home page.
good first issue medium priority
We should design an example sketch or diagram for what we want the home page to look like. We could use Visio or just a normal drawing. We could then upload it to Github and add it to our project in a resource folder. We should make sure to label the diagram so we know what each piece and widget does.
1.0
Create an example mockup of the home page. - We should design an example sketch or diagram for what we want the home page to look like. We could use Visio or just a normal drawing. We could then upload it to Github and add it to our project in a resource folder. We should make sure to label the diagram so we know what each piece and widget does.
priority
create an example mockup of the home page we should design an example sketch or diagram for what we want the home page to look like we could use visio or just a normal drawing we could then upload it to github and add it to our project in a resource folder we should make sure to label the diagram so we know what each piece and widget does
1
634,424
20,361,023,107
IssuesEvent
2022-02-20 17:35:01
moonstar-x/discord-tts-bot
https://api.github.com/repos/moonstar-x/discord-tts-bot
opened
Add support for TTS Tool - Microsoft and Amazon voices.
Priority: Medium Type: Feature Request
#### :zap: Describe the New Feature The site https://ttstool.com/ exposes a public API to use with some voices from Microsoft and Amazon engines. I think maybe the Amazon ones would be duplicate as specified in #71, however this tool supports SSML and some options for volume, pitch, and rate. To request this API, send a POST to: ```text https://support.readaloud.app/ttstool/createParts ``` With a body shaped like this: ```text [{ "voiceId": "Amazon Polish (Ewa)", "ssml": "<speak version=\"1.0\" xml:lang=\"pl-PL\"><prosody volume='x-loud' rate='x-fast' pitch='x-high'>yeah</prosody></speak>" }] ``` This returns a an JSON array with the IDs of the voices. You can technically pass in multiple items to the previous array to get multiple TTS streams in the same request. The response looks like this: ```text [ "efh7l3ll" ] ``` You can then use this ID in a GET request like this: ```text https://support.readaloud.app/ttstool/getParts?q=efh7l3ll ``` This returns an `audio/mpeg` stream.
1.0
Add support for TTS Tool - Microsoft and Amazon voices. - #### :zap: Describe the New Feature The site https://ttstool.com/ exposes a public API to use with some voices from Microsoft and Amazon engines. I think maybe the Amazon ones would be duplicate as specified in #71, however this tool supports SSML and some options for volume, pitch, and rate. To request this API, send a POST to: ```text https://support.readaloud.app/ttstool/createParts ``` With a body shaped like this: ```text [{ "voiceId": "Amazon Polish (Ewa)", "ssml": "<speak version=\"1.0\" xml:lang=\"pl-PL\"><prosody volume='x-loud' rate='x-fast' pitch='x-high'>yeah</prosody></speak>" }] ``` This returns a an JSON array with the IDs of the voices. You can technically pass in multiple items to the previous array to get multiple TTS streams in the same request. The response looks like this: ```text [ "efh7l3ll" ] ``` You can then use this ID in a GET request like this: ```text https://support.readaloud.app/ttstool/getParts?q=efh7l3ll ``` This returns an `audio/mpeg` stream.
priority
add support for tts tool microsoft and amazon voices zap describe the new feature the site exposes a public api to use with some voices from microsoft and amazon engines i think maybe the amazon ones would be duplicate as specified in however this tool supports ssml and some options for volume pitch and rate to request this api send a post to text with a body shaped like this text voiceid amazon polish ewa ssml yeah this returns a an json array with the ids of the voices you can technically pass in multiple items to the previous array to get multiple tts streams in the same request the response looks like this text you can then use this id in a get request like this text this returns an audio mpeg stream
1
91,942
3,863,517,812
IssuesEvent
2016-04-08 09:45:57
iamxavier/elmah
https://api.github.com/repos/iamxavier/elmah
closed
Support Logging to Azure Table Storage
auto-migrated Priority-Medium Type-Enhancement
``` What new or enhanced feature are you proposing? Add support for logging to Azure Table Storage What goal would this enhancement help you achieve? Ability to use ELMAH with Windows Azure + Table Storage. ``` Original issue reported on code.google.com by `pbsbac...@gmail.com` on 27 May 2009 at 1:30
1.0
Support Logging to Azure Table Storage - ``` What new or enhanced feature are you proposing? Add support for logging to Azure Table Storage What goal would this enhancement help you achieve? Ability to use ELMAH with Windows Azure + Table Storage. ``` Original issue reported on code.google.com by `pbsbac...@gmail.com` on 27 May 2009 at 1:30
priority
support logging to azure table storage what new or enhanced feature are you proposing add support for logging to azure table storage what goal would this enhancement help you achieve ability to use elmah with windows azure table storage original issue reported on code google com by pbsbac gmail com on may at
1
38,662
2,849,832,411
IssuesEvent
2015-05-31 01:49:36
GiraffaFS/giraffa
https://api.github.com/repos/GiraffaFS/giraffa
closed
Release Giraffa 0.2.0
Priority-Medium Type-Task
Release Steps: * Move remaining issues from Milestone 0.2.0 to Milestone 0.3.0 * Update poms to 0.2.0 * Create release tag, named release_0.2.0 * Update poms to 0..3.0-SNAPSHOT * Update Wiki documentation to include new release. Include Release notes generated from git messages * Run Jenkins RELEASE build for new release option.
1.0
Release Giraffa 0.2.0 - Release Steps: * Move remaining issues from Milestone 0.2.0 to Milestone 0.3.0 * Update poms to 0.2.0 * Create release tag, named release_0.2.0 * Update poms to 0..3.0-SNAPSHOT * Update Wiki documentation to include new release. Include Release notes generated from git messages * Run Jenkins RELEASE build for new release option.
priority
release giraffa release steps move remaining issues from milestone to milestone update poms to create release tag named release update poms to snapshot update wiki documentation to include new release include release notes generated from git messages run jenkins release build for new release option
1
599,578
18,277,556,473
IssuesEvent
2021-10-04 20:49:22
poissonconsulting/ypr
https://api.github.com/repos/poissonconsulting/ypr
opened
ypr_tabulate_yield add population name column to table
Effort: 2 Medium Priority: 4 Low Type: Enhancement Difficulty: 1 Simple
## Your Idea The table that is outputted by ypr_tabulate_yield.ypr_populations could have a column that adds the names of each population to the output of the table so you can easily tell which is which
1.0
ypr_tabulate_yield add population name column to table - ## Your Idea The table that is outputted by ypr_tabulate_yield.ypr_populations could have a column that adds the names of each population to the output of the table so you can easily tell which is which
priority
ypr tabulate yield add population name column to table your idea the table that is outputted by ypr tabulate yield ypr populations could have a column that adds the names of each population to the output of the table so you can easily tell which is which
1
805,570
29,577,045,728
IssuesEvent
2023-06-07 00:16:51
TimeWalkOrg/Manhattan
https://api.github.com/repos/TimeWalkOrg/Manhattan
closed
Creek origins should be more natural
2 Medium Priority
The ends of the creeks (where they appear to originate) look unnatural. The water is springing out of the rock like a powerful spring, though this is a small island. Can we add a few smaller grooves uphill of the "origin" that are dry that appear to "feed" the creeks? Slowing down the water (per Issue #3 ) will help. Can we maybe cover the origins of the creeks with dead leaf and small rock cover of some kind? ![1609_manhattan_ue5_south_tip - ISSUES](https://github.com/TimeWalkOrg/Manhattan/assets/694013/fb460bac-db50-4684-9fc7-9ed3e7fbe2af)
1.0
Creek origins should be more natural - The ends of the creeks (where they appear to originate) look unnatural. The water is springing out of the rock like a powerful spring, though this is a small island. Can we add a few smaller grooves uphill of the "origin" that are dry that appear to "feed" the creeks? Slowing down the water (per Issue #3 ) will help. Can we maybe cover the origins of the creeks with dead leaf and small rock cover of some kind? ![1609_manhattan_ue5_south_tip - ISSUES](https://github.com/TimeWalkOrg/Manhattan/assets/694013/fb460bac-db50-4684-9fc7-9ed3e7fbe2af)
priority
creek origins should be more natural the ends of the creeks where they appear to originate look unnatural the water is springing out of the rock like a powerful spring though this is a small island can we add a few smaller grooves uphill of the origin that are dry that appear to feed the creeks slowing down the water per issue will help can we maybe cover the origins of the creeks with dead leaf and small rock cover of some kind
1
248,665
7,934,747,908
IssuesEvent
2018-07-08 23:00:17
SETI/pds-opus
https://api.github.com/repos/SETI/pds-opus
opened
Sort on multiple fields
A-Enhancement B-OPUS Django B-OPUS JS Effort 2 Medium Priority TBD
Currently it's possible to sort on a single field by using the "view table" view and clicking on a column header. However, there are times you want to sort on multiple columns, since columns like "planet" or "mission" can have a large number of entries that are all identical. For example, you might want to sort on "Target Body" in ascending order, then on "Filter" in ascending order, then on "Observation Duration" in descending order. The JPL Imaging Node "Image Atlas" allows this kind of searching. We should too.
1.0
Sort on multiple fields - Currently it's possible to sort on a single field by using the "view table" view and clicking on a column header. However, there are times you want to sort on multiple columns, since columns like "planet" or "mission" can have a large number of entries that are all identical. For example, you might want to sort on "Target Body" in ascending order, then on "Filter" in ascending order, then on "Observation Duration" in descending order. The JPL Imaging Node "Image Atlas" allows this kind of searching. We should too.
priority
sort on multiple fields currently it s possible to sort on a single field by using the view table view and clicking on a column header however there are times you want to sort on multiple columns since columns like planet or mission can have a large number of entries that are all identical for example you might want to sort on target body in ascending order then on filter in ascending order then on observation duration in descending order the jpl imaging node image atlas allows this kind of searching we should too
1
417,369
12,158,999,347
IssuesEvent
2020-04-26 07:04:25
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
opened
Check for repeating error events from pods in openshift-storage namespace
Medium Priority
This check can be done as a single test case or at the end of relevant test runs. Relevant test runs means those runs which will not introduce any errors which will in turn trigger error events in pods. eg: tier1 runs. This will also cover bugs like this - [1777591](https://bugzilla.redhat.com/show_bug.cgi?id=1777591)
1.0
Check for repeating error events from pods in openshift-storage namespace - This check can be done as a single test case or at the end of relevant test runs. Relevant test runs means those runs which will not introduce any errors which will in turn trigger error events in pods. eg: tier1 runs. This will also cover bugs like this - [1777591](https://bugzilla.redhat.com/show_bug.cgi?id=1777591)
priority
check for repeating error events from pods in openshift storage namespace this check can be done as a single test case or at the end of relevant test runs relevant test runs means those runs which will not introduce any errors which will in turn trigger error events in pods eg runs this will also cover bugs like this
1
500,904
14,516,983,689
IssuesEvent
2020-12-13 17:53:14
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Some Installations of ui_next do not render anything in UI on MS Edge Classic and IE11
component:ui_next priority:medium state:needs_devel type:bug
##### ISSUE TYPE - Bug Report ##### SUMMARY In the bug bash, we set up a dev server to point to a remote tower api server. When a user of IE11 or Edge Classic attempts to load the page, they see nothing. No errors in the network or console logs. This has a self signed cert pointed to a yolo'd tower with it's own cert. This page loads in the chromium-based browsers and firefox. ##### ENVIRONMENT * AWX version: 9.2.0 * AWX install method: ui_next local dev install on a remote ec2 target * Operating System: Windows 10 * Web Browser: IE11 and Edge Classic (18) ##### STEPS TO REPRODUCE 1. Log in to web UI on IE11 or Edge Classic ##### EXPECTED RESULTS Functional UI ##### ACTUAL RESULTS Blank page. No console or network errors
1.0
Some Installations of ui_next do not render anything in UI on MS Edge Classic and IE11 - ##### ISSUE TYPE - Bug Report ##### SUMMARY In the bug bash, we set up a dev server to point to a remote tower api server. When a user of IE11 or Edge Classic attempts to load the page, they see nothing. No errors in the network or console logs. This has a self signed cert pointed to a yolo'd tower with it's own cert. This page loads in the chromium-based browsers and firefox. ##### ENVIRONMENT * AWX version: 9.2.0 * AWX install method: ui_next local dev install on a remote ec2 target * Operating System: Windows 10 * Web Browser: IE11 and Edge Classic (18) ##### STEPS TO REPRODUCE 1. Log in to web UI on IE11 or Edge Classic ##### EXPECTED RESULTS Functional UI ##### ACTUAL RESULTS Blank page. No console or network errors
priority
some installations of ui next do not render anything in ui on ms edge classic and issue type bug report summary in the bug bash we set up a dev server to point to a remote tower api server when a user of or edge classic attempts to load the page they see nothing no errors in the network or console logs this has a self signed cert pointed to a yolo d tower with it s own cert this page loads in the chromium based browsers and firefox environment awx version awx install method ui next local dev install on a remote target operating system windows web browser and edge classic steps to reproduce log in to web ui on or edge classic expected results functional ui actual results blank page no console or network errors
1
652,791
21,561,960,306
IssuesEvent
2022-05-01 09:47:45
xournalpp/xournalpp
https://api.github.com/repos/xournalpp/xournalpp
closed
Make colors configurable with only stylus input and option to add configured color bins to toolbar
enhancement priority::medium
I think additional configurable color slots, that keep the colour, after you have selected a standard color, would be nice. A colour wheel would also be interesting
1.0
Make colors configurable with only stylus input and option to add configured color bins to toolbar - I think additional configurable color slots, that keep the colour, after you have selected a standard color, would be nice. A colour wheel would also be interesting
priority
make colors configurable with only stylus input and option to add configured color bins to toolbar i think additional configurable color slots that keep the colour after you have selected a standard color would be nice a colour wheel would also be interesting
1
230,706
7,613,018,389
IssuesEvent
2018-05-01 19:40:15
fgpv-vpgf/fgpv-vpgf
https://api.github.com/repos/fgpv-vpgf/fgpv-vpgf
closed
Legend toggling and data table viewing issue with attributes that contain apostrophes (single quotes)
bug-type: regression priority: medium problem: bug
Layers displayed with queries on stings that contain apostrophes (single quotes) throw an error. ![apostrophe](https://user-images.githubusercontent.com/2006029/36858303-c8d9bd78-1d48-11e8-858d-34e6d45c7ea4.gif) `http://geoappext.nrcan.gc.ca/arcgis/rest/services/MMS/IMA_F/MapServer/0/query?f=json&where=(TYPE_ENTENTE_NOM%20%3D%20'Protocole%20d'entente')&returnGeometry=true&spatialRel=esriSpatialRelIntersects&geometry=%7B%22xmin%22%3A-4074491.517018034%2C%22ymin%22%3A-192876.4845059691%2C%22xmax%22%3A-10483.389001777861%2C%22ymax%22%3A3871131.643510287%2C%22spatialReference%22%3A%7B%22wkid%22%3A3978%7D%7D&geometryType=esriGeometryEnvelope&inSR=3978&outFields=OBJECTID%2CTYPE_ENTENTE_NOM%2CPROJECT_NAME_FR&outSR=3978` `{"error":{"code":400,"message":"Unable to complete operation.","details":[]}}`
1.0
Legend toggling and data table viewing issue with attributes that contain apostrophes (single quotes) - Layers displayed with queries on stings that contain apostrophes (single quotes) throw an error. ![apostrophe](https://user-images.githubusercontent.com/2006029/36858303-c8d9bd78-1d48-11e8-858d-34e6d45c7ea4.gif) `http://geoappext.nrcan.gc.ca/arcgis/rest/services/MMS/IMA_F/MapServer/0/query?f=json&where=(TYPE_ENTENTE_NOM%20%3D%20'Protocole%20d'entente')&returnGeometry=true&spatialRel=esriSpatialRelIntersects&geometry=%7B%22xmin%22%3A-4074491.517018034%2C%22ymin%22%3A-192876.4845059691%2C%22xmax%22%3A-10483.389001777861%2C%22ymax%22%3A3871131.643510287%2C%22spatialReference%22%3A%7B%22wkid%22%3A3978%7D%7D&geometryType=esriGeometryEnvelope&inSR=3978&outFields=OBJECTID%2CTYPE_ENTENTE_NOM%2CPROJECT_NAME_FR&outSR=3978` `{"error":{"code":400,"message":"Unable to complete operation.","details":[]}}`
priority
legend toggling and data table viewing issue with attributes that contain apostrophes single quotes layers displayed with queries on stings that contain apostrophes single quotes throw an error error code message unable to complete operation details
1
807,897
30,023,443,319
IssuesEvent
2023-06-27 02:49:15
AnOpenSauceDev/Methane-mod
https://api.github.com/repos/AnOpenSauceDev/Methane-mod
closed
[Issue] 1.20 is broken
enhancement compat Methane-Next priority:medium
## Description of what happened methane worked fine till 23w14a (by editing fabric.mod.json), but it broke for 23w16a ## Mods used methane 1.7 cloth config 10.0.96 fabric api for 23w16a ## possible ways to replicate this bug join a server; then it crashes maybe u already want to update methane to 1.20 might be an issue with MatrixStack screen. not sure tho it seems to be a fundamental change, because all 1.19.4 mods that worked fine for 23w14a broke for 23w16a (game crashes while launching/joining server) ## crashreport ---- Minecraft Crash Report ---- // Shall we play a game? Time: 2023-04-21 01:03:34 Description: Unexpected error java.lang.AbstractMethodError: Receiver class me.wolfie.methane.client.HudRenderListener does not define or inherit an implementation of the resolved method 'abstract void onHudRender(net.minecraft.class_332, float)' of interface net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback. at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27) at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393) at net.minecraft.class_329.method_1753(class_329.java:371) at net.minecraft.class_757.method_3192(class_757.java:918) at net.minecraft.class_310.method_1523(class_310.java:1203) at net.minecraft.class_310.method_1514(class_310.java:786) at net.minecraft.client.main.Main.main(Main.java:240) at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462) at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74) at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23) A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- Head -- Thread: Render thread Stacktrace: at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27) at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393) at net.minecraft.class_329.method_1753(class_329.java:371) -- Affected level -- Details: All players: 1 total; [class_746['Cyclopropinon'/2857, l='ClientLevel', x=192.87, y=133.00, z=-246.02]] Chunk stats: 729, 248 Level dimension: minecraft:the_nether Level spawn location: World: (0,73,0), Section: (at 0,9,0 in 0,4,0; chunk contains blocks 0,0,0 to 15,255,15), Region: (0,0; contains chunks 0,0 to 31,31, blocks 0,0,0 to 511,255,511) Level time: 119784655 game time, 129434541 day time Server brand: fabric Server type: Non-integrated multiplayer server Stacktrace: at net.minecraft.class_638.method_8538(class_638.java:455) at net.minecraft.class_310.method_1587(class_310.java:2394) at net.minecraft.class_310.method_1514(class_310.java:810) at net.minecraft.client.main.Main.main(Main.java:240) at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462) at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74) at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23) -- Last reload -- Details: Reload number: 1 Reload reason: initial Finished: Yes Packs: vanilla, Fabric Mods -- System Details -- Details: Minecraft Version: 23w16a Minecraft Version ID: 23w16a Operating System: Windows 10 (amd64) version 10.0 Java Version: 17.0.3, Microsoft Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Microsoft Memory: 359664912 bytes (343 MiB) / 671088640 bytes (640 MiB) up to 2147483648 bytes (2048 MiB) CPUs: 4 Processor Vendor: GenuineIntel Processor Name: Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz Identifier: Intel64 Family 6 Model 37 Stepping 5 Microarchitecture: Westmere (Client) Frequency (GHz): 2.66 Number of physical packages: 1 Number of physical CPUs: 2 Number of logical CPUs: 4 Graphics card #0 name: NVIDIA NVS 3100M Graphics card #0 vendor: NVIDIA (0x10de) Graphics card #0 VRAM (MB): 512.00 Graphics card #0 deviceId: 0x0a6c Graphics card #0 versionInfo: DriverVersion=21.21.13.4201 Memory slot #0 capacity (MB): 4096.00 Memory slot #0 clockSpeed (GHz): 1.33 Memory slot #0 type: DDR3 Virtual memory max (MB): 15219.67 Virtual memory used (MB): 10379.44 Swap memory total (MB): 11264.00 Swap memory used (MB): 2109.98 JVM Flags: 9 total; -XX:HeapDumpPath=MojangTricksIntelDriversForPerformance_javaw.exe_minecraft.exe.heapdump -Xss1M -Xmx2G -XX:+UnlockExperimentalVMOptions -XX:+UseG1GC -XX:G1NewSizePercent=20 -XX:G1ReservePercent=20 -XX:MaxGCPauseMillis=50 -XX:G1HeapRegionSize=32M Fabric Mods: cloth-config: Cloth Config v10 10.0.96 cloth-basic-math: cloth-basic-math 0.6.1 fabric-api: Fabric API 0.78.1+1.20 fabric-api-base: Fabric API Base 0.4.23+9ff28bce67 fabric-api-lookup-api-v1: Fabric API Lookup API (v1) 1.6.26+eff2638667 fabric-biome-api-v1: Fabric Biome API (v1) 13.0.6+348a9c6467 fabric-block-api-v1: Fabric Block API (v1) 1.0.5+e022e5d167 fabric-blockrenderlayer-v1: Fabric BlockRenderLayer Registration (v1) 1.1.33+c2e6f67467 fabric-client-tags-api-v1: Fabric Client Tags 1.0.14+1134c5b867 fabric-command-api-v1: Fabric Command API (v1) 1.2.26+f71b366f67 fabric-command-api-v2: Fabric Command API (v2) 2.2.5+df5b2a9d67 fabric-commands-v0: Fabric Commands (v0) 0.2.43+df3654b367 fabric-containers-v0: Fabric Containers (v0) 0.1.55+df3654b367 fabric-content-registries-v0: Fabric Content Registries (v0) 4.0.0+eff2638667 fabric-convention-tags-v1: Fabric Convention Tags 1.4.0+9a7c5daa67 fabric-crash-report-info-v1: Fabric Crash Report Info (v1) 0.2.14+aeb40ebe67 fabric-data-generation-api-v1: Fabric Data Generation API (v1) 12.0.0+eff2638667 fabric-dimensions-v1: Fabric Dimensions API (v1) 2.1.44+7f87f8fa67 fabric-entity-events-v1: Fabric Entity Events (v1) 1.5.14+eff2638667 fabric-events-interaction-v0: Fabric Events Interaction (v0) 0.4.42+a1ccd7bf67 fabric-events-lifecycle-v0: Fabric Events Lifecycle (v0) 0.2.53+df3654b367 fabric-game-rule-api-v1: Fabric Game Rule API (v1) 1.0.33+eff2638667 fabric-item-api-v1: Fabric Item API (v1) 2.1.18+eff2638667 fabric-item-group-api-v1: Fabric Item Group API (v1) 4.0.0+eff2638667 fabric-key-binding-api-v1: Fabric Key Binding API (v1) 1.0.32+c477957e67 fabric-keybindings-v0: Fabric Key Bindings (v0) 0.2.30+df3654b367 fabric-lifecycle-events-v1: Fabric Lifecycle Events (v1) 2.2.14+5da15ca167 fabric-loot-api-v2: Fabric Loot API (v2) 1.1.29+eff2638667 fabric-loot-tables-v1: Fabric Loot Tables (v1) 1.1.33+9e7660c667 fabric-message-api-v1: Fabric Message API (v1) 5.1.0+1ee8be4067 fabric-mining-level-api-v1: Fabric Mining Level API (v1) 2.1.39+eff2638667 fabric-models-v0: Fabric Models (v0) 0.3.29+11ba9c3b67 fabric-networking-api-v1: Fabric Networking API (v1) 1.3.2+eff2638667 fabric-networking-v0: Fabric Networking (v0) 0.3.42+df3654b367 fabric-object-builder-api-v1: Fabric Object Builder API (v1) 10.0.0+eff2638667 fabric-particles-v1: Fabric Particles (v1) 1.0.22+f1e4495b67 fabric-recipe-api-v1: Fabric Recipe API (v1) 1.0.9+a1ccd7bf67 fabric-registry-sync-v0: Fabric Registry Sync (v0) 2.1.5+eff2638667 fabric-renderer-api-v1: Fabric Renderer API (v1) 2.2.5+eff2638667 fabric-renderer-indigo: Fabric Renderer - Indigo 1.1.1+81e8c57667 fabric-renderer-registries-v1: Fabric Renderer Registries (v1) 3.2.38+df3654b367 fabric-rendering-data-attachment-v1: Fabric Rendering Data Attachment (v1) 0.3.27+afca2f3e67 fabric-rendering-fluids-v1: Fabric Rendering Fluids (v1) 3.0.20+f1e4495b67 fabric-rendering-v0: Fabric Rendering (v0) 1.1.41+df3654b367 fabric-rendering-v1: Fabric Rendering (v1) 3.0.0+eff2638667 fabric-resource-conditions-api-v1: Fabric Resource Conditions API (v1) 2.3.0+e6c7d4ee67 fabric-resource-loader-v0: Fabric Resource Loader (v0) 0.11.1+03ffe37867 fabric-screen-api-v1: Fabric Screen API (v1) 2.0.0+eff2638667 fabric-screen-handler-api-v1: Fabric Screen Handler API (v1) 1.3.21+eff2638667 fabric-sound-api-v1: Fabric Sound API (v1) 1.0.8+75e9821167 fabric-transfer-api-v1: Fabric Transfer API (v1) 3.1.1+eff2638667 fabric-transitive-access-wideners-v1: Fabric Transitive Access Wideners (v1) 4.0.1+848ffaab67 fabricloader: Fabric Loader 0.14.19 java: OpenJDK 64-Bit Server VM 17 methane: Methane 1.7 minecraft: Minecraft 1.20-alpha.23.16.a Launched Version: fabric-loader-0.14.19-23w16a Backend library: LWJGL version 3.3.1 SNAPSHOT Backend API: NVS 3100M/PCIe/SSE2 GL version 3.2.0, NVIDIA Corporation Window size: 1440x837 GL Caps: Using framebuffer using OpenGL 3.2 GL debug messages: Using VBOs: Yes Is Modded: Definitely; Client brand changed to 'fabric' Type: Client (map_client.txt) Graphics mode: fancy Resource Packs: fabric Current Language: en_us CPU: 4x Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz
1.0
[Issue] 1.20 is broken - ## Description of what happened methane worked fine till 23w14a (by editing fabric.mod.json), but it broke for 23w16a ## Mods used methane 1.7 cloth config 10.0.96 fabric api for 23w16a ## possible ways to replicate this bug join a server; then it crashes maybe u already want to update methane to 1.20 might be an issue with MatrixStack screen. not sure tho it seems to be a fundamental change, because all 1.19.4 mods that worked fine for 23w14a broke for 23w16a (game crashes while launching/joining server) ## crashreport ---- Minecraft Crash Report ---- // Shall we play a game? Time: 2023-04-21 01:03:34 Description: Unexpected error java.lang.AbstractMethodError: Receiver class me.wolfie.methane.client.HudRenderListener does not define or inherit an implementation of the resolved method 'abstract void onHudRender(net.minecraft.class_332, float)' of interface net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback. at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27) at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393) at net.minecraft.class_329.method_1753(class_329.java:371) at net.minecraft.class_757.method_3192(class_757.java:918) at net.minecraft.class_310.method_1523(class_310.java:1203) at net.minecraft.class_310.method_1514(class_310.java:786) at net.minecraft.client.main.Main.main(Main.java:240) at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462) at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74) at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23) A detailed walkthrough of the error, its code path and all known details is as follows: --------------------------------------------------------------------------------------- -- Head -- Thread: Render thread Stacktrace: at net.fabricmc.fabric.api.client.rendering.v1.HudRenderCallback.lambda$static$0(HudRenderCallback.java:27) at net.minecraft.class_329.handler$zhd000$fabric-rendering-v1$render(class_329.java:1393) at net.minecraft.class_329.method_1753(class_329.java:371) -- Affected level -- Details: All players: 1 total; [class_746['Cyclopropinon'/2857, l='ClientLevel', x=192.87, y=133.00, z=-246.02]] Chunk stats: 729, 248 Level dimension: minecraft:the_nether Level spawn location: World: (0,73,0), Section: (at 0,9,0 in 0,4,0; chunk contains blocks 0,0,0 to 15,255,15), Region: (0,0; contains chunks 0,0 to 31,31, blocks 0,0,0 to 511,255,511) Level time: 119784655 game time, 129434541 day time Server brand: fabric Server type: Non-integrated multiplayer server Stacktrace: at net.minecraft.class_638.method_8538(class_638.java:455) at net.minecraft.class_310.method_1587(class_310.java:2394) at net.minecraft.class_310.method_1514(class_310.java:810) at net.minecraft.client.main.Main.main(Main.java:240) at net.fabricmc.loader.impl.game.minecraft.MinecraftGameProvider.launch(MinecraftGameProvider.java:462) at net.fabricmc.loader.impl.launch.knot.Knot.launch(Knot.java:74) at net.fabricmc.loader.impl.launch.knot.KnotClient.main(KnotClient.java:23) -- Last reload -- Details: Reload number: 1 Reload reason: initial Finished: Yes Packs: vanilla, Fabric Mods -- System Details -- Details: Minecraft Version: 23w16a Minecraft Version ID: 23w16a Operating System: Windows 10 (amd64) version 10.0 Java Version: 17.0.3, Microsoft Java VM Version: OpenJDK 64-Bit Server VM (mixed mode), Microsoft Memory: 359664912 bytes (343 MiB) / 671088640 bytes (640 MiB) up to 2147483648 bytes (2048 MiB) CPUs: 4 Processor Vendor: GenuineIntel Processor Name: Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz Identifier: Intel64 Family 6 Model 37 Stepping 5 Microarchitecture: Westmere (Client) Frequency (GHz): 2.66 Number of physical packages: 1 Number of physical CPUs: 2 Number of logical CPUs: 4 Graphics card #0 name: NVIDIA NVS 3100M Graphics card #0 vendor: NVIDIA (0x10de) Graphics card #0 VRAM (MB): 512.00 Graphics card #0 deviceId: 0x0a6c Graphics card #0 versionInfo: DriverVersion=21.21.13.4201 Memory slot #0 capacity (MB): 4096.00 Memory slot #0 clockSpeed (GHz): 1.33 Memory slot #0 type: DDR3 Virtual memory max (MB): 15219.67 Virtual memory used (MB): 10379.44 Swap memory total (MB): 11264.00 Swap memory used (MB): 2109.98 JVM Flags: 9 total; -XX:HeapDumpPath=MojangTricksIntelDriversForPerformance_javaw.exe_minecraft.exe.heapdump -Xss1M -Xmx2G -XX:+UnlockExperimentalVMOptions -XX:+UseG1GC -XX:G1NewSizePercent=20 -XX:G1ReservePercent=20 -XX:MaxGCPauseMillis=50 -XX:G1HeapRegionSize=32M Fabric Mods: cloth-config: Cloth Config v10 10.0.96 cloth-basic-math: cloth-basic-math 0.6.1 fabric-api: Fabric API 0.78.1+1.20 fabric-api-base: Fabric API Base 0.4.23+9ff28bce67 fabric-api-lookup-api-v1: Fabric API Lookup API (v1) 1.6.26+eff2638667 fabric-biome-api-v1: Fabric Biome API (v1) 13.0.6+348a9c6467 fabric-block-api-v1: Fabric Block API (v1) 1.0.5+e022e5d167 fabric-blockrenderlayer-v1: Fabric BlockRenderLayer Registration (v1) 1.1.33+c2e6f67467 fabric-client-tags-api-v1: Fabric Client Tags 1.0.14+1134c5b867 fabric-command-api-v1: Fabric Command API (v1) 1.2.26+f71b366f67 fabric-command-api-v2: Fabric Command API (v2) 2.2.5+df5b2a9d67 fabric-commands-v0: Fabric Commands (v0) 0.2.43+df3654b367 fabric-containers-v0: Fabric Containers (v0) 0.1.55+df3654b367 fabric-content-registries-v0: Fabric Content Registries (v0) 4.0.0+eff2638667 fabric-convention-tags-v1: Fabric Convention Tags 1.4.0+9a7c5daa67 fabric-crash-report-info-v1: Fabric Crash Report Info (v1) 0.2.14+aeb40ebe67 fabric-data-generation-api-v1: Fabric Data Generation API (v1) 12.0.0+eff2638667 fabric-dimensions-v1: Fabric Dimensions API (v1) 2.1.44+7f87f8fa67 fabric-entity-events-v1: Fabric Entity Events (v1) 1.5.14+eff2638667 fabric-events-interaction-v0: Fabric Events Interaction (v0) 0.4.42+a1ccd7bf67 fabric-events-lifecycle-v0: Fabric Events Lifecycle (v0) 0.2.53+df3654b367 fabric-game-rule-api-v1: Fabric Game Rule API (v1) 1.0.33+eff2638667 fabric-item-api-v1: Fabric Item API (v1) 2.1.18+eff2638667 fabric-item-group-api-v1: Fabric Item Group API (v1) 4.0.0+eff2638667 fabric-key-binding-api-v1: Fabric Key Binding API (v1) 1.0.32+c477957e67 fabric-keybindings-v0: Fabric Key Bindings (v0) 0.2.30+df3654b367 fabric-lifecycle-events-v1: Fabric Lifecycle Events (v1) 2.2.14+5da15ca167 fabric-loot-api-v2: Fabric Loot API (v2) 1.1.29+eff2638667 fabric-loot-tables-v1: Fabric Loot Tables (v1) 1.1.33+9e7660c667 fabric-message-api-v1: Fabric Message API (v1) 5.1.0+1ee8be4067 fabric-mining-level-api-v1: Fabric Mining Level API (v1) 2.1.39+eff2638667 fabric-models-v0: Fabric Models (v0) 0.3.29+11ba9c3b67 fabric-networking-api-v1: Fabric Networking API (v1) 1.3.2+eff2638667 fabric-networking-v0: Fabric Networking (v0) 0.3.42+df3654b367 fabric-object-builder-api-v1: Fabric Object Builder API (v1) 10.0.0+eff2638667 fabric-particles-v1: Fabric Particles (v1) 1.0.22+f1e4495b67 fabric-recipe-api-v1: Fabric Recipe API (v1) 1.0.9+a1ccd7bf67 fabric-registry-sync-v0: Fabric Registry Sync (v0) 2.1.5+eff2638667 fabric-renderer-api-v1: Fabric Renderer API (v1) 2.2.5+eff2638667 fabric-renderer-indigo: Fabric Renderer - Indigo 1.1.1+81e8c57667 fabric-renderer-registries-v1: Fabric Renderer Registries (v1) 3.2.38+df3654b367 fabric-rendering-data-attachment-v1: Fabric Rendering Data Attachment (v1) 0.3.27+afca2f3e67 fabric-rendering-fluids-v1: Fabric Rendering Fluids (v1) 3.0.20+f1e4495b67 fabric-rendering-v0: Fabric Rendering (v0) 1.1.41+df3654b367 fabric-rendering-v1: Fabric Rendering (v1) 3.0.0+eff2638667 fabric-resource-conditions-api-v1: Fabric Resource Conditions API (v1) 2.3.0+e6c7d4ee67 fabric-resource-loader-v0: Fabric Resource Loader (v0) 0.11.1+03ffe37867 fabric-screen-api-v1: Fabric Screen API (v1) 2.0.0+eff2638667 fabric-screen-handler-api-v1: Fabric Screen Handler API (v1) 1.3.21+eff2638667 fabric-sound-api-v1: Fabric Sound API (v1) 1.0.8+75e9821167 fabric-transfer-api-v1: Fabric Transfer API (v1) 3.1.1+eff2638667 fabric-transitive-access-wideners-v1: Fabric Transitive Access Wideners (v1) 4.0.1+848ffaab67 fabricloader: Fabric Loader 0.14.19 java: OpenJDK 64-Bit Server VM 17 methane: Methane 1.7 minecraft: Minecraft 1.20-alpha.23.16.a Launched Version: fabric-loader-0.14.19-23w16a Backend library: LWJGL version 3.3.1 SNAPSHOT Backend API: NVS 3100M/PCIe/SSE2 GL version 3.2.0, NVIDIA Corporation Window size: 1440x837 GL Caps: Using framebuffer using OpenGL 3.2 GL debug messages: Using VBOs: Yes Is Modded: Definitely; Client brand changed to 'fabric' Type: Client (map_client.txt) Graphics mode: fancy Resource Packs: fabric Current Language: en_us CPU: 4x Intel(R) Core(TM) i5 CPU M 560 @ 2.67GHz
priority
is broken description of what happened methane worked fine till by editing fabric mod json but it broke for mods used methane cloth config fabric api for possible ways to replicate this bug join a server then it crashes maybe u already want to update methane to might be an issue with matrixstack screen not sure tho it seems to be a fundamental change because all mods that worked fine for broke for game crashes while launching joining server crashreport minecraft crash report shall we play a game time description unexpected error java lang abstractmethoderror receiver class me wolfie methane client hudrenderlistener does not define or inherit an implementation of the resolved method abstract void onhudrender net minecraft class float of interface net fabricmc fabric api client rendering hudrendercallback at net fabricmc fabric api client rendering hudrendercallback lambda static hudrendercallback java at net minecraft class handler fabric rendering render class java at net minecraft class method class java at net minecraft class method class java at net minecraft class method class java at net minecraft class method class java at net minecraft client main main main main java at net fabricmc loader impl game minecraft minecraftgameprovider launch minecraftgameprovider java at net fabricmc loader impl launch knot knot launch knot java at net fabricmc loader impl launch knot knotclient main knotclient java a detailed walkthrough of the error its code path and all known details is as follows head thread render thread stacktrace at net fabricmc fabric api client rendering hudrendercallback lambda static hudrendercallback java at net minecraft class handler fabric rendering render class java at net minecraft class method class java affected level details all players total chunk stats level dimension minecraft the nether level spawn location world section at in chunk contains blocks to region contains chunks to blocks to level time game time day time server brand fabric server type non integrated multiplayer server stacktrace at net minecraft class method class java at net minecraft class method class java at net minecraft class method class java at net minecraft client main main main main java at net fabricmc loader impl game minecraft minecraftgameprovider launch minecraftgameprovider java at net fabricmc loader impl launch knot knot launch knot java at net fabricmc loader impl launch knot knotclient main knotclient java last reload details reload number reload reason initial finished yes packs vanilla fabric mods system details details minecraft version minecraft version id operating system windows version java version microsoft java vm version openjdk bit server vm mixed mode microsoft memory bytes mib bytes mib up to bytes mib cpus processor vendor genuineintel processor name intel r core tm cpu m identifier family model stepping microarchitecture westmere client frequency ghz number of physical packages number of physical cpus number of logical cpus graphics card name nvidia nvs graphics card vendor nvidia graphics card vram mb graphics card deviceid graphics card versioninfo driverversion memory slot capacity mb memory slot clockspeed ghz memory slot type virtual memory max mb virtual memory used mb swap memory total mb swap memory used mb jvm flags total xx heapdumppath mojangtricksinteldriversforperformance javaw exe minecraft exe heapdump xx unlockexperimentalvmoptions xx xx xx xx maxgcpausemillis xx fabric mods cloth config cloth config cloth basic math cloth basic math fabric api fabric api fabric api base fabric api base fabric api lookup api fabric api lookup api fabric biome api fabric biome api fabric block api fabric block api fabric blockrenderlayer fabric blockrenderlayer registration fabric client tags api fabric client tags fabric command api fabric command api fabric command api fabric command api fabric commands fabric commands fabric containers fabric containers fabric content registries fabric content registries fabric convention tags fabric convention tags fabric crash report info fabric crash report info fabric data generation api fabric data generation api fabric dimensions fabric dimensions api fabric entity events fabric entity events fabric events interaction fabric events interaction fabric events lifecycle fabric events lifecycle fabric game rule api fabric game rule api fabric item api fabric item api fabric item group api fabric item group api fabric key binding api fabric key binding api fabric keybindings fabric key bindings fabric lifecycle events fabric lifecycle events fabric loot api fabric loot api fabric loot tables fabric loot tables fabric message api fabric message api fabric mining level api fabric mining level api fabric models fabric models fabric networking api fabric networking api fabric networking fabric networking fabric object builder api fabric object builder api fabric particles fabric particles fabric recipe api fabric recipe api fabric registry sync fabric registry sync fabric renderer api fabric renderer api fabric renderer indigo fabric renderer indigo fabric renderer registries fabric renderer registries fabric rendering data attachment fabric rendering data attachment fabric rendering fluids fabric rendering fluids fabric rendering fabric rendering fabric rendering fabric rendering fabric resource conditions api fabric resource conditions api fabric resource loader fabric resource loader fabric screen api fabric screen api fabric screen handler api fabric screen handler api fabric sound api fabric sound api fabric transfer api fabric transfer api fabric transitive access wideners fabric transitive access wideners fabricloader fabric loader java openjdk bit server vm methane methane minecraft minecraft alpha a launched version fabric loader backend library lwjgl version snapshot backend api nvs pcie gl version nvidia corporation window size gl caps using framebuffer using opengl gl debug messages using vbos yes is modded definitely client brand changed to fabric type client map client txt graphics mode fancy resource packs fabric current language en us cpu intel r core tm cpu m
1
247,842
7,924,251,041
IssuesEvent
2018-07-05 16:19:01
uracreative/task-management
https://api.github.com/repos/uracreative/task-management
closed
OTF report June 2018
Internal: Operations Internal: Overdue Priority: Medium
Documentation for monthly report of activities of Ura Design as part of OTF. Please add here on bullet points every time we have done tasks related to OTF. Deadline: 26.06.2018.
1.0
OTF report June 2018 - Documentation for monthly report of activities of Ura Design as part of OTF. Please add here on bullet points every time we have done tasks related to OTF. Deadline: 26.06.2018.
priority
otf report june documentation for monthly report of activities of ura design as part of otf please add here on bullet points every time we have done tasks related to otf deadline
1
501,248
14,524,178,703
IssuesEvent
2020-12-14 11:05:05
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
[0.9.2 staging-1872] Industrial elecator collider problem
Category: Gameplay Priority: Medium
This part doesn't have collider: ![image](https://user-images.githubusercontent.com/45708377/102073534-81083980-3e14-11eb-9d31-b6f367d861f1.png) Therefore, if the block under the elevator does not have a square edge, you will fall by car: ![image](https://user-images.githubusercontent.com/45708377/102073634-a5641600-3e14-11eb-91ef-d90154930c3d.png) ![image](https://user-images.githubusercontent.com/45708377/102073646-ad23ba80-3e14-11eb-8b5f-2a1c33cecd94.png) It wiill be harder but you can fall with square edgge too: ![image](https://user-images.githubusercontent.com/45708377/102073929-160b3280-3e15-11eb-9278-408839e3dceb.png) I think that at least visually we have a gap, but it should be removed: ![image](https://user-images.githubusercontent.com/45708377/102073968-20c5c780-3e15-11eb-9923-0e0d9a4d79ca.png)
1.0
[0.9.2 staging-1872] Industrial elecator collider problem - This part doesn't have collider: ![image](https://user-images.githubusercontent.com/45708377/102073534-81083980-3e14-11eb-9d31-b6f367d861f1.png) Therefore, if the block under the elevator does not have a square edge, you will fall by car: ![image](https://user-images.githubusercontent.com/45708377/102073634-a5641600-3e14-11eb-91ef-d90154930c3d.png) ![image](https://user-images.githubusercontent.com/45708377/102073646-ad23ba80-3e14-11eb-8b5f-2a1c33cecd94.png) It wiill be harder but you can fall with square edgge too: ![image](https://user-images.githubusercontent.com/45708377/102073929-160b3280-3e15-11eb-9278-408839e3dceb.png) I think that at least visually we have a gap, but it should be removed: ![image](https://user-images.githubusercontent.com/45708377/102073968-20c5c780-3e15-11eb-9923-0e0d9a4d79ca.png)
priority
industrial elecator collider problem this part doesn t have collider therefore if the block under the elevator does not have a square edge you will fall by car it wiill be harder but you can fall with square edgge too i think that at least visually we have a gap but it should be removed
1
54,999
3,071,798,917
IssuesEvent
2015-08-19 14:05:04
RobotiumTech/robotium
https://api.github.com/repos/RobotiumTech/robotium
closed
ClickonButton fails on buttons only with background images
bug imported Priority-Medium
_From [yair.oha...@gmail.com](https://code.google.com/u/103769070358968325894/) on February 09, 2011 13:58:24_ What steps will reproduce the problem? 1.open dialog with buttons which have no text and backround specifics. 2.execute solo.clickonbutton(index) 3. What is the expected output? What do you see instead? I expect to see a click - nothing happens. What version of the product are you using? On what operating system? version 2.1 on Win 7 ,64 bit Please provide any additional information below. **Attachment:** [servings_edit_dialog.xml](http://code.google.com/p/robotium/issues/detail?id=66) _Original issue: http://code.google.com/p/robotium/issues/detail?id=66_
1.0
ClickonButton fails on buttons only with background images - _From [yair.oha...@gmail.com](https://code.google.com/u/103769070358968325894/) on February 09, 2011 13:58:24_ What steps will reproduce the problem? 1.open dialog with buttons which have no text and backround specifics. 2.execute solo.clickonbutton(index) 3. What is the expected output? What do you see instead? I expect to see a click - nothing happens. What version of the product are you using? On what operating system? version 2.1 on Win 7 ,64 bit Please provide any additional information below. **Attachment:** [servings_edit_dialog.xml](http://code.google.com/p/robotium/issues/detail?id=66) _Original issue: http://code.google.com/p/robotium/issues/detail?id=66_
priority
clickonbutton fails on buttons only with background images from on february what steps will reproduce the problem open dialog with buttons which have no text and backround specifics execute solo clickonbutton index what is the expected output what do you see instead i expect to see a click nothing happens what version of the product are you using on what operating system version on win bit please provide any additional information below attachment original issue
1
684,886
23,436,680,018
IssuesEvent
2022-08-15 10:38:50
conan-io/conan
https://api.github.com/repos/conan-io/conan
closed
[feature] Access to license and homepage for dependencies in conanfile.py
type: feature priority: medium complex: low
For some automated license report we would need to have access to the attributes `description`, `license` and `homepage` which can be set in the conan recipe. As self.deps_cpp_info, self.deps_env_info or self.deps_user_info will be removed in conan 2.0, this should probably be added to [self.dependencies](https://docs.conan.io/en/latest/reference/conanfile/dependencies.html#dependencies-interface). > Do not use self.deps_cpp_info, self.deps_env_info or self.deps_user_info, these have been removed in 2.0. During this change it would be great to just have access to all attributes for the dependencies
1.0
[feature] Access to license and homepage for dependencies in conanfile.py - For some automated license report we would need to have access to the attributes `description`, `license` and `homepage` which can be set in the conan recipe. As self.deps_cpp_info, self.deps_env_info or self.deps_user_info will be removed in conan 2.0, this should probably be added to [self.dependencies](https://docs.conan.io/en/latest/reference/conanfile/dependencies.html#dependencies-interface). > Do not use self.deps_cpp_info, self.deps_env_info or self.deps_user_info, these have been removed in 2.0. During this change it would be great to just have access to all attributes for the dependencies
priority
access to license and homepage for dependencies in conanfile py for some automated license report we would need to have access to the attributes description license and homepage which can be set in the conan recipe as self deps cpp info self deps env info or self deps user info will be removed in conan this should probably be added to do not use self deps cpp info self deps env info or self deps user info these have been removed in during this change it would be great to just have access to all attributes for the dependencies
1
682,857
23,359,948,951
IssuesEvent
2022-08-10 10:46:25
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Bluetooth controller extended advertisement crashes in lll layer
bug priority: medium area: Bluetooth area: Bluetooth Controller
Zephyr version: NCS 2.0.0 - Zephyr tag v3.0.99-ncs1 I'm running and Bluetooth host and controller combination of nRF9160(Host) and nRF52833(Controller) using the BT_LL_SW_SPLIT variant. The application is very complicated so i cannot provide a reproducible sample. I have managed to capture an stack trace using Memfault. The problem occurs rather random once in a few hours. ![image](https://user-images.githubusercontent.com/54847502/183609500-a238f4fe-62c9-4546-8db5-da570c6e6f3a.png) [lll.c](https://github.com/nrfconnect/sdk-zephyr/blob/v3.0.99-ncs1/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c) Note that the trace is capture with disabled CONFIG_BT_ASSERT=n on the controller. We had issues with advertisement raising radio tx not ready. In production we disabled the assert to remove unnecessary restarts with cost of some advertisements not getting transmitted. I can try to catch some debug outputs using the BT_ASSERT, however the assert string is usually lost due to LOG_DEFFER. To catch the issue with memfault is also possibility with the BT_ASSERT but requires more work. Afterwards i tried to optimize IRQ times by these settings(replaced Zero Latency Interrupts) and haven't tested the assert variant ever since: ``` CONFIG_BT_CTLR_ULL_HIGH_PRIO=1 CONFIG_BT_CTLR_ULL_LOW_PRIO=2 ``` Basic description of the application. It broadcasts a large amount of advertisements sets 6 sets of advertisements using BT legacy(1 connectable). 6 sets of advertisements using BT long range (Coded phy). Interval of all advertisements are 4HZ. Application performs regular ADV_DATA updates for the advertisements with interval 1Hz-4Hz. Communication between Controller and Host is implemented by custom layer. I can provide HCI traces using RTT BT debug if necessary. However, we are using same layer to implement other communication between the chips and we had no problems with it. It is basically H5(Confirmed messages, and retransmission) over multiplexed UART with Flow Control. The controller and host does not seem to generate any error logs. Configuration of the host(BT part): ``` # Enable Bluetooth stack and libraries CONFIG_BT=y CONFIG_BT_PERIPHERAL=y CONFIG_BT_NO_DRIVER=y CONFIG_BT_H4=n CONFIG_BT_SMP=y CONFIG_BT_KEYS_OVERWRITE_OLDEST=y CONFIG_BT_HCI_VS=y CONFIG_BT_HCI_VS_EXT=y CONFIG_BT_WAIT_NOP=n CONFIG_BT_DEVICE_NAME="Mini" CONFIG_BT_DEVICE_APPEARANCE=666 CONFIG_BT_EXT_ADV=y CONFIG_BT_EXT_ADV_MAX_ADV_SET=15 CONFIG_BT_USER_PHY_UPDATE=y CONFIG_BT_MAX_CONN=4 # Allow for large Bluetooth data packets. CONFIG_BT_BUF_ACL_RX_SIZE=255 CONFIG_BT_BUF_CMD_TX_SIZE=255 CONFIG_BT_L2CAP_TX_MTU=247 CONFIG_BT_L2CAP_TX_BUF_COUNT=10 CONFIG_BT_L2CAP_DYNAMIC_CHANNEL=y CONFIG_BT_ATT_PREPARE_COUNT=2 ``` Configuration of the controller(BT part): ``` #BT Driver CONFIG_BT=y CONFIG_BT_CTLR=y CONFIG_BT_CTLR_TX_PWR_PLUS_8=y CONFIG_NET_BUF=y CONFIG_BT_HCI_RAW=y CONFIG_BT_CENTRAL=n CONFIG_BT_EXT_ADV=y CONFIG_BT_CTLR_ADV_EXT=y CONFIG_BT_CTLR_ADV_SET=15 CONFIG_BT_CTLR_ADV_DATA_LEN_MAX=100 CONFIG_BT_CTLR_LE_ENC=y CONFIG_BT_CTLR_ADVANCED_FEATURES=y CONFIG_BT_CTLR_OPTIMIZE_FOR_SPEED=y CONFIG_BT_CTLR_SCHED_ADVANCED=y CONFIG_BT_MAX_CONN=4 CONFIG_BT_CTLR_TX_PWR_DYNAMIC_CONTROL=y CONFIG_BT_CTLR_RX_BUFFERS=16 CONFIG_BT_CTLR_PHY_CODED=y CONFIG_BT_CTLR_ULL_HIGH_PRIO=1 CONFIG_BT_CTLR_ULL_LOW_PRIO=2 CONFIG_BT_CTLR_LOW_LAT=n CONFIG_BT_ASSERT=n CONFIG_BT_LL_SOFTDEVICE=n CONFIG_BT_LL_SW_SPLIT=y CONFIG_MPSL=n #BT Minimizing CONFIG_BT_CTLR_PHY_2M=n CONFIG_BT_CTLR_PRIVACY=n CONFIG_BT_CTLR_LE_PING=n CONFIG_BT_CTLR_LE_ENC=n CONFIG_BT_CTLR_CRYPTO=n ``` What other outputs might be helpful to identify the origin of the problem ?
1.0
Bluetooth controller extended advertisement crashes in lll layer - Zephyr version: NCS 2.0.0 - Zephyr tag v3.0.99-ncs1 I'm running and Bluetooth host and controller combination of nRF9160(Host) and nRF52833(Controller) using the BT_LL_SW_SPLIT variant. The application is very complicated so i cannot provide a reproducible sample. I have managed to capture an stack trace using Memfault. The problem occurs rather random once in a few hours. ![image](https://user-images.githubusercontent.com/54847502/183609500-a238f4fe-62c9-4546-8db5-da570c6e6f3a.png) [lll.c](https://github.com/nrfconnect/sdk-zephyr/blob/v3.0.99-ncs1/subsys/bluetooth/controller/ll_sw/nordic/lll/lll.c) Note that the trace is capture with disabled CONFIG_BT_ASSERT=n on the controller. We had issues with advertisement raising radio tx not ready. In production we disabled the assert to remove unnecessary restarts with cost of some advertisements not getting transmitted. I can try to catch some debug outputs using the BT_ASSERT, however the assert string is usually lost due to LOG_DEFFER. To catch the issue with memfault is also possibility with the BT_ASSERT but requires more work. Afterwards i tried to optimize IRQ times by these settings(replaced Zero Latency Interrupts) and haven't tested the assert variant ever since: ``` CONFIG_BT_CTLR_ULL_HIGH_PRIO=1 CONFIG_BT_CTLR_ULL_LOW_PRIO=2 ``` Basic description of the application. It broadcasts a large amount of advertisements sets 6 sets of advertisements using BT legacy(1 connectable). 6 sets of advertisements using BT long range (Coded phy). Interval of all advertisements are 4HZ. Application performs regular ADV_DATA updates for the advertisements with interval 1Hz-4Hz. Communication between Controller and Host is implemented by custom layer. I can provide HCI traces using RTT BT debug if necessary. However, we are using same layer to implement other communication between the chips and we had no problems with it. It is basically H5(Confirmed messages, and retransmission) over multiplexed UART with Flow Control. The controller and host does not seem to generate any error logs. Configuration of the host(BT part): ``` # Enable Bluetooth stack and libraries CONFIG_BT=y CONFIG_BT_PERIPHERAL=y CONFIG_BT_NO_DRIVER=y CONFIG_BT_H4=n CONFIG_BT_SMP=y CONFIG_BT_KEYS_OVERWRITE_OLDEST=y CONFIG_BT_HCI_VS=y CONFIG_BT_HCI_VS_EXT=y CONFIG_BT_WAIT_NOP=n CONFIG_BT_DEVICE_NAME="Mini" CONFIG_BT_DEVICE_APPEARANCE=666 CONFIG_BT_EXT_ADV=y CONFIG_BT_EXT_ADV_MAX_ADV_SET=15 CONFIG_BT_USER_PHY_UPDATE=y CONFIG_BT_MAX_CONN=4 # Allow for large Bluetooth data packets. CONFIG_BT_BUF_ACL_RX_SIZE=255 CONFIG_BT_BUF_CMD_TX_SIZE=255 CONFIG_BT_L2CAP_TX_MTU=247 CONFIG_BT_L2CAP_TX_BUF_COUNT=10 CONFIG_BT_L2CAP_DYNAMIC_CHANNEL=y CONFIG_BT_ATT_PREPARE_COUNT=2 ``` Configuration of the controller(BT part): ``` #BT Driver CONFIG_BT=y CONFIG_BT_CTLR=y CONFIG_BT_CTLR_TX_PWR_PLUS_8=y CONFIG_NET_BUF=y CONFIG_BT_HCI_RAW=y CONFIG_BT_CENTRAL=n CONFIG_BT_EXT_ADV=y CONFIG_BT_CTLR_ADV_EXT=y CONFIG_BT_CTLR_ADV_SET=15 CONFIG_BT_CTLR_ADV_DATA_LEN_MAX=100 CONFIG_BT_CTLR_LE_ENC=y CONFIG_BT_CTLR_ADVANCED_FEATURES=y CONFIG_BT_CTLR_OPTIMIZE_FOR_SPEED=y CONFIG_BT_CTLR_SCHED_ADVANCED=y CONFIG_BT_MAX_CONN=4 CONFIG_BT_CTLR_TX_PWR_DYNAMIC_CONTROL=y CONFIG_BT_CTLR_RX_BUFFERS=16 CONFIG_BT_CTLR_PHY_CODED=y CONFIG_BT_CTLR_ULL_HIGH_PRIO=1 CONFIG_BT_CTLR_ULL_LOW_PRIO=2 CONFIG_BT_CTLR_LOW_LAT=n CONFIG_BT_ASSERT=n CONFIG_BT_LL_SOFTDEVICE=n CONFIG_BT_LL_SW_SPLIT=y CONFIG_MPSL=n #BT Minimizing CONFIG_BT_CTLR_PHY_2M=n CONFIG_BT_CTLR_PRIVACY=n CONFIG_BT_CTLR_LE_PING=n CONFIG_BT_CTLR_LE_ENC=n CONFIG_BT_CTLR_CRYPTO=n ``` What other outputs might be helpful to identify the origin of the problem ?
priority
bluetooth controller extended advertisement crashes in lll layer zephyr version ncs zephyr tag i m running and bluetooth host and controller combination of host and controller using the bt ll sw split variant the application is very complicated so i cannot provide a reproducible sample i have managed to capture an stack trace using memfault the problem occurs rather random once in a few hours note that the trace is capture with disabled config bt assert n on the controller we had issues with advertisement raising radio tx not ready in production we disabled the assert to remove unnecessary restarts with cost of some advertisements not getting transmitted i can try to catch some debug outputs using the bt assert however the assert string is usually lost due to log deffer to catch the issue with memfault is also possibility with the bt assert but requires more work afterwards i tried to optimize irq times by these settings replaced zero latency interrupts and haven t tested the assert variant ever since config bt ctlr ull high prio config bt ctlr ull low prio basic description of the application it broadcasts a large amount of advertisements sets sets of advertisements using bt legacy connectable sets of advertisements using bt long range coded phy interval of all advertisements are application performs regular adv data updates for the advertisements with interval communication between controller and host is implemented by custom layer i can provide hci traces using rtt bt debug if necessary however we are using same layer to implement other communication between the chips and we had no problems with it it is basically confirmed messages and retransmission over multiplexed uart with flow control the controller and host does not seem to generate any error logs configuration of the host bt part enable bluetooth stack and libraries config bt y config bt peripheral y config bt no driver y config bt n config bt smp y config bt keys overwrite oldest y config bt hci vs y config bt hci vs ext y config bt wait nop n config bt device name mini config bt device appearance config bt ext adv y config bt ext adv max adv set config bt user phy update y config bt max conn allow for large bluetooth data packets config bt buf acl rx size config bt buf cmd tx size config bt tx mtu config bt tx buf count config bt dynamic channel y config bt att prepare count configuration of the controller bt part bt driver config bt y config bt ctlr y config bt ctlr tx pwr plus y config net buf y config bt hci raw y config bt central n config bt ext adv y config bt ctlr adv ext y config bt ctlr adv set config bt ctlr adv data len max config bt ctlr le enc y config bt ctlr advanced features y config bt ctlr optimize for speed y config bt ctlr sched advanced y config bt max conn config bt ctlr tx pwr dynamic control y config bt ctlr rx buffers config bt ctlr phy coded y config bt ctlr ull high prio config bt ctlr ull low prio config bt ctlr low lat n config bt assert n config bt ll softdevice n config bt ll sw split y config mpsl n bt minimizing config bt ctlr phy n config bt ctlr privacy n config bt ctlr le ping n config bt ctlr le enc n config bt ctlr crypto n what other outputs might be helpful to identify the origin of the problem
1
736,842
25,490,042,703
IssuesEvent
2022-11-26 23:52:53
Thenlie/Streamability
https://api.github.com/repos/Thenlie/Streamability
closed
📝Install and configure Tailwind CSS
type: feature priority: medium
Set up tailwind as suggested in the documentation. Make sure this is running painlessly for dev, it should auto update with the site when developing but remove unused styles when bundling for production. This should be set up initially with the understanding we will need to make some custom classes and possible using the `@apply` directive to add them to the DOM. References #63
1.0
📝Install and configure Tailwind CSS - Set up tailwind as suggested in the documentation. Make sure this is running painlessly for dev, it should auto update with the site when developing but remove unused styles when bundling for production. This should be set up initially with the understanding we will need to make some custom classes and possible using the `@apply` directive to add them to the DOM. References #63
priority
📝install and configure tailwind css set up tailwind as suggested in the documentation make sure this is running painlessly for dev it should auto update with the site when developing but remove unused styles when bundling for production this should be set up initially with the understanding we will need to make some custom classes and possible using the apply directive to add them to the dom references
1
289,026
8,854,282,774
IssuesEvent
2019-01-09 00:37:36
visit-dav/issues-test
https://api.github.com/repos/visit-dav/issues-test
closed
Gadget reader failure on re-open
bug crash likelihood medium priority reviewed severity high wrong results
Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1352 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: Gadget reader failure on re-open Assigned to: Kathleen Biagas Category: - Target version: 2.6.2 Author: Kathleen Biagas Start: 02/26/2013 Due date: % Done: 0% Estimated time: 2.00 hours Created: 02/26/2013 06:19 pm Updated: 02/27/2013 07:10 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.6.0 Impact: Expected Use: OS: All Support Group: Any Description: Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database. Comments: Made the reader close its file handles.M databases/Gadget/avtGadgetFileFormat.CM databases/Gadget/avtGadgetFileFormat.h
1.0
Gadget reader failure on re-open - Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1352 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: Gadget reader failure on re-open Assigned to: Kathleen Biagas Category: - Target version: 2.6.2 Author: Kathleen Biagas Start: 02/26/2013 Due date: % Done: 0% Estimated time: 2.00 hours Created: 02/26/2013 06:19 pm Updated: 02/27/2013 07:10 pm Likelihood: 3 - Occasional Severity: 4 - Crash / Wrong Results Found in version: 2.6.0 Impact: Expected Use: OS: All Support Group: Any Description: Allen Harvey reported strangeness with the Gadget reader.He has a lot of files (> 400) grouped in a virtual database.If he saves settings from a session where he has opened this database, then restarts visit and makes a plot, clicking 'Reopen' will yield in the information window: ===========VisIt could not read from the file "/home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database". The generated error message was: There was an error opening /home/biagas2/CustomerData/AllenHarvey/gadget/snap_* database. It may be an invalid file. VisIt tried using the following file format readers to open the file: Gadget The following error(s) may be helpful in identifying the problem:The selected database type was not correct for the given file. The message from the database was Cant open file ============================ I added code to print out the errno and message from the failed fopen attempt in the Gadget reader (which is in the constructor, by the way) , and got this: Error: 24: Too many open files Strangely enough, if you delete the config files and retry, VisIt does not complain when reopening the database. Comments: Made the reader close its file handles.M databases/Gadget/avtGadgetFileFormat.CM databases/Gadget/avtGadgetFileFormat.h
priority
gadget reader failure on re open allen harvey reported strangeness with the gadget reader he has a lot of files grouped in a virtual database if he saves settings from a session where he has opened this database then restarts visit and makes a plot clicking reopen will yield in the information window visit could not read from the file home customerdata allenharvey gadget snap database the generated error message was there was an error opening home customerdata allenharvey gadget snap database it may be an invalid file visit tried using the following file format readers to open the file gadget the following error s may be helpful in identifying the problem the selected database type was not correct for the given file the message from the database was cant open file i added code to print out the errno and message from the failed fopen attempt in the gadget reader which is in the constructor by the way and got this error too many open files strangely enough if you delete the config files and retry visit does not complain when reopening the database redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject gadget reader failure on re open assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time hours created pm updated pm likelihood occasional severity crash wrong results found in version impact expected use os all support group any description allen harvey reported strangeness with the gadget reader he has a lot of files grouped in a virtual database if he saves settings from a session where he has opened this database then restarts visit and makes a plot clicking reopen will yield in the information window visit could not read from the file home customerdata allenharvey gadget snap database the generated error message was there was an error opening home customerdata allenharvey gadget snap database it may be an invalid file visit tried using the following file format readers to open the file gadget the following error s may be helpful in identifying the problem the selected database type was not correct for the given file the message from the database was cant open file i added code to print out the errno and message from the failed fopen attempt in the gadget reader which is in the constructor by the way and got this error too many open files strangely enough if you delete the config files and retry visit does not complain when reopening the database comments made the reader close its file handles m databases gadget avtgadgetfileformat cm databases gadget avtgadgetfileformat h
1
150,833
5,791,823,112
IssuesEvent
2017-05-02 07:34:24
Caleydo/taggle
https://api.github.com/repos/Caleydo/taggle
opened
Apply column stratification of first categorical column correctly
bug medium priority
* Release number or git hash: * Web browser version and OS: * Environment (local or deployed): ### Steps to reproduce 1. Add a matrix (e.g. Number1Hits) 2. Add a categorical column for this matrix (e.g. continent) 3. Split matrix by min/mean/max (type doesn't matter) ### Observed behavior * The matrix is not splitted into 3 vectors as expected * The user first has to click the stratify button of the categorical column and then split ![splitting_matrix](https://cloud.githubusercontent.com/assets/5851088/25608238/6ed5f052-2f1a-11e7-8a20-ba762b1ae272.gif) ### Expected behavior * The stratification should be applied automatically when adding the first categorical column to a matrix
1.0
Apply column stratification of first categorical column correctly - * Release number or git hash: * Web browser version and OS: * Environment (local or deployed): ### Steps to reproduce 1. Add a matrix (e.g. Number1Hits) 2. Add a categorical column for this matrix (e.g. continent) 3. Split matrix by min/mean/max (type doesn't matter) ### Observed behavior * The matrix is not splitted into 3 vectors as expected * The user first has to click the stratify button of the categorical column and then split ![splitting_matrix](https://cloud.githubusercontent.com/assets/5851088/25608238/6ed5f052-2f1a-11e7-8a20-ba762b1ae272.gif) ### Expected behavior * The stratification should be applied automatically when adding the first categorical column to a matrix
priority
apply column stratification of first categorical column correctly release number or git hash web browser version and os environment local or deployed steps to reproduce add a matrix e g add a categorical column for this matrix e g continent split matrix by min mean max type doesn t matter observed behavior the matrix is not splitted into vectors as expected the user first has to click the stratify button of the categorical column and then split expected behavior the stratification should be applied automatically when adding the first categorical column to a matrix
1
320,980
9,791,862,032
IssuesEvent
2019-06-10 15:58:29
OperationCode/front-end
https://api.github.com/repos/OperationCode/front-end
closed
Prop-type error in Select + Input component
Priority: Medium Type: Developer Experience / Refactoring help wanted
# Bug Report ## What is the current behavior? I've duct-taped the behavior in https://github.com/OperationCode/front-end/pull/506 , but if you undo that PR and run forms with a Select where `isMulti` is true, you'll see a prop-type error where `form.touched.$fieldName` is an array and it should be a boolean. **Error Stack** ``` Error: Uncaught [Error: Warning: Failed prop type: Invalid prop `form.touched.programmingLanguages` of type `array` supplied to `Select`, expected `boolean`. in Select (created by FieldInner) in FieldInner (created by Context.Consumer) in FormikConnect(FieldInner) (at Technology.js:71) in div (at Technology.js:70) in Technology (at Technology.test.js:39) in form (at Form.js:5) in Unknown (created by Context.Consumer) in FormikConnectedForm (at Technology.test.js:38) in Formik (created by WrapperComponent) in WrapperComponent] ``` **Implementation** ```javascript <Field className={styles.fullWidth} name="programmingLanguages" label="Programming Languages That Interest You" component={Select} isMulti options={[{ label: 'JavaScript', value: 'javascript' }, { label: 'Python', value: 'python' }, /* etc */]} disabled={isSubmitting} placeholder="Select at least one..." /> ``` ## What is the expected behavior? Essentially `form.touched` should be an object where ever form field name is represented in the keys and the values are a boolean on whether or not they've been "touched". ## What steps did you take to get this behavior? Undo changes in aforementioned duct tape PR, play around inside profile update forms, and see run-time prop-type error. ## Additional Context I started doubting the source code of Formik, and was able to get what I believe is a troubling result (copied code from `formik/src/utils.ts` into console on lodash's website): ![image](https://user-images.githubusercontent.com/9523719/57975160-eebe5c00-7978-11e9-9227-70097ce5e690.png) ![image](https://user-images.githubusercontent.com/9523719/57975168-10b7de80-7979-11e9-84c0-cc5e2cc23506.png) [Here's a Code Sandbox](https://codesandbox.io/embed/710qrr0qjx?fontsize=14) with a `react-select` component inside a Formik context where `isMulti` is true and `touched` is always a boolean.
1.0
Prop-type error in Select + Input component - # Bug Report ## What is the current behavior? I've duct-taped the behavior in https://github.com/OperationCode/front-end/pull/506 , but if you undo that PR and run forms with a Select where `isMulti` is true, you'll see a prop-type error where `form.touched.$fieldName` is an array and it should be a boolean. **Error Stack** ``` Error: Uncaught [Error: Warning: Failed prop type: Invalid prop `form.touched.programmingLanguages` of type `array` supplied to `Select`, expected `boolean`. in Select (created by FieldInner) in FieldInner (created by Context.Consumer) in FormikConnect(FieldInner) (at Technology.js:71) in div (at Technology.js:70) in Technology (at Technology.test.js:39) in form (at Form.js:5) in Unknown (created by Context.Consumer) in FormikConnectedForm (at Technology.test.js:38) in Formik (created by WrapperComponent) in WrapperComponent] ``` **Implementation** ```javascript <Field className={styles.fullWidth} name="programmingLanguages" label="Programming Languages That Interest You" component={Select} isMulti options={[{ label: 'JavaScript', value: 'javascript' }, { label: 'Python', value: 'python' }, /* etc */]} disabled={isSubmitting} placeholder="Select at least one..." /> ``` ## What is the expected behavior? Essentially `form.touched` should be an object where ever form field name is represented in the keys and the values are a boolean on whether or not they've been "touched". ## What steps did you take to get this behavior? Undo changes in aforementioned duct tape PR, play around inside profile update forms, and see run-time prop-type error. ## Additional Context I started doubting the source code of Formik, and was able to get what I believe is a troubling result (copied code from `formik/src/utils.ts` into console on lodash's website): ![image](https://user-images.githubusercontent.com/9523719/57975160-eebe5c00-7978-11e9-9227-70097ce5e690.png) ![image](https://user-images.githubusercontent.com/9523719/57975168-10b7de80-7979-11e9-84c0-cc5e2cc23506.png) [Here's a Code Sandbox](https://codesandbox.io/embed/710qrr0qjx?fontsize=14) with a `react-select` component inside a Formik context where `isMulti` is true and `touched` is always a boolean.
priority
prop type error in select input component bug report what is the current behavior i ve duct taped the behavior in but if you undo that pr and run forms with a select where ismulti is true you ll see a prop type error where form touched fieldname is an array and it should be a boolean error stack error uncaught error warning failed prop type invalid prop form touched programminglanguages of type array supplied to select expected boolean in select created by fieldinner in fieldinner created by context consumer in formikconnect fieldinner at technology js in div at technology js in technology at technology test js in form at form js in unknown created by context consumer in formikconnectedform at technology test js in formik created by wrappercomponent in wrappercomponent implementation javascript field classname styles fullwidth name programminglanguages label programming languages that interest you component select ismulti options disabled issubmitting placeholder select at least one what is the expected behavior essentially form touched should be an object where ever form field name is represented in the keys and the values are a boolean on whether or not they ve been touched what steps did you take to get this behavior undo changes in aforementioned duct tape pr play around inside profile update forms and see run time prop type error additional context i started doubting the source code of formik and was able to get what i believe is a troubling result copied code from formik src utils ts into console on lodash s website with a react select component inside a formik context where ismulti is true and touched is always a boolean
1
730,600
25,181,228,053
IssuesEvent
2022-11-11 13:52:05
eclipse/lsp4jakarta
https://api.github.com/repos/eclipse/lsp4jakarta
closed
Update snippets with more relevant names, prefixes and descriptions for Jakarta EE 9+
1 medium priority
The Jakarta RESTful Web Services snippets still refer to "JAX-RS" in the description: https://github.com/eclipse/lsp4jakarta/blob/7fb1fe4fdfac031819f426240b6227b06ab712d1/jakarta.ls/src/main/resources/jax-rs.json#L2-L25 Update to use more relevant names, prefixes and descriptions for Jakarta EE 9+ Also consider if any of the other snippets should be updated in the same way: https://github.com/eclipse/lsp4jakarta/tree/main/jakarta.ls/src/main/resources
1.0
Update snippets with more relevant names, prefixes and descriptions for Jakarta EE 9+ - The Jakarta RESTful Web Services snippets still refer to "JAX-RS" in the description: https://github.com/eclipse/lsp4jakarta/blob/7fb1fe4fdfac031819f426240b6227b06ab712d1/jakarta.ls/src/main/resources/jax-rs.json#L2-L25 Update to use more relevant names, prefixes and descriptions for Jakarta EE 9+ Also consider if any of the other snippets should be updated in the same way: https://github.com/eclipse/lsp4jakarta/tree/main/jakarta.ls/src/main/resources
priority
update snippets with more relevant names prefixes and descriptions for jakarta ee the jakarta restful web services snippets still refer to jax rs in the description update to use more relevant names prefixes and descriptions for jakarta ee also consider if any of the other snippets should be updated in the same way
1
235,138
7,734,949,419
IssuesEvent
2018-05-27 09:17:30
Radarr/Radarr
https://api.github.com/repos/Radarr/Radarr
closed
Add AHD_Internal Flag
confirmed enhancement indexer priority:medium
Would it be possible to add an AHD indexer? There is a search api available at `https://awesome-hd.me/searchapi.php?action=imdbsearch&passkey=%s&imdb=%s&internal=%s` In this passkey is the users passkey, imdb is the IMDb ID int he form tt1234567 and internal is a boolean that will set the results to only include AHD internals if true or all releases if false or not included. This will return the following for a search for Bride of Chucky (IMDb ID tt0144120) ``` <?xml version="1.0"?> <searchresults> <authkey>AUTHKEY REMOVED</authkey> <torrent> <id>18809</id> <groupid>9687</groupid> <time>2014-03-18 22:29:33</time> <userid>33249</userid> <size>22661245411</size> <snatched>35</snatched> <seeders>1</seeders> <leechers>0</leechers> <releasegroup>HiFi</releasegroup> <resolution>1080p</resolution> <media>Blu-ray</media> <format>MKV</format> <encoding>VC-1 Remux</encoding> <audioformat>DTS-HD MA</audioformat> <audiobitrate>Variable</audiobitrate> <audiochannels>6</audiochannels> <subtitles>English, French, Spanish</subtitles> <encodestatus>AHDGolddxva</encodestatus> <freeleech>0.50</freeleech> <cover> http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </cover> <smallcover> http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </smallcover> <year>1998</year> <name>Bride of Chucky</name> <imdb>tt0144120</imdb> <type>Movie</type> <plotoutline> Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll. </plotoutline> </torrent> <torrent> <id>17500</id> <groupid>9687</groupid> <time>2014-02-02 06:07:05</time> <userid>37948</userid> <size>22661245137</size> <snatched>6</snatched> <seeders>0</seeders> <leechers>0</leechers> <releasegroup>User-Upload</releasegroup> <resolution>1080p</resolution> <media>Blu-ray</media> <format>MKV</format> <encoding>VC-1 Remux</encoding> <audioformat>DTS-HD MA</audioformat> <audiobitrate>Variable</audiobitrate> <audiochannels>6</audiochannels> <subtitles>English, Spanish, French</subtitles> <encodestatus>AHDGolddxva</encodestatus> <freeleech>0.50</freeleech> <cover> http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </cover> <smallcover> http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </smallcover> <year>1998</year> <name>Bride of Chucky</name> <imdb>tt0144120</imdb> <type>Movie</type> <plotoutline> Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll. </plotoutline> </torrent> </searchresults> ``` The download link for the above then would be `https://awesome-hd.me/torrents.php?action=download&id=%s&authkey=%s&torrent_pass=%s'` the authkey and id are returned in the search results above and the torrent_pass is the passkey as used earlier.
1.0
Add AHD_Internal Flag - Would it be possible to add an AHD indexer? There is a search api available at `https://awesome-hd.me/searchapi.php?action=imdbsearch&passkey=%s&imdb=%s&internal=%s` In this passkey is the users passkey, imdb is the IMDb ID int he form tt1234567 and internal is a boolean that will set the results to only include AHD internals if true or all releases if false or not included. This will return the following for a search for Bride of Chucky (IMDb ID tt0144120) ``` <?xml version="1.0"?> <searchresults> <authkey>AUTHKEY REMOVED</authkey> <torrent> <id>18809</id> <groupid>9687</groupid> <time>2014-03-18 22:29:33</time> <userid>33249</userid> <size>22661245411</size> <snatched>35</snatched> <seeders>1</seeders> <leechers>0</leechers> <releasegroup>HiFi</releasegroup> <resolution>1080p</resolution> <media>Blu-ray</media> <format>MKV</format> <encoding>VC-1 Remux</encoding> <audioformat>DTS-HD MA</audioformat> <audiobitrate>Variable</audiobitrate> <audiochannels>6</audiochannels> <subtitles>English, French, Spanish</subtitles> <encodestatus>AHDGolddxva</encodestatus> <freeleech>0.50</freeleech> <cover> http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </cover> <smallcover> http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </smallcover> <year>1998</year> <name>Bride of Chucky</name> <imdb>tt0144120</imdb> <type>Movie</type> <plotoutline> Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll. </plotoutline> </torrent> <torrent> <id>17500</id> <groupid>9687</groupid> <time>2014-02-02 06:07:05</time> <userid>37948</userid> <size>22661245137</size> <snatched>6</snatched> <seeders>0</seeders> <leechers>0</leechers> <releasegroup>User-Upload</releasegroup> <resolution>1080p</resolution> <media>Blu-ray</media> <format>MKV</format> <encoding>VC-1 Remux</encoding> <audioformat>DTS-HD MA</audioformat> <audiobitrate>Variable</audiobitrate> <audiochannels>6</audiochannels> <subtitles>English, Spanish, French</subtitles> <encodestatus>AHDGolddxva</encodestatus> <freeleech>0.50</freeleech> <cover> http://image.tmdb.org/t/p/w342/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </cover> <smallcover> http://image.tmdb.org/t/p/w154/mAAWRvrTQJQa5EFevZ67MgCVylI.jpg </smallcover> <year>1998</year> <name>Bride of Chucky</name> <imdb>tt0144120</imdb> <type>Movie</type> <plotoutline> Chucky, the doll possessed by a serial killer, discovers the perfect mate to kill and revive into the body of another doll. </plotoutline> </torrent> </searchresults> ``` The download link for the above then would be `https://awesome-hd.me/torrents.php?action=download&id=%s&authkey=%s&torrent_pass=%s'` the authkey and id are returned in the search results above and the torrent_pass is the passkey as used earlier.
priority
add ahd internal flag would it be possible to add an ahd indexer there is a search api available at in this passkey is the users passkey imdb is the imdb id int he form and internal is a boolean that will set the results to only include ahd internals if true or all releases if false or not included this will return the following for a search for bride of chucky imdb id authkey removed hifi blu ray mkv vc remux dts hd ma variable english french spanish ahdgolddxva bride of chucky movie chucky the doll possessed by a serial killer discovers the perfect mate to kill and revive into the body of another doll user upload blu ray mkv vc remux dts hd ma variable english spanish french ahdgolddxva bride of chucky movie chucky the doll possessed by a serial killer discovers the perfect mate to kill and revive into the body of another doll the download link for the above then would be the authkey and id are returned in the search results above and the torrent pass is the passkey as used earlier
1
26,463
2,684,555,866
IssuesEvent
2015-03-29 03:30:53
gtcasl/gpuocelot
https://api.github.com/repos/gtcasl/gpuocelot
closed
Ocelot does not build from source
bug imported Priority-Medium
_From [lokendra...@gmail.com](https://code.google.com/u/102201010740957725244/) on January 10, 2012 01:41:00_ What steps will reproduce the problem? 1. ./build.py --install -p /home/zenome/local --no_llvm What is the expected output? What do you see instead? Expected: compilation successful Output: Installing ocelot... config_h_build([".release_build/configure.h"], ["configure.h.in"]) g++ -o .release_build/ocelot/tools/OcelotConfig.o -c -O2 -Wall -Werror -std=c++0x -DOCELOT_CXXFLAGS="\"-O2 -Wall -Werror -std=c++0x\"" -DPACKAGE="\"ocelot\"" -DVERSION="\"2.1.1624\"" -DOCELOT_PREFIX_PATH="\"/home/zenome/local\"" -DOCELOT_LDFLAGS="\"-locelot\"" -L"\"/home/zenome/local/lib\"" -DOCELOT_INCLUDE_PATH="\"/home/zenome/local/include\"" -DOCELOT_LIB_PATH="\"/home/zenome/local/lib\"" -DOCELOT_BIN_PATH="\"/home/zenome/local/bin\"" -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/OcelotConfig.cpp g++ -o .release_build/ocelot/tools/CFG.o -c -O2 -Wall -Werror -std=c++0x -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/CFG.cpp ocelot/tools/CFG.cpp: In function 'void analyze(const char*)': ocelot/tools/CFG.cpp:56:11: error: 'class ir::PTXKernel' has no member named 'dom_tree' ocelot/tools/CFG.cpp:61:11: error: 'class ir::PTXKernel' has no member named 'pdom_tree' ocelot/tools/CFG.cpp:66:11: error: 'class ir::PTXKernel' has no member named 'ctrl_tree' What version of the product are you using? On what operating system? Ocelot: 1649 Ubuntu-11.04 GCC/++ : 4.5.2 Boost: 1.46.0 Please provide any additional information below. _Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=61_
1.0
Ocelot does not build from source - _From [lokendra...@gmail.com](https://code.google.com/u/102201010740957725244/) on January 10, 2012 01:41:00_ What steps will reproduce the problem? 1. ./build.py --install -p /home/zenome/local --no_llvm What is the expected output? What do you see instead? Expected: compilation successful Output: Installing ocelot... config_h_build([".release_build/configure.h"], ["configure.h.in"]) g++ -o .release_build/ocelot/tools/OcelotConfig.o -c -O2 -Wall -Werror -std=c++0x -DOCELOT_CXXFLAGS="\"-O2 -Wall -Werror -std=c++0x\"" -DPACKAGE="\"ocelot\"" -DVERSION="\"2.1.1624\"" -DOCELOT_PREFIX_PATH="\"/home/zenome/local\"" -DOCELOT_LDFLAGS="\"-locelot\"" -L"\"/home/zenome/local/lib\"" -DOCELOT_INCLUDE_PATH="\"/home/zenome/local/include\"" -DOCELOT_LIB_PATH="\"/home/zenome/local/lib\"" -DOCELOT_BIN_PATH="\"/home/zenome/local/bin\"" -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/OcelotConfig.cpp g++ -o .release_build/ocelot/tools/CFG.o -c -O2 -Wall -Werror -std=c++0x -I. -I.release_build -I/home/zenome/local/share/boost_1_46_0_install/include -I/usr/include -I/usr/include ocelot/tools/CFG.cpp ocelot/tools/CFG.cpp: In function 'void analyze(const char*)': ocelot/tools/CFG.cpp:56:11: error: 'class ir::PTXKernel' has no member named 'dom_tree' ocelot/tools/CFG.cpp:61:11: error: 'class ir::PTXKernel' has no member named 'pdom_tree' ocelot/tools/CFG.cpp:66:11: error: 'class ir::PTXKernel' has no member named 'ctrl_tree' What version of the product are you using? On what operating system? Ocelot: 1649 Ubuntu-11.04 GCC/++ : 4.5.2 Boost: 1.46.0 Please provide any additional information below. _Original issue: http://code.google.com/p/gpuocelot/issues/detail?id=61_
priority
ocelot does not build from source from on january what steps will reproduce the problem build py install p home zenome local no llvm what is the expected output what do you see instead expected compilation successful output installing ocelot config h build g o release build ocelot tools ocelotconfig o c wall werror std c docelot cxxflags wall werror std c dpackage ocelot dversion docelot prefix path home zenome local docelot ldflags locelot l home zenome local lib docelot include path home zenome local include docelot lib path home zenome local lib docelot bin path home zenome local bin i i release build i home zenome local share boost install include i usr include i usr include ocelot tools ocelotconfig cpp g o release build ocelot tools cfg o c wall werror std c i i release build i home zenome local share boost install include i usr include i usr include ocelot tools cfg cpp ocelot tools cfg cpp in function void analyze const char ocelot tools cfg cpp error class ir ptxkernel has no member named dom tree ocelot tools cfg cpp error class ir ptxkernel has no member named pdom tree ocelot tools cfg cpp error class ir ptxkernel has no member named ctrl tree what version of the product are you using on what operating system ocelot ubuntu gcc boost please provide any additional information below original issue
1
22,293
2,648,595,179
IssuesEvent
2015-03-14 02:15:07
dhamp/eiskaltdcpp
https://api.github.com/repos/dhamp/eiskaltdcpp
opened
feature to log bot private messages
imported Priority-Medium Type-FeatureRequest
_From [hg2rbw](https://code.google.com/u/hg2rbw/) on May 07, 2011 20:42:37_ feature to log bot private messages like opchat _Original issue: http://code.google.com/p/eiskaltdc/issues/detail?id=1067_
1.0
feature to log bot private messages - _From [hg2rbw](https://code.google.com/u/hg2rbw/) on May 07, 2011 20:42:37_ feature to log bot private messages like opchat _Original issue: http://code.google.com/p/eiskaltdc/issues/detail?id=1067_
priority
feature to log bot private messages from on may feature to log bot private messages like opchat original issue
1
449,328
12,967,346,953
IssuesEvent
2020-07-21 02:53:46
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Improve Diff Output When Specifying Show Changes in Job Template
component:ui priority:medium state:needs_devel type:enhancement
##### ISSUE TYPE - Feature Idea ##### SUMMARY <!-- Briefly describe the problem or desired enhancement. --> Per an user's request, they would like to see an improved output when specifying Show Changes in a job template. For example, if a job has thousands of lines of output, they would like to be able to filter for just the diff changes (i.e. For reporting purposes).
1.0
Improve Diff Output When Specifying Show Changes in Job Template - ##### ISSUE TYPE - Feature Idea ##### SUMMARY <!-- Briefly describe the problem or desired enhancement. --> Per an user's request, they would like to see an improved output when specifying Show Changes in a job template. For example, if a job has thousands of lines of output, they would like to be able to filter for just the diff changes (i.e. For reporting purposes).
priority
improve diff output when specifying show changes in job template issue type feature idea summary per an user s request they would like to see an improved output when specifying show changes in a job template for example if a job has thousands of lines of output they would like to be able to filter for just the diff changes i e for reporting purposes
1
229,995
7,603,084,074
IssuesEvent
2018-04-29 10:33:52
resin-io/resin-supervisor
https://api.github.com/repos/resin-io/resin-supervisor
closed
Trailing slash in WORKDIR causes a service restart loop
Low-hanging fruit Medium Priority flow/in-progress type/bug
If the image for a service has `Config.WorkingDir = "/usr/src/app/"`, the resulting container will have `Config.WorkingDir = "/usr/src/app"` - so the isSameContainer check in src/compose/service.coffee will always be false. This results in an endless loop of the supervisor killing and restarting the affected service. Workaround is to define WORKDIR without trailing slashes (i.e. `/usr/src/app` instead of `/usr/src/app/`). The fix would be to remove trailing slashes from `@workingDir` in the Service constructor. <img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_bbyz)
1.0
Trailing slash in WORKDIR causes a service restart loop - If the image for a service has `Config.WorkingDir = "/usr/src/app/"`, the resulting container will have `Config.WorkingDir = "/usr/src/app"` - so the isSameContainer check in src/compose/service.coffee will always be false. This results in an endless loop of the supervisor killing and restarting the affected service. Workaround is to define WORKDIR without trailing slashes (i.e. `/usr/src/app` instead of `/usr/src/app/`). The fix would be to remove trailing slashes from `@workingDir` in the Service constructor. <img src="https://frontapp.com/assets/img/icons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_bbyz)
priority
trailing slash in workdir causes a service restart loop if the image for a service has config workingdir usr src app the resulting container will have config workingdir usr src app so the issamecontainer check in src compose service coffee will always be false this results in an endless loop of the supervisor killing and restarting the affected service workaround is to define workdir without trailing slashes i e usr src app instead of usr src app the fix would be to remove trailing slashes from workingdir in the service constructor
1
689,195
23,611,361,487
IssuesEvent
2022-08-24 12:41:41
Knowledge-Management-Capstone/knowledge-management-dashboard
https://api.github.com/repos/Knowledge-Management-Capstone/knowledge-management-dashboard
opened
KMND-133 Revamp `DisscussionPage` UI
type:enhancement estimated-sp:2 priority:medium
## Description The problem with current UI is hard to display when the message is sent. I propose to use Slack/Discord like UI to overcome this. Please refer to [this repository](https://github.com/stuyy/chat-platform-react)
1.0
KMND-133 Revamp `DisscussionPage` UI - ## Description The problem with current UI is hard to display when the message is sent. I propose to use Slack/Discord like UI to overcome this. Please refer to [this repository](https://github.com/stuyy/chat-platform-react)
priority
kmnd revamp disscussionpage ui description the problem with current ui is hard to display when the message is sent i propose to use slack discord like ui to overcome this please refer to
1
129,428
5,096,915,235
IssuesEvent
2017-01-03 19:45:26
dteviot/WebToEpub
https://api.github.com/repos/dteviot/WebToEpub
reopened
[imgur] grab images from imgur
bug medium priority
I'm pretty sure there was talk about this but I can't seem to find it any more so I made a separate issue for sake of visibility. An example would be "https://crimsonmagic.me/2016/11/16/gifting-10-1/" [FIXED LINK], where there are various image -links- like "http://imgur.com/K4CZyyP.jpg" that don't get grabbed (or shown). There's also the more complex case of linked imgur galleries like on "http://skythewoodtl.com/fanfic-gifting/" [ADDED LINK], where you have a link to an imgur gallery with all the illustration before the novel ("Images" at the end).
1.0
[imgur] grab images from imgur - I'm pretty sure there was talk about this but I can't seem to find it any more so I made a separate issue for sake of visibility. An example would be "https://crimsonmagic.me/2016/11/16/gifting-10-1/" [FIXED LINK], where there are various image -links- like "http://imgur.com/K4CZyyP.jpg" that don't get grabbed (or shown). There's also the more complex case of linked imgur galleries like on "http://skythewoodtl.com/fanfic-gifting/" [ADDED LINK], where you have a link to an imgur gallery with all the illustration before the novel ("Images" at the end).
priority
grab images from imgur i m pretty sure there was talk about this but i can t seem to find it any more so i made a separate issue for sake of visibility an example would be where there are various image links like that don t get grabbed or shown there s also the more complex case of linked imgur galleries like on where you have a link to an imgur gallery with all the illustration before the novel images at the end
1
249,429
7,961,728,018
IssuesEvent
2018-07-13 11:55:30
magicDGS/ReadTools
https://api.github.com/repos/magicDGS/ReadTools
closed
Check usages of GATKRead defensive copy methods and substitute for performance improvement
Priority: Medium Status: Accepted Type: Enhancement
For example, `CheckQualityReadTransformer` makes always a defensive copy for the qualities, when it is not required because there is only an assertion and not change of qualities. This might reduce performance (concretely in the example case, it is performed quite often).
1.0
Check usages of GATKRead defensive copy methods and substitute for performance improvement - For example, `CheckQualityReadTransformer` makes always a defensive copy for the qualities, when it is not required because there is only an assertion and not change of qualities. This might reduce performance (concretely in the example case, it is performed quite often).
priority
check usages of gatkread defensive copy methods and substitute for performance improvement for example checkqualityreadtransformer makes always a defensive copy for the qualities when it is not required because there is only an assertion and not change of qualities this might reduce performance concretely in the example case it is performed quite often
1
167,287
6,335,898,146
IssuesEvent
2017-07-26 19:53:06
ngageoint/hootenanny
https://api.github.com/repos/ngageoint/hootenanny
opened
Error conflating single Rafah tile with HG workflow
Category: Core Priority: Medium Status: Defined Type: Bug
`scripts/core/ServiceOsmApiDbHootApiDbConflate.sh tmp/rafah.osm tmp/Rafah_UTDS_5_FCODE_noemptyFC_v61.osm "" "unifying" ServiceOsmApiDbHootApiDbRafahConflateTest2 true 1 Error running apply-changeset: Error executing query: ERROR: duplicate key value violates unique constraint "changesets_pkey" DETAIL: Key (id)=(6) already exists. (23505) QPSQL: Unable to create query (INSERT INTO changesets (id, user_id, created_at, closed_at) VALUES (6, 1, (now() at time zone 'utc'), (now() at time zone 'utc'));)` This actually happens for any random tile selected.
1.0
Error conflating single Rafah tile with HG workflow - `scripts/core/ServiceOsmApiDbHootApiDbConflate.sh tmp/rafah.osm tmp/Rafah_UTDS_5_FCODE_noemptyFC_v61.osm "" "unifying" ServiceOsmApiDbHootApiDbRafahConflateTest2 true 1 Error running apply-changeset: Error executing query: ERROR: duplicate key value violates unique constraint "changesets_pkey" DETAIL: Key (id)=(6) already exists. (23505) QPSQL: Unable to create query (INSERT INTO changesets (id, user_id, created_at, closed_at) VALUES (6, 1, (now() at time zone 'utc'), (now() at time zone 'utc'));)` This actually happens for any random tile selected.
priority
error conflating single rafah tile with hg workflow scripts core serviceosmapidbhootapidbconflate sh tmp rafah osm tmp rafah utds fcode noemptyfc osm unifying true error running apply changeset error executing query error duplicate key value violates unique constraint changesets pkey detail key id already exists qpsql unable to create query insert into changesets id user id created at closed at values now at time zone utc now at time zone utc this actually happens for any random tile selected
1
439,219
12,679,487,608
IssuesEvent
2020-06-19 11:54:20
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
closed
LT.61006
:beetle: bug :beetle: :grey_exclamation: priority medium
**Mod Version** f63a4a0d **What expansions do you have installed?** All **Please explain your issue in as much detail as possible:** LT.61006 can make no sense for a lot of characters **Steps to reproduce the issue:** Fire LT.61006 **Upload an attachment below: .zip of your save, or screenshots:** <details> <summary>Click to expand</summary> ![image](https://user-images.githubusercontent.com/46576860/67620770-fe3cad80-f812-11e9-9511-b00bd77eef61.png) </details>
1.0
LT.61006 - **Mod Version** f63a4a0d **What expansions do you have installed?** All **Please explain your issue in as much detail as possible:** LT.61006 can make no sense for a lot of characters **Steps to reproduce the issue:** Fire LT.61006 **Upload an attachment below: .zip of your save, or screenshots:** <details> <summary>Click to expand</summary> ![image](https://user-images.githubusercontent.com/46576860/67620770-fe3cad80-f812-11e9-9511-b00bd77eef61.png) </details>
priority
lt mod version what expansions do you have installed all please explain your issue in as much detail as possible lt can make no sense for a lot of characters steps to reproduce the issue fire lt upload an attachment below zip of your save or screenshots click to expand
1
509,438
14,731,027,317
IssuesEvent
2021-01-06 14:07:28
profclems/glab
https://api.github.com/repos/profclems/glab
reopened
Friendlier recovery when issue/mr create fails
help wanted kind: enhancement priority: medium
**Describe the feature or problem you'd like to solve** When `glab mr create` fails or cancelled, the body content typed using the editor is lost. Sometimes the user may type a long MR description and just lose it due to a simple mistake, such as not adding any commit to the MR branch yet (e.g., when creating draft MRs). **Propose a Solution** To me, the best user experience would be if the person can just resume where they left off so whenever someone authors an MR or issue, we save their title & body in a file scoped to that repository. If the create operation succeeded, we delete the file. Otherwise, the save file stays on disk and is used to pre-fill the next `issue/mr` create invocation, which is most likely a retry after a server or validation failure. Or instead of prefilling, we provide `--recover` flag to continue from the last session **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
1.0
Friendlier recovery when issue/mr create fails - **Describe the feature or problem you'd like to solve** When `glab mr create` fails or cancelled, the body content typed using the editor is lost. Sometimes the user may type a long MR description and just lose it due to a simple mistake, such as not adding any commit to the MR branch yet (e.g., when creating draft MRs). **Propose a Solution** To me, the best user experience would be if the person can just resume where they left off so whenever someone authors an MR or issue, we save their title & body in a file scoped to that repository. If the create operation succeeded, we delete the file. Otherwise, the save file stays on disk and is used to pre-fill the next `issue/mr` create invocation, which is most likely a retry after a server or validation failure. Or instead of prefilling, we provide `--recover` flag to continue from the last session **Additional context** <!-- Add any other context or screenshots about the feature request here. -->
priority
friendlier recovery when issue mr create fails describe the feature or problem you d like to solve when glab mr create fails or cancelled the body content typed using the editor is lost sometimes the user may type a long mr description and just lose it due to a simple mistake such as not adding any commit to the mr branch yet e g when creating draft mrs propose a solution to me the best user experience would be if the person can just resume where they left off so whenever someone authors an mr or issue we save their title body in a file scoped to that repository if the create operation succeeded we delete the file otherwise the save file stays on disk and is used to pre fill the next issue mr create invocation which is most likely a retry after a server or validation failure or instead of prefilling we provide recover flag to continue from the last session additional context
1
17,084
2,615,130,537
IssuesEvent
2015-03-01 06:00:25
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
opened
Maps Coordinate sample
auto-migrated Priority-Medium Type-Sample
``` Which Google API and version: coordinate v1 Java environment (e.g. Java 6, Android 2.3, App Engine)? Java6 External references, such as API reference guide? https://developers.google.com/coordinate/ Please provide any additional information below. Service accounts are not currently supported on Google Maps Coordinate API, so the web server flow should demonstrated. Just like in the Fusion Tables API sample - http://code.google.com/p/google-api-java-client/wiki/APIs#Fusion_Tables_API ``` Original issue reported on code.google.com by `nicol...@google.com` on 4 Dec 2012 at 7:33
1.0
Maps Coordinate sample - ``` Which Google API and version: coordinate v1 Java environment (e.g. Java 6, Android 2.3, App Engine)? Java6 External references, such as API reference guide? https://developers.google.com/coordinate/ Please provide any additional information below. Service accounts are not currently supported on Google Maps Coordinate API, so the web server flow should demonstrated. Just like in the Fusion Tables API sample - http://code.google.com/p/google-api-java-client/wiki/APIs#Fusion_Tables_API ``` Original issue reported on code.google.com by `nicol...@google.com` on 4 Dec 2012 at 7:33
priority
maps coordinate sample which google api and version coordinate java environment e g java android app engine external references such as api reference guide please provide any additional information below service accounts are not currently supported on google maps coordinate api so the web server flow should demonstrated just like in the fusion tables api sample original issue reported on code google com by nicol google com on dec at
1
766,175
26,873,659,755
IssuesEvent
2023-02-04 19:33:42
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
closed
[API.Task] "Status 500 Internal server error" response on attempt to create task with valid values
bug Priority: Medium API
**Environment:** macOS High Sierra Version 10.13.6 Postman for Mac Version 10.8.7 **Reproducible:** always **Preconditions** 1. Run Postman 2. Create new request > Choose 'POST' in 'Method' drop-down 3. Enter request URL: https://speak-ukrainian.org.ua/dev/api/challenge/x/task 4. Navigate to 'Authorization' tab, choose 'Bearer Token' in 'Type' drop-down 5. Set 'accessToken' of user with admin rights "eyJhbGciOiJIUzUxMiJ9.eyJzdWIiOiJhZG1pbkBnbWFpbC5jb20iLCJqdGkiOiIxIiwiZXhwIjoxNjc1Mzc0ODczfQ.dyThALXsOb7qrrG2cr2iUEVAUKloiGTei7Z5GCgaVZcpjPL0a3E-RPBb2Q75b8bBPvXb1YURPgqPKqBNRV5NiA" * You need to replace "x" in URL with the existing challenge ID (you can check the site or DB) TD1: "{ "name": " namenamename1213#$% ", "description": " descriptiondescriptiondescriptiondescriptiondescription12345$%%^$# ", "picture": "/upload/test/test.png", "startDate": "2021-12-03" }" **Steps to reproduce** 1. Choose body>raw>JSON and enter data(TD1) in field 2. Click on 'Send' button **Actual result** Server is responded with status 500 **Expected result** Server is responded with status 200. Task is successfully created. **User story and test case links** [Test case](https://jira.softserve.academy/browse/TUA-441) **Labels to be added** "Bug", Priority ("pri: Medium "), Severity ("severity: Major"), Type "API" ![tua-441](https://user-images.githubusercontent.com/122798739/216449145-7bb1b86d-df05-4606-9778-5bf43e8e6d1a.png)
1.0
[API.Task] "Status 500 Internal server error" response on attempt to create task with valid values - **Environment:** macOS High Sierra Version 10.13.6 Postman for Mac Version 10.8.7 **Reproducible:** always **Preconditions** 1. Run Postman 2. Create new request > Choose 'POST' in 'Method' drop-down 3. Enter request URL: https://speak-ukrainian.org.ua/dev/api/challenge/x/task 4. Navigate to 'Authorization' tab, choose 'Bearer Token' in 'Type' drop-down 5. Set 'accessToken' of user with admin rights "eyJhbGciOiJIUzUxMiJ9.eyJzdWIiOiJhZG1pbkBnbWFpbC5jb20iLCJqdGkiOiIxIiwiZXhwIjoxNjc1Mzc0ODczfQ.dyThALXsOb7qrrG2cr2iUEVAUKloiGTei7Z5GCgaVZcpjPL0a3E-RPBb2Q75b8bBPvXb1YURPgqPKqBNRV5NiA" * You need to replace "x" in URL with the existing challenge ID (you can check the site or DB) TD1: "{ "name": " namenamename1213#$% ", "description": " descriptiondescriptiondescriptiondescriptiondescription12345$%%^$# ", "picture": "/upload/test/test.png", "startDate": "2021-12-03" }" **Steps to reproduce** 1. Choose body>raw>JSON and enter data(TD1) in field 2. Click on 'Send' button **Actual result** Server is responded with status 500 **Expected result** Server is responded with status 200. Task is successfully created. **User story and test case links** [Test case](https://jira.softserve.academy/browse/TUA-441) **Labels to be added** "Bug", Priority ("pri: Medium "), Severity ("severity: Major"), Type "API" ![tua-441](https://user-images.githubusercontent.com/122798739/216449145-7bb1b86d-df05-4606-9778-5bf43e8e6d1a.png)
priority
status internal server error response on attempt to create task with valid values environment macos high sierra version postman for mac version reproducible always preconditions run postman create new request choose post in method drop down enter request url navigate to authorization tab choose bearer token in type drop down set accesstoken of user with admin rights you need to replace x in url with the existing challenge id you can check the site or db name description picture upload test test png startdate steps to reproduce choose body raw json and enter data in field click on send button actual result server is responded with status expected result server is responded with status task is successfully created user story and test case links labels to be added bug priority pri medium severity severity major type api
1
5,725
2,578,768,550
IssuesEvent
2015-02-13 02:13:13
phetsims/website
https://api.github.com/repos/phetsims/website
closed
Header Focus Color Change
Medium Priority
Can the outline be a light color (like white) for focus objects in the blue header area? For example, it’s hard to tell when the PhET logos are focused, and impossible to tell when the “search” button is focused (for me at least).
1.0
Header Focus Color Change - Can the outline be a light color (like white) for focus objects in the blue header area? For example, it’s hard to tell when the PhET logos are focused, and impossible to tell when the “search” button is focused (for me at least).
priority
header focus color change can the outline be a light color like white for focus objects in the blue header area for example it’s hard to tell when the phet logos are focused and impossible to tell when the “search” button is focused for me at least
1
87,641
3,756,436,865
IssuesEvent
2016-03-13 10:26:19
kotcrab/vis-editor
https://api.github.com/repos/kotcrab/vis-editor
opened
Split FileChooser into separate extension
priority: medium ui
It uses AWT imports which creates lint error when doing gradle build, probably would be the best to split it into separate module.
1.0
Split FileChooser into separate extension - It uses AWT imports which creates lint error when doing gradle build, probably would be the best to split it into separate module.
priority
split filechooser into separate extension it uses awt imports which creates lint error when doing gradle build probably would be the best to split it into separate module
1
22,250
2,648,473,741
IssuesEvent
2015-03-14 00:07:22
prikhi/pencil
https://api.github.com/repos/prikhi/pencil
closed
"Export as PNG" broken in Firefox 36
1 star bug imported Priority-Medium
_From [m.overme...@yahoo.ca](https://code.google.com/u/108010164991601365011/) on March 03, 2015 08:33:14_ What steps will reproduce the problem? 1. Update to Firefox 36 2. Open Pencil Sketching Firefox add-on 3. Try to "Export as PNG" What is the expected output? What do you see instead? The file never appears on the file system. What version of the product are you using? On what operating system? I'm using the Firefox addon in Firefox 36, on Windows 7 Professional. Please provide any additional information below. -"Export as PNG" works as expected in Firefox 35.0.1. -This is probably due to the changes Firefox made to addon compatibility ( https://blog.mozilla.org/addons/2015/01/13/compatibility-for-firefox-36/ ) -The workaround is to use the Windows build of Pencil. _Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=668_
1.0
"Export as PNG" broken in Firefox 36 - _From [m.overme...@yahoo.ca](https://code.google.com/u/108010164991601365011/) on March 03, 2015 08:33:14_ What steps will reproduce the problem? 1. Update to Firefox 36 2. Open Pencil Sketching Firefox add-on 3. Try to "Export as PNG" What is the expected output? What do you see instead? The file never appears on the file system. What version of the product are you using? On what operating system? I'm using the Firefox addon in Firefox 36, on Windows 7 Professional. Please provide any additional information below. -"Export as PNG" works as expected in Firefox 35.0.1. -This is probably due to the changes Firefox made to addon compatibility ( https://blog.mozilla.org/addons/2015/01/13/compatibility-for-firefox-36/ ) -The workaround is to use the Windows build of Pencil. _Original issue: http://code.google.com/p/evoluspencil/issues/detail?id=668_
priority
export as png broken in firefox from on march what steps will reproduce the problem update to firefox open pencil sketching firefox add on try to export as png what is the expected output what do you see instead the file never appears on the file system what version of the product are you using on what operating system i m using the firefox addon in firefox on windows professional please provide any additional information below export as png works as expected in firefox this is probably due to the changes firefox made to addon compatibility the workaround is to use the windows build of pencil original issue
1
443,589
12,796,649,795
IssuesEvent
2020-07-02 10:50:27
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
[0.9.0 staging-1636] Gigot sleeve shirt icon/model issue
Priority: Medium
So, now Gigot sleeve shirt is looking like that? ![bandicam 2020-07-02 13-05-02-050](https://user-images.githubusercontent.com/27898520/86350058-8bc27e00-bc7b-11ea-939e-4b8196ecafa7.jpg) But the icon looks different with the vest part
1.0
[0.9.0 staging-1636] Gigot sleeve shirt icon/model issue - So, now Gigot sleeve shirt is looking like that? ![bandicam 2020-07-02 13-05-02-050](https://user-images.githubusercontent.com/27898520/86350058-8bc27e00-bc7b-11ea-939e-4b8196ecafa7.jpg) But the icon looks different with the vest part
priority
gigot sleeve shirt icon model issue so now gigot sleeve shirt is looking like that but the icon looks different with the vest part
1
612,092
18,990,667,406
IssuesEvent
2021-11-22 06:44:47
turbot/steampipe-plugin-aws
https://api.github.com/repos/turbot/steampipe-plugin-aws
closed
Add table aws_elasticache_redis_metric_cpu_utilization_daily
enhancement new table priority:medium
**References** Add any related links that will help us understand the resource, including vendor documentation, related GitHub issues, and Go SDK documentation. Engine level cloudwatch metric of EngineCPUUtilization can be considered https://docs.aws.amazon.com/AmazonElastiCache/latest/red-ug/CacheMetrics.Redis.html
1.0
Add table aws_elasticache_redis_metric_cpu_utilization_daily - **References** Add any related links that will help us understand the resource, including vendor documentation, related GitHub issues, and Go SDK documentation. Engine level cloudwatch metric of EngineCPUUtilization can be considered https://docs.aws.amazon.com/AmazonElastiCache/latest/red-ug/CacheMetrics.Redis.html
priority
add table aws elasticache redis metric cpu utilization daily references add any related links that will help us understand the resource including vendor documentation related github issues and go sdk documentation engine level cloudwatch metric of enginecpuutilization can be considered
1
800,592
28,372,154,628
IssuesEvent
2023-04-12 17:50:01
svthalia/Reaxit
https://api.github.com/repos/svthalia/Reaxit
opened
Double-click unlikes is odd
bug priority: medium
### Describe the bug Double-click unlikes is odd ### How to reproduce Steps to reproduce the behavior: 1. Double click a liked photo 2. Observe ### Expected behaviour It would like the photo (again), but not unlike. This is unusual, look at instagram ### Screenshots N/A ### Environment: N/A ### Additional context N/A
1.0
Double-click unlikes is odd - ### Describe the bug Double-click unlikes is odd ### How to reproduce Steps to reproduce the behavior: 1. Double click a liked photo 2. Observe ### Expected behaviour It would like the photo (again), but not unlike. This is unusual, look at instagram ### Screenshots N/A ### Environment: N/A ### Additional context N/A
priority
double click unlikes is odd describe the bug double click unlikes is odd how to reproduce steps to reproduce the behavior double click a liked photo observe expected behaviour it would like the photo again but not unlike this is unusual look at instagram screenshots n a environment n a additional context n a
1
25,235
2,678,312,524
IssuesEvent
2015-03-26 09:53:11
hials/HTML5-GroupProject
https://api.github.com/repos/hials/HTML5-GroupProject
closed
Javascript applications for the "JavaScript Fun" page
medium-priority
We need one or more javascript applications for the "JavaScript Fun" page
1.0
Javascript applications for the "JavaScript Fun" page - We need one or more javascript applications for the "JavaScript Fun" page
priority
javascript applications for the javascript fun page we need one or more javascript applications for the javascript fun page
1
692,872
23,752,056,104
IssuesEvent
2022-08-31 21:44:06
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] Strict inequality conditions are pushed down as non-strict inequalities to DocDB
kind/bug area/ysql priority/medium
Jira Link: [DB-1170](https://yugabyte.atlassian.net/browse/DB-1170) ### Description It seems that if a YSQL user requests a filter of the form `c < 3`, the corresponding request sent to DocDB will have `c <= 3`, causing an extra suite of seeks. The extra rows produced are filtered at the YSQL layer. However, this filtering is not reflected in the `EXPLAIN ANALYZE` output. Consider the following steps to reproduce this issue: 1) `CREATE TABLE rrr(r1 INT, r2 INT, v INT, PRIMARY KEY(r1 ASC, r2 ASC));` 2) `INSERT INTO rrr SELECT i/10, i%10, i FROM generate_series(1, 100) i;` 3) `set yb_debug_log_docdb_requests=true;` 4) `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 < 2 and r2 = 1;` 5) Observe how the RPC sent by this has a `QL_OP_LESS_THAN_EQUAL` operator in the condition specifier instead of `QL_OP_LESS_THAN` 6) Observe, how `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 <= 2 and r2 = 1;` also takes less time than the query in step 4 as no extra filtering has to be done. In step 4, note that there were no displayed rows filtered by recheck in the output. You can also verify that rows were indeed filtered by putting a breakpoint on the codepath that returns false over [here](https://github.com/yugabyte/yugabyte-db/blob/0307f302e99737ad6cd0411657e593880bc53c10/src/postgres/src/backend/access/yb_access/yb_scan.c#L1303). The extra seeks resulting from this become more non-negligible as you add more columns and strict inequality conditions.
1.0
[YSQL] Strict inequality conditions are pushed down as non-strict inequalities to DocDB - Jira Link: [DB-1170](https://yugabyte.atlassian.net/browse/DB-1170) ### Description It seems that if a YSQL user requests a filter of the form `c < 3`, the corresponding request sent to DocDB will have `c <= 3`, causing an extra suite of seeks. The extra rows produced are filtered at the YSQL layer. However, this filtering is not reflected in the `EXPLAIN ANALYZE` output. Consider the following steps to reproduce this issue: 1) `CREATE TABLE rrr(r1 INT, r2 INT, v INT, PRIMARY KEY(r1 ASC, r2 ASC));` 2) `INSERT INTO rrr SELECT i/10, i%10, i FROM generate_series(1, 100) i;` 3) `set yb_debug_log_docdb_requests=true;` 4) `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 < 2 and r2 = 1;` 5) Observe how the RPC sent by this has a `QL_OP_LESS_THAN_EQUAL` operator in the condition specifier instead of `QL_OP_LESS_THAN` 6) Observe, how `EXPLAIN ANALYZE SELECT * FROM rrr WHERE r1 <= 2 and r2 = 1;` also takes less time than the query in step 4 as no extra filtering has to be done. In step 4, note that there were no displayed rows filtered by recheck in the output. You can also verify that rows were indeed filtered by putting a breakpoint on the codepath that returns false over [here](https://github.com/yugabyte/yugabyte-db/blob/0307f302e99737ad6cd0411657e593880bc53c10/src/postgres/src/backend/access/yb_access/yb_scan.c#L1303). The extra seeks resulting from this become more non-negligible as you add more columns and strict inequality conditions.
priority
strict inequality conditions are pushed down as non strict inequalities to docdb jira link description it seems that if a ysql user requests a filter of the form c the corresponding request sent to docdb will have c causing an extra suite of seeks the extra rows produced are filtered at the ysql layer however this filtering is not reflected in the explain analyze output consider the following steps to reproduce this issue create table rrr int int v int primary key asc asc insert into rrr select i i i from generate series i set yb debug log docdb requests true explain analyze select from rrr where and observe how the rpc sent by this has a ql op less than equal operator in the condition specifier instead of ql op less than observe how explain analyze select from rrr where and also takes less time than the query in step as no extra filtering has to be done in step note that there were no displayed rows filtered by recheck in the output you can also verify that rows were indeed filtered by putting a breakpoint on the codepath that returns false over the extra seeks resulting from this become more non negligible as you add more columns and strict inequality conditions
1
270,921
8,474,674,241
IssuesEvent
2018-10-24 16:46:51
CS2103-AY1819S1-T13-4/main
https://api.github.com/repos/CS2103-AY1819S1-T13-4/main
opened
Implement Map DS in UniqueModuleList for O(1) look up
priority.Medium severity.Low status.Ongoing type.Enhancement
Improve look up efficiency of UniqueModuleList
1.0
Implement Map DS in UniqueModuleList for O(1) look up - Improve look up efficiency of UniqueModuleList
priority
implement map ds in uniquemodulelist for o look up improve look up efficiency of uniquemodulelist
1
454,151
13,095,675,985
IssuesEvent
2020-08-03 14:29:19
HabitRPG/habitica
https://api.github.com/repos/HabitRPG/habitica
closed
Modals should open without being scrolled
help wanted priority: medium section: Achievements/Popups/Notifications
Often when long modals pop up, they're scrolled to the very bottom by default, like this: <img width="533" alt="screen shot 2017-10-27 at 1 40 52 pm" src="https://user-images.githubusercontent.com/3494081/32124586-24485edc-bb1d-11e7-8dfe-abf0f8915283.png"> They should instead pop up with the top of the modal displaying, like this: <img width="574" alt="screen shot 2017-10-27 at 1 40 55 pm" src="https://user-images.githubusercontent.com/3494081/32124559-0ffbd6e8-bb1d-11e7-90ea-1c8d925c322d.png">
1.0
Modals should open without being scrolled - Often when long modals pop up, they're scrolled to the very bottom by default, like this: <img width="533" alt="screen shot 2017-10-27 at 1 40 52 pm" src="https://user-images.githubusercontent.com/3494081/32124586-24485edc-bb1d-11e7-8dfe-abf0f8915283.png"> They should instead pop up with the top of the modal displaying, like this: <img width="574" alt="screen shot 2017-10-27 at 1 40 55 pm" src="https://user-images.githubusercontent.com/3494081/32124559-0ffbd6e8-bb1d-11e7-90ea-1c8d925c322d.png">
priority
modals should open without being scrolled often when long modals pop up they re scrolled to the very bottom by default like this img width alt screen shot at pm src they should instead pop up with the top of the modal displaying like this img width alt screen shot at pm src
1
651,067
21,464,249,123
IssuesEvent
2022-04-26 00:49:48
Dragon1320/anime-notifier
https://api.github.com/repos/Dragon1320/anime-notifier
closed
Establish database connection
priority: medium type: feature module: database
We should be able to establish a connection to MongoDB given some parameters. Credentials should be optional since we probably won't need them when testing locally. You can interpret 'configurable' however you like - be it function arguments or reading a config file. If it is the latter though, I can recommend some useful libraries. The following should be configurable: - address - port - database Additionally, the following should be optional: - username - password
1.0
Establish database connection - We should be able to establish a connection to MongoDB given some parameters. Credentials should be optional since we probably won't need them when testing locally. You can interpret 'configurable' however you like - be it function arguments or reading a config file. If it is the latter though, I can recommend some useful libraries. The following should be configurable: - address - port - database Additionally, the following should be optional: - username - password
priority
establish database connection we should be able to establish a connection to mongodb given some parameters credentials should be optional since we probably won t need them when testing locally you can interpret configurable however you like be it function arguments or reading a config file if it is the latter though i can recommend some useful libraries the following should be configurable address port database additionally the following should be optional username password
1
693,805
23,791,209,389
IssuesEvent
2022-09-02 14:42:01
meisnate12/Plex-Meta-Manager
https://api.github.com/repos/meisnate12/Plex-Meta-Manager
closed
Feature Request: I'd like to see a way to publish various collections on the home screen for a set number of days.
enhancement priority:medium
<!--- Please make sure you submit all Pull Requests to the develop branch not the master branch. ---> **Is your feature request related to a problem? Please describe.** Similar to how Netflix has lists you can scroll through, I'd like to do this with Plex so users might be presented with an option they hadn't considered before. **Describe the solution you'd like** I'd like to be able to either have a random collection assigned to the screen for a set period of time, or a list of collections that could change on a rotating basis **Describe alternatives you've considered** I can make it work on a limited basis using a method similar to the holiday collections that are able to be pinned to the home screen. I have a list of collections and gave them all a date range. My biggest issues are how they are ordered since you can only use custom sorting on one collection, and how to have them change out. Currently I have dates ranges attached to each one.
1.0
Feature Request: I'd like to see a way to publish various collections on the home screen for a set number of days. - <!--- Please make sure you submit all Pull Requests to the develop branch not the master branch. ---> **Is your feature request related to a problem? Please describe.** Similar to how Netflix has lists you can scroll through, I'd like to do this with Plex so users might be presented with an option they hadn't considered before. **Describe the solution you'd like** I'd like to be able to either have a random collection assigned to the screen for a set period of time, or a list of collections that could change on a rotating basis **Describe alternatives you've considered** I can make it work on a limited basis using a method similar to the holiday collections that are able to be pinned to the home screen. I have a list of collections and gave them all a date range. My biggest issues are how they are ordered since you can only use custom sorting on one collection, and how to have them change out. Currently I have dates ranges attached to each one.
priority
feature request i d like to see a way to publish various collections on the home screen for a set number of days please make sure you submit all pull requests to the develop branch not the master branch is your feature request related to a problem please describe similar to how netflix has lists you can scroll through i d like to do this with plex so users might be presented with an option they hadn t considered before describe the solution you d like i d like to be able to either have a random collection assigned to the screen for a set period of time or a list of collections that could change on a rotating basis describe alternatives you ve considered i can make it work on a limited basis using a method similar to the holiday collections that are able to be pinned to the home screen i have a list of collections and gave them all a date range my biggest issues are how they are ordered since you can only use custom sorting on one collection and how to have them change out currently i have dates ranges attached to each one
1
159,231
6,042,453,965
IssuesEvent
2017-06-11 13:11:05
psi-plus/main
https://api.github.com/repos/psi-plus/main
closed
Косметический фикс на кликабельное фото из vCard.
OpSys-All Priority-Medium Type-Enhancement Usability
Original [issue 12](https://code.google.com/p/psi-dev/issues/detail?id=12) created by psi-plus on 2008-12-24T09:00:58.000Z: 1. Запрещает изменять размер окна с фото. 2. Открывает единственное окно. 3. Окно типа ToolWindow.
1.0
Косметический фикс на кликабельное фото из vCard. - Original [issue 12](https://code.google.com/p/psi-dev/issues/detail?id=12) created by psi-plus on 2008-12-24T09:00:58.000Z: 1. Запрещает изменять размер окна с фото. 2. Открывает единственное окно. 3. Окно типа ToolWindow.
priority
косметический фикс на кликабельное фото из vcard original created by psi plus on запрещает изменять размер окна с фото открывает единственное окно окно типа toolwindow
1
462,468
13,247,788,889
IssuesEvent
2020-08-19 17:51:02
radical-cybertools/radical.pilot
https://api.github.com/repos/radical-cybertools/radical.pilot
closed
Configurable SMT level
comp:agent:executor comp:pmgr:launcher layer:rp layer:saga priority:medium topic:resource type:enhancement
https://github.com/radical-cybertools/radical.saga/blob/devel/src/radical/saga/adaptors/lsf/lsfjob.py#L38 `SMT` can be adjusted by a user in some way by either resource config, parameters, or ENV variables (e.g. `export RADICAL_SAGA_SMT`). I was in a conversation and this was raised if a user can specify different SMT level instead of fixed SMT4.
1.0
Configurable SMT level - https://github.com/radical-cybertools/radical.saga/blob/devel/src/radical/saga/adaptors/lsf/lsfjob.py#L38 `SMT` can be adjusted by a user in some way by either resource config, parameters, or ENV variables (e.g. `export RADICAL_SAGA_SMT`). I was in a conversation and this was raised if a user can specify different SMT level instead of fixed SMT4.
priority
configurable smt level smt can be adjusted by a user in some way by either resource config parameters or env variables e g export radical saga smt i was in a conversation and this was raised if a user can specify different smt level instead of fixed
1