Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
31,497
2,733,156,095
IssuesEvent
2015-04-17 12:09:56
tjcsl/cslbot
https://api.github.com/repos/tjcsl/cslbot
closed
gizoogle filter strips anything after an opening HTML (<) tag
bot bug high priority
No description given. Issue created by jwoglom!~jwoglom@unaffiliated/jwoglom
1.0
gizoogle filter strips anything after an opening HTML (<) tag - No description given. Issue created by jwoglom!~jwoglom@unaffiliated/jwoglom
priority
gizoogle filter strips anything after an opening html tag no description given issue created by jwoglom jwoglom unaffiliated jwoglom
1
304,193
9,322,383,408
IssuesEvent
2019-03-27 08:00:55
rucio/rucio
https://api.github.com/repos/rucio/rucio
opened
traceback error in conveyor-submitter and judge*
Priority: High bug
Motivation ---------- When no scheme is specified in the configuration, the conveyor fails to find a common scheme for the source and destination files, here `gsiftp`: ``` 2019-03-27 08:51:47,386 3214 CRITICAL Exception happened when trying to get transfer for request 5d9c6d99ebf04465ad3bef4d8078219c: Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/rucio/core/transfer.py", line 624, in get_transfer_requests_and_source_replicas 'schemes': __add_compatible_schemes(schemes=[matching_scheme[0]], allowed_schemes=current_schemes), File "/usr/local/lib/python2.7/dist-packages/rucio/core/transfer.py", line 923, in __add_compatible_schemes if scheme in allowed_schemes: TypeError: argument of type 'NoneType' is not iterable ``` The respective rules stay in the STUCK state when asked for reevaluation: ``` 2019-03-27 08:49:23,677 3213 INFO rule_repairer[0/0]: Repairing rule 32d6775ead104b2c8be50cc946f01782 2019-03-27 08:49:23,749 3213 INFO Rule 32d6775ead104b2c8be50cc946f01782 [0/0/1] state=STUCK ```
1.0
traceback error in conveyor-submitter and judge* - Motivation ---------- When no scheme is specified in the configuration, the conveyor fails to find a common scheme for the source and destination files, here `gsiftp`: ``` 2019-03-27 08:51:47,386 3214 CRITICAL Exception happened when trying to get transfer for request 5d9c6d99ebf04465ad3bef4d8078219c: Traceback (most recent call last): File "/usr/local/lib/python2.7/dist-packages/rucio/core/transfer.py", line 624, in get_transfer_requests_and_source_replicas 'schemes': __add_compatible_schemes(schemes=[matching_scheme[0]], allowed_schemes=current_schemes), File "/usr/local/lib/python2.7/dist-packages/rucio/core/transfer.py", line 923, in __add_compatible_schemes if scheme in allowed_schemes: TypeError: argument of type 'NoneType' is not iterable ``` The respective rules stay in the STUCK state when asked for reevaluation: ``` 2019-03-27 08:49:23,677 3213 INFO rule_repairer[0/0]: Repairing rule 32d6775ead104b2c8be50cc946f01782 2019-03-27 08:49:23,749 3213 INFO Rule 32d6775ead104b2c8be50cc946f01782 [0/0/1] state=STUCK ```
priority
traceback error in conveyor submitter and judge motivation when no scheme is specified in the configuration the conveyor fails to find a common scheme for the source and destination files here gsiftp critical exception happened when trying to get transfer for request traceback most recent call last file usr local lib dist packages rucio core transfer py line in get transfer requests and source replicas schemes add compatible schemes schemes allowed schemes current schemes file usr local lib dist packages rucio core transfer py line in add compatible schemes if scheme in allowed schemes typeerror argument of type nonetype is not iterable the respective rules stay in the stuck state when asked for reevaluation info rule repairer repairing rule info rule state stuck
1
173,031
6,519,215,041
IssuesEvent
2017-08-28 11:47:49
CanberraOceanRacingClub/namadgi3
https://api.github.com/repos/CanberraOceanRacingClub/namadgi3
opened
Bar crossing SOP
priority 1: High
Many members will be unfamiliar with crossing river bars in a large vessel. The Clyde River Bar presents a special problem for CORC member.
1.0
Bar crossing SOP - Many members will be unfamiliar with crossing river bars in a large vessel. The Clyde River Bar presents a special problem for CORC member.
priority
bar crossing sop many members will be unfamiliar with crossing river bars in a large vessel the clyde river bar presents a special problem for corc member
1
637,571
20,672,128,676
IssuesEvent
2022-03-10 04:11:41
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Cannot subscribe to webhook APIs in APIM configured with Postgres
Type/Bug Severity/Blocker Priority/Highest WUM Affected/4.0.0
### Description: Cannot subscribe to webhook API. the following error is thrown. The reason for this error is timestamp is set to null in WebhooksDAO.addSubscription() in https://github.com/wso2/carbon-apimgt/blob/master/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/WebhooksDAO.java#L182 **SecretValidationTestCase** and **WebSubAPITestCase** test cases are failing due to this [2021-10-13 12:08:29,456] ERROR - NotifyApiServiceImpl Error while processing notification org.wso2.carbon.apimgt.api.APIManagementException: Error while adding subscriptions request for callbackhttp://www.google.com for the API 35e130c1-d5a4-47b5-9e7d-8909e4fdebbe at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription_aroundBody10(WebhooksDAO.java:187) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription(WebhooksDAO.java:163) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription_aroundBody2(WebhooksDAO.java:87) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription(WebhooksDAO.java:75) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.notification.WebhooksSubscriptionEventHandler.handleEvent(WebhooksSubscriptionEventHandler.java:51) ~[org.wso2.carbon.apimgt.notification_9.0.174.jar:?] at org.wso2.carbon.apimgt.notification.NotificationEventService.processEvent(NotificationEventService.java:44) ~[org.wso2.carbon.apimgt.notification_9.0.174.jar:?] at org.wso2.carbon.apimgt.internal.service.impl.NotifyApiServiceImpl.notifyPost(NotifyApiServiceImpl.java:29) [classes/:?] at org.wso2.carbon.apimgt.internal.service.NotifyApi.notifyPost(NotifyApi.java:47) [classes/:?] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_161] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_161] : : at org.apache.tomcat.util.threads.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:659) [tomcat_9.0.52.wso2v2.jar:?] at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.52.wso2v2.jar:?] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_161] Caused by: org.postgresql.util.PSQLException: ERROR: column "delivered_at" is of type timestamp without time zone but expression is of type character varying Hint: You will need to rewrite or cast the expression. Position: 228 at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2284) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2003) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:200) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:424) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:161) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.jdbc.PgPreparedStatement.executeUpdate(PgPreparedStatement.java:133) ~[postgresql-9.4.1208.jar:9.4.1208] at sun.reflect.GeneratedMethodAccessor143.invoke(Unknown Source) ~[?:?] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_161] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_161] at org.apache.tomcat.jdbc.pool.StatementFacade$StatementProxy.invoke(StatementFacade.java:114) ~[jdbc-pool_9.0.35.wso2v1.jar:?] at com.sun.proxy.$Proxy47.executeUpdate(Unknown Source) ~[?:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription_aroundBody10(WebhooksDAO.java:185) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] ... 59 more ### Steps to reproduce: 1. Setup webook and subscribe as mentioned in https://apim.docs.wso2.com/en/latest/tutorials/streaming-api/create-and-publish-websub-api/ ### Affected Product Version: APIM 4.0.0 ### Environment details (with versions): DB: Postgres 10.x
1.0
Cannot subscribe to webhook APIs in APIM configured with Postgres - ### Description: Cannot subscribe to webhook API. the following error is thrown. The reason for this error is timestamp is set to null in WebhooksDAO.addSubscription() in https://github.com/wso2/carbon-apimgt/blob/master/components/apimgt/org.wso2.carbon.apimgt.impl/src/main/java/org/wso2/carbon/apimgt/impl/dao/WebhooksDAO.java#L182 **SecretValidationTestCase** and **WebSubAPITestCase** test cases are failing due to this [2021-10-13 12:08:29,456] ERROR - NotifyApiServiceImpl Error while processing notification org.wso2.carbon.apimgt.api.APIManagementException: Error while adding subscriptions request for callbackhttp://www.google.com for the API 35e130c1-d5a4-47b5-9e7d-8909e4fdebbe at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription_aroundBody10(WebhooksDAO.java:187) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription(WebhooksDAO.java:163) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription_aroundBody2(WebhooksDAO.java:87) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription(WebhooksDAO.java:75) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] at org.wso2.carbon.apimgt.notification.WebhooksSubscriptionEventHandler.handleEvent(WebhooksSubscriptionEventHandler.java:51) ~[org.wso2.carbon.apimgt.notification_9.0.174.jar:?] at org.wso2.carbon.apimgt.notification.NotificationEventService.processEvent(NotificationEventService.java:44) ~[org.wso2.carbon.apimgt.notification_9.0.174.jar:?] at org.wso2.carbon.apimgt.internal.service.impl.NotifyApiServiceImpl.notifyPost(NotifyApiServiceImpl.java:29) [classes/:?] at org.wso2.carbon.apimgt.internal.service.NotifyApi.notifyPost(NotifyApi.java:47) [classes/:?] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:1.8.0_161] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:1.8.0_161] : : at org.apache.tomcat.util.threads.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:659) [tomcat_9.0.52.wso2v2.jar:?] at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.52.wso2v2.jar:?] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_161] Caused by: org.postgresql.util.PSQLException: ERROR: column "delivered_at" is of type timestamp without time zone but expression is of type character varying Hint: You will need to rewrite or cast the expression. Position: 228 at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2284) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:2003) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:200) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.jdbc.PgStatement.execute(PgStatement.java:424) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.jdbc.PgPreparedStatement.executeWithFlags(PgPreparedStatement.java:161) ~[postgresql-9.4.1208.jar:9.4.1208] at org.postgresql.jdbc.PgPreparedStatement.executeUpdate(PgPreparedStatement.java:133) ~[postgresql-9.4.1208.jar:9.4.1208] at sun.reflect.GeneratedMethodAccessor143.invoke(Unknown Source) ~[?:?] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_161] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_161] at org.apache.tomcat.jdbc.pool.StatementFacade$StatementProxy.invoke(StatementFacade.java:114) ~[jdbc-pool_9.0.35.wso2v1.jar:?] at com.sun.proxy.$Proxy47.executeUpdate(Unknown Source) ~[?:?] at org.wso2.carbon.apimgt.impl.dao.WebhooksDAO.addSubscription_aroundBody10(WebhooksDAO.java:185) ~[org.wso2.carbon.apimgt.impl_9.0.174.51.jar:?] ... 59 more ### Steps to reproduce: 1. Setup webook and subscribe as mentioned in https://apim.docs.wso2.com/en/latest/tutorials/streaming-api/create-and-publish-websub-api/ ### Affected Product Version: APIM 4.0.0 ### Environment details (with versions): DB: Postgres 10.x
priority
cannot subscribe to webhook apis in apim configured with postgres description cannot subscribe to webhook api the following error is thrown the reason for this error is timestamp is set to null in webhooksdao addsubscription in secretvalidationtestcase and websubapitestcase test cases are failing due to this error notifyapiserviceimpl error while processing notification org carbon apimgt api apimanagementexception error while adding subscriptions request for callback for the api at org carbon apimgt impl dao webhooksdao addsubscription webhooksdao java at org carbon apimgt impl dao webhooksdao addsubscription webhooksdao java at org carbon apimgt impl dao webhooksdao addsubscription webhooksdao java at org carbon apimgt impl dao webhooksdao addsubscription webhooksdao java at org carbon apimgt notification webhookssubscriptioneventhandler handleevent webhookssubscriptioneventhandler java at org carbon apimgt notification notificationeventservice processevent notificationeventservice java at org carbon apimgt internal service impl notifyapiserviceimpl notifypost notifyapiserviceimpl java at org carbon apimgt internal service notifyapi notifypost notifyapi java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at org apache tomcat util threads threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java caused by org postgresql util psqlexception error column delivered at is of type timestamp without time zone but expression is of type character varying hint you will need to rewrite or cast the expression position at org postgresql core queryexecutorimpl receiveerrorresponse queryexecutorimpl java at org postgresql core queryexecutorimpl processresults queryexecutorimpl java at org postgresql core queryexecutorimpl execute queryexecutorimpl java at org postgresql jdbc pgstatement execute pgstatement java at org postgresql jdbc pgpreparedstatement executewithflags pgpreparedstatement java at org postgresql jdbc pgpreparedstatement executeupdate pgpreparedstatement java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache tomcat jdbc pool statementfacade statementproxy invoke statementfacade java at com sun proxy executeupdate unknown source at org carbon apimgt impl dao webhooksdao addsubscription webhooksdao java more steps to reproduce setup webook and subscribe as mentioned in affected product version apim environment details with versions db postgres x
1
578,829
17,155,783,599
IssuesEvent
2021-07-14 06:36:35
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
The User_Code in OAuth Device Flow cannot be configured to have a user defined keyset
Component/OAuth Priority/High Severity/Major improvement
**Description:** The User_Code in device flow is generated from a static key set [1] and hence, cannot be configured by the admin. [1] - https://github.com/wso2-extensions/identity-inbound-auth-oauth/blob/b211bbcddaa6006ae0b40f89c363ea4480a009e9/components/org.wso2.carbon.identity.oauth/src/main/java/org/wso2/carbon/identity/oauth2/device/constants/Constants.java#L59
1.0
The User_Code in OAuth Device Flow cannot be configured to have a user defined keyset - **Description:** The User_Code in device flow is generated from a static key set [1] and hence, cannot be configured by the admin. [1] - https://github.com/wso2-extensions/identity-inbound-auth-oauth/blob/b211bbcddaa6006ae0b40f89c363ea4480a009e9/components/org.wso2.carbon.identity.oauth/src/main/java/org/wso2/carbon/identity/oauth2/device/constants/Constants.java#L59
priority
the user code in oauth device flow cannot be configured to have a user defined keyset description the user code in device flow is generated from a static key set and hence cannot be configured by the admin
1
603,993
18,675,296,000
IssuesEvent
2021-10-31 13:05:25
FantasticoFox/DataAccounting
https://api.github.com/repos/FantasticoFox/DataAccounting
opened
[Feature] Implement Chunking to keep remote-verification feasable
high priority ux
It takes time to verify revisions, it takes longer time with Signatures as it is an additional compute step and it takes longest with remote lockups on-chain to check if a time-stamp is correct. If a page has more then 50 revisions this can take 30 s+ on a remote server (especially if the server is on the other side of the world and you have ~250 ms latency). To avoid unmanageable chain-verification issues we can 'chunk' every 50 revisions (it's arbitrary but goes along with the blob of loading 50 revisions in MW). Implementation specification: * There is a manual way to do chunking by just 'moving a page and renaming it with the last page_verification hash in the title as an identifier' - this allows when linking the old page to have the state of the page as well. If 'correct linking' is implemented this becomes obsolete as the pages will link with page_verification hashes. * An automated implemented would detect how many verified revisions exist and move the page as a consequence of succeeding 50 pages with auto-rename and should be turned into 'read only' [Action:Protect] to restrict the page from any further manipulation. The original page should have (in the content section) a new footer added called 'Data Accounting Chunking' with a drop down menu, implemented like in the signatures. ![image](https://user-images.githubusercontent.com/45313235/139584818-69c1322f-f688-42db-ab9d-45784ae2affb.png)
1.0
[Feature] Implement Chunking to keep remote-verification feasable - It takes time to verify revisions, it takes longer time with Signatures as it is an additional compute step and it takes longest with remote lockups on-chain to check if a time-stamp is correct. If a page has more then 50 revisions this can take 30 s+ on a remote server (especially if the server is on the other side of the world and you have ~250 ms latency). To avoid unmanageable chain-verification issues we can 'chunk' every 50 revisions (it's arbitrary but goes along with the blob of loading 50 revisions in MW). Implementation specification: * There is a manual way to do chunking by just 'moving a page and renaming it with the last page_verification hash in the title as an identifier' - this allows when linking the old page to have the state of the page as well. If 'correct linking' is implemented this becomes obsolete as the pages will link with page_verification hashes. * An automated implemented would detect how many verified revisions exist and move the page as a consequence of succeeding 50 pages with auto-rename and should be turned into 'read only' [Action:Protect] to restrict the page from any further manipulation. The original page should have (in the content section) a new footer added called 'Data Accounting Chunking' with a drop down menu, implemented like in the signatures. ![image](https://user-images.githubusercontent.com/45313235/139584818-69c1322f-f688-42db-ab9d-45784ae2affb.png)
priority
implement chunking to keep remote verification feasable it takes time to verify revisions it takes longer time with signatures as it is an additional compute step and it takes longest with remote lockups on chain to check if a time stamp is correct if a page has more then revisions this can take s on a remote server especially if the server is on the other side of the world and you have ms latency to avoid unmanageable chain verification issues we can chunk every revisions it s arbitrary but goes along with the blob of loading revisions in mw implementation specification there is a manual way to do chunking by just moving a page and renaming it with the last page verification hash in the title as an identifier this allows when linking the old page to have the state of the page as well if correct linking is implemented this becomes obsolete as the pages will link with page verification hashes an automated implemented would detect how many verified revisions exist and move the page as a consequence of succeeding pages with auto rename and should be turned into read only to restrict the page from any further manipulation the original page should have in the content section a new footer added called data accounting chunking with a drop down menu implemented like in the signatures
1
348,709
10,451,736,885
IssuesEvent
2019-09-19 13:26:43
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
Custom post description migrates to image description
Urgent [Priority: HIGH] bug
Excerpts are showing in captions To re-occur, import WordPress.2019-08-28.xml from the attachments, and see the post "Стратегии на E3 2019" and then view it's AMP version. You will find this text added below the image as caption "Обзор всех новостей стратегий с выставки" - THis text is actually the excerpt of the post. [Downloads.zip](https://github.com/ahmedkaludi/accelerated-mobile-pages/files/3562443/Downloads.zip) REF: https://wordpress.org/support/topic/custom-post-description-migrates-to-image-description/#post-11875201
1.0
Custom post description migrates to image description - Excerpts are showing in captions To re-occur, import WordPress.2019-08-28.xml from the attachments, and see the post "Стратегии на E3 2019" and then view it's AMP version. You will find this text added below the image as caption "Обзор всех новостей стратегий с выставки" - THis text is actually the excerpt of the post. [Downloads.zip](https://github.com/ahmedkaludi/accelerated-mobile-pages/files/3562443/Downloads.zip) REF: https://wordpress.org/support/topic/custom-post-description-migrates-to-image-description/#post-11875201
priority
custom post description migrates to image description excerpts are showing in captions to re occur import wordpress xml from the attachments and see the post стратегии на and then view it s amp version you will find this text added below the image as caption обзор всех новостей стратегий с выставки this text is actually the excerpt of the post ref
1
742,237
25,844,469,958
IssuesEvent
2022-12-13 04:45:29
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
Mobile version is blank in mSupply on the fresh install
Priority: high Bug: production Effort: small Solution: Proposed
## Describe the bug There are many mobile sites that don't have version information on the desktop. In the desktop, the mobile version is populated from message.toVersionString with type = 'mobile_upgrade' As discussed with @sworup when the mobile is newly installed, no message seems to send to the desktop. Only when the mobile is upgraded, the message is sent. ### To reproduce Steps to reproduce the behavior: 1. Do a fresh install of mSupply mobile and enter the sync username password to sync data 2. At this point the mobile app should have a site set. 3. In mSupply desktop go to site table to find the site of the current mobile app 4. See current mobile app version is empty ![SCR-20221116-ebo](https://user-images.githubusercontent.com/683452/202127435-f0bbbfc6-8ffb-489a-a9de-64f6fe7c9bd3.png) ### Expected behavior Even during a fresh install, the desktop site table record for the mobile app's site should have the `version` field populated with the current version of the mobile app. ### Proposed Solution We have to do two things: - During fresh install, the mobile version should be sent through message to desktop to update `site.version` properly. - For fresh install too, add a ` createRecord(database, 'UpgradeMessage', fromVersion, toVersion);` call where fromVersion would be an empty string and toVersion would be the current version. This will tell the dashboard user on which day a fresh installation was done, along with showing the current version. At this part of the code: https://github.com/openmsupply/mobile/blob/9ecf8683e74d9cad797191f44c500ca3cd6f5fe6/src/dataMigration.js#L35-L53 Make sure the line below is not triggering a return. https://github.com/openmsupply/mobile/blob/9ecf8683e74d9cad797191f44c500ca3cd6f5fe6/src/dataMigration.js#L39 If it is triggering a return, then the update of the Setting would not trigger, hence not triggering the message that would update `site.version` on desktop. https://github.com/openmsupply/mobile/blob/9ecf8683e74d9cad797191f44c500ca3cd6f5fe6/src/dataMigration.js#L54-L56 This is where I would start. ### Version and device info - App version: 8.6.1 - Tablet model: - OS version: ### Additional context Add any other context about the problem here.
1.0
Mobile version is blank in mSupply on the fresh install - ## Describe the bug There are many mobile sites that don't have version information on the desktop. In the desktop, the mobile version is populated from message.toVersionString with type = 'mobile_upgrade' As discussed with @sworup when the mobile is newly installed, no message seems to send to the desktop. Only when the mobile is upgraded, the message is sent. ### To reproduce Steps to reproduce the behavior: 1. Do a fresh install of mSupply mobile and enter the sync username password to sync data 2. At this point the mobile app should have a site set. 3. In mSupply desktop go to site table to find the site of the current mobile app 4. See current mobile app version is empty ![SCR-20221116-ebo](https://user-images.githubusercontent.com/683452/202127435-f0bbbfc6-8ffb-489a-a9de-64f6fe7c9bd3.png) ### Expected behavior Even during a fresh install, the desktop site table record for the mobile app's site should have the `version` field populated with the current version of the mobile app. ### Proposed Solution We have to do two things: - During fresh install, the mobile version should be sent through message to desktop to update `site.version` properly. - For fresh install too, add a ` createRecord(database, 'UpgradeMessage', fromVersion, toVersion);` call where fromVersion would be an empty string and toVersion would be the current version. This will tell the dashboard user on which day a fresh installation was done, along with showing the current version. At this part of the code: https://github.com/openmsupply/mobile/blob/9ecf8683e74d9cad797191f44c500ca3cd6f5fe6/src/dataMigration.js#L35-L53 Make sure the line below is not triggering a return. https://github.com/openmsupply/mobile/blob/9ecf8683e74d9cad797191f44c500ca3cd6f5fe6/src/dataMigration.js#L39 If it is triggering a return, then the update of the Setting would not trigger, hence not triggering the message that would update `site.version` on desktop. https://github.com/openmsupply/mobile/blob/9ecf8683e74d9cad797191f44c500ca3cd6f5fe6/src/dataMigration.js#L54-L56 This is where I would start. ### Version and device info - App version: 8.6.1 - Tablet model: - OS version: ### Additional context Add any other context about the problem here.
priority
mobile version is blank in msupply on the fresh install describe the bug there are many mobile sites that don t have version information on the desktop in the desktop the mobile version is populated from message toversionstring with type mobile upgrade as discussed with sworup when the mobile is newly installed no message seems to send to the desktop only when the mobile is upgraded the message is sent to reproduce steps to reproduce the behavior do a fresh install of msupply mobile and enter the sync username password to sync data at this point the mobile app should have a site set in msupply desktop go to site table to find the site of the current mobile app see current mobile app version is empty expected behavior even during a fresh install the desktop site table record for the mobile app s site should have the version field populated with the current version of the mobile app proposed solution we have to do two things during fresh install the mobile version should be sent through message to desktop to update site version properly for fresh install too add a createrecord database upgrademessage fromversion toversion call where fromversion would be an empty string and toversion would be the current version this will tell the dashboard user on which day a fresh installation was done along with showing the current version at this part of the code make sure the line below is not triggering a return if it is triggering a return then the update of the setting would not trigger hence not triggering the message that would update site version on desktop this is where i would start version and device info app version tablet model os version additional context add any other context about the problem here
1
178,951
6,620,332,716
IssuesEvent
2017-09-21 15:14:14
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Name days: Convert to structured answer and move code to template
Category: Highest Impact Tasks Improvement Priority: High Topic: Other
### Description <!-- Describe the bug or suggestion in detail --> The name days IA needs to be updated. It's got a lot of legacy code, injecting html into the duckbar which shouldn't be happening. For some reason the styles are messed up now (below is what it should look like) Can you please update this @kirkins? ![screen shot 2017-07-18 at 10 15 09 am](https://user-images.githubusercontent.com/8960296/28309688-06395f38-6ba2-11e7-97da-30e3339fc424.png) ## Steps to recreate <!-- Describe the steps, or provide a link to an example search --> ## People to notify <!-- Please @mention any relevant people/organizations here:--> <!-- LANGUAGE LEADERS ONLY: REMOVE THIS LINE ## Get Started - [ ] 1) Claim this issue by commenting below - [ ] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-goodies/blob/master/CONTRIBUTING.md) - [ ] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository - [ ] 4) Create a Pull Request ## Resources - Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions - Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics - Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help <!-- DO NOT REMOVE --> --- <!-- The Instant Answer ID can be found by clicking the `?` icon beside the Instant Answer result on DuckDuckGo.com --> Instant Answer Page: https://duck.co/ia/view/name_days <!-- FILL THIS IN: ^^^^ -->
1.0
Name days: Convert to structured answer and move code to template - ### Description <!-- Describe the bug or suggestion in detail --> The name days IA needs to be updated. It's got a lot of legacy code, injecting html into the duckbar which shouldn't be happening. For some reason the styles are messed up now (below is what it should look like) Can you please update this @kirkins? ![screen shot 2017-07-18 at 10 15 09 am](https://user-images.githubusercontent.com/8960296/28309688-06395f38-6ba2-11e7-97da-30e3339fc424.png) ## Steps to recreate <!-- Describe the steps, or provide a link to an example search --> ## People to notify <!-- Please @mention any relevant people/organizations here:--> <!-- LANGUAGE LEADERS ONLY: REMOVE THIS LINE ## Get Started - [ ] 1) Claim this issue by commenting below - [ ] 2) Review our [Contributing Guide](https://github.com/duckduckgo/zeroclickinfo-goodies/blob/master/CONTRIBUTING.md) - [ ] 3) [Set up your development environment](https://docs.duckduckhack.com/welcome/setup-dev-environment.html), and fork this repository - [ ] 4) Create a Pull Request ## Resources - Join [DuckDuckHack Slack](https://quackslack.herokuapp.com/) to ask questions - Join the [DuckDuckHack Forum](https://forum.duckduckhack.com/) to discuss project planning and Instant Answer metrics - Read the [DuckDuckHack Documentation](https://docs.duckduckhack.com/) for technical help <!-- DO NOT REMOVE --> --- <!-- The Instant Answer ID can be found by clicking the `?` icon beside the Instant Answer result on DuckDuckGo.com --> Instant Answer Page: https://duck.co/ia/view/name_days <!-- FILL THIS IN: ^^^^ -->
priority
name days convert to structured answer and move code to template description the name days ia needs to be updated it s got a lot of legacy code injecting html into the duckbar which shouldn t be happening for some reason the styles are messed up now below is what it should look like can you please update this kirkins steps to recreate people to notify language leaders only remove this line get started claim this issue by commenting below review our and fork this repository create a pull request resources join to ask questions join the to discuss project planning and instant answer metrics read the for technical help instant answer page
1
402,715
11,813,428,877
IssuesEvent
2020-03-19 22:23:09
sqlalchemy/alembic
https://api.github.com/repos/sqlalchemy/alembic
closed
more regressions with type comaprisons
autogenerate - detection bug high priority
this comaprison should return no change, because "Unicode" is the generic type: ``` diff --git a/tests/test_autogen_diffs.py b/tests/test_autogen_diffs.py index e1e5c8d..7fa4465 100644 --- a/tests/test_autogen_diffs.py +++ b/tests/test_autogen_diffs.py @@ -795,6 +795,8 @@ class CompareMetadataToInspectorTest(TestBase): (Unicode(32), VARCHAR(32), False, config.requirements.unicode_string), (VARCHAR(6), VARCHAR(12), True), (VARCHAR(6), String(12), True), + (mysql.VARCHAR(200, charset='utf8'), Unicode(200), False, ), + (String(255, collation='utf8_bin'), String(255), False) ) def test_string_comparisons(self, cola, colb, expect_changes): is_(self._compare_columns(cola, colb), expect_changes) ```
1.0
more regressions with type comaprisons - this comaprison should return no change, because "Unicode" is the generic type: ``` diff --git a/tests/test_autogen_diffs.py b/tests/test_autogen_diffs.py index e1e5c8d..7fa4465 100644 --- a/tests/test_autogen_diffs.py +++ b/tests/test_autogen_diffs.py @@ -795,6 +795,8 @@ class CompareMetadataToInspectorTest(TestBase): (Unicode(32), VARCHAR(32), False, config.requirements.unicode_string), (VARCHAR(6), VARCHAR(12), True), (VARCHAR(6), String(12), True), + (mysql.VARCHAR(200, charset='utf8'), Unicode(200), False, ), + (String(255, collation='utf8_bin'), String(255), False) ) def test_string_comparisons(self, cola, colb, expect_changes): is_(self._compare_columns(cola, colb), expect_changes) ```
priority
more regressions with type comaprisons this comaprison should return no change because unicode is the generic type diff git a tests test autogen diffs py b tests test autogen diffs py index a tests test autogen diffs py b tests test autogen diffs py class comparemetadatatoinspectortest testbase unicode varchar false config requirements unicode string varchar varchar true varchar string true mysql varchar charset unicode false string collation bin string false def test string comparisons self cola colb expect changes is self compare columns cola colb expect changes
1
379,221
11,217,732,466
IssuesEvent
2020-01-07 09:54:22
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Godot 3.1 RC2 create_trimesh_collision() and create_convex_collision() lead to instant crash on Android
bug high priority platform:android topic:core
<!-- Please search existing issues for potential duplicates before filing yours: https://github.com/godotengine/godot/issues?q=is%3Aissue --> **Godot version:** 3.1 RC2 **OS/device including version:** Oukitel K6000 Pro, Android 6.0 Samsung Galaxy J5, Android 7 **Issue description:** The methods `create_convex_collision()` and `create_trimesh_collision()` of `MeshInstance` (https://docs.godotengine.org/en/latest/classes/class_meshinstance.html lead to an instant crash on Android. The last version I tested was 3.1 beta 7, there it worked, so it seems it was introduced in some of the last betas. I think this is the relevant error description from logcat: > > 03-12 17:30:47.191 32301 32317 E godot : **ERROR**: OpenGL ES 2.0 does not allow retrieving mesh array data > 03-12 17:30:47.191 32301 32317 E godot : At: drivers/gles2/rasterizer_storage_gles2.cpp:2541:mesh_surface_get_array() - OpenGL ES 2.0 does not allow retrieving mesh array data > 03-12 17:30:47.191 32301 32317 E godot : **ERROR**: Condition ' vertex_data.size() == 0 ' is true. returned: Array() > 03-12 17:30:47.191 32301 32317 E godot : At: servers/visual_server.cpp:1606:mesh_surface_get_arrays() - Condition ' vertex_data.size() == 0 ' is true. returned: Array() > 03-12 17:30:47.191 32301 32317 E godot : **ERROR**: FATAL: Index p_index=0 out of size (((Vector<T> *)(this))->_cowdata.size()=0) > 03-12 17:30:47.191 32301 32317 E godot : At: ./core/vector.h:49:operator[]() - FATAL: Index p_index=0 out of size (((Vector<T> *)(this))->_cowdata.size()=0) > 03-12 17:30:47.277 32343 32343 I AEE/AED : #00 pc 000000000164c5d8 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #01 pc 00000000010946ac /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #02 pc 0000000000c1270c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #03 pc 0000000000c12854 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #04 pc 000000000034c1e4 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #05 pc 00000000016ce8b8 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #06 pc 0000000001781c80 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #07 pc 000000000028d6e0 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #08 pc 000000000024a738 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #09 pc 00000000008b58fc /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #10 pc 0000000000bdd92c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #11 pc 0000000000c15654 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #12 pc 00000000016cb16c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #13 pc 00000000008b72dc /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #14 pc 00000000008b726c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #15 pc 00000000008b726c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #16 pc 00000000008bc6a0 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #17 pc 00000000008e97b8 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #18 pc 00000000001a5a84 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so (Java_org_godotengine_godot_GodotLib_step+80) > 03-12 17:30:47.278 32343 32343 I AEE/AED : #19 pc 000000000027d0a8 /data/app/com.example.crashTest-1/oat/arm64/base.odex (offset 0x147000) (void org.godotengine.godot.GodotLib.step()+124) > 03-12 17:30:47.279 32343 32343 I AEE/AED : #20 pc 0000000000281f80 /data/app/com.example.crashTest-1/oat/arm64/base.odex (offset 0x147000) (void org.godotengine.godot.GodotView$Renderer.onDrawFrame(javax.microedition.khronos.opengles.GL10)+84) > 03-12 17:30:47.594 942 32344 W ActivityManager: Force finishing activity com.example.crashTest/org.godotengine.godot.Godot > 03-12 17:30:47.649 942 959 V WindowManager: Changing focus from Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} to Window{3be8bb7 u0 Application Error: com.example.crashTest} Callers=com.android.server.wm.WindowManagerService.addWindow:2821 com.android.server.wm.Session.addToDisplay:171 android.view.ViewRootImpl.setView:647 android.view.WindowManagerGlobal.addView:319 > 03-12 17:30:47.649 942 963 I WindowManager: Focus moving from Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} to Window{3be8bb7 u0 Application Error: com.example.crashTest} > 03-12 17:30:47.703 942 963 I WindowManager: Losing delayed focus: Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} > 03-12 17:30:47.714 268 970 I BufferQueueProducer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:268) disconnect(P): api 1 > 03-12 17:30:47.715 268 970 I BufferQueueConsumer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:268) getReleasedBuffers: returning mask 0xffffffffffffffff > 03-12 17:30:47.717 942 1815 I WindowState: WIN DEATH: Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} > 03-12 17:30:47.777 942 1678 V WindowManager: Removing focused app token:AppWindowToken{edbb8cd token=Token{df0be64 ActivityRecord{97c8ff7 u0 com.example.crashTest/org.godotengine.godot.Godot t148}}} > 03-12 17:30:47.837 268 268 I BufferQueueConsumer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:268) setDefaultBufferSize: width=1080 height=1920 > 03-12 17:30:47.838 942 1678 W WindowState: Failed to report 'resized' to the client of Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot}, removing this window. > 03-12 17:30:47.841 268 384 D SurfaceFlinger: remove: com.example.crashTest/org.godotengine.godot.Godot > 03-12 17:30:47.855 268 268 I BufferQueueConsumer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:-1) disconnect(C) > 03-12 17:30:47.857 268 268 I BufferQueue: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:-1) ~BufferQueueCore > **Steps to reproduce:** - Create a new project - Add a MeshInstance Object to the tree - Execute `create_convex_collision()` or `create_trimesh_collision()` **Minimal reproduction project:** <!-- Recommended as it greatly speeds up debugging. Drag and drop a zip archive to upload it. --> [AndroidCrash.zip](https://github.com/godotengine/godot/files/2957886/AndroidCrash.zip)
1.0
Godot 3.1 RC2 create_trimesh_collision() and create_convex_collision() lead to instant crash on Android - <!-- Please search existing issues for potential duplicates before filing yours: https://github.com/godotengine/godot/issues?q=is%3Aissue --> **Godot version:** 3.1 RC2 **OS/device including version:** Oukitel K6000 Pro, Android 6.0 Samsung Galaxy J5, Android 7 **Issue description:** The methods `create_convex_collision()` and `create_trimesh_collision()` of `MeshInstance` (https://docs.godotengine.org/en/latest/classes/class_meshinstance.html lead to an instant crash on Android. The last version I tested was 3.1 beta 7, there it worked, so it seems it was introduced in some of the last betas. I think this is the relevant error description from logcat: > > 03-12 17:30:47.191 32301 32317 E godot : **ERROR**: OpenGL ES 2.0 does not allow retrieving mesh array data > 03-12 17:30:47.191 32301 32317 E godot : At: drivers/gles2/rasterizer_storage_gles2.cpp:2541:mesh_surface_get_array() - OpenGL ES 2.0 does not allow retrieving mesh array data > 03-12 17:30:47.191 32301 32317 E godot : **ERROR**: Condition ' vertex_data.size() == 0 ' is true. returned: Array() > 03-12 17:30:47.191 32301 32317 E godot : At: servers/visual_server.cpp:1606:mesh_surface_get_arrays() - Condition ' vertex_data.size() == 0 ' is true. returned: Array() > 03-12 17:30:47.191 32301 32317 E godot : **ERROR**: FATAL: Index p_index=0 out of size (((Vector<T> *)(this))->_cowdata.size()=0) > 03-12 17:30:47.191 32301 32317 E godot : At: ./core/vector.h:49:operator[]() - FATAL: Index p_index=0 out of size (((Vector<T> *)(this))->_cowdata.size()=0) > 03-12 17:30:47.277 32343 32343 I AEE/AED : #00 pc 000000000164c5d8 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #01 pc 00000000010946ac /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #02 pc 0000000000c1270c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #03 pc 0000000000c12854 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #04 pc 000000000034c1e4 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #05 pc 00000000016ce8b8 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.277 32343 32343 I AEE/AED : #06 pc 0000000001781c80 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #07 pc 000000000028d6e0 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #08 pc 000000000024a738 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #09 pc 00000000008b58fc /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #10 pc 0000000000bdd92c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #11 pc 0000000000c15654 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #12 pc 00000000016cb16c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #13 pc 00000000008b72dc /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #14 pc 00000000008b726c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #15 pc 00000000008b726c /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #16 pc 00000000008bc6a0 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #17 pc 00000000008e97b8 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so > 03-12 17:30:47.278 32343 32343 I AEE/AED : #18 pc 00000000001a5a84 /data/app/com.example.crashTest-1/lib/arm64/libgodot_android.so (Java_org_godotengine_godot_GodotLib_step+80) > 03-12 17:30:47.278 32343 32343 I AEE/AED : #19 pc 000000000027d0a8 /data/app/com.example.crashTest-1/oat/arm64/base.odex (offset 0x147000) (void org.godotengine.godot.GodotLib.step()+124) > 03-12 17:30:47.279 32343 32343 I AEE/AED : #20 pc 0000000000281f80 /data/app/com.example.crashTest-1/oat/arm64/base.odex (offset 0x147000) (void org.godotengine.godot.GodotView$Renderer.onDrawFrame(javax.microedition.khronos.opengles.GL10)+84) > 03-12 17:30:47.594 942 32344 W ActivityManager: Force finishing activity com.example.crashTest/org.godotengine.godot.Godot > 03-12 17:30:47.649 942 959 V WindowManager: Changing focus from Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} to Window{3be8bb7 u0 Application Error: com.example.crashTest} Callers=com.android.server.wm.WindowManagerService.addWindow:2821 com.android.server.wm.Session.addToDisplay:171 android.view.ViewRootImpl.setView:647 android.view.WindowManagerGlobal.addView:319 > 03-12 17:30:47.649 942 963 I WindowManager: Focus moving from Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} to Window{3be8bb7 u0 Application Error: com.example.crashTest} > 03-12 17:30:47.703 942 963 I WindowManager: Losing delayed focus: Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} > 03-12 17:30:47.714 268 970 I BufferQueueProducer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:268) disconnect(P): api 1 > 03-12 17:30:47.715 268 970 I BufferQueueConsumer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:268) getReleasedBuffers: returning mask 0xffffffffffffffff > 03-12 17:30:47.717 942 1815 I WindowState: WIN DEATH: Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot} > 03-12 17:30:47.777 942 1678 V WindowManager: Removing focused app token:AppWindowToken{edbb8cd token=Token{df0be64 ActivityRecord{97c8ff7 u0 com.example.crashTest/org.godotengine.godot.Godot t148}}} > 03-12 17:30:47.837 268 268 I BufferQueueConsumer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:268) setDefaultBufferSize: width=1080 height=1920 > 03-12 17:30:47.838 942 1678 W WindowState: Failed to report 'resized' to the client of Window{e32c239 u0 com.example.crashTest/org.godotengine.godot.Godot}, removing this window. > 03-12 17:30:47.841 268 384 D SurfaceFlinger: remove: com.example.crashTest/org.godotengine.godot.Godot > 03-12 17:30:47.855 268 268 I BufferQueueConsumer: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:-1) disconnect(C) > 03-12 17:30:47.857 268 268 I BufferQueue: [com.example.crashTest/org.godotengine.godot.Godot](this:0x7f847fb800,id:2482,api:1,p:-1,c:-1) ~BufferQueueCore > **Steps to reproduce:** - Create a new project - Add a MeshInstance Object to the tree - Execute `create_convex_collision()` or `create_trimesh_collision()` **Minimal reproduction project:** <!-- Recommended as it greatly speeds up debugging. Drag and drop a zip archive to upload it. --> [AndroidCrash.zip](https://github.com/godotengine/godot/files/2957886/AndroidCrash.zip)
priority
godot create trimesh collision and create convex collision lead to instant crash on android please search existing issues for potential duplicates before filing yours godot version os device including version oukitel pro android samsung galaxy android issue description the methods create convex collision and create trimesh collision of meshinstance lead to an instant crash on android the last version i tested was beta there it worked so it seems it was introduced in some of the last betas i think this is the relevant error description from logcat e godot error opengl es does not allow retrieving mesh array data e godot at drivers rasterizer storage cpp mesh surface get array opengl es does not allow retrieving mesh array data e godot error condition vertex data size is true returned array e godot at servers visual server cpp mesh surface get arrays condition vertex data size is true returned array e godot error fatal index p index out of size vector this cowdata size e godot at core vector h operator fatal index p index out of size vector this cowdata size i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so i aee aed pc data app com example crashtest lib libgodot android so java org godotengine godot godotlib step i aee aed pc data app com example crashtest oat base odex offset void org godotengine godot godotlib step i aee aed pc data app com example crashtest oat base odex offset void org godotengine godot godotview renderer ondrawframe javax microedition khronos opengles w activitymanager force finishing activity com example crashtest org godotengine godot godot v windowmanager changing focus from window com example crashtest org godotengine godot godot to window application error com example crashtest callers com android server wm windowmanagerservice addwindow com android server wm session addtodisplay android view viewrootimpl setview android view windowmanagerglobal addview i windowmanager focus moving from window com example crashtest org godotengine godot godot to window application error com example crashtest i windowmanager losing delayed focus window com example crashtest org godotengine godot godot i bufferqueueproducer this id api p c disconnect p api i bufferqueueconsumer this id api p c getreleasedbuffers returning mask i windowstate win death window com example crashtest org godotengine godot godot v windowmanager removing focused app token appwindowtoken token token activityrecord com example crashtest org godotengine godot godot i bufferqueueconsumer this id api p c setdefaultbuffersize width height w windowstate failed to report resized to the client of window com example crashtest org godotengine godot godot removing this window d surfaceflinger remove com example crashtest org godotengine godot godot i bufferqueueconsumer this id api p c disconnect c i bufferqueue this id api p c bufferqueuecore steps to reproduce create a new project add a meshinstance object to the tree execute create convex collision or create trimesh collision minimal reproduction project
1
264,672
8,318,215,548
IssuesEvent
2018-09-25 14:10:14
layersoflondon/application
https://api.github.com/repos/layersoflondon/application
opened
Map: Map popover doesn't feel very clickable
High priority
I've seen and heard about users not expecting to click though to the full record.
1.0
Map: Map popover doesn't feel very clickable - I've seen and heard about users not expecting to click though to the full record.
priority
map map popover doesn t feel very clickable i ve seen and heard about users not expecting to click though to the full record
1
288,091
8,824,985,716
IssuesEvent
2019-01-02 19:07:46
spacetelescope/specviz
https://api.github.com/repos/spacetelescope/specviz
closed
specviz crash: model fitting
bug priority-high
If I load a spectrum (e.g. example_stis.fits), select a region, click on add model and select Const1d specviz crashes with the following message ``` Traceback (most recent call last): File "/Users/gderosa/Desktop/specviz/specviz/plugins/model_editor/model_editor.py", line 41, in <lambda> action.triggered.connect(lambda x, m=v: self._add_fittable_model(m)) File "/Users/gderosa/Desktop/specviz/specviz/plugins/model_editor/model_editor.py", line 88, in _add_fittable_model idx = self.model_tree_view.model().add_model(model()) AttributeError: 'NoneType' object has no attribute 'add_model' Abort trap: 6 ```
1.0
specviz crash: model fitting - If I load a spectrum (e.g. example_stis.fits), select a region, click on add model and select Const1d specviz crashes with the following message ``` Traceback (most recent call last): File "/Users/gderosa/Desktop/specviz/specviz/plugins/model_editor/model_editor.py", line 41, in <lambda> action.triggered.connect(lambda x, m=v: self._add_fittable_model(m)) File "/Users/gderosa/Desktop/specviz/specviz/plugins/model_editor/model_editor.py", line 88, in _add_fittable_model idx = self.model_tree_view.model().add_model(model()) AttributeError: 'NoneType' object has no attribute 'add_model' Abort trap: 6 ```
priority
specviz crash model fitting if i load a spectrum e g example stis fits select a region click on add model and select specviz crashes with the following message traceback most recent call last file users gderosa desktop specviz specviz plugins model editor model editor py line in action triggered connect lambda x m v self add fittable model m file users gderosa desktop specviz specviz plugins model editor model editor py line in add fittable model idx self model tree view model add model model attributeerror nonetype object has no attribute add model abort trap
1
445,166
12,827,115,688
IssuesEvent
2020-07-06 17:50:41
unitymakesus/ednc-2020
https://api.github.com/repos/unitymakesus/ednc-2020
closed
Articles only showing up for authors on localhost??
Priority: High Type: Bug
Local Test: https://ednc.test/author/mrash/ ![image](https://user-images.githubusercontent.com/928176/86299291-e30a1500-bbcd-11ea-912a-6608ba01fa85.png) Localhost (`yarn start`): https://localhost:3000/author/mrash/ ![image](https://user-images.githubusercontent.com/928176/86299319-f2895e00-bbcd-11ea-87c9-9b80041b872b.png)
1.0
Articles only showing up for authors on localhost?? - Local Test: https://ednc.test/author/mrash/ ![image](https://user-images.githubusercontent.com/928176/86299291-e30a1500-bbcd-11ea-912a-6608ba01fa85.png) Localhost (`yarn start`): https://localhost:3000/author/mrash/ ![image](https://user-images.githubusercontent.com/928176/86299319-f2895e00-bbcd-11ea-87c9-9b80041b872b.png)
priority
articles only showing up for authors on localhost local test localhost yarn start
1
553,003
16,332,804,168
IssuesEvent
2021-05-12 11:23:21
lutraconsulting/input
https://api.github.com/repos/lutraconsulting/input
closed
Refactor form-related models
enhancement forms high priority
There are several models linked together making things quite complex - and buggy in some more advanced scenarios, e.g. when using conditional visibility. It would be good to simplify the whole approach - maybe something like this: - do not link models together - e.g. the dreaded QgsQuickSubModel - use just simple list models (not hierarchical) - have one central controller class for all the form logic to avoid spaghetti of signals - remove QgsQuickAttributeModel if possible - if we don't need that item model then let's keep the logic in the central controller ![image](https://user-images.githubusercontent.com/193367/100242571-4148e300-2f35-11eb-90b9-9745f1eb8f18.png)
1.0
Refactor form-related models - There are several models linked together making things quite complex - and buggy in some more advanced scenarios, e.g. when using conditional visibility. It would be good to simplify the whole approach - maybe something like this: - do not link models together - e.g. the dreaded QgsQuickSubModel - use just simple list models (not hierarchical) - have one central controller class for all the form logic to avoid spaghetti of signals - remove QgsQuickAttributeModel if possible - if we don't need that item model then let's keep the logic in the central controller ![image](https://user-images.githubusercontent.com/193367/100242571-4148e300-2f35-11eb-90b9-9745f1eb8f18.png)
priority
refactor form related models there are several models linked together making things quite complex and buggy in some more advanced scenarios e g when using conditional visibility it would be good to simplify the whole approach maybe something like this do not link models together e g the dreaded qgsquicksubmodel use just simple list models not hierarchical have one central controller class for all the form logic to avoid spaghetti of signals remove qgsquickattributemodel if possible if we don t need that item model then let s keep the logic in the central controller
1
559,239
16,553,432,079
IssuesEvent
2021-05-28 11:15:33
ChainSafe/chainbridge-utils
https://api.github.com/repos/ChainSafe/chainbridge-utils
closed
Update deprecated package "golang.org/x/crypto/ssh/terminal" --> "golang.org/x/term"
Priority: 2 - High bug
The package `golang.org/x/crypto/ssh/terminal` has been deprecated and should be replaced by `golang.org/x/term`. The deprecation of this package **may** be linked to an [issue](https://discord.com/channels/593655374469660673/713076180374519859/846056268489424896) experienced by a Discord user (XanMan) whereby he encounters an error when attempting to input the password for his encrypted Keystore: ```bash invalid input: The handle is invalid. Enter password to encrypt keystore file: ``` ## Expected Behavior Program should accept command line input without error ## Current Behavior **Example command:** `chainbridge accounts generate` - This successfully creates the `keys` directory within ChainBridge repository root, then fails upon prompting the user for password (input). Program is throwing error: `The handle is invalid` which is originating from the `golang.org/x/crypto/ssh/terminal` package and propagated by our method `keystore.GetPassword` as shown [here](https://github.com/ChainSafe/chainbridge-utils/blob/2aba2e18b4cb636b58c353d7d9cd4820393d49bb/keystore/encrypt.go#L112). ## Possible Solution Update package `golang.org/x/crypto/ssh/terminal` to `golang.org/x/term`. ## Steps to Reproduce (for bugs) ***Use windows machine** 1. `chainbridge accounts generate` 2. attempt inputting keystore password ## Related [Gossamer-1599](https://github.com/ChainSafe/gossamer/issues/1599) ## Versions ChainBridge commit (or docker tag): `v1.1.1` chainbridge-solidity version: chainbridge-substrate version: Go version: `1.16.4`
1.0
Update deprecated package "golang.org/x/crypto/ssh/terminal" --> "golang.org/x/term" - The package `golang.org/x/crypto/ssh/terminal` has been deprecated and should be replaced by `golang.org/x/term`. The deprecation of this package **may** be linked to an [issue](https://discord.com/channels/593655374469660673/713076180374519859/846056268489424896) experienced by a Discord user (XanMan) whereby he encounters an error when attempting to input the password for his encrypted Keystore: ```bash invalid input: The handle is invalid. Enter password to encrypt keystore file: ``` ## Expected Behavior Program should accept command line input without error ## Current Behavior **Example command:** `chainbridge accounts generate` - This successfully creates the `keys` directory within ChainBridge repository root, then fails upon prompting the user for password (input). Program is throwing error: `The handle is invalid` which is originating from the `golang.org/x/crypto/ssh/terminal` package and propagated by our method `keystore.GetPassword` as shown [here](https://github.com/ChainSafe/chainbridge-utils/blob/2aba2e18b4cb636b58c353d7d9cd4820393d49bb/keystore/encrypt.go#L112). ## Possible Solution Update package `golang.org/x/crypto/ssh/terminal` to `golang.org/x/term`. ## Steps to Reproduce (for bugs) ***Use windows machine** 1. `chainbridge accounts generate` 2. attempt inputting keystore password ## Related [Gossamer-1599](https://github.com/ChainSafe/gossamer/issues/1599) ## Versions ChainBridge commit (or docker tag): `v1.1.1` chainbridge-solidity version: chainbridge-substrate version: Go version: `1.16.4`
priority
update deprecated package golang org x crypto ssh terminal golang org x term the package golang org x crypto ssh terminal has been deprecated and should be replaced by golang org x term the deprecation of this package may be linked to an experienced by a discord user xanman whereby he encounters an error when attempting to input the password for his encrypted keystore bash invalid input the handle is invalid enter password to encrypt keystore file expected behavior program should accept command line input without error current behavior example command chainbridge accounts generate this successfully creates the keys directory within chainbridge repository root then fails upon prompting the user for password input program is throwing error the handle is invalid which is originating from the golang org x crypto ssh terminal package and propagated by our method keystore getpassword as shown possible solution update package golang org x crypto ssh terminal to golang org x term steps to reproduce for bugs use windows machine chainbridge accounts generate attempt inputting keystore password related versions chainbridge commit or docker tag chainbridge solidity version chainbridge substrate version go version
1
468,158
13,462,412,482
IssuesEvent
2020-09-09 16:02:15
AY2021S1-CS2103-W14/tp
https://api.github.com/repos/AY2021S1-CS2103-W14/tp
opened
As a new/forgetful user, I want to access the command list/user guide
priority.High type.Story
... easily refer to instructions for commands and usage instructions
1.0
As a new/forgetful user, I want to access the command list/user guide - ... easily refer to instructions for commands and usage instructions
priority
as a new forgetful user i want to access the command list user guide easily refer to instructions for commands and usage instructions
1
205,916
7,107,248,082
IssuesEvent
2018-01-16 19:18:25
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Placing blocks against a wall makes a second rock appear, and the placed block cannot be picked up.
High Priority
Something to do with prediction allowing placement on walls, but the server not.
1.0
Placing blocks against a wall makes a second rock appear, and the placed block cannot be picked up. - Something to do with prediction allowing placement on walls, but the server not.
priority
placing blocks against a wall makes a second rock appear and the placed block cannot be picked up something to do with prediction allowing placement on walls but the server not
1
774,248
27,189,333,233
IssuesEvent
2023-02-19 16:01:46
AY2223S2-CS2103T-T09-4/tp
https://api.github.com/repos/AY2223S2-CS2103T-T09-4/tp
opened
Add FAQ and command summary to the User Guide for v1.1
type.Story priority.High guide.User
As a user, I can see the FAQ and command summary in the User Guide.
1.0
Add FAQ and command summary to the User Guide for v1.1 - As a user, I can see the FAQ and command summary in the User Guide.
priority
add faq and command summary to the user guide for as a user i can see the faq and command summary in the user guide
1
270,103
8,452,349,577
IssuesEvent
2018-10-20 02:38:07
Sage-Bionetworks/Agora
https://api.github.com/repos/Sage-Bionetworks/Agora
closed
Update Background Images
high priority
When I was exporting the smaller images I noticed so inconsistencies so here's the updated files for all of the backgrounds, let me know if you need another format etc. You can also find these in the Images page of Figma. [UpdatedImages.zip](https://github.com/Sage-Bionetworks/Agora/files/2493057/UpdatedImages.zip)
1.0
Update Background Images - When I was exporting the smaller images I noticed so inconsistencies so here's the updated files for all of the backgrounds, let me know if you need another format etc. You can also find these in the Images page of Figma. [UpdatedImages.zip](https://github.com/Sage-Bionetworks/Agora/files/2493057/UpdatedImages.zip)
priority
update background images when i was exporting the smaller images i noticed so inconsistencies so here s the updated files for all of the backgrounds let me know if you need another format etc you can also find these in the images page of figma
1
394,764
11,648,464,001
IssuesEvent
2020-03-01 20:52:03
openmsupply/mobile
https://api.github.com/repos/openmsupply/mobile
closed
Two decimal places showing
Bug: development Docs: not needed Effort: small Module: dispensary Priority: high
## Describe the bug When paying for a prescription, two decimal places show instead of one in the payment amount ### To reproduce 1. Go to pay for a prescription 4. See error ### Expected behaviour Should only show one ### Proposed Solution Show one ### Version and device info N/A ### Additional context nN/A
1.0
Two decimal places showing - ## Describe the bug When paying for a prescription, two decimal places show instead of one in the payment amount ### To reproduce 1. Go to pay for a prescription 4. See error ### Expected behaviour Should only show one ### Proposed Solution Show one ### Version and device info N/A ### Additional context nN/A
priority
two decimal places showing describe the bug when paying for a prescription two decimal places show instead of one in the payment amount to reproduce go to pay for a prescription see error expected behaviour should only show one proposed solution show one version and device info n a additional context nn a
1
424,902
12,324,780,283
IssuesEvent
2020-05-13 14:11:41
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
closed
Can not add WMS background in WGS84
Priority: High bug
## Description When you have a map in WGS84, adding a background shows an error: ![image](https://user-images.githubusercontent.com/1279510/81160586-8baf4600-8f8a-11ea-8710-e8a522d62282.png) It works with CSW or WMTS catalogs. ## How to reproduce - Login as admin (to show crs selector) - create a new map and switch to WGS84 - Try to add a new background from WMS service (tested with https://gs-stable.geo-solutions.it/geoserver/wms) *Expected Result* The layer is added if the CRS is supported *Current Result* An error tells the CRS is not compatible - [x] Not browser related <details><summary> <b>Browser info</b> </summary> <!-- If browser related, please compile the following table --> <!-- If your browser is not in the list please add a new row to the table with the version --> (use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users) | Browser Affected | Version | |---|---| |Internet Explorer| | |Edge| | |Chrome| | |Firefox| | |Safari| | </details>
1.0
Can not add WMS background in WGS84 - ## Description When you have a map in WGS84, adding a background shows an error: ![image](https://user-images.githubusercontent.com/1279510/81160586-8baf4600-8f8a-11ea-8710-e8a522d62282.png) It works with CSW or WMTS catalogs. ## How to reproduce - Login as admin (to show crs selector) - create a new map and switch to WGS84 - Try to add a new background from WMS service (tested with https://gs-stable.geo-solutions.it/geoserver/wms) *Expected Result* The layer is added if the CRS is supported *Current Result* An error tells the CRS is not compatible - [x] Not browser related <details><summary> <b>Browser info</b> </summary> <!-- If browser related, please compile the following table --> <!-- If your browser is not in the list please add a new row to the table with the version --> (use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users) | Browser Affected | Version | |---|---| |Internet Explorer| | |Edge| | |Chrome| | |Firefox| | |Safari| | </details>
priority
can not add wms background in description when you have a map in adding a background shows an error it works with csw or wmts catalogs how to reproduce login as admin to show crs selector create a new map and switch to try to add a new background from wms service tested with expected result the layer is added if the crs is supported current result an error tells the crs is not compatible not browser related browser info use this site a href for non expert users browser affected version internet explorer edge chrome firefox safari
1
127,694
5,038,507,733
IssuesEvent
2016-12-18 09:23:30
fossasia/asksusi.com
https://api.github.com/repos/fossasia/asksusi.com
closed
Responsive UI for asksusi
bug enhancement High Priority
Currently the css is not mobile compatible. ![screenshot_2016-08-21-10-17-34_com android chrome](https://cloud.githubusercontent.com/assets/6227784/17835584/eff20646-6790-11e6-832f-24a93cfce11a.png) The css needs to be tested on various devices and make use of bootstrap.
1.0
Responsive UI for asksusi - Currently the css is not mobile compatible. ![screenshot_2016-08-21-10-17-34_com android chrome](https://cloud.githubusercontent.com/assets/6227784/17835584/eff20646-6790-11e6-832f-24a93cfce11a.png) The css needs to be tested on various devices and make use of bootstrap.
priority
responsive ui for asksusi currently the css is not mobile compatible the css needs to be tested on various devices and make use of bootstrap
1
739,254
25,587,753,662
IssuesEvent
2022-12-01 10:36:32
nf-core/tools
https://api.github.com/repos/nf-core/tools
opened
Module lint: Check container syntax
linting high-priority
### Description of feature x-ref https://github.com/nf-core/tools/issues/1627 Check that the `container` syntax is correct and can be parsed properly. eg. single quotes vs double quotes.
1.0
Module lint: Check container syntax - ### Description of feature x-ref https://github.com/nf-core/tools/issues/1627 Check that the `container` syntax is correct and can be parsed properly. eg. single quotes vs double quotes.
priority
module lint check container syntax description of feature x ref check that the container syntax is correct and can be parsed properly eg single quotes vs double quotes
1
685,590
23,461,649,250
IssuesEvent
2022-08-16 13:34:49
kubermatic/dashboard
https://api.github.com/repos/kubermatic/dashboard
closed
Remove the warning for delete AKS/GKE external cluster dialog
kind/bug priority/high sig/ui externalcluster
### What happened Remove the warning fro AKS/GKE externalcluster delete: - [ ] remove the warning sign to information sign - [ ] keep the first line informing user of attached mds - [ ] remove the line `Please delete nodegroups....` ** No changes required for EKS ![delete](https://user-images.githubusercontent.com/78943096/184630753-49714229-ed09-4ae4-8756-9e9cc362ddd4.png) ### Expected behavior ### How to reproduce ### Environment - UI Version: - API Version: - Domain: - Others: ### Current workaround ### Affected user persona ### Business goal to be improved ### Metric to be improved
1.0
Remove the warning for delete AKS/GKE external cluster dialog - ### What happened Remove the warning fro AKS/GKE externalcluster delete: - [ ] remove the warning sign to information sign - [ ] keep the first line informing user of attached mds - [ ] remove the line `Please delete nodegroups....` ** No changes required for EKS ![delete](https://user-images.githubusercontent.com/78943096/184630753-49714229-ed09-4ae4-8756-9e9cc362ddd4.png) ### Expected behavior ### How to reproduce ### Environment - UI Version: - API Version: - Domain: - Others: ### Current workaround ### Affected user persona ### Business goal to be improved ### Metric to be improved
priority
remove the warning for delete aks gke external cluster dialog what happened remove the warning fro aks gke externalcluster delete remove the warning sign to information sign keep the first line informing user of attached mds remove the line please delete nodegroups no changes required for eks expected behavior how to reproduce environment ui version api version domain others current workaround affected user persona business goal to be improved metric to be improved
1
590,690
17,784,728,542
IssuesEvent
2021-08-31 09:39:28
ita-social-projects/OoS-Backend
https://api.github.com/repos/ita-social-projects/OoS-Backend
opened
[SP Class] Drop-down box with data is missing on the 'Новий гурток' create page in 'Опис' tab in 'Напрямок за яким здійснюється навчання' field
bug priority:high sev:minor Type:Functional
**Environment:** Windows 10 Home, Google Chrome 92.0.4515.159. **Reproducible:** always. **Build found:** 31/08/2021 12:26 Preconditions A user (Service Provider) has been registered already in the system. (e.g.Login/pass=7gakor46280@asmm5.com/Qwer1234?) **Steps to reproduce** 1. Loggin as Service Provider. 2. Click on the drop-down arrow to the right of the 'User Name' -> 'Мої гуртки'. 3. Click on the 'Додати гурток' button. 4. Fill mandatory fields on the 'Про гурток' tab -> click on 'Опис' tab. 5. Take a look at 'Напрямок за яким здійснюється навчання' field. **Actual result** The drop-down box with data is missing in 'Напрямок за яким здійснюється навчання' field. ![SP_class_DDBox_mis](https://user-images.githubusercontent.com/88823354/131479386-0fe13414-c41e-4c3d-a782-4898012ea2af.png) **Expected result** The drop-down box with data is present in 'Напрямок за яким здійснюється навчання' field. ![SP_class_DDBox_mis_figma](https://user-images.githubusercontent.com/88823354/131479566-7bbc3ddc-22b1-4e44-931c-5e4c795bcabd.png) **User story and test case links** ([#55 Story](https://github.com/ita-social-projects/OoS-Backend/issues/55))" **Labels to be added** "Bug", Priority ("pri: "), Severity ("severity:"), Type ("UI, "Functional"), "API" (for back-end bugs).
1.0
[SP Class] Drop-down box with data is missing on the 'Новий гурток' create page in 'Опис' tab in 'Напрямок за яким здійснюється навчання' field - **Environment:** Windows 10 Home, Google Chrome 92.0.4515.159. **Reproducible:** always. **Build found:** 31/08/2021 12:26 Preconditions A user (Service Provider) has been registered already in the system. (e.g.Login/pass=7gakor46280@asmm5.com/Qwer1234?) **Steps to reproduce** 1. Loggin as Service Provider. 2. Click on the drop-down arrow to the right of the 'User Name' -> 'Мої гуртки'. 3. Click on the 'Додати гурток' button. 4. Fill mandatory fields on the 'Про гурток' tab -> click on 'Опис' tab. 5. Take a look at 'Напрямок за яким здійснюється навчання' field. **Actual result** The drop-down box with data is missing in 'Напрямок за яким здійснюється навчання' field. ![SP_class_DDBox_mis](https://user-images.githubusercontent.com/88823354/131479386-0fe13414-c41e-4c3d-a782-4898012ea2af.png) **Expected result** The drop-down box with data is present in 'Напрямок за яким здійснюється навчання' field. ![SP_class_DDBox_mis_figma](https://user-images.githubusercontent.com/88823354/131479566-7bbc3ddc-22b1-4e44-931c-5e4c795bcabd.png) **User story and test case links** ([#55 Story](https://github.com/ita-social-projects/OoS-Backend/issues/55))" **Labels to be added** "Bug", Priority ("pri: "), Severity ("severity:"), Type ("UI, "Functional"), "API" (for back-end bugs).
priority
drop down box with data is missing on the новий гурток create page in опис tab in напрямок за яким здійснюється навчання field environment windows home google chrome reproducible always build found preconditions a user service provider has been registered already in the system e g login pass com steps to reproduce loggin as service provider click on the drop down arrow to the right of the user name мої гуртки click on the додати гурток button fill mandatory fields on the про гурток tab click on опис tab take a look at напрямок за яким здійснюється навчання field actual result the drop down box with data is missing in напрямок за яким здійснюється навчання field expected result the drop down box with data is present in напрямок за яким здійснюється навчання field user story and test case links labels to be added bug priority pri severity severity type ui functional api for back end bugs
1
783,404
27,528,948,912
IssuesEvent
2023-03-06 20:22:15
chirila/FileLingR
https://api.github.com/repos/chirila/FileLingR
opened
return names of duplicate files
high priority New feature
per Daisy Rosenblum's comment about transcripts from multiple sources, version control
1.0
return names of duplicate files - per Daisy Rosenblum's comment about transcripts from multiple sources, version control
priority
return names of duplicate files per daisy rosenblum s comment about transcripts from multiple sources version control
1
345,048
10,352,127,087
IssuesEvent
2019-09-05 08:35:26
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
opened
Impossible to eliminate layer groups
Priority: High bug
### Description There's no button that allow to eliminate layer groups in TOC ### In case of Bug (otherwise remove this paragraph) *Browser Affected* (use this site: https://www.whatsmybrowser.org/ for non expert users) - [X] Internet Explorer - [X] Chrome - [X] Firefox - [ ] Safari *Browser Version Affected* - Not browser affected *Steps to reproduce* - Open Mapstore - Open a Tial Map - Open TOC - Add a New Group - Select the group *Expected Result* - Ther's the buttom that allow you to delete the group *Current Result* - Ther's not the buttom that allow you to delete the group ### Other useful information (optional): If a group has a layer, selecting that group there's the buttom that allows you to delete his layer, but is still missing the possibility to eliminate the group
1.0
Impossible to eliminate layer groups - ### Description There's no button that allow to eliminate layer groups in TOC ### In case of Bug (otherwise remove this paragraph) *Browser Affected* (use this site: https://www.whatsmybrowser.org/ for non expert users) - [X] Internet Explorer - [X] Chrome - [X] Firefox - [ ] Safari *Browser Version Affected* - Not browser affected *Steps to reproduce* - Open Mapstore - Open a Tial Map - Open TOC - Add a New Group - Select the group *Expected Result* - Ther's the buttom that allow you to delete the group *Current Result* - Ther's not the buttom that allow you to delete the group ### Other useful information (optional): If a group has a layer, selecting that group there's the buttom that allows you to delete his layer, but is still missing the possibility to eliminate the group
priority
impossible to eliminate layer groups description there s no button that allow to eliminate layer groups in toc in case of bug otherwise remove this paragraph browser affected use this site for non expert users internet explorer chrome firefox safari browser version affected not browser affected steps to reproduce open mapstore open a tial map open toc add a new group select the group expected result ther s the buttom that allow you to delete the group current result ther s not the buttom that allow you to delete the group other useful information optional if a group has a layer selecting that group there s the buttom that allows you to delete his layer but is still missing the possibility to eliminate the group
1
240,749
7,805,428,374
IssuesEvent
2018-06-11 10:48:22
erezam/DNA-PRIMER-BUILDER
https://api.github.com/repos/erezam/DNA-PRIMER-BUILDER
closed
implementation a method to calculate accurate temperature
AMIR In progress OR priority 1(higher)
we will use linear regression with vector of letter pairs
1.0
implementation a method to calculate accurate temperature - we will use linear regression with vector of letter pairs
priority
implementation a method to calculate accurate temperature we will use linear regression with vector of letter pairs
1
265,719
8,357,899,389
IssuesEvent
2018-10-02 23:34:51
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Possible duplicated loc. strings
Fixed High Priority Localization
https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#49772 > A towering metal lightpost that requires electricity to run. https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#52093 > A towering metal light post that requires electricity to run. @Nanders
1.0
Possible duplicated loc. strings - https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#49772 > A towering metal lightpost that requires electricity to run. https://crowdin.com/translate/eco-by-strange-loop-games/24/en-ru#52093 > A towering metal light post that requires electricity to run. @Nanders
priority
possible duplicated loc strings a towering metal lightpost that requires electricity to run a towering metal light post that requires electricity to run nanders
1
286,901
8,795,665,752
IssuesEvent
2018-12-22 18:32:52
nextcloud/user_sql
https://api.github.com/repos/nextcloud/user_sql
closed
Revers active column checkbox
enhancement high priority
No problem, A lot of users(like me) ask for features but we use this software for free and you make it in your own spare time. So thank you But to let my instruction work one small change is necessary; a reverse option, Joomla works with a block column 1= blocked option instead of 1=active. _Originally posted by @fsdfasf in https://github.com/nextcloud/user_sql/issues/84#issuecomment-449557902_
1.0
Revers active column checkbox - No problem, A lot of users(like me) ask for features but we use this software for free and you make it in your own spare time. So thank you But to let my instruction work one small change is necessary; a reverse option, Joomla works with a block column 1= blocked option instead of 1=active. _Originally posted by @fsdfasf in https://github.com/nextcloud/user_sql/issues/84#issuecomment-449557902_
priority
revers active column checkbox no problem a lot of users like me ask for features but we use this software for free and you make it in your own spare time so thank you but to let my instruction work one small change is necessary a reverse option joomla works with a block column blocked option instead of active originally posted by fsdfasf in
1
496,153
14,333,809,917
IssuesEvent
2020-11-27 06:46:57
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
xhamster.com - site is not usable
browser-fenix engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/62588 --> <!-- @extra_labels: browser-fenix --> **URL**: https://xhamster.com **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201117192430</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/11/481811e5-8b4b-486a-b643-38610e8ee00f) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
xhamster.com - site is not usable - <!-- @browser: Firefox Mobile 84.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:84.0) Gecko/84.0 Firefox/84.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/62588 --> <!-- @extra_labels: browser-fenix --> **URL**: https://xhamster.com **Browser / Version**: Firefox Mobile 84.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Buttons or links not working **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201117192430</li><li>channel: beta</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/11/481811e5-8b4b-486a-b643-38610e8ee00f) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
xhamster com site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description buttons or links not working steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
230,882
7,620,838,749
IssuesEvent
2018-05-03 05:45:13
steedos/creator
https://api.github.com/repos/steedos/creator
closed
列表查询界面开发
fix:Done priority:High
- [ ] "查询“属于全局的actions,类似于“新建”,所有按钮放在“新建”按钮旁边。 - [ ] 不同的对象得到的查询界面不一样,查询界面上可查询的字段,根据字段search_able属性决定 - [ ] 查询样式:如下图 ![_1524558401936](https://user-images.githubusercontent.com/27475169/39175246-62fd94d4-47dc-11e8-8975-65d2be53af03.png)
1.0
列表查询界面开发 - - [ ] "查询“属于全局的actions,类似于“新建”,所有按钮放在“新建”按钮旁边。 - [ ] 不同的对象得到的查询界面不一样,查询界面上可查询的字段,根据字段search_able属性决定 - [ ] 查询样式:如下图 ![_1524558401936](https://user-images.githubusercontent.com/27475169/39175246-62fd94d4-47dc-11e8-8975-65d2be53af03.png)
priority
列表查询界面开发 查询“属于全局的actions 类似于“新建”,所有按钮放在“新建”按钮旁边。 不同的对象得到的查询界面不一样,查询界面上可查询的字段,根据字段search able属性决定 查询样式:如下图
1
123,557
4,864,855,606
IssuesEvent
2016-11-14 19:11:14
BinPar/PPD
https://api.github.com/repos/BinPar/PPD
opened
FECHA PREVISTA PAIS IMPRESION NO APARECE EN FILIAL
Priority: High
![image](https://cloud.githubusercontent.com/assets/22589031/20278704/0213d676-aaa6-11e6-94a0-e3fa545b3e79.png) esto es lo que yo veo con FILIAL ESPAÑA (Producto libro MEDICINA DE URGENCIAS COPY MEX) ![image](https://cloud.githubusercontent.com/assets/22589031/20278753/29246e1a-aaa6-11e6-86d1-9ea63c5932a3.png) Esto es lo que aparece en la pantalla de COPY MEX. La fecha PREVISTA PAIS DE IMPRESIÓN tiene que aparecer en la información de todas las filiales, ya que quien la indica es el pais COPY. Las filiales no puede introducir esa fecha @CristianBinpar
1.0
FECHA PREVISTA PAIS IMPRESION NO APARECE EN FILIAL - ![image](https://cloud.githubusercontent.com/assets/22589031/20278704/0213d676-aaa6-11e6-94a0-e3fa545b3e79.png) esto es lo que yo veo con FILIAL ESPAÑA (Producto libro MEDICINA DE URGENCIAS COPY MEX) ![image](https://cloud.githubusercontent.com/assets/22589031/20278753/29246e1a-aaa6-11e6-86d1-9ea63c5932a3.png) Esto es lo que aparece en la pantalla de COPY MEX. La fecha PREVISTA PAIS DE IMPRESIÓN tiene que aparecer en la información de todas las filiales, ya que quien la indica es el pais COPY. Las filiales no puede introducir esa fecha @CristianBinpar
priority
fecha prevista pais impresion no aparece en filial esto es lo que yo veo con filial españa producto libro medicina de urgencias copy mex esto es lo que aparece en la pantalla de copy mex la fecha prevista pais de impresión tiene que aparecer en la información de todas las filiales ya que quien la indica es el pais copy las filiales no puede introducir esa fecha cristianbinpar
1
682,438
23,345,280,103
IssuesEvent
2022-08-09 17:20:42
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
opened
Remove new/assignment-approve statuses from the list of ACTIVE status
High Priority WMStats ReqMgr2 ReqMgr2MS Feature change
**Impact of the new feature** ReqMgr2/WMStats and other services consuming this data **Is your feature request related to a problem? Please describe.** We have been discussing on how to alleviate the load on WMStats - and anything/anyone consuming ACTIVE data - and one of the ideas is to reduce the number of workflows that are constantly monitored by WMStats (and cached by DataCache WMStatsServer application). That could be done by removing `new` and `assignment-approved` from the list of ACTIVE status defined here: https://github.com/dmwm/WMCore/blob/master/src/python/WMCore/ReqMgr/DataStructs/RequestStatus.py#L71 however, this will impact the data that is served by `wmstatsserver` app and the content provided, e.g. the `protectedlfns` and `globallocks`, used in the microservices for some dangerous and sensitive actions. **Describe the solution you'd like** Before we can shorten this list of ACTIVE status, we need to: * investigate what the impact is on the system * investigate if it could cause problems for those consuming the list of protected LFNs (output) * investigate if it could cause problems for those consuming the global locks (input + output) if no negative side effects are expected, we should then stop monitoring workflows in status `new` or `assignment-approvement` in WMStats. This will also affect those querying ReqMgr2 with this query string `status=ACTIVE`. **Describe alternatives you've considered** At the current load (30k requests), the system is still behaving properly, but we could see a degradation at 40k or so. So we need to find ways to reduce the footprint on WMStats. **Additional context** Likely dependent on: https://github.com/dmwm/WMCore/issues/11241 and https://github.com/dmwm/WMCore/issues/11243
1.0
Remove new/assignment-approve statuses from the list of ACTIVE status - **Impact of the new feature** ReqMgr2/WMStats and other services consuming this data **Is your feature request related to a problem? Please describe.** We have been discussing on how to alleviate the load on WMStats - and anything/anyone consuming ACTIVE data - and one of the ideas is to reduce the number of workflows that are constantly monitored by WMStats (and cached by DataCache WMStatsServer application). That could be done by removing `new` and `assignment-approved` from the list of ACTIVE status defined here: https://github.com/dmwm/WMCore/blob/master/src/python/WMCore/ReqMgr/DataStructs/RequestStatus.py#L71 however, this will impact the data that is served by `wmstatsserver` app and the content provided, e.g. the `protectedlfns` and `globallocks`, used in the microservices for some dangerous and sensitive actions. **Describe the solution you'd like** Before we can shorten this list of ACTIVE status, we need to: * investigate what the impact is on the system * investigate if it could cause problems for those consuming the list of protected LFNs (output) * investigate if it could cause problems for those consuming the global locks (input + output) if no negative side effects are expected, we should then stop monitoring workflows in status `new` or `assignment-approvement` in WMStats. This will also affect those querying ReqMgr2 with this query string `status=ACTIVE`. **Describe alternatives you've considered** At the current load (30k requests), the system is still behaving properly, but we could see a degradation at 40k or so. So we need to find ways to reduce the footprint on WMStats. **Additional context** Likely dependent on: https://github.com/dmwm/WMCore/issues/11241 and https://github.com/dmwm/WMCore/issues/11243
priority
remove new assignment approve statuses from the list of active status impact of the new feature wmstats and other services consuming this data is your feature request related to a problem please describe we have been discussing on how to alleviate the load on wmstats and anything anyone consuming active data and one of the ideas is to reduce the number of workflows that are constantly monitored by wmstats and cached by datacache wmstatsserver application that could be done by removing new and assignment approved from the list of active status defined here however this will impact the data that is served by wmstatsserver app and the content provided e g the protectedlfns and globallocks used in the microservices for some dangerous and sensitive actions describe the solution you d like before we can shorten this list of active status we need to investigate what the impact is on the system investigate if it could cause problems for those consuming the list of protected lfns output investigate if it could cause problems for those consuming the global locks input output if no negative side effects are expected we should then stop monitoring workflows in status new or assignment approvement in wmstats this will also affect those querying with this query string status active describe alternatives you ve considered at the current load requests the system is still behaving properly but we could see a degradation at or so so we need to find ways to reduce the footprint on wmstats additional context likely dependent on and
1
233,719
7,703,784,114
IssuesEvent
2018-05-21 09:41:52
yldio/joyent-portal
https://api.github.com/repos/yldio/joyent-portal
closed
Bugs from user-testing
bug engineering high priority
- [x] Package table, vCPU is missing - [x] Package table, Storage Optimised icon is incorrect - [x] Tags, delete cross still visible and clickable when the Tags card is collapsed. This should not be the case - [x] User scripts, code block component should only have 5 empty lines when opened - [x] Firewalls, implement checkboxes in the place of the toggles for "Enable firewalls" and "View inactive rules" - [x] CNS, remove the browser default disabled cursor when hovering over the CNS urls - [ ] CNS, 'Network hostnames' not showing despite instance being attached to additional networks - [x] CNS, CNS Service Name tag should go below the input - [x] General, 'Read the docs' link collapses the card - [x] General, On click, anchor state goes grey, consult design systems for correct colours
1.0
Bugs from user-testing - - [x] Package table, vCPU is missing - [x] Package table, Storage Optimised icon is incorrect - [x] Tags, delete cross still visible and clickable when the Tags card is collapsed. This should not be the case - [x] User scripts, code block component should only have 5 empty lines when opened - [x] Firewalls, implement checkboxes in the place of the toggles for "Enable firewalls" and "View inactive rules" - [x] CNS, remove the browser default disabled cursor when hovering over the CNS urls - [ ] CNS, 'Network hostnames' not showing despite instance being attached to additional networks - [x] CNS, CNS Service Name tag should go below the input - [x] General, 'Read the docs' link collapses the card - [x] General, On click, anchor state goes grey, consult design systems for correct colours
priority
bugs from user testing package table vcpu is missing package table storage optimised icon is incorrect tags delete cross still visible and clickable when the tags card is collapsed this should not be the case user scripts code block component should only have empty lines when opened firewalls implement checkboxes in the place of the toggles for enable firewalls and view inactive rules cns remove the browser default disabled cursor when hovering over the cns urls cns network hostnames not showing despite instance being attached to additional networks cns cns service name tag should go below the input general read the docs link collapses the card general on click anchor state goes grey consult design systems for correct colours
1
394,104
11,631,964,119
IssuesEvent
2020-02-28 03:16:21
SesameCrew/sesame_issues
https://api.github.com/repos/SesameCrew/sesame_issues
closed
Contact actions add-on does nothing
bug high priority
I have purchased the contact actions add-on, but nothing is changed... I have multiple shortcuts for the same account (contact, whatsapp, dialer) and only the default 3 shortcuts for contact shortcuts, even if I disable them from settings. Plus -> long pressing the default shortcuts they disappear ![Screenshot_2019-12-22-18-03-58-684_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324930-88d51c80-24e5-11ea-88d6-429fae48f671.png) ![Screenshot_2019-12-22-18-04-20-400_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324931-896db300-24e5-11ea-95a6-aa3f4bf91740.png) ![Screenshot_2019-12-22-18-04-16-405_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324932-896db300-24e5-11ea-86b5-09ba353947c0.png) ![Screenshot_2019-12-22-18-03-44-218_ninja sesame app edge](https://user-images.githubusercontent.com/9931444/71324933-896db300-24e5-11ea-95b5-76f63c7ea8b2.png) ![Screenshot_2019-12-22-18-04-25-146_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324934-8a064980-24e5-11ea-9935-7042cf87c7ce.png)
1.0
Contact actions add-on does nothing - I have purchased the contact actions add-on, but nothing is changed... I have multiple shortcuts for the same account (contact, whatsapp, dialer) and only the default 3 shortcuts for contact shortcuts, even if I disable them from settings. Plus -> long pressing the default shortcuts they disappear ![Screenshot_2019-12-22-18-03-58-684_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324930-88d51c80-24e5-11ea-88d6-429fae48f671.png) ![Screenshot_2019-12-22-18-04-20-400_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324931-896db300-24e5-11ea-95a6-aa3f4bf91740.png) ![Screenshot_2019-12-22-18-04-16-405_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324932-896db300-24e5-11ea-86b5-09ba353947c0.png) ![Screenshot_2019-12-22-18-03-44-218_ninja sesame app edge](https://user-images.githubusercontent.com/9931444/71324933-896db300-24e5-11ea-95b5-76f63c7ea8b2.png) ![Screenshot_2019-12-22-18-04-25-146_com teslacoilsw launcher](https://user-images.githubusercontent.com/9931444/71324934-8a064980-24e5-11ea-9935-7042cf87c7ce.png)
priority
contact actions add on does nothing i have purchased the contact actions add on but nothing is changed i have multiple shortcuts for the same account contact whatsapp dialer and only the default shortcuts for contact shortcuts even if i disable them from settings plus long pressing the default shortcuts they disappear
1
274,319
8,559,651,100
IssuesEvent
2018-11-08 21:56:59
johannbrehmer/madminer
https://api.github.com/repos/johannbrehmer/madminer
closed
Clean up madminer.plotting
cosmetic high priority
The `madminer.plotting` functions `plot_fisherinfo_barplot()` and (to a lesser extent) `kinematic_distribution_of_information()` are messy. Docstrings are missing, there is code that was written for specific processes, some objects are never used at all. Generally the coding style is different from the rest of MadMiner. I think we should either clean up these functions, or remove them from the package (in that case we can move them to the parton-level tutorial notebook). Thoughts?
1.0
Clean up madminer.plotting - The `madminer.plotting` functions `plot_fisherinfo_barplot()` and (to a lesser extent) `kinematic_distribution_of_information()` are messy. Docstrings are missing, there is code that was written for specific processes, some objects are never used at all. Generally the coding style is different from the rest of MadMiner. I think we should either clean up these functions, or remove them from the package (in that case we can move them to the parton-level tutorial notebook). Thoughts?
priority
clean up madminer plotting the madminer plotting functions plot fisherinfo barplot and to a lesser extent kinematic distribution of information are messy docstrings are missing there is code that was written for specific processes some objects are never used at all generally the coding style is different from the rest of madminer i think we should either clean up these functions or remove them from the package in that case we can move them to the parton level tutorial notebook thoughts
1
516,322
14,979,421,011
IssuesEvent
2021-01-28 12:14:23
jesus-collective/mobile
https://api.github.com/repos/jesus-collective/mobile
closed
Since beta update, users face a blank screen
High Priority login
The url ends in /payment3. As an admin, I was able to get around this by logging out and logging back in, but this wouldn't be ideal for other users as they may not know to do that. As a non-admin user, the url ends in /payment1. Payment page flashes quickly and then the user faces blank screen and can't get in.
1.0
Since beta update, users face a blank screen - The url ends in /payment3. As an admin, I was able to get around this by logging out and logging back in, but this wouldn't be ideal for other users as they may not know to do that. As a non-admin user, the url ends in /payment1. Payment page flashes quickly and then the user faces blank screen and can't get in.
priority
since beta update users face a blank screen the url ends in as an admin i was able to get around this by logging out and logging back in but this wouldn t be ideal for other users as they may not know to do that as a non admin user the url ends in payment page flashes quickly and then the user faces blank screen and can t get in
1
133,923
5,217,115,345
IssuesEvent
2017-01-26 12:49:50
ActionFPS/ActionFPS-Game
https://api.github.com/repos/ActionFPS/ActionFPS-Game
opened
Version number unification
high-priority question
Should make version number easy to set and make consistent across all packages https://github.com/ActionFPS/ActionFPS-Game/blob/master/source/release/package_assaultcube.sh#L136 What's the best way to handle that ?
1.0
Version number unification - Should make version number easy to set and make consistent across all packages https://github.com/ActionFPS/ActionFPS-Game/blob/master/source/release/package_assaultcube.sh#L136 What's the best way to handle that ?
priority
version number unification should make version number easy to set and make consistent across all packages what s the best way to handle that
1
490,920
14,142,104,256
IssuesEvent
2020-11-10 13:38:44
reichlab/covidHubUtils
https://api.github.com/repos/reichlab/covidHubUtils
closed
load_forecasts() should allow for querying forecasts from multiple timezeroes at once
High Priority
right now, load_forecasts() uses a fairly different query structure for forecast_dates/timezeroes than the native zoltr `do_zoltar_query()` does. The difference is that we load_forecasts() specifies a single date and a window, but the native zoltar queries allow for all timezeroes to be queried (timezeroes left empty) or a specific set of timezeroes to be specified. my suggestion is that we should change the specification of load_forecasts() to accept a `forecast_date` argument that can be passed directly to `do_zoltar_query()`.
1.0
load_forecasts() should allow for querying forecasts from multiple timezeroes at once - right now, load_forecasts() uses a fairly different query structure for forecast_dates/timezeroes than the native zoltr `do_zoltar_query()` does. The difference is that we load_forecasts() specifies a single date and a window, but the native zoltar queries allow for all timezeroes to be queried (timezeroes left empty) or a specific set of timezeroes to be specified. my suggestion is that we should change the specification of load_forecasts() to accept a `forecast_date` argument that can be passed directly to `do_zoltar_query()`.
priority
load forecasts should allow for querying forecasts from multiple timezeroes at once right now load forecasts uses a fairly different query structure for forecast dates timezeroes than the native zoltr do zoltar query does the difference is that we load forecasts specifies a single date and a window but the native zoltar queries allow for all timezeroes to be queried timezeroes left empty or a specific set of timezeroes to be specified my suggestion is that we should change the specification of load forecasts to accept a forecast date argument that can be passed directly to do zoltar query
1
328,903
10,001,295,525
IssuesEvent
2019-07-12 15:17:55
Qiskit/qiskit-terra
https://api.github.com/repos/Qiskit/qiskit-terra
closed
There needs to be a ApplyLayout pass
enhancement priority: high
<!-- ⚠️ If you do not respect this template, your issue will be closed --> <!-- ⚠️ Make sure to browse the opened and closed issues to confirm this idea does not exist. --> ### What is the expected enhancement? The default `StochasticSwap` pass does not map a circuit from the logical (virtual) representation to the physical representation for running on hardware. The previous `swap_mapper` pass used to do this. As such, there needs to be an `ApplyLayout` pass that flattens the quantum registers and applies the layout to map to hardware for running on a physical backend.
1.0
There needs to be a ApplyLayout pass - <!-- ⚠️ If you do not respect this template, your issue will be closed --> <!-- ⚠️ Make sure to browse the opened and closed issues to confirm this idea does not exist. --> ### What is the expected enhancement? The default `StochasticSwap` pass does not map a circuit from the logical (virtual) representation to the physical representation for running on hardware. The previous `swap_mapper` pass used to do this. As such, there needs to be an `ApplyLayout` pass that flattens the quantum registers and applies the layout to map to hardware for running on a physical backend.
priority
there needs to be a applylayout pass what is the expected enhancement the default stochasticswap pass does not map a circuit from the logical virtual representation to the physical representation for running on hardware the previous swap mapper pass used to do this as such there needs to be an applylayout pass that flattens the quantum registers and applies the layout to map to hardware for running on a physical backend
1
716,767
24,647,315,482
IssuesEvent
2022-10-17 15:44:24
xHyroM/hykord
https://api.github.com/repos/xHyroM/hykord
closed
Refactor @hykord/structures
@hykord/core priority: high 📕
Make @hykord/hooks with simple `$plugin` and `$theme` function for defining theme/plugin For themes you can use css
1.0
Refactor @hykord/structures - Make @hykord/hooks with simple `$plugin` and `$theme` function for defining theme/plugin For themes you can use css
priority
refactor hykord structures make hykord hooks with simple plugin and theme function for defining theme plugin for themes you can use css
1
274,712
8,564,541,949
IssuesEvent
2018-11-09 17:01:20
prettier/prettier
https://api.github.com/repos/prettier/prettier
closed
HTML: Breaking Angular input with a cast to number
lang:angular priority:high status:has pr type:bug
**Prettier 1.15.1** [Playground link](https://prettier.io/playground/#.....) ```sh "htmlWhitespaceSensitivity": "ignore" ``` **Input:** ```html <div class="tool-svg" fxFlex="10%" fxLayout="center center"> <cnc-tool-assembly-svg [tool]="toolAssembly.tool" [stickout]="+toolAssembly.stickoutMm" ></cnc-tool-assembly-svg> </div> ``` **Output:** ```html <div class="tool-svg" fxFlex="10%" fxLayout="center center"> <cnc-tool-assembly-svg [tool]="toolAssembly.tool" [stickout]="toolAssembly.stickoutMm - "></cnc-tool-assembly-svg> </div> ``` **Problem here:** ``` [stickout]="+toolAssembly.stickoutMm" ``` is updated to: ``` [stickout]="toolAssembly.stickoutMm - " ``` **Expected behavior:** :warning: Behavior shouldn't be changed :warning:
1.0
HTML: Breaking Angular input with a cast to number - **Prettier 1.15.1** [Playground link](https://prettier.io/playground/#.....) ```sh "htmlWhitespaceSensitivity": "ignore" ``` **Input:** ```html <div class="tool-svg" fxFlex="10%" fxLayout="center center"> <cnc-tool-assembly-svg [tool]="toolAssembly.tool" [stickout]="+toolAssembly.stickoutMm" ></cnc-tool-assembly-svg> </div> ``` **Output:** ```html <div class="tool-svg" fxFlex="10%" fxLayout="center center"> <cnc-tool-assembly-svg [tool]="toolAssembly.tool" [stickout]="toolAssembly.stickoutMm - "></cnc-tool-assembly-svg> </div> ``` **Problem here:** ``` [stickout]="+toolAssembly.stickoutMm" ``` is updated to: ``` [stickout]="toolAssembly.stickoutMm - " ``` **Expected behavior:** :warning: Behavior shouldn't be changed :warning:
priority
html breaking angular input with a cast to number prettier sh htmlwhitespacesensitivity ignore input html cnc tool assembly svg toolassembly tool toolassembly stickoutmm output html problem here toolassembly stickoutmm is updated to toolassembly stickoutmm expected behavior warning behavior shouldn t be changed warning
1
5,991
2,581,663,436
IssuesEvent
2015-02-14 08:56:12
AutomationSolutionz/Framework_0.1
https://api.github.com/repos/AutomationSolutionz/Framework_0.1
opened
Search is not showing unique item
bug Priority 2 (High) Shetu
If there are two passed items in the run detail, it is showing passed twice. It needs to show unique item only. This bug was found in run id detail page. But make sure it is solved in all other drop downs.
1.0
Search is not showing unique item - If there are two passed items in the run detail, it is showing passed twice. It needs to show unique item only. This bug was found in run id detail page. But make sure it is solved in all other drop downs.
priority
search is not showing unique item if there are two passed items in the run detail it is showing passed twice it needs to show unique item only this bug was found in run id detail page but make sure it is solved in all other drop downs
1
64,246
3,206,531,300
IssuesEvent
2015-10-05 01:56:41
cs2103aug2015-f09-2j/main
https://api.github.com/repos/cs2103aug2015-f09-2j/main
closed
Design Entity Class: Task
priority.high type.task
Attributes: ArrayList<Notes> id: String deadline: String (if null: floating task) description: String category priority
1.0
Design Entity Class: Task - Attributes: ArrayList<Notes> id: String deadline: String (if null: floating task) description: String category priority
priority
design entity class task attributes arraylist id string deadline string if null floating task description string category priority
1
567,923
16,919,651,016
IssuesEvent
2021-06-25 02:16:44
Guilherme-Vasconcelos/Aqua
https://api.github.com/repos/Guilherme-Vasconcelos/Aqua
closed
Improve the pre-commit hook
enhancement high priority
**Describe your idea** The current pre-commit hook (located at `hooks/pre-commit`) is not completely reliable because it doesn't check for tricks such as: - Commented authorize; - authorize inside string For instance, these two snippets would be considered OK to the current pre-commit hook: ```python # @authorize def start(...): ... ``` ```python '@authorize' def start(...): ... ``` The task is to create a pre-commit hook that actually checks: - If all command handlers (i.e. files in aqua/extensions/commands) and the talk handler (aqua/extensions/talk/talk.py) are actually using a valid authorize (the authorize must be present, correctly placed and not inside a string or comment); - If all command handlers are described in the `/help` command: aqua/extensions/commands/help.py; **Extra information** - If the new pre-commit hook is not written in bash, remember to update the GitHub workflow accordingly (`.github/workflows/pre-commit.yml`). - Only the function that has the same name as the file needs an authorize (because that's the function which is executed by the bot). For example, the function `def start(...)` needs an authorize if it is located inside the `start.py` file. - Internal functions (those which begin with `_` such as `/_invalid`, which is to be merged in `command-abreviations` branch) should not be documented in `/help` or README, though they must still have an authorize. - https://docs.python.org/3/library/ast.html + https://docs.pytest.org/en/6.2.x/ (possibility of writing tests and using AST to verify decorators. With this it is no longer required to have a pre-commit hook. These tests will act as some kind of static analysis.)
1.0
Improve the pre-commit hook - **Describe your idea** The current pre-commit hook (located at `hooks/pre-commit`) is not completely reliable because it doesn't check for tricks such as: - Commented authorize; - authorize inside string For instance, these two snippets would be considered OK to the current pre-commit hook: ```python # @authorize def start(...): ... ``` ```python '@authorize' def start(...): ... ``` The task is to create a pre-commit hook that actually checks: - If all command handlers (i.e. files in aqua/extensions/commands) and the talk handler (aqua/extensions/talk/talk.py) are actually using a valid authorize (the authorize must be present, correctly placed and not inside a string or comment); - If all command handlers are described in the `/help` command: aqua/extensions/commands/help.py; **Extra information** - If the new pre-commit hook is not written in bash, remember to update the GitHub workflow accordingly (`.github/workflows/pre-commit.yml`). - Only the function that has the same name as the file needs an authorize (because that's the function which is executed by the bot). For example, the function `def start(...)` needs an authorize if it is located inside the `start.py` file. - Internal functions (those which begin with `_` such as `/_invalid`, which is to be merged in `command-abreviations` branch) should not be documented in `/help` or README, though they must still have an authorize. - https://docs.python.org/3/library/ast.html + https://docs.pytest.org/en/6.2.x/ (possibility of writing tests and using AST to verify decorators. With this it is no longer required to have a pre-commit hook. These tests will act as some kind of static analysis.)
priority
improve the pre commit hook describe your idea the current pre commit hook located at hooks pre commit is not completely reliable because it doesn t check for tricks such as commented authorize authorize inside string for instance these two snippets would be considered ok to the current pre commit hook python authorize def start python authorize def start the task is to create a pre commit hook that actually checks if all command handlers i e files in aqua extensions commands and the talk handler aqua extensions talk talk py are actually using a valid authorize the authorize must be present correctly placed and not inside a string or comment if all command handlers are described in the help command aqua extensions commands help py extra information if the new pre commit hook is not written in bash remember to update the github workflow accordingly github workflows pre commit yml only the function that has the same name as the file needs an authorize because that s the function which is executed by the bot for example the function def start needs an authorize if it is located inside the start py file internal functions those which begin with such as invalid which is to be merged in command abreviations branch should not be documented in help or readme though they must still have an authorize possibility of writing tests and using ast to verify decorators with this it is no longer required to have a pre commit hook these tests will act as some kind of static analysis
1
804,785
29,501,256,880
IssuesEvent
2023-06-02 22:09:38
aws/s2n-tls
https://api.github.com/repos/aws/s2n-tls
closed
Crash on thread termination when s2n has been unloaded
priority/high size/medium
### Problem: I've encountered a couple of issues with s2n's thread local state clean up. The first is a crash on thread exit. Pre-requisites: s2n built as a shared object on a glibc-based system The following simple C++ program crashes while resolving the call to helper1.join(): ``` #include <thread> #include <dlfcn.h> void foo() { void *s2n_so = dlopen("<path to libs2n.so>", RTLD_NOW); int (*s2n_init)(void) = NULL; *(void **)(&s2n_init) = dlsym(s2n_so, "s2n_init"); int (*s2n_cleanup)(void) = NULL; *(void **)(&s2n_cleanup) = dlsym(s2n_so, "s2n_cleanup"); (*s2n_init)(); (*s2n_cleanup)(); dlclose(s2n_so); } int main(int argc, char *argv[]) { std::thread helper1(foo); helper1.join(); return 0; } ``` The crash call stack looks like: ``` #0 0x00007ffff6d8ee79 in ?? () #1 0x00007ffff7891711 in __GI___nptl_deallocate_tsd () at ./nptl/nptl_deallocate_tsd.c:73 #2 __GI___nptl_deallocate_tsd () at ./nptl/nptl_deallocate_tsd.c:22 #3 0x00007ffff78949ca in start_thread (arg=<optimized out>) at ./nptl/pthread_create.c:453 #4 0x00007ffff7926a00 in clone3 () at ../sysdeps/unix/sysv/linux/x86_64/clone3.S:81 ``` In particular, nptl_deallocate_ts.c:73 is the place in glibc where thread local slot destructors are invoked (https://fossies.org/linux/glibc/nptl/nptl_deallocate_tsd.c). In this case, the destructor call refers to code that has been unloaded due to the call to dlclose(). The root cause is the initialization of `s2n_per_thread_rand_state_key` in `s2n_drbg_make_rand_state_key`: https://github.com/aws/s2n-tls/blob/v1.3.43/utils/s2n_random.c#L147. The slot is allocated with a destructor, but even if s2n is shutdown properly beforehand, the destructor is still getting invoked at thread end, after the shared object unload. The setup may seem contrived, but it's actually a simplification of a scenario we are experiencing when a managed runtime (like node) is using a native module (the CRT, that includes s2n) in a worker thread. Before terminating itself, the worker thread unloads the module which in turn unloads s2n. ### Solution: https://github.com/aws/s2n-tls/pull/3988 Since destructors are only called if the slot contains a non-null value, my tentative fix proposal is to zero the slot in `s2n_rand_cleanup_thread` after the bits have been wiped. The crash disappears once this is done. I don't understand s2n's thread local storage usage well enough to know if that is sufficient or if there is a potential for `s2n_rand_cleanup_thread` to be called while the key is uninitialized (so perhaps don't fail if the `pthread_setspecific` call fails). * **Does this change what S2N sends over the wire?** No * **Does this change any public APIs?** No * **Which versions of TLS will this impact?** All/universal ### Requirements / Acceptance Criteria: * **Testing:** Ideally a standalone test that essentially repeats the above program and doesn't crash would be a useful test. Converting the std::thread to pthreads API would let the test stay pure C. Since it requires a known fixed path to a shared build of s2n (and no link time dependency), the setup may be a bit messier than existing tests. ### Out of scope: N/A
1.0
Crash on thread termination when s2n has been unloaded - ### Problem: I've encountered a couple of issues with s2n's thread local state clean up. The first is a crash on thread exit. Pre-requisites: s2n built as a shared object on a glibc-based system The following simple C++ program crashes while resolving the call to helper1.join(): ``` #include <thread> #include <dlfcn.h> void foo() { void *s2n_so = dlopen("<path to libs2n.so>", RTLD_NOW); int (*s2n_init)(void) = NULL; *(void **)(&s2n_init) = dlsym(s2n_so, "s2n_init"); int (*s2n_cleanup)(void) = NULL; *(void **)(&s2n_cleanup) = dlsym(s2n_so, "s2n_cleanup"); (*s2n_init)(); (*s2n_cleanup)(); dlclose(s2n_so); } int main(int argc, char *argv[]) { std::thread helper1(foo); helper1.join(); return 0; } ``` The crash call stack looks like: ``` #0 0x00007ffff6d8ee79 in ?? () #1 0x00007ffff7891711 in __GI___nptl_deallocate_tsd () at ./nptl/nptl_deallocate_tsd.c:73 #2 __GI___nptl_deallocate_tsd () at ./nptl/nptl_deallocate_tsd.c:22 #3 0x00007ffff78949ca in start_thread (arg=<optimized out>) at ./nptl/pthread_create.c:453 #4 0x00007ffff7926a00 in clone3 () at ../sysdeps/unix/sysv/linux/x86_64/clone3.S:81 ``` In particular, nptl_deallocate_ts.c:73 is the place in glibc where thread local slot destructors are invoked (https://fossies.org/linux/glibc/nptl/nptl_deallocate_tsd.c). In this case, the destructor call refers to code that has been unloaded due to the call to dlclose(). The root cause is the initialization of `s2n_per_thread_rand_state_key` in `s2n_drbg_make_rand_state_key`: https://github.com/aws/s2n-tls/blob/v1.3.43/utils/s2n_random.c#L147. The slot is allocated with a destructor, but even if s2n is shutdown properly beforehand, the destructor is still getting invoked at thread end, after the shared object unload. The setup may seem contrived, but it's actually a simplification of a scenario we are experiencing when a managed runtime (like node) is using a native module (the CRT, that includes s2n) in a worker thread. Before terminating itself, the worker thread unloads the module which in turn unloads s2n. ### Solution: https://github.com/aws/s2n-tls/pull/3988 Since destructors are only called if the slot contains a non-null value, my tentative fix proposal is to zero the slot in `s2n_rand_cleanup_thread` after the bits have been wiped. The crash disappears once this is done. I don't understand s2n's thread local storage usage well enough to know if that is sufficient or if there is a potential for `s2n_rand_cleanup_thread` to be called while the key is uninitialized (so perhaps don't fail if the `pthread_setspecific` call fails). * **Does this change what S2N sends over the wire?** No * **Does this change any public APIs?** No * **Which versions of TLS will this impact?** All/universal ### Requirements / Acceptance Criteria: * **Testing:** Ideally a standalone test that essentially repeats the above program and doesn't crash would be a useful test. Converting the std::thread to pthreads API would let the test stay pure C. Since it requires a known fixed path to a shared build of s2n (and no link time dependency), the setup may be a bit messier than existing tests. ### Out of scope: N/A
priority
crash on thread termination when has been unloaded problem i ve encountered a couple of issues with s thread local state clean up the first is a crash on thread exit pre requisites built as a shared object on a glibc based system the following simple c program crashes while resolving the call to join include include void foo void so dlopen rtld now int init void null void init dlsym so init int cleanup void null void cleanup dlsym so cleanup init cleanup dlclose so int main int argc char argv std thread foo join return the crash call stack looks like in in gi nptl deallocate tsd at nptl nptl deallocate tsd c gi nptl deallocate tsd at nptl nptl deallocate tsd c in start thread arg at nptl pthread create c in at sysdeps unix sysv linux s in particular nptl deallocate ts c is the place in glibc where thread local slot destructors are invoked in this case the destructor call refers to code that has been unloaded due to the call to dlclose the root cause is the initialization of per thread rand state key in drbg make rand state key the slot is allocated with a destructor but even if is shutdown properly beforehand the destructor is still getting invoked at thread end after the shared object unload the setup may seem contrived but it s actually a simplification of a scenario we are experiencing when a managed runtime like node is using a native module the crt that includes in a worker thread before terminating itself the worker thread unloads the module which in turn unloads solution since destructors are only called if the slot contains a non null value my tentative fix proposal is to zero the slot in rand cleanup thread after the bits have been wiped the crash disappears once this is done i don t understand s thread local storage usage well enough to know if that is sufficient or if there is a potential for rand cleanup thread to be called while the key is uninitialized so perhaps don t fail if the pthread setspecific call fails does this change what sends over the wire no does this change any public apis no which versions of tls will this impact all universal requirements acceptance criteria testing ideally a standalone test that essentially repeats the above program and doesn t crash would be a useful test converting the std thread to pthreads api would let the test stay pure c since it requires a known fixed path to a shared build of and no link time dependency the setup may be a bit messier than existing tests out of scope n a
1
522,337
15,158,406,496
IssuesEvent
2021-02-12 01:04:44
NOAA-GSL/MATS
https://api.github.com/repos/NOAA-GSL/MATS
opened
New plot for MATS: Normalized errors
Priority: High Project: MATS Type: Task
--- Author Name: **dave.turner** (@davidturner53) Original Redmine Issue: 85519, https://vlab.ncep.noaa.gov/redmine/issues/85519 Original Date: 2020-12-02 Original Assignee: molly.b.smith --- Request from Stan, Eric, et al. Desire to have normalized errors (NE), computed as NE = 100* (exp – cntrl) / cntrl, and uncertainties in both the cntrl and exp data are propagated to provide the error bars. Assume that the first curve is the control, and that any other curves are the experiment. Need to be able to show this for sure as a profile (upper air and aircraft) and time-series (all apps); but this can be done for any of the other plot types (e.g., die off, valid time, etc). I'd start with continuous variable apps first, then work on contingency tables second.
1.0
New plot for MATS: Normalized errors - --- Author Name: **dave.turner** (@davidturner53) Original Redmine Issue: 85519, https://vlab.ncep.noaa.gov/redmine/issues/85519 Original Date: 2020-12-02 Original Assignee: molly.b.smith --- Request from Stan, Eric, et al. Desire to have normalized errors (NE), computed as NE = 100* (exp – cntrl) / cntrl, and uncertainties in both the cntrl and exp data are propagated to provide the error bars. Assume that the first curve is the control, and that any other curves are the experiment. Need to be able to show this for sure as a profile (upper air and aircraft) and time-series (all apps); but this can be done for any of the other plot types (e.g., die off, valid time, etc). I'd start with continuous variable apps first, then work on contingency tables second.
priority
new plot for mats normalized errors author name dave turner original redmine issue original date original assignee molly b smith request from stan eric et al desire to have normalized errors ne computed as ne exp – cntrl cntrl and uncertainties in both the cntrl and exp data are propagated to provide the error bars assume that the first curve is the control and that any other curves are the experiment need to be able to show this for sure as a profile upper air and aircraft and time series all apps but this can be done for any of the other plot types e g die off valid time etc i d start with continuous variable apps first then work on contingency tables second
1
142,918
5,480,048,737
IssuesEvent
2017-03-13 05:52:47
fossasia/open-event-webapp
https://api.github.com/repos/fossasia/open-event-webapp
opened
Images of speakers not showing up on tracks and other pages
bug Priority: High Priority: Urgent
After the revert of the lazy loading the images do not show up anymore on a number of pages.
2.0
Images of speakers not showing up on tracks and other pages - After the revert of the lazy loading the images do not show up anymore on a number of pages.
priority
images of speakers not showing up on tracks and other pages after the revert of the lazy loading the images do not show up anymore on a number of pages
1
53,860
3,052,025,666
IssuesEvent
2015-08-12 12:36:54
OCHA-DAP/hdx-ckan
https://api.github.com/repos/OCHA-DAP/hdx-ckan
closed
COD Migration Geo Preview thingy
CODmigration GeoPreview Priority-High
Making a placeholder for this for discussion. We have closed the issue about triggering geopreview for datasets made from the API, but gospawn won't know when to apply the magic words (it won't know if a dataset is a zipped shapefile, zipped geodatabase, or zipped cat vidoes). Which means that for most of the migrated CODs we would have to trigger it manually. Options (open to more ideas): 1. Add something to gospawn to inspect zip files for .shp and set file format to "zipped shapefile" if found. KML (uncommon in CODs) and geojson (rare to nonexistant) should be triggered automatically by their file format (though this needs to be tested)
1.0
COD Migration Geo Preview thingy - Making a placeholder for this for discussion. We have closed the issue about triggering geopreview for datasets made from the API, but gospawn won't know when to apply the magic words (it won't know if a dataset is a zipped shapefile, zipped geodatabase, or zipped cat vidoes). Which means that for most of the migrated CODs we would have to trigger it manually. Options (open to more ideas): 1. Add something to gospawn to inspect zip files for .shp and set file format to "zipped shapefile" if found. KML (uncommon in CODs) and geojson (rare to nonexistant) should be triggered automatically by their file format (though this needs to be tested)
priority
cod migration geo preview thingy making a placeholder for this for discussion we have closed the issue about triggering geopreview for datasets made from the api but gospawn won t know when to apply the magic words it won t know if a dataset is a zipped shapefile zipped geodatabase or zipped cat vidoes which means that for most of the migrated cods we would have to trigger it manually options open to more ideas add something to gospawn to inspect zip files for shp and set file format to zipped shapefile if found kml uncommon in cods and geojson rare to nonexistant should be triggered automatically by their file format though this needs to be tested
1
58,205
3,088,042,290
IssuesEvent
2015-08-25 14:51:44
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
closed
iOS 8 cvvideocamera
affected: 2.4 auto-transferred bug category: ios priority: high
Transferred from http://code.opencv.org/issues/3893 ``` || Mohammed Hgras on 2014-09-02 08:00 || Priority: High || Affected: 2.4.9 (latest release) || Category: ios || Tracker: Bug || Difficulty: || PR: || Platform: ARM / iOS ``` iOS 8 cvvideocamera ----------- ``` Hey, I am using Xcode 6 beta, I am developing an app with opencv my target was 7.1. Things were going so good. I am changing the target to iOS 8 beta. Unfortunately my code is not going well. The button that starts the camera do not do any thing. I got these warnings 2014-08-10 14:06:48.494 Detect with curves[258:10575] 17545849:_UIScreenEdgePanRecognizerEdgeSettings.edgeRegionSize=13.000000 2014-08-10 14:06:48.538 Detect with curves[258:10575] camera available: YES 2014-08-10 14:06:53.459 Detect with curves[258:10575] [Camera] device connected? YES 2014-08-10 14:06:53.459 Detect with curves[258:10575] [Camera] device position back 2014-08-10 14:06:53.463 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> isVideoMinFrameDurationSupported] is deprecated. Please use AVCaptureDevice activeFormat.videoSupportedFrameRateRanges 2014-08-10 14:06:53.464 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> setVideoMinFrameDuration:] is deprecated. Please use AVCaptureDevice setActiveVideoMinFrameDuration 2014-08-10 14:06:53.464 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> isVideoMaxFrameDurationSupported] is deprecated. Please use AVCaptureDevice activeFormat.videoSupportedFrameRateRanges 2014-08-10 14:06:53.465 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> setVideoMaxFrameDuration:] is deprecated. Please use AVCaptureDevice setActiveVideoMaxFrameDuration 2014-08-10 14:06:53.466 Detect with curves[258:10575] layout preview layer 2014-08-10 14:06:53.467 Detect with curves[258:10575] [Camera] created AVCaptureVideoDataOutput at 70 FPS ``` History ------- ##### Andrey Pavlenko on 2014-09-04 13:19 ``` Mohammed, thank you for the report. OpenCV community would really appreciate if you also provide a fix for this issue via a pull-request (see [[How to contribute]]). If not, please wait for somebody else to take this task. - Category set to ios ```
1.0
iOS 8 cvvideocamera - Transferred from http://code.opencv.org/issues/3893 ``` || Mohammed Hgras on 2014-09-02 08:00 || Priority: High || Affected: 2.4.9 (latest release) || Category: ios || Tracker: Bug || Difficulty: || PR: || Platform: ARM / iOS ``` iOS 8 cvvideocamera ----------- ``` Hey, I am using Xcode 6 beta, I am developing an app with opencv my target was 7.1. Things were going so good. I am changing the target to iOS 8 beta. Unfortunately my code is not going well. The button that starts the camera do not do any thing. I got these warnings 2014-08-10 14:06:48.494 Detect with curves[258:10575] 17545849:_UIScreenEdgePanRecognizerEdgeSettings.edgeRegionSize=13.000000 2014-08-10 14:06:48.538 Detect with curves[258:10575] camera available: YES 2014-08-10 14:06:53.459 Detect with curves[258:10575] [Camera] device connected? YES 2014-08-10 14:06:53.459 Detect with curves[258:10575] [Camera] device position back 2014-08-10 14:06:53.463 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> isVideoMinFrameDurationSupported] is deprecated. Please use AVCaptureDevice activeFormat.videoSupportedFrameRateRanges 2014-08-10 14:06:53.464 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> setVideoMinFrameDuration:] is deprecated. Please use AVCaptureDevice setActiveVideoMinFrameDuration 2014-08-10 14:06:53.464 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> isVideoMaxFrameDurationSupported] is deprecated. Please use AVCaptureDevice activeFormat.videoSupportedFrameRateRanges 2014-08-10 14:06:53.465 Detect with curves[258:10575] WARNING: -[<AVCaptureConnection: 0x78007160> setVideoMaxFrameDuration:] is deprecated. Please use AVCaptureDevice setActiveVideoMaxFrameDuration 2014-08-10 14:06:53.466 Detect with curves[258:10575] layout preview layer 2014-08-10 14:06:53.467 Detect with curves[258:10575] [Camera] created AVCaptureVideoDataOutput at 70 FPS ``` History ------- ##### Andrey Pavlenko on 2014-09-04 13:19 ``` Mohammed, thank you for the report. OpenCV community would really appreciate if you also provide a fix for this issue via a pull-request (see [[How to contribute]]). If not, please wait for somebody else to take this task. - Category set to ios ```
priority
ios cvvideocamera transferred from mohammed hgras on priority high affected latest release category ios tracker bug difficulty pr platform arm ios ios cvvideocamera hey i am using xcode beta i am developing an app with opencv my target was things were going so good i am changing the target to ios beta unfortunately my code is not going well the button that starts the camera do not do any thing i got these warnings detect with curves uiscreenedgepanrecognizeredgesettings edgeregionsize detect with curves camera available yes detect with curves device connected yes detect with curves device position back detect with curves warning is deprecated please use avcapturedevice activeformat videosupportedframerateranges detect with curves warning is deprecated please use avcapturedevice setactivevideominframeduration detect with curves warning is deprecated please use avcapturedevice activeformat videosupportedframerateranges detect with curves warning is deprecated please use avcapturedevice setactivevideomaxframeduration detect with curves layout preview layer detect with curves created avcapturevideodataoutput at fps history andrey pavlenko on mohammed thank you for the report opencv community would really appreciate if you also provide a fix for this issue via a pull request see if not please wait for somebody else to take this task category set to ios
1
124,640
4,928,794,374
IssuesEvent
2016-11-27 14:19:27
Citadel-Station-13/Citadel-Station-13
https://api.github.com/repos/Citadel-Station-13/Citadel-Station-13
closed
Space Parallax Overlays All Turf and Mobs
Bug Icon/Art Priority: High
Issue concerning the Space Parallax for animated stars and what not. -Issue caused just by a base compile from 510.1332 version of Dream Maker, which is the earliest version of Dream Maker able to compile Sixth Port without code issues. Runs on latest Dream Daemon. ![c05eadb4069f6e4ae5623a3b0b3c4d91](https://cloud.githubusercontent.com/assets/5521322/20647134/dd7f03b6-b451-11e6-9c7d-8936b2fdbef1.png) ![921d8addfdb37badcb3d616ce9eb3a66](https://cloud.githubusercontent.com/assets/5521322/20647136/dfc9a5f4-b451-11e6-970c-8335bd0b1d8a.png) -Current workaround: Disable parallax through game preferences and restart Dream Seeker. Stops the parallax from being shown, thus removing it from overlaying everything.
1.0
Space Parallax Overlays All Turf and Mobs - Issue concerning the Space Parallax for animated stars and what not. -Issue caused just by a base compile from 510.1332 version of Dream Maker, which is the earliest version of Dream Maker able to compile Sixth Port without code issues. Runs on latest Dream Daemon. ![c05eadb4069f6e4ae5623a3b0b3c4d91](https://cloud.githubusercontent.com/assets/5521322/20647134/dd7f03b6-b451-11e6-9c7d-8936b2fdbef1.png) ![921d8addfdb37badcb3d616ce9eb3a66](https://cloud.githubusercontent.com/assets/5521322/20647136/dfc9a5f4-b451-11e6-970c-8335bd0b1d8a.png) -Current workaround: Disable parallax through game preferences and restart Dream Seeker. Stops the parallax from being shown, thus removing it from overlaying everything.
priority
space parallax overlays all turf and mobs issue concerning the space parallax for animated stars and what not issue caused just by a base compile from version of dream maker which is the earliest version of dream maker able to compile sixth port without code issues runs on latest dream daemon current workaround disable parallax through game preferences and restart dream seeker stops the parallax from being shown thus removing it from overlaying everything
1
549,935
16,101,985,664
IssuesEvent
2021-04-27 10:27:58
snowplow/snowplow-android-tracker
https://api.github.com/repos/snowplow/snowplow-android-tracker
closed
Store sessions info separately per tracker namespace
priority:high status:completed type:enhancement
Multiple tracker instances needs to have independent session info. The tracker 1.x used to store some information about the session in the filesystem. That would cause interferences between trackers in a multi-tracker app.
1.0
Store sessions info separately per tracker namespace - Multiple tracker instances needs to have independent session info. The tracker 1.x used to store some information about the session in the filesystem. That would cause interferences between trackers in a multi-tracker app.
priority
store sessions info separately per tracker namespace multiple tracker instances needs to have independent session info the tracker x used to store some information about the session in the filesystem that would cause interferences between trackers in a multi tracker app
1
90,358
3,814,804,062
IssuesEvent
2016-03-28 15:11:17
cs2102-database-systems-g11/stuffy
https://api.github.com/repos/cs2102-database-systems-g11/stuffy
closed
User can bid for advertisements
priority.high type.feature
- [ ] User is allowed to bid on items if deadline is not passed. - [ ] If bid >= buyout, the user immediately wins. - [ ] User can choose to buyout at any time and immediately win. Prereq: #4, #8
1.0
User can bid for advertisements - - [ ] User is allowed to bid on items if deadline is not passed. - [ ] If bid >= buyout, the user immediately wins. - [ ] User can choose to buyout at any time and immediately win. Prereq: #4, #8
priority
user can bid for advertisements user is allowed to bid on items if deadline is not passed if bid buyout the user immediately wins user can choose to buyout at any time and immediately win prereq
1
612,505
19,014,325,388
IssuesEvent
2021-11-23 12:50:57
wso2/product-microgateway
https://api.github.com/repos/wso2/product-microgateway
closed
Remove `service_envoyproxy_io` cluster from the router config
Type/Bug Priority/Highest
### Description: <!-- Describe the issue --> cluster called `service_envoyproxy_io` is added to the envoy configuration to handle `/` route. This cluster's endpoint is configured as `envoyproxy.io` which is a third party endpoint for a Choreo Connect user. This is not intentional and we should remove this cluster from the router. ### Steps to reproduce: - Start choreo connect and check the configured clusters in the router configured dump - You'll see `service_envoyproxy_io` cluster ### Affected Product Version: <!-- Members can use Affected/*** labels --> 1.0.0-beta-snapshot ### Environment details (with versions): - OS: - Client: - Env (Docker/K8s):
1.0
Remove `service_envoyproxy_io` cluster from the router config - ### Description: <!-- Describe the issue --> cluster called `service_envoyproxy_io` is added to the envoy configuration to handle `/` route. This cluster's endpoint is configured as `envoyproxy.io` which is a third party endpoint for a Choreo Connect user. This is not intentional and we should remove this cluster from the router. ### Steps to reproduce: - Start choreo connect and check the configured clusters in the router configured dump - You'll see `service_envoyproxy_io` cluster ### Affected Product Version: <!-- Members can use Affected/*** labels --> 1.0.0-beta-snapshot ### Environment details (with versions): - OS: - Client: - Env (Docker/K8s):
priority
remove service envoyproxy io cluster from the router config description cluster called service envoyproxy io is added to the envoy configuration to handle route this cluster s endpoint is configured as envoyproxy io which is a third party endpoint for a choreo connect user this is not intentional and we should remove this cluster from the router steps to reproduce start choreo connect and check the configured clusters in the router configured dump you ll see service envoyproxy io cluster affected product version beta snapshot environment details with versions os client env docker
1
467,460
13,449,010,932
IssuesEvent
2020-09-08 16:14:55
flameshot-org/flameshot
https://api.github.com/repos/flameshot-org/flameshot
closed
Fix packages (rpm, deb, flatpak, snap)
Bug High Priority
Moving to cmake broke the package builds. These need to be fixed and added to CI.
1.0
Fix packages (rpm, deb, flatpak, snap) - Moving to cmake broke the package builds. These need to be fixed and added to CI.
priority
fix packages rpm deb flatpak snap moving to cmake broke the package builds these need to be fixed and added to ci
1
270,340
8,454,840,455
IssuesEvent
2018-10-21 08:44:02
CS2103-AY1819S1-W17-3/main
https://api.github.com/repos/CS2103-AY1819S1-W17-3/main
opened
Update the person and group details in the meeting book
priority.High status.Ongoing type.Enhancement
The person and group details are modifiable and can be updated to allow greater flexibility in managing the MeetingBook
1.0
Update the person and group details in the meeting book - The person and group details are modifiable and can be updated to allow greater flexibility in managing the MeetingBook
priority
update the person and group details in the meeting book the person and group details are modifiable and can be updated to allow greater flexibility in managing the meetingbook
1
217,773
7,328,126,890
IssuesEvent
2018-03-04 17:41:54
plotify/plotify
https://api.github.com/repos/plotify/plotify
closed
Lizenzdateien werden dem Installationsprogramm nicht hinzugefügt
Bug Platform: macOS Priority: High
/Applications/Plotify-app/Contents/LICENSES.dependencies.txt
1.0
Lizenzdateien werden dem Installationsprogramm nicht hinzugefügt - /Applications/Plotify-app/Contents/LICENSES.dependencies.txt
priority
lizenzdateien werden dem installationsprogramm nicht hinzugefügt applications plotify app contents licenses dependencies txt
1
631,561
20,154,217,990
IssuesEvent
2022-02-09 15:06:31
carbon-design-system/carbon-for-ibm-dotcom
https://api.github.com/repos/carbon-design-system/carbon-for-ibm-dotcom
closed
React e2e testing foundation
priority: high epic
This epic tracks the setup of wider e2e testing of react components and its downstream test applications (NextJS test application). This should include the ability to test multiple states (hover, active, carbon themes, component variations, etc), as well as automated visual regression.
1.0
React e2e testing foundation - This epic tracks the setup of wider e2e testing of react components and its downstream test applications (NextJS test application). This should include the ability to test multiple states (hover, active, carbon themes, component variations, etc), as well as automated visual regression.
priority
react testing foundation this epic tracks the setup of wider testing of react components and its downstream test applications nextjs test application this should include the ability to test multiple states hover active carbon themes component variations etc as well as automated visual regression
1
209,888
7,180,849,192
IssuesEvent
2018-02-01 01:23:12
sul-dlss/preservation_catalog
https://api.github.com/repos/sul-dlss/preservation_catalog
closed
ship prod logs off to afs or elk
devops practice high priority in progress
we need to long term retain logs from perservation-catalog-prod. "/opt/app/pres/preservation_catalog/shared/log/production.log-20171118.gz" etc.
1.0
ship prod logs off to afs or elk - we need to long term retain logs from perservation-catalog-prod. "/opt/app/pres/preservation_catalog/shared/log/production.log-20171118.gz" etc.
priority
ship prod logs off to afs or elk we need to long term retain logs from perservation catalog prod opt app pres preservation catalog shared log production log gz etc
1
359,060
10,653,576,230
IssuesEvent
2019-10-17 14:40:01
woocommerce/woocommerce-admin
https://api.github.com/repos/woocommerce/woocommerce-admin
closed
Fix broken dashboard.
Dashboard [Many] Large [Much] Large [Priority] High [Type] Bug
Discovered on the latest version of master (as of a2194da08f6) that the dashboard is broken. The following errors exist in the console: ``` Uncaught TypeError: Cannot read property 'includes' of null at getTasks (index.js:15886) at index.js:10319 at ComponentWithSelect.getNextMergeProps (index.js:29841) at new ComponentWithSelect (index.js:29789) at constructClassInstance (react-dom.8b3dda97.js:13446) at updateClassComponent (react-dom.8b3dda97.js:17123) at beginWork$1 (react-dom.8b3dda97.js:18643) at HTMLUnknownElement.callCallback (react-dom.8b3dda97.js:341) at Object.invokeGuardedCallbackDev (react-dom.8b3dda97.js:391) at invokeGuardedCallback (react-dom.8b3dda97.js:448) ``` ``` The above error occurred in the <ComponentWithSelect> component: in ComponentWithSelect (created by Context.Consumer) in WithSelect(WithDispatch(CustomizableDashboard)) (created by Dashboard) in Dashboard (created by ComponentWithSelect) in ComponentWithSelect (created by Context.Consumer) in WithSelect(Dashboard) (created by Controller) in Controller (created by Layout) in div (created by Layout) in div (created by PrimaryLayout) in PrimaryLayout (created by Layout) in div (created by Layout) in Layout (created by Context.Consumer) in Route (created by _PageLayout) in Switch (created by _PageLayout) in Router (created by _PageLayout) in _PageLayout (created by FilteredComponent) in FilteredComponent ``` As a sidenote, it might be good to add error boundaries to the app to catch problems like this and display a more user-friendly error item instead of a blank screen?
1.0
Fix broken dashboard. - Discovered on the latest version of master (as of a2194da08f6) that the dashboard is broken. The following errors exist in the console: ``` Uncaught TypeError: Cannot read property 'includes' of null at getTasks (index.js:15886) at index.js:10319 at ComponentWithSelect.getNextMergeProps (index.js:29841) at new ComponentWithSelect (index.js:29789) at constructClassInstance (react-dom.8b3dda97.js:13446) at updateClassComponent (react-dom.8b3dda97.js:17123) at beginWork$1 (react-dom.8b3dda97.js:18643) at HTMLUnknownElement.callCallback (react-dom.8b3dda97.js:341) at Object.invokeGuardedCallbackDev (react-dom.8b3dda97.js:391) at invokeGuardedCallback (react-dom.8b3dda97.js:448) ``` ``` The above error occurred in the <ComponentWithSelect> component: in ComponentWithSelect (created by Context.Consumer) in WithSelect(WithDispatch(CustomizableDashboard)) (created by Dashboard) in Dashboard (created by ComponentWithSelect) in ComponentWithSelect (created by Context.Consumer) in WithSelect(Dashboard) (created by Controller) in Controller (created by Layout) in div (created by Layout) in div (created by PrimaryLayout) in PrimaryLayout (created by Layout) in div (created by Layout) in Layout (created by Context.Consumer) in Route (created by _PageLayout) in Switch (created by _PageLayout) in Router (created by _PageLayout) in _PageLayout (created by FilteredComponent) in FilteredComponent ``` As a sidenote, it might be good to add error boundaries to the app to catch problems like this and display a more user-friendly error item instead of a blank screen?
priority
fix broken dashboard discovered on the latest version of master as of that the dashboard is broken the following errors exist in the console uncaught typeerror cannot read property includes of null at gettasks index js at index js at componentwithselect getnextmergeprops index js at new componentwithselect index js at constructclassinstance react dom js at updateclasscomponent react dom js at beginwork react dom js at htmlunknownelement callcallback react dom js at object invokeguardedcallbackdev react dom js at invokeguardedcallback react dom js the above error occurred in the component in componentwithselect created by context consumer in withselect withdispatch customizabledashboard created by dashboard in dashboard created by componentwithselect in componentwithselect created by context consumer in withselect dashboard created by controller in controller created by layout in div created by layout in div created by primarylayout in primarylayout created by layout in div created by layout in layout created by context consumer in route created by pagelayout in switch created by pagelayout in router created by pagelayout in pagelayout created by filteredcomponent in filteredcomponent as a sidenote it might be good to add error boundaries to the app to catch problems like this and display a more user friendly error item instead of a blank screen
1
532,451
15,556,913,619
IssuesEvent
2021-03-16 08:28:28
AY2021S2-CS2113-W10-2/tp
https://api.github.com/repos/AY2021S2-CS2113-W10-2/tp
closed
Count number of emails in inbox
priority.High type.Story
As a user, I want to count the number of emails in box so that I can have a general idea about my total emails
1.0
Count number of emails in inbox - As a user, I want to count the number of emails in box so that I can have a general idea about my total emails
priority
count number of emails in inbox as a user i want to count the number of emails in box so that i can have a general idea about my total emails
1
692,621
23,742,989,950
IssuesEvent
2022-08-31 13:53:45
slsdetectorgroup/slsDetectorPackage
https://api.github.com/repos/slsdetectorgroup/slsDetectorPackage
closed
Receiver print out mulfunction for file writing
action - Bug priority - High status - resolved
<!-- Preview changes before submitting --> <!-- Please fill out everything that is applicable and edit title to match --> <!-- This is a comment, the syntax is a bit different from c++ or bash --> ##### *Describe the bug <!-- A clear and concise description of what the bug is --> When saving G2 data into disk (in continuous mode), the receiver prints the following error: ... - 16:28:48.384 ERROR: 0 : Write to file failed for image number 2717. Wrote 1 bytes instead of 2672 - 16:28:48.384 ERROR: 0 : Write to file failed for image number 2718. Wrote 1 bytes instead of 2672 - 16:28:48.384 ERROR: 0 : Write to file failed for image number 2719. Wrote 1 bytes instead of 2672 - 16:28:48.390 INFO: Master File: /external_pool2/gotthard_data/datadir_gotthardII_module/SSRF/GH2-0122/ADC_cal//ADC_cal_0x22_100kHz_master_29.json - 16:28:48.395 INFO: Status: finished - 16:28:48.395 INFO: Summary of Port 50015 Missing Packets : 0 Complete Frames : 2720 Last Frame Caught : 1436003 All packets have been collected from the receiver and saved into the disk, but the receiver prints ERROR. ##### Distribution: <!-- RHEL7, RHEL6, Fedora, etc --> Any ##### Detector type or Receiver: <!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard, Receiver --> G2 ##### Software Package Version: <!-- developer, 4.2.0, 4.1.1, etc --> developer, the one to be released ##### Priority: <!-- Super Low, Low, Medium, High, Super High --> High ##### Expected behavior <!-- A clear and concise description of what you expected to happen. --> ##### To Reproduce <!-- Steps to reproduce the behavior: --> <!-- 1. Go to '...' --> <!-- 2. Click on '....' --> <!-- 3. Scroll down to '....' --> <!-- 4. See error --> ##### Screenshots <!-- If applicable, add screenshots to help explain your problem. --> ##### Additional context <!-- Add any other context about the problem here. -->
1.0
Receiver print out mulfunction for file writing - <!-- Preview changes before submitting --> <!-- Please fill out everything that is applicable and edit title to match --> <!-- This is a comment, the syntax is a bit different from c++ or bash --> ##### *Describe the bug <!-- A clear and concise description of what the bug is --> When saving G2 data into disk (in continuous mode), the receiver prints the following error: ... - 16:28:48.384 ERROR: 0 : Write to file failed for image number 2717. Wrote 1 bytes instead of 2672 - 16:28:48.384 ERROR: 0 : Write to file failed for image number 2718. Wrote 1 bytes instead of 2672 - 16:28:48.384 ERROR: 0 : Write to file failed for image number 2719. Wrote 1 bytes instead of 2672 - 16:28:48.390 INFO: Master File: /external_pool2/gotthard_data/datadir_gotthardII_module/SSRF/GH2-0122/ADC_cal//ADC_cal_0x22_100kHz_master_29.json - 16:28:48.395 INFO: Status: finished - 16:28:48.395 INFO: Summary of Port 50015 Missing Packets : 0 Complete Frames : 2720 Last Frame Caught : 1436003 All packets have been collected from the receiver and saved into the disk, but the receiver prints ERROR. ##### Distribution: <!-- RHEL7, RHEL6, Fedora, etc --> Any ##### Detector type or Receiver: <!-- If applicable, Eiger, Jungfrau, Mythen3, Gotthard2, Gotthard, Moench, ChipTestBoard, Receiver --> G2 ##### Software Package Version: <!-- developer, 4.2.0, 4.1.1, etc --> developer, the one to be released ##### Priority: <!-- Super Low, Low, Medium, High, Super High --> High ##### Expected behavior <!-- A clear and concise description of what you expected to happen. --> ##### To Reproduce <!-- Steps to reproduce the behavior: --> <!-- 1. Go to '...' --> <!-- 2. Click on '....' --> <!-- 3. Scroll down to '....' --> <!-- 4. See error --> ##### Screenshots <!-- If applicable, add screenshots to help explain your problem. --> ##### Additional context <!-- Add any other context about the problem here. -->
priority
receiver print out mulfunction for file writing describe the bug when saving data into disk in continuous mode the receiver prints the following error error write to file failed for image number wrote bytes instead of error write to file failed for image number wrote bytes instead of error write to file failed for image number wrote bytes instead of info master file external gotthard data datadir gotthardii module ssrf adc cal adc cal master json info status finished info summary of port missing packets complete frames last frame caught all packets have been collected from the receiver and saved into the disk but the receiver prints error distribution any detector type or receiver software package version developer the one to be released priority high expected behavior to reproduce screenshots additional context
1
402,072
11,801,731,558
IssuesEvent
2020-03-18 20:03:42
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[videocenter] Add support for Video On Demand
new feature priority: high
A new video content type has been added to VC: Video On Demand (`/component/video-on-demand`). We need to add UI support for this new kind of video. The content type is like the YouTube Video in but it uses a Video picker that will use MediaConvert to generate a single HLS URL.
1.0
[videocenter] Add support for Video On Demand - A new video content type has been added to VC: Video On Demand (`/component/video-on-demand`). We need to add UI support for this new kind of video. The content type is like the YouTube Video in but it uses a Video picker that will use MediaConvert to generate a single HLS URL.
priority
add support for video on demand a new video content type has been added to vc video on demand component video on demand we need to add ui support for this new kind of video the content type is like the youtube video in but it uses a video picker that will use mediaconvert to generate a single hls url
1
329,325
10,014,703,091
IssuesEvent
2019-07-15 18:11:57
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
closed
Classification Metadata
Display/Interface Function-Taxonomy/Identification Priority-High
@DerekSikes suggested that we change the title for this section from "Non-Classification Terms" to "Classification Metadata" ![image](https://user-images.githubusercontent.com/5725767/53131666-a4207580-352a-11e9-9195-e66b91ed00d1.png) @ArctosDB/taxonomy agrees! <!-- probot = {"768919":{"who":"dustymc","what":"","when":"2019-05-02T09:00:00.000Z"}} -->
1.0
Classification Metadata - @DerekSikes suggested that we change the title for this section from "Non-Classification Terms" to "Classification Metadata" ![image](https://user-images.githubusercontent.com/5725767/53131666-a4207580-352a-11e9-9195-e66b91ed00d1.png) @ArctosDB/taxonomy agrees! <!-- probot = {"768919":{"who":"dustymc","what":"","when":"2019-05-02T09:00:00.000Z"}} -->
priority
classification metadata dereksikes suggested that we change the title for this section from non classification terms to classification metadata arctosdb taxonomy agrees
1
735,334
25,390,044,293
IssuesEvent
2022-11-22 02:42:46
skyportal/skyportal
https://api.github.com/repos/skyportal/skyportal
closed
Follow-up requests are being assigned to the wrong instruments
bug critical high priority
Fritz assignment for follow-up runs is currently not working correctly. When a user requests that an object is assigned to a particular run, it ends up being assigned to a completely different run/instrument/night. The priority is not correct. Worse: after assigning an object, the GUI changes to show a collapsed list of requests. So users see that the request "went through". However, they do not see the actual request itself by default, so they will probably not notice that the request was wrong. I think this bug, if not fixed quickly, will lead to a lot of missed targets/wasted science. See video for an example: https://user-images.githubusercontent.com/14312838/203143949-09ab4ae3-8635-4d1e-a3d6-ba7fbf968a13.mp4
1.0
Follow-up requests are being assigned to the wrong instruments - Fritz assignment for follow-up runs is currently not working correctly. When a user requests that an object is assigned to a particular run, it ends up being assigned to a completely different run/instrument/night. The priority is not correct. Worse: after assigning an object, the GUI changes to show a collapsed list of requests. So users see that the request "went through". However, they do not see the actual request itself by default, so they will probably not notice that the request was wrong. I think this bug, if not fixed quickly, will lead to a lot of missed targets/wasted science. See video for an example: https://user-images.githubusercontent.com/14312838/203143949-09ab4ae3-8635-4d1e-a3d6-ba7fbf968a13.mp4
priority
follow up requests are being assigned to the wrong instruments fritz assignment for follow up runs is currently not working correctly when a user requests that an object is assigned to a particular run it ends up being assigned to a completely different run instrument night the priority is not correct worse after assigning an object the gui changes to show a collapsed list of requests so users see that the request went through however they do not see the actual request itself by default so they will probably not notice that the request was wrong i think this bug if not fixed quickly will lead to a lot of missed targets wasted science see video for an example
1
70,816
3,343,863,331
IssuesEvent
2015-11-15 20:52:28
R4stl1n/allianceauth
https://api.github.com/repos/R4stl1n/allianceauth
closed
index out of range exceptions thrown when processing empty lists
bug High Priority
In services.managers.util.ts3, line 142 throws an exception for empty lists. This occurs frequently after issuing a "use sid=1" command. The exception prevents adding users and goes unhandled.
1.0
index out of range exceptions thrown when processing empty lists - In services.managers.util.ts3, line 142 throws an exception for empty lists. This occurs frequently after issuing a "use sid=1" command. The exception prevents adding users and goes unhandled.
priority
index out of range exceptions thrown when processing empty lists in services managers util line throws an exception for empty lists this occurs frequently after issuing a use sid command the exception prevents adding users and goes unhandled
1
50,824
3,007,178,890
IssuesEvent
2015-07-27 14:52:52
pufexi/multiorder
https://api.github.com/repos/pufexi/multiorder
closed
Pri vytvareni obj v MO, kdyz nevlozim tracking to padne
bug high priority
1) Pokud tam neco napisu, tak to projde, ale to neni ucel. 2) Napadlo me ze pokud tam je nejaky tracking a ja importuji trackingy z Posty, tak co se stane? Zadouci by to bylo tracking prepsat tim novejsim z importu. Tak to tak zarid pokud se tak nedeje ci zda to nahodou nepadne nebo neco.
1.0
Pri vytvareni obj v MO, kdyz nevlozim tracking to padne - 1) Pokud tam neco napisu, tak to projde, ale to neni ucel. 2) Napadlo me ze pokud tam je nejaky tracking a ja importuji trackingy z Posty, tak co se stane? Zadouci by to bylo tracking prepsat tim novejsim z importu. Tak to tak zarid pokud se tak nedeje ci zda to nahodou nepadne nebo neco.
priority
pri vytvareni obj v mo kdyz nevlozim tracking to padne pokud tam neco napisu tak to projde ale to neni ucel napadlo me ze pokud tam je nejaky tracking a ja importuji trackingy z posty tak co se stane zadouci by to bylo tracking prepsat tim novejsim z importu tak to tak zarid pokud se tak nedeje ci zda to nahodou nepadne nebo neco
1
2,360
2,525,694,620
IssuesEvent
2015-01-21 04:18:55
Rapptz/jsonpp
https://api.github.com/repos/Rapptz/jsonpp
opened
to_string does not properly dump real JSON.
bug priority: high status: confirmed
Since it doesn't dump real JSON, it's probably not a good way to dump JSON. So it should probably be removed or renamed for a better API. The reason it doesn't dump real JSON is because when dumping strings, it ignores the `\uxxxx` literals and even the escape sequences which makes it wrong. It might be worth looking into how Python does JSON dumping. This allows it to be relevant to issue #5.
1.0
to_string does not properly dump real JSON. - Since it doesn't dump real JSON, it's probably not a good way to dump JSON. So it should probably be removed or renamed for a better API. The reason it doesn't dump real JSON is because when dumping strings, it ignores the `\uxxxx` literals and even the escape sequences which makes it wrong. It might be worth looking into how Python does JSON dumping. This allows it to be relevant to issue #5.
priority
to string does not properly dump real json since it doesn t dump real json it s probably not a good way to dump json so it should probably be removed or renamed for a better api the reason it doesn t dump real json is because when dumping strings it ignores the uxxxx literals and even the escape sequences which makes it wrong it might be worth looking into how python does json dumping this allows it to be relevant to issue
1
155,607
5,957,609,780
IssuesEvent
2017-05-29 03:23:21
quaich-serverless/quaich
https://api.github.com/repos/quaich-serverless/quaich
opened
Add handling of path based parameters
Priority: High Type: Feature
Currently dynamic path parameters are handled only via `resource`. That is placeholders in path (e.g. `/user/{userId}`) would only work if they are captured by HTTP api. As basic integration with HTTP API is via `catchAll` single resource (as described here: https://github.com/quaich-serverless/quartercask/issues/14 ) it would be useful to have also dynamic path parts working with such configuration (it essentially boils down to the fact that API gateway won't be giving us params in `pathParameters` param).
1.0
Add handling of path based parameters - Currently dynamic path parameters are handled only via `resource`. That is placeholders in path (e.g. `/user/{userId}`) would only work if they are captured by HTTP api. As basic integration with HTTP API is via `catchAll` single resource (as described here: https://github.com/quaich-serverless/quartercask/issues/14 ) it would be useful to have also dynamic path parts working with such configuration (it essentially boils down to the fact that API gateway won't be giving us params in `pathParameters` param).
priority
add handling of path based parameters currently dynamic path parameters are handled only via resource that is placeholders in path e g user userid would only work if they are captured by http api as basic integration with http api is via catchall single resource as described here it would be useful to have also dynamic path parts working with such configuration it essentially boils down to the fact that api gateway won t be giving us params in pathparameters param
1
411,311
12,016,754,585
IssuesEvent
2020-04-10 16:48:46
rbanffy/3270font
https://api.github.com/repos/rbanffy/3270font
closed
Unicode13 "symbols for legacy computing"
Missing glyph enhancement high priority
Unicode 13 adds a new block, at [U1FB00](https://www.unicode.org/charts/PDF/U1FB00.pdf), that seems to be close to 3270font's focus. Well, it's about 8-bit computers and teletext rather than mainframes, but it's still a good fit. The symbols are simple geometrical shapes, and would require little work per codepoint (but there's still a lot of them). Being the first (or one of the first) fonts to release them would get you a lot of new installs. ☺
1.0
Unicode13 "symbols for legacy computing" - Unicode 13 adds a new block, at [U1FB00](https://www.unicode.org/charts/PDF/U1FB00.pdf), that seems to be close to 3270font's focus. Well, it's about 8-bit computers and teletext rather than mainframes, but it's still a good fit. The symbols are simple geometrical shapes, and would require little work per codepoint (but there's still a lot of them). Being the first (or one of the first) fonts to release them would get you a lot of new installs. ☺
priority
symbols for legacy computing unicode adds a new block at that seems to be close to s focus well it s about bit computers and teletext rather than mainframes but it s still a good fit the symbols are simple geometrical shapes and would require little work per codepoint but there s still a lot of them being the first or one of the first fonts to release them would get you a lot of new installs ☺
1
166,417
6,304,320,334
IssuesEvent
2017-07-21 15:38:40
reymesson1/primerejercicio
https://api.github.com/repos/reymesson1/primerejercicio
closed
Make order from Deals Pizza
In Progress Priority: High
Priority: High Description: Add a field discount to each ingredients of the pizza. 200*15=3000 ~= 30.00 100*10=1000~=10.00 100*5=500~=5.00 200*2=400 ~=4.00
1.0
Make order from Deals Pizza - Priority: High Description: Add a field discount to each ingredients of the pizza. 200*15=3000 ~= 30.00 100*10=1000~=10.00 100*5=500~=5.00 200*2=400 ~=4.00
priority
make order from deals pizza priority high description add a field discount to each ingredients of the pizza
1
744,822
25,957,040,886
IssuesEvent
2022-12-18 11:26:04
dogukanarslan/software-development-practice
https://api.github.com/repos/dogukanarslan/software-development-practice
closed
Add timestamps
enhancement priority: medium effort: high state: pending
Currently, pins don't have `created_at` and `updated_at` fields. These fields should be listed. - [ ] Add `created_at` and `updated_at` fields to Pin model and User model - [ ] Show `created_at` on pin detail - [ ] Sort pind by newest to oldest on tables - [ ] Update existing records on database by finding their creation date on MongoDB
1.0
Add timestamps - Currently, pins don't have `created_at` and `updated_at` fields. These fields should be listed. - [ ] Add `created_at` and `updated_at` fields to Pin model and User model - [ ] Show `created_at` on pin detail - [ ] Sort pind by newest to oldest on tables - [ ] Update existing records on database by finding their creation date on MongoDB
priority
add timestamps currently pins don t have created at and updated at fields these fields should be listed add created at and updated at fields to pin model and user model show created at on pin detail sort pind by newest to oldest on tables update existing records on database by finding their creation date on mongodb
1
783,875
27,549,843,506
IssuesEvent
2023-03-07 14:17:16
AY2223S2-CS2103T-T12-2/tp
https://api.github.com/repos/AY2223S2-CS2103T-T12-2/tp
opened
Create basic info fields for Patient class
type.Task priority.High
To add the following fields for v1.2 (basic info): - [ ] NRIC (String) - [ ] Name (String) - [ ] Status (Enum) Remember to test inputs for the respective data types later
1.0
Create basic info fields for Patient class - To add the following fields for v1.2 (basic info): - [ ] NRIC (String) - [ ] Name (String) - [ ] Status (Enum) Remember to test inputs for the respective data types later
priority
create basic info fields for patient class to add the following fields for basic info nric string name string status enum remember to test inputs for the respective data types later
1
16,172
2,612,250,369
IssuesEvent
2015-02-27 13:09:48
Krasnyanskiy/jrsh
https://api.github.com/repos/Krasnyanskiy/jrsh
opened
As User I want to login in ssh style
feature high priority
#### Example 1 ```bash >>> jrsh superuser|organization_1@localhost:8080/jasperserver-pro ``` #### Example 2 ```bash >>> jrsh superuser|organization_1%@https://localhost:8080/jasperserver-pro ``` #### Example 3 ```bash >>> jrsh superuser%password@localhost:8080/jasperserver-pro ```
1.0
As User I want to login in ssh style - #### Example 1 ```bash >>> jrsh superuser|organization_1@localhost:8080/jasperserver-pro ``` #### Example 2 ```bash >>> jrsh superuser|organization_1%@https://localhost:8080/jasperserver-pro ``` #### Example 3 ```bash >>> jrsh superuser%password@localhost:8080/jasperserver-pro ```
priority
as user i want to login in ssh style example bash jrsh superuser organization localhost jasperserver pro example bash jrsh superuser organization example bash jrsh superuser password localhost jasperserver pro
1
580,310
17,226,348,387
IssuesEvent
2021-07-20 02:36:15
geolonia/app.geolonia.com
https://api.github.com/repos/geolonia/app.geolonia.com
closed
港区の区役所・総合支所 GeoJSONをアップロードした時に画面が真っ白になる
Priority: High bug
東京オープンデータカタログの港区の区役所・総合支所 GeoJSONを、GeoJSON APIにアップすると画面が真っ白になってしまう。 ### アップロードしたGeoJSON https://opendata.city.minato.tokyo.jp/dataset/74c06ebb-47dd-4fe1-8ba7-a5be60d2a448/resource/f1a1056b-a00e-4c12-8a78-288e0eee7ba0/download/minatokushisetsujoho_kuyakusyo.json ### スクリーンショット ![スクリーンショット 2021-06-25 11 29 53](https://user-images.githubusercontent.com/8760841/123360766-d5048f80-d5a8-11eb-8da3-cc21e92d7a58.png)
1.0
港区の区役所・総合支所 GeoJSONをアップロードした時に画面が真っ白になる - 東京オープンデータカタログの港区の区役所・総合支所 GeoJSONを、GeoJSON APIにアップすると画面が真っ白になってしまう。 ### アップロードしたGeoJSON https://opendata.city.minato.tokyo.jp/dataset/74c06ebb-47dd-4fe1-8ba7-a5be60d2a448/resource/f1a1056b-a00e-4c12-8a78-288e0eee7ba0/download/minatokushisetsujoho_kuyakusyo.json ### スクリーンショット ![スクリーンショット 2021-06-25 11 29 53](https://user-images.githubusercontent.com/8760841/123360766-d5048f80-d5a8-11eb-8da3-cc21e92d7a58.png)
priority
港区の区役所・総合支所 geojsonをアップロードした時に画面が真っ白になる 東京オープンデータカタログの港区の区役所・総合支所 geojsonを、geojson apiにアップすると画面が真っ白になってしまう。 アップロードしたgeojson スクリーンショット
1
677,135
23,152,474,846
IssuesEvent
2022-07-29 09:39:46
apache/dolphinscheduler
https://api.github.com/repos/apache/dolphinscheduler
closed
[Bug] [Workflow Definition] The network layout of the workflow is unusable, and the front end reports an error: h is not function, and the checked is: Full workflow
bug UI need to verify priority:high
### Search before asking - [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues. ### What happened ![image](https://user-images.githubusercontent.com/59044203/169323045-3178a969-dcef-422b-b44f-5b0a98af0ddc.png) ### What you expected to happen The network layout works normally ### How to reproduce 1.Go to the workflow network layout 2.Enter a row 1, Enter the column 1 3.Click OK ### Anything else _No response_ ### Version dev ### Are you willing to submit PR? - [ ] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
1.0
[Bug] [Workflow Definition] The network layout of the workflow is unusable, and the front end reports an error: h is not function, and the checked is: Full workflow - ### Search before asking - [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues. ### What happened ![image](https://user-images.githubusercontent.com/59044203/169323045-3178a969-dcef-422b-b44f-5b0a98af0ddc.png) ### What you expected to happen The network layout works normally ### How to reproduce 1.Go to the workflow network layout 2.Enter a row 1, Enter the column 1 3.Click OK ### Anything else _No response_ ### Version dev ### Are you willing to submit PR? - [ ] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
priority
the network layout of the workflow is unusable and the front end reports an error h is not function and the checked is full workflow search before asking i had searched in the and found no similar issues what happened what you expected to happen the network layout works normally how to reproduce go to the workflow network layout enter a row enter the column click ok anything else no response version dev are you willing to submit pr yes i am willing to submit a pr code of conduct i agree to follow this project s
1
506,273
14,661,237,644
IssuesEvent
2020-12-29 02:53:39
TeamSTEP/project-witch-one
https://api.github.com/repos/TeamSTEP/project-witch-one
closed
Refactor game manager object spawning
High Priority Hoon Kim add feature
## Required Tasks: - [x] remove `LoaderInEditor` prefab from all scenes (and the project) - [x] make all game manager scene spawned from the client before they join a game session - [x] allow the game manager to manager game scenes and sessions across the network ## Description: Currently, we need the `LoaderInEditor` prefab to be part of all the game scenes in order to spawn the game manager, camera, and the HUD. However, we want to remove all of this and make sure that the game scene only contains the tilemap grid (plus collision) and nothing else. Spawning all management related additive scenes before the room loading is easy, but there is a bug where this would only work on master clients but not for remote clients. This bug may be stemming from different network callbacks depending on the client type, we need to fix this so that all manager objects are loaded from the menu before the game starts. ## Difficulty 6/10 ## Estimated Implementation Time - Optimistic - 3 days - Normal - 1 week - Pessimistic - 2 weeks ## Work Start Date {when will the implementation will start}
1.0
Refactor game manager object spawning - ## Required Tasks: - [x] remove `LoaderInEditor` prefab from all scenes (and the project) - [x] make all game manager scene spawned from the client before they join a game session - [x] allow the game manager to manager game scenes and sessions across the network ## Description: Currently, we need the `LoaderInEditor` prefab to be part of all the game scenes in order to spawn the game manager, camera, and the HUD. However, we want to remove all of this and make sure that the game scene only contains the tilemap grid (plus collision) and nothing else. Spawning all management related additive scenes before the room loading is easy, but there is a bug where this would only work on master clients but not for remote clients. This bug may be stemming from different network callbacks depending on the client type, we need to fix this so that all manager objects are loaded from the menu before the game starts. ## Difficulty 6/10 ## Estimated Implementation Time - Optimistic - 3 days - Normal - 1 week - Pessimistic - 2 weeks ## Work Start Date {when will the implementation will start}
priority
refactor game manager object spawning required tasks remove loaderineditor prefab from all scenes and the project make all game manager scene spawned from the client before they join a game session allow the game manager to manager game scenes and sessions across the network description currently we need the loaderineditor prefab to be part of all the game scenes in order to spawn the game manager camera and the hud however we want to remove all of this and make sure that the game scene only contains the tilemap grid plus collision and nothing else spawning all management related additive scenes before the room loading is easy but there is a bug where this would only work on master clients but not for remote clients this bug may be stemming from different network callbacks depending on the client type we need to fix this so that all manager objects are loaded from the menu before the game starts difficulty estimated implementation time optimistic days normal week pessimistic weeks work start date when will the implementation will start
1
164,892
6,258,386,603
IssuesEvent
2017-07-14 15:27:56
fossasia/open-event-orga-server
https://api.github.com/repos/fossasia/open-event-orga-server
closed
IntegrityError: (psycopg2.IntegrityError) duplicate key value violates unique constraint "event_sub_topics_slug_key"
bug has-PR Priority: High
https://sentry.eventyay.com/eventyay/api/issues/16/ ``` IntegrityError: (psycopg2.IntegrityError) duplicate key value violates unique constraint "event_sub_topics_slug_key" DETAIL: Key (slug)=(other) already exists. [SQL: "INSERT INTO event_sub_topics(name, slug, event_topic_id) SELECT DISTINCT event_sub_topic_id, lower(replace(regexp_replace(event_sub_topic_id, '& |,', '', 'g'), ' ', '-')), event_topic_id FROM events where not exists (SELECT 1 FROM event_sub_topics where event_sub_topics.name=events.event_sub_topic_id) and event_sub_topic_id is not null"] (25 additional frame(s) were not displayed) ... File "sqlalchemy/engine/base.py", line 1189, in _execute_context context) File "sqlalchemy/engine/base.py", line 1402, in _handle_dbapi_exception exc_info File "sqlalchemy/util/compat.py", line 203, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "sqlalchemy/engine/base.py", line 1182, in _execute_context context) File "sqlalchemy/engine/default.py", line 470, in do_execute cursor.execute(statement, parameters) IntegrityError: (psycopg2.IntegrityError) duplicate key value violates unique constraint "event_sub_topics_slug_key" DETAIL: Key (slug)=(other) already exists. [SQL: "INSERT INTO event_sub_topics(name, slug, event_topic_id) SELECT DISTINCT event_sub_topic_id, lower(replace(regexp_replace(event_sub_topic_id, '& |,', '', 'g'), ' ', '-')), event_topic_id FROM events where not exists (SELECT 1 FROM event_sub_topics where event_sub_topics.name=events.event_sub_topic_id) and event_sub_topic_id is not null"] ``` > This occurred while running the migrations against a copy of the vintage production database The slug should be unique along with event_topic_id. (ie) a same slug=event_topic_id pair cannot exist. Eg. Valid entries: ``` slug = other event_topic_id = 1 slug = other event_topic_id = 2 ``` Invalid entries. ``` slug = other event_topic_id = 1 slug = other event_topic_id = 1 ```
1.0
IntegrityError: (psycopg2.IntegrityError) duplicate key value violates unique constraint "event_sub_topics_slug_key" - https://sentry.eventyay.com/eventyay/api/issues/16/ ``` IntegrityError: (psycopg2.IntegrityError) duplicate key value violates unique constraint "event_sub_topics_slug_key" DETAIL: Key (slug)=(other) already exists. [SQL: "INSERT INTO event_sub_topics(name, slug, event_topic_id) SELECT DISTINCT event_sub_topic_id, lower(replace(regexp_replace(event_sub_topic_id, '& |,', '', 'g'), ' ', '-')), event_topic_id FROM events where not exists (SELECT 1 FROM event_sub_topics where event_sub_topics.name=events.event_sub_topic_id) and event_sub_topic_id is not null"] (25 additional frame(s) were not displayed) ... File "sqlalchemy/engine/base.py", line 1189, in _execute_context context) File "sqlalchemy/engine/base.py", line 1402, in _handle_dbapi_exception exc_info File "sqlalchemy/util/compat.py", line 203, in raise_from_cause reraise(type(exception), exception, tb=exc_tb, cause=cause) File "sqlalchemy/engine/base.py", line 1182, in _execute_context context) File "sqlalchemy/engine/default.py", line 470, in do_execute cursor.execute(statement, parameters) IntegrityError: (psycopg2.IntegrityError) duplicate key value violates unique constraint "event_sub_topics_slug_key" DETAIL: Key (slug)=(other) already exists. [SQL: "INSERT INTO event_sub_topics(name, slug, event_topic_id) SELECT DISTINCT event_sub_topic_id, lower(replace(regexp_replace(event_sub_topic_id, '& |,', '', 'g'), ' ', '-')), event_topic_id FROM events where not exists (SELECT 1 FROM event_sub_topics where event_sub_topics.name=events.event_sub_topic_id) and event_sub_topic_id is not null"] ``` > This occurred while running the migrations against a copy of the vintage production database The slug should be unique along with event_topic_id. (ie) a same slug=event_topic_id pair cannot exist. Eg. Valid entries: ``` slug = other event_topic_id = 1 slug = other event_topic_id = 2 ``` Invalid entries. ``` slug = other event_topic_id = 1 slug = other event_topic_id = 1 ```
priority
integrityerror integrityerror duplicate key value violates unique constraint event sub topics slug key integrityerror integrityerror duplicate key value violates unique constraint event sub topics slug key detail key slug other already exists additional frame s were not displayed file sqlalchemy engine base py line in execute context context file sqlalchemy engine base py line in handle dbapi exception exc info file sqlalchemy util compat py line in raise from cause reraise type exception exception tb exc tb cause cause file sqlalchemy engine base py line in execute context context file sqlalchemy engine default py line in do execute cursor execute statement parameters integrityerror integrityerror duplicate key value violates unique constraint event sub topics slug key detail key slug other already exists this occurred while running the migrations against a copy of the vintage production database the slug should be unique along with event topic id ie a same slug event topic id pair cannot exist eg valid entries slug other event topic id slug other event topic id invalid entries slug other event topic id slug other event topic id
1
458,748
13,181,048,216
IssuesEvent
2020-08-12 13:45:38
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
0.9 issue with text in specialities interface
Category: Localization Priority: High Status: Fixed
open specialities (z) select any category, select any profession available to add star. Text about star is broken ![изображение](https://user-images.githubusercontent.com/4980243/89529018-acd73b00-d7f4-11ea-9515-746637423137.png) select any non-researched profession. Specialities names broken ![изображение](https://user-images.githubusercontent.com/4980243/89529059-bf517480-d7f4-11ea-8f04-2b11398fc4ed.png)
1.0
0.9 issue with text in specialities interface - open specialities (z) select any category, select any profession available to add star. Text about star is broken ![изображение](https://user-images.githubusercontent.com/4980243/89529018-acd73b00-d7f4-11ea-9515-746637423137.png) select any non-researched profession. Specialities names broken ![изображение](https://user-images.githubusercontent.com/4980243/89529059-bf517480-d7f4-11ea-8f04-2b11398fc4ed.png)
priority
issue with text in specialities interface open specialities z select any category select any profession available to add star text about star is broken select any non researched profession specialities names broken
1
99,177
4,049,037,267
IssuesEvent
2016-05-23 12:49:52
ow2-proactive/scheduling
https://api.github.com/repos/ow2-proactive/scheduling
closed
Low priority tasks can be scheduled over high priority ones
priority:high severity:major type:bug
Due to jobs locking mechanism, SchedulingMethodImpl can receive an incomplete number of jobs. In that case, a low priority job can have its tasks scheduled, while high priority jobs are simply locked during the current scheduling loop. Additionnaly the InternalPolicy filter can remove high-priority tasks from the list of tasks to be scheduled, making priorities irrelevant.
1.0
Low priority tasks can be scheduled over high priority ones - Due to jobs locking mechanism, SchedulingMethodImpl can receive an incomplete number of jobs. In that case, a low priority job can have its tasks scheduled, while high priority jobs are simply locked during the current scheduling loop. Additionnaly the InternalPolicy filter can remove high-priority tasks from the list of tasks to be scheduled, making priorities irrelevant.
priority
low priority tasks can be scheduled over high priority ones due to jobs locking mechanism schedulingmethodimpl can receive an incomplete number of jobs in that case a low priority job can have its tasks scheduled while high priority jobs are simply locked during the current scheduling loop additionnaly the internalpolicy filter can remove high priority tasks from the list of tasks to be scheduled making priorities irrelevant
1
792,876
27,976,223,556
IssuesEvent
2023-03-25 16:11:09
AY2223S2-CS2113-T13-2/tp
https://api.github.com/repos/AY2223S2-CS2113-T13-2/tp
opened
variable category never assigned in CommandCategory
type.Bug priority.High
![Screenshot 2023-03-26 at 00 01 21](https://user-images.githubusercontent.com/79389472/227728694-44464df9-00c2-46f9-8343-1e81536814a6.png) User input "category" will never enter the final else statement as `category` is not assigned. ![Screenshot 2023-03-26 at 00 09 39](https://user-images.githubusercontent.com/79389472/227728776-fa70841a-f13f-4dc4-8444-c443ede23bbf.png) The unassigned `category` also results in unintended output for the else statement.
1.0
variable category never assigned in CommandCategory - ![Screenshot 2023-03-26 at 00 01 21](https://user-images.githubusercontent.com/79389472/227728694-44464df9-00c2-46f9-8343-1e81536814a6.png) User input "category" will never enter the final else statement as `category` is not assigned. ![Screenshot 2023-03-26 at 00 09 39](https://user-images.githubusercontent.com/79389472/227728776-fa70841a-f13f-4dc4-8444-c443ede23bbf.png) The unassigned `category` also results in unintended output for the else statement.
priority
variable category never assigned in commandcategory user input category will never enter the final else statement as category is not assigned the unassigned category also results in unintended output for the else statement
1
1,673
2,517,447,705
IssuesEvent
2015-01-16 14:54:26
geosolutions-it/MapStore
https://api.github.com/repos/geosolutions-it/MapStore
closed
Bug between SpatialSelectorQueryForm and FeatureEditor
bug Priority: High
When the user sets a Region Of Interest (i.e. circle), performs a search and then uses the 'Edit Feature' action of the FeatureEditor plugin, the spatial selection control of the SpatialSelectorQueryForm plugin stays activated. So in this case a new spatial selection (and the consequence search) is performed before showing the editor's popup. To replicate the bug: 1) enable the SpatialSelectorQueryForm and FeatureEditor (the FeatureManager is needed) 2) Perform a search with a 'circle' selection (ROI) 3) Click on the toogle button Edit Feature 4) Click on a feature in the map in order to visualize the editor's popup.
1.0
Bug between SpatialSelectorQueryForm and FeatureEditor - When the user sets a Region Of Interest (i.e. circle), performs a search and then uses the 'Edit Feature' action of the FeatureEditor plugin, the spatial selection control of the SpatialSelectorQueryForm plugin stays activated. So in this case a new spatial selection (and the consequence search) is performed before showing the editor's popup. To replicate the bug: 1) enable the SpatialSelectorQueryForm and FeatureEditor (the FeatureManager is needed) 2) Perform a search with a 'circle' selection (ROI) 3) Click on the toogle button Edit Feature 4) Click on a feature in the map in order to visualize the editor's popup.
priority
bug between spatialselectorqueryform and featureeditor when the user sets a region of interest i e circle performs a search and then uses the edit feature action of the featureeditor plugin the spatial selection control of the spatialselectorqueryform plugin stays activated so in this case a new spatial selection and the consequence search is performed before showing the editor s popup to replicate the bug enable the spatialselectorqueryform and featureeditor the featuremanager is needed perform a search with a circle selection roi click on the toogle button edit feature click on a feature in the map in order to visualize the editor s popup
1
34,381
2,777,348,944
IssuesEvent
2015-05-05 07:24:10
punongbayan-araullo/tickets
https://api.github.com/repos/punongbayan-araullo/tickets
opened
Include the target companies and their related projects in the Company and contacts grid
priority - high status - accepted system - clients type - enhancement
Include the target companies and the related projects in the Company and contacts grid
1.0
Include the target companies and their related projects in the Company and contacts grid - Include the target companies and the related projects in the Company and contacts grid
priority
include the target companies and their related projects in the company and contacts grid include the target companies and the related projects in the company and contacts grid
1
347,836
10,435,002,938
IssuesEvent
2019-09-17 16:19:01
openstax/domo
https://api.github.com/repos/openstax/domo
opened
Content Management Q3 Tracking Time
High Priority - Need with 5 business days High priority
**Name:** Alina Slavik **Contact Emai:** ams14@rice.edu **Slack ID:** alina **Description:** Please create a pie chart showing where time was tracked for the following individuals between the following dates: July 1, 2019 - today. Alina Slavik Larissa Chu Ryan Stickney Marc Jeannott Theresa Masciale Colby Powers **How are you planning to use this data?:** Presentation for Leadership **Grant proposal or funding report:** No **Sensitive to individuals:** No **Due by:** 9/18/2019 **Who and how will data be shared:** This will be shared with the Leadership presentation on a slide at tomorrow's presentation. **Type of data:** Tracking Time **Visualization format:** pie chart **Primary OpenStax Team:** Content Production **Why do you need it by this date:** My presentation to leadership is this date. Due to a miscommunication, an earlier request for this did not yield the information needed. **Is the due date flexible?:** No **Primary OpenStax Team:** Content Production
2.0
Content Management Q3 Tracking Time - **Name:** Alina Slavik **Contact Emai:** ams14@rice.edu **Slack ID:** alina **Description:** Please create a pie chart showing where time was tracked for the following individuals between the following dates: July 1, 2019 - today. Alina Slavik Larissa Chu Ryan Stickney Marc Jeannott Theresa Masciale Colby Powers **How are you planning to use this data?:** Presentation for Leadership **Grant proposal or funding report:** No **Sensitive to individuals:** No **Due by:** 9/18/2019 **Who and how will data be shared:** This will be shared with the Leadership presentation on a slide at tomorrow's presentation. **Type of data:** Tracking Time **Visualization format:** pie chart **Primary OpenStax Team:** Content Production **Why do you need it by this date:** My presentation to leadership is this date. Due to a miscommunication, an earlier request for this did not yield the information needed. **Is the due date flexible?:** No **Primary OpenStax Team:** Content Production
priority
content management tracking time name alina slavik contact emai rice edu slack id alina description please create a pie chart showing where time was tracked for the following individuals between the following dates july today alina slavik larissa chu ryan stickney marc jeannott theresa masciale colby powers how are you planning to use this data presentation for leadership grant proposal or funding report no sensitive to individuals no due by who and how will data be shared this will be shared with the leadership presentation on a slide at tomorrow s presentation type of data tracking time visualization format pie chart primary openstax team content production why do you need it by this date my presentation to leadership is this date due to a miscommunication an earlier request for this did not yield the information needed is the due date flexible no primary openstax team content production
1
85,845
3,699,388,497
IssuesEvent
2016-02-28 22:49:18
uEssentials/uEssentials
https://api.github.com/repos/uEssentials/uEssentials
closed
Memory leak detected.
high-priority unverified bug
> Time Description (additional details on these messages) > 08:24:03 PM The application has been restarted (run #4). > 08:24:02 PM Run Statistics: 2016/02/21 19:24:47 - 2016/02/21 20:24:00 (59.2 minutes) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 313MB / 412MB > Warning 08:24:02 PM AlwaysUpService has detected that the application has stopped. > 07:24:41 PM The application has been restarted (run #3). > 07:24:40 PM Run Statistics: 2016/02/21 16:17:23 - 2016/02/21 19:24:33 (3.1 hours) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 354MB / 399MB > Warning 07:24:40 PM AlwaysUpService has detected that the application has stopped. > 04:17:17 PM The application has been restarted (run #2). > 04:17:16 PM Run Statistics: 2016/02/21 12:30:54 - 2016/02/21 16:17:10 (3.8 hours) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 374MB / 453MB > Warning 04:17:16 PM AlwaysUpService has detected that the application has stopped. > 12:30:46 PM The application has been started. > 12:30:35 PM The service has started. > 12:26:56 PM The service has stopped. > 12:26:56 PM The application has been stopped. > 12:26:55 PM Run Statistics: 2016/02/21 10:34:31 - 2016/02/21 12:26:53 (1.9 hours) CPU (Avg / Max): 26% / 37% Memory (Avg / Max): 325MB / 365MB > 10:34:25 AM The application has been restarted (run #2). > 10:34:24 AM Run Statistics: 2016/02/21 06:47:47 - 2016/02/21 10:34:22 (3.8 hours) CPU (Avg / Max): 25% / 30% Memory (Avg / Max): 317MB / 390MB > Warning 10:34:24 AM AlwaysUpService has detected that the application has stopped. > 06:47:39 AM The application has been started. > 06:47:27 AM The service has started (at boot). > 06:45:05 AM The service has stopped. > 06:45:05 AM The application has been stopped. > 06:45:03 AM Run Statistics: 2016/02/21 03:30:17 - 2016/02/21 06:44:58 (3.2 hours) CPU (Avg / Max): 25% / 28% Memory (Avg / Max): 301MB / 341MB > 03:30:11 AM The application has been restarted (run #7). > 03:30:10 AM Run-Day Statistics: 2016/02/21 00:00:01 - 2016/02/21 03:30:05 (3.5 hours) CPU (Avg / Max): 25% / 30% Memory (Avg / Max): 337MB / 380MB > 03:30:10 AM Run Statistics: 2016/02/20 23:44:13 - 2016/02/21 03:30:05 (3.8 hours) CPU (Avg / Max): 25% / 30% Memory (Avg / Max): 333MB / 380MB > Warning 03:30:10 AM AlwaysUpService has detected that the application has stopped. > 12:00:01 AM Run-Day Statistics: 2016/02/20 23:44:13 - 2016/02/20 23:59:55 (15.7 minutes) CPU (Avg / Max): 23% / 26% Memory (Avg / Max): 277MB / 291MB > > Saturday 02/20/16 > > Time Description (additional details on these messages) > 11:44:07 PM The application has been restarted (run #6). > 11:44:06 PM Run Statistics: 2016/02/20 21:15:04 - 2016/02/20 23:44:02 (2.5 hours) CPU (Avg / Max): 25% / 39% Memory (Avg / Max): 328MB / 365MB > Warning 11:44:06 PM AlwaysUpService has detected that the application has stopped. > 09:14:58 PM The application has been restarted (run #5). > 09:14:57 PM Run Statistics: 2016/02/20 17:58:12 - 2016/02/20 21:14:51 (3.3 hours) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 330MB / 388MB > Warning 09:14:57 PM AlwaysUpService has detected that the application has stopped. > 05:58:07 PM The application has been restarted (run #4). > 05:58:06 PM Run Statistics: 2016/02/20 17:47:32 - 2016/02/20 17:58:00 (10.5 minutes) CPU (Avg / Max): 25% / 28% Memory (Avg / Max): 317MB / 327MB > Warning 05:58:06 PM AlwaysUpService has detected that the application has stopped. > 05:47:26 PM The application has been restarted (run #3). > 05:47:25 PM Run Statistics: 2016/02/20 15:53:59 - 2016/02/20 17:47:23 (1.9 hours) CPU (Avg / Max): 28% / 34% Memory (Avg / Max): 361MB / 399MB > Warning 05:47:25 PM AlwaysUpService has detected that the application has stopped. > 03:53:53 PM The application has been restarted (run #2). > 03:53:52 PM Run Statistics: 2016/02/20 12:59:00 - 2016/02/20 15:53:46 (2.9 hours) CPU (Avg / Max): 28% / 35% Memory (Avg / Max): 371MB / 420MB > Warning 03:53:52 PM AlwaysUpService has detected that the application has stopped. > 12:58:52 PM The application has been started. > 12:58:41 PM The service has started. > 12:58:34 PM The service has stopped. > 12:58:34 PM The application has been stopped. > 12:58:32 PM Run Statistics: 2016/02/20 12:54:33 - 2016/02/20 12:58:28 (3.9 minutes) CPU (Avg / Max): 17% / 24% Memory (Avg / Max): 317MB / 327MB > 12:54:25 PM The application has been started. > 12:54:14 PM The service has started. > 12:53:34 PM The service has stopped. > 12:53:34 PM The application has been stopped. > 12:53:33 PM Run Statistics: 2016/02/20 10:34:42 - 2016/02/20 12:53:27 (2.3 hours) CPU (Avg / Max): 28% / 34% Memory (Avg / Max): 384MB / 446MB > 10:34:35 AM The application has been restarted (run #2). > 10:34:34 AM Run Statistics: 2016/02/20 06:47:58 - 2016/02/20 10:34:28 (3.8 hours) CPU (Avg / Max): 21% / 32% Memory (Avg / Max): 360MB / 432MB > Warning 10:34:34 AM AlwaysUpService has detected that the application has stopped. > 06:47:50 AM The application has been started. > 06:47:38 AM The service has started (at boot). > 06:45:04 AM The service has stopped. > 06:45:04 AM The application has been stopped. > 06:45:03 AM Run Statistics: 2016/02/20 05:14:50 - 2016/02/20 06:45:02 (1.5 hours) CPU (Avg / Max): 13% / 21% Memory (Avg / Max): 327MB / 347MB > 05:14:44 AM The application has been restarted (run #5). > 05:14:43 AM Run Statistics: 2016/02/20 01:28:25 - 2016/02/20 05:14:37 (3.8 hours) CPU (Avg / Max): 17% / 28% Memory (Avg / Max): 356MB / 397MB > Warning 05:14:43 AM AlwaysUpService has detected that the application has stopped. > 01:28:19 AM The application has been restarted (run #4). > 01:28:18 AM Run-Day Statistics: 2016/02/20 00:00:04 - 2016/02/20 01:28:16 (1.5 hours) CPU (Avg / Max): 30% / 34% Memory (Avg / Max): 512MB / 534MB > 01:28:18 AM Run Statistics: 2016/02/19 19:29:09 - 2016/02/20 01:28:16 (6.0 hours) CPU (Avg / Max): 29% / 35% Memory (Avg / Max): 440MB / 534MB > Warning 01:28:18 AM AlwaysUpService has detected that the application has stopped. > 12:00:04 AM Run-Day Statistics: 2016/02/19 19:29:09 - 2016/02/19 23:59:58 (4.5 hours) CPU (Avg / Max): 28% / 35% Memory (Avg / Max): 416MB / 494MB removed u essentials and its the complete opposite. i will attach my logs to here. I am labeling this as a high priority as this has caused some issues. I really want to continue to use this mod but by the looks of it and the direction its going I wont be able to do it until you solve it. Ill attach my logs from what my server has logged.
1.0
Memory leak detected. - > Time Description (additional details on these messages) > 08:24:03 PM The application has been restarted (run #4). > 08:24:02 PM Run Statistics: 2016/02/21 19:24:47 - 2016/02/21 20:24:00 (59.2 minutes) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 313MB / 412MB > Warning 08:24:02 PM AlwaysUpService has detected that the application has stopped. > 07:24:41 PM The application has been restarted (run #3). > 07:24:40 PM Run Statistics: 2016/02/21 16:17:23 - 2016/02/21 19:24:33 (3.1 hours) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 354MB / 399MB > Warning 07:24:40 PM AlwaysUpService has detected that the application has stopped. > 04:17:17 PM The application has been restarted (run #2). > 04:17:16 PM Run Statistics: 2016/02/21 12:30:54 - 2016/02/21 16:17:10 (3.8 hours) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 374MB / 453MB > Warning 04:17:16 PM AlwaysUpService has detected that the application has stopped. > 12:30:46 PM The application has been started. > 12:30:35 PM The service has started. > 12:26:56 PM The service has stopped. > 12:26:56 PM The application has been stopped. > 12:26:55 PM Run Statistics: 2016/02/21 10:34:31 - 2016/02/21 12:26:53 (1.9 hours) CPU (Avg / Max): 26% / 37% Memory (Avg / Max): 325MB / 365MB > 10:34:25 AM The application has been restarted (run #2). > 10:34:24 AM Run Statistics: 2016/02/21 06:47:47 - 2016/02/21 10:34:22 (3.8 hours) CPU (Avg / Max): 25% / 30% Memory (Avg / Max): 317MB / 390MB > Warning 10:34:24 AM AlwaysUpService has detected that the application has stopped. > 06:47:39 AM The application has been started. > 06:47:27 AM The service has started (at boot). > 06:45:05 AM The service has stopped. > 06:45:05 AM The application has been stopped. > 06:45:03 AM Run Statistics: 2016/02/21 03:30:17 - 2016/02/21 06:44:58 (3.2 hours) CPU (Avg / Max): 25% / 28% Memory (Avg / Max): 301MB / 341MB > 03:30:11 AM The application has been restarted (run #7). > 03:30:10 AM Run-Day Statistics: 2016/02/21 00:00:01 - 2016/02/21 03:30:05 (3.5 hours) CPU (Avg / Max): 25% / 30% Memory (Avg / Max): 337MB / 380MB > 03:30:10 AM Run Statistics: 2016/02/20 23:44:13 - 2016/02/21 03:30:05 (3.8 hours) CPU (Avg / Max): 25% / 30% Memory (Avg / Max): 333MB / 380MB > Warning 03:30:10 AM AlwaysUpService has detected that the application has stopped. > 12:00:01 AM Run-Day Statistics: 2016/02/20 23:44:13 - 2016/02/20 23:59:55 (15.7 minutes) CPU (Avg / Max): 23% / 26% Memory (Avg / Max): 277MB / 291MB > > Saturday 02/20/16 > > Time Description (additional details on these messages) > 11:44:07 PM The application has been restarted (run #6). > 11:44:06 PM Run Statistics: 2016/02/20 21:15:04 - 2016/02/20 23:44:02 (2.5 hours) CPU (Avg / Max): 25% / 39% Memory (Avg / Max): 328MB / 365MB > Warning 11:44:06 PM AlwaysUpService has detected that the application has stopped. > 09:14:58 PM The application has been restarted (run #5). > 09:14:57 PM Run Statistics: 2016/02/20 17:58:12 - 2016/02/20 21:14:51 (3.3 hours) CPU (Avg / Max): 26% / 31% Memory (Avg / Max): 330MB / 388MB > Warning 09:14:57 PM AlwaysUpService has detected that the application has stopped. > 05:58:07 PM The application has been restarted (run #4). > 05:58:06 PM Run Statistics: 2016/02/20 17:47:32 - 2016/02/20 17:58:00 (10.5 minutes) CPU (Avg / Max): 25% / 28% Memory (Avg / Max): 317MB / 327MB > Warning 05:58:06 PM AlwaysUpService has detected that the application has stopped. > 05:47:26 PM The application has been restarted (run #3). > 05:47:25 PM Run Statistics: 2016/02/20 15:53:59 - 2016/02/20 17:47:23 (1.9 hours) CPU (Avg / Max): 28% / 34% Memory (Avg / Max): 361MB / 399MB > Warning 05:47:25 PM AlwaysUpService has detected that the application has stopped. > 03:53:53 PM The application has been restarted (run #2). > 03:53:52 PM Run Statistics: 2016/02/20 12:59:00 - 2016/02/20 15:53:46 (2.9 hours) CPU (Avg / Max): 28% / 35% Memory (Avg / Max): 371MB / 420MB > Warning 03:53:52 PM AlwaysUpService has detected that the application has stopped. > 12:58:52 PM The application has been started. > 12:58:41 PM The service has started. > 12:58:34 PM The service has stopped. > 12:58:34 PM The application has been stopped. > 12:58:32 PM Run Statistics: 2016/02/20 12:54:33 - 2016/02/20 12:58:28 (3.9 minutes) CPU (Avg / Max): 17% / 24% Memory (Avg / Max): 317MB / 327MB > 12:54:25 PM The application has been started. > 12:54:14 PM The service has started. > 12:53:34 PM The service has stopped. > 12:53:34 PM The application has been stopped. > 12:53:33 PM Run Statistics: 2016/02/20 10:34:42 - 2016/02/20 12:53:27 (2.3 hours) CPU (Avg / Max): 28% / 34% Memory (Avg / Max): 384MB / 446MB > 10:34:35 AM The application has been restarted (run #2). > 10:34:34 AM Run Statistics: 2016/02/20 06:47:58 - 2016/02/20 10:34:28 (3.8 hours) CPU (Avg / Max): 21% / 32% Memory (Avg / Max): 360MB / 432MB > Warning 10:34:34 AM AlwaysUpService has detected that the application has stopped. > 06:47:50 AM The application has been started. > 06:47:38 AM The service has started (at boot). > 06:45:04 AM The service has stopped. > 06:45:04 AM The application has been stopped. > 06:45:03 AM Run Statistics: 2016/02/20 05:14:50 - 2016/02/20 06:45:02 (1.5 hours) CPU (Avg / Max): 13% / 21% Memory (Avg / Max): 327MB / 347MB > 05:14:44 AM The application has been restarted (run #5). > 05:14:43 AM Run Statistics: 2016/02/20 01:28:25 - 2016/02/20 05:14:37 (3.8 hours) CPU (Avg / Max): 17% / 28% Memory (Avg / Max): 356MB / 397MB > Warning 05:14:43 AM AlwaysUpService has detected that the application has stopped. > 01:28:19 AM The application has been restarted (run #4). > 01:28:18 AM Run-Day Statistics: 2016/02/20 00:00:04 - 2016/02/20 01:28:16 (1.5 hours) CPU (Avg / Max): 30% / 34% Memory (Avg / Max): 512MB / 534MB > 01:28:18 AM Run Statistics: 2016/02/19 19:29:09 - 2016/02/20 01:28:16 (6.0 hours) CPU (Avg / Max): 29% / 35% Memory (Avg / Max): 440MB / 534MB > Warning 01:28:18 AM AlwaysUpService has detected that the application has stopped. > 12:00:04 AM Run-Day Statistics: 2016/02/19 19:29:09 - 2016/02/19 23:59:58 (4.5 hours) CPU (Avg / Max): 28% / 35% Memory (Avg / Max): 416MB / 494MB removed u essentials and its the complete opposite. i will attach my logs to here. I am labeling this as a high priority as this has caused some issues. I really want to continue to use this mod but by the looks of it and the direction its going I wont be able to do it until you solve it. Ill attach my logs from what my server has logged.
priority
memory leak detected time description additional details on these messages pm the application has been restarted run pm run statistics minutes cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been restarted run pm run statistics hours cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been restarted run pm run statistics hours cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been started pm the service has started pm the service has stopped pm the application has been stopped pm run statistics hours cpu avg max memory avg max am the application has been restarted run am run statistics hours cpu avg max memory avg max warning am alwaysupservice has detected that the application has stopped am the application has been started am the service has started at boot am the service has stopped am the application has been stopped am run statistics hours cpu avg max memory avg max am the application has been restarted run am run day statistics hours cpu avg max memory avg max am run statistics hours cpu avg max memory avg max warning am alwaysupservice has detected that the application has stopped am run day statistics minutes cpu avg max memory avg max saturday time description additional details on these messages pm the application has been restarted run pm run statistics hours cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been restarted run pm run statistics hours cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been restarted run pm run statistics minutes cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been restarted run pm run statistics hours cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been restarted run pm run statistics hours cpu avg max memory avg max warning pm alwaysupservice has detected that the application has stopped pm the application has been started pm the service has started pm the service has stopped pm the application has been stopped pm run statistics minutes cpu avg max memory avg max pm the application has been started pm the service has started pm the service has stopped pm the application has been stopped pm run statistics hours cpu avg max memory avg max am the application has been restarted run am run statistics hours cpu avg max memory avg max warning am alwaysupservice has detected that the application has stopped am the application has been started am the service has started at boot am the service has stopped am the application has been stopped am run statistics hours cpu avg max memory avg max am the application has been restarted run am run statistics hours cpu avg max memory avg max warning am alwaysupservice has detected that the application has stopped am the application has been restarted run am run day statistics hours cpu avg max memory avg max am run statistics hours cpu avg max memory avg max warning am alwaysupservice has detected that the application has stopped am run day statistics hours cpu avg max memory avg max removed u essentials and its the complete opposite i will attach my logs to here i am labeling this as a high priority as this has caused some issues i really want to continue to use this mod but by the looks of it and the direction its going i wont be able to do it until you solve it ill attach my logs from what my server has logged
1
286,716
8,792,066,421
IssuesEvent
2018-12-21 14:56:07
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
GLES2 Light2D fail compiling shader ("lowp" undefined)
bug high priority topic:rendering
**Godot version:** f3cb236f9d0e559e759dabc7cb5b6c1368d45e6e **OS:** macOS, Intel HD 5000 Light2D is still not working with GLES2. I've tested with latest commit(f3cb236f9d0e559e759dabc7cb5b6c1368d45e6e) and got below error. ``` 0: /Users/lazywierdo/godot/bin/godot.osx.opt.tools.64 1: --path 2: /Users/lazywierdo/godot-test-projects/light2d-gles2 3: -e 4: res://main.tscn Current path: /Users/lazywierdo/godot-test-projects/light2d-gles2 OpenGL ES 2.0 Renderer: Intel HD Graphics 5000 OpenGL Engine ERROR: initialize: Directional shadow framebuffer status invalid At: drivers/gles2/rasterizer_scene_gles2.cpp:3178. 1: #version 120 2: #define USE_GLES_OVER_GL 3: #define USE_LIGHTING 4: 5: #ifdef USE_GLES_OVER_GL 6: #define mediump 7: #define highp 8: #else 9: precision mediump float; 10: precision mediump int; 11: #endif 12: 13: uniform sampler2D color_texture; // texunit:-1 14: /* clang-format on */ 15: uniform highp vec2 color_texpixel_size; 16: uniform mediump sampler2D normal_texture; // texunit:-2 17: 18: varying mediump vec2 uv_interp; 19: varying mediump vec4 color_interp; 20: 21: uniform highp float time; 22: 23: uniform vec4 final_modulate; 24: 25: #ifdef SCREEN_TEXTURE_USED 26: 27: uniform sampler2D screen_texture; // texunit:-3 28: 29: #endif 30: 31: #ifdef SCREEN_UV_USED 32: 33: uniform vec2 screen_pixel_size; 34: 35: #endif 36: 37: #ifdef USE_LIGHTING 38: 39: uniform highp mat4 light_matrix; 40: uniform highp mat4 light_local_matrix; 41: uniform highp mat4 shadow_matrix; 42: uniform highp vec4 light_color; 43: uniform highp vec4 light_shadow_color; 44: uniform highp vec2 light_pos; 45: uniform highp float shadowpixel_size; 46: uniform highp float shadow_gradient; 47: uniform highp float light_height; 48: uniform highp float light_outside_alpha; 49: uniform highp float shadow_distance_mult; 50: 51: uniform lowp sampler2D light_texture; // texunit:-3 52: varying vec4 light_uv_interp; 53: varying vec2 transformed_light_uv; 54: 55: varying vec4 local_rot; 56: 57: #ifdef USE_SHADOWS 58: 59: uniform highp sampler2D shadow_texture; // texunit:-4 60: varying highp vec2 pos; 61: 62: #endif 63: 64: const bool at_light_pass = true; 65: #else 66: const bool at_light_pass = false; 67: #endif 68: 69: uniform bool use_default_normal; 70: 71: 72: /* clang-format off */ 73: 74: 75: /* clang-format on */ 76: 77: void main() { 78: 79: vec4 color = color_interp; 80: 81: #if !defined(COLOR_USED) 82: //default behavior, texture by color 83: color *= texture2D(color_texture, uv_interp); 84: #endif 85: 86: #ifdef SCREEN_UV_USED 87: vec2 screen_uv = gl_FragCoord.xy * screen_pixel_size; 88: #endif 89: 90: vec3 normal; 91: 92: #if defined(NORMAL_USED) 93: 94: bool normal_used = true; 95: #else 96: bool normal_used = false; 97: #endif 98: 99: if (use_default_normal) { 100: normal.xy = texture2D(normal_texture, uv_interp).xy * 2.0 - 1.0; 101: normal.z = sqrt(1.0 - dot(normal.xy, normal.xy)); 102: normal_used = true; 103: } else { 104: normal = vec3(0.0, 0.0, 1.0); 105: } 106: 107: { 108: float normal_depth = 1.0; 109: 110: #if defined(NORMALMAP_USED) 111: vec3 normal_map = vec3(0.0, 0.0, 1.0); 112: normal_used = true; 113: #endif 114: 115: /* clang-format off */ 116: 117: 118: /* clang-format on */ 119: 120: #if defined(NORMALMAP_USED) 121: normal = mix(vec3(0.0, 0.0, 1.0), normal_map * vec3(2.0, -2.0, 1.0) - vec3(1.0, -1.0, 0.0), normal_depth); 122: #endif 123: } 124: color *= final_modulate; 125: 126: #ifdef USE_LIGHTING 127: 128: vec2 light_vec = transformed_light_uv; 129: 130: if (normal_used) { 131: normal.xy = mat2(local_rot.xy, local_rot.zw) * normal.xy; 132: } 133: 134: float att = 1.0; 135: 136: vec2 light_uv = light_uv_interp.xy; 137: vec4 light = texture2D(light_texture, light_uv); 138: 139: if (any(lessThan(light_uv_interp.xy, vec2(0.0, 0.0))) || any(greaterThanEqual(light_uv_interp.xy, vec2(1.0, 1.0)))) { 140: color.a *= light_outside_alpha; //invisible 141: 142: } else { 143: float real_light_height = light_height; 144: vec4 real_light_color = light_color; 145: vec4 real_light_shadow_color = light_shadow_color; 146: 147: #if defined(USE_LIGHT_SHADER_CODE) 148: //light is written by the light shader 149: light_compute( 150: light, 151: light_vec, 152: real_light_height, 153: real_light_color, 154: light_uv, 155: real_light_shadow_color, 156: normal, 157: uv, 158: #if defined(SCREEN_UV_USED) 159: screen_uv, 160: #endif 161: color); 162: #endif 163: 164: light *= real_light_color; 165: 166: if (normal_used) { 167: vec3 light_normal = normalize(vec3(light_vec, -real_light_height)); 168: light *= max(dot(-light_normal, normal), 0.0); 169: } 170: 171: color *= light; 172: 173: #ifdef USE_SHADOWS 174: light_vec = light_uv_interp.zw; //for shadows 175: float angle_to_light = -atan(light_vec.x, light_vec.y); 176: float PI = 3.14159265358979323846264; 177: /*int i = int(mod(floor((angle_to_light+7.0*PI/6.0)/(4.0*PI/6.0))+1.0, 3.0)); // +1 pq os indices estao em ordem 2,0,1 nos arrays 178: float ang*/ 179: 180: float su, sz; 181: 182: float abs_angle = abs(angle_to_light); 183: vec2 point; 184: float sh; 185: if (abs_angle < 45.0 * PI / 180.0) { 186: point = light_vec; 187: sh = 0.0 + (1.0 / 8.0); 188: } else if (abs_angle > 135.0 * PI / 180.0) { 189: point = -light_vec; 190: sh = 0.5 + (1.0 / 8.0); 191: } else if (angle_to_light > 0.0) { 192: 193: point = vec2(light_vec.y, -light_vec.x); 194: sh = 0.25 + (1.0 / 8.0); 195: } else { 196: 197: point = vec2(-light_vec.y, light_vec.x); 198: sh = 0.75 + (1.0 / 8.0); 199: } 200: 201: highp vec4 s = shadow_matrix * vec4(point, 0.0, 1.0); 202: s.xyz /= s.w; 203: su = s.x * 0.5 + 0.5; 204: sz = s.z * 0.5 + 0.5; 205: //sz=lightlength(light_vec); 206: 207: highp float shadow_attenuation = 0.0; 208: 209: #ifdef USE_RGBA_SHADOWS 210: 211: #define SHADOW_DEPTH(m_tex, m_uv) dot(texture2D((m_tex), (m_uv)), vec4(1.0 / (256.0 * 256.0 * 256.0), 1.0 / (256.0 * 256.0), 1.0 / 256.0, 1)) 212: 213: #else 214: 215: #define SHADOW_DEPTH(m_tex, m_uv) (texture2D((m_tex), (m_uv)).r) 216: 217: #endif 218: 219: #ifdef SHADOW_USE_GRADIENT 220: 221: #define SHADOW_TEST(m_ofs) \ 222: { \ 223: highp float sd = SHADOW_DEPTH(shadow_texture, vec2(m_ofs, sh)); \ 224: shadow_attenuation += 1.0 - smoothstep(sd, sd + shadow_gradient, sz); \ 225: } 226: 227: #else 228: 229: #define SHADOW_TEST(m_ofs) \ 230: { \ 231: highp float sd = SHADOW_DEPTH(shadow_texture, vec2(m_ofs, sh)); \ 232: shadow_attenuation += step(sz, sd); \ 233: } 234: 235: #endif 236: 237: #ifdef SHADOW_FILTER_NEAREST 238: 239: SHADOW_TEST(su); 240: 241: #endif 242: 243: #ifdef SHADOW_FILTER_PCF3 244: 245: SHADOW_TEST(su + shadowpixel_size); 246: SHADOW_TEST(su); 247: SHADOW_TEST(su - shadowpixel_size); 248: shadow_attenuation /= 3.0; 249: 250: #endif 251: 252: #ifdef SHADOW_FILTER_PCF5 253: 254: SHADOW_TEST(su + shadowpixel_size * 2.0); 255: SHADOW_TEST(su + shadowpixel_size); 256: SHADOW_TEST(su); 257: SHADOW_TEST(su - shadowpixel_size); 258: SHADOW_TEST(su - shadowpixel_size * 2.0); 259: shadow_attenuation /= 5.0; 260: 261: #endif 262: 263: #ifdef SHADOW_FILTER_PCF7 264: 265: SHADOW_TEST(su + shadowpixel_size * 3.0); 266: SHADOW_TEST(su + shadowpixel_size * 2.0); 267: SHADOW_TEST(su + shadowpixel_size); 268: SHADOW_TEST(su); 269: SHADOW_TEST(su - shadowpixel_size); 270: SHADOW_TEST(su - shadowpixel_size * 2.0); 271: SHADOW_TEST(su - shadowpixel_size * 3.0); 272: shadow_attenuation /= 7.0; 273: 274: #endif 275: 276: #ifdef SHADOW_FILTER_PCF9 277: 278: SHADOW_TEST(su + shadowpixel_size * 4.0); 279: SHADOW_TEST(su + shadowpixel_size * 3.0); 280: SHADOW_TEST(su + shadowpixel_size * 2.0); 281: SHADOW_TEST(su + shadowpixel_size); 282: SHADOW_TEST(su); 283: SHADOW_TEST(su - shadowpixel_size); 284: SHADOW_TEST(su - shadowpixel_size * 2.0); 285: SHADOW_TEST(su - shadowpixel_size * 3.0); 286: SHADOW_TEST(su - shadowpixel_size * 4.0); 287: shadow_attenuation /= 9.0; 288: 289: #endif 290: 291: #ifdef SHADOW_FILTER_PCF13 292: 293: SHADOW_TEST(su + shadowpixel_size * 6.0); 294: SHADOW_TEST(su + shadowpixel_size * 5.0); 295: SHADOW_TEST(su + shadowpixel_size * 4.0); 296: SHADOW_TEST(su + shadowpixel_size * 3.0); 297: SHADOW_TEST(su + shadowpixel_size * 2.0); 298: SHADOW_TEST(su + shadowpixel_size); 299: SHADOW_TEST(su); 300: SHADOW_TEST(su - shadowpixel_size); 301: SHADOW_TEST(su - shadowpixel_size * 2.0); 302: SHADOW_TEST(su - shadowpixel_size * 3.0); 303: SHADOW_TEST(su - shadowpixel_size * 4.0); 304: SHADOW_TEST(su - shadowpixel_size * 5.0); 305: SHADOW_TEST(su - shadowpixel_size * 6.0); 306: shadow_attenuation /= 13.0; 307: 308: #endif 309: 310: //color *= shadow_attenuation; 311: color = mix(real_light_shadow_color, color, shadow_attenuation); 312: //use shadows 313: #endif 314: } 315: 316: //use lighting 317: #endif 318: 319: gl_FragColor = color; 320: } 321: ERROR: _display_error_with_code: CanvasShaderGLES2: Fragment shader compilation failed: ERROR: 0:51: 'lowp' : Reserved word. ERROR: 0:51: 'lowp' : syntax error: syntax error At: drivers/gles2/shader_gles2.cpp:196. ERROR: get_current_version: Method/Function Failed, returning: __null At: drivers/gles2/shader_gles2.cpp:426. ERROR: bind: Condition ' !version ' is true. returned: false At: drivers/gles2/shader_gles2.cpp:133. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ``` _Originally posted by @kiidmale in https://github.com/godotengine/godot/issues/21853#issuecomment-449328028_
1.0
GLES2 Light2D fail compiling shader ("lowp" undefined) - **Godot version:** f3cb236f9d0e559e759dabc7cb5b6c1368d45e6e **OS:** macOS, Intel HD 5000 Light2D is still not working with GLES2. I've tested with latest commit(f3cb236f9d0e559e759dabc7cb5b6c1368d45e6e) and got below error. ``` 0: /Users/lazywierdo/godot/bin/godot.osx.opt.tools.64 1: --path 2: /Users/lazywierdo/godot-test-projects/light2d-gles2 3: -e 4: res://main.tscn Current path: /Users/lazywierdo/godot-test-projects/light2d-gles2 OpenGL ES 2.0 Renderer: Intel HD Graphics 5000 OpenGL Engine ERROR: initialize: Directional shadow framebuffer status invalid At: drivers/gles2/rasterizer_scene_gles2.cpp:3178. 1: #version 120 2: #define USE_GLES_OVER_GL 3: #define USE_LIGHTING 4: 5: #ifdef USE_GLES_OVER_GL 6: #define mediump 7: #define highp 8: #else 9: precision mediump float; 10: precision mediump int; 11: #endif 12: 13: uniform sampler2D color_texture; // texunit:-1 14: /* clang-format on */ 15: uniform highp vec2 color_texpixel_size; 16: uniform mediump sampler2D normal_texture; // texunit:-2 17: 18: varying mediump vec2 uv_interp; 19: varying mediump vec4 color_interp; 20: 21: uniform highp float time; 22: 23: uniform vec4 final_modulate; 24: 25: #ifdef SCREEN_TEXTURE_USED 26: 27: uniform sampler2D screen_texture; // texunit:-3 28: 29: #endif 30: 31: #ifdef SCREEN_UV_USED 32: 33: uniform vec2 screen_pixel_size; 34: 35: #endif 36: 37: #ifdef USE_LIGHTING 38: 39: uniform highp mat4 light_matrix; 40: uniform highp mat4 light_local_matrix; 41: uniform highp mat4 shadow_matrix; 42: uniform highp vec4 light_color; 43: uniform highp vec4 light_shadow_color; 44: uniform highp vec2 light_pos; 45: uniform highp float shadowpixel_size; 46: uniform highp float shadow_gradient; 47: uniform highp float light_height; 48: uniform highp float light_outside_alpha; 49: uniform highp float shadow_distance_mult; 50: 51: uniform lowp sampler2D light_texture; // texunit:-3 52: varying vec4 light_uv_interp; 53: varying vec2 transformed_light_uv; 54: 55: varying vec4 local_rot; 56: 57: #ifdef USE_SHADOWS 58: 59: uniform highp sampler2D shadow_texture; // texunit:-4 60: varying highp vec2 pos; 61: 62: #endif 63: 64: const bool at_light_pass = true; 65: #else 66: const bool at_light_pass = false; 67: #endif 68: 69: uniform bool use_default_normal; 70: 71: 72: /* clang-format off */ 73: 74: 75: /* clang-format on */ 76: 77: void main() { 78: 79: vec4 color = color_interp; 80: 81: #if !defined(COLOR_USED) 82: //default behavior, texture by color 83: color *= texture2D(color_texture, uv_interp); 84: #endif 85: 86: #ifdef SCREEN_UV_USED 87: vec2 screen_uv = gl_FragCoord.xy * screen_pixel_size; 88: #endif 89: 90: vec3 normal; 91: 92: #if defined(NORMAL_USED) 93: 94: bool normal_used = true; 95: #else 96: bool normal_used = false; 97: #endif 98: 99: if (use_default_normal) { 100: normal.xy = texture2D(normal_texture, uv_interp).xy * 2.0 - 1.0; 101: normal.z = sqrt(1.0 - dot(normal.xy, normal.xy)); 102: normal_used = true; 103: } else { 104: normal = vec3(0.0, 0.0, 1.0); 105: } 106: 107: { 108: float normal_depth = 1.0; 109: 110: #if defined(NORMALMAP_USED) 111: vec3 normal_map = vec3(0.0, 0.0, 1.0); 112: normal_used = true; 113: #endif 114: 115: /* clang-format off */ 116: 117: 118: /* clang-format on */ 119: 120: #if defined(NORMALMAP_USED) 121: normal = mix(vec3(0.0, 0.0, 1.0), normal_map * vec3(2.0, -2.0, 1.0) - vec3(1.0, -1.0, 0.0), normal_depth); 122: #endif 123: } 124: color *= final_modulate; 125: 126: #ifdef USE_LIGHTING 127: 128: vec2 light_vec = transformed_light_uv; 129: 130: if (normal_used) { 131: normal.xy = mat2(local_rot.xy, local_rot.zw) * normal.xy; 132: } 133: 134: float att = 1.0; 135: 136: vec2 light_uv = light_uv_interp.xy; 137: vec4 light = texture2D(light_texture, light_uv); 138: 139: if (any(lessThan(light_uv_interp.xy, vec2(0.0, 0.0))) || any(greaterThanEqual(light_uv_interp.xy, vec2(1.0, 1.0)))) { 140: color.a *= light_outside_alpha; //invisible 141: 142: } else { 143: float real_light_height = light_height; 144: vec4 real_light_color = light_color; 145: vec4 real_light_shadow_color = light_shadow_color; 146: 147: #if defined(USE_LIGHT_SHADER_CODE) 148: //light is written by the light shader 149: light_compute( 150: light, 151: light_vec, 152: real_light_height, 153: real_light_color, 154: light_uv, 155: real_light_shadow_color, 156: normal, 157: uv, 158: #if defined(SCREEN_UV_USED) 159: screen_uv, 160: #endif 161: color); 162: #endif 163: 164: light *= real_light_color; 165: 166: if (normal_used) { 167: vec3 light_normal = normalize(vec3(light_vec, -real_light_height)); 168: light *= max(dot(-light_normal, normal), 0.0); 169: } 170: 171: color *= light; 172: 173: #ifdef USE_SHADOWS 174: light_vec = light_uv_interp.zw; //for shadows 175: float angle_to_light = -atan(light_vec.x, light_vec.y); 176: float PI = 3.14159265358979323846264; 177: /*int i = int(mod(floor((angle_to_light+7.0*PI/6.0)/(4.0*PI/6.0))+1.0, 3.0)); // +1 pq os indices estao em ordem 2,0,1 nos arrays 178: float ang*/ 179: 180: float su, sz; 181: 182: float abs_angle = abs(angle_to_light); 183: vec2 point; 184: float sh; 185: if (abs_angle < 45.0 * PI / 180.0) { 186: point = light_vec; 187: sh = 0.0 + (1.0 / 8.0); 188: } else if (abs_angle > 135.0 * PI / 180.0) { 189: point = -light_vec; 190: sh = 0.5 + (1.0 / 8.0); 191: } else if (angle_to_light > 0.0) { 192: 193: point = vec2(light_vec.y, -light_vec.x); 194: sh = 0.25 + (1.0 / 8.0); 195: } else { 196: 197: point = vec2(-light_vec.y, light_vec.x); 198: sh = 0.75 + (1.0 / 8.0); 199: } 200: 201: highp vec4 s = shadow_matrix * vec4(point, 0.0, 1.0); 202: s.xyz /= s.w; 203: su = s.x * 0.5 + 0.5; 204: sz = s.z * 0.5 + 0.5; 205: //sz=lightlength(light_vec); 206: 207: highp float shadow_attenuation = 0.0; 208: 209: #ifdef USE_RGBA_SHADOWS 210: 211: #define SHADOW_DEPTH(m_tex, m_uv) dot(texture2D((m_tex), (m_uv)), vec4(1.0 / (256.0 * 256.0 * 256.0), 1.0 / (256.0 * 256.0), 1.0 / 256.0, 1)) 212: 213: #else 214: 215: #define SHADOW_DEPTH(m_tex, m_uv) (texture2D((m_tex), (m_uv)).r) 216: 217: #endif 218: 219: #ifdef SHADOW_USE_GRADIENT 220: 221: #define SHADOW_TEST(m_ofs) \ 222: { \ 223: highp float sd = SHADOW_DEPTH(shadow_texture, vec2(m_ofs, sh)); \ 224: shadow_attenuation += 1.0 - smoothstep(sd, sd + shadow_gradient, sz); \ 225: } 226: 227: #else 228: 229: #define SHADOW_TEST(m_ofs) \ 230: { \ 231: highp float sd = SHADOW_DEPTH(shadow_texture, vec2(m_ofs, sh)); \ 232: shadow_attenuation += step(sz, sd); \ 233: } 234: 235: #endif 236: 237: #ifdef SHADOW_FILTER_NEAREST 238: 239: SHADOW_TEST(su); 240: 241: #endif 242: 243: #ifdef SHADOW_FILTER_PCF3 244: 245: SHADOW_TEST(su + shadowpixel_size); 246: SHADOW_TEST(su); 247: SHADOW_TEST(su - shadowpixel_size); 248: shadow_attenuation /= 3.0; 249: 250: #endif 251: 252: #ifdef SHADOW_FILTER_PCF5 253: 254: SHADOW_TEST(su + shadowpixel_size * 2.0); 255: SHADOW_TEST(su + shadowpixel_size); 256: SHADOW_TEST(su); 257: SHADOW_TEST(su - shadowpixel_size); 258: SHADOW_TEST(su - shadowpixel_size * 2.0); 259: shadow_attenuation /= 5.0; 260: 261: #endif 262: 263: #ifdef SHADOW_FILTER_PCF7 264: 265: SHADOW_TEST(su + shadowpixel_size * 3.0); 266: SHADOW_TEST(su + shadowpixel_size * 2.0); 267: SHADOW_TEST(su + shadowpixel_size); 268: SHADOW_TEST(su); 269: SHADOW_TEST(su - shadowpixel_size); 270: SHADOW_TEST(su - shadowpixel_size * 2.0); 271: SHADOW_TEST(su - shadowpixel_size * 3.0); 272: shadow_attenuation /= 7.0; 273: 274: #endif 275: 276: #ifdef SHADOW_FILTER_PCF9 277: 278: SHADOW_TEST(su + shadowpixel_size * 4.0); 279: SHADOW_TEST(su + shadowpixel_size * 3.0); 280: SHADOW_TEST(su + shadowpixel_size * 2.0); 281: SHADOW_TEST(su + shadowpixel_size); 282: SHADOW_TEST(su); 283: SHADOW_TEST(su - shadowpixel_size); 284: SHADOW_TEST(su - shadowpixel_size * 2.0); 285: SHADOW_TEST(su - shadowpixel_size * 3.0); 286: SHADOW_TEST(su - shadowpixel_size * 4.0); 287: shadow_attenuation /= 9.0; 288: 289: #endif 290: 291: #ifdef SHADOW_FILTER_PCF13 292: 293: SHADOW_TEST(su + shadowpixel_size * 6.0); 294: SHADOW_TEST(su + shadowpixel_size * 5.0); 295: SHADOW_TEST(su + shadowpixel_size * 4.0); 296: SHADOW_TEST(su + shadowpixel_size * 3.0); 297: SHADOW_TEST(su + shadowpixel_size * 2.0); 298: SHADOW_TEST(su + shadowpixel_size); 299: SHADOW_TEST(su); 300: SHADOW_TEST(su - shadowpixel_size); 301: SHADOW_TEST(su - shadowpixel_size * 2.0); 302: SHADOW_TEST(su - shadowpixel_size * 3.0); 303: SHADOW_TEST(su - shadowpixel_size * 4.0); 304: SHADOW_TEST(su - shadowpixel_size * 5.0); 305: SHADOW_TEST(su - shadowpixel_size * 6.0); 306: shadow_attenuation /= 13.0; 307: 308: #endif 309: 310: //color *= shadow_attenuation; 311: color = mix(real_light_shadow_color, color, shadow_attenuation); 312: //use shadows 313: #endif 314: } 315: 316: //use lighting 317: #endif 318: 319: gl_FragColor = color; 320: } 321: ERROR: _display_error_with_code: CanvasShaderGLES2: Fragment shader compilation failed: ERROR: 0:51: 'lowp' : Reserved word. ERROR: 0:51: 'lowp' : syntax error: syntax error At: drivers/gles2/shader_gles2.cpp:196. ERROR: get_current_version: Method/Function Failed, returning: __null At: drivers/gles2/shader_gles2.cpp:426. ERROR: bind: Condition ' !version ' is true. returned: false At: drivers/gles2/shader_gles2.cpp:133. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ERROR: _get_uniform: Condition ' !version ' is true. returned: -1 At: drivers/gles2/shader_gles2.h:524. ``` _Originally posted by @kiidmale in https://github.com/godotengine/godot/issues/21853#issuecomment-449328028_
priority
fail compiling shader lowp undefined godot version os macos intel hd is still not working with i ve tested with latest commit and got below error users lazywierdo godot bin godot osx opt tools path users lazywierdo godot test projects e res main tscn current path users lazywierdo godot test projects opengl es renderer intel hd graphics opengl engine error initialize directional shadow framebuffer status invalid at drivers rasterizer scene cpp version define use gles over gl define use lighting ifdef use gles over gl define mediump define highp else precision mediump float precision mediump int endif uniform color texture texunit clang format on uniform highp color texpixel size uniform mediump normal texture texunit varying mediump uv interp varying mediump color interp uniform highp float time uniform final modulate ifdef screen texture used uniform screen texture texunit endif ifdef screen uv used uniform screen pixel size endif ifdef use lighting uniform highp light matrix uniform highp light local matrix uniform highp shadow matrix uniform highp light color uniform highp light shadow color uniform highp light pos uniform highp float shadowpixel size uniform highp float shadow gradient uniform highp float light height uniform highp float light outside alpha uniform highp float shadow distance mult uniform lowp light texture texunit varying light uv interp varying transformed light uv varying local rot ifdef use shadows uniform highp shadow texture texunit varying highp pos endif const bool at light pass true else const bool at light pass false endif uniform bool use default normal clang format off clang format on void main color color interp if defined color used default behavior texture by color color color texture uv interp endif ifdef screen uv used screen uv gl fragcoord xy screen pixel size endif normal if defined normal used bool normal used true else bool normal used false endif if use default normal normal xy normal texture uv interp xy normal z sqrt dot normal xy normal xy normal used true else normal float normal depth if defined normalmap used normal map normal used true endif clang format off clang format on if defined normalmap used normal mix normal map normal depth endif color final modulate ifdef use lighting light vec transformed light uv if normal used normal xy local rot xy local rot zw normal xy float att light uv light uv interp xy light light texture light uv if any lessthan light uv interp xy any greaterthanequal light uv interp xy color a light outside alpha invisible else float real light height light height real light color light color real light shadow color light shadow color if defined use light shader code light is written by the light shader light compute light light vec real light height real light color light uv real light shadow color normal uv if defined screen uv used screen uv endif color endif light real light color if normal used light normal normalize light vec real light height light max dot light normal normal color light ifdef use shadows light vec light uv interp zw for shadows float angle to light atan light vec x light vec y float pi int i int mod floor angle to light pi pi pq os indices estao em ordem nos arrays float ang float su sz float abs angle abs angle to light point float sh if abs angle pi point light vec sh else if abs angle pi point light vec sh else if angle to light point light vec y light vec x sh else point light vec y light vec x sh highp s shadow matrix point s xyz s w su s x sz s z sz lightlength light vec highp float shadow attenuation ifdef use rgba shadows define shadow depth m tex m uv dot m tex m uv else define shadow depth m tex m uv m tex m uv r endif ifdef shadow use gradient define shadow test m ofs highp float sd shadow depth shadow texture m ofs sh shadow attenuation smoothstep sd sd shadow gradient sz else define shadow test m ofs highp float sd shadow depth shadow texture m ofs sh shadow attenuation step sz sd endif ifdef shadow filter nearest shadow test su endif ifdef shadow filter shadow test su shadowpixel size shadow test su shadow test su shadowpixel size shadow attenuation endif ifdef shadow filter shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadow test su shadowpixel size shadow test su shadowpixel size shadow attenuation endif ifdef shadow filter shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow attenuation endif ifdef shadow filter shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow attenuation endif ifdef shadow filter shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow test su shadowpixel size shadow attenuation endif color shadow attenuation color mix real light shadow color color shadow attenuation use shadows endif use lighting endif gl fragcolor color error display error with code fragment shader compilation failed error lowp reserved word error lowp syntax error syntax error at drivers shader cpp error get current version method function failed returning null at drivers shader cpp error bind condition version is true returned false at drivers shader cpp error get uniform condition version is true returned at drivers shader h error get uniform condition version is true returned at drivers shader h error get uniform condition version is true returned at drivers shader h error get uniform condition version is true returned at drivers shader h error get uniform condition version is true returned at drivers shader h error get uniform condition version is true returned at drivers shader h error get uniform condition version is true returned at drivers shader h originally posted by kiidmale in
1
613,809
19,099,326,413
IssuesEvent
2021-11-29 20:24:30
stackabletech/t2
https://api.github.com/repos/stackabletech/t2
closed
bring OPA back
type/bug priority/high
During the K8s4n, I somehow dropped the OPA operator. Time to bring it back on stage!
1.0
bring OPA back - During the K8s4n, I somehow dropped the OPA operator. Time to bring it back on stage!
priority
bring opa back during the i somehow dropped the opa operator time to bring it back on stage
1
766,517
26,886,328,221
IssuesEvent
2023-02-06 03:46:43
wso2/api-manager
https://api.github.com/repos/wso2/api-manager
closed
[MSSQL] Error while retrieving correlation configs | Cannot create API
Type/Bug Priority/Highest Resolution/Invalid Component/APIM 4.2.0-beta
Observed following in pre beta back - carbon-apimgt v9.28.71 ``` [2023-02-05 23:43:10,728] ERROR - APIManagerComponent Error while initializing the API manager component org.wso2.carbon.apimgt.api.APIManagementException: Error while retrieving correlation configs at org.wso2.carbon.apimgt.impl.dao.CorrelationConfigDAO.isConfigExist_aroundBody6(CorrelationConfigDAO.java:193) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.dao.CorrelationConfigDAO.isConfigExist(CorrelationConfigDAO.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.addDefaultCorrelationConfigs_aroundBody854(APIUtil.java:9938) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.addDefaultCorrelationConfigs(APIUtil.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.internal.APIManagerComponent.activate_aroundBody0(APIManagerComponent.java:284) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.internal.APIManagerComponent.activate(APIManagerComponent.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] ... Caused by: com.microsoft.sqlserver.jdbc.SQLServerException: Invalid object name 'AM_CORRELATION_CONFIGS'. at com.microsoft.sqlserver.jdbc.SQLServerException.makeFromDatabaseError(SQLServerException.java:262) ~[mssql-jdbc-7.4.1.jre11.jar:?] at com.microsoft.sqlserver.jdbc.SQLServerStatement.getNextResult(SQLServerStatement.java:1624) ~[mssql-jdbc-7.4.1.jre11.jar:?] at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement.doExecutePreparedStatement(SQLServerPreparedStatement.java:594) ~[mssql-jdbc-7.4.1.jre11.jar:?] ``` This was followed by ``` [2023-02-05 23:44:06,594] ERROR - Framework FrameworkEvent ERROR java.lang.NullPointerException: null at org.wso2.carbon.apimgt.common.gateway.util.CommonAPIUtil.getHttpClient(CommonAPIUtil.java:84) ~[org.wso2.carbon.apimgt.common.gateway_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient_aroundBody402(APIUtil.java:5124) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient(APIUtil.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.correlation.CorrelationConfigManager.invokeService(CorrelationConfigManager.java:161) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.correlation.CorrelationConfigManager.initializeCorrelationComponentList(CorrelationConfigManager.java:74) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] ``` Because of the above, publisher fails to create an API with following error. ``` [2023-02-05 23:48:29,817] ERROR - GlobalThrowableMapper An unknown exception has been captured by the global exception mapper. java.lang.NullPointerException: null at org.wso2.carbon.apimgt.common.gateway.util.CommonAPIUtil.getHttpClient(CommonAPIUtil.java:84) ~[org.wso2.carbon.apimgt.common.gateway_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient_aroundBody402(APIUtil.java:5124) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient(APIUtil.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.loader.KeyManagerConfigurationDataRetriever.run_aroundBody0(KeyManagerConfigurationDataRetriever.java:56) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.loader.KeyManagerConfigurationDataRetriever.run(KeyManagerConfigurationDataRetriever.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagerDto_aroundBody14(KeyManagerHolder.java:260) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagerDto(KeyManagerHolder.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagers_aroundBody2(KeyManagerHolder.java:125) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagers(KeyManagerHolder.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.validateSharedScopes_aroundBody356(APIProviderImpl.java:4530) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.validateSharedScopes(APIProviderImpl.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.validateSharedScopes(UserAwareAPIProvider.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.common.mappings.PublisherCommonUtils.validateScopes(PublisherCommonUtils.java:773) ~[org.wso2.carbon.apimgt.rest.api.publisher.v1.common_9.28.71.jar:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.common.mappings.PublisherCommonUtils.addAPIWithGeneratedSwaggerDefinition(PublisherCommonUtils.java:844) ~[org.wso2.carbon.apimgt.rest.api.publisher.v1.common_9.28.71.jar:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.impl.ApisApiServiceImpl.createAPI(ApisApiServiceImpl.java:170) ~[?:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.ApisApi.createAPI(ApisApi.java:319) ~[?:?] ```
1.0
[MSSQL] Error while retrieving correlation configs | Cannot create API - Observed following in pre beta back - carbon-apimgt v9.28.71 ``` [2023-02-05 23:43:10,728] ERROR - APIManagerComponent Error while initializing the API manager component org.wso2.carbon.apimgt.api.APIManagementException: Error while retrieving correlation configs at org.wso2.carbon.apimgt.impl.dao.CorrelationConfigDAO.isConfigExist_aroundBody6(CorrelationConfigDAO.java:193) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.dao.CorrelationConfigDAO.isConfigExist(CorrelationConfigDAO.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.addDefaultCorrelationConfigs_aroundBody854(APIUtil.java:9938) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.addDefaultCorrelationConfigs(APIUtil.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.internal.APIManagerComponent.activate_aroundBody0(APIManagerComponent.java:284) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.internal.APIManagerComponent.activate(APIManagerComponent.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) ~[?:?] at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) ~[?:?] at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:?] at java.lang.reflect.Method.invoke(Method.java:566) ~[?:?] ... Caused by: com.microsoft.sqlserver.jdbc.SQLServerException: Invalid object name 'AM_CORRELATION_CONFIGS'. at com.microsoft.sqlserver.jdbc.SQLServerException.makeFromDatabaseError(SQLServerException.java:262) ~[mssql-jdbc-7.4.1.jre11.jar:?] at com.microsoft.sqlserver.jdbc.SQLServerStatement.getNextResult(SQLServerStatement.java:1624) ~[mssql-jdbc-7.4.1.jre11.jar:?] at com.microsoft.sqlserver.jdbc.SQLServerPreparedStatement.doExecutePreparedStatement(SQLServerPreparedStatement.java:594) ~[mssql-jdbc-7.4.1.jre11.jar:?] ``` This was followed by ``` [2023-02-05 23:44:06,594] ERROR - Framework FrameworkEvent ERROR java.lang.NullPointerException: null at org.wso2.carbon.apimgt.common.gateway.util.CommonAPIUtil.getHttpClient(CommonAPIUtil.java:84) ~[org.wso2.carbon.apimgt.common.gateway_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient_aroundBody402(APIUtil.java:5124) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient(APIUtil.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.correlation.CorrelationConfigManager.invokeService(CorrelationConfigManager.java:161) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.correlation.CorrelationConfigManager.initializeCorrelationComponentList(CorrelationConfigManager.java:74) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] ``` Because of the above, publisher fails to create an API with following error. ``` [2023-02-05 23:48:29,817] ERROR - GlobalThrowableMapper An unknown exception has been captured by the global exception mapper. java.lang.NullPointerException: null at org.wso2.carbon.apimgt.common.gateway.util.CommonAPIUtil.getHttpClient(CommonAPIUtil.java:84) ~[org.wso2.carbon.apimgt.common.gateway_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient_aroundBody402(APIUtil.java:5124) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.utils.APIUtil.getHttpClient(APIUtil.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.loader.KeyManagerConfigurationDataRetriever.run_aroundBody0(KeyManagerConfigurationDataRetriever.java:56) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.loader.KeyManagerConfigurationDataRetriever.run(KeyManagerConfigurationDataRetriever.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagerDto_aroundBody14(KeyManagerHolder.java:260) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagerDto(KeyManagerHolder.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagers_aroundBody2(KeyManagerHolder.java:125) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.factory.KeyManagerHolder.getTenantKeyManagers(KeyManagerHolder.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.validateSharedScopes_aroundBody356(APIProviderImpl.java:4530) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.validateSharedScopes(APIProviderImpl.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.validateSharedScopes(UserAwareAPIProvider.java:1) ~[org.wso2.carbon.apimgt.impl_9.28.71.jar:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.common.mappings.PublisherCommonUtils.validateScopes(PublisherCommonUtils.java:773) ~[org.wso2.carbon.apimgt.rest.api.publisher.v1.common_9.28.71.jar:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.common.mappings.PublisherCommonUtils.addAPIWithGeneratedSwaggerDefinition(PublisherCommonUtils.java:844) ~[org.wso2.carbon.apimgt.rest.api.publisher.v1.common_9.28.71.jar:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.impl.ApisApiServiceImpl.createAPI(ApisApiServiceImpl.java:170) ~[?:?] at org.wso2.carbon.apimgt.rest.api.publisher.v1.ApisApi.createAPI(ApisApi.java:319) ~[?:?] ```
priority
error while retrieving correlation configs cannot create api observed following in pre beta back carbon apimgt error apimanagercomponent error while initializing the api manager component org carbon apimgt api apimanagementexception error while retrieving correlation configs at org carbon apimgt impl dao correlationconfigdao isconfigexist correlationconfigdao java at org carbon apimgt impl dao correlationconfigdao isconfigexist correlationconfigdao java at org carbon apimgt impl utils apiutil adddefaultcorrelationconfigs apiutil java at org carbon apimgt impl utils apiutil adddefaultcorrelationconfigs apiutil java at org carbon apimgt impl internal apimanagercomponent activate apimanagercomponent java at org carbon apimgt impl internal apimanagercomponent activate apimanagercomponent java at jdk internal reflect nativemethodaccessorimpl native method at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java caused by com microsoft sqlserver jdbc sqlserverexception invalid object name am correlation configs at com microsoft sqlserver jdbc sqlserverexception makefromdatabaseerror sqlserverexception java at com microsoft sqlserver jdbc sqlserverstatement getnextresult sqlserverstatement java at com microsoft sqlserver jdbc sqlserverpreparedstatement doexecutepreparedstatement sqlserverpreparedstatement java this was followed by error framework frameworkevent error java lang nullpointerexception null at org carbon apimgt common gateway util commonapiutil gethttpclient commonapiutil java at org carbon apimgt impl utils apiutil gethttpclient apiutil java at org carbon apimgt impl utils apiutil gethttpclient apiutil java at org carbon apimgt impl correlation correlationconfigmanager invokeservice correlationconfigmanager java at org carbon apimgt impl correlation correlationconfigmanager initializecorrelationcomponentlist correlationconfigmanager java because of the above publisher fails to create an api with following error error globalthrowablemapper an unknown exception has been captured by the global exception mapper java lang nullpointerexception null at org carbon apimgt common gateway util commonapiutil gethttpclient commonapiutil java at org carbon apimgt impl utils apiutil gethttpclient apiutil java at org carbon apimgt impl utils apiutil gethttpclient apiutil java at org carbon apimgt impl loader keymanagerconfigurationdataretriever run keymanagerconfigurationdataretriever java at org carbon apimgt impl loader keymanagerconfigurationdataretriever run keymanagerconfigurationdataretriever java at org carbon apimgt impl factory keymanagerholder gettenantkeymanagerdto keymanagerholder java at org carbon apimgt impl factory keymanagerholder gettenantkeymanagerdto keymanagerholder java at org carbon apimgt impl factory keymanagerholder gettenantkeymanagers keymanagerholder java at org carbon apimgt impl factory keymanagerholder gettenantkeymanagers keymanagerholder java at org carbon apimgt impl apiproviderimpl validatesharedscopes apiproviderimpl java at org carbon apimgt impl apiproviderimpl validatesharedscopes apiproviderimpl java at org carbon apimgt impl userawareapiprovider validatesharedscopes userawareapiprovider java at org carbon apimgt rest api publisher common mappings publishercommonutils validatescopes publishercommonutils java at org carbon apimgt rest api publisher common mappings publishercommonutils addapiwithgeneratedswaggerdefinition publishercommonutils java at org carbon apimgt rest api publisher impl apisapiserviceimpl createapi apisapiserviceimpl java at org carbon apimgt rest api publisher apisapi createapi apisapi java
1
347,179
10,426,042,954
IssuesEvent
2019-09-16 16:39:48
jetrails/magento2-cloudflare
https://api.github.com/repos/jetrails/magento2-cloudflare
closed
ACL Not Used Properly In Magento 2.2?
bug priority: high
### Info Magento Version: 2.2.8 Extension Version: ^1.1 || ^1.2 ### Problem Magento 2.2 Cloudflare extension might not respect the ACL resources that it created. Resources are not selected, but still, show up for everyone regardless of role. Did not confirm if it is the same behavior for 2.3.
1.0
ACL Not Used Properly In Magento 2.2? - ### Info Magento Version: 2.2.8 Extension Version: ^1.1 || ^1.2 ### Problem Magento 2.2 Cloudflare extension might not respect the ACL resources that it created. Resources are not selected, but still, show up for everyone regardless of role. Did not confirm if it is the same behavior for 2.3.
priority
acl not used properly in magento info magento version extension version problem magento cloudflare extension might not respect the acl resources that it created resources are not selected but still show up for everyone regardless of role did not confirm if it is the same behavior for
1
206,828
7,121,572,707
IssuesEvent
2018-01-19 08:26:17
wso2/message-broker
https://api.github.com/repos/wso2/message-broker
opened
Implement exchanges rest apo
Complexity/Moderate Module/broker-core Priority/High Severity/Major Type/New Feature
### Description Implement /exchanges REST api to create/delete exchanges bindings from the broker.
1.0
Implement exchanges rest apo - ### Description Implement /exchanges REST api to create/delete exchanges bindings from the broker.
priority
implement exchanges rest apo description implement exchanges rest api to create delete exchanges bindings from the broker
1
239,561
7,799,809,445
IssuesEvent
2018-06-09 00:45:17
tine20/Tine-2.0-Open-Source-Groupware-and-CRM
https://api.github.com/repos/tine20/Tine-2.0-Open-Source-Groupware-and-CRM
closed
0005554: start duplicate resolve dialog from ctx menu
Addressbook Bug Mantis high priority
**Reported by cweiss on 7 Feb 2012 15:37** - enable option if exactly two contacts are selected - new &quot;outer&quot; dialog which handles update, delete and buttons - in duplicate resolveGridPanel: -- don&#039;t show &quot;my record&quot; ( when client record is missing) -- show numbers before duplicates -- actions: keep existing no. n
1.0
0005554: start duplicate resolve dialog from ctx menu - **Reported by cweiss on 7 Feb 2012 15:37** - enable option if exactly two contacts are selected - new &quot;outer&quot; dialog which handles update, delete and buttons - in duplicate resolveGridPanel: -- don&#039;t show &quot;my record&quot; ( when client record is missing) -- show numbers before duplicates -- actions: keep existing no. n
priority
start duplicate resolve dialog from ctx menu reported by cweiss on feb enable option if exactly two contacts are selected new quot outer quot dialog which handles update delete and buttons in duplicate resolvegridpanel don t show quot my record quot when client record is missing show numbers before duplicates actions keep existing no n
1
391,514
11,574,950,190
IssuesEvent
2020-02-21 08:42:35
waellet/waellet
https://api.github.com/repos/waellet/waellet
closed
Connection status not set to DISCONNECTED when Aepp closed
area/sdk kind/bug priority/high
There is problem with client disconnect when closing Aepp, which leads to error when changing accounts and networks from Wallet. The error is ![image](https://user-images.githubusercontent.com/8629465/73162733-27612880-40f7-11ea-8d1a-29275d5cfb09.png) because port (client) is closed but it's status is still **CONNECTED**: I think the problem is comming from incorrect client instance passed in `onDisconnect` method in `RpcWallet`. ![image](https://user-images.githubusercontent.com/8629465/73164066-d999ef80-40f9-11ea-8c59-51e4ddfc3690.png) - When calling `disconnectWallet` (https://github.com/aeternity/aepp-sdk-js/blob/develop/es/utils/aepp-wallet-communication/rpc/aepp-rpc.js#L131) from Aepp the client status is set to disconnected and in `onDisconnect` method the `client` param is instance of `RpcClient ` (https://github.com/aeternity/aepp-sdk-js/blob/develop/es/utils/aepp-wallet-communication/rpc/rpc-clients.js#L108) ![image](https://user-images.githubusercontent.com/8629465/73163546-e407b980-40f8-11ea-9630-94bdebd02700.png) - When Aepp is disconnected from window or tab close the client status is not set to disconnected and in `onDisconnect` method the `client` param is instance of `BrowserRuntimeConnection` (https://github.com/aeternity/aepp-sdk-js/blob/develop/es/utils/aepp-wallet-communication/connection/browser-runtime.js#L102) ![image](https://user-images.githubusercontent.com/8629465/73163568-f2ee6c00-40f8-11ea-9576-55ba529c2afd.png)
1.0
Connection status not set to DISCONNECTED when Aepp closed - There is problem with client disconnect when closing Aepp, which leads to error when changing accounts and networks from Wallet. The error is ![image](https://user-images.githubusercontent.com/8629465/73162733-27612880-40f7-11ea-8d1a-29275d5cfb09.png) because port (client) is closed but it's status is still **CONNECTED**: I think the problem is comming from incorrect client instance passed in `onDisconnect` method in `RpcWallet`. ![image](https://user-images.githubusercontent.com/8629465/73164066-d999ef80-40f9-11ea-8c59-51e4ddfc3690.png) - When calling `disconnectWallet` (https://github.com/aeternity/aepp-sdk-js/blob/develop/es/utils/aepp-wallet-communication/rpc/aepp-rpc.js#L131) from Aepp the client status is set to disconnected and in `onDisconnect` method the `client` param is instance of `RpcClient ` (https://github.com/aeternity/aepp-sdk-js/blob/develop/es/utils/aepp-wallet-communication/rpc/rpc-clients.js#L108) ![image](https://user-images.githubusercontent.com/8629465/73163546-e407b980-40f8-11ea-9630-94bdebd02700.png) - When Aepp is disconnected from window or tab close the client status is not set to disconnected and in `onDisconnect` method the `client` param is instance of `BrowserRuntimeConnection` (https://github.com/aeternity/aepp-sdk-js/blob/develop/es/utils/aepp-wallet-communication/connection/browser-runtime.js#L102) ![image](https://user-images.githubusercontent.com/8629465/73163568-f2ee6c00-40f8-11ea-9576-55ba529c2afd.png)
priority
connection status not set to disconnected when aepp closed there is problem with client disconnect when closing aepp which leads to error when changing accounts and networks from wallet the error is because port client is closed but it s status is still connected i think the problem is comming from incorrect client instance passed in ondisconnect method in rpcwallet when calling disconnectwallet from aepp the client status is set to disconnected and in ondisconnect method the client param is instance of rpcclient when aepp is disconnected from window or tab close the client status is not set to disconnected and in ondisconnect method the client param is instance of browserruntimeconnection
1
445,253
12,828,092,850
IssuesEvent
2020-07-06 19:49:52
cthit/Gamma
https://api.github.com/repos/cthit/Gamma
opened
In production, some errors is not caught and logged properly
Priority: High Status: Accepted Type: Bug Where: Backend
We had a situation where a group could not be edited or deleted. It just said `PUT not supported` or `DELETE not supported`. Improved logging could have helped in this situation.
1.0
In production, some errors is not caught and logged properly - We had a situation where a group could not be edited or deleted. It just said `PUT not supported` or `DELETE not supported`. Improved logging could have helped in this situation.
priority
in production some errors is not caught and logged properly we had a situation where a group could not be edited or deleted it just said put not supported or delete not supported improved logging could have helped in this situation
1
421,141
12,254,240,266
IssuesEvent
2020-05-06 08:07:22
RTradeLtd/s3x
https://api.github.com/repos/RTradeLtd/s3x
closed
A breaking change is coming, lets make the transition as seamless as possible!
high-priority
The goal of this project is to add TemporalX support to minio, while the goal has not changed, the original approach has shown it's downsides. Maintains our fork with upstream has become a regular time sink, time we could use bug fixing and developing new features. So we are working towards reorganizing s3x only as a minio gateway instead of a full fork. Removing the core minio code also gives us a chance to shink this repo if we rewrite master with a fresh history, ~~should this be done?~~ (yes) Most of the work will be towards adopting the CI to concentrate on testing our gateway instead of minio as a whole. With this change, what other breaking changes should we also make? What do you foresee that we should avoid breaking? Such as anything that we should avoid removing from s3x for your use cases. This issue is to collect user feedback on this change, while #62 is for development.
1.0
A breaking change is coming, lets make the transition as seamless as possible! - The goal of this project is to add TemporalX support to minio, while the goal has not changed, the original approach has shown it's downsides. Maintains our fork with upstream has become a regular time sink, time we could use bug fixing and developing new features. So we are working towards reorganizing s3x only as a minio gateway instead of a full fork. Removing the core minio code also gives us a chance to shink this repo if we rewrite master with a fresh history, ~~should this be done?~~ (yes) Most of the work will be towards adopting the CI to concentrate on testing our gateway instead of minio as a whole. With this change, what other breaking changes should we also make? What do you foresee that we should avoid breaking? Such as anything that we should avoid removing from s3x for your use cases. This issue is to collect user feedback on this change, while #62 is for development.
priority
a breaking change is coming lets make the transition as seamless as possible the goal of this project is to add temporalx support to minio while the goal has not changed the original approach has shown it s downsides maintains our fork with upstream has become a regular time sink time we could use bug fixing and developing new features so we are working towards reorganizing only as a minio gateway instead of a full fork removing the core minio code also gives us a chance to shink this repo if we rewrite master with a fresh history should this be done yes most of the work will be towards adopting the ci to concentrate on testing our gateway instead of minio as a whole with this change what other breaking changes should we also make what do you foresee that we should avoid breaking such as anything that we should avoid removing from for your use cases this issue is to collect user feedback on this change while is for development
1
737,315
25,511,118,716
IssuesEvent
2022-11-28 13:11:50
node-schedule/node-schedule
https://api.github.com/repos/node-schedule/node-schedule
closed
Cancelling all jobs still leaves long running process
help wanted high priority
in the case of an application, shutdown node-schedule holds the application up because of long-running timers, which generally makes sense in cases except for an explicit shutdown. fairly repeatable, especially in a mocha test create a recurring schedule then ```javascript import { scheduledJobs as scheduledCronJobs, cancelJob as cancelCronJob } from 'node-schedule'; import wtf from 'wtfnode'; after(() => { wtf.dump(); Object.keys(scheduledCronJobs).forEach(jobName => cancelCronJob(jobName)); console.info(scheduledCronJobs); // prints empty wtf.dump(); // still has long running timer. }); ``` ideally, we should actually have a safe shutdown function for use in apps (not other libraries): ```javascript function cancelAllJobs(triggered = false, cb = () => {}) { Object.keys(scheduledJobs).forEach(name => canceJob(name)); if (triggered) { // cancel jobs that are still processing } else { // wait for triggered jobs to complete } cb(); } ```
1.0
Cancelling all jobs still leaves long running process - in the case of an application, shutdown node-schedule holds the application up because of long-running timers, which generally makes sense in cases except for an explicit shutdown. fairly repeatable, especially in a mocha test create a recurring schedule then ```javascript import { scheduledJobs as scheduledCronJobs, cancelJob as cancelCronJob } from 'node-schedule'; import wtf from 'wtfnode'; after(() => { wtf.dump(); Object.keys(scheduledCronJobs).forEach(jobName => cancelCronJob(jobName)); console.info(scheduledCronJobs); // prints empty wtf.dump(); // still has long running timer. }); ``` ideally, we should actually have a safe shutdown function for use in apps (not other libraries): ```javascript function cancelAllJobs(triggered = false, cb = () => {}) { Object.keys(scheduledJobs).forEach(name => canceJob(name)); if (triggered) { // cancel jobs that are still processing } else { // wait for triggered jobs to complete } cb(); } ```
priority
cancelling all jobs still leaves long running process in the case of an application shutdown node schedule holds the application up because of long running timers which generally makes sense in cases except for an explicit shutdown fairly repeatable especially in a mocha test create a recurring schedule then javascript import scheduledjobs as scheduledcronjobs canceljob as cancelcronjob from node schedule import wtf from wtfnode after wtf dump object keys scheduledcronjobs foreach jobname cancelcronjob jobname console info scheduledcronjobs prints empty wtf dump still has long running timer ideally we should actually have a safe shutdown function for use in apps not other libraries javascript function cancelalljobs triggered false cb object keys scheduledjobs foreach name cancejob name if triggered cancel jobs that are still processing else wait for triggered jobs to complete cb
1
173,409
6,524,640,158
IssuesEvent
2017-08-29 13:24:44
nbnuk/nbnatlas-issues
https://api.github.com/repos/nbnuk/nbnatlas-issues
closed
Licence information isn't downloading for all of the records.
bug high-priority
I downloaded all of the Japanese knotweed records, and in the 'citation' file, column E 'Rights' doesn't always have a licence listed, and on these rows, column B 'Name' has the data partner name, not the dataset name (and the 'more info' link (column F) takes you to the whole data partner page, not the dataset page for the specific dataset that the records came from). Therefore, data users will have a hard time finding the licence info for the whole download.
1.0
Licence information isn't downloading for all of the records. - I downloaded all of the Japanese knotweed records, and in the 'citation' file, column E 'Rights' doesn't always have a licence listed, and on these rows, column B 'Name' has the data partner name, not the dataset name (and the 'more info' link (column F) takes you to the whole data partner page, not the dataset page for the specific dataset that the records came from). Therefore, data users will have a hard time finding the licence info for the whole download.
priority
licence information isn t downloading for all of the records i downloaded all of the japanese knotweed records and in the citation file column e rights doesn t always have a licence listed and on these rows column b name has the data partner name not the dataset name and the more info link column f takes you to the whole data partner page not the dataset page for the specific dataset that the records came from therefore data users will have a hard time finding the licence info for the whole download
1
505,700
14,643,953,495
IssuesEvent
2020-12-25 19:52:58
pterodactyl/panel
https://api.github.com/repos/pterodactyl/panel
closed
High iowait and memory usage when creating backups
bug help wanted :wave: high priority
**Is your feature request related to a problem? Please describe.** Backing up to an HDD causes huge spikes in iowait, which can slow servers down. **Describe the solution you'd like** I'd like the limit the speed at which they read/write to the HDD. wings is currently trying to sustain 100-200M/s reads/writes for an HDD backup, which just isn't viable. There may be a better way to handle this than simply limiting read/write speeds, but I'm not sure what it would be. **Describe alternatives you've considered** I think the alternatives would be to either disable backups for users with an HDD or to simply pay a few bucks to upgrade from an HDD to an SSD. Still, I think it'd make sense for wings to handle this in a way that won't lag servers. **Additional context** iotop output for two separate backups: ![image](https://user-images.githubusercontent.com/43528123/101324854-76f0a480-3830-11eb-81aa-fefa54cf8a34.png) ![image](https://user-images.githubusercontent.com/43528123/101324872-7b1cc200-3830-11eb-98dc-ada9a8929d71.png)
1.0
High iowait and memory usage when creating backups - **Is your feature request related to a problem? Please describe.** Backing up to an HDD causes huge spikes in iowait, which can slow servers down. **Describe the solution you'd like** I'd like the limit the speed at which they read/write to the HDD. wings is currently trying to sustain 100-200M/s reads/writes for an HDD backup, which just isn't viable. There may be a better way to handle this than simply limiting read/write speeds, but I'm not sure what it would be. **Describe alternatives you've considered** I think the alternatives would be to either disable backups for users with an HDD or to simply pay a few bucks to upgrade from an HDD to an SSD. Still, I think it'd make sense for wings to handle this in a way that won't lag servers. **Additional context** iotop output for two separate backups: ![image](https://user-images.githubusercontent.com/43528123/101324854-76f0a480-3830-11eb-81aa-fefa54cf8a34.png) ![image](https://user-images.githubusercontent.com/43528123/101324872-7b1cc200-3830-11eb-98dc-ada9a8929d71.png)
priority
high iowait and memory usage when creating backups is your feature request related to a problem please describe backing up to an hdd causes huge spikes in iowait which can slow servers down describe the solution you d like i d like the limit the speed at which they read write to the hdd wings is currently trying to sustain s reads writes for an hdd backup which just isn t viable there may be a better way to handle this than simply limiting read write speeds but i m not sure what it would be describe alternatives you ve considered i think the alternatives would be to either disable backups for users with an hdd or to simply pay a few bucks to upgrade from an hdd to an ssd still i think it d make sense for wings to handle this in a way that won t lag servers additional context iotop output for two separate backups
1
73,467
3,412,442,536
IssuesEvent
2015-12-05 22:05:16
websharks/wp-kb-articles
https://api.github.com/repos/websharks/wp-kb-articles
closed
Possible WP 4.4 Compatibility Issue
bug high priority
## Overview WordPress v4.4 introduces some new objects for `WP_Term` and `WP_Comment`. This is mostly OK with WPKBA. However, it recently came to my attention that in PHP 7.0 there are some flags being thrown with respect to type hints that are looking for `\stdClass` instances instead of the new `WP_Term` class. ## What needs to change? We need to search the codebase for `\stdClass $` and remove `\stdClass` as it pertains to `WP_Term`.
1.0
Possible WP 4.4 Compatibility Issue - ## Overview WordPress v4.4 introduces some new objects for `WP_Term` and `WP_Comment`. This is mostly OK with WPKBA. However, it recently came to my attention that in PHP 7.0 there are some flags being thrown with respect to type hints that are looking for `\stdClass` instances instead of the new `WP_Term` class. ## What needs to change? We need to search the codebase for `\stdClass $` and remove `\stdClass` as it pertains to `WP_Term`.
priority
possible wp compatibility issue overview wordpress introduces some new objects for wp term and wp comment this is mostly ok with wpkba however it recently came to my attention that in php there are some flags being thrown with respect to type hints that are looking for stdclass instances instead of the new wp term class what needs to change we need to search the codebase for stdclass and remove stdclass as it pertains to wp term
1