Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
512,705
14,907,845,605
IssuesEvent
2021-01-22 04:17:00
Plaxy-Technologies-Inc/YouPlanets-Bug-Report
https://api.github.com/repos/Plaxy-Technologies-Inc/YouPlanets-Bug-Report
closed
Bug/Feature Missing: Can't connect to Stripe without Social Security Number - creators can't get paid without one?
Emergency Priority: High
![image](https://user-images.githubusercontent.com/77167655/105377111-1515af80-5bd0-11eb-9797-840a35694cbd.png)
1.0
Bug/Feature Missing: Can't connect to Stripe without Social Security Number - creators can't get paid without one? - ![image](https://user-images.githubusercontent.com/77167655/105377111-1515af80-5bd0-11eb-9797-840a35694cbd.png)
priority
bug feature missing can t connect to stripe without social security number creators can t get paid without one
1
631,183
20,146,999,115
IssuesEvent
2022-02-09 08:39:28
Disfactory/Disfactory
https://api.github.com/repos/Disfactory/Disfactory
closed
無法上傳照片
bug high priority
**Describe the bug** User couldn't upload the photo. **To Reproduce** Steps to reproduce the behavior: 1. Go to disfactory.tw 2. Click on "我想新增可疑工廠" 3. Click on '新增照片', choose one 4. See error **Screenshots** <img width="1440" alt="截圖 2022-02-07 上午10 51 32" src="https://user-images.githubusercontent.com/60970217/152717064-b8c8a2af-ba4d-4b9d-9308-d11770edb287.png"> **Desktop (please complete the following information):** - OS: MacOS - Browser: chrome - Version: 97.0.4692.99 **Additional context** There are two users respond that they experienced the same problem through troubleshooting form respectively on 1/30 and 2/5.
1.0
無法上傳照片 - **Describe the bug** User couldn't upload the photo. **To Reproduce** Steps to reproduce the behavior: 1. Go to disfactory.tw 2. Click on "我想新增可疑工廠" 3. Click on '新增照片', choose one 4. See error **Screenshots** <img width="1440" alt="截圖 2022-02-07 上午10 51 32" src="https://user-images.githubusercontent.com/60970217/152717064-b8c8a2af-ba4d-4b9d-9308-d11770edb287.png"> **Desktop (please complete the following information):** - OS: MacOS - Browser: chrome - Version: 97.0.4692.99 **Additional context** There are two users respond that they experienced the same problem through troubleshooting form respectively on 1/30 and 2/5.
priority
無法上傳照片 describe the bug user couldn t upload the photo to reproduce steps to reproduce the behavior go to disfactory tw click on 我想新增可疑工廠 click on 新增照片 choose one see error screenshots img width alt 截圖 src desktop please complete the following information os macos browser chrome version additional context there are two users respond that they experienced the same problem through troubleshooting form respectively on and
1
335,813
10,167,022,266
IssuesEvent
2019-08-07 17:11:39
worldmaking/msvr
https://api.github.com/repos/worldmaking/msvr
opened
running max causes module instancing error
Priority: High bug
@Zodsmar told me that he's able to reproduce an error where modules will not instantiate whenever max's audio rendering is turned on.
1.0
running max causes module instancing error - @Zodsmar told me that he's able to reproduce an error where modules will not instantiate whenever max's audio rendering is turned on.
priority
running max causes module instancing error zodsmar told me that he s able to reproduce an error where modules will not instantiate whenever max s audio rendering is turned on
1
41,328
2,868,998,255
IssuesEvent
2015-06-05 22:28:20
dart-lang/pub-dartlang
https://api.github.com/repos/dart-lang/pub-dartlang
closed
pub search not finding library that exists
bug notplanned Priority-High
<a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)** _Originally opened as dart-lang/sdk#20755_ ---- This has happened a few times now, there's some systemic going on. This time, I can't find stagehand. It's first version has been up since 7/26. Yet, search can't find it. See screenshots. There have been other bugs like this, and somehow it just starts working. But in this case, it's been over a month. I think we should take a look at this. ______ **Attachments:** [Screen Shot 2014-08-30 at 5.19.53 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.53 PM.png) (32.68 KB) [Screen Shot 2014-08-30 at 5.19.48 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.48 PM.png) (40.56 KB)
1.0
pub search not finding library that exists - <a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)** _Originally opened as dart-lang/sdk#20755_ ---- This has happened a few times now, there's some systemic going on. This time, I can't find stagehand. It's first version has been up since 7/26. Yet, search can't find it. See screenshots. There have been other bugs like this, and somehow it just starts working. But in this case, it's been over a month. I think we should take a look at this. ______ **Attachments:** [Screen Shot 2014-08-30 at 5.19.53 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.53 PM.png) (32.68 KB) [Screen Shot 2014-08-30 at 5.19.48 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.48 PM.png) (40.56 KB)
priority
pub search not finding library that exists issue by originally opened as dart lang sdk this has happened a few times now there s some systemic going on this time i can t find stagehand it s first version has been up since yet search can t find it see screenshots there have been other bugs like this and somehow it just starts working but in this case it s been over a month i think we should take a look at this attachments shot at pm png kb shot at pm png kb
1
114,685
4,642,618,547
IssuesEvent
2016-09-30 10:18:05
armadito/armadito-av
https://api.github.com/repos/armadito/armadito-av
opened
improve/simplify REST API
enhancement high priority
update REST API for the following enhancements: - remove '/api/register' call and integrate returning the token in the JSON objects returned by other requests such as '/api/scan' or '/api/status' - add arguments to '/api/scan': generate progress events, progress event frequency (may be merge in one parameter: 0 for no progress event, > 0 for frequency in seconds)
1.0
improve/simplify REST API - update REST API for the following enhancements: - remove '/api/register' call and integrate returning the token in the JSON objects returned by other requests such as '/api/scan' or '/api/status' - add arguments to '/api/scan': generate progress events, progress event frequency (may be merge in one parameter: 0 for no progress event, > 0 for frequency in seconds)
priority
improve simplify rest api update rest api for the following enhancements remove api register call and integrate returning the token in the json objects returned by other requests such as api scan or api status add arguments to api scan generate progress events progress event frequency may be merge in one parameter for no progress event for frequency in seconds
1
591,477
17,840,705,137
IssuesEvent
2021-09-03 09:41:53
francheska-vicente/cssweng
https://api.github.com/repos/francheska-vicente/cssweng
opened
Rooms that offer monthly rate should not have extra persons
bug priority: high issue: back-end severity: medium issue: validation
### Summary - Rooms that offer monthly rates should not entertain extra pax. ### Steps to Reproduce 1. login 2. choose any date for booking 3. choose room 305 4. input 100 in the number of pax field ### Visual Proof ![image](https://user-images.githubusercontent.com/74450482/131984778-e9ee807c-8b8d-497a-a225-61de857b3af3.png) ### Expected Results: - Number of persons should be limited to at most 4 pax for twin bed (rm. 305). ### Actual Results: - There is no limit to the number of extra pax for the monthly rate of twin bed (rm. 305). | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
1.0
Rooms that offer monthly rate should not have extra persons - ### Summary - Rooms that offer monthly rates should not entertain extra pax. ### Steps to Reproduce 1. login 2. choose any date for booking 3. choose room 305 4. input 100 in the number of pax field ### Visual Proof ![image](https://user-images.githubusercontent.com/74450482/131984778-e9ee807c-8b8d-497a-a225-61de857b3af3.png) ### Expected Results: - Number of persons should be limited to at most 4 pax for twin bed (rm. 305). ### Actual Results: - There is no limit to the number of extra pax for the monthly rate of twin bed (rm. 305). | Additional Information | | | ----------- | ----------- | | Platform | V8 engine (Google) | | Operating System | Windows 10 |
priority
rooms that offer monthly rate should not have extra persons summary rooms that offer monthly rates should not entertain extra pax steps to reproduce login choose any date for booking choose room input in the number of pax field visual proof expected results number of persons should be limited to at most pax for twin bed rm actual results there is no limit to the number of extra pax for the monthly rate of twin bed rm additional information platform engine google operating system windows
1
363,060
10,736,972,952
IssuesEvent
2019-10-29 12:07:04
AY1920S1-CS2103T-W11-4/main
https://api.github.com/repos/AY1920S1-CS2103T-W11-4/main
closed
As a new user I want to see usage instructions
priority.High type.Story
So that I can refer to instructions when I forget how to use the App
1.0
As a new user I want to see usage instructions - So that I can refer to instructions when I forget how to use the App
priority
as a new user i want to see usage instructions so that i can refer to instructions when i forget how to use the app
1
107,178
4,290,704,000
IssuesEvent
2016-07-18 10:54:05
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
opened
Disable current help tool Mobile
enhancement Mobile Priority: High
As a consequence of #845 we are going for the moment to remove the current help widget/tool from the mobile layout waiting for a more mobile friendly widget to be developed in future releases.
1.0
Disable current help tool Mobile - As a consequence of #845 we are going for the moment to remove the current help widget/tool from the mobile layout waiting for a more mobile friendly widget to be developed in future releases.
priority
disable current help tool mobile as a consequence of we are going for the moment to remove the current help widget tool from the mobile layout waiting for a more mobile friendly widget to be developed in future releases
1
560,816
16,605,418,674
IssuesEvent
2021-06-02 02:43:04
QuantEcon/quantecon-book-theme
https://api.github.com/repos/QuantEcon/quantecon-book-theme
closed
[google analytics] Check if Google Analytics is Setup
bug high-priority
@DrDrij @AakashGfude this is an issue to track the status for looking to see if google analytics is setup correctly for - [ ] python-programming.quantecon.org - [ ] python.quantecon.org
1.0
[google analytics] Check if Google Analytics is Setup - @DrDrij @AakashGfude this is an issue to track the status for looking to see if google analytics is setup correctly for - [ ] python-programming.quantecon.org - [ ] python.quantecon.org
priority
check if google analytics is setup drdrij aakashgfude this is an issue to track the status for looking to see if google analytics is setup correctly for python programming quantecon org python quantecon org
1
323,563
9,856,515,180
IssuesEvent
2019-06-19 22:26:40
NCIOCPL/cgov-digital-platform
https://api.github.com/repos/NCIOCPL/cgov-digital-platform
closed
Spanish CGOV - Missing Images and Infographics
High priority Migration bug
Images are missing from Spanish pages. These got migrated into Drupal, but they are not appearing on the front-end. Some examples pages include: - http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/que-es - http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/sobrellevar/supervivencia/infancia - http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/radioterapia/haz-externo - Infographic on http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades --- AC - Adding Videos from #2020 http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/inmunoterapia VERSUS https://www.cancer.gov/espanol/cancer/tratamiento/tipos/inmunoterapia • http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades VERSUS https://www.cancer.gov/espanol/cancer/naturaleza/desigualdades • http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/seguridad-paciente • http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/pago (edited)
1.0
Spanish CGOV - Missing Images and Infographics - Images are missing from Spanish pages. These got migrated into Drupal, but they are not appearing on the front-end. Some examples pages include: - http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/que-es - http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/sobrellevar/supervivencia/infancia - http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/radioterapia/haz-externo - Infographic on http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades --- AC - Adding Videos from #2020 http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/inmunoterapia VERSUS https://www.cancer.gov/espanol/cancer/tratamiento/tipos/inmunoterapia • http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades VERSUS https://www.cancer.gov/espanol/cancer/naturaleza/desigualdades • http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/seguridad-paciente • http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/pago (edited)
priority
spanish cgov missing images and infographics images are missing from spanish pages these got migrated into drupal but they are not appearing on the front end some examples pages include infographic on ac adding videos from versus • versus • • edited
1
683,803
23,394,799,847
IssuesEvent
2022-08-11 21:47:03
savvy-coders/sc-curriculum
https://api.github.com/repos/savvy-coders/sc-curriculum
closed
Investigate Parcel 2.7 potential 404 Error on npm run serve
bug high priority
When trying to build out SPA, ran npm run serve as test to see if it would build, it did build however the browser gave a 404 error. When uninstalling Parcel 2.7 and installed Parcel 2.5, the error seemed to "fix itself". Need to investigate if this was an isolated issue or if conflicts exist with new version of Parcel
1.0
Investigate Parcel 2.7 potential 404 Error on npm run serve - When trying to build out SPA, ran npm run serve as test to see if it would build, it did build however the browser gave a 404 error. When uninstalling Parcel 2.7 and installed Parcel 2.5, the error seemed to "fix itself". Need to investigate if this was an isolated issue or if conflicts exist with new version of Parcel
priority
investigate parcel potential error on npm run serve when trying to build out spa ran npm run serve as test to see if it would build it did build however the browser gave a error when uninstalling parcel and installed parcel the error seemed to fix itself need to investigate if this was an isolated issue or if conflicts exist with new version of parcel
1
240,704
7,804,801,283
IssuesEvent
2018-06-11 08:45:38
nimble-platform/common
https://api.github.com/repos/nimble-platform/common
closed
Each product within catalogue needs an assignment to a concept (domain ontology or eclass)
data schema very high priority
The current indexed data set (catalogue2) contains products which have no assignment to a concrete class. Current proposal is to add the field item_commodity_classification_uri in catalogue2. For release 3: Each product within the catalogue2 needs a value for the field item_commodity_classification_uri
1.0
Each product within catalogue needs an assignment to a concept (domain ontology or eclass) - The current indexed data set (catalogue2) contains products which have no assignment to a concrete class. Current proposal is to add the field item_commodity_classification_uri in catalogue2. For release 3: Each product within the catalogue2 needs a value for the field item_commodity_classification_uri
priority
each product within catalogue needs an assignment to a concept domain ontology or eclass the current indexed data set contains products which have no assignment to a concrete class current proposal is to add the field item commodity classification uri in for release each product within the needs a value for the field item commodity classification uri
1
610,960
18,941,119,878
IssuesEvent
2021-11-18 03:03:50
wso2/product-microgateway
https://api.github.com/repos/wso2/product-microgateway
closed
Error Appling API Key auth for petstore API
Type/Bug Priority/High
### Description: Error Appling API Key auth for petstore API ```log adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event REMOVE_API_FROM_GATEWAY is received adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event DEPLOY_API_IN_GATEWAY is received adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:163] - [synchronizer.FetchAPIsFromControlPlane] [-] API 2653577f-2fb4-4066-887c-0f887f85af76 is added/updated to APIList for label [Default] adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:64] - [synchronizer.FetchAPIs] [-] Fetching APIs from Control Plane. adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:171] - [synchronizer.FetchAPIsFromControlPlane] [-] Pushing data to router and enforcer adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:58] - [synchronizer.PushAPIProjects] [-] Start Deploying 1 API/s... adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:91] - [synchronizer.PushAPIProjects] [-] Start deploying api from file (API_ID:REVISION_ID).zip : 2653577f-2fb4-4066-887c-0f887f85af76-f9073d74-6da6-4d5a-a3cf-4017c930d430.zip adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:230] - [api.ApplyAPIProjectFromAPIM] [-] Deploying api SwaggerPetstore:1.0.5 in Organization carbon.super adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:240] - [api.ApplyAPIProjectFromAPIM] [-] API SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76" already deployed to vhost: localhost adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[petstore_auth:[write:pets read:pets]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[default:[write:pets read:pets]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 ERRO [apis_impl.go:222] - [api.ApplyAPIProjectFromAPIM.func1] [-] Recovered from panic. Error encountered while applying API SwaggerPetstore:1.0.5 to localhost. adapter_1 | 2021-11-03 17:45:03 ERRO [apis_fetcher.go:105] - [synchronizer.PushAPIProjects] [-] Error occurred while applying project SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76" adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:113] - [synchronizer.PushAPIProjects] [-] Successfully deployed 1 API/s ``` ### Steps to reproduce: - Create API from https://petstore.swagger.io/v2/swagger.json - Add API Key auth from runtime API Configs in APIM - Deploy API - check adapter logs ### Affected Product Version: 1.0.0-beta-snapshot ### Environment details (with versions): - OS: mac - Client: curl - Env (Docker/K8s): docker compose --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
1.0
Error Appling API Key auth for petstore API - ### Description: Error Appling API Key auth for petstore API ```log adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event REMOVE_API_FROM_GATEWAY is received adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event DEPLOY_API_IN_GATEWAY is received adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:163] - [synchronizer.FetchAPIsFromControlPlane] [-] API 2653577f-2fb4-4066-887c-0f887f85af76 is added/updated to APIList for label [Default] adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:64] - [synchronizer.FetchAPIs] [-] Fetching APIs from Control Plane. adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:171] - [synchronizer.FetchAPIsFromControlPlane] [-] Pushing data to router and enforcer adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:58] - [synchronizer.PushAPIProjects] [-] Start Deploying 1 API/s... adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:91] - [synchronizer.PushAPIProjects] [-] Start deploying api from file (API_ID:REVISION_ID).zip : 2653577f-2fb4-4066-887c-0f887f85af76-f9073d74-6da6-4d5a-a3cf-4017c930d430.zip adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:230] - [api.ApplyAPIProjectFromAPIM] [-] Deploying api SwaggerPetstore:1.0.5 in Organization carbon.super adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:240] - [api.ApplyAPIProjectFromAPIM] [-] API SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76" already deployed to vhost: localhost adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[petstore_auth:[write:pets read:pets]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[default:[write:pets read:pets]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }]. adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]] adapter_1 | 2021-11-03 17:45:03 ERRO [apis_impl.go:222] - [api.ApplyAPIProjectFromAPIM.func1] [-] Recovered from panic. Error encountered while applying API SwaggerPetstore:1.0.5 to localhost. adapter_1 | 2021-11-03 17:45:03 ERRO [apis_fetcher.go:105] - [synchronizer.PushAPIProjects] [-] Error occurred while applying project SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76" adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:113] - [synchronizer.PushAPIProjects] [-] Successfully deployed 1 API/s ``` ### Steps to reproduce: - Create API from https://petstore.swagger.io/v2/swagger.json - Add API Key auth from runtime API Configs in APIM - Deploy API - check adapter logs ### Affected Product Version: 1.0.0-beta-snapshot ### Environment details (with versions): - OS: mac - Client: curl - Env (Docker/K8s): docker compose --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
priority
error appling api key auth for petstore api description error appling api key auth for petstore api log adapter info event remove api from gateway is received adapter info event deploy api in gateway is received adapter info api is added updated to apilist for label adapter info fetching apis from control plane adapter info pushing data to router and enforcer adapter info start deploying api s adapter info start deploying api from file api id revision id zip zip adapter info deploying api swaggerpetstore in organization carbon super adapter info api swaggerpetstore with uuid already deployed to vhost localhost adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info new method key value map adapter erro recovered from panic error encountered while applying api swaggerpetstore to localhost adapter erro error occurred while applying project swaggerpetstore with uuid adapter info successfully deployed api s steps to reproduce create api from add api key auth from runtime api configs in apim deploy api check adapter logs affected product version beta snapshot environment details with versions os mac client curl env docker docker compose optional fields related issues suggested labels suggested assignees
1
149,889
5,730,556,549
IssuesEvent
2017-04-21 09:44:13
RestComm/mediaserver
https://api.github.com/repos/RestComm/mediaserver
closed
MGCP Channel must filter incoming packets
High-Priority MGCP2 netty
Add a network guard to the new MGCP Channel pipeline that filters incoming packets. Only packets coming from local network are accepted, while packets coming from unknown external networks are discarded.
1.0
MGCP Channel must filter incoming packets - Add a network guard to the new MGCP Channel pipeline that filters incoming packets. Only packets coming from local network are accepted, while packets coming from unknown external networks are discarded.
priority
mgcp channel must filter incoming packets add a network guard to the new mgcp channel pipeline that filters incoming packets only packets coming from local network are accepted while packets coming from unknown external networks are discarded
1
317,619
9,667,001,322
IssuesEvent
2019-05-21 12:16:34
sunpy/sunpy
https://api.github.com/repos/sunpy/sunpy
closed
Prepare for diff-rotations from different points of view.
Effort High Feature Request Package Intermediate Priority Medium coordinates
I've got this from a [draft document](https://issues.cosmos.esa.int/solarorbiterwiki/download/attachments/5801215/Triplet-TN%20SOL-SGS-TN-0020%20v0_2.pdf?version=1&modificationDate=1499950338000&api=v2) from [Solar Orbiter team](https://issues.cosmos.esa.int/solarorbiterwiki/display/SOSP/SOC+Documents): > 4.3 SOC handling of the differential rotation > SOC will use the following model of differential rotation to propagate from the triplet > epoch: > ω(Φ) = A + B sin2(Φ) + C sin4(Φ) > Where ω is the rotation rate (in deg/day) > And Φ is the solar latitude > > SOC will choose one of two sets of parameters: > For magnetic features, meaning sunspots/active regions (derived from “Magnetic” > in [DIFF]) > > A = 14.252 > B = -1.678 > C = -2.401 > For non-magnetic features, e.g. coronal holes > A = 14.705 > B = 0.0 > C = 0.0 > Note that this is a “rigid” rotation corresponding to 26.24 day synodic period from > Earth (=> 24.48 day sidereal period). None of the current values we've got matches what they are planing. ```python >>> howard.to(u.deg / u.day) <Quantity [ 14.32632838, -2.11875209, -1.83163148] deg / d> >>> snodgrass.to(u.deg / u.day) <Quantity [ 14.1134631 , -1.69797189, -2.34646844] deg / d> >>> allen.to(u.deg / u.day) <Quantity [ 14.44, -3. , 0. ] deg / d> ``` We do have a `synodic` correction as an option: `rotation -= 0.9856 * u.deg / u.day * duration`
1.0
Prepare for diff-rotations from different points of view. - I've got this from a [draft document](https://issues.cosmos.esa.int/solarorbiterwiki/download/attachments/5801215/Triplet-TN%20SOL-SGS-TN-0020%20v0_2.pdf?version=1&modificationDate=1499950338000&api=v2) from [Solar Orbiter team](https://issues.cosmos.esa.int/solarorbiterwiki/display/SOSP/SOC+Documents): > 4.3 SOC handling of the differential rotation > SOC will use the following model of differential rotation to propagate from the triplet > epoch: > ω(Φ) = A + B sin2(Φ) + C sin4(Φ) > Where ω is the rotation rate (in deg/day) > And Φ is the solar latitude > > SOC will choose one of two sets of parameters: > For magnetic features, meaning sunspots/active regions (derived from “Magnetic” > in [DIFF]) > > A = 14.252 > B = -1.678 > C = -2.401 > For non-magnetic features, e.g. coronal holes > A = 14.705 > B = 0.0 > C = 0.0 > Note that this is a “rigid” rotation corresponding to 26.24 day synodic period from > Earth (=> 24.48 day sidereal period). None of the current values we've got matches what they are planing. ```python >>> howard.to(u.deg / u.day) <Quantity [ 14.32632838, -2.11875209, -1.83163148] deg / d> >>> snodgrass.to(u.deg / u.day) <Quantity [ 14.1134631 , -1.69797189, -2.34646844] deg / d> >>> allen.to(u.deg / u.day) <Quantity [ 14.44, -3. , 0. ] deg / d> ``` We do have a `synodic` correction as an option: `rotation -= 0.9856 * u.deg / u.day * duration`
priority
prepare for diff rotations from different points of view i ve got this from a from soc handling of the differential rotation soc will use the following model of differential rotation to propagate from the triplet epoch ω φ a b φ c φ where ω is the rotation rate in deg day and φ is the solar latitude soc will choose one of two sets of parameters for magnetic features meaning sunspots active regions derived from “magnetic” in a b c for non magnetic features e g coronal holes a b c note that this is a “rigid” rotation corresponding to day synodic period from earth day sidereal period none of the current values we ve got matches what they are planing python howard to u deg u day snodgrass to u deg u day allen to u deg u day we do have a synodic correction as an option rotation u deg u day duration
1
513,964
14,930,045,222
IssuesEvent
2021-01-25 01:43:21
grannypron/uaf_levels
https://api.github.com/repos/grannypron/uaf_levels
closed
Error message when Quick is used in combat
Waiting For Approval bug game39 high priority
Get the following error when Q is pressed in combat for Quick: Un-implemented code Error Code = 00551B8C ![uaf_00132](https://user-images.githubusercontent.com/19980049/104625168-71be1b00-5662-11eb-8b42-90799213e5ed.png)
1.0
Error message when Quick is used in combat - Get the following error when Q is pressed in combat for Quick: Un-implemented code Error Code = 00551B8C ![uaf_00132](https://user-images.githubusercontent.com/19980049/104625168-71be1b00-5662-11eb-8b42-90799213e5ed.png)
priority
error message when quick is used in combat get the following error when q is pressed in combat for quick un implemented code error code
1
447,153
12,884,600,776
IssuesEvent
2020-07-13 03:34:17
mistifiedwarrior/Hackthon-Project
https://api.github.com/repos/mistifiedwarrior/Hackthon-Project
closed
Shopkeeper: see all bookings
easy high priority
As a shopkeeper I want to see my all bookings So that i can predict the today's selling Acceptance criteria - [x] criteria 1 Given in shopkeeper home page When i click on the bookings Then i can see all bookings - [x] criteria 2 Given in booking details When i want to look at the booking history Then i can see all history
1.0
Shopkeeper: see all bookings - As a shopkeeper I want to see my all bookings So that i can predict the today's selling Acceptance criteria - [x] criteria 1 Given in shopkeeper home page When i click on the bookings Then i can see all bookings - [x] criteria 2 Given in booking details When i want to look at the booking history Then i can see all history
priority
shopkeeper see all bookings as a shopkeeper i want to see my all bookings so that i can predict the today s selling acceptance criteria criteria given in shopkeeper home page when i click on the bookings then i can see all bookings criteria given in booking details when i want to look at the booking history then i can see all history
1
338,817
10,237,788,662
IssuesEvent
2019-08-19 14:35:15
IBM/carbon-addons-iot-react
https://api.github.com/repos/IBM/carbon-addons-iot-react
closed
[v2]: Upgrade Components to Carbon v10
:computer: Development :fire: High priority v2
Containing issue for the migration of our components to Carbon v10 High level steps/pieces: - [x] upgrade carbon deps - [x] upgrade storybook to v5 to support new prop-type definitions - [x] Update/Fix Components - [x] `ButtonEnhanced` refactored to `Button` wrapper - [ ] ~`Table` renamed to `DataTable`, ensure sub components exported in the same fashion as Carbon to support custom composition of components. Avoid any breaking change from Carbon's documentation/usage for `DataTable`~ - [x] review visually for defects, fix as necessary - [ ] ... Maximo needs it and we need to support downstream teams
1.0
[v2]: Upgrade Components to Carbon v10 - Containing issue for the migration of our components to Carbon v10 High level steps/pieces: - [x] upgrade carbon deps - [x] upgrade storybook to v5 to support new prop-type definitions - [x] Update/Fix Components - [x] `ButtonEnhanced` refactored to `Button` wrapper - [ ] ~`Table` renamed to `DataTable`, ensure sub components exported in the same fashion as Carbon to support custom composition of components. Avoid any breaking change from Carbon's documentation/usage for `DataTable`~ - [x] review visually for defects, fix as necessary - [ ] ... Maximo needs it and we need to support downstream teams
priority
upgrade components to carbon containing issue for the migration of our components to carbon high level steps pieces upgrade carbon deps upgrade storybook to to support new prop type definitions update fix components buttonenhanced refactored to button wrapper table renamed to datatable ensure sub components exported in the same fashion as carbon to support custom composition of components avoid any breaking change from carbon s documentation usage for datatable review visually for defects fix as necessary maximo needs it and we need to support downstream teams
1
133,397
5,202,535,556
IssuesEvent
2017-01-24 09:52:25
Promact/promact-oauth-server
https://api.github.com/repos/Promact/promact-oauth-server
closed
Restructure OAuth External Login Flow
Done high-priority OAuth Ready Support
Right now its not flowing the rule of OAuth 2.0 flow. So I need to restructure the OAuth flow and implement flow as OAuth 2.0 rules. Following this link - https://www.digitalocean.com/community/tutorials/an-introduction-to-oauth-2 This is the current flow: 1. When user click the link, Login with Promact, then it will call the point 2 2. Authorization Code Link : The above link call a API and it will redirect to https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId 3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId will be done then user need to login. 4. OAuth Request to Application : After successfully completion of point 3, An request(server side call) will be send from OAuth to application with refresh token for getting app’s secret and redirect Uri then OAuth will receive the response and check for app secret. 5. Application Receives User Details and Access token : After successfully completion of point 4, Application will receive user details and add user to application and store the data. My proposal for new structure: 1. When user click the link, Login with Promact, then it will call the point 2. 2. Authorization Code Link : The above link call a API and it will redirect to https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId&redirectUri=redirectUri Example: https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId={PromactAppClientId}&redirectUri=”https://promactslack.azurewebsites.net/OAuth/Authorization?code={AuthorizationCode}” PromactAppClientId = Promact app’s client Id redirectUri = external login call Url 3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId and redirectUri will be done then user need to login. 4. Application Receives Authorization Code : After successfully completion of point 3, server will redirect to Point 2’s redirectUri with authorization code. Example : https://promactslack.azurewebsites.net/OAuth/Authorization?code=DFSD45FGD45FG11 5. Application Requests Access Token : After successfully completion of point 4, Server will call a API and application will request the OAuth for access token and user details Http Request Url will be like this https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId=PromactAppClientId&authorizationcode=AuthorizationCode&clientSecret=PromactAppClientSecret Example : https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId={PromactAppClientId}&authorizationcode={AuthorizationCode}&clientSecret={PromactAppClientSecret} PromactAppClientId=Promact app’s clientId AuthorizationCode = Point 4 code PromactAppClientSecret – Promact app’s client secret 6. Application Receives Access Token : After successfully completion of point 5 and authorization code, client secret and details match then, Application will receive all required details of user and create and user for external login and store its access token of Promact.
1.0
Restructure OAuth External Login Flow - Right now its not flowing the rule of OAuth 2.0 flow. So I need to restructure the OAuth flow and implement flow as OAuth 2.0 rules. Following this link - https://www.digitalocean.com/community/tutorials/an-introduction-to-oauth-2 This is the current flow: 1. When user click the link, Login with Promact, then it will call the point 2 2. Authorization Code Link : The above link call a API and it will redirect to https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId 3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId will be done then user need to login. 4. OAuth Request to Application : After successfully completion of point 3, An request(server side call) will be send from OAuth to application with refresh token for getting app’s secret and redirect Uri then OAuth will receive the response and check for app secret. 5. Application Receives User Details and Access token : After successfully completion of point 4, Application will receive user details and add user to application and store the data. My proposal for new structure: 1. When user click the link, Login with Promact, then it will call the point 2. 2. Authorization Code Link : The above link call a API and it will redirect to https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId&redirectUri=redirectUri Example: https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId={PromactAppClientId}&redirectUri=”https://promactslack.azurewebsites.net/OAuth/Authorization?code={AuthorizationCode}” PromactAppClientId = Promact app’s client Id redirectUri = external login call Url 3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId and redirectUri will be done then user need to login. 4. Application Receives Authorization Code : After successfully completion of point 3, server will redirect to Point 2’s redirectUri with authorization code. Example : https://promactslack.azurewebsites.net/OAuth/Authorization?code=DFSD45FGD45FG11 5. Application Requests Access Token : After successfully completion of point 4, Server will call a API and application will request the OAuth for access token and user details Http Request Url will be like this https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId=PromactAppClientId&authorizationcode=AuthorizationCode&clientSecret=PromactAppClientSecret Example : https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId={PromactAppClientId}&authorizationcode={AuthorizationCode}&clientSecret={PromactAppClientSecret} PromactAppClientId=Promact app’s clientId AuthorizationCode = Point 4 code PromactAppClientSecret – Promact app’s client secret 6. Application Receives Access Token : After successfully completion of point 5 and authorization code, client secret and details match then, Application will receive all required details of user and create and user for external login and store its access token of Promact.
priority
restructure oauth external login flow right now its not flowing the rule of oauth flow so i need to restructure the oauth flow and implement flow as oauth rules following this link this is the current flow when user click the link login with promact then it will call the point authorization code link the above link call a api and it will redirect to user authorizes application after successfully completion of point server will redirect to oauth server in oauth checking of app details with clientid will be done then user need to login oauth request to application after successfully completion of point an request server side call will be send from oauth to application with refresh token for getting app’s secret and redirect uri then oauth will receive the response and check for app secret application receives user details and access token after successfully completion of point application will receive user details and add user to application and store the data my proposal for new structure when user click the link login with promact then it will call the point authorization code link the above link call a api and it will redirect to example promactappclientid promact app’s client id redirecturi external login call url user authorizes application after successfully completion of point server will redirect to oauth server in oauth checking of app details with clientid and redirecturi will be done then user need to login application receives authorization code after successfully completion of point server will redirect to point ’s redirecturi with authorization code example application requests access token after successfully completion of point server will call a api and application will request the oauth for access token and user details http request url will be like this example promactappclientid promact app’s clientid authorizationcode point code promactappclientsecret – promact app’s client secret application receives access token after successfully completion of point and authorization code client secret and details match then application will receive all required details of user and create and user for external login and store its access token of promact
1
309,360
9,473,837,112
IssuesEvent
2019-04-19 04:13:48
ga4gh/data-repository-service-schemas
https://api.github.com/repos/ga4gh/data-repository-service-schemas
closed
refine the list of AccessMethod type values
Due: Apr Priority: Critical Priority: High Project: DRS
Following up on the initial `type` list for `AccessMethod`, as enumerated in #236: - how do we want to support htsget? My first thought is it should be a `type` = `htsget` of its own, where the `access_url` is required to support the htsget protocol, vs. just being a place to fetch bytes. - do we want our initial `type` list to only include methods that we expect somebody to have implemented by v1 launch, or should we lean further into the hypothetical future? I'm inclined to trim it back to only things that will be live in 2019, once we know what that list is. - can we combine `http` and `https` into one? I expect most servers will only use one of them, and as a client I don't think I would ever have interesting logic to pick one vs. the other -- if I intend to use HTTP, I'll pick that `type` and use whatever `access_url` I'm given.
2.0
refine the list of AccessMethod type values - Following up on the initial `type` list for `AccessMethod`, as enumerated in #236: - how do we want to support htsget? My first thought is it should be a `type` = `htsget` of its own, where the `access_url` is required to support the htsget protocol, vs. just being a place to fetch bytes. - do we want our initial `type` list to only include methods that we expect somebody to have implemented by v1 launch, or should we lean further into the hypothetical future? I'm inclined to trim it back to only things that will be live in 2019, once we know what that list is. - can we combine `http` and `https` into one? I expect most servers will only use one of them, and as a client I don't think I would ever have interesting logic to pick one vs. the other -- if I intend to use HTTP, I'll pick that `type` and use whatever `access_url` I'm given.
priority
refine the list of accessmethod type values following up on the initial type list for accessmethod as enumerated in how do we want to support htsget my first thought is it should be a type htsget of its own where the access url is required to support the htsget protocol vs just being a place to fetch bytes do we want our initial type list to only include methods that we expect somebody to have implemented by launch or should we lean further into the hypothetical future i m inclined to trim it back to only things that will be live in once we know what that list is can we combine http and https into one i expect most servers will only use one of them and as a client i don t think i would ever have interesting logic to pick one vs the other if i intend to use http i ll pick that type and use whatever access url i m given
1
323,103
9,842,860,435
IssuesEvent
2019-06-18 10:12:46
stageosu/Kaguya
https://api.github.com/repos/stageosu/Kaguya
closed
Music player frequently skips songs on its own.
Bug Command Related Help wanted High Priority Music
More often than not, Kaguya will seemingly automatically skip songs that are queued (when it is their turn to play). In the log, it shows that there are no more items in the queue, so that is the reason why it has stopped playing (when it's clearly not the case).
1.0
Music player frequently skips songs on its own. - More often than not, Kaguya will seemingly automatically skip songs that are queued (when it is their turn to play). In the log, it shows that there are no more items in the queue, so that is the reason why it has stopped playing (when it's clearly not the case).
priority
music player frequently skips songs on its own more often than not kaguya will seemingly automatically skip songs that are queued when it is their turn to play in the log it shows that there are no more items in the queue so that is the reason why it has stopped playing when it s clearly not the case
1
790,611
27,830,476,095
IssuesEvent
2023-03-20 04:05:08
AY2223S2-CS2103-F10-2/tp
https://api.github.com/repos/AY2223S2-CS2103-F10-2/tp
opened
As a user, I can navigate to a specific module/lecture and search for contents within it's scope
type.Story priority.High
Integrate nav and find command to work together
1.0
As a user, I can navigate to a specific module/lecture and search for contents within it's scope - Integrate nav and find command to work together
priority
as a user i can navigate to a specific module lecture and search for contents within it s scope integrate nav and find command to work together
1
263,578
8,292,960,972
IssuesEvent
2018-09-20 03:54:52
aspnet/websdk
https://api.github.com/repos/aspnet/websdk
reopened
Missing features when creating a WebDeploy package
High Priority
This is the next step of https://github.com/dotnet/cli/issues/6598. I am now able to create a WebDeploy package using MSBuild v15. However, I am not able to accomplish the same things than before (with ASP.NET "classic" and MSBuild v14): - [ ] How can I add parameters to the `WebApplication.Full.Parameters.xml` file (produced by the build) ? - [x] There is no `.cmd` file generated. - [x] In v14, the parameter `PackageLocation` defined the folder where to generate the package. Now, I have to utilize `DesktopBuildPackageLocation` and specify a file path. The former was very interresting for solutions that have multiple web applications (each web app generates its own package). With this new parameter, the zip file contains both the sites, but configured for only one.
1.0
Missing features when creating a WebDeploy package - This is the next step of https://github.com/dotnet/cli/issues/6598. I am now able to create a WebDeploy package using MSBuild v15. However, I am not able to accomplish the same things than before (with ASP.NET "classic" and MSBuild v14): - [ ] How can I add parameters to the `WebApplication.Full.Parameters.xml` file (produced by the build) ? - [x] There is no `.cmd` file generated. - [x] In v14, the parameter `PackageLocation` defined the folder where to generate the package. Now, I have to utilize `DesktopBuildPackageLocation` and specify a file path. The former was very interresting for solutions that have multiple web applications (each web app generates its own package). With this new parameter, the zip file contains both the sites, but configured for only one.
priority
missing features when creating a webdeploy package this is the next step of i am now able to create a webdeploy package using msbuild however i am not able to accomplish the same things than before with asp net classic and msbuild how can i add parameters to the webapplication full parameters xml file produced by the build there is no cmd file generated in the parameter packagelocation defined the folder where to generate the package now i have to utilize desktopbuildpackagelocation and specify a file path the former was very interresting for solutions that have multiple web applications each web app generates its own package with this new parameter the zip file contains both the sites but configured for only one
1
310,266
9,487,814,890
IssuesEvent
2019-04-22 17:55:57
CosminNechifor/IKHNAIE
https://api.github.com/repos/CosminNechifor/IKHNAIE
closed
Implement a market contract
High Priority
The market contract should take care of keeping track of the ``Components`` that are being sent for sale. Functionalities: **TBD** This is block by the development fo the ``ComponentContract`` #11 #12. The functions ``submitedForSale`` and ``buyComponent`` should be accessible only from the ``Manager`` contract and it should take care if the user who performs the action has enough ``ether`` or is the ``owner`` of the component.
1.0
Implement a market contract - The market contract should take care of keeping track of the ``Components`` that are being sent for sale. Functionalities: **TBD** This is block by the development fo the ``ComponentContract`` #11 #12. The functions ``submitedForSale`` and ``buyComponent`` should be accessible only from the ``Manager`` contract and it should take care if the user who performs the action has enough ``ether`` or is the ``owner`` of the component.
priority
implement a market contract the market contract should take care of keeping track of the components that are being sent for sale functionalities tbd this is block by the development fo the componentcontract the functions submitedforsale and buycomponent should be accessible only from the manager contract and it should take care if the user who performs the action has enough ether or is the owner of the component
1
359,276
10,667,527,206
IssuesEvent
2019-10-19 13:08:26
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Need separate Save button when assigning subscription tiers for an API for better user experience.
3.0.0-beta Priority/Highest Type/Improvement
Create an API and publish it. By default, we assign the Bronze tier. If someone needs it to have unlimited tier only, they need to first tick "Unlimited" tier and then untick the "Bronze" tier. But usually, people will untick the "Bronze tier" first and then tick the tier they want. In such a case, when the Bronze tier is unticked, an error is thrown, saying API should have a tier assigned if it is not in created or if it is not a prototype API. Hence for better UX, a separate Save button would be ideal to save after ticking/unticking the subscription policies.
1.0
Need separate Save button when assigning subscription tiers for an API for better user experience. - Create an API and publish it. By default, we assign the Bronze tier. If someone needs it to have unlimited tier only, they need to first tick "Unlimited" tier and then untick the "Bronze" tier. But usually, people will untick the "Bronze tier" first and then tick the tier they want. In such a case, when the Bronze tier is unticked, an error is thrown, saying API should have a tier assigned if it is not in created or if it is not a prototype API. Hence for better UX, a separate Save button would be ideal to save after ticking/unticking the subscription policies.
priority
need separate save button when assigning subscription tiers for an api for better user experience create an api and publish it by default we assign the bronze tier if someone needs it to have unlimited tier only they need to first tick unlimited tier and then untick the bronze tier but usually people will untick the bronze tier first and then tick the tier they want in such a case when the bronze tier is unticked an error is thrown saying api should have a tier assigned if it is not in created or if it is not a prototype api hence for better ux a separate save button would be ideal to save after ticking unticking the subscription policies
1
164,983
6,259,769,831
IssuesEvent
2017-07-14 18:52:23
wpninjas/ninja-forms
https://api.github.com/repos/wpninjas/ninja-forms
closed
reCaptcha - Submitting form without validating locks the user from submitting.
DIFFICULTY: Involved PRIORITY: High VALUE: Friendly VALUE: Painless
As the title suggests, submitting without validating reCaptcha locks the form with an error, rendering the user unable to submit the form. The only solution in this case is a complete reload of the page, and resubmission whilst appropriately validating reCaptcha.
1.0
reCaptcha - Submitting form without validating locks the user from submitting. - As the title suggests, submitting without validating reCaptcha locks the form with an error, rendering the user unable to submit the form. The only solution in this case is a complete reload of the page, and resubmission whilst appropriately validating reCaptcha.
priority
recaptcha submitting form without validating locks the user from submitting as the title suggests submitting without validating recaptcha locks the form with an error rendering the user unable to submit the form the only solution in this case is a complete reload of the page and resubmission whilst appropriately validating recaptcha
1
554,839
16,440,319,445
IssuesEvent
2021-05-20 13:43:55
anguaive/typelonger
https://api.github.com/repos/anguaive/typelonger
opened
New specification and basic feature set
cleanup high-priority
## To do: - [ ] Examine old specification, rethink what to include and what to not include - [ ] Decide on the basic feature set
1.0
New specification and basic feature set - ## To do: - [ ] Examine old specification, rethink what to include and what to not include - [ ] Decide on the basic feature set
priority
new specification and basic feature set to do examine old specification rethink what to include and what to not include decide on the basic feature set
1
586,436
17,577,481,689
IssuesEvent
2021-08-15 22:09:41
ErnestoFGonzalez/django-amazon-sns-mobile-push-notification
https://api.github.com/repos/ErnestoFGonzalez/django-amazon-sns-mobile-push-notification
closed
Add `badge` argument to `client.Client.publish_to_ios` method
enhancement priority:high difficulty:low
From [Local and Remote Notification Programming Guide: Creating the Remote Notification Payload](https://developer.apple.com/library/archive/documentation/NetworkingInternet/Conceptual/RemoteNotificationsPG/CreatingtheNotificationPayload.html#//apple_ref/doc/uid/TP40008194-CH10-SW1), we see that the `"aps"` entry includes a `"badge"` entry itself, which accepts an integer to set the app's devices badge number. We'll had an optional `badge` argument to `client.Client.publish_to_ios` in order to be able to pass a payload like ``` { "aps" : { "alert" : { "title" : string, "body" : string, "sound": string }, "badge" : number }, "id" : string, "type": string, "serializer": object } ``` In order to pass `badge` down to `publish_to_ios` we nee to add the `badge` param to the following methods: - [ ] `tasks.send_sns_mobile_push_notification_to_device` - [ ] `models.Device.send` - [ ] and finally `client.Client.publish_to_ios`.
1.0
Add `badge` argument to `client.Client.publish_to_ios` method - From [Local and Remote Notification Programming Guide: Creating the Remote Notification Payload](https://developer.apple.com/library/archive/documentation/NetworkingInternet/Conceptual/RemoteNotificationsPG/CreatingtheNotificationPayload.html#//apple_ref/doc/uid/TP40008194-CH10-SW1), we see that the `"aps"` entry includes a `"badge"` entry itself, which accepts an integer to set the app's devices badge number. We'll had an optional `badge` argument to `client.Client.publish_to_ios` in order to be able to pass a payload like ``` { "aps" : { "alert" : { "title" : string, "body" : string, "sound": string }, "badge" : number }, "id" : string, "type": string, "serializer": object } ``` In order to pass `badge` down to `publish_to_ios` we nee to add the `badge` param to the following methods: - [ ] `tasks.send_sns_mobile_push_notification_to_device` - [ ] `models.Device.send` - [ ] and finally `client.Client.publish_to_ios`.
priority
add badge argument to client client publish to ios method from we see that the aps entry includes a badge entry itself which accepts an integer to set the app s devices badge number we ll had an optional badge argument to client client publish to ios in order to be able to pass a payload like aps alert title string body string sound string badge number id string type string serializer object in order to pass badge down to publish to ios we nee to add the badge param to the following methods tasks send sns mobile push notification to device models device send and finally client client publish to ios
1
681,168
23,299,414,375
IssuesEvent
2022-08-07 04:37:26
AlexanderDefuria/FRC-Scouting
https://api.github.com/repos/AlexanderDefuria/FRC-Scouting
opened
Optimize Match Data Loading
Bug High Priority Back End
Currently matchdata is taking a long long time to load. Investigate.
1.0
Optimize Match Data Loading - Currently matchdata is taking a long long time to load. Investigate.
priority
optimize match data loading currently matchdata is taking a long long time to load investigate
1
266,364
8,366,359,087
IssuesEvent
2018-10-04 08:54:24
CS2113-AY1819S1-F10-3/Main
https://api.github.com/repos/CS2113-AY1819S1-F10-3/Main
closed
Morphing and implementing project specific classes
priority.high status.ongoing
Subject,PO and HQP classes (HQP is a sub-class of PO)
1.0
Morphing and implementing project specific classes - Subject,PO and HQP classes (HQP is a sub-class of PO)
priority
morphing and implementing project specific classes subject po and hqp classes hqp is a sub class of po
1
550,229
16,107,446,812
IssuesEvent
2021-04-27 16:31:45
pokt-network/pocket-core
https://api.github.com/repos/pokt-network/pocket-core
closed
Shift Tx Indexer to pocket core
enhancement high priority
**Is your feature request related to a problem? Please describe.** Currently by using tendermint's `txIndexer` can lead to miss indexing of a tx as evidenced in #1188. This happens because the event does not necessarily match the signer. **Describe the solution you'd like** A custom `txIndexer` that allows us to specify how we wanna index and store tx's **Describe alternatives you've considered** Modifying the tx Indexer on tendermint was considered however it would've led to breaking the ABCI standard in order to be able to pass down signer for comparassion of event & signer
1.0
Shift Tx Indexer to pocket core - **Is your feature request related to a problem? Please describe.** Currently by using tendermint's `txIndexer` can lead to miss indexing of a tx as evidenced in #1188. This happens because the event does not necessarily match the signer. **Describe the solution you'd like** A custom `txIndexer` that allows us to specify how we wanna index and store tx's **Describe alternatives you've considered** Modifying the tx Indexer on tendermint was considered however it would've led to breaking the ABCI standard in order to be able to pass down signer for comparassion of event & signer
priority
shift tx indexer to pocket core is your feature request related to a problem please describe currently by using tendermint s txindexer can lead to miss indexing of a tx as evidenced in this happens because the event does not necessarily match the signer describe the solution you d like a custom txindexer that allows us to specify how we wanna index and store tx s describe alternatives you ve considered modifying the tx indexer on tendermint was considered however it would ve led to breaking the abci standard in order to be able to pass down signer for comparassion of event signer
1
443,261
12,769,462,407
IssuesEvent
2020-06-30 03:41:26
ArkEcosystem/core
https://api.github.com/repos/ArkEcosystem/core
closed
Remove container getters
Priority: High
When the migration from awilix to inversify started I added 2 helpers in the form of `app.log` and `app.events`. These should be removed and replaced with injection where the logger or event dispatcher is needed. This will ensure that developers use injection and don't use internal helpers. https://github.com/ArkEcosystem/core/blob/develop/packages/core-kernel/src/application.ts#L340-L360
1.0
Remove container getters - When the migration from awilix to inversify started I added 2 helpers in the form of `app.log` and `app.events`. These should be removed and replaced with injection where the logger or event dispatcher is needed. This will ensure that developers use injection and don't use internal helpers. https://github.com/ArkEcosystem/core/blob/develop/packages/core-kernel/src/application.ts#L340-L360
priority
remove container getters when the migration from awilix to inversify started i added helpers in the form of app log and app events these should be removed and replaced with injection where the logger or event dispatcher is needed this will ensure that developers use injection and don t use internal helpers
1
248,258
7,928,597,720
IssuesEvent
2018-07-06 12:20:36
jncc/topcat
https://api.github.com/repos/jncc/topcat
closed
Copy some updated data files into http://data.jncc.gov.uk/data/
high priority
I will be redacting some data from a large number of data files that are currently stored in http://data.jncc.gov.uk/data/ for access from Data.gov.uk. Once this work is complete I will need developer assistance to replace the existing files with the updated files. Sometime in mid-May would be ideal. Thank you.
1.0
Copy some updated data files into http://data.jncc.gov.uk/data/ - I will be redacting some data from a large number of data files that are currently stored in http://data.jncc.gov.uk/data/ for access from Data.gov.uk. Once this work is complete I will need developer assistance to replace the existing files with the updated files. Sometime in mid-May would be ideal. Thank you.
priority
copy some updated data files into i will be redacting some data from a large number of data files that are currently stored in for access from data gov uk once this work is complete i will need developer assistance to replace the existing files with the updated files sometime in mid may would be ideal thank you
1
525,472
15,254,346,161
IssuesEvent
2021-02-20 11:36:51
epam/Indigo
https://api.github.com/repos/epam/Indigo
closed
Calculate method does not return errors if structure is not valid
Bug High priority
**Step to reproduce** 1. Call indigo.calculate() for [R-Member.zip](https://github.com/epam/Indigo/files/5952295/R-Member.zip) **Expected result** Method throws not empty exception **Actual result** exception is empty
1.0
Calculate method does not return errors if structure is not valid - **Step to reproduce** 1. Call indigo.calculate() for [R-Member.zip](https://github.com/epam/Indigo/files/5952295/R-Member.zip) **Expected result** Method throws not empty exception **Actual result** exception is empty
priority
calculate method does not return errors if structure is not valid step to reproduce call indigo calculate for expected result method throws not empty exception actual result exception is empty
1
570,925
17,023,209,413
IssuesEvent
2021-07-03 00:52:10
microsoft/winget-create
https://api.github.com/repos/microsoft/winget-create
closed
Multiple URL support
High-Priority In-PR Issue-Feature
# Description of the new feature/enhancement Add the possibility to add more than one installer in wingetcreate new and either wingetcreate update # Proposed technical implementation details (optional) ~I think this is a low priority feature request, so keep calm and develop the rest 😆~ I didn't saw the source code, so I'm ignorant
1.0
Multiple URL support - # Description of the new feature/enhancement Add the possibility to add more than one installer in wingetcreate new and either wingetcreate update # Proposed technical implementation details (optional) ~I think this is a low priority feature request, so keep calm and develop the rest 😆~ I didn't saw the source code, so I'm ignorant
priority
multiple url support description of the new feature enhancement add the possibility to add more than one installer in wingetcreate new and either wingetcreate update proposed technical implementation details optional i think this is a low priority feature request so keep calm and develop the rest 😆 i didn t saw the source code so i m ignorant
1
497,814
14,394,341,895
IssuesEvent
2020-12-03 01:03:37
OpenPrinting/cups
https://api.github.com/repos/OpenPrinting/cups
closed
Canon ts6200 with IPP everywhere has duplicated grayscale option in color mode
bug priority-high
os : arch linux cups : 2.3.3op1 I installed canon ts6200 series with IPP everywhere and there are one color and two grayscale options in color mode. previously I installed this printer with IPP everywhere on ubuntu 20.04, it had one color and one grayscale in color mode. Please check that, thanks.
1.0
Canon ts6200 with IPP everywhere has duplicated grayscale option in color mode - os : arch linux cups : 2.3.3op1 I installed canon ts6200 series with IPP everywhere and there are one color and two grayscale options in color mode. previously I installed this printer with IPP everywhere on ubuntu 20.04, it had one color and one grayscale in color mode. Please check that, thanks.
priority
canon with ipp everywhere has duplicated grayscale option in color mode os arch linux cups i installed canon series with ipp everywhere and there are one color and two grayscale options in color mode previously i installed this printer with ipp everywhere on ubuntu it had one color and one grayscale in color mode please check that thanks
1
548,944
16,082,091,978
IssuesEvent
2021-04-26 06:41:11
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.outlook.com - see bug description
browser-fixme ml-needsdiagnosis-false ml-probability-high priority-normal
<!-- @browser: Apple Mail 605.1.15 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/605.1.15 (KHTML, like Gecko) --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/71941 --> **URL**: https://www.outlook.com **Browser / Version**: Apple Mail 605.1.15 **Operating System**: Mac OS X 10.15.6 **Tested Another Browser**: Yes Safari **Problem type**: Something else **Description**: Cannot enlarge text to a readable size **Steps to Reproduce**: I've been trying to read my email (Outlook) on Firefox Focus. My eyesight is not good and the text is too small to read. On Safari I can use finger and thumb to enlarge the page easily, but that feature isn't working on Firefox Focus with Outlook. To make sure I am being clear: I can enlarge text on Firefox, ( just as I can on Safari ) when viewing other pages, but not when viewing Outlook. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.outlook.com - see bug description - <!-- @browser: Apple Mail 605.1.15 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/605.1.15 (KHTML, like Gecko) --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/71941 --> **URL**: https://www.outlook.com **Browser / Version**: Apple Mail 605.1.15 **Operating System**: Mac OS X 10.15.6 **Tested Another Browser**: Yes Safari **Problem type**: Something else **Description**: Cannot enlarge text to a readable size **Steps to Reproduce**: I've been trying to read my email (Outlook) on Firefox Focus. My eyesight is not good and the text is too small to read. On Safari I can use finger and thumb to enlarge the page easily, but that feature isn't working on Firefox Focus with Outlook. To make sure I am being clear: I can enlarge text on Firefox, ( just as I can on Safari ) when viewing other pages, but not when viewing Outlook. <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
see bug description url browser version apple mail operating system mac os x tested another browser yes safari problem type something else description cannot enlarge text to a readable size steps to reproduce i ve been trying to read my email outlook on firefox focus my eyesight is not good and the text is too small to read on safari i can use finger and thumb to enlarge the page easily but that feature isn t working on firefox focus with outlook to make sure i am being clear i can enlarge text on firefox just as i can on safari when viewing other pages but not when viewing outlook browser configuration none from with ❤️
1
675,953
23,112,657,422
IssuesEvent
2022-07-27 14:12:45
icecube/ASTERIA
https://api.github.com/repos/icecube/ASTERIA
closed
Replace internal `Source` and mass mixing classes with models from snewpy.
enhancement high priority
Since [snewpy](https://github.com/SNEWS2/snewpy) has standardized access to a number of CCSN models and flavor transformation scenarios, we should pull it in as a dependency. This would entail deleting the `Source` class in the `source` module, but keeping the functionality of the `photonic_energy_per_vol` member function. About flavor transformations, we can probably completely remove the `oscillations` module and use the transformation classes implemented in snewpy.
1.0
Replace internal `Source` and mass mixing classes with models from snewpy. - Since [snewpy](https://github.com/SNEWS2/snewpy) has standardized access to a number of CCSN models and flavor transformation scenarios, we should pull it in as a dependency. This would entail deleting the `Source` class in the `source` module, but keeping the functionality of the `photonic_energy_per_vol` member function. About flavor transformations, we can probably completely remove the `oscillations` module and use the transformation classes implemented in snewpy.
priority
replace internal source and mass mixing classes with models from snewpy since has standardized access to a number of ccsn models and flavor transformation scenarios we should pull it in as a dependency this would entail deleting the source class in the source module but keeping the functionality of the photonic energy per vol member function about flavor transformations we can probably completely remove the oscillations module and use the transformation classes implemented in snewpy
1
71,758
3,367,799,855
IssuesEvent
2015-11-22 13:57:21
jgirald/ES2015C
https://api.github.com/repos/jgirald/ES2015C
closed
Allow selecting units which are occluded by buildings
Development High Priority Team C
Sometimes, it's possible that a unit is totally or partially occluded by a building, e.g. it's behind a building. While moving the camera provides a solution for this (even tough this feature is not implemented yet, too!), sometimes it's more convenient to be able to select the unit without moving the camera. A way to implement this is the following: Instead of executing a single ray trace to detect the target object when the mouse is clicked, we will execute two ray traces. The first one will exclude the currently selected object from the targets, in order to be able to detect the objects behind the currently selected object. If we are able to detect such an object, we consider that the target of the click. Otherwise, the regular ray tracing procedure is done. Example unit occluded by building: https://i.imgur.com/drGtvNc.png NOTE: Only the selection code will be changed to implement this feature. The display/drawing code will not be changed in order to notify that a unit is behind a building. **Product backlog item:** As a player, I want to see the features of my units/buildings selecting them with clicks. [High Priority] **Estimated effort:** 2h **Ending condition:** When a unit is behind a building, it should be possible to select it by clicking twice over it. The first click will select the building, and the second click will select the unit behind it. Clicking a third time will reselect the building, and a fourth will reselect the unit, etc.
1.0
Allow selecting units which are occluded by buildings - Sometimes, it's possible that a unit is totally or partially occluded by a building, e.g. it's behind a building. While moving the camera provides a solution for this (even tough this feature is not implemented yet, too!), sometimes it's more convenient to be able to select the unit without moving the camera. A way to implement this is the following: Instead of executing a single ray trace to detect the target object when the mouse is clicked, we will execute two ray traces. The first one will exclude the currently selected object from the targets, in order to be able to detect the objects behind the currently selected object. If we are able to detect such an object, we consider that the target of the click. Otherwise, the regular ray tracing procedure is done. Example unit occluded by building: https://i.imgur.com/drGtvNc.png NOTE: Only the selection code will be changed to implement this feature. The display/drawing code will not be changed in order to notify that a unit is behind a building. **Product backlog item:** As a player, I want to see the features of my units/buildings selecting them with clicks. [High Priority] **Estimated effort:** 2h **Ending condition:** When a unit is behind a building, it should be possible to select it by clicking twice over it. The first click will select the building, and the second click will select the unit behind it. Clicking a third time will reselect the building, and a fourth will reselect the unit, etc.
priority
allow selecting units which are occluded by buildings sometimes it s possible that a unit is totally or partially occluded by a building e g it s behind a building while moving the camera provides a solution for this even tough this feature is not implemented yet too sometimes it s more convenient to be able to select the unit without moving the camera a way to implement this is the following instead of executing a single ray trace to detect the target object when the mouse is clicked we will execute two ray traces the first one will exclude the currently selected object from the targets in order to be able to detect the objects behind the currently selected object if we are able to detect such an object we consider that the target of the click otherwise the regular ray tracing procedure is done example unit occluded by building note only the selection code will be changed to implement this feature the display drawing code will not be changed in order to notify that a unit is behind a building product backlog item as a player i want to see the features of my units buildings selecting them with clicks estimated effort ending condition when a unit is behind a building it should be possible to select it by clicking twice over it the first click will select the building and the second click will select the unit behind it clicking a third time will reselect the building and a fourth will reselect the unit etc
1
707,653
24,313,034,011
IssuesEvent
2022-09-30 01:44:51
UC-Davis-molecular-computing/scadnano
https://api.github.com/repos/UC-Davis-molecular-computing/scadnano
closed
fix bug in assigning DNA sequence to strand bound to another strand with an extension
bug high priority closed in dev
Create two bound strands, one with an extension, and assign DNA to the other with "assign complement to bound strands" selected: ![image](https://user-images.githubusercontent.com/19274365/189195166-bffcaf08-884d-42c8-9242-c01c277df677.png) ![image](https://user-images.githubusercontent.com/19274365/189195205-07c71b26-90a6-42ce-96a0-600168b186ea.png) It generates this error: ``` Uncaught (in promise) TypeError: Cannot read properties of null (reading 'length') at merge_wildcards (util.dart:1257:9) at Object.compute_dna_complement_from (assign_or_remove_dna_reducer.dart:198:14) at Object.assign_dna_reducer (assign_or_remove_dna_reducer.dart:89:26) at DevToolsMiddleware.new.assign_dna_middleware (assign_dna.dart:16:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.throttle_middleware (throttle.dart:27:9) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.example_design_selected_middleware (example_design_selected.dart:9:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.export_cadnano_or_codenano_file_middleware (export_cadnano_or_codenano_file.dart:25:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.load_file_middleware (load_file.dart:49:9) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.save_file_middleware (save_file.dart:11:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.export_svg_middleware (export_svg.dart:42:9) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.forbid_create_circular_strand_no_crossovers_middleware$ (forbid_create_circular_strand_no_crossovers_middleware.dart:88:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.move_ensure_all_in_same_helix_group_middleware (move_ensure_same_group.dart:21:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.local_storage_middleware (local_storage.dart:130:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.reset_local_storage_middleware (reset_local_storage.dart:19:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at Store.new.dispatch (store.dart:267:25) at DevToolsStore.new.dispatch (store.dart:74:29) at app.App.new.dispatch (app.dart:140:13) at ask_for_assign_dna_sequence (design_main_strand.dart:1124:7) at ask_for_assign_dna_sequence.next (<anonymous>) at async_patch.dart:45:50 at _RootZone.runUnary (zone.dart:1613:54) at _FutureListener.thenAwait.handleValue (future_impl.dart:155:18) at handleValueCallback (future_impl.dart:707:44) at _Future._propagateToListeners (future_impl.dart:736:13) at [_completeWithValue] (future_impl.dart:542:5) at async._AsyncCallbackEntry.new.callback (future_impl.dart:580:7) at Object._microtaskLoop (schedule_microtask.dart:40:11) at _startMicrotaskLoop (schedule_microtask.dart:49:5) at async_patch.dart:166:15 ```
1.0
fix bug in assigning DNA sequence to strand bound to another strand with an extension - Create two bound strands, one with an extension, and assign DNA to the other with "assign complement to bound strands" selected: ![image](https://user-images.githubusercontent.com/19274365/189195166-bffcaf08-884d-42c8-9242-c01c277df677.png) ![image](https://user-images.githubusercontent.com/19274365/189195205-07c71b26-90a6-42ce-96a0-600168b186ea.png) It generates this error: ``` Uncaught (in promise) TypeError: Cannot read properties of null (reading 'length') at merge_wildcards (util.dart:1257:9) at Object.compute_dna_complement_from (assign_or_remove_dna_reducer.dart:198:14) at Object.assign_dna_reducer (assign_or_remove_dna_reducer.dart:89:26) at DevToolsMiddleware.new.assign_dna_middleware (assign_dna.dart:16:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.throttle_middleware (throttle.dart:27:9) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.example_design_selected_middleware (example_design_selected.dart:9:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.export_cadnano_or_codenano_file_middleware (export_cadnano_or_codenano_file.dart:25:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.load_file_middleware (load_file.dart:49:9) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.save_file_middleware (save_file.dart:11:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.export_svg_middleware (export_svg.dart:42:9) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.forbid_create_circular_strand_no_crossovers_middleware$ (forbid_create_circular_strand_no_crossovers_middleware.dart:88:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.move_ensure_all_in_same_helix_group_middleware (move_ensure_same_group.dart:21:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.local_storage_middleware (local_storage.dart:130:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at middleware.dart:35:20 at DevToolsMiddleware.new.reset_local_storage_middleware (reset_local_storage.dart:19:7) at DevToolsMiddleware.new.call (middleware.dart:39:25) at store.dart:255:43 at Store.new.dispatch (store.dart:267:25) at DevToolsStore.new.dispatch (store.dart:74:29) at app.App.new.dispatch (app.dart:140:13) at ask_for_assign_dna_sequence (design_main_strand.dart:1124:7) at ask_for_assign_dna_sequence.next (<anonymous>) at async_patch.dart:45:50 at _RootZone.runUnary (zone.dart:1613:54) at _FutureListener.thenAwait.handleValue (future_impl.dart:155:18) at handleValueCallback (future_impl.dart:707:44) at _Future._propagateToListeners (future_impl.dart:736:13) at [_completeWithValue] (future_impl.dart:542:5) at async._AsyncCallbackEntry.new.callback (future_impl.dart:580:7) at Object._microtaskLoop (schedule_microtask.dart:40:11) at _startMicrotaskLoop (schedule_microtask.dart:49:5) at async_patch.dart:166:15 ```
priority
fix bug in assigning dna sequence to strand bound to another strand with an extension create two bound strands one with an extension and assign dna to the other with assign complement to bound strands selected it generates this error uncaught in promise typeerror cannot read properties of null reading length at merge wildcards util dart at object compute dna complement from assign or remove dna reducer dart at object assign dna reducer assign or remove dna reducer dart at devtoolsmiddleware new assign dna middleware assign dna dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new throttle middleware throttle dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new example design selected middleware example design selected dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new export cadnano or codenano file middleware export cadnano or codenano file dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new load file middleware load file dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new save file middleware save file dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new export svg middleware export svg dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new forbid create circular strand no crossovers middleware forbid create circular strand no crossovers middleware dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new move ensure all in same helix group middleware move ensure same group dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new local storage middleware local storage dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new reset local storage middleware reset local storage dart at devtoolsmiddleware new call middleware dart at store dart at store new dispatch store dart at devtoolsstore new dispatch store dart at app app new dispatch app dart at ask for assign dna sequence design main strand dart at ask for assign dna sequence next at async patch dart at rootzone rununary zone dart at futurelistener thenawait handlevalue future impl dart at handlevaluecallback future impl dart at future propagatetolisteners future impl dart at future impl dart at async asynccallbackentry new callback future impl dart at object microtaskloop schedule microtask dart at startmicrotaskloop schedule microtask dart at async patch dart
1
288,019
8,824,106,942
IssuesEvent
2019-01-02 15:55:52
ScottLogic/data-engineering-generator
https://api.github.com/repos/ScottLogic/data-engineering-generator
closed
Cucumber Exhaustive Combination: GC overhead limit errors thrown for small requests
bug generator high-priority
## Bug Report Exhaustive combination tests across three fields, each with a set of five data options, were run. This was expected to generate 124 rows of data, not expected to be a large request, but failed with a memory error. [Test Example 2.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648567/Test.Example.2.txt) A second test expecting 25 roes from three fields (one set of 1 data option and two sets of five data options) also failed with the same memory error. [Test Example 1.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648569/Test.Example.1.txt) ### Steps to Reproduce: 1. Run tests from the feature file ExhaustiveCombination.feature 2. See that there is a failure for the test **Scenario: Running an exhaustive combination strategy across three fields with five data options each should be successful** 3. Note that the test is long running and then fails with an out of memory error ### Expected Result: These tests should generate the expected 124 / 25 rows of data based on the exhaustive combination strategy. ### Actual Result: The tests fail with a memory error: > java.lang.OutOfMemoryError: GC overhead limit exceeded > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:35) > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.Iterator.forEachRemaining(Iterator.java:116) > at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721) > at java.util.stream.AbstractTask.compute(AbstractTask.java:316) > at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731) > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) > at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401) > at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734) > at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714) > at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233) > at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) > at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) > at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.permute(ExhaustiveCombinationStrategy.java:19) > at com.scottlogic.deg.generator.generation.databags.MultiplexingDataBagSource.generate(MultiplexingDataBagSource.java:17) > at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:379) > > > java.lang.OutOfMemoryError: Java heap space > at java.util.HashMap.resize(HashMap.java:704) > at java.util.HashMap.putVal(HashMap.java:629) > at java.util.HashMap.put(HashMap.java:612) > at com.scottlogic.deg.generator.generation.databags.DataBag$DataBagBuilder.set(DataBag.java:76) > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:42) > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.Iterator.forEachRemaining(Iterator.java:116) > at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721) > at java.util.stream.AbstractTask.compute(AbstractTask.java:316) > at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731) > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) > at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401) > at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734) > at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714) > at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233) > at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) > at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:528) ### Environment: ExhaustiveCombination branch running in full generation strategy & exhaustive combination strategy on IntelliJ. ### Additional Context: Related to existing performance issues.
1.0
Cucumber Exhaustive Combination: GC overhead limit errors thrown for small requests - ## Bug Report Exhaustive combination tests across three fields, each with a set of five data options, were run. This was expected to generate 124 rows of data, not expected to be a large request, but failed with a memory error. [Test Example 2.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648567/Test.Example.2.txt) A second test expecting 25 roes from three fields (one set of 1 data option and two sets of five data options) also failed with the same memory error. [Test Example 1.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648569/Test.Example.1.txt) ### Steps to Reproduce: 1. Run tests from the feature file ExhaustiveCombination.feature 2. See that there is a failure for the test **Scenario: Running an exhaustive combination strategy across three fields with five data options each should be successful** 3. Note that the test is long running and then fails with an out of memory error ### Expected Result: These tests should generate the expected 124 / 25 rows of data based on the exhaustive combination strategy. ### Actual Result: The tests fail with a memory error: > java.lang.OutOfMemoryError: GC overhead limit exceeded > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:35) > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.Iterator.forEachRemaining(Iterator.java:116) > at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721) > at java.util.stream.AbstractTask.compute(AbstractTask.java:316) > at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731) > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) > at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401) > at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734) > at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714) > at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233) > at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) > at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234) > at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.permute(ExhaustiveCombinationStrategy.java:19) > at com.scottlogic.deg.generator.generation.databags.MultiplexingDataBagSource.generate(MultiplexingDataBagSource.java:17) > at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:379) > > > java.lang.OutOfMemoryError: Java heap space > at java.util.HashMap.resize(HashMap.java:704) > at java.util.HashMap.putVal(HashMap.java:629) > at java.util.HashMap.put(HashMap.java:612) > at com.scottlogic.deg.generator.generation.databags.DataBag$DataBagBuilder.set(DataBag.java:76) > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:42) > at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.Iterator.forEachRemaining(Iterator.java:116) > at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747) > at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721) > at java.util.stream.AbstractTask.compute(AbstractTask.java:316) > at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731) > at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289) > at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401) > at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734) > at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714) > at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233) > at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18) > at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193) > at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382) > at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481) > at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471) > at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708) > at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:528) ### Environment: ExhaustiveCombination branch running in full generation strategy & exhaustive combination strategy on IntelliJ. ### Additional Context: Related to existing performance issues.
priority
cucumber exhaustive combination gc overhead limit errors thrown for small requests bug report exhaustive combination tests across three fields each with a set of five data options were run this was expected to generate rows of data not expected to be a large request but failed with a memory error a second test expecting roes from three fields one set of data option and two sets of five data options also failed with the same memory error steps to reproduce run tests from the feature file exhaustivecombination feature see that there is a failure for the test scenario running an exhaustive combination strategy across three fields with five data options each should be successful note that the test is long running and then fails with an out of memory error expected result these tests should generate the expected rows of data based on the exhaustive combination strategy actual result the tests fail with a memory error java lang outofmemoryerror gc overhead limit exceeded at com scottlogic deg generator generation fieldspecfulfiller lambda generate fieldspecfulfiller java at com scottlogic deg generator generation fieldspecfulfiller lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util iterator foreachremaining iterator java at java util spliterators iteratorspliterator foreachremaining spliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream streamspliterators wrappingspliterator foreachremaining streamspliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reducetask doleaf reduceops java at java util stream reduceops reducetask doleaf reduceops java at java util stream abstracttask compute abstracttask java at java util concurrent countedcompleter exec countedcompleter java at java util concurrent forkjointask doexec forkjointask java at java util concurrent forkjointask doinvoke forkjointask java at java util concurrent forkjointask invoke forkjointask java at java util stream reduceops reduceop evaluateparallel reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda permute exhaustivecombinationstrategy java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util stream referencepipeline accept referencepipeline java at java util arraylist arraylistspliterator foreachremaining arraylist java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy permute exhaustivecombinationstrategy java at com scottlogic deg generator generation databags multiplexingdatabagsource generate multiplexingdatabagsource java at ✽ the following data should be generated c datagenerator data engineering generator generator src test java com scottlogic deg generator cucumber exhaustivecombination feature java lang outofmemoryerror java heap space at java util hashmap resize hashmap java at java util hashmap putval hashmap java at java util hashmap put hashmap java at com scottlogic deg generator generation databags databag databagbuilder set databag java at com scottlogic deg generator generation fieldspecfulfiller lambda generate fieldspecfulfiller java at com scottlogic deg generator generation fieldspecfulfiller lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util iterator foreachremaining iterator java at java util spliterators iteratorspliterator foreachremaining spliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream streamspliterators wrappingspliterator foreachremaining streamspliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reducetask doleaf reduceops java at java util stream reduceops reducetask doleaf reduceops java at java util stream abstracttask compute abstracttask java at java util concurrent countedcompleter exec countedcompleter java at java util concurrent forkjointask doexec forkjointask java at java util concurrent forkjointask doinvoke forkjointask java at java util concurrent forkjointask invoke forkjointask java at java util stream reduceops reduceop evaluateparallel reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda permute exhaustivecombinationstrategy java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util stream referencepipeline accept referencepipeline java at java util arraylist arraylistspliterator foreachremaining arraylist java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at ✽ the following data should be generated c datagenerator data engineering generator generator src test java com scottlogic deg generator cucumber exhaustivecombination feature environment exhaustivecombination branch running in full generation strategy exhaustive combination strategy on intellij additional context related to existing performance issues
1
143,589
5,520,549,714
IssuesEvent
2017-03-19 06:36:21
nim-lang/Nim
https://api.github.com/repos/nim-lang/Nim
closed
A version of httpclient that uses streams in the stdlib
High Priority Stdlib
I decided I would build a script in `nim` that would download some rather large files, as you do. I wanted to use `httpclient` but it bothered me that there were no streams for the body, just strings. This meant to me that the downloaded data is stored into memory. After reading the source of `httpclient`, this is indeed the case. I can see how this would not be a problem for small files but when downloading rather large files - that is files that would easily deplete any RAM - it would be a bit of a problem. So, in my naive attempt, I changed all usage of strings, when downloading the body, into streams; and here is my gist for it [httpclientstreams.nim](https://gist.github.com/NebulaFox/89926ebfa56840ca49a0202cc39d9de3). I successfully tested my changes on downloading some images and other media with `httpclientstreams`. As a result of using streams, I removed all the `*Content` methods, because as a user, if I just wanted the content, I would provide a `Stream` and forget about the `Response`. As an additional consequence of using streams for the body of the response, I have also removed the `body` field from `Response`. As this was strictly a naive attempt, there are leftover `proc`s from `httpclient` that should be deleted, but in saying that is `httpclientstreams` an acceptable implementation to be included into the stdlib? A couple of other things to note is a benefit of using `streams` is that it would make it possible to process the incoming body of data on the fly. And, I can see `httpclient` easily using the streamed-version encapsulating the methods with `StringStream`, however I do understand that `streams` is not supported for JavaScript.
1.0
A version of httpclient that uses streams in the stdlib - I decided I would build a script in `nim` that would download some rather large files, as you do. I wanted to use `httpclient` but it bothered me that there were no streams for the body, just strings. This meant to me that the downloaded data is stored into memory. After reading the source of `httpclient`, this is indeed the case. I can see how this would not be a problem for small files but when downloading rather large files - that is files that would easily deplete any RAM - it would be a bit of a problem. So, in my naive attempt, I changed all usage of strings, when downloading the body, into streams; and here is my gist for it [httpclientstreams.nim](https://gist.github.com/NebulaFox/89926ebfa56840ca49a0202cc39d9de3). I successfully tested my changes on downloading some images and other media with `httpclientstreams`. As a result of using streams, I removed all the `*Content` methods, because as a user, if I just wanted the content, I would provide a `Stream` and forget about the `Response`. As an additional consequence of using streams for the body of the response, I have also removed the `body` field from `Response`. As this was strictly a naive attempt, there are leftover `proc`s from `httpclient` that should be deleted, but in saying that is `httpclientstreams` an acceptable implementation to be included into the stdlib? A couple of other things to note is a benefit of using `streams` is that it would make it possible to process the incoming body of data on the fly. And, I can see `httpclient` easily using the streamed-version encapsulating the methods with `StringStream`, however I do understand that `streams` is not supported for JavaScript.
priority
a version of httpclient that uses streams in the stdlib i decided i would build a script in nim that would download some rather large files as you do i wanted to use httpclient but it bothered me that there were no streams for the body just strings this meant to me that the downloaded data is stored into memory after reading the source of httpclient this is indeed the case i can see how this would not be a problem for small files but when downloading rather large files that is files that would easily deplete any ram it would be a bit of a problem so in my naive attempt i changed all usage of strings when downloading the body into streams and here is my gist for it i successfully tested my changes on downloading some images and other media with httpclientstreams as a result of using streams i removed all the content methods because as a user if i just wanted the content i would provide a stream and forget about the response as an additional consequence of using streams for the body of the response i have also removed the body field from response as this was strictly a naive attempt there are leftover proc s from httpclient that should be deleted but in saying that is httpclientstreams an acceptable implementation to be included into the stdlib a couple of other things to note is a benefit of using streams is that it would make it possible to process the incoming body of data on the fly and i can see httpclient easily using the streamed version encapsulating the methods with stringstream however i do understand that streams is not supported for javascript
1
235,966
7,744,442,565
IssuesEvent
2018-05-29 15:24:00
Gloirin/m2gTest
https://api.github.com/repos/Gloirin/m2gTest
closed
0004932: add id and label to filter objects + filter groups
Feature Request Tinebase high priority
**Reported by pschuele on 6 Oct 2011 17:28** add id and label to filter objects + filter groups - label should be saved in persistent filter, id not
1.0
0004932: add id and label to filter objects + filter groups - **Reported by pschuele on 6 Oct 2011 17:28** add id and label to filter objects + filter groups - label should be saved in persistent filter, id not
priority
add id and label to filter objects filter groups reported by pschuele on oct add id and label to filter objects filter groups label should be saved in persistent filter id not
1
174,236
6,538,052,542
IssuesEvent
2017-09-01 02:57:32
enforcer574/smashclub
https://api.github.com/repos/enforcer574/smashclub
opened
Investigate use of special ASCII characters in profile fields
Incident Priority 2 (High) Vulnerability
One account has used a special HTML character in their catch phrase. This is most likely not an issue, but it warrants an investigation into how the current input sanitization handles these characters and if it presents an injection vulnerability.
1.0
Investigate use of special ASCII characters in profile fields - One account has used a special HTML character in their catch phrase. This is most likely not an issue, but it warrants an investigation into how the current input sanitization handles these characters and if it presents an injection vulnerability.
priority
investigate use of special ascii characters in profile fields one account has used a special html character in their catch phrase this is most likely not an issue but it warrants an investigation into how the current input sanitization handles these characters and if it presents an injection vulnerability
1
328,920
10,001,478,950
IssuesEvent
2019-07-12 15:43:29
databio/refgenconf
https://api.github.com/repos/databio/refgenconf
closed
adding genome attributes to config file
enhancement priority-high
related to #7 (but for genomes instead of for assets) what if we want to add some other attributes about a genome? examples include a description, URL to where it came from, what species it comes from, how long it is, etc. maybe the config format should introduce an "assets" attribute under genome so they are not right under it?
1.0
adding genome attributes to config file - related to #7 (but for genomes instead of for assets) what if we want to add some other attributes about a genome? examples include a description, URL to where it came from, what species it comes from, how long it is, etc. maybe the config format should introduce an "assets" attribute under genome so they are not right under it?
priority
adding genome attributes to config file related to but for genomes instead of for assets what if we want to add some other attributes about a genome examples include a description url to where it came from what species it comes from how long it is etc maybe the config format should introduce an assets attribute under genome so they are not right under it
1
787,797
27,731,352,788
IssuesEvent
2023-03-15 08:17:45
AY2223S2-CS2103T-W15-1/tp
https://api.github.com/repos/AY2223S2-CS2103T-W15-1/tp
closed
As a user I want to be able to add a new contact with missing telegram
type.Story priority.High
...so I can add people when I don’t know their telegram
1.0
As a user I want to be able to add a new contact with missing telegram - ...so I can add people when I don’t know their telegram
priority
as a user i want to be able to add a new contact with missing telegram so i can add people when i don’t know their telegram
1
248,195
7,928,255,472
IssuesEvent
2018-07-06 10:56:31
checkorg1/test
https://api.github.com/repos/checkorg1/test
closed
Token UI to support display of variants and to show meta-content on hover
Highest Priority
_From sardhanari on 16-07-2017 08:19 IST_ <img width="130" alt="screen shot 2017-07-15 at 10 02 36 pm" src="https://user-images.githubusercontent.com/28209236/28243954-0995c710-69aa-11e7-8a07-5a951d112e4a.png"> Selecting the **chevron**(down arrow symbol) shows the list of variants. Selecting a variant replaces the token. So in the compare page, one can check the association strengths against the different variants. Similarly in the neighborhood page, we could cycle through the different variants to see how it affects the neighborhood. In the neighborhood page, the chevron symbol shows up only for the token whose neighborhood is being looked at. Hovering on the **info icon**(i in a circle) will display the meta-content specific to the token. The info icon should be displayed on any token for which there is meta-content. That way one does not have to go to the neighborhood page explicitly to see the meta-content. HTML/CSS to render the above button design: ```html <div class="chip nearest-word ellipsis draggable-token" id="chip-ranking" nearest-word="pd_1" style=" box-shadow: rgba(200,200,200,1) 1px 0px 0px 0px, rgba(200,200,200,1) -1px 0px 0px 0px, rgba(200,200,200,1) 0px 1px 0px 0px, rgba(200,200,200,1) 0px -1px 0px 0px; "> <div>pd_1</div> <i class="fa fa-chevron-down" aria-hidden="true" style=" line-height: inherit; margin-left: 8px; "></i> <div style=" border-left: 1px solid #f7f7f7; border-right: 1px solid #a2a2a2; margin-left: 5px; "></div> <i class="fa fa-info-circle" aria-hidden="true" style=" line-height: inherit; font-size: 15px; margin-left: 9px; x solid #38546d; "></i> </div> ``` **Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/109]
1.0
Token UI to support display of variants and to show meta-content on hover - _From sardhanari on 16-07-2017 08:19 IST_ <img width="130" alt="screen shot 2017-07-15 at 10 02 36 pm" src="https://user-images.githubusercontent.com/28209236/28243954-0995c710-69aa-11e7-8a07-5a951d112e4a.png"> Selecting the **chevron**(down arrow symbol) shows the list of variants. Selecting a variant replaces the token. So in the compare page, one can check the association strengths against the different variants. Similarly in the neighborhood page, we could cycle through the different variants to see how it affects the neighborhood. In the neighborhood page, the chevron symbol shows up only for the token whose neighborhood is being looked at. Hovering on the **info icon**(i in a circle) will display the meta-content specific to the token. The info icon should be displayed on any token for which there is meta-content. That way one does not have to go to the neighborhood page explicitly to see the meta-content. HTML/CSS to render the above button design: ```html <div class="chip nearest-word ellipsis draggable-token" id="chip-ranking" nearest-word="pd_1" style=" box-shadow: rgba(200,200,200,1) 1px 0px 0px 0px, rgba(200,200,200,1) -1px 0px 0px 0px, rgba(200,200,200,1) 0px 1px 0px 0px, rgba(200,200,200,1) 0px -1px 0px 0px; "> <div>pd_1</div> <i class="fa fa-chevron-down" aria-hidden="true" style=" line-height: inherit; margin-left: 8px; "></i> <div style=" border-left: 1px solid #f7f7f7; border-right: 1px solid #a2a2a2; margin-left: 5px; "></div> <i class="fa fa-info-circle" aria-hidden="true" style=" line-height: inherit; font-size: 15px; margin-left: 9px; x solid #38546d; "></i> </div> ``` **Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/109]
priority
token ui to support display of variants and to show meta content on hover from sardhanari on ist img width alt screen shot at pm src selecting the chevron down arrow symbol shows the list of variants selecting a variant replaces the token so in the compare page one can check the association strengths against the different variants similarly in the neighborhood page we could cycle through the different variants to see how it affects the neighborhood in the neighborhood page the chevron symbol shows up only for the token whose neighborhood is being looked at hovering on the info icon i in a circle will display the meta content specific to the token the info icon should be displayed on any token for which there is meta content that way one does not have to go to the neighborhood page explicitly to see the meta content html css to render the above button design html div class chip nearest word ellipsis draggable token id chip ranking nearest word pd style box shadow rgba rgba rgba rgba pd i class fa fa chevron down aria hidden true style line height inherit margin left div style border left solid border right solid margin left i class fa fa info circle aria hidden true style line height inherit font size margin left x solid copied from original issue
1
285,263
8,756,713,256
IssuesEvent
2018-12-14 18:40:35
Datapolitan-Training/data-analysis-r
https://api.github.com/repos/Datapolitan-Training/data-analysis-r
closed
Add explicit sort and filter instructions
High Priority enhancement
Add the code to do the sorting with basic `sort()` or `order()` operations
1.0
Add explicit sort and filter instructions - Add the code to do the sorting with basic `sort()` or `order()` operations
priority
add explicit sort and filter instructions add the code to do the sorting with basic sort or order operations
1
808,036
30,031,256,633
IssuesEvent
2023-06-27 09:44:16
tweag/nixpkgs-graph-explorer
https://api.github.com/repos/tweag/nixpkgs-graph-explorer
closed
Display query error
component:backend component:ui priority:high
**Is your feature request related to a problem? Please describe.** When a request is done using an invalid query, I want to see the error to fix the query. For now, it only displays "Error fetching graph data :( Please check your query and try again later". <img width="704" alt="image" src="https://github.com/tweag/nixpkgs-graph-explorer/assets/1882000/f1a8df02-9feb-49a0-a721-cef717e1df74"> **Describe the solution you'd like** When the query is invalid, I want to see the query error.
1.0
Display query error - **Is your feature request related to a problem? Please describe.** When a request is done using an invalid query, I want to see the error to fix the query. For now, it only displays "Error fetching graph data :( Please check your query and try again later". <img width="704" alt="image" src="https://github.com/tweag/nixpkgs-graph-explorer/assets/1882000/f1a8df02-9feb-49a0-a721-cef717e1df74"> **Describe the solution you'd like** When the query is invalid, I want to see the query error.
priority
display query error is your feature request related to a problem please describe when a request is done using an invalid query i want to see the error to fix the query for now it only displays error fetching graph data please check your query and try again later img width alt image src describe the solution you d like when the query is invalid i want to see the query error
1
724,836
24,943,297,227
IssuesEvent
2022-10-31 20:56:34
bounswe/bounswe2022group4
https://api.github.com/repos/bounswe/bounswe2022group4
closed
Frontend: Lifting State Up Login Information
Category - To Do Priority - High Status: In Progress whom: individual Difficulty - Hard Language - React.js Team - Frontend
Description: After user log in, he/she should not view sign in and sign up buttons anymore. In order to prevent that i need to store login information in a global variable. Steps: 1) Make research about how to lifting a state up in React. 2) Store login information of the user in global variable. 3) Use the variable in navigation bar 4) Hide profile page if user not logged in 5) Hide sign in and sign up button and show sign out button if user logged in Reviewer: @BeratDamar Deadline: 31.10.2022 23.59
1.0
Frontend: Lifting State Up Login Information - Description: After user log in, he/she should not view sign in and sign up buttons anymore. In order to prevent that i need to store login information in a global variable. Steps: 1) Make research about how to lifting a state up in React. 2) Store login information of the user in global variable. 3) Use the variable in navigation bar 4) Hide profile page if user not logged in 5) Hide sign in and sign up button and show sign out button if user logged in Reviewer: @BeratDamar Deadline: 31.10.2022 23.59
priority
frontend lifting state up login information description after user log in he she should not view sign in and sign up buttons anymore in order to prevent that i need to store login information in a global variable steps make research about how to lifting a state up in react store login information of the user in global variable use the variable in navigation bar hide profile page if user not logged in hide sign in and sign up button and show sign out button if user logged in reviewer beratdamar deadline
1
635,114
20,379,366,256
IssuesEvent
2022-02-21 19:23:40
cds-snc/notification-planning
https://api.github.com/repos/cds-snc/notification-planning
opened
Add page to UI when a session times out or user logs out
High Priority | Haute priorité Security | Sécurité UX
_Add page to UI when a session times out or user logs out_ # Description As a (user), I need to be able to do (X) so that I can achieve (Y) outcome. WHY are we building? WHAT are we building? VALUE created by our solution # Acceptance Criteria** (Definition of done) _To be refined through discussion with the team_ Given some context, when (X) action occurs, then (Y) outcome is achieved If this user story emerged from User Research insights: - [ ] Link research insight back to hypothesis from Epic or Objective in Airtable - [ ] Once change/fix/feature is implemented, mark insight as "resolved" in Airtable - [ ] Once change/fix/feature is implements, link insight to design artifacts (Figma) in Airtable * A11y * Bilingualism * Privacy considerations * Security controls in place * Measuring success and metrics
1.0
Add page to UI when a session times out or user logs out - _Add page to UI when a session times out or user logs out_ # Description As a (user), I need to be able to do (X) so that I can achieve (Y) outcome. WHY are we building? WHAT are we building? VALUE created by our solution # Acceptance Criteria** (Definition of done) _To be refined through discussion with the team_ Given some context, when (X) action occurs, then (Y) outcome is achieved If this user story emerged from User Research insights: - [ ] Link research insight back to hypothesis from Epic or Objective in Airtable - [ ] Once change/fix/feature is implemented, mark insight as "resolved" in Airtable - [ ] Once change/fix/feature is implements, link insight to design artifacts (Figma) in Airtable * A11y * Bilingualism * Privacy considerations * Security controls in place * Measuring success and metrics
priority
add page to ui when a session times out or user logs out add page to ui when a session times out or user logs out description as a user i need to be able to do x so that i can achieve y outcome why are we building what are we building value created by our solution acceptance criteria definition of done to be refined through discussion with the team given some context when x action occurs then y outcome is achieved if this user story emerged from user research insights link research insight back to hypothesis from epic or objective in airtable once change fix feature is implemented mark insight as resolved in airtable once change fix feature is implements link insight to design artifacts figma in airtable bilingualism privacy considerations security controls in place measuring success and metrics
1
339,850
10,263,145,722
IssuesEvent
2019-08-22 13:49:05
deep-learning-indaba/Baobab
https://api.github.com/repos/deep-learning-indaba/Baobab
closed
Attendance Admin: User with both normal registration and guest_registration are not indicated as "guests"
High Priority back-end front-end
Attendance Admin: User with both normal registration and guest_registration are not indicated as "Invited Guests" ![image](https://user-images.githubusercontent.com/5547095/63518680-db7bd900-c4f1-11e9-8c97-bb76012e1fb7.png)
1.0
Attendance Admin: User with both normal registration and guest_registration are not indicated as "guests" - Attendance Admin: User with both normal registration and guest_registration are not indicated as "Invited Guests" ![image](https://user-images.githubusercontent.com/5547095/63518680-db7bd900-c4f1-11e9-8c97-bb76012e1fb7.png)
priority
attendance admin user with both normal registration and guest registration are not indicated as guests attendance admin user with both normal registration and guest registration are not indicated as invited guests
1
418,607
12,200,645,654
IssuesEvent
2020-04-30 05:20:34
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
[0.9.0 staging-1531] Strange disconnects
Priority: High Status: Fixed
Can't reproduce by 100%, but today I have it 3 times: 1st - I created laws, 2nd and 3rd I checked work party/contracts. When disconnect start in log exceptions spam: ``` Exception: Expected to receive type key, since there is no view mapping for ID 4475, but type key was not given at Eco.Shared.View.ViewManager.GetView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.UnpackageView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Unity.Properties.TypeConversion+Converter`2+ConvertDelegate[TSource,TDestination].Invoke (TSource value) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Utils.ListExtensions.FromBson[T] (System.Collections.Generic.IList`1[T] list, Eco.Shared.Serialization.BSONArray bsonArray) [0x00000] in <00000000000000000000000000000000>:0 at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0 at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0 at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0 at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0 Rethrow as TargetInvocationException: Exception has been thrown by the target of an invocation. at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0 at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0 at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0 at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0 UnityEngine.Logger:LogException(Exception, Object) UnityEngine.Debug:LogException(Exception) FramePlanner.PlannerGroup:OnUpdate() FramePlanner.FramePlannerSystem:OnUpdate() Unity.Entities.ComponentSystem:InternalUpdate() Unity.Entities.ComponentSystemGroup:OnUpdate() Unity.Entities.ComponentSystem:InternalUpdate() UnityEngine.Experimental.LowLevel.UpdateFunction:Invoke() ``` https://drive.google.com/file/d/1NYKGOIEf2XpbGXydUDLAT3hN550Se_ot/view?usp=sharing then I can't connect to server, i need to restart game
1.0
[0.9.0 staging-1531] Strange disconnects - Can't reproduce by 100%, but today I have it 3 times: 1st - I created laws, 2nd and 3rd I checked work party/contracts. When disconnect start in log exceptions spam: ``` Exception: Expected to receive type key, since there is no view mapping for ID 4475, but type key was not given at Eco.Shared.View.ViewManager.GetView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.UnpackageView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Unity.Properties.TypeConversion+Converter`2+ConvertDelegate[TSource,TDestination].Invoke (TSource value) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Utils.ListExtensions.FromBson[T] (System.Collections.Generic.IList`1[T] list, Eco.Shared.Serialization.BSONArray bsonArray) [0x00000] in <00000000000000000000000000000000>:0 at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0 at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0 at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0 at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0 Rethrow as TargetInvocationException: Exception has been thrown by the target of an invocation. at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0 at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0 at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0 at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0 at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0 at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0 at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0 UnityEngine.Logger:LogException(Exception, Object) UnityEngine.Debug:LogException(Exception) FramePlanner.PlannerGroup:OnUpdate() FramePlanner.FramePlannerSystem:OnUpdate() Unity.Entities.ComponentSystem:InternalUpdate() Unity.Entities.ComponentSystemGroup:OnUpdate() Unity.Entities.ComponentSystem:InternalUpdate() UnityEngine.Experimental.LowLevel.UpdateFunction:Invoke() ``` https://drive.google.com/file/d/1NYKGOIEf2XpbGXydUDLAT3hN550Se_ot/view?usp=sharing then I can't connect to server, i need to restart game
priority
strange disconnects can t reproduce by but today i have it times i created laws and i checked work party contracts when disconnect start in log exceptions spam exception expected to receive type key since there is no view mapping for id but type key was not given at eco shared view viewmanager getview eco shared serialization bsonobject bson in at eco shared view viewmanager unpackageview eco shared serialization bsonobject bson in at unity properties typeconversion converter convertdelegate invoke tsource value in at eco shared serialization bsonmanipulator frombson eco shared serialization bsonvalue value system type type in at eco shared utils listextensions frombson system collections generic ilist list eco shared serialization bsonarray bsonarray in at system reflection monomethod invoke system object obj system reflection bindingflags invokeattr system reflection binder binder system object parameters system globalization cultureinfo culture in at system reflection methodbase invoke system object obj system object parameters in at eco shared serialization bsonmanipulator frombsonilist system type type eco shared serialization bsonvalue value in at eco shared serialization bsonmanipulator frombsonidictionary system type type eco shared serialization bsonobject obj in at eco shared serialization bsonmanipulator frombson eco shared serialization bsonvalue value system type type in at eco shared view view updateproperty eco shared serialization bsonvalue propnameid eco shared serialization bsonobject bson in at eco shared view viewmanager viewnotify eco shared serialization bsonobject bson eco shared view view view system collections generic list notifications in at eco shared view viewmanager applyproperties eco shared view view view eco shared serialization bsonobject bson system collections generic list notifications in at eco shared view viewmanager receiveviewupdate eco shared serialization bsonarray updates in at clientpackethandler receivewhile eco shared networking networkclient client system func condition in at system action invoke t obj in at frameplanner plannergroup onupdate in at frameplanner frameplannersystem onupdate in at unity entities componentsystem internalupdate in at unity entities componentsystemgroup onupdate in at unity entities componentsystem internalupdate in at unityengine experimental lowlevel playerloopsystem updatefunction invoke in rethrow as targetinvocationexception exception has been thrown by the target of an invocation at system reflection monomethod invoke system object obj system reflection bindingflags invokeattr system reflection binder binder system object parameters system globalization cultureinfo culture in at system reflection methodbase invoke system object obj system object parameters in at eco shared serialization bsonmanipulator frombsonilist system type type eco shared serialization bsonvalue value in at eco shared serialization bsonmanipulator frombsonidictionary system type type eco shared serialization bsonobject obj in at eco shared serialization bsonmanipulator frombson eco shared serialization bsonvalue value system type type in at eco shared view view updateproperty eco shared serialization bsonvalue propnameid eco shared serialization bsonobject bson in at eco shared view viewmanager viewnotify eco shared serialization bsonobject bson eco shared view view view system collections generic list notifications in at eco shared view viewmanager applyproperties eco shared view view view eco shared serialization bsonobject bson system collections generic list notifications in at eco shared view viewmanager receiveviewupdate eco shared serialization bsonarray updates in at clientpackethandler receivewhile eco shared networking networkclient client system func condition in at system action invoke t obj in at frameplanner plannergroup onupdate in at frameplanner frameplannersystem onupdate in at unity entities componentsystem internalupdate in at unity entities componentsystemgroup onupdate in at unity entities componentsystem internalupdate in at unityengine experimental lowlevel playerloopsystem updatefunction invoke in unityengine logger logexception exception object unityengine debug logexception exception frameplanner plannergroup onupdate frameplanner frameplannersystem onupdate unity entities componentsystem internalupdate unity entities componentsystemgroup onupdate unity entities componentsystem internalupdate unityengine experimental lowlevel updatefunction invoke then i can t connect to server i need to restart game
1
640,101
20,773,477,543
IssuesEvent
2022-03-16 08:13:52
wso2/product-apim-tooling
https://api.github.com/repos/wso2/product-apim-tooling
closed
apictl certificate has expired
Type/Bug Priority/Highest Affected/4.0.1
$_apictl import api --file ./PetstoreAPI --environment dev_ Login to APIM in dev Username:admin Password: apictl: Error in connecting. Reason: Post "https://localhost:9443/client-registration/v0.17/register": x509: certificate has expired or is not yet valid: current time 2022-01-26T23:07:52+05:30 is after 2022-01-25T07:30:43Z Exit status 1 How to get rid of this error? Using apictl-4.0.1 and wso2am-4.0.0
1.0
apictl certificate has expired - $_apictl import api --file ./PetstoreAPI --environment dev_ Login to APIM in dev Username:admin Password: apictl: Error in connecting. Reason: Post "https://localhost:9443/client-registration/v0.17/register": x509: certificate has expired or is not yet valid: current time 2022-01-26T23:07:52+05:30 is after 2022-01-25T07:30:43Z Exit status 1 How to get rid of this error? Using apictl-4.0.1 and wso2am-4.0.0
priority
apictl certificate has expired apictl import api file petstoreapi environment dev login to apim in dev username admin password apictl error in connecting reason post certificate has expired or is not yet valid current time is after exit status how to get rid of this error using apictl and
1
279,391
8,664,782,218
IssuesEvent
2018-11-28 21:13:35
CzolgIT/servertanksgame
https://api.github.com/repos/CzolgIT/servertanksgame
closed
Receiving and sending UDP data (server side)
enhancement priority: highest
As described above, allow server to send data about the current state of the player.
1.0
Receiving and sending UDP data (server side) - As described above, allow server to send data about the current state of the player.
priority
receiving and sending udp data server side as described above allow server to send data about the current state of the player
1
145,509
5,576,964,436
IssuesEvent
2017-03-28 08:27:01
onaio/onadata
https://api.github.com/repos/onaio/onadata
closed
BacklogLimitExceeded on csv import
Error Priority: High
```python Internal Server Error: /api/v1/forms/[pk]/csv_import.json BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json ea3c17c4-373e-446f-9e29-e5af265f9641 Request Method: GET Request URL: https://api.ona.io/api/v1/forms/[pk]/csv_import.json?job_uuid=ea3c17c4-373e-446f-9e29-e5af265f9641 Django Version: 1.9.5 Python Executable: /..../.virtualenvs/onadata/bin/uwsgi Python Version: 2.7.6 Python Path: ['/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter', '/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter/cWriterow', '.', '', '/..../.virtualenvs/onadata/src/django-nose', '/..../.virtualenvs/onadata/src/django-digest', '/..../.virtualenvs/onadata/src/j2xclient', '/..../.virtualenvs/onadata/src/cors-headers', '/..../.virtualenvs/onadata/src/pricing', '/..../.virtualenvs/onadata/src/savreaderwriter', '/..../.virtualenvs/onadata/src/google-export', '/..../.virtualenvs/onadata/lib/python2.7', '/..../.virtualenvs/onadata/lib/python2.7/plat-x86_64-linux-gnu', '/..../.virtualenvs/onadata/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/lib/python2.7/lib-old', '/..../.virtualenvs/onadata/lib/python2.7/lib-dynload', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/local/lib/python2.7/site-packages'] Server time: Mon, 6 Feb 2017 09:55:35 -0500 Installed Applications: ('google_export', 'pricing', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.messages', 'django.contrib.staticfiles', 'django.contrib.humanize', 'django.contrib.admin', 'django.contrib.admindocs', 'django.contrib.gis', 'registration', 'django_nose', 'django_digest', 'corsheaders', 'oauth2_provider', 'rest_framework', 'rest_framework.authtoken', 'taggit', 'readonly', 'onadata.apps.logger', 'onadata.apps.viewer', 'onadata.apps.main', 'onadata.apps.restservice', 'onadata.apps.api', 'guardian', 'djcelery', 'onadata.apps.sms_support', 'onadata.libs', 'reversion') Installed Middleware: ('multidb.middleware.PinningRouterMiddleware', 'onadata.libs.profiling.sql.SqlTimingMiddleware', 'django.middleware.http.ConditionalGetMiddleware', 'django.middleware.cache.UpdateCacheMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.cache.FetchFromCacheMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'onadata.libs.utils.middleware.LocaleMiddlewareWithTweaks', 'django.middleware.csrf.CsrfViewMiddleware', 'corsheaders.middleware.CorsMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'onadata.libs.utils.middleware.HTTPResponseNotAllowedMiddleware', 'readonly.middleware.DatabaseReadOnlyMiddleware') Traceback: File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response 149. response = self.process_exception_by_middleware(e, request) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response 147. response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/views/decorators/csrf.py" in wrapped_view 58. return view_func(*args, **kwargs) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/viewsets.py" in view 87. return self.dispatch(request, *args, **kwargs) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch 466. response = self.handle_exception(exc) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch 463. response = handler(request, *args, **kwargs) File "./onadata/apps/api/viewsets/xform_viewset.py" in csv_import 549. request.query_params.get('job_uuid'))) File "./onadata/libs/utils/csv_import.py" in get_async_csv_submission_status 303. result = (job.result or job.state) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in result 356. return self._get_task_meta()['result'] File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in _get_task_meta 339. return self._maybe_set_cache(self.backend.get_task_meta(self.id)) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/backends/amqp.py" in get_task_meta 180. raise self.BacklogLimitExceeded(task_id) Exception Type: BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json Exception Value: ea3c17c4-373e-446f-9e29-e5af265f9641 Request information: GET: job_uuid = u'ea3c17c4-373e-446f-9e29-e5af265f9641' POST: No POST data FILES: No FILES data COOKIES: No cookie data ```
1.0
BacklogLimitExceeded on csv import - ```python Internal Server Error: /api/v1/forms/[pk]/csv_import.json BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json ea3c17c4-373e-446f-9e29-e5af265f9641 Request Method: GET Request URL: https://api.ona.io/api/v1/forms/[pk]/csv_import.json?job_uuid=ea3c17c4-373e-446f-9e29-e5af265f9641 Django Version: 1.9.5 Python Executable: /..../.virtualenvs/onadata/bin/uwsgi Python Version: 2.7.6 Python Path: ['/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter', '/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter/cWriterow', '.', '', '/..../.virtualenvs/onadata/src/django-nose', '/..../.virtualenvs/onadata/src/django-digest', '/..../.virtualenvs/onadata/src/j2xclient', '/..../.virtualenvs/onadata/src/cors-headers', '/..../.virtualenvs/onadata/src/pricing', '/..../.virtualenvs/onadata/src/savreaderwriter', '/..../.virtualenvs/onadata/src/google-export', '/..../.virtualenvs/onadata/lib/python2.7', '/..../.virtualenvs/onadata/lib/python2.7/plat-x86_64-linux-gnu', '/..../.virtualenvs/onadata/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/lib/python2.7/lib-old', '/..../.virtualenvs/onadata/lib/python2.7/lib-dynload', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/local/lib/python2.7/site-packages'] Server time: Mon, 6 Feb 2017 09:55:35 -0500 Installed Applications: ('google_export', 'pricing', 'django.contrib.auth', 'django.contrib.contenttypes', 'django.contrib.sessions', 'django.contrib.sites', 'django.contrib.messages', 'django.contrib.staticfiles', 'django.contrib.humanize', 'django.contrib.admin', 'django.contrib.admindocs', 'django.contrib.gis', 'registration', 'django_nose', 'django_digest', 'corsheaders', 'oauth2_provider', 'rest_framework', 'rest_framework.authtoken', 'taggit', 'readonly', 'onadata.apps.logger', 'onadata.apps.viewer', 'onadata.apps.main', 'onadata.apps.restservice', 'onadata.apps.api', 'guardian', 'djcelery', 'onadata.apps.sms_support', 'onadata.libs', 'reversion') Installed Middleware: ('multidb.middleware.PinningRouterMiddleware', 'onadata.libs.profiling.sql.SqlTimingMiddleware', 'django.middleware.http.ConditionalGetMiddleware', 'django.middleware.cache.UpdateCacheMiddleware', 'django.middleware.common.CommonMiddleware', 'django.middleware.cache.FetchFromCacheMiddleware', 'django.contrib.sessions.middleware.SessionMiddleware', 'onadata.libs.utils.middleware.LocaleMiddlewareWithTweaks', 'django.middleware.csrf.CsrfViewMiddleware', 'corsheaders.middleware.CorsMiddleware', 'django.contrib.auth.middleware.AuthenticationMiddleware', 'django.contrib.messages.middleware.MessageMiddleware', 'onadata.libs.utils.middleware.HTTPResponseNotAllowedMiddleware', 'readonly.middleware.DatabaseReadOnlyMiddleware') Traceback: File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response 149. response = self.process_exception_by_middleware(e, request) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response 147. response = wrapped_callback(request, *callback_args, **callback_kwargs) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/views/decorators/csrf.py" in wrapped_view 58. return view_func(*args, **kwargs) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/viewsets.py" in view 87. return self.dispatch(request, *args, **kwargs) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch 466. response = self.handle_exception(exc) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch 463. response = handler(request, *args, **kwargs) File "./onadata/apps/api/viewsets/xform_viewset.py" in csv_import 549. request.query_params.get('job_uuid'))) File "./onadata/libs/utils/csv_import.py" in get_async_csv_submission_status 303. result = (job.result or job.state) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in result 356. return self._get_task_meta()['result'] File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in _get_task_meta 339. return self._maybe_set_cache(self.backend.get_task_meta(self.id)) File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/backends/amqp.py" in get_task_meta 180. raise self.BacklogLimitExceeded(task_id) Exception Type: BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json Exception Value: ea3c17c4-373e-446f-9e29-e5af265f9641 Request information: GET: job_uuid = u'ea3c17c4-373e-446f-9e29-e5af265f9641' POST: No POST data FILES: No FILES data COOKIES: No cookie data ```
priority
backloglimitexceeded on csv import python internal server error api forms csv import json backloglimitexceeded at api forms csv import json request method get request url csv import json job uuid django version python executable virtualenvs onadata bin uwsgi python version python path server time mon feb installed applications google export pricing django contrib auth django contrib contenttypes django contrib sessions django contrib sites django contrib messages django contrib staticfiles django contrib humanize django contrib admin django contrib admindocs django contrib gis registration django nose django digest corsheaders provider rest framework rest framework authtoken taggit readonly onadata apps logger onadata apps viewer onadata apps main onadata apps restservice onadata apps api guardian djcelery onadata apps sms support onadata libs reversion installed middleware multidb middleware pinningroutermiddleware onadata libs profiling sql sqltimingmiddleware django middleware http conditionalgetmiddleware django middleware cache updatecachemiddleware django middleware common commonmiddleware django middleware cache fetchfromcachemiddleware django contrib sessions middleware sessionmiddleware onadata libs utils middleware localemiddlewarewithtweaks django middleware csrf csrfviewmiddleware corsheaders middleware corsmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware onadata libs utils middleware httpresponsenotallowedmiddleware readonly middleware databasereadonlymiddleware traceback file virtualenvs onadata local lib site packages django core handlers base py in get response response self process exception by middleware e request file virtualenvs onadata local lib site packages django core handlers base py in get response response wrapped callback request callback args callback kwargs file virtualenvs onadata local lib site packages django views decorators csrf py in wrapped view return view func args kwargs file virtualenvs onadata local lib site packages rest framework viewsets py in view return self dispatch request args kwargs file virtualenvs onadata local lib site packages rest framework views py in dispatch response self handle exception exc file virtualenvs onadata local lib site packages rest framework views py in dispatch response handler request args kwargs file onadata apps api viewsets xform viewset py in csv import request query params get job uuid file onadata libs utils csv import py in get async csv submission status result job result or job state file virtualenvs onadata local lib site packages celery result py in result return self get task meta file virtualenvs onadata local lib site packages celery result py in get task meta return self maybe set cache self backend get task meta self id file virtualenvs onadata local lib site packages celery backends amqp py in get task meta raise self backloglimitexceeded task id exception type backloglimitexceeded at api forms csv import json exception value request information get job uuid u post no post data files no files data cookies no cookie data
1
735,980
25,451,660,447
IssuesEvent
2022-11-24 10:55:50
owncloud/web
https://api.github.com/repos/owncloud/web
closed
Recipients disappear after re-sharing to the group
Type:Bug Priority:p2-high
ownCloud Web UI 6.0.0-rc.6 Infinite Scale 2.0.0-rc.1+736f513cf Community -local Steps: - admin creates file and shares to marie - marie accepts file and re-shares it to the users and the group Actual: after re-sharing to the group again, all recipients or public links disappear, but still remain after reloading the page https://user-images.githubusercontent.com/84779829/201134209-e8df26b8-2a62-4616-a90d-ac0aba9ae550.mov
1.0
Recipients disappear after re-sharing to the group - ownCloud Web UI 6.0.0-rc.6 Infinite Scale 2.0.0-rc.1+736f513cf Community -local Steps: - admin creates file and shares to marie - marie accepts file and re-shares it to the users and the group Actual: after re-sharing to the group again, all recipients or public links disappear, but still remain after reloading the page https://user-images.githubusercontent.com/84779829/201134209-e8df26b8-2a62-4616-a90d-ac0aba9ae550.mov
priority
recipients disappear after re sharing to the group owncloud web ui rc infinite scale rc community local steps admin creates file and shares to marie marie accepts file and re shares it to the users and the group actual after re sharing to the group again all recipients or public links disappear but still remain after reloading the page
1
73,348
3,410,993,068
IssuesEvent
2015-12-04 22:58:34
rurban/perl-compiler
https://api.github.com/repos/rurban/perl-compiler
opened
POSIX: Constant subroutine MB_CUR_MAX redefined warning
Priority-High
This is similar to issue #335, this is happening in many binaries ``` > perlcc -e 'use POSIX; print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")' pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:15052: warning: integer constant is too large for ‘unsigned long’ type pccOZTKF.c:15053: warning: integer constant is too large for ‘unsigned long’ type Constant subroutine MB_CUR_MAX redefined at ./a.out line 4294967295. ok ``` Notice that if ENV is not reset then we have no warnings... ``` > perl -e 'exec("./a.out")' ok > ./a.out ok ``` one workaround is to disable POSIX import {code} > perlcc -e 'use POSIX (); print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")' ok {code}
1.0
POSIX: Constant subroutine MB_CUR_MAX redefined warning - This is similar to issue #335, this is happening in many binaries ``` > perlcc -e 'use POSIX; print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")' pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90 pccOZTKF.c:15052: warning: integer constant is too large for ‘unsigned long’ type pccOZTKF.c:15053: warning: integer constant is too large for ‘unsigned long’ type Constant subroutine MB_CUR_MAX redefined at ./a.out line 4294967295. ok ``` Notice that if ENV is not reset then we have no warnings... ``` > perl -e 'exec("./a.out")' ok > ./a.out ok ``` one workaround is to disable POSIX import {code} > perlcc -e 'use POSIX (); print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")' ok {code}
priority
posix constant subroutine mb cur max redefined warning this is similar to issue this is happening in many binaries perlcc e use posix print qq ok n perl e local env exec a out pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning integer constant is too large for ‘unsigned long’ type pccoztkf c warning integer constant is too large for ‘unsigned long’ type constant subroutine mb cur max redefined at a out line ok notice that if env is not reset then we have no warnings perl e exec a out ok a out ok one workaround is to disable posix import code perlcc e use posix print qq ok n perl e local env exec a out ok code
1
212,529
7,238,203,450
IssuesEvent
2018-02-13 13:53:41
SELinuxProject/selinux-kernel
https://api.github.com/repos/SELinuxProject/selinux-kernel
opened
kernel softlockup due to too many SIDs/contexts
bug priority/high
As reported by yangjhong1 on selinux list, when too many SIDs/contexts have been allocated (e.g. 300000+ as a result of repeated docker container creations for 2 days), sidtab_search_context becomes very slow and can cause a kernel softlockup warning. docker randomly selects a category pair for every container creation, so this can occur just from creating containers over time, even if old containers are removed promptly (category set reuse for removed containers will eventually occur but each selection is random). It can also occur from any other activity that allocates SIDs/contexts, even those that simply probe for context validity. sidtab_search_context() is a reverse lookup in the sidtab and presently just walks the entire hash table. At a minimum, we need to add a reverse hash table to help mitigate this, possibly using a SELinux hashtab or the core kernel's hashtable.h or rhashtable.h data structures. We might also want a fast check of the context category set to see if it has ever been previously used (i.e. maintain a ebitmap of used categories, and check whether it contains the context's category set) so that we can fail fast on a lookup of a new category set. However, the fact that we might need to support 300000+ SIDs/contexts also suggests that we should likely revisit the sidtab forward hash table since it is too small to efficiently handle that. That too is a candidate to be replaced by e.g. hashtable or rhashtable. I can see both short term and long term fixes for this bug; short term might just be adding simple reverse hash table and perhaps a category ebitmap test; longer term might be reworking the forward hash and switching over to hashtable or rhashtable structures.
1.0
kernel softlockup due to too many SIDs/contexts - As reported by yangjhong1 on selinux list, when too many SIDs/contexts have been allocated (e.g. 300000+ as a result of repeated docker container creations for 2 days), sidtab_search_context becomes very slow and can cause a kernel softlockup warning. docker randomly selects a category pair for every container creation, so this can occur just from creating containers over time, even if old containers are removed promptly (category set reuse for removed containers will eventually occur but each selection is random). It can also occur from any other activity that allocates SIDs/contexts, even those that simply probe for context validity. sidtab_search_context() is a reverse lookup in the sidtab and presently just walks the entire hash table. At a minimum, we need to add a reverse hash table to help mitigate this, possibly using a SELinux hashtab or the core kernel's hashtable.h or rhashtable.h data structures. We might also want a fast check of the context category set to see if it has ever been previously used (i.e. maintain a ebitmap of used categories, and check whether it contains the context's category set) so that we can fail fast on a lookup of a new category set. However, the fact that we might need to support 300000+ SIDs/contexts also suggests that we should likely revisit the sidtab forward hash table since it is too small to efficiently handle that. That too is a candidate to be replaced by e.g. hashtable or rhashtable. I can see both short term and long term fixes for this bug; short term might just be adding simple reverse hash table and perhaps a category ebitmap test; longer term might be reworking the forward hash and switching over to hashtable or rhashtable structures.
priority
kernel softlockup due to too many sids contexts as reported by on selinux list when too many sids contexts have been allocated e g as a result of repeated docker container creations for days sidtab search context becomes very slow and can cause a kernel softlockup warning docker randomly selects a category pair for every container creation so this can occur just from creating containers over time even if old containers are removed promptly category set reuse for removed containers will eventually occur but each selection is random it can also occur from any other activity that allocates sids contexts even those that simply probe for context validity sidtab search context is a reverse lookup in the sidtab and presently just walks the entire hash table at a minimum we need to add a reverse hash table to help mitigate this possibly using a selinux hashtab or the core kernel s hashtable h or rhashtable h data structures we might also want a fast check of the context category set to see if it has ever been previously used i e maintain a ebitmap of used categories and check whether it contains the context s category set so that we can fail fast on a lookup of a new category set however the fact that we might need to support sids contexts also suggests that we should likely revisit the sidtab forward hash table since it is too small to efficiently handle that that too is a candidate to be replaced by e g hashtable or rhashtable i can see both short term and long term fixes for this bug short term might just be adding simple reverse hash table and perhaps a category ebitmap test longer term might be reworking the forward hash and switching over to hashtable or rhashtable structures
1
629,046
20,021,983,715
IssuesEvent
2022-02-01 17:12:18
dtcenter/METplus
https://api.github.com/repos/dtcenter/METplus
closed
Implement various enhancements to climatology settings
type: enhancement component: use case wrapper priority: high requestor: NOAA/EMC alert: NEED ACCOUNT KEY required: FOR OFFICIAL RELEASE METplus: Configuration
Various enhancements are needed to improve the logic for setting climatology variables in the climo_mean and climo_stdev dictionaries. ## Describe the Enhancements ## - [x] Add support for setting a list of field levels in climo variables Setting the following: `GRID_STAT_CLIMO_MEAN_FIELD = {name="UGRD"; level=["P850","P500","P250"];}` Results in an improperly formatted value: METPLUS_CLIMO_MEAN_DICT=climo_mean = {file_name = ["/scratch1/NCEPDEV/global/glopara/fix/fix_verif/climo_fi les/cmean_1d.19591203"];field = [{name="UGRD"; level=["P850",**P500,P250**];}];} Add a test for setting climo field info and enhance the logic to properly format a list of levels. Make sure the behavior works for all apps that use climatology variables. - [x] Add support for using forecast or observation field info for climatology It is often the case that the climatology field info is the same as the forecast field info. If \<APP\>\_CLIMO_[MEAN/STDEV]_FIELD_NAME is not set, then the wrapper should check if \<APP\>\_CLIMO\_[MEAN/STDEV]\_USE\_[FCST/OBS] is set to True. Error out if both are True. If one is True, then use the field info from the appropriate input as the field info for the climo variable - [x] Add support for setting file_type for climo_[mean/stdev] Add optional \<APP\>\_CLIMO_[MEAN/STDEV]_FILE_TYPE to set file_type in the appropriate climo dictionary -- this was implemented in PR #1353 ### Time Estimate ### ~1 day ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [X] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Enhancement Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
1.0
Implement various enhancements to climatology settings - Various enhancements are needed to improve the logic for setting climatology variables in the climo_mean and climo_stdev dictionaries. ## Describe the Enhancements ## - [x] Add support for setting a list of field levels in climo variables Setting the following: `GRID_STAT_CLIMO_MEAN_FIELD = {name="UGRD"; level=["P850","P500","P250"];}` Results in an improperly formatted value: METPLUS_CLIMO_MEAN_DICT=climo_mean = {file_name = ["/scratch1/NCEPDEV/global/glopara/fix/fix_verif/climo_fi les/cmean_1d.19591203"];field = [{name="UGRD"; level=["P850",**P500,P250**];}];} Add a test for setting climo field info and enhance the logic to properly format a list of levels. Make sure the behavior works for all apps that use climatology variables. - [x] Add support for using forecast or observation field info for climatology It is often the case that the climatology field info is the same as the forecast field info. If \<APP\>\_CLIMO_[MEAN/STDEV]_FIELD_NAME is not set, then the wrapper should check if \<APP\>\_CLIMO\_[MEAN/STDEV]\_USE\_[FCST/OBS] is set to True. Error out if both are True. If one is True, then use the field info from the appropriate input as the field info for the climo variable - [x] Add support for setting file_type for climo_[mean/stdev] Add optional \<APP\>\_CLIMO_[MEAN/STDEV]_FILE_TYPE to set file_type in the appropriate climo dictionary -- this was implemented in PR #1353 ### Time Estimate ### ~1 day ### Sub-Issues ### Consider breaking the enhancement down into sub-issues. - [X] *Add a checkbox for each sub-issue here.* ### Relevant Deadlines ### *List relevant project deadlines here or state NONE.* ### Funding Source ### *Define the source of funding and account keys here or state NONE.* ## Define the Metadata ## ### Assignee ### - [ ] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [ ] Select **component(s)** - [ ] Select **priority** - [ ] Select **requestor(s)** ### Projects and Milestone ### - [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label - [ ] Select **Milestone** as the next official version or **Future Versions** ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Enhancement Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **develop**. Branch name: `feature_<Issue Number>_<Description>` - [ ] Complete the development and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **develop**. Pull request: `feature <Issue Number> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Close this issue.
priority
implement various enhancements to climatology settings various enhancements are needed to improve the logic for setting climatology variables in the climo mean and climo stdev dictionaries describe the enhancements add support for setting a list of field levels in climo variables setting the following grid stat climo mean field name ugrd level results in an improperly formatted value metplus climo mean dict climo mean file name ncepdev global glopara fix fix verif climo fi les cmean field add a test for setting climo field info and enhance the logic to properly format a list of levels make sure the behavior works for all apps that use climatology variables add support for using forecast or observation field info for climatology it is often the case that the climatology field info is the same as the forecast field info if climo field name is not set then the wrapper should check if climo use is set to true error out if both are true if one is true then use the field info from the appropriate input as the field info for the climo variable add support for setting file type for climo add optional climo file type to set file type in the appropriate climo dictionary this was implemented in pr time estimate day sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue
1
733,938
25,330,966,316
IssuesEvent
2022-11-18 13:13:35
csmith1188/formbar
https://api.github.com/repos/csmith1188/formbar
opened
Major Bug TUTD
bug high priority frontend backend large
Sometimes, a student will cause the formbar alternate adding or removing a selected "thumb" from a student every second. Each time it plays the sound for adding a thumb (even when it is removing the thumb) and updates the formbar. The permissions of the student makes no difference, so they cannot be blocked by being made a guest. This seems to coincide with the logpoint that logs the students while in the /controlpanel. May be related to new websocket features.
1.0
Major Bug TUTD - Sometimes, a student will cause the formbar alternate adding or removing a selected "thumb" from a student every second. Each time it plays the sound for adding a thumb (even when it is removing the thumb) and updates the formbar. The permissions of the student makes no difference, so they cannot be blocked by being made a guest. This seems to coincide with the logpoint that logs the students while in the /controlpanel. May be related to new websocket features.
priority
major bug tutd sometimes a student will cause the formbar alternate adding or removing a selected thumb from a student every second each time it plays the sound for adding a thumb even when it is removing the thumb and updates the formbar the permissions of the student makes no difference so they cannot be blocked by being made a guest this seems to coincide with the logpoint that logs the students while in the controlpanel may be related to new websocket features
1
601,030
18,363,895,785
IssuesEvent
2021-10-09 18:09:50
alec-roberson/MS-CV2
https://api.github.com/repos/alec-roberson/MS-CV2
closed
Get the training process working on the server
bug good first issue high priority
Being a different machine, the server has inevitably run into various problems and bugs when trying to execute the training code for our networks.
1.0
Get the training process working on the server - Being a different machine, the server has inevitably run into various problems and bugs when trying to execute the training code for our networks.
priority
get the training process working on the server being a different machine the server has inevitably run into various problems and bugs when trying to execute the training code for our networks
1
681,648
23,319,543,446
IssuesEvent
2022-08-08 15:11:29
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Performance numbers decrease due to extra DB queries executing in IS 6.0.0
Priority/Highest Severity/Blocker bug QA-Reported product-performance
**Describe the issue:** Following bold and italic extra DB queries can be observed at the IS 6.0.0 compared to 5.11.0-GA. Currently this is identified as the major reason for performance decrease compared to previous version. IS 6.0.0-alpha2 - "CaseInsensitiveUernames" property set to false - SELECT * FROM IDN_RECOVERY_DATA WHERE LOWER(USER_NAME)=LOWER(?) AND USER_DOMAIN = ? AND TENANT_ID = ? AND SCENARIO NOT IN ('EMAIL_VERIFICATION_ON_UPDATE', 'MOBILE_VERIFICATION_ON_UPDATE') - _**SELECT WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME,WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID FROM WF_REQUEST,WF_REQUEST_ENTITY_RELATIONSHIP WHERE WF_REQUEST.UUID = WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME = ? AND WF_REQUEST.STATUS = ? AND WF_REQUEST.OPERATION_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.TENANT_ID = ?**_ - SELECT UM_USER_ID, UM_USER_NAME, UM_USER_PASSWORD, UM_SALT_VALUE, UM_REQUIRE_CHANGE, UM_CHANGED_TIME FROM UM_USER WHERE LOWER(UM_USER_NAME)=LOWER(?) AND UM_TENANT_ID=? - _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_ - _**SELECT DATA_KEY, DATA_VALUE FROM IDN_IDENTITY_USER_DATA WHERE TENANT_ID = ? AND LOWER(USER_NAME) = LOWER(?)**_ - _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_ [1] https://docs.google.com/spreadsheets/d/1I79C69KFB6lFypLh7esmKxeYMJRl5eQQjd1LyFZweRI/edit?usp=sharing - tab "Comparison-Summery-Detail" **How to reproduce:** 1. Enable correlation logs in IS 6.0.0 and IS 5.11.0-GA product versions - https://is.docs.wso2.com/en/latest/setup/working-with-product-observability/?query=CaseInsensitiveUernames 2. Execute performance test "User authentication for the super tenant" - https://github.com/aaujayasena/performance-is/blob/master/common/jmeter/authenticate/Authenticate_Super_Tenant_User.jmx 3. Compare performance numbers between product versions 4. Compare DB queries filtering correlation ID **Expected behavior:** Performance numbers of IS 6.0.0-alpha2 >= 5.11.0-GA **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: IS 6.0.0-alpha2 and IS 5.11.0-GA - OS: MAC for local testing and ubuntu for aws environment - Database: MSSQL - Userstore: JDBC --- ### Optional Fields **Related issues:** https://github.com/wso2/product-is/issues/14372 **Suggested labels:** <!-- Only to be used by non-members -->
1.0
Performance numbers decrease due to extra DB queries executing in IS 6.0.0 - **Describe the issue:** Following bold and italic extra DB queries can be observed at the IS 6.0.0 compared to 5.11.0-GA. Currently this is identified as the major reason for performance decrease compared to previous version. IS 6.0.0-alpha2 - "CaseInsensitiveUernames" property set to false - SELECT * FROM IDN_RECOVERY_DATA WHERE LOWER(USER_NAME)=LOWER(?) AND USER_DOMAIN = ? AND TENANT_ID = ? AND SCENARIO NOT IN ('EMAIL_VERIFICATION_ON_UPDATE', 'MOBILE_VERIFICATION_ON_UPDATE') - _**SELECT WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME,WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID FROM WF_REQUEST,WF_REQUEST_ENTITY_RELATIONSHIP WHERE WF_REQUEST.UUID = WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME = ? AND WF_REQUEST.STATUS = ? AND WF_REQUEST.OPERATION_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.TENANT_ID = ?**_ - SELECT UM_USER_ID, UM_USER_NAME, UM_USER_PASSWORD, UM_SALT_VALUE, UM_REQUIRE_CHANGE, UM_CHANGED_TIME FROM UM_USER WHERE LOWER(UM_USER_NAME)=LOWER(?) AND UM_TENANT_ID=? - _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_ - _**SELECT DATA_KEY, DATA_VALUE FROM IDN_IDENTITY_USER_DATA WHERE TENANT_ID = ? AND LOWER(USER_NAME) = LOWER(?)**_ - _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_ [1] https://docs.google.com/spreadsheets/d/1I79C69KFB6lFypLh7esmKxeYMJRl5eQQjd1LyFZweRI/edit?usp=sharing - tab "Comparison-Summery-Detail" **How to reproduce:** 1. Enable correlation logs in IS 6.0.0 and IS 5.11.0-GA product versions - https://is.docs.wso2.com/en/latest/setup/working-with-product-observability/?query=CaseInsensitiveUernames 2. Execute performance test "User authentication for the super tenant" - https://github.com/aaujayasena/performance-is/blob/master/common/jmeter/authenticate/Authenticate_Super_Tenant_User.jmx 3. Compare performance numbers between product versions 4. Compare DB queries filtering correlation ID **Expected behavior:** Performance numbers of IS 6.0.0-alpha2 >= 5.11.0-GA **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: IS 6.0.0-alpha2 and IS 5.11.0-GA - OS: MAC for local testing and ubuntu for aws environment - Database: MSSQL - Userstore: JDBC --- ### Optional Fields **Related issues:** https://github.com/wso2/product-is/issues/14372 **Suggested labels:** <!-- Only to be used by non-members -->
priority
performance numbers decrease due to extra db queries executing in is describe the issue following bold and italic extra db queries can be observed at the is compared to ga currently this is identified as the major reason for performance decrease compared to previous version is caseinsensitiveuernames property set to false select from idn recovery data where lower user name lower and user domain and tenant id and scenario not in email verification on update mobile verification on update select wf request entity relationship entity name wf request entity relationship request id from wf request wf request entity relationship where wf request uuid wf request entity relationship request id and wf request entity relationship entity type and wf request entity relationship entity name and wf request status and wf request operation type and wf request entity relationship tenant id select um user id um user name um user password um salt value um require change um changed time from um user where lower um user name lower and um tenant id select um attr name um attr value from um user attribute um user where um user um id um user attribute um user id and um user um user id and um profile id and um user attribute um tenant id and um user um tenant id select data key data value from idn identity user data where tenant id and lower user name lower select um attr name um attr value from um user attribute um user where um user um id um user attribute um user id and um user um user id and um profile id and um user attribute um tenant id and um user um tenant id tab comparison summery detail how to reproduce enable correlation logs in is and is ga product versions execute performance test user authentication for the super tenant compare performance numbers between product versions compare db queries filtering correlation id expected behavior performance numbers of is ga environment information please complete the following information remove any unnecessary fields product version is and is ga os mac for local testing and ubuntu for aws environment database mssql userstore jdbc optional fields related issues suggested labels
1
87,563
3,755,754,261
IssuesEvent
2016-03-12 21:46:58
johnnyflowers/schoolcalendars
https://api.github.com/repos/johnnyflowers/schoolcalendars
closed
Google calendar URLs should be formatted consistently
high priority
We need to be sure that for every google calendar we host, the URL is constructed to include: ?showTitle=1&showNav=1&showDate=1 in the actual URL itself. Otherwise, users will not be able to browse into future months. This is likely something you already did last year, but we have no way of knowing because of the lack of git, heh.
1.0
Google calendar URLs should be formatted consistently - We need to be sure that for every google calendar we host, the URL is constructed to include: ?showTitle=1&showNav=1&showDate=1 in the actual URL itself. Otherwise, users will not be able to browse into future months. This is likely something you already did last year, but we have no way of knowing because of the lack of git, heh.
priority
google calendar urls should be formatted consistently we need to be sure that for every google calendar we host the url is constructed to include showtitle shownav showdate in the actual url itself otherwise users will not be able to browse into future months this is likely something you already did last year but we have no way of knowing because of the lack of git heh
1
829,271
31,862,042,017
IssuesEvent
2023-09-15 11:40:14
Automattic/woocommerce-payments
https://api.github.com/repos/Automattic/woocommerce-payments
opened
Enable Deferred Intent UPE for legacy UPE users
type: enhancement priority: high component: checkout category: projects
### Description Similar to https://github.com/Automattic/woocommerce-payments/issues/7003, but this time instead of newly onboarded merchants, legacy UPE merchants should have the deferred UPE feature flag enabled by default. 💡 The solution for this task **should be merged between Sep 19 and October 6,** so that it is released as part of WooPayments 6.6.0
1.0
Enable Deferred Intent UPE for legacy UPE users - ### Description Similar to https://github.com/Automattic/woocommerce-payments/issues/7003, but this time instead of newly onboarded merchants, legacy UPE merchants should have the deferred UPE feature flag enabled by default. 💡 The solution for this task **should be merged between Sep 19 and October 6,** so that it is released as part of WooPayments 6.6.0
priority
enable deferred intent upe for legacy upe users description similar to but this time instead of newly onboarded merchants legacy upe merchants should have the deferred upe feature flag enabled by default 💡 the solution for this task should be merged between sep and october so that it is released as part of woopayments
1
819,920
30,755,433,802
IssuesEvent
2023-07-29 02:04:40
GC-spigot/AdvancedEnchantments
https://api.github.com/repos/GC-spigot/AdvancedEnchantments
closed
Bug on sacred white parchment
Priority: High Bug: Confirmed Status: In development
### Describe the bug The keep after death is in 'true', the scroll is removed and the item remains. The test I did resulted in the scroll still remaining on the item even after death. ### How to reproduce Tests were carried out through death by drowning, still the item was not dropped but the holy white scroll was not removed from the item ### Screenshots / Videos Test keep after death: false https://drive.google.com/file/d/1-BOYeVFQepheGxEKvt7lakjlR-LsWG7G/view?usp=drivesdk Test keep after death: true https://drive.google.com/file/d/1-YYD5GXm23wlTf24NI-yqid_tKE_-On7/view?usp=drivesdk ### "/ae plinfo" link https://paste.md-5.net/joharazodo ### Server Log https://mcpaste.io/2015c6aaba30d058
1.0
Bug on sacred white parchment - ### Describe the bug The keep after death is in 'true', the scroll is removed and the item remains. The test I did resulted in the scroll still remaining on the item even after death. ### How to reproduce Tests were carried out through death by drowning, still the item was not dropped but the holy white scroll was not removed from the item ### Screenshots / Videos Test keep after death: false https://drive.google.com/file/d/1-BOYeVFQepheGxEKvt7lakjlR-LsWG7G/view?usp=drivesdk Test keep after death: true https://drive.google.com/file/d/1-YYD5GXm23wlTf24NI-yqid_tKE_-On7/view?usp=drivesdk ### "/ae plinfo" link https://paste.md-5.net/joharazodo ### Server Log https://mcpaste.io/2015c6aaba30d058
priority
bug on sacred white parchment describe the bug the keep after death is in true the scroll is removed and the item remains the test i did resulted in the scroll still remaining on the item even after death how to reproduce tests were carried out through death by drowning still the item was not dropped but the holy white scroll was not removed from the item screenshots videos test keep after death false test keep after death true ae plinfo link server log
1
488,009
14,073,374,306
IssuesEvent
2020-11-04 04:38:07
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
SAML back channel SLO not working
Priority/Highest Severity/Critical bug
**Describe the issue:** 1. Configured SAML back channel SLO using the doc [1] 2. Once logout form one application other application doesn't get logged out. 3. Getting following errors in the console. ``` [2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo [2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error ``` [1] https://is.docs.wso2.com/en/5.11.0/learn/handling-saml-single-logout-requests-from-federated-identity-providers/ **How to reproduce:** 1. Configured SAML back channel SLO using the doc [1] 2. Once logout form one application other application doesn't get logged out. 3. Getting following errors in the console. ``` [2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo [2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error ``` **Expected behavior:** First application should automatically logged out **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version:IS-5.11.0-beta2 - OS: Linux - Database: POSTGRE - Userstore: JDBC ---
1.0
SAML back channel SLO not working - **Describe the issue:** 1. Configured SAML back channel SLO using the doc [1] 2. Once logout form one application other application doesn't get logged out. 3. Getting following errors in the console. ``` [2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo [2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error ``` [1] https://is.docs.wso2.com/en/5.11.0/learn/handling-saml-single-logout-requests-from-federated-identity-providers/ **How to reproduce:** 1. Configured SAML back channel SLO using the doc [1] 2. Once logout form one application other application doesn't get logged out. 3. Getting following errors in the console. ``` [2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo [2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error ``` **Expected behavior:** First application should automatically logged out **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version:IS-5.11.0-beta2 - OS: Linux - Database: POSTGRE - Userstore: JDBC ---
priority
saml back channel slo not working describe the issue configured saml back channel slo using the doc once logout form one application other application doesn t get logged out getting following errors in the console info org carbon identity sso saml logout logoutrequestsender sending single log out request again with retry count after waiting for milli seconds to warn org carbon identity sso saml logout logoutrequestsender failed single logout response from with status code internal server error how to reproduce configured saml back channel slo using the doc once logout form one application other application doesn t get logged out getting following errors in the console info org carbon identity sso saml logout logoutrequestsender sending single log out request again with retry count after waiting for milli seconds to warn org carbon identity sso saml logout logoutrequestsender failed single logout response from with status code internal server error expected behavior first application should automatically logged out environment information please complete the following information remove any unnecessary fields product version is os linux database postgre userstore jdbc
1
296,766
9,125,860,252
IssuesEvent
2019-02-24 17:06:39
pixijs/pixi.js
https://api.github.com/repos/pixijs/pixi.js
closed
TilingSprite ignoring realtime texture updates with Canvas renderer.
Difficulty: Easy Domain: API Plugin: Tiling Sprite Priority: High Renderer: Canvas Resolution: Won't Fix Status: Accepting PRs Type: Bug Version: v4.x
So I'm using [PIXI.extras.TilingSprite](https://pixijs.github.io/docs/PIXI.extras.TilingSprite.html) with the new 4.0.0-RC2. I updated to this version after seeing this bug with 3.0.11 while using PIXI.js within [nwjs](http://nwjs.io/), which is pretty much Chromium, in both OS X 10.11 and Ubuntu 16.04. I'm using TilingSprites to allow for sprite sheets, which works great if everything is on a single sprite sheet, but when changing sheets (walking, jumping, etc.) the following update seems to be ignored when using the Canvas renderer. `spriteObj.texture = textureObj;` The update works perfectly when using the WebGL renderer though. It seems to work when the object is first created and the initial texture is applied, but seems to stop working after the object is attached to a stage. I'll work on getting more source code together for a better example, but there isn't a lot to do here to reproduce it: just do an assignment as above and see if it's caught in the next render call. Try that with both the WebGL and Canvas renderers. We're working with the WebGL renderer for the time being, but I'm not sure that our target platform will support this renderer (currently a debian os running nwjs).
1.0
TilingSprite ignoring realtime texture updates with Canvas renderer. - So I'm using [PIXI.extras.TilingSprite](https://pixijs.github.io/docs/PIXI.extras.TilingSprite.html) with the new 4.0.0-RC2. I updated to this version after seeing this bug with 3.0.11 while using PIXI.js within [nwjs](http://nwjs.io/), which is pretty much Chromium, in both OS X 10.11 and Ubuntu 16.04. I'm using TilingSprites to allow for sprite sheets, which works great if everything is on a single sprite sheet, but when changing sheets (walking, jumping, etc.) the following update seems to be ignored when using the Canvas renderer. `spriteObj.texture = textureObj;` The update works perfectly when using the WebGL renderer though. It seems to work when the object is first created and the initial texture is applied, but seems to stop working after the object is attached to a stage. I'll work on getting more source code together for a better example, but there isn't a lot to do here to reproduce it: just do an assignment as above and see if it's caught in the next render call. Try that with both the WebGL and Canvas renderers. We're working with the WebGL renderer for the time being, but I'm not sure that our target platform will support this renderer (currently a debian os running nwjs).
priority
tilingsprite ignoring realtime texture updates with canvas renderer so i m using with the new i updated to this version after seeing this bug with while using pixi js within which is pretty much chromium in both os x and ubuntu i m using tilingsprites to allow for sprite sheets which works great if everything is on a single sprite sheet but when changing sheets walking jumping etc the following update seems to be ignored when using the canvas renderer spriteobj texture textureobj the update works perfectly when using the webgl renderer though it seems to work when the object is first created and the initial texture is applied but seems to stop working after the object is attached to a stage i ll work on getting more source code together for a better example but there isn t a lot to do here to reproduce it just do an assignment as above and see if it s caught in the next render call try that with both the webgl and canvas renderers we re working with the webgl renderer for the time being but i m not sure that our target platform will support this renderer currently a debian os running nwjs
1
330,164
10,035,664,491
IssuesEvent
2019-07-18 08:53:06
netdata/netdata
https://api.github.com/repos/netdata/netdata
closed
Provide an include configuration mechanism
area/daemon feature request priority/high
It would great if one could modify the configuration of netdata without touching the configuration shipped by the distribution. That way, when upgrades come along, those files are seen as not modified by the upgrade process and you get smoother upgrades, no diffs to read and new options as a bonus. I don't know if there is a formalisation of this, but the include.d mechanism is great for that. That way we can edit (or ship via salt/ansible/puppet/chef/etc.) atomic files for atomic bits of configuration that don't interfere with the default configuration. Netdata already uses that for python.d health.d charts.d etc. It would be awesome (or is already the case ?) to be able to have the same for the other configurations: * netdata.conf -> add netdata.d/*.conf * stream.conf -> add stream.d/*.conf * fping.conf -> add fping.d/*.conf * etc.
1.0
Provide an include configuration mechanism - It would great if one could modify the configuration of netdata without touching the configuration shipped by the distribution. That way, when upgrades come along, those files are seen as not modified by the upgrade process and you get smoother upgrades, no diffs to read and new options as a bonus. I don't know if there is a formalisation of this, but the include.d mechanism is great for that. That way we can edit (or ship via salt/ansible/puppet/chef/etc.) atomic files for atomic bits of configuration that don't interfere with the default configuration. Netdata already uses that for python.d health.d charts.d etc. It would be awesome (or is already the case ?) to be able to have the same for the other configurations: * netdata.conf -> add netdata.d/*.conf * stream.conf -> add stream.d/*.conf * fping.conf -> add fping.d/*.conf * etc.
priority
provide an include configuration mechanism it would great if one could modify the configuration of netdata without touching the configuration shipped by the distribution that way when upgrades come along those files are seen as not modified by the upgrade process and you get smoother upgrades no diffs to read and new options as a bonus i don t know if there is a formalisation of this but the include d mechanism is great for that that way we can edit or ship via salt ansible puppet chef etc atomic files for atomic bits of configuration that don t interfere with the default configuration netdata already uses that for python d health d charts d etc it would be awesome or is already the case to be able to have the same for the other configurations netdata conf add netdata d conf stream conf add stream d conf fping conf add fping d conf etc
1
64,862
3,219,042,002
IssuesEvent
2015-10-08 07:21:11
cs2103aug2015-t16-2j/main
https://api.github.com/repos/cs2103aug2015-t16-2j/main
closed
As a user, I want to edit/ update my entries.
priority.high type.story
So I do not have to delete and retype when I want to change details of my scheduled entry.
1.0
As a user, I want to edit/ update my entries. - So I do not have to delete and retype when I want to change details of my scheduled entry.
priority
as a user i want to edit update my entries so i do not have to delete and retype when i want to change details of my scheduled entry
1
186,248
6,734,648,638
IssuesEvent
2017-10-18 18:47:38
geosolutions-it/eumetsat-EOWS
https://api.github.com/repos/geosolutions-it/eumetsat-EOWS
opened
Improve NetCDF serving performance
geoserver Priority: High user story
A few things come to my mind here: - [ ] remove intermediate indexes - [ ] review and speed up code to reuse internal readers
1.0
Improve NetCDF serving performance - A few things come to my mind here: - [ ] remove intermediate indexes - [ ] review and speed up code to reuse internal readers
priority
improve netcdf serving performance a few things come to my mind here remove intermediate indexes review and speed up code to reuse internal readers
1
277,868
8,633,694,499
IssuesEvent
2018-11-22 14:36:30
BBC-News/psammead
https://api.github.com/repos/BBC-News/psammead
opened
GEL Typography, GEL Grid, GEL Spacing and Colour utilities
alpha-2 high priority refinement needed
**Is your feature request related to a problem? Please describe.** We currently don't have utility files in this repository. We should add them. **Describe the solution you'd like** We should have a top-level `utilities` directory. Currently in Simorgh, we have this structure lib ├── constants │   ├── styles.js │   └── typography.js ├── globalStyles.js ├── globalStyles.test.js └── layoutGrid.js helpers ├── mediaQueries.js Importing this structure here into Psammead is not ideal. We should create a clear structure and define what each file should contain. I propose having these files in the `utilities` directory: ``` gel-breakpoints.js gel-spacing.js gel-grid-styled-components.js gel-typography-styled-components.js articles-colours.js articles-fonts.js articles-svgs.js articles-amp-boilerplate.js ``` Here is the mapping from existing code to new component files: `gel-spacing.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L30-L50 `articles-colours.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L4-L17 `articles-fonts.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L19-L28 `articles-svgs.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L83-L89 `articles-amp-boilerplate.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L90-L95 `gel-breakpoints.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L51-L82 & https://github.com/BBC-News/simorgh/blob/latest/src/app/helpers/mediaQueries.js#L7-L13 `gel-grid-styled-components.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/layoutGrid.js#L16-L82 `gel-typography-styled-components.js`: https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/typography.js#L12-L73 ``` I would particularly like to draw attention to the last two files - since they export Styled Components, not just constants. **Describe alternatives you've considered** N/A **Testing notes** The files should have the correct content for the relevant names. - [x] Initially labelled with ["Refinement needed"](https://github.com/BBC-News/psammead/labels/Refinement%20Needed)
1.0
GEL Typography, GEL Grid, GEL Spacing and Colour utilities - **Is your feature request related to a problem? Please describe.** We currently don't have utility files in this repository. We should add them. **Describe the solution you'd like** We should have a top-level `utilities` directory. Currently in Simorgh, we have this structure lib ├── constants │   ├── styles.js │   └── typography.js ├── globalStyles.js ├── globalStyles.test.js └── layoutGrid.js helpers ├── mediaQueries.js Importing this structure here into Psammead is not ideal. We should create a clear structure and define what each file should contain. I propose having these files in the `utilities` directory: ``` gel-breakpoints.js gel-spacing.js gel-grid-styled-components.js gel-typography-styled-components.js articles-colours.js articles-fonts.js articles-svgs.js articles-amp-boilerplate.js ``` Here is the mapping from existing code to new component files: `gel-spacing.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L30-L50 `articles-colours.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L4-L17 `articles-fonts.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L19-L28 `articles-svgs.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L83-L89 `articles-amp-boilerplate.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L90-L95 `gel-breakpoints.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L51-L82 & https://github.com/BBC-News/simorgh/blob/latest/src/app/helpers/mediaQueries.js#L7-L13 `gel-grid-styled-components.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/layoutGrid.js#L16-L82 `gel-typography-styled-components.js`: https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/typography.js#L12-L73 ``` I would particularly like to draw attention to the last two files - since they export Styled Components, not just constants. **Describe alternatives you've considered** N/A **Testing notes** The files should have the correct content for the relevant names. - [x] Initially labelled with ["Refinement needed"](https://github.com/BBC-News/psammead/labels/Refinement%20Needed)
priority
gel typography gel grid gel spacing and colour utilities is your feature request related to a problem please describe we currently don t have utility files in this repository we should add them describe the solution you d like we should have a top level utilities directory currently in simorgh we have this structure lib ├── constants │   ├── styles js │   └── typography js ├── globalstyles js ├── globalstyles test js └── layoutgrid js helpers ├── mediaqueries js importing this structure here into psammead is not ideal we should create a clear structure and define what each file should contain i propose having these files in the utilities directory gel breakpoints js gel spacing js gel grid styled components js gel typography styled components js articles colours js articles fonts js articles svgs js articles amp boilerplate js here is the mapping from existing code to new component files gel spacing js articles colours js articles fonts js articles svgs js articles amp boilerplate js gel breakpoints js gel grid styled components js gel typography styled components js i would particularly like to draw attention to the last two files since they export styled components not just constants describe alternatives you ve considered n a testing notes the files should have the correct content for the relevant names initially labelled with
1
67,967
3,283,354,962
IssuesEvent
2015-10-28 12:14:29
leeensminger/OED_Wetlands
https://api.github.com/repos/leeensminger/OED_Wetlands
closed
Projects - Associated Mitigation Sites do not display in the datagrid
bug - high priority
The Mitigation Sites tab does not display any records. I confirmed within the database that project ID 8cf9ce93-1f20-4a24-b32b-ddb97f19c49f (MD 124 Lane extension) is associated to 2 mitigation sites: 20f37616-5d1d-4d20-a09a-f3f01b8a6b0d 33556824-9b44-46f5-8dc7-8d85a9ed9475 No records are displayed within the mitigation sites tab. ![image](https://cloud.githubusercontent.com/assets/4103847/10715163/cf9269dc-7ae1-11e5-82e8-4c321889f3ac.png)
1.0
Projects - Associated Mitigation Sites do not display in the datagrid - The Mitigation Sites tab does not display any records. I confirmed within the database that project ID 8cf9ce93-1f20-4a24-b32b-ddb97f19c49f (MD 124 Lane extension) is associated to 2 mitigation sites: 20f37616-5d1d-4d20-a09a-f3f01b8a6b0d 33556824-9b44-46f5-8dc7-8d85a9ed9475 No records are displayed within the mitigation sites tab. ![image](https://cloud.githubusercontent.com/assets/4103847/10715163/cf9269dc-7ae1-11e5-82e8-4c321889f3ac.png)
priority
projects associated mitigation sites do not display in the datagrid the mitigation sites tab does not display any records i confirmed within the database that project id md lane extension is associated to mitigation sites no records are displayed within the mitigation sites tab
1
336,004
10,169,769,939
IssuesEvent
2019-08-08 02:07:16
treehouses/cli
https://api.github.com/repos/treehouses/cli
closed
Treehouses tor external/internal ports are switched
high priority
`treehouses tor add 80 8080` sets external to 80, local to 8080, but displays the opposite
1.0
Treehouses tor external/internal ports are switched - `treehouses tor add 80 8080` sets external to 80, local to 8080, but displays the opposite
priority
treehouses tor external internal ports are switched treehouses tor add sets external to local to but displays the opposite
1
211,686
7,203,548,571
IssuesEvent
2018-02-06 09:35:34
vmware/vic-product
https://api.github.com/repos/vmware/vic-product
closed
Document 1.3.0 to 1.3.1 upgrade
component/ova kind/user-doc priority/high pub/vsphere
Following Slack chat with @andrewtchin, there are differences between upgrading from 1.2.x to 1.3.x and upgrading from 1.3.0 to 1.3.1. This is due to the migration from 2 disks to 4 disks during upgrade from 1.2.x, meaning that in 1.3.x to 1.3.y upgrades, different disks need to be added/removed from the appliances. @andrewtchin and @morris-jason can you please point me to any existing issues that provide details of the 1.3.x to 1.3.y procedure? Thanks!
1.0
Document 1.3.0 to 1.3.1 upgrade - Following Slack chat with @andrewtchin, there are differences between upgrading from 1.2.x to 1.3.x and upgrading from 1.3.0 to 1.3.1. This is due to the migration from 2 disks to 4 disks during upgrade from 1.2.x, meaning that in 1.3.x to 1.3.y upgrades, different disks need to be added/removed from the appliances. @andrewtchin and @morris-jason can you please point me to any existing issues that provide details of the 1.3.x to 1.3.y procedure? Thanks!
priority
document to upgrade following slack chat with andrewtchin there are differences between upgrading from x to x and upgrading from to this is due to the migration from disks to disks during upgrade from x meaning that in x to y upgrades different disks need to be added removed from the appliances andrewtchin and morris jason can you please point me to any existing issues that provide details of the x to y procedure thanks
1
176,312
6,558,208,966
IssuesEvent
2017-09-06 20:27:37
zehro/UAH-Theatre
https://api.github.com/repos/zehro/UAH-Theatre
closed
Set up the SQL Database
Feature Priority - High Team - Database
## Description We need a database to hold all the inventory items/data. ## Tasks - [x] Create an inventory database using a SQL database - [x] Connect to the database via Python Flask - [x] Set up basic CRUM operations to interact with the database ## Done Done Criteria 1. Test that the SQL database is live, interactable, and editable
1.0
Set up the SQL Database - ## Description We need a database to hold all the inventory items/data. ## Tasks - [x] Create an inventory database using a SQL database - [x] Connect to the database via Python Flask - [x] Set up basic CRUM operations to interact with the database ## Done Done Criteria 1. Test that the SQL database is live, interactable, and editable
priority
set up the sql database description we need a database to hold all the inventory items data tasks create an inventory database using a sql database connect to the database via python flask set up basic crum operations to interact with the database done done criteria test that the sql database is live interactable and editable
1
470,172
13,533,984,629
IssuesEvent
2020-09-16 04:31:46
SchizoDuckie/DuckieTV
https://api.github.com/repos/SchizoDuckie/DuckieTV
opened
Trakt TV updated shows API is now hard limited to 100 results meaning shows aren't updating
bug high priority
Not sure when this change happened but I noticed recently that the calendar hasn't been updating, this issue #1272 is also probably related. When DuckieTV checks for updated shows it calls `https://api-v2launch.trakt.tv/shows/updates/{date}?limit=10000`, since this was added it there was never a limit on the amount of results that could be returned but it is now capped at 100 which means there is basically 0 chance any of your shows will be getting updated. Changing the update period also isn't likely to help as the changes we request are only `yyyy-mm-dd` which means you'll only be getting changes for the whole day so even if you check every hour, you'll still only be getting all changes for that day and if there is already over 100 then you're not going to get any updates. Only way to fix this is the code is going to have to change to load every page of changes while the total results returned is less than 100, unless there is another way to get changes.
1.0
Trakt TV updated shows API is now hard limited to 100 results meaning shows aren't updating - Not sure when this change happened but I noticed recently that the calendar hasn't been updating, this issue #1272 is also probably related. When DuckieTV checks for updated shows it calls `https://api-v2launch.trakt.tv/shows/updates/{date}?limit=10000`, since this was added it there was never a limit on the amount of results that could be returned but it is now capped at 100 which means there is basically 0 chance any of your shows will be getting updated. Changing the update period also isn't likely to help as the changes we request are only `yyyy-mm-dd` which means you'll only be getting changes for the whole day so even if you check every hour, you'll still only be getting all changes for that day and if there is already over 100 then you're not going to get any updates. Only way to fix this is the code is going to have to change to load every page of changes while the total results returned is less than 100, unless there is another way to get changes.
priority
trakt tv updated shows api is now hard limited to results meaning shows aren t updating not sure when this change happened but i noticed recently that the calendar hasn t been updating this issue is also probably related when duckietv checks for updated shows it calls since this was added it there was never a limit on the amount of results that could be returned but it is now capped at which means there is basically chance any of your shows will be getting updated changing the update period also isn t likely to help as the changes we request are only yyyy mm dd which means you ll only be getting changes for the whole day so even if you check every hour you ll still only be getting all changes for that day and if there is already over then you re not going to get any updates only way to fix this is the code is going to have to change to load every page of changes while the total results returned is less than unless there is another way to get changes
1
407,751
11,936,946,385
IssuesEvent
2020-04-02 11:15:25
AY1920S2-CS2103T-F10-1/main
https://api.github.com/repos/AY1920S2-CS2103T-F10-1/main
closed
As a user, I want to generate a pdf version of my resume.
priority.High status.Ongoing
The generated resume needn't be very well formatted. There should be, however, clear division between each section.
1.0
As a user, I want to generate a pdf version of my resume. - The generated resume needn't be very well formatted. There should be, however, clear division between each section.
priority
as a user i want to generate a pdf version of my resume the generated resume needn t be very well formatted there should be however clear division between each section
1
402,178
11,805,960,489
IssuesEvent
2020-03-19 08:35:12
luna/ide
https://api.github.com/repos/luna/ide
closed
After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS
Category: IDE Priority: Highest Type: Bug
<!-- Please ensure that you are using the latest version of BaseGL before reporting the bug! It may have been fixed since. --> ### General Summary <!-- - Please include a high-level description of your bug here. --> After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS. It doesn't matter if the text is copied from the editor or from anywhere else. Pasting on other systems works ok. There is no new errors in dev console. ### Steps to Reproduce <!-- Please list the reproduction steps for your bug. --> 1. Open Enso Studio and File MAnager Mock on MacOS 2. Write any text 3. Copy part of it 4. Paste it ### Expected Result <!-- - A description of the results you expected from the reproduction steps. --> Changed text and fully working IDE ### Actual Result <!-- - A description of what actually happens when you perform these steps. - Please include any error output if relevant. --> Navigation with arrows is no longer working ### Luna Version <!-- - Please include the version of BaseGL you are using here. -->
1.0
After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS - <!-- Please ensure that you are using the latest version of BaseGL before reporting the bug! It may have been fixed since. --> ### General Summary <!-- - Please include a high-level description of your bug here. --> After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS. It doesn't matter if the text is copied from the editor or from anywhere else. Pasting on other systems works ok. There is no new errors in dev console. ### Steps to Reproduce <!-- Please list the reproduction steps for your bug. --> 1. Open Enso Studio and File MAnager Mock on MacOS 2. Write any text 3. Copy part of it 4. Paste it ### Expected Result <!-- - A description of the results you expected from the reproduction steps. --> Changed text and fully working IDE ### Actual Result <!-- - A description of what actually happens when you perform these steps. - Please include any error output if relevant. --> Navigation with arrows is no longer working ### Luna Version <!-- - Please include the version of BaseGL you are using here. -->
priority
after pasting text to text editor navigation with keyboard arrows stops working on macos please ensure that you are using the latest version of basegl before reporting the bug it may have been fixed since general summary please include a high level description of your bug here after pasting text to text editor navigation with keyboard arrows stops working on macos it doesn t matter if the text is copied from the editor or from anywhere else pasting on other systems works ok there is no new errors in dev console steps to reproduce please list the reproduction steps for your bug open enso studio and file manager mock on macos write any text copy part of it paste it expected result a description of the results you expected from the reproduction steps changed text and fully working ide actual result a description of what actually happens when you perform these steps please include any error output if relevant navigation with arrows is no longer working luna version please include the version of basegl you are using here
1
79,080
3,520,196,820
IssuesEvent
2016-01-12 19:49:20
Thaerious/NERVE
https://api.github.com/repos/Thaerious/NERVE
reopened
Although logged into CWRC, can't get list of CWRC entity links
bug - can not replicate bug - high priority
I'm having no problems with the Geonames lookup. When trying to link to names, sometimes I get the message saying that I haven't logged into the CWRC beta site, even though I have, and sometimes I get this message: java.lang.NullPointerException: TEC not set. TEC not set. - Throwable.java:116 - Exception.java:29 - RuntimeException.java:29 - NullPointerException.java:29 - GenericLinkForm.java:40 - GenericLinkForm.java:39 - PopupPanel.java:512 - DropFlowListener.java:256 - CWRCLookup.java:38 - Request.java:250 - RequestBuilder.java:412 - XMLHttpRequest.java:329 - Impl.java:247 - Impl.java:299 - Impl.java:72
1.0
Although logged into CWRC, can't get list of CWRC entity links - I'm having no problems with the Geonames lookup. When trying to link to names, sometimes I get the message saying that I haven't logged into the CWRC beta site, even though I have, and sometimes I get this message: java.lang.NullPointerException: TEC not set. TEC not set. - Throwable.java:116 - Exception.java:29 - RuntimeException.java:29 - NullPointerException.java:29 - GenericLinkForm.java:40 - GenericLinkForm.java:39 - PopupPanel.java:512 - DropFlowListener.java:256 - CWRCLookup.java:38 - Request.java:250 - RequestBuilder.java:412 - XMLHttpRequest.java:329 - Impl.java:247 - Impl.java:299 - Impl.java:72
priority
although logged into cwrc can t get list of cwrc entity links i m having no problems with the geonames lookup when trying to link to names sometimes i get the message saying that i haven t logged into the cwrc beta site even though i have and sometimes i get this message java lang nullpointerexception tec not set tec not set throwable java exception java runtimeexception java nullpointerexception java genericlinkform java genericlinkform java popuppanel java dropflowlistener java cwrclookup java request java requestbuilder java xmlhttprequest java impl java impl java impl java
1
28,564
2,707,929,279
IssuesEvent
2015-04-08 03:42:52
openfarmcc/OpenFarm
https://api.github.com/repos/openfarmcc/OpenFarm
closed
Subscribe to Newsletter on User creation flow does not work
bug High Priority User Accounts
Mailchimp receives no email address and Active Admin shows that the boolean has not been switched to true.
1.0
Subscribe to Newsletter on User creation flow does not work - Mailchimp receives no email address and Active Admin shows that the boolean has not been switched to true.
priority
subscribe to newsletter on user creation flow does not work mailchimp receives no email address and active admin shows that the boolean has not been switched to true
1
519,614
15,054,636,425
IssuesEvent
2021-02-03 17:41:28
mredwin1/DivisionManagementSystem
https://api.github.com/repos/mredwin1/DivisionManagementSystem
closed
When generating the attendance document only look at history up to the incident date
Area: Backend Priority: High Type: Bug
Currently the history includes their entire history but should only include thing up to the incident they are receiving encase something gets edited and there is a future point.
1.0
When generating the attendance document only look at history up to the incident date - Currently the history includes their entire history but should only include thing up to the incident they are receiving encase something gets edited and there is a future point.
priority
when generating the attendance document only look at history up to the incident date currently the history includes their entire history but should only include thing up to the incident they are receiving encase something gets edited and there is a future point
1
620,604
19,565,652,922
IssuesEvent
2022-01-03 23:37:01
bounswe/2021SpringGroup6
https://api.github.com/repos/bounswe/2021SpringGroup6
closed
Android - User's created events
Type: Task Status: In Progress Platform: Mobile Priority: High
The events created by the user will be displayed on home page. The user will be able to navigate to the Event Detail pages of his/her created events. On Event Detail Page, the user will be able to reject or accept the interested users of the event.
1.0
Android - User's created events - The events created by the user will be displayed on home page. The user will be able to navigate to the Event Detail pages of his/her created events. On Event Detail Page, the user will be able to reject or accept the interested users of the event.
priority
android user s created events the events created by the user will be displayed on home page the user will be able to navigate to the event detail pages of his her created events on event detail page the user will be able to reject or accept the interested users of the event
1
188,132
6,773,076,747
IssuesEvent
2017-10-27 03:10:52
xcat2/xcat-core
https://api.github.com/repos/xcat2/xcat-core
closed
[WSC] trying to put xcat common openbmc attributes into group def fails when createing node def
component:coral priority:high sprint2 type:usability
we have the following group definition: ```` [root@c699mgt00 create_cnodes]# lsdef -t group ws_compute -z # <xCAT data object stanza file> ws_compute: objtype=group cons=openbmc consoleondemand=1 grouptype=static mgt=openbmc netboot=petitboot nichostnamesuffixes.ib2=-ib2 nichostnamesuffixes.ib0=-ib0 nichostnamesuffixes.ib1=-ib1 nichostnamesuffixes.ib3=-ib3 nicnetworks.ib2=IB02 nicnetworks.ib0=IB00 nicnetworks.ib1=IB01 nicnetworks.ib3=IB03 nictypes.ib2=Infiniband nictypes.ib0=Infiniband nictypes.ib1=Infiniband nictypes.ib3=Infiniband primarynic=mac profile=compute usercomment=[offline] -- from ws_compute group def ```` We then try to do: ```` chdef -z <<EOF c699c001: objtype=node nodetype=mp bmc=10.3.1.101 groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01 hostnames=c699f01-U04 ip=10.3.1.1 nicips.ib0=10.10.1.1 nicips.ib1=10.11.1.1 nicips.ib2=10.12.1.1 nicips.ib3=10.13.1.1 rack=c699f01 switch=c699tor01 switchport=1 unit=u04 EOF ```` it then tells us: Cannot set the 'bmc' attribute unless a value is provided for 'mgt'. mgt => The method to use to do general hardware management of the node. This attribute is used as the default if power or getmac is not set. Valid values: openbmc, ipmi, blade, hmc, ivm, fsp, bpa, kvm, esx, rhevm. See the power attribute for more details. Error: Cannot set the attr='bmc' attribute unless 'mgt=openbmc'. 1 object definitions have been created or modified. New object definitions 'c699c001' have been created. The resultant definition is: ```` [root@c699mgt00 create_cnodes]# lsdef c699c001 Object name: c699c001 cons=openbmc consoleondemand=1 groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01 hostnames=c699f01-U04 ip=10.3.1.1 mgt=openbmc netboot=petitboot nichostnamesuffixes.ib2=-ib2 nichostnamesuffixes.ib0=-ib0 nichostnamesuffixes.ib1=-ib1 nichostnamesuffixes.ib3=-ib3 nicips.ib2=10.12.1.1 nicips.ib0=10.10.1.1 nicips.ib1=10.11.1.1 nicips.ib3=10.13.1.1 nicnetworks.ib2=IB02 nicnetworks.ib0=IB00 nicnetworks.ib1=IB01 nicnetworks.ib3=IB03 nictypes.ib2=Infiniband nictypes.ib0=Infiniband nictypes.ib1=Infiniband nictypes.ib3=Infiniband nodetype=mp postbootscripts=otherpkgs primarynic=mac profile=compute rack=c699f01 switch=c699tor01 switchport=1 unit=u04 ```` note, If i repeat the command a second time it works: ```` [root@c699mgt00 create_cnodes]# chdef -z <<EOF c699c001: objtype=node nodetype=mp bmc=10.3.1.101 groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01 hostnames=c699f01-U04 ip=10.3.1.1 nicips.ib0=10.10.1.1 nicips.ib1=10.11.1.1 nicips.ib2=10.12.1.1 nicips.ib3=10.13.1.1 rack=c699f01 switch=c699tor01 switchport=1 unit=u04 EOF 1 object definitions have been created or modified. ````
1.0
[WSC] trying to put xcat common openbmc attributes into group def fails when createing node def - we have the following group definition: ```` [root@c699mgt00 create_cnodes]# lsdef -t group ws_compute -z # <xCAT data object stanza file> ws_compute: objtype=group cons=openbmc consoleondemand=1 grouptype=static mgt=openbmc netboot=petitboot nichostnamesuffixes.ib2=-ib2 nichostnamesuffixes.ib0=-ib0 nichostnamesuffixes.ib1=-ib1 nichostnamesuffixes.ib3=-ib3 nicnetworks.ib2=IB02 nicnetworks.ib0=IB00 nicnetworks.ib1=IB01 nicnetworks.ib3=IB03 nictypes.ib2=Infiniband nictypes.ib0=Infiniband nictypes.ib1=Infiniband nictypes.ib3=Infiniband primarynic=mac profile=compute usercomment=[offline] -- from ws_compute group def ```` We then try to do: ```` chdef -z <<EOF c699c001: objtype=node nodetype=mp bmc=10.3.1.101 groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01 hostnames=c699f01-U04 ip=10.3.1.1 nicips.ib0=10.10.1.1 nicips.ib1=10.11.1.1 nicips.ib2=10.12.1.1 nicips.ib3=10.13.1.1 rack=c699f01 switch=c699tor01 switchport=1 unit=u04 EOF ```` it then tells us: Cannot set the 'bmc' attribute unless a value is provided for 'mgt'. mgt => The method to use to do general hardware management of the node. This attribute is used as the default if power or getmac is not set. Valid values: openbmc, ipmi, blade, hmc, ivm, fsp, bpa, kvm, esx, rhevm. See the power attribute for more details. Error: Cannot set the attr='bmc' attribute unless 'mgt=openbmc'. 1 object definitions have been created or modified. New object definitions 'c699c001' have been created. The resultant definition is: ```` [root@c699mgt00 create_cnodes]# lsdef c699c001 Object name: c699c001 cons=openbmc consoleondemand=1 groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01 hostnames=c699f01-U04 ip=10.3.1.1 mgt=openbmc netboot=petitboot nichostnamesuffixes.ib2=-ib2 nichostnamesuffixes.ib0=-ib0 nichostnamesuffixes.ib1=-ib1 nichostnamesuffixes.ib3=-ib3 nicips.ib2=10.12.1.1 nicips.ib0=10.10.1.1 nicips.ib1=10.11.1.1 nicips.ib3=10.13.1.1 nicnetworks.ib2=IB02 nicnetworks.ib0=IB00 nicnetworks.ib1=IB01 nicnetworks.ib3=IB03 nictypes.ib2=Infiniband nictypes.ib0=Infiniband nictypes.ib1=Infiniband nictypes.ib3=Infiniband nodetype=mp postbootscripts=otherpkgs primarynic=mac profile=compute rack=c699f01 switch=c699tor01 switchport=1 unit=u04 ```` note, If i repeat the command a second time it works: ```` [root@c699mgt00 create_cnodes]# chdef -z <<EOF c699c001: objtype=node nodetype=mp bmc=10.3.1.101 groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01 hostnames=c699f01-U04 ip=10.3.1.1 nicips.ib0=10.10.1.1 nicips.ib1=10.11.1.1 nicips.ib2=10.12.1.1 nicips.ib3=10.13.1.1 rack=c699f01 switch=c699tor01 switchport=1 unit=u04 EOF 1 object definitions have been created or modified. ````
priority
trying to put xcat common openbmc attributes into group def fails when createing node def we have the following group definition lsdef t group ws compute z ws compute objtype group cons openbmc consoleondemand grouptype static mgt openbmc netboot petitboot nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nicnetworks nicnetworks nicnetworks nicnetworks nictypes infiniband nictypes infiniband nictypes infiniband nictypes infiniband primarynic mac profile compute usercomment from ws compute group def we then try to do chdef z eof objtype node nodetype mp bmc groups all compute lsf compute witherspoon ws compute ws hostnames ip nicips nicips nicips nicips rack switch switchport unit eof it then tells us cannot set the bmc attribute unless a value is provided for mgt mgt the method to use to do general hardware management of the node this attribute is used as the default if power or getmac is not set valid values openbmc ipmi blade hmc ivm fsp bpa kvm esx rhevm see the power attribute for more details error cannot set the attr bmc attribute unless mgt openbmc object definitions have been created or modified new object definitions have been created the resultant definition is lsdef object name cons openbmc consoleondemand groups all compute lsf compute witherspoon ws compute ws hostnames ip mgt openbmc netboot petitboot nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nicips nicips nicips nicips nicnetworks nicnetworks nicnetworks nicnetworks nictypes infiniband nictypes infiniband nictypes infiniband nictypes infiniband nodetype mp postbootscripts otherpkgs primarynic mac profile compute rack switch switchport unit note if i repeat the command a second time it works chdef z eof objtype node nodetype mp bmc groups all compute lsf compute witherspoon ws compute ws hostnames ip nicips nicips nicips nicips rack switch switchport unit eof object definitions have been created or modified
1
705,114
24,221,961,680
IssuesEvent
2022-09-26 11:38:49
dnd-side-project/dnd-7th-7-frontend
https://api.github.com/repos/dnd-side-project/dnd-7th-7-frontend
closed
[feat] 경로 등록 API form data 가공 및 post 작업
🔧 Type: Feature 🔴 Priority: High
## 🔥 설명 (이슈에 대한 자세한 설명을 적어주세요!) - <br> ## 🔥 ETC (기타 사항을 적어주세요!) - <br> >#### types >- feat : 새로운 기능 >- update : 로직 변경 >- bug : 버그 수정 >- refactor : 파일 및 폴더 리팩토링 >- style : 코드 형식 변경 >- docs : 문서 관련 >- test : 테스트 코드 >- etc : 기타 변경 사항 >
1.0
[feat] 경로 등록 API form data 가공 및 post 작업 - ## 🔥 설명 (이슈에 대한 자세한 설명을 적어주세요!) - <br> ## 🔥 ETC (기타 사항을 적어주세요!) - <br> >#### types >- feat : 새로운 기능 >- update : 로직 변경 >- bug : 버그 수정 >- refactor : 파일 및 폴더 리팩토링 >- style : 코드 형식 변경 >- docs : 문서 관련 >- test : 테스트 코드 >- etc : 기타 변경 사항 >
priority
경로 등록 api form data 가공 및 post 작업 🔥 설명 이슈에 대한 자세한 설명을 적어주세요 🔥 etc 기타 사항을 적어주세요 types feat 새로운 기능 update 로직 변경 bug 버그 수정 refactor 파일 및 폴더 리팩토링 style 코드 형식 변경 docs 문서 관련 test 테스트 코드 etc 기타 변경 사항
1
648,104
21,165,233,292
IssuesEvent
2022-04-07 13:04:37
Aam-Digital/ndb-core
https://api.github.com/repos/Aam-Digital/ndb-core
opened
new notes added on child details tab are only displayed after reload
Type: Bug Status: High Priority
**Describe the bug** Creating a new note from the "Notes" tab of the Child Details view opens a popup form and works correctly, but the new note is not displayed in the table after saving. Only after opening that child's details again is the new note shown in the table.
1.0
new notes added on child details tab are only displayed after reload - **Describe the bug** Creating a new note from the "Notes" tab of the Child Details view opens a popup form and works correctly, but the new note is not displayed in the table after saving. Only after opening that child's details again is the new note shown in the table.
priority
new notes added on child details tab are only displayed after reload describe the bug creating a new note from the notes tab of the child details view opens a popup form and works correctly but the new note is not displayed in the table after saving only after opening that child s details again is the new note shown in the table
1
554,889
16,441,572,844
IssuesEvent
2021-05-20 14:53:41
hunchat/hunchat-api
https://api.github.com/repos/hunchat/hunchat-api
opened
Make `user.username` the default for `user.name`
priority:high
In the creation of a `User` object, if no `name` is provided, set `username` as `name`. Note: I might have to set `blank=True` in the model field definition.
1.0
Make `user.username` the default for `user.name` - In the creation of a `User` object, if no `name` is provided, set `username` as `name`. Note: I might have to set `blank=True` in the model field definition.
priority
make user username the default for user name in the creation of a user object if no name is provided set username as name note i might have to set blank true in the model field definition
1
324,639
9,906,793,878
IssuesEvent
2019-06-27 14:33:07
infor-design/enterprise
https://api.github.com/repos/infor-design/enterprise
reopened
Dropdown: 'startsWith' filter works by word rather than start of a String
[2] focus: datagrid priority: high type: bug :bug:
<!-- Please be aware that this is a publicly visible bug report. Do not post any credentials, screenshots with proprietary information, or anything you think shouldn't be visible to the world. If private information is required to be shared for a quality bug report, please email one of the [code owners](https://github.com/infor-design/enterprise/blob/master/.github/CODEOWNERS) directly. --> **Describe the bug** The 'startsWith' filter currently works on words of a String, rather than start of a String (see below for example). Currently, there is no way to filter on start of a String. Example list: ![list](https://user-images.githubusercontent.com/45232365/52408963-05c4f680-2acc-11e9-94f2-d090bd6d2cbe.png) If I enter 'C' I expect 'Correction' to be selected. This works ![filtered-list-1](https://user-images.githubusercontent.com/45232365/52408976-0cec0480-2acc-11e9-87b2-85e77c0525ed.png) However, if I enter 'A' I expect ''Allocated' to be selected but instead 'Not Allocated' is selected. It seems that the 'startsWith' works on word splits. ![filtered-list-2](https://user-images.githubusercontent.com/45232365/52409004-1f663e00-2acc-11e9-95a3-9489dc6aa071.png) **To Reproduce** Steps to reproduce the behavior: 1. Go to https://design.infor.com/code/ids-enterprise/latest/demo/components/dropdown/test-filter-types.html 2. Enter 'S' into 'StartsWith Filter' dropdown. 3. 'American Samoa' selected by default **Expected behavior** Expected 'South Carolina' to be selected by default **Version** Found in 4.14.0, exists in 4.15.0 **Platform** - Device: PC - OS Version: Windows 10 - Browser Name: Chrome - Browser Version: Version 72.0.3626.81 (Official Build) (64-bit) **Additional context** Discussion had with @tmcconechy and thought process was to change existing 'startsWith' behavior to filter based on start of String, and to add a new filter option such as 'startsWithWord' to filter based on the current behavior of 'startsWith'.
1.0
Dropdown: 'startsWith' filter works by word rather than start of a String - <!-- Please be aware that this is a publicly visible bug report. Do not post any credentials, screenshots with proprietary information, or anything you think shouldn't be visible to the world. If private information is required to be shared for a quality bug report, please email one of the [code owners](https://github.com/infor-design/enterprise/blob/master/.github/CODEOWNERS) directly. --> **Describe the bug** The 'startsWith' filter currently works on words of a String, rather than start of a String (see below for example). Currently, there is no way to filter on start of a String. Example list: ![list](https://user-images.githubusercontent.com/45232365/52408963-05c4f680-2acc-11e9-94f2-d090bd6d2cbe.png) If I enter 'C' I expect 'Correction' to be selected. This works ![filtered-list-1](https://user-images.githubusercontent.com/45232365/52408976-0cec0480-2acc-11e9-87b2-85e77c0525ed.png) However, if I enter 'A' I expect ''Allocated' to be selected but instead 'Not Allocated' is selected. It seems that the 'startsWith' works on word splits. ![filtered-list-2](https://user-images.githubusercontent.com/45232365/52409004-1f663e00-2acc-11e9-95a3-9489dc6aa071.png) **To Reproduce** Steps to reproduce the behavior: 1. Go to https://design.infor.com/code/ids-enterprise/latest/demo/components/dropdown/test-filter-types.html 2. Enter 'S' into 'StartsWith Filter' dropdown. 3. 'American Samoa' selected by default **Expected behavior** Expected 'South Carolina' to be selected by default **Version** Found in 4.14.0, exists in 4.15.0 **Platform** - Device: PC - OS Version: Windows 10 - Browser Name: Chrome - Browser Version: Version 72.0.3626.81 (Official Build) (64-bit) **Additional context** Discussion had with @tmcconechy and thought process was to change existing 'startsWith' behavior to filter based on start of String, and to add a new filter option such as 'startsWithWord' to filter based on the current behavior of 'startsWith'.
priority
dropdown startswith filter works by word rather than start of a string describe the bug the startswith filter currently works on words of a string rather than start of a string see below for example currently there is no way to filter on start of a string example list if i enter c i expect correction to be selected this works however if i enter a i expect allocated to be selected but instead not allocated is selected it seems that the startswith works on word splits to reproduce steps to reproduce the behavior go to enter s into startswith filter dropdown american samoa selected by default expected behavior expected south carolina to be selected by default version found in exists in platform device pc os version windows browser name chrome browser version version official build bit additional context discussion had with tmcconechy and thought process was to change existing startswith behavior to filter based on start of string and to add a new filter option such as startswithword to filter based on the current behavior of startswith
1
396,361
11,708,394,350
IssuesEvent
2020-03-08 13:03:38
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
opened
Graph failed to attach when adding comment
Priority: High
![image](https://user-images.githubusercontent.com/3536496/76163345-34564a80-6180-11ea-8e28-6775b90d83eb.png) This comment had both a map and a graph, but only one showed.
1.0
Graph failed to attach when adding comment - ![image](https://user-images.githubusercontent.com/3536496/76163345-34564a80-6180-11ea-8e28-6775b90d83eb.png) This comment had both a map and a graph, but only one showed.
priority
graph failed to attach when adding comment this comment had both a map and a graph but only one showed
1
669,333
22,620,275,596
IssuesEvent
2022-06-30 05:21:58
heading1/WYLSBingsu
https://api.github.com/repos/heading1/WYLSBingsu
closed
[FE] 알림 컴포넌트 생성
🖥 Frontend ❗️high-priority 🔨 Feature
## 🔨 기능 설명 알림 컴포넌트 생성 ## 📑 완료 조건 알림 컴포넌트 생성 ## 💭 관련 백로그 [대분류]-[중분류]-[소분류] 공통 컴포넌트 - 알림 컴포넌트 - 알림컴포넌트 생성 ## 💭 예상 작업 시간 1h
1.0
[FE] 알림 컴포넌트 생성 - ## 🔨 기능 설명 알림 컴포넌트 생성 ## 📑 완료 조건 알림 컴포넌트 생성 ## 💭 관련 백로그 [대분류]-[중분류]-[소분류] 공통 컴포넌트 - 알림 컴포넌트 - 알림컴포넌트 생성 ## 💭 예상 작업 시간 1h
priority
알림 컴포넌트 생성 🔨 기능 설명 알림 컴포넌트 생성 📑 완료 조건 알림 컴포넌트 생성 💭 관련 백로그 공통 컴포넌트 알림 컴포넌트 알림컴포넌트 생성 💭 예상 작업 시간
1
114,648
4,642,039,090
IssuesEvent
2016-09-30 08:08:47
rndsolutions/hawkcd
https://api.github.com/repos/rndsolutions/hawkcd
closed
Fetch Material task doesn't show all available materials
awaiting verification bug high priority ui
**Steps to reproduce** 1. Log in as Admin 2. Edit an existing Pipeline's tasks 3. Add a Fetch Material task **Expected behavior** All materials should be displayed in Material selection of the Fetch Material task. **Actual behavior** Only the assigned Material is shown.
1.0
Fetch Material task doesn't show all available materials - **Steps to reproduce** 1. Log in as Admin 2. Edit an existing Pipeline's tasks 3. Add a Fetch Material task **Expected behavior** All materials should be displayed in Material selection of the Fetch Material task. **Actual behavior** Only the assigned Material is shown.
priority
fetch material task doesn t show all available materials steps to reproduce log in as admin edit an existing pipeline s tasks add a fetch material task expected behavior all materials should be displayed in material selection of the fetch material task actual behavior only the assigned material is shown
1
760,322
26,636,771,265
IssuesEvent
2023-01-24 22:48:12
yuri2303/portifolio
https://api.github.com/repos/yuri2303/portifolio
closed
criar estrutura inicia do projeto
priority: high weight:2 type: feature
## Descrição Seleciona um template pronto e adicionar os arquivos para o projeto.
1.0
criar estrutura inicia do projeto - ## Descrição Seleciona um template pronto e adicionar os arquivos para o projeto.
priority
criar estrutura inicia do projeto descrição seleciona um template pronto e adicionar os arquivos para o projeto
1
660,814
22,031,983,042
IssuesEvent
2022-05-28 02:09:29
uwlib-cams/sinopia_maps
https://api.github.com/repos/uwlib-cams/sinopia_maps
closed
RECONFIGURE OUTPUT TO RTs because many things have changed
high priority xslt messy issue
ack 🤢 🤢 🤢 RDF RT structure has changed ![image](https://user-images.githubusercontent.com/27664966/170805458-b5b7c53d-50ab-4237-ae60-373fa8a40f70.png) 😢 😢 😢
1.0
RECONFIGURE OUTPUT TO RTs because many things have changed - ack 🤢 🤢 🤢 RDF RT structure has changed ![image](https://user-images.githubusercontent.com/27664966/170805458-b5b7c53d-50ab-4237-ae60-373fa8a40f70.png) 😢 😢 😢
priority
reconfigure output to rts because many things have changed ack 🤢 🤢 🤢 rdf rt structure has changed 😢 😢 😢
1
176,906
6,569,079,614
IssuesEvent
2017-09-09 02:01:44
ianroberts131/visual-algorithms
https://api.github.com/repos/ianroberts131/visual-algorithms
closed
Fix Speed Buttons for Merge Sort
Priority: High Type: Bug
# Bug ## What is the current behavior? When changing the speed for Merge Sort, the algorithm doesn't run properly ## What is the expected behavior? Should be able to change the speed and have the algorithm function properly
1.0
Fix Speed Buttons for Merge Sort - # Bug ## What is the current behavior? When changing the speed for Merge Sort, the algorithm doesn't run properly ## What is the expected behavior? Should be able to change the speed and have the algorithm function properly
priority
fix speed buttons for merge sort bug what is the current behavior when changing the speed for merge sort the algorithm doesn t run properly what is the expected behavior should be able to change the speed and have the algorithm function properly
1
237,981
7,768,810,133
IssuesEvent
2018-06-03 22:05:27
7547-G2/hoy-como-backend
https://api.github.com/repos/7547-G2/hoy-como-backend
closed
Crear get de extras para un plato en Mobile
Backend High-Priority Sprint 4
se envia id del plato en el path se espera recibir jsonArray: [{"extra_id": 99, "nombre_extra": "string", "precio": 999}, {...}, ...]
1.0
Crear get de extras para un plato en Mobile - se envia id del plato en el path se espera recibir jsonArray: [{"extra_id": 99, "nombre_extra": "string", "precio": 999}, {...}, ...]
priority
crear get de extras para un plato en mobile se envia id del plato en el path se espera recibir jsonarray
1
384,782
11,403,270,934
IssuesEvent
2020-01-31 06:38:28
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
API: unable to create a switch with a MAC address as ID
API Priority: High Type: Bug
**Describe the bug** When you want to create a switch with MAC address as ID, you can't. **To Reproduce** Steps to reproduce the behavior: ```console # pfperl-api get -M POST /api/v1/config/switches -c '{"id":"ff:ff:ff:ff:ff:ff","description":"test"}' | python -m json.tool { "errors": [ { "field": "id", "message": "Wrong value" } ], "message": "Unable to validate", "status": 422 } ``` **Expected behavior** Switch creation should occur. EDIT: For this switch: - `DELETE` is working - `PATCH` is not working - `GET` is working
1.0
API: unable to create a switch with a MAC address as ID - **Describe the bug** When you want to create a switch with MAC address as ID, you can't. **To Reproduce** Steps to reproduce the behavior: ```console # pfperl-api get -M POST /api/v1/config/switches -c '{"id":"ff:ff:ff:ff:ff:ff","description":"test"}' | python -m json.tool { "errors": [ { "field": "id", "message": "Wrong value" } ], "message": "Unable to validate", "status": 422 } ``` **Expected behavior** Switch creation should occur. EDIT: For this switch: - `DELETE` is working - `PATCH` is not working - `GET` is working
priority
api unable to create a switch with a mac address as id describe the bug when you want to create a switch with mac address as id you can t to reproduce steps to reproduce the behavior console pfperl api get m post api config switches c id ff ff ff ff ff ff description test python m json tool errors field id message wrong value message unable to validate status expected behavior switch creation should occur edit for this switch delete is working patch is not working get is working
1
801,997
28,564,889,169
IssuesEvent
2023-04-21 00:33:19
WordPress/wporg-two-factor
https://api.github.com/repos/WordPress/wporg-two-factor
closed
Temporarily updating the email address but not savings breaks password view.
bug ui priority: high
## GIF ![Screen Capture on 2023-04-20 at 11-29-59](https://user-images.githubusercontent.com/1657336/233242578-65754d5a-9e6c-4c2c-b05b-87f22b05ca12.gif) **Steps to reproduce** 1. Visit `users/{username}/edit/account/?screen=email` 2. Change the email address to an invalid email address 3. Click "Back" 4. Click "Password" list item 5. Notice the UI is broken.
1.0
Temporarily updating the email address but not savings breaks password view. - ## GIF ![Screen Capture on 2023-04-20 at 11-29-59](https://user-images.githubusercontent.com/1657336/233242578-65754d5a-9e6c-4c2c-b05b-87f22b05ca12.gif) **Steps to reproduce** 1. Visit `users/{username}/edit/account/?screen=email` 2. Change the email address to an invalid email address 3. Click "Back" 4. Click "Password" list item 5. Notice the UI is broken.
priority
temporarily updating the email address but not savings breaks password view gif steps to reproduce visit users username edit account screen email change the email address to an invalid email address click back click password list item notice the ui is broken
1
224,606
7,471,942,370
IssuesEvent
2018-04-03 10:56:26
ballerina-lang/composer
https://api.github.com/repos/ballerina-lang/composer
closed
Change the border color of the selected element
0.94-pre-release Imported Priority/Highest Severity/Major component/Composer
Can we change the color of the border of a code segment? If e have a large code block e.g: a while loop. It is difficult to recognize the boundary of the loop. Can we change the color when the element is selected?
1.0
Change the border color of the selected element - Can we change the color of the border of a code segment? If e have a large code block e.g: a while loop. It is difficult to recognize the boundary of the loop. Can we change the color when the element is selected?
priority
change the border color of the selected element can we change the color of the border of a code segment if e have a large code block e g a while loop it is difficult to recognize the boundary of the loop can we change the color when the element is selected
1
167,728
6,345,281,142
IssuesEvent
2017-07-27 21:52:32
ualbertalib/DMPonline_v4
https://api.github.com/repos/ualbertalib/DMPonline_v4
closed
Guidance is visible even it's set to be not published.
bug Priority:High
Reported by UQAM: 1. ![image](https://user-images.githubusercontent.com/2047614/28326337-ba2a664a-6b9d-11e7-89a4-b3b12c4b91e8.png) 2. Plans created in English and in French ![image](https://user-images.githubusercontent.com/2047614/28326351-c14e600c-6b9d-11e7-8b1b-1b0b12bdd9e0.png) 3. Display ![image](https://user-images.githubusercontent.com/2047614/28326360-cc0c8c12-6b9d-11e7-8fc6-ea524f57396c.png)
1.0
Guidance is visible even it's set to be not published. - Reported by UQAM: 1. ![image](https://user-images.githubusercontent.com/2047614/28326337-ba2a664a-6b9d-11e7-89a4-b3b12c4b91e8.png) 2. Plans created in English and in French ![image](https://user-images.githubusercontent.com/2047614/28326351-c14e600c-6b9d-11e7-8b1b-1b0b12bdd9e0.png) 3. Display ![image](https://user-images.githubusercontent.com/2047614/28326360-cc0c8c12-6b9d-11e7-8fc6-ea524f57396c.png)
priority
guidance is visible even it s set to be not published reported by uqam plans created in english and in french display
1
630,244
20,102,605,431
IssuesEvent
2022-02-07 07:00:23
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
reopened
Wrong operation in stft's center attribute
high priority triage review module: correctness (silent) module: fft
### 🐛 Describe the bug According to the pytorch's document, if I set `center` to `True` the first frame should be centered in the beginning of stft: https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L620 And the following code shows that the stft pads by `n_fft//2` https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L692 This works fine in most cases but if the `win_length` and `n_fft` are different padding `n_fft//2` does not gurantee the first frame be centered. I think `win_length//2` is the right choice for the proper operation. ### Versions Collecting environment information... PyTorch version: 1.10.1+cu113 Is debug build: False CUDA used to build PyTorch: 11.3 ROCM used to build PyTorch: N/A OS: Ubuntu 20.04.1 LTS (x86_64) GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Clang version: Could not collect CMake version: version 3.16.3 Libc version: glibc-2.31 Python version: 3.8.10 (default, Sep 28 2021, 16:10:42) [GCC 9.3.0] (64-bit runtime) Python platform: Linux-3.10.0-693.5.2.el7.x86_64-x86_64-with-glibc2.29 Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: Tesla P40 GPU 1: Tesla P40 GPU 2: Tesla P40 GPU 3: Tesla P40 GPU 4: Tesla P40 GPU 5: Tesla P40 GPU 6: Tesla P40 GPU 7: Tesla P40 Nvidia driver version: 455.32.00 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.0.5 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.21.4 [pip3] pytorch-lightning==1.4.9 [pip3] torch==1.10.1+cu113 [pip3] torchaudio==0.10.1+cu113 [pip3] torchmetrics==0.6.1 [pip3] torchvision==0.11.2+cu113 [conda] Could not collect cc @ezyang @gchanan @zou3519 @anjali411 @dylanbespalko @mruberry @Lezcano @nikitaved @peterbell10
1.0
Wrong operation in stft's center attribute - ### 🐛 Describe the bug According to the pytorch's document, if I set `center` to `True` the first frame should be centered in the beginning of stft: https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L620 And the following code shows that the stft pads by `n_fft//2` https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L692 This works fine in most cases but if the `win_length` and `n_fft` are different padding `n_fft//2` does not gurantee the first frame be centered. I think `win_length//2` is the right choice for the proper operation. ### Versions Collecting environment information... PyTorch version: 1.10.1+cu113 Is debug build: False CUDA used to build PyTorch: 11.3 ROCM used to build PyTorch: N/A OS: Ubuntu 20.04.1 LTS (x86_64) GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Clang version: Could not collect CMake version: version 3.16.3 Libc version: glibc-2.31 Python version: 3.8.10 (default, Sep 28 2021, 16:10:42) [GCC 9.3.0] (64-bit runtime) Python platform: Linux-3.10.0-693.5.2.el7.x86_64-x86_64-with-glibc2.29 Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: Tesla P40 GPU 1: Tesla P40 GPU 2: Tesla P40 GPU 3: Tesla P40 GPU 4: Tesla P40 GPU 5: Tesla P40 GPU 6: Tesla P40 GPU 7: Tesla P40 Nvidia driver version: 455.32.00 cuDNN version: Probably one of the following: /usr/lib/x86_64-linux-gnu/libcudnn.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.0.5 /usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.0.5 HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] numpy==1.21.4 [pip3] pytorch-lightning==1.4.9 [pip3] torch==1.10.1+cu113 [pip3] torchaudio==0.10.1+cu113 [pip3] torchmetrics==0.6.1 [pip3] torchvision==0.11.2+cu113 [conda] Could not collect cc @ezyang @gchanan @zou3519 @anjali411 @dylanbespalko @mruberry @Lezcano @nikitaved @peterbell10
priority
wrong operation in stft s center attribute 🐛 describe the bug according to the pytorch s document if i set center to true the first frame should be centered in the beginning of stft and the following code shows that the stft pads by n fft this works fine in most cases but if the win length and n fft are different padding n fft does not gurantee the first frame be centered i think win length is the right choice for the proper operation versions collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version libc version glibc python version default sep bit runtime python platform linux with is cuda available true cuda runtime version could not collect gpu models and configuration gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla nvidia driver version cudnn version probably one of the following usr lib linux gnu libcudnn so usr lib linux gnu libcudnn adv infer so usr lib linux gnu libcudnn adv train so usr lib linux gnu libcudnn cnn infer so usr lib linux gnu libcudnn cnn train so usr lib linux gnu libcudnn ops infer so usr lib linux gnu libcudnn ops train so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy pytorch lightning torch torchaudio torchmetrics torchvision could not collect cc ezyang gchanan dylanbespalko mruberry lezcano nikitaved
1