Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
512,705 | 14,907,845,605 | IssuesEvent | 2021-01-22 04:17:00 | Plaxy-Technologies-Inc/YouPlanets-Bug-Report | https://api.github.com/repos/Plaxy-Technologies-Inc/YouPlanets-Bug-Report | closed | Bug/Feature Missing: Can't connect to Stripe without Social Security Number - creators can't get paid without one? | Emergency Priority: High | 
| 1.0 | Bug/Feature Missing: Can't connect to Stripe without Social Security Number - creators can't get paid without one? - 
| priority | bug feature missing can t connect to stripe without social security number creators can t get paid without one | 1 |
631,183 | 20,146,999,115 | IssuesEvent | 2022-02-09 08:39:28 | Disfactory/Disfactory | https://api.github.com/repos/Disfactory/Disfactory | closed | 無法上傳照片 | bug high priority | **Describe the bug**
User couldn't upload the photo.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to disfactory.tw
2. Click on "我想新增可疑工廠"
3. Click on '新增照片', choose one
4. See error
**Screenshots**
<img width="1440" alt="截圖 2022-02-07 上午10 51 32" src="https://user-images.githubusercontent.com/60970217/152717064-b8c8a2af-ba4d-4b9d-9308-d11770edb287.png">
**Desktop (please complete the following information):**
- OS: MacOS
- Browser: chrome
- Version: 97.0.4692.99
**Additional context**
There are two users respond that they experienced the same problem through troubleshooting form respectively on 1/30 and 2/5. | 1.0 | 無法上傳照片 - **Describe the bug**
User couldn't upload the photo.
**To Reproduce**
Steps to reproduce the behavior:
1. Go to disfactory.tw
2. Click on "我想新增可疑工廠"
3. Click on '新增照片', choose one
4. See error
**Screenshots**
<img width="1440" alt="截圖 2022-02-07 上午10 51 32" src="https://user-images.githubusercontent.com/60970217/152717064-b8c8a2af-ba4d-4b9d-9308-d11770edb287.png">
**Desktop (please complete the following information):**
- OS: MacOS
- Browser: chrome
- Version: 97.0.4692.99
**Additional context**
There are two users respond that they experienced the same problem through troubleshooting form respectively on 1/30 and 2/5. | priority | 無法上傳照片 describe the bug user couldn t upload the photo to reproduce steps to reproduce the behavior go to disfactory tw click on 我想新增可疑工廠 click on 新增照片 choose one see error screenshots img width alt 截圖 src desktop please complete the following information os macos browser chrome version additional context there are two users respond that they experienced the same problem through troubleshooting form respectively on and | 1 |
335,813 | 10,167,022,266 | IssuesEvent | 2019-08-07 17:11:39 | worldmaking/msvr | https://api.github.com/repos/worldmaking/msvr | opened | running max causes module instancing error | Priority: High bug | @Zodsmar told me that he's able to reproduce an error where modules will not instantiate whenever max's audio rendering is turned on. | 1.0 | running max causes module instancing error - @Zodsmar told me that he's able to reproduce an error where modules will not instantiate whenever max's audio rendering is turned on. | priority | running max causes module instancing error zodsmar told me that he s able to reproduce an error where modules will not instantiate whenever max s audio rendering is turned on | 1 |
41,328 | 2,868,998,255 | IssuesEvent | 2015-06-05 22:28:20 | dart-lang/pub-dartlang | https://api.github.com/repos/dart-lang/pub-dartlang | closed | pub search not finding library that exists | bug notplanned Priority-High | <a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)**
_Originally opened as dart-lang/sdk#20755_
----
This has happened a few times now, there's some systemic going on.
This time, I can't find stagehand. It's first version has been up since 7/26. Yet, search can't find it. See screenshots.
There have been other bugs like this, and somehow it just starts working. But in this case, it's been over a month.
I think we should take a look at this.
______
**Attachments:**
[Screen Shot 2014-08-30 at 5.19.53 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.53 PM.png) (32.68 KB)
[Screen Shot 2014-08-30 at 5.19.48 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.48 PM.png) (40.56 KB) | 1.0 | pub search not finding library that exists - <a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)**
_Originally opened as dart-lang/sdk#20755_
----
This has happened a few times now, there's some systemic going on.
This time, I can't find stagehand. It's first version has been up since 7/26. Yet, search can't find it. See screenshots.
There have been other bugs like this, and somehow it just starts working. But in this case, it's been over a month.
I think we should take a look at this.
______
**Attachments:**
[Screen Shot 2014-08-30 at 5.19.53 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.53 PM.png) (32.68 KB)
[Screen Shot 2014-08-30 at 5.19.48 PM.png](https://storage.googleapis.com/google-code-attachments/dart/issue-20755/comment-0/Screen Shot 2014-08-30 at 5.19.48 PM.png) (40.56 KB) | priority | pub search not finding library that exists issue by originally opened as dart lang sdk this has happened a few times now there s some systemic going on this time i can t find stagehand it s first version has been up since yet search can t find it see screenshots there have been other bugs like this and somehow it just starts working but in this case it s been over a month i think we should take a look at this attachments shot at pm png kb shot at pm png kb | 1 |
114,685 | 4,642,618,547 | IssuesEvent | 2016-09-30 10:18:05 | armadito/armadito-av | https://api.github.com/repos/armadito/armadito-av | opened | improve/simplify REST API | enhancement high priority | update REST API for the following enhancements:
- remove '/api/register' call and integrate returning the token in the JSON objects returned by other requests such as '/api/scan' or '/api/status'
- add arguments to '/api/scan': generate progress events, progress event frequency (may be merge in one parameter: 0 for no progress event, > 0 for frequency in seconds)
| 1.0 | improve/simplify REST API - update REST API for the following enhancements:
- remove '/api/register' call and integrate returning the token in the JSON objects returned by other requests such as '/api/scan' or '/api/status'
- add arguments to '/api/scan': generate progress events, progress event frequency (may be merge in one parameter: 0 for no progress event, > 0 for frequency in seconds)
| priority | improve simplify rest api update rest api for the following enhancements remove api register call and integrate returning the token in the json objects returned by other requests such as api scan or api status add arguments to api scan generate progress events progress event frequency may be merge in one parameter for no progress event for frequency in seconds | 1 |
591,477 | 17,840,705,137 | IssuesEvent | 2021-09-03 09:41:53 | francheska-vicente/cssweng | https://api.github.com/repos/francheska-vicente/cssweng | opened | Rooms that offer monthly rate should not have extra persons | bug priority: high issue: back-end severity: medium issue: validation | ### Summary
- Rooms that offer monthly rates should not entertain extra pax.
### Steps to Reproduce
1. login
2. choose any date for booking
3. choose room 305
4. input 100 in the number of pax field
### Visual Proof

### Expected Results:
- Number of persons should be limited to at most 4 pax for twin bed (rm. 305).
### Actual Results:
- There is no limit to the number of extra pax for the monthly rate of twin bed (rm. 305).
| Additional Information | |
| ----------- | ----------- |
| Platform | V8 engine (Google) |
| Operating System | Windows 10 | | 1.0 | Rooms that offer monthly rate should not have extra persons - ### Summary
- Rooms that offer monthly rates should not entertain extra pax.
### Steps to Reproduce
1. login
2. choose any date for booking
3. choose room 305
4. input 100 in the number of pax field
### Visual Proof

### Expected Results:
- Number of persons should be limited to at most 4 pax for twin bed (rm. 305).
### Actual Results:
- There is no limit to the number of extra pax for the monthly rate of twin bed (rm. 305).
| Additional Information | |
| ----------- | ----------- |
| Platform | V8 engine (Google) |
| Operating System | Windows 10 | | priority | rooms that offer monthly rate should not have extra persons summary rooms that offer monthly rates should not entertain extra pax steps to reproduce login choose any date for booking choose room input in the number of pax field visual proof expected results number of persons should be limited to at most pax for twin bed rm actual results there is no limit to the number of extra pax for the monthly rate of twin bed rm additional information platform engine google operating system windows | 1 |
363,060 | 10,736,972,952 | IssuesEvent | 2019-10-29 12:07:04 | AY1920S1-CS2103T-W11-4/main | https://api.github.com/repos/AY1920S1-CS2103T-W11-4/main | closed | As a new user I want to see usage instructions | priority.High type.Story | So that I can refer to instructions when I forget how to use the App | 1.0 | As a new user I want to see usage instructions - So that I can refer to instructions when I forget how to use the App | priority | as a new user i want to see usage instructions so that i can refer to instructions when i forget how to use the app | 1 |
107,178 | 4,290,704,000 | IssuesEvent | 2016-07-18 10:54:05 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | opened | Disable current help tool Mobile | enhancement Mobile Priority: High | As a consequence of #845 we are going for the moment to remove the current help widget/tool from the mobile layout waiting for a more mobile friendly widget to be developed in future releases. | 1.0 | Disable current help tool Mobile - As a consequence of #845 we are going for the moment to remove the current help widget/tool from the mobile layout waiting for a more mobile friendly widget to be developed in future releases. | priority | disable current help tool mobile as a consequence of we are going for the moment to remove the current help widget tool from the mobile layout waiting for a more mobile friendly widget to be developed in future releases | 1 |
560,816 | 16,605,418,674 | IssuesEvent | 2021-06-02 02:43:04 | QuantEcon/quantecon-book-theme | https://api.github.com/repos/QuantEcon/quantecon-book-theme | closed | [google analytics] Check if Google Analytics is Setup | bug high-priority | @DrDrij @AakashGfude this is an issue to track the status for looking to see if google analytics is setup correctly for
- [ ] python-programming.quantecon.org
- [ ] python.quantecon.org | 1.0 | [google analytics] Check if Google Analytics is Setup - @DrDrij @AakashGfude this is an issue to track the status for looking to see if google analytics is setup correctly for
- [ ] python-programming.quantecon.org
- [ ] python.quantecon.org | priority | check if google analytics is setup drdrij aakashgfude this is an issue to track the status for looking to see if google analytics is setup correctly for python programming quantecon org python quantecon org | 1 |
323,563 | 9,856,515,180 | IssuesEvent | 2019-06-19 22:26:40 | NCIOCPL/cgov-digital-platform | https://api.github.com/repos/NCIOCPL/cgov-digital-platform | closed | Spanish CGOV - Missing Images and Infographics | High priority Migration bug | Images are missing from Spanish pages. These got migrated into Drupal, but they are not appearing on the front-end. Some examples pages include:
- http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/que-es
- http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/sobrellevar/supervivencia/infancia
- http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/radioterapia/haz-externo
- Infographic on http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades
---
AC - Adding Videos from #2020
http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/inmunoterapia
VERSUS https://www.cancer.gov/espanol/cancer/tratamiento/tipos/inmunoterapia
• http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades
VERSUS https://www.cancer.gov/espanol/cancer/naturaleza/desigualdades
• http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/seguridad-paciente
• http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/pago (edited)
| 1.0 | Spanish CGOV - Missing Images and Infographics - Images are missing from Spanish pages. These got migrated into Drupal, but they are not appearing on the front-end. Some examples pages include:
- http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/que-es
- http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/sobrellevar/supervivencia/infancia
- http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/radioterapia/haz-externo
- Infographic on http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades
---
AC - Adding Videos from #2020
http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/tipos/inmunoterapia
VERSUS https://www.cancer.gov/espanol/cancer/tratamiento/tipos/inmunoterapia
• http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/naturaleza/desigualdades
VERSUS https://www.cancer.gov/espanol/cancer/naturaleza/desigualdades
• http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/seguridad-paciente
• http://ncigovcdode176.prod.acquia-sites.com/espanol/cancer/tratamiento/estudios-clinicos/pago (edited)
| priority | spanish cgov missing images and infographics images are missing from spanish pages these got migrated into drupal but they are not appearing on the front end some examples pages include infographic on ac adding videos from versus • versus • • edited | 1 |
683,803 | 23,394,799,847 | IssuesEvent | 2022-08-11 21:47:03 | savvy-coders/sc-curriculum | https://api.github.com/repos/savvy-coders/sc-curriculum | closed | Investigate Parcel 2.7 potential 404 Error on npm run serve | bug high priority | When trying to build out SPA, ran npm run serve as test to see if it would build, it did build however the browser gave a 404 error. When uninstalling Parcel 2.7 and installed Parcel 2.5, the error seemed to "fix itself". Need to investigate if this was an isolated issue or if conflicts exist with new version of Parcel | 1.0 | Investigate Parcel 2.7 potential 404 Error on npm run serve - When trying to build out SPA, ran npm run serve as test to see if it would build, it did build however the browser gave a 404 error. When uninstalling Parcel 2.7 and installed Parcel 2.5, the error seemed to "fix itself". Need to investigate if this was an isolated issue or if conflicts exist with new version of Parcel | priority | investigate parcel potential error on npm run serve when trying to build out spa ran npm run serve as test to see if it would build it did build however the browser gave a error when uninstalling parcel and installed parcel the error seemed to fix itself need to investigate if this was an isolated issue or if conflicts exist with new version of parcel | 1 |
240,704 | 7,804,801,283 | IssuesEvent | 2018-06-11 08:45:38 | nimble-platform/common | https://api.github.com/repos/nimble-platform/common | closed | Each product within catalogue needs an assignment to a concept (domain ontology or eclass) | data schema very high priority | The current indexed data set (catalogue2) contains products which have no assignment to a concrete class. Current proposal is to add the field item_commodity_classification_uri in catalogue2.
For release 3: Each product within the catalogue2 needs a value for the field item_commodity_classification_uri | 1.0 | Each product within catalogue needs an assignment to a concept (domain ontology or eclass) - The current indexed data set (catalogue2) contains products which have no assignment to a concrete class. Current proposal is to add the field item_commodity_classification_uri in catalogue2.
For release 3: Each product within the catalogue2 needs a value for the field item_commodity_classification_uri | priority | each product within catalogue needs an assignment to a concept domain ontology or eclass the current indexed data set contains products which have no assignment to a concrete class current proposal is to add the field item commodity classification uri in for release each product within the needs a value for the field item commodity classification uri | 1 |
610,960 | 18,941,119,878 | IssuesEvent | 2021-11-18 03:03:50 | wso2/product-microgateway | https://api.github.com/repos/wso2/product-microgateway | closed | Error Appling API Key auth for petstore API | Type/Bug Priority/High | ### Description:
Error Appling API Key auth for petstore API
```log
adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event REMOVE_API_FROM_GATEWAY is received
adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event DEPLOY_API_IN_GATEWAY is received
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:163] - [synchronizer.FetchAPIsFromControlPlane] [-] API 2653577f-2fb4-4066-887c-0f887f85af76 is added/updated to APIList for label [Default]
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:64] - [synchronizer.FetchAPIs] [-] Fetching APIs from Control Plane.
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:171] - [synchronizer.FetchAPIsFromControlPlane] [-] Pushing data to router and enforcer
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:58] - [synchronizer.PushAPIProjects] [-] Start Deploying 1 API/s...
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:91] - [synchronizer.PushAPIProjects] [-] Start deploying api from file (API_ID:REVISION_ID).zip : 2653577f-2fb4-4066-887c-0f887f85af76-f9073d74-6da6-4d5a-a3cf-4017c930d430.zip
adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:230] - [api.ApplyAPIProjectFromAPIM] [-] Deploying api SwaggerPetstore:1.0.5 in Organization carbon.super
adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:240] - [api.ApplyAPIProjectFromAPIM] [-] API SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76" already deployed to vhost: localhost
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[petstore_auth:[write:pets read:pets]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[default:[write:pets read:pets]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 ERRO [apis_impl.go:222] - [api.ApplyAPIProjectFromAPIM.func1] [-] Recovered from panic. Error encountered while applying API SwaggerPetstore:1.0.5 to localhost.
adapter_1 | 2021-11-03 17:45:03 ERRO [apis_fetcher.go:105] - [synchronizer.PushAPIProjects] [-] Error occurred while applying project SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76"
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:113] - [synchronizer.PushAPIProjects] [-] Successfully deployed 1 API/s
```
### Steps to reproduce:
- Create API from https://petstore.swagger.io/v2/swagger.json
- Add API Key auth from runtime API Configs in APIM
- Deploy API
- check adapter logs
### Affected Product Version:
1.0.0-beta-snapshot
### Environment details (with versions):
- OS: mac
- Client: curl
- Env (Docker/K8s): docker compose
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| 1.0 | Error Appling API Key auth for petstore API - ### Description:
Error Appling API Key auth for petstore API
```log
adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event REMOVE_API_FROM_GATEWAY is received
adapter_1 | 2021-11-03 17:45:03 INFO [notification_listener.go:79] - [messaging.handleNotification] [-] Event DEPLOY_API_IN_GATEWAY is received
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:163] - [synchronizer.FetchAPIsFromControlPlane] [-] API 2653577f-2fb4-4066-887c-0f887f85af76 is added/updated to APIList for label [Default]
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:64] - [synchronizer.FetchAPIs] [-] Fetching APIs from Control Plane.
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:171] - [synchronizer.FetchAPIsFromControlPlane] [-] Pushing data to router and enforcer
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:58] - [synchronizer.PushAPIProjects] [-] Start Deploying 1 API/s...
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:91] - [synchronizer.PushAPIProjects] [-] Start deploying api from file (API_ID:REVISION_ID).zip : 2653577f-2fb4-4066-887c-0f887f85af76-f9073d74-6da6-4d5a-a3cf-4017c930d430.zip
adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:230] - [api.ApplyAPIProjectFromAPIM] [-] Deploying api SwaggerPetstore:1.0.5 in Organization carbon.super
adapter_1 | 2021-11-03 17:45:03 INFO [apis_impl.go:240] - [api.ApplyAPIProjectFromAPIM] [-] API SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76" already deployed to vhost: localhost
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[default:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[petstore_auth:[write:pets read:pets]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[default:[write:pets read:pets]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:213] - [model.checkAPIKeyInOperationArray] [-] Inside security scheme &[{api_key api_key api_key } {default oauth2 } {api_key apiKey api_key header} {petstore_auth oauth2 }].
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 0. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 1. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 INFO [open_api.go:217] - [model.checkAPIKeyInOperationArray] [-] New method key 2. Value: map[api_key:[]]
adapter_1 | 2021-11-03 17:45:03 ERRO [apis_impl.go:222] - [api.ApplyAPIProjectFromAPIM.func1] [-] Recovered from panic. Error encountered while applying API SwaggerPetstore:1.0.5 to localhost.
adapter_1 | 2021-11-03 17:45:03 ERRO [apis_fetcher.go:105] - [synchronizer.PushAPIProjects] [-] Error occurred while applying project SwaggerPetstore:1.0.5 with UUID "2653577f-2fb4-4066-887c-0f887f85af76"
adapter_1 | 2021-11-03 17:45:03 INFO [apis_fetcher.go:113] - [synchronizer.PushAPIProjects] [-] Successfully deployed 1 API/s
```
### Steps to reproduce:
- Create API from https://petstore.swagger.io/v2/swagger.json
- Add API Key auth from runtime API Configs in APIM
- Deploy API
- check adapter logs
### Affected Product Version:
1.0.0-beta-snapshot
### Environment details (with versions):
- OS: mac
- Client: curl
- Env (Docker/K8s): docker compose
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| priority | error appling api key auth for petstore api description error appling api key auth for petstore api log adapter info event remove api from gateway is received adapter info event deploy api in gateway is received adapter info api is added updated to apilist for label adapter info fetching apis from control plane adapter info pushing data to router and enforcer adapter info start deploying api s adapter info start deploying api from file api id revision id zip zip adapter info deploying api swaggerpetstore in organization carbon super adapter info api swaggerpetstore with uuid already deployed to vhost localhost adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info new method key value map adapter info inside security scheme adapter info new method key value map adapter info new method key value map adapter info new method key value map adapter erro recovered from panic error encountered while applying api swaggerpetstore to localhost adapter erro error occurred while applying project swaggerpetstore with uuid adapter info successfully deployed api s steps to reproduce create api from add api key auth from runtime api configs in apim deploy api check adapter logs affected product version beta snapshot environment details with versions os mac client curl env docker docker compose optional fields related issues suggested labels suggested assignees | 1 |
149,889 | 5,730,556,549 | IssuesEvent | 2017-04-21 09:44:13 | RestComm/mediaserver | https://api.github.com/repos/RestComm/mediaserver | closed | MGCP Channel must filter incoming packets | High-Priority MGCP2 netty | Add a network guard to the new MGCP Channel pipeline that filters incoming packets.
Only packets coming from local network are accepted, while packets coming from unknown external networks are discarded. | 1.0 | MGCP Channel must filter incoming packets - Add a network guard to the new MGCP Channel pipeline that filters incoming packets.
Only packets coming from local network are accepted, while packets coming from unknown external networks are discarded. | priority | mgcp channel must filter incoming packets add a network guard to the new mgcp channel pipeline that filters incoming packets only packets coming from local network are accepted while packets coming from unknown external networks are discarded | 1 |
317,619 | 9,667,001,322 | IssuesEvent | 2019-05-21 12:16:34 | sunpy/sunpy | https://api.github.com/repos/sunpy/sunpy | closed | Prepare for diff-rotations from different points of view. | Effort High Feature Request Package Intermediate Priority Medium coordinates | I've got this from a [draft document](https://issues.cosmos.esa.int/solarorbiterwiki/download/attachments/5801215/Triplet-TN%20SOL-SGS-TN-0020%20v0_2.pdf?version=1&modificationDate=1499950338000&api=v2) from [Solar Orbiter team](https://issues.cosmos.esa.int/solarorbiterwiki/display/SOSP/SOC+Documents):
> 4.3 SOC handling of the differential rotation
> SOC will use the following model of differential rotation to propagate from the triplet
> epoch:
> ω(Φ) = A + B sin2(Φ) + C sin4(Φ)
> Where ω is the rotation rate (in deg/day)
> And Φ is the solar latitude
>
> SOC will choose one of two sets of parameters:
> For magnetic features, meaning sunspots/active regions (derived from “Magnetic”
> in [DIFF])
>
> A = 14.252
> B = -1.678
> C = -2.401
> For non-magnetic features, e.g. coronal holes
> A = 14.705
> B = 0.0
> C = 0.0
> Note that this is a “rigid” rotation corresponding to 26.24 day synodic period from
> Earth (=> 24.48 day sidereal period).
None of the current values we've got matches what they are planing.
```python
>>> howard.to(u.deg / u.day)
<Quantity [ 14.32632838, -2.11875209, -1.83163148] deg / d>
>>> snodgrass.to(u.deg / u.day)
<Quantity [ 14.1134631 , -1.69797189, -2.34646844] deg / d>
>>> allen.to(u.deg / u.day)
<Quantity [ 14.44, -3. , 0. ] deg / d>
```
We do have a `synodic` correction as an option: `rotation -= 0.9856 * u.deg / u.day * duration` | 1.0 | Prepare for diff-rotations from different points of view. - I've got this from a [draft document](https://issues.cosmos.esa.int/solarorbiterwiki/download/attachments/5801215/Triplet-TN%20SOL-SGS-TN-0020%20v0_2.pdf?version=1&modificationDate=1499950338000&api=v2) from [Solar Orbiter team](https://issues.cosmos.esa.int/solarorbiterwiki/display/SOSP/SOC+Documents):
> 4.3 SOC handling of the differential rotation
> SOC will use the following model of differential rotation to propagate from the triplet
> epoch:
> ω(Φ) = A + B sin2(Φ) + C sin4(Φ)
> Where ω is the rotation rate (in deg/day)
> And Φ is the solar latitude
>
> SOC will choose one of two sets of parameters:
> For magnetic features, meaning sunspots/active regions (derived from “Magnetic”
> in [DIFF])
>
> A = 14.252
> B = -1.678
> C = -2.401
> For non-magnetic features, e.g. coronal holes
> A = 14.705
> B = 0.0
> C = 0.0
> Note that this is a “rigid” rotation corresponding to 26.24 day synodic period from
> Earth (=> 24.48 day sidereal period).
None of the current values we've got matches what they are planing.
```python
>>> howard.to(u.deg / u.day)
<Quantity [ 14.32632838, -2.11875209, -1.83163148] deg / d>
>>> snodgrass.to(u.deg / u.day)
<Quantity [ 14.1134631 , -1.69797189, -2.34646844] deg / d>
>>> allen.to(u.deg / u.day)
<Quantity [ 14.44, -3. , 0. ] deg / d>
```
We do have a `synodic` correction as an option: `rotation -= 0.9856 * u.deg / u.day * duration` | priority | prepare for diff rotations from different points of view i ve got this from a from soc handling of the differential rotation soc will use the following model of differential rotation to propagate from the triplet epoch ω φ a b φ c φ where ω is the rotation rate in deg day and φ is the solar latitude soc will choose one of two sets of parameters for magnetic features meaning sunspots active regions derived from “magnetic” in a b c for non magnetic features e g coronal holes a b c note that this is a “rigid” rotation corresponding to day synodic period from earth day sidereal period none of the current values we ve got matches what they are planing python howard to u deg u day snodgrass to u deg u day allen to u deg u day we do have a synodic correction as an option rotation u deg u day duration | 1 |
513,964 | 14,930,045,222 | IssuesEvent | 2021-01-25 01:43:21 | grannypron/uaf_levels | https://api.github.com/repos/grannypron/uaf_levels | closed | Error message when Quick is used in combat | Waiting For Approval bug game39 high priority | Get the following error when Q is pressed in combat for Quick:
Un-implemented code
Error Code = 00551B8C

| 1.0 | Error message when Quick is used in combat - Get the following error when Q is pressed in combat for Quick:
Un-implemented code
Error Code = 00551B8C

| priority | error message when quick is used in combat get the following error when q is pressed in combat for quick un implemented code error code | 1 |
447,153 | 12,884,600,776 | IssuesEvent | 2020-07-13 03:34:17 | mistifiedwarrior/Hackthon-Project | https://api.github.com/repos/mistifiedwarrior/Hackthon-Project | closed | Shopkeeper: see all bookings | easy high priority | As a shopkeeper
I want to see my all bookings
So that i can predict the today's selling
Acceptance criteria
- [x] criteria 1
Given in shopkeeper home page
When i click on the bookings
Then i can see all bookings
- [x] criteria 2
Given in booking details
When i want to look at the booking history
Then i can see all history | 1.0 | Shopkeeper: see all bookings - As a shopkeeper
I want to see my all bookings
So that i can predict the today's selling
Acceptance criteria
- [x] criteria 1
Given in shopkeeper home page
When i click on the bookings
Then i can see all bookings
- [x] criteria 2
Given in booking details
When i want to look at the booking history
Then i can see all history | priority | shopkeeper see all bookings as a shopkeeper i want to see my all bookings so that i can predict the today s selling acceptance criteria criteria given in shopkeeper home page when i click on the bookings then i can see all bookings criteria given in booking details when i want to look at the booking history then i can see all history | 1 |
338,817 | 10,237,788,662 | IssuesEvent | 2019-08-19 14:35:15 | IBM/carbon-addons-iot-react | https://api.github.com/repos/IBM/carbon-addons-iot-react | closed | [v2]: Upgrade Components to Carbon v10 | :computer: Development :fire: High priority v2 | Containing issue for the migration of our components to Carbon v10
High level steps/pieces:
- [x] upgrade carbon deps
- [x] upgrade storybook to v5 to support new prop-type definitions
- [x] Update/Fix Components
- [x] `ButtonEnhanced` refactored to `Button` wrapper
- [ ] ~`Table` renamed to `DataTable`, ensure sub components exported in the same fashion as Carbon to support custom composition of components. Avoid any breaking change from Carbon's documentation/usage for `DataTable`~
- [x] review visually for defects, fix as necessary
- [ ] ...
Maximo needs it and we need to support downstream teams | 1.0 | [v2]: Upgrade Components to Carbon v10 - Containing issue for the migration of our components to Carbon v10
High level steps/pieces:
- [x] upgrade carbon deps
- [x] upgrade storybook to v5 to support new prop-type definitions
- [x] Update/Fix Components
- [x] `ButtonEnhanced` refactored to `Button` wrapper
- [ ] ~`Table` renamed to `DataTable`, ensure sub components exported in the same fashion as Carbon to support custom composition of components. Avoid any breaking change from Carbon's documentation/usage for `DataTable`~
- [x] review visually for defects, fix as necessary
- [ ] ...
Maximo needs it and we need to support downstream teams | priority | upgrade components to carbon containing issue for the migration of our components to carbon high level steps pieces upgrade carbon deps upgrade storybook to to support new prop type definitions update fix components buttonenhanced refactored to button wrapper table renamed to datatable ensure sub components exported in the same fashion as carbon to support custom composition of components avoid any breaking change from carbon s documentation usage for datatable review visually for defects fix as necessary maximo needs it and we need to support downstream teams | 1 |
133,397 | 5,202,535,556 | IssuesEvent | 2017-01-24 09:52:25 | Promact/promact-oauth-server | https://api.github.com/repos/Promact/promact-oauth-server | closed | Restructure OAuth External Login Flow | Done high-priority OAuth Ready Support | Right now its not flowing the rule of OAuth 2.0 flow. So I need to restructure the OAuth flow and implement flow as OAuth 2.0 rules.
Following this link - https://www.digitalocean.com/community/tutorials/an-introduction-to-oauth-2
This is the current flow:
1. When user click the link, Login with Promact, then it will call the point 2
2. Authorization Code Link : The above link call a API and it will redirect to https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId
3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId will be done then user need to login.
4. OAuth Request to Application : After successfully completion of point 3, An request(server side call) will be send from OAuth to application with refresh token for getting app’s secret and redirect Uri then OAuth will receive the response and check for app secret.
5. Application Receives User Details and Access token : After successfully completion of point 4, Application will receive user details and add user to application and store the data.
My proposal for new structure:
1. When user click the link, Login with Promact, then it will call the point 2.
2. Authorization Code Link : The above link call a API and it will redirect to
https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId&redirectUri=redirectUri
Example:
https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId={PromactAppClientId}&redirectUri=”https://promactslack.azurewebsites.net/OAuth/Authorization?code={AuthorizationCode}”
PromactAppClientId = Promact app’s client Id
redirectUri = external login call Url
3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId and redirectUri will be done then user need to login.
4. Application Receives Authorization Code : After successfully completion of point 3, server will redirect to
Point 2’s redirectUri with authorization code.
Example : https://promactslack.azurewebsites.net/OAuth/Authorization?code=DFSD45FGD45FG11
5. Application Requests Access Token : After successfully completion of point 4, Server will call a API and application will request the OAuth for access token and user details
Http Request Url will be like this
https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId=PromactAppClientId&authorizationcode=AuthorizationCode&clientSecret=PromactAppClientSecret
Example :
https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId={PromactAppClientId}&authorizationcode={AuthorizationCode}&clientSecret={PromactAppClientSecret}
PromactAppClientId=Promact app’s clientId
AuthorizationCode = Point 4 code
PromactAppClientSecret – Promact app’s client secret
6. Application Receives Access Token : After successfully completion of point 5 and authorization code, client secret and details match then, Application will receive all required details of user and create and user for external login and store its access token of Promact.
| 1.0 | Restructure OAuth External Login Flow - Right now its not flowing the rule of OAuth 2.0 flow. So I need to restructure the OAuth flow and implement flow as OAuth 2.0 rules.
Following this link - https://www.digitalocean.com/community/tutorials/an-introduction-to-oauth-2
This is the current flow:
1. When user click the link, Login with Promact, then it will call the point 2
2. Authorization Code Link : The above link call a API and it will redirect to https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId
3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId will be done then user need to login.
4. OAuth Request to Application : After successfully completion of point 3, An request(server side call) will be send from OAuth to application with refresh token for getting app’s secret and redirect Uri then OAuth will receive the response and check for app secret.
5. Application Receives User Details and Access token : After successfully completion of point 4, Application will receive user details and add user to application and store the data.
My proposal for new structure:
1. When user click the link, Login with Promact, then it will call the point 2.
2. Authorization Code Link : The above link call a API and it will redirect to
https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId=PromactAppClientId&redirectUri=redirectUri
Example:
https://promactoauth.azurewebsites.net/OAuth/ExternalLogin?clientId={PromactAppClientId}&redirectUri=”https://promactslack.azurewebsites.net/OAuth/Authorization?code={AuthorizationCode}”
PromactAppClientId = Promact app’s client Id
redirectUri = external login call Url
3. User Authorizes Application : After successfully completion of point 2, server will redirect to OAuth server. In OAuth checking of app details with clientId and redirectUri will be done then user need to login.
4. Application Receives Authorization Code : After successfully completion of point 3, server will redirect to
Point 2’s redirectUri with authorization code.
Example : https://promactslack.azurewebsites.net/OAuth/Authorization?code=DFSD45FGD45FG11
5. Application Requests Access Token : After successfully completion of point 4, Server will call a API and application will request the OAuth for access token and user details
Http Request Url will be like this
https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId=PromactAppClientId&authorizationcode=AuthorizationCode&clientSecret=PromactAppClientSecret
Example :
https://promactoauth.azurewebsites.net/OAuth/AccessToken?clientId={PromactAppClientId}&authorizationcode={AuthorizationCode}&clientSecret={PromactAppClientSecret}
PromactAppClientId=Promact app’s clientId
AuthorizationCode = Point 4 code
PromactAppClientSecret – Promact app’s client secret
6. Application Receives Access Token : After successfully completion of point 5 and authorization code, client secret and details match then, Application will receive all required details of user and create and user for external login and store its access token of Promact.
| priority | restructure oauth external login flow right now its not flowing the rule of oauth flow so i need to restructure the oauth flow and implement flow as oauth rules following this link this is the current flow when user click the link login with promact then it will call the point authorization code link the above link call a api and it will redirect to user authorizes application after successfully completion of point server will redirect to oauth server in oauth checking of app details with clientid will be done then user need to login oauth request to application after successfully completion of point an request server side call will be send from oauth to application with refresh token for getting app’s secret and redirect uri then oauth will receive the response and check for app secret application receives user details and access token after successfully completion of point application will receive user details and add user to application and store the data my proposal for new structure when user click the link login with promact then it will call the point authorization code link the above link call a api and it will redirect to example promactappclientid promact app’s client id redirecturi external login call url user authorizes application after successfully completion of point server will redirect to oauth server in oauth checking of app details with clientid and redirecturi will be done then user need to login application receives authorization code after successfully completion of point server will redirect to point ’s redirecturi with authorization code example application requests access token after successfully completion of point server will call a api and application will request the oauth for access token and user details http request url will be like this example promactappclientid promact app’s clientid authorizationcode point code promactappclientsecret – promact app’s client secret application receives access token after successfully completion of point and authorization code client secret and details match then application will receive all required details of user and create and user for external login and store its access token of promact | 1 |
309,360 | 9,473,837,112 | IssuesEvent | 2019-04-19 04:13:48 | ga4gh/data-repository-service-schemas | https://api.github.com/repos/ga4gh/data-repository-service-schemas | closed | refine the list of AccessMethod type values | Due: Apr Priority: Critical Priority: High Project: DRS | Following up on the initial `type` list for `AccessMethod`, as enumerated in #236:
- how do we want to support htsget? My first thought is it should be a `type` = `htsget` of its own, where the `access_url` is required to support the htsget protocol, vs. just being a place to fetch bytes.
- do we want our initial `type` list to only include methods that we expect somebody to have implemented by v1 launch, or should we lean further into the hypothetical future? I'm inclined to trim it back to only things that will be live in 2019, once we know what that list is.
- can we combine `http` and `https` into one? I expect most servers will only use one of them, and as a client I don't think I would ever have interesting logic to pick one vs. the other -- if I intend to use HTTP, I'll pick that `type` and use whatever `access_url` I'm given. | 2.0 | refine the list of AccessMethod type values - Following up on the initial `type` list for `AccessMethod`, as enumerated in #236:
- how do we want to support htsget? My first thought is it should be a `type` = `htsget` of its own, where the `access_url` is required to support the htsget protocol, vs. just being a place to fetch bytes.
- do we want our initial `type` list to only include methods that we expect somebody to have implemented by v1 launch, or should we lean further into the hypothetical future? I'm inclined to trim it back to only things that will be live in 2019, once we know what that list is.
- can we combine `http` and `https` into one? I expect most servers will only use one of them, and as a client I don't think I would ever have interesting logic to pick one vs. the other -- if I intend to use HTTP, I'll pick that `type` and use whatever `access_url` I'm given. | priority | refine the list of accessmethod type values following up on the initial type list for accessmethod as enumerated in how do we want to support htsget my first thought is it should be a type htsget of its own where the access url is required to support the htsget protocol vs just being a place to fetch bytes do we want our initial type list to only include methods that we expect somebody to have implemented by launch or should we lean further into the hypothetical future i m inclined to trim it back to only things that will be live in once we know what that list is can we combine http and https into one i expect most servers will only use one of them and as a client i don t think i would ever have interesting logic to pick one vs the other if i intend to use http i ll pick that type and use whatever access url i m given | 1 |
323,103 | 9,842,860,435 | IssuesEvent | 2019-06-18 10:12:46 | stageosu/Kaguya | https://api.github.com/repos/stageosu/Kaguya | closed | Music player frequently skips songs on its own. | Bug Command Related Help wanted High Priority Music | More often than not, Kaguya will seemingly automatically skip songs that are queued (when it is their turn to play). In the log, it shows that there are no more items in the queue, so that is the reason why it has stopped playing (when it's clearly not the case). | 1.0 | Music player frequently skips songs on its own. - More often than not, Kaguya will seemingly automatically skip songs that are queued (when it is their turn to play). In the log, it shows that there are no more items in the queue, so that is the reason why it has stopped playing (when it's clearly not the case). | priority | music player frequently skips songs on its own more often than not kaguya will seemingly automatically skip songs that are queued when it is their turn to play in the log it shows that there are no more items in the queue so that is the reason why it has stopped playing when it s clearly not the case | 1 |
790,611 | 27,830,476,095 | IssuesEvent | 2023-03-20 04:05:08 | AY2223S2-CS2103-F10-2/tp | https://api.github.com/repos/AY2223S2-CS2103-F10-2/tp | opened | As a user, I can navigate to a specific module/lecture and search for contents within it's scope | type.Story priority.High | Integrate nav and find command to work together | 1.0 | As a user, I can navigate to a specific module/lecture and search for contents within it's scope - Integrate nav and find command to work together | priority | as a user i can navigate to a specific module lecture and search for contents within it s scope integrate nav and find command to work together | 1 |
263,578 | 8,292,960,972 | IssuesEvent | 2018-09-20 03:54:52 | aspnet/websdk | https://api.github.com/repos/aspnet/websdk | reopened | Missing features when creating a WebDeploy package | High Priority | This is the next step of https://github.com/dotnet/cli/issues/6598.
I am now able to create a WebDeploy package using MSBuild v15.
However, I am not able to accomplish the same things than before (with ASP.NET "classic" and MSBuild v14):
- [ ] How can I add parameters to the `WebApplication.Full.Parameters.xml` file (produced by the build) ?
- [x] There is no `.cmd` file generated.
- [x] In v14, the parameter `PackageLocation` defined the folder where to generate the package. Now, I have to utilize `DesktopBuildPackageLocation` and specify a file path. The former was very interresting for solutions that have multiple web applications (each web app generates its own package). With this new parameter, the zip file contains both the sites, but configured for only one. | 1.0 | Missing features when creating a WebDeploy package - This is the next step of https://github.com/dotnet/cli/issues/6598.
I am now able to create a WebDeploy package using MSBuild v15.
However, I am not able to accomplish the same things than before (with ASP.NET "classic" and MSBuild v14):
- [ ] How can I add parameters to the `WebApplication.Full.Parameters.xml` file (produced by the build) ?
- [x] There is no `.cmd` file generated.
- [x] In v14, the parameter `PackageLocation` defined the folder where to generate the package. Now, I have to utilize `DesktopBuildPackageLocation` and specify a file path. The former was very interresting for solutions that have multiple web applications (each web app generates its own package). With this new parameter, the zip file contains both the sites, but configured for only one. | priority | missing features when creating a webdeploy package this is the next step of i am now able to create a webdeploy package using msbuild however i am not able to accomplish the same things than before with asp net classic and msbuild how can i add parameters to the webapplication full parameters xml file produced by the build there is no cmd file generated in the parameter packagelocation defined the folder where to generate the package now i have to utilize desktopbuildpackagelocation and specify a file path the former was very interresting for solutions that have multiple web applications each web app generates its own package with this new parameter the zip file contains both the sites but configured for only one | 1 |
310,266 | 9,487,814,890 | IssuesEvent | 2019-04-22 17:55:57 | CosminNechifor/IKHNAIE | https://api.github.com/repos/CosminNechifor/IKHNAIE | closed | Implement a market contract | High Priority | The market contract should take care of keeping track of the ``Components`` that are being sent for sale.
Functionalities: **TBD**
This is block by the development fo the ``ComponentContract`` #11 #12. The functions ``submitedForSale`` and ``buyComponent`` should be accessible only from the ``Manager`` contract and it should take care if the user who performs the action has enough ``ether`` or is the ``owner`` of the component.
| 1.0 | Implement a market contract - The market contract should take care of keeping track of the ``Components`` that are being sent for sale.
Functionalities: **TBD**
This is block by the development fo the ``ComponentContract`` #11 #12. The functions ``submitedForSale`` and ``buyComponent`` should be accessible only from the ``Manager`` contract and it should take care if the user who performs the action has enough ``ether`` or is the ``owner`` of the component.
| priority | implement a market contract the market contract should take care of keeping track of the components that are being sent for sale functionalities tbd this is block by the development fo the componentcontract the functions submitedforsale and buycomponent should be accessible only from the manager contract and it should take care if the user who performs the action has enough ether or is the owner of the component | 1 |
359,276 | 10,667,527,206 | IssuesEvent | 2019-10-19 13:08:26 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Need separate Save button when assigning subscription tiers for an API for better user experience. | 3.0.0-beta Priority/Highest Type/Improvement | Create an API and publish it. By default, we assign the Bronze tier. If someone needs it to have unlimited tier only, they need to first tick "Unlimited" tier and then untick the "Bronze" tier. But usually, people will untick the "Bronze tier" first and then tick the tier they want. In such a case, when the Bronze tier is unticked, an error is thrown, saying API should have a tier assigned if it is not in created or if it is not a prototype API. Hence for better UX, a separate Save button would be ideal to save after ticking/unticking the subscription policies. | 1.0 | Need separate Save button when assigning subscription tiers for an API for better user experience. - Create an API and publish it. By default, we assign the Bronze tier. If someone needs it to have unlimited tier only, they need to first tick "Unlimited" tier and then untick the "Bronze" tier. But usually, people will untick the "Bronze tier" first and then tick the tier they want. In such a case, when the Bronze tier is unticked, an error is thrown, saying API should have a tier assigned if it is not in created or if it is not a prototype API. Hence for better UX, a separate Save button would be ideal to save after ticking/unticking the subscription policies. | priority | need separate save button when assigning subscription tiers for an api for better user experience create an api and publish it by default we assign the bronze tier if someone needs it to have unlimited tier only they need to first tick unlimited tier and then untick the bronze tier but usually people will untick the bronze tier first and then tick the tier they want in such a case when the bronze tier is unticked an error is thrown saying api should have a tier assigned if it is not in created or if it is not a prototype api hence for better ux a separate save button would be ideal to save after ticking unticking the subscription policies | 1 |
164,983 | 6,259,769,831 | IssuesEvent | 2017-07-14 18:52:23 | wpninjas/ninja-forms | https://api.github.com/repos/wpninjas/ninja-forms | closed | reCaptcha - Submitting form without validating locks the user from submitting. | DIFFICULTY: Involved PRIORITY: High VALUE: Friendly VALUE: Painless | As the title suggests, submitting without validating reCaptcha locks the form with an error, rendering the user unable to submit the form.
The only solution in this case is a complete reload of the page, and resubmission whilst appropriately validating reCaptcha. | 1.0 | reCaptcha - Submitting form without validating locks the user from submitting. - As the title suggests, submitting without validating reCaptcha locks the form with an error, rendering the user unable to submit the form.
The only solution in this case is a complete reload of the page, and resubmission whilst appropriately validating reCaptcha. | priority | recaptcha submitting form without validating locks the user from submitting as the title suggests submitting without validating recaptcha locks the form with an error rendering the user unable to submit the form the only solution in this case is a complete reload of the page and resubmission whilst appropriately validating recaptcha | 1 |
554,839 | 16,440,319,445 | IssuesEvent | 2021-05-20 13:43:55 | anguaive/typelonger | https://api.github.com/repos/anguaive/typelonger | opened | New specification and basic feature set | cleanup high-priority | ## To do:
- [ ] Examine old specification, rethink what to include and what to not include
- [ ] Decide on the basic feature set | 1.0 | New specification and basic feature set - ## To do:
- [ ] Examine old specification, rethink what to include and what to not include
- [ ] Decide on the basic feature set | priority | new specification and basic feature set to do examine old specification rethink what to include and what to not include decide on the basic feature set | 1 |
586,436 | 17,577,481,689 | IssuesEvent | 2021-08-15 22:09:41 | ErnestoFGonzalez/django-amazon-sns-mobile-push-notification | https://api.github.com/repos/ErnestoFGonzalez/django-amazon-sns-mobile-push-notification | closed | Add `badge` argument to `client.Client.publish_to_ios` method | enhancement priority:high difficulty:low | From [Local and Remote Notification Programming Guide: Creating the Remote Notification Payload](https://developer.apple.com/library/archive/documentation/NetworkingInternet/Conceptual/RemoteNotificationsPG/CreatingtheNotificationPayload.html#//apple_ref/doc/uid/TP40008194-CH10-SW1), we see that the `"aps"` entry includes a `"badge"` entry itself, which accepts an integer to set the app's devices badge number.
We'll had an optional `badge` argument to `client.Client.publish_to_ios` in order to be able to pass a payload like
```
{
"aps" : {
"alert" : {
"title" : string,
"body" : string,
"sound": string
},
"badge" : number
},
"id" : string,
"type": string,
"serializer": object
}
```
In order to pass `badge` down to `publish_to_ios` we nee to add the `badge` param to the following methods:
- [ ] `tasks.send_sns_mobile_push_notification_to_device`
- [ ] `models.Device.send`
- [ ] and finally `client.Client.publish_to_ios`.
| 1.0 | Add `badge` argument to `client.Client.publish_to_ios` method - From [Local and Remote Notification Programming Guide: Creating the Remote Notification Payload](https://developer.apple.com/library/archive/documentation/NetworkingInternet/Conceptual/RemoteNotificationsPG/CreatingtheNotificationPayload.html#//apple_ref/doc/uid/TP40008194-CH10-SW1), we see that the `"aps"` entry includes a `"badge"` entry itself, which accepts an integer to set the app's devices badge number.
We'll had an optional `badge` argument to `client.Client.publish_to_ios` in order to be able to pass a payload like
```
{
"aps" : {
"alert" : {
"title" : string,
"body" : string,
"sound": string
},
"badge" : number
},
"id" : string,
"type": string,
"serializer": object
}
```
In order to pass `badge` down to `publish_to_ios` we nee to add the `badge` param to the following methods:
- [ ] `tasks.send_sns_mobile_push_notification_to_device`
- [ ] `models.Device.send`
- [ ] and finally `client.Client.publish_to_ios`.
| priority | add badge argument to client client publish to ios method from we see that the aps entry includes a badge entry itself which accepts an integer to set the app s devices badge number we ll had an optional badge argument to client client publish to ios in order to be able to pass a payload like aps alert title string body string sound string badge number id string type string serializer object in order to pass badge down to publish to ios we nee to add the badge param to the following methods tasks send sns mobile push notification to device models device send and finally client client publish to ios | 1 |
681,168 | 23,299,414,375 | IssuesEvent | 2022-08-07 04:37:26 | AlexanderDefuria/FRC-Scouting | https://api.github.com/repos/AlexanderDefuria/FRC-Scouting | opened | Optimize Match Data Loading | Bug High Priority Back End | Currently matchdata is taking a long long time to load. Investigate. | 1.0 | Optimize Match Data Loading - Currently matchdata is taking a long long time to load. Investigate. | priority | optimize match data loading currently matchdata is taking a long long time to load investigate | 1 |
266,364 | 8,366,359,087 | IssuesEvent | 2018-10-04 08:54:24 | CS2113-AY1819S1-F10-3/Main | https://api.github.com/repos/CS2113-AY1819S1-F10-3/Main | closed | Morphing and implementing project specific classes | priority.high status.ongoing | Subject,PO and HQP classes (HQP is a sub-class of PO) | 1.0 | Morphing and implementing project specific classes - Subject,PO and HQP classes (HQP is a sub-class of PO) | priority | morphing and implementing project specific classes subject po and hqp classes hqp is a sub class of po | 1 |
550,229 | 16,107,446,812 | IssuesEvent | 2021-04-27 16:31:45 | pokt-network/pocket-core | https://api.github.com/repos/pokt-network/pocket-core | closed | Shift Tx Indexer to pocket core | enhancement high priority | **Is your feature request related to a problem? Please describe.**
Currently by using tendermint's `txIndexer` can lead to miss indexing of a tx as evidenced in #1188. This happens because the event does not necessarily match the signer.
**Describe the solution you'd like**
A custom `txIndexer` that allows us to specify how we wanna index and store tx's
**Describe alternatives you've considered**
Modifying the tx Indexer on tendermint was considered however it would've led to breaking the ABCI standard in order to be able to pass down signer for comparassion of event & signer
| 1.0 | Shift Tx Indexer to pocket core - **Is your feature request related to a problem? Please describe.**
Currently by using tendermint's `txIndexer` can lead to miss indexing of a tx as evidenced in #1188. This happens because the event does not necessarily match the signer.
**Describe the solution you'd like**
A custom `txIndexer` that allows us to specify how we wanna index and store tx's
**Describe alternatives you've considered**
Modifying the tx Indexer on tendermint was considered however it would've led to breaking the ABCI standard in order to be able to pass down signer for comparassion of event & signer
| priority | shift tx indexer to pocket core is your feature request related to a problem please describe currently by using tendermint s txindexer can lead to miss indexing of a tx as evidenced in this happens because the event does not necessarily match the signer describe the solution you d like a custom txindexer that allows us to specify how we wanna index and store tx s describe alternatives you ve considered modifying the tx indexer on tendermint was considered however it would ve led to breaking the abci standard in order to be able to pass down signer for comparassion of event signer | 1 |
443,261 | 12,769,462,407 | IssuesEvent | 2020-06-30 03:41:26 | ArkEcosystem/core | https://api.github.com/repos/ArkEcosystem/core | closed | Remove container getters | Priority: High | When the migration from awilix to inversify started I added 2 helpers in the form of `app.log` and `app.events`. These should be removed and replaced with injection where the logger or event dispatcher is needed. This will ensure that developers use injection and don't use internal helpers.
https://github.com/ArkEcosystem/core/blob/develop/packages/core-kernel/src/application.ts#L340-L360 | 1.0 | Remove container getters - When the migration from awilix to inversify started I added 2 helpers in the form of `app.log` and `app.events`. These should be removed and replaced with injection where the logger or event dispatcher is needed. This will ensure that developers use injection and don't use internal helpers.
https://github.com/ArkEcosystem/core/blob/develop/packages/core-kernel/src/application.ts#L340-L360 | priority | remove container getters when the migration from awilix to inversify started i added helpers in the form of app log and app events these should be removed and replaced with injection where the logger or event dispatcher is needed this will ensure that developers use injection and don t use internal helpers | 1 |
248,258 | 7,928,597,720 | IssuesEvent | 2018-07-06 12:20:36 | jncc/topcat | https://api.github.com/repos/jncc/topcat | closed | Copy some updated data files into http://data.jncc.gov.uk/data/ | high priority | I will be redacting some data from a large number of data files that are currently stored in http://data.jncc.gov.uk/data/ for access from Data.gov.uk.
Once this work is complete I will need developer assistance to replace the existing files with the updated files. Sometime in mid-May would be ideal. Thank you. | 1.0 | Copy some updated data files into http://data.jncc.gov.uk/data/ - I will be redacting some data from a large number of data files that are currently stored in http://data.jncc.gov.uk/data/ for access from Data.gov.uk.
Once this work is complete I will need developer assistance to replace the existing files with the updated files. Sometime in mid-May would be ideal. Thank you. | priority | copy some updated data files into i will be redacting some data from a large number of data files that are currently stored in for access from data gov uk once this work is complete i will need developer assistance to replace the existing files with the updated files sometime in mid may would be ideal thank you | 1 |
525,472 | 15,254,346,161 | IssuesEvent | 2021-02-20 11:36:51 | epam/Indigo | https://api.github.com/repos/epam/Indigo | closed | Calculate method does not return errors if structure is not valid | Bug High priority | **Step to reproduce**
1. Call indigo.calculate() for [R-Member.zip](https://github.com/epam/Indigo/files/5952295/R-Member.zip)
**Expected result**
Method throws not empty exception
**Actual result**
exception is empty
| 1.0 | Calculate method does not return errors if structure is not valid - **Step to reproduce**
1. Call indigo.calculate() for [R-Member.zip](https://github.com/epam/Indigo/files/5952295/R-Member.zip)
**Expected result**
Method throws not empty exception
**Actual result**
exception is empty
| priority | calculate method does not return errors if structure is not valid step to reproduce call indigo calculate for expected result method throws not empty exception actual result exception is empty | 1 |
570,925 | 17,023,209,413 | IssuesEvent | 2021-07-03 00:52:10 | microsoft/winget-create | https://api.github.com/repos/microsoft/winget-create | closed | Multiple URL support | High-Priority In-PR Issue-Feature | # Description of the new feature/enhancement
Add the possibility to add more than one installer in wingetcreate new and either wingetcreate update
# Proposed technical implementation details (optional)
~I think this is a low priority feature request, so keep calm and develop the rest 😆~
I didn't saw the source code, so I'm ignorant | 1.0 | Multiple URL support - # Description of the new feature/enhancement
Add the possibility to add more than one installer in wingetcreate new and either wingetcreate update
# Proposed technical implementation details (optional)
~I think this is a low priority feature request, so keep calm and develop the rest 😆~
I didn't saw the source code, so I'm ignorant | priority | multiple url support description of the new feature enhancement add the possibility to add more than one installer in wingetcreate new and either wingetcreate update proposed technical implementation details optional i think this is a low priority feature request so keep calm and develop the rest 😆 i didn t saw the source code so i m ignorant | 1 |
497,814 | 14,394,341,895 | IssuesEvent | 2020-12-03 01:03:37 | OpenPrinting/cups | https://api.github.com/repos/OpenPrinting/cups | closed | Canon ts6200 with IPP everywhere has duplicated grayscale option in color mode | bug priority-high | os : arch linux
cups : 2.3.3op1
I installed canon ts6200 series with IPP everywhere and there are one color and two grayscale options in color mode.
previously I installed this printer with IPP everywhere on ubuntu 20.04, it had one color and one grayscale in color mode.
Please check that, thanks. | 1.0 | Canon ts6200 with IPP everywhere has duplicated grayscale option in color mode - os : arch linux
cups : 2.3.3op1
I installed canon ts6200 series with IPP everywhere and there are one color and two grayscale options in color mode.
previously I installed this printer with IPP everywhere on ubuntu 20.04, it had one color and one grayscale in color mode.
Please check that, thanks. | priority | canon with ipp everywhere has duplicated grayscale option in color mode os arch linux cups i installed canon series with ipp everywhere and there are one color and two grayscale options in color mode previously i installed this printer with ipp everywhere on ubuntu it had one color and one grayscale in color mode please check that thanks | 1 |
548,944 | 16,082,091,978 | IssuesEvent | 2021-04-26 06:41:11 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.outlook.com - see bug description | browser-fixme ml-needsdiagnosis-false ml-probability-high priority-normal | <!-- @browser: Apple Mail 605.1.15 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/605.1.15 (KHTML, like Gecko) -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/71941 -->
**URL**: https://www.outlook.com
**Browser / Version**: Apple Mail 605.1.15
**Operating System**: Mac OS X 10.15.6
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Cannot enlarge text to a readable size
**Steps to Reproduce**:
I've been trying to read my email (Outlook) on Firefox Focus. My eyesight is not good and the text is too small to read. On Safari I can use finger and thumb to enlarge the page easily, but that feature isn't working on Firefox Focus with Outlook.
To make sure I am being clear: I can enlarge text on Firefox, ( just as I can on Safari ) when viewing other pages, but not when viewing Outlook.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.outlook.com - see bug description - <!-- @browser: Apple Mail 605.1.15 -->
<!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_6) AppleWebKit/605.1.15 (KHTML, like Gecko) -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/71941 -->
**URL**: https://www.outlook.com
**Browser / Version**: Apple Mail 605.1.15
**Operating System**: Mac OS X 10.15.6
**Tested Another Browser**: Yes Safari
**Problem type**: Something else
**Description**: Cannot enlarge text to a readable size
**Steps to Reproduce**:
I've been trying to read my email (Outlook) on Firefox Focus. My eyesight is not good and the text is too small to read. On Safari I can use finger and thumb to enlarge the page easily, but that feature isn't working on Firefox Focus with Outlook.
To make sure I am being clear: I can enlarge text on Firefox, ( just as I can on Safari ) when viewing other pages, but not when viewing Outlook.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | see bug description url browser version apple mail operating system mac os x tested another browser yes safari problem type something else description cannot enlarge text to a readable size steps to reproduce i ve been trying to read my email outlook on firefox focus my eyesight is not good and the text is too small to read on safari i can use finger and thumb to enlarge the page easily but that feature isn t working on firefox focus with outlook to make sure i am being clear i can enlarge text on firefox just as i can on safari when viewing other pages but not when viewing outlook browser configuration none from with ❤️ | 1 |
675,953 | 23,112,657,422 | IssuesEvent | 2022-07-27 14:12:45 | icecube/ASTERIA | https://api.github.com/repos/icecube/ASTERIA | closed | Replace internal `Source` and mass mixing classes with models from snewpy. | enhancement high priority | Since [snewpy](https://github.com/SNEWS2/snewpy) has standardized access to a number of CCSN models and flavor transformation scenarios, we should pull it in as a dependency. This would entail deleting the `Source` class in the `source` module, but keeping the functionality of the `photonic_energy_per_vol` member function.
About flavor transformations, we can probably completely remove the `oscillations` module and use the transformation classes implemented in snewpy. | 1.0 | Replace internal `Source` and mass mixing classes with models from snewpy. - Since [snewpy](https://github.com/SNEWS2/snewpy) has standardized access to a number of CCSN models and flavor transformation scenarios, we should pull it in as a dependency. This would entail deleting the `Source` class in the `source` module, but keeping the functionality of the `photonic_energy_per_vol` member function.
About flavor transformations, we can probably completely remove the `oscillations` module and use the transformation classes implemented in snewpy. | priority | replace internal source and mass mixing classes with models from snewpy since has standardized access to a number of ccsn models and flavor transformation scenarios we should pull it in as a dependency this would entail deleting the source class in the source module but keeping the functionality of the photonic energy per vol member function about flavor transformations we can probably completely remove the oscillations module and use the transformation classes implemented in snewpy | 1 |
71,758 | 3,367,799,855 | IssuesEvent | 2015-11-22 13:57:21 | jgirald/ES2015C | https://api.github.com/repos/jgirald/ES2015C | closed | Allow selecting units which are occluded by buildings | Development High Priority Team C | Sometimes, it's possible that a unit is totally or partially occluded by a building, e.g. it's behind a building.
While moving the camera provides a solution for this (even tough this feature is not implemented yet, too!), sometimes it's more convenient to be able to select the unit without moving the camera.
A way to implement this is the following: Instead of executing a single ray trace to detect the target object when the mouse is clicked, we will execute two ray traces. The first one will exclude the currently selected object from the targets, in order to be able to detect the objects behind the currently selected object. If we are able to detect such an object, we consider that the target of the click. Otherwise, the regular ray tracing procedure is done.
Example unit occluded by building: https://i.imgur.com/drGtvNc.png
NOTE: Only the selection code will be changed to implement this feature. The display/drawing code will not be changed in order to notify that a unit is behind a building.
**Product backlog item:** As a player, I want to see the features of my units/buildings selecting them with clicks. [High Priority]
**Estimated effort:** 2h
**Ending condition:** When a unit is behind a building, it should be possible to select it by clicking twice over it. The first click will select the building, and the second click will select the unit behind it. Clicking a third time will reselect the building, and a fourth will reselect the unit, etc. | 1.0 | Allow selecting units which are occluded by buildings - Sometimes, it's possible that a unit is totally or partially occluded by a building, e.g. it's behind a building.
While moving the camera provides a solution for this (even tough this feature is not implemented yet, too!), sometimes it's more convenient to be able to select the unit without moving the camera.
A way to implement this is the following: Instead of executing a single ray trace to detect the target object when the mouse is clicked, we will execute two ray traces. The first one will exclude the currently selected object from the targets, in order to be able to detect the objects behind the currently selected object. If we are able to detect such an object, we consider that the target of the click. Otherwise, the regular ray tracing procedure is done.
Example unit occluded by building: https://i.imgur.com/drGtvNc.png
NOTE: Only the selection code will be changed to implement this feature. The display/drawing code will not be changed in order to notify that a unit is behind a building.
**Product backlog item:** As a player, I want to see the features of my units/buildings selecting them with clicks. [High Priority]
**Estimated effort:** 2h
**Ending condition:** When a unit is behind a building, it should be possible to select it by clicking twice over it. The first click will select the building, and the second click will select the unit behind it. Clicking a third time will reselect the building, and a fourth will reselect the unit, etc. | priority | allow selecting units which are occluded by buildings sometimes it s possible that a unit is totally or partially occluded by a building e g it s behind a building while moving the camera provides a solution for this even tough this feature is not implemented yet too sometimes it s more convenient to be able to select the unit without moving the camera a way to implement this is the following instead of executing a single ray trace to detect the target object when the mouse is clicked we will execute two ray traces the first one will exclude the currently selected object from the targets in order to be able to detect the objects behind the currently selected object if we are able to detect such an object we consider that the target of the click otherwise the regular ray tracing procedure is done example unit occluded by building note only the selection code will be changed to implement this feature the display drawing code will not be changed in order to notify that a unit is behind a building product backlog item as a player i want to see the features of my units buildings selecting them with clicks estimated effort ending condition when a unit is behind a building it should be possible to select it by clicking twice over it the first click will select the building and the second click will select the unit behind it clicking a third time will reselect the building and a fourth will reselect the unit etc | 1 |
707,653 | 24,313,034,011 | IssuesEvent | 2022-09-30 01:44:51 | UC-Davis-molecular-computing/scadnano | https://api.github.com/repos/UC-Davis-molecular-computing/scadnano | closed | fix bug in assigning DNA sequence to strand bound to another strand with an extension | bug high priority closed in dev | Create two bound strands, one with an extension, and assign DNA to the other with "assign complement to bound strands" selected:


It generates this error:
```
Uncaught (in promise) TypeError: Cannot read properties of null (reading 'length')
at merge_wildcards (util.dart:1257:9)
at Object.compute_dna_complement_from (assign_or_remove_dna_reducer.dart:198:14)
at Object.assign_dna_reducer (assign_or_remove_dna_reducer.dart:89:26)
at DevToolsMiddleware.new.assign_dna_middleware (assign_dna.dart:16:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.throttle_middleware (throttle.dart:27:9)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.example_design_selected_middleware (example_design_selected.dart:9:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.export_cadnano_or_codenano_file_middleware (export_cadnano_or_codenano_file.dart:25:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.load_file_middleware (load_file.dart:49:9)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.save_file_middleware (save_file.dart:11:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.export_svg_middleware (export_svg.dart:42:9)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.forbid_create_circular_strand_no_crossovers_middleware$ (forbid_create_circular_strand_no_crossovers_middleware.dart:88:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.move_ensure_all_in_same_helix_group_middleware (move_ensure_same_group.dart:21:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.local_storage_middleware (local_storage.dart:130:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.reset_local_storage_middleware (reset_local_storage.dart:19:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at Store.new.dispatch (store.dart:267:25)
at DevToolsStore.new.dispatch (store.dart:74:29)
at app.App.new.dispatch (app.dart:140:13)
at ask_for_assign_dna_sequence (design_main_strand.dart:1124:7)
at ask_for_assign_dna_sequence.next (<anonymous>)
at async_patch.dart:45:50
at _RootZone.runUnary (zone.dart:1613:54)
at _FutureListener.thenAwait.handleValue (future_impl.dart:155:18)
at handleValueCallback (future_impl.dart:707:44)
at _Future._propagateToListeners (future_impl.dart:736:13)
at [_completeWithValue] (future_impl.dart:542:5)
at async._AsyncCallbackEntry.new.callback (future_impl.dart:580:7)
at Object._microtaskLoop (schedule_microtask.dart:40:11)
at _startMicrotaskLoop (schedule_microtask.dart:49:5)
at async_patch.dart:166:15
``` | 1.0 | fix bug in assigning DNA sequence to strand bound to another strand with an extension - Create two bound strands, one with an extension, and assign DNA to the other with "assign complement to bound strands" selected:


It generates this error:
```
Uncaught (in promise) TypeError: Cannot read properties of null (reading 'length')
at merge_wildcards (util.dart:1257:9)
at Object.compute_dna_complement_from (assign_or_remove_dna_reducer.dart:198:14)
at Object.assign_dna_reducer (assign_or_remove_dna_reducer.dart:89:26)
at DevToolsMiddleware.new.assign_dna_middleware (assign_dna.dart:16:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.throttle_middleware (throttle.dart:27:9)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.example_design_selected_middleware (example_design_selected.dart:9:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.export_cadnano_or_codenano_file_middleware (export_cadnano_or_codenano_file.dart:25:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.load_file_middleware (load_file.dart:49:9)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.save_file_middleware (save_file.dart:11:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.export_svg_middleware (export_svg.dart:42:9)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.forbid_create_circular_strand_no_crossovers_middleware$ (forbid_create_circular_strand_no_crossovers_middleware.dart:88:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.move_ensure_all_in_same_helix_group_middleware (move_ensure_same_group.dart:21:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.local_storage_middleware (local_storage.dart:130:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at middleware.dart:35:20
at DevToolsMiddleware.new.reset_local_storage_middleware (reset_local_storage.dart:19:7)
at DevToolsMiddleware.new.call (middleware.dart:39:25)
at store.dart:255:43
at Store.new.dispatch (store.dart:267:25)
at DevToolsStore.new.dispatch (store.dart:74:29)
at app.App.new.dispatch (app.dart:140:13)
at ask_for_assign_dna_sequence (design_main_strand.dart:1124:7)
at ask_for_assign_dna_sequence.next (<anonymous>)
at async_patch.dart:45:50
at _RootZone.runUnary (zone.dart:1613:54)
at _FutureListener.thenAwait.handleValue (future_impl.dart:155:18)
at handleValueCallback (future_impl.dart:707:44)
at _Future._propagateToListeners (future_impl.dart:736:13)
at [_completeWithValue] (future_impl.dart:542:5)
at async._AsyncCallbackEntry.new.callback (future_impl.dart:580:7)
at Object._microtaskLoop (schedule_microtask.dart:40:11)
at _startMicrotaskLoop (schedule_microtask.dart:49:5)
at async_patch.dart:166:15
``` | priority | fix bug in assigning dna sequence to strand bound to another strand with an extension create two bound strands one with an extension and assign dna to the other with assign complement to bound strands selected it generates this error uncaught in promise typeerror cannot read properties of null reading length at merge wildcards util dart at object compute dna complement from assign or remove dna reducer dart at object assign dna reducer assign or remove dna reducer dart at devtoolsmiddleware new assign dna middleware assign dna dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new throttle middleware throttle dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new example design selected middleware example design selected dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new export cadnano or codenano file middleware export cadnano or codenano file dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new load file middleware load file dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new save file middleware save file dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new export svg middleware export svg dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new forbid create circular strand no crossovers middleware forbid create circular strand no crossovers middleware dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new move ensure all in same helix group middleware move ensure same group dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new local storage middleware local storage dart at devtoolsmiddleware new call middleware dart at store dart at middleware dart at devtoolsmiddleware new reset local storage middleware reset local storage dart at devtoolsmiddleware new call middleware dart at store dart at store new dispatch store dart at devtoolsstore new dispatch store dart at app app new dispatch app dart at ask for assign dna sequence design main strand dart at ask for assign dna sequence next at async patch dart at rootzone rununary zone dart at futurelistener thenawait handlevalue future impl dart at handlevaluecallback future impl dart at future propagatetolisteners future impl dart at future impl dart at async asynccallbackentry new callback future impl dart at object microtaskloop schedule microtask dart at startmicrotaskloop schedule microtask dart at async patch dart | 1 |
288,019 | 8,824,106,942 | IssuesEvent | 2019-01-02 15:55:52 | ScottLogic/data-engineering-generator | https://api.github.com/repos/ScottLogic/data-engineering-generator | closed | Cucumber Exhaustive Combination: GC overhead limit errors thrown for small requests | bug generator high-priority | ## Bug Report
Exhaustive combination tests across three fields, each with a set of five data options, were run. This was expected to generate 124 rows of data, not expected to be a large request, but failed with a memory error.
[Test Example 2.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648567/Test.Example.2.txt)
A second test expecting 25 roes from three fields (one set of 1 data option and two sets of five data options) also failed with the same memory error.
[Test Example 1.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648569/Test.Example.1.txt)
### Steps to Reproduce:
1. Run tests from the feature file ExhaustiveCombination.feature
2. See that there is a failure for the test **Scenario: Running an exhaustive combination strategy across three fields with five data options each should be successful**
3. Note that the test is long running and then fails with an out of memory error
### Expected Result:
These tests should generate the expected 124 / 25 rows of data based on the exhaustive combination strategy.
### Actual Result:
The tests fail with a memory error:
> java.lang.OutOfMemoryError: GC overhead limit exceeded
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:35)
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.Iterator.forEachRemaining(Iterator.java:116)
> at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721)
> at java.util.stream.AbstractTask.compute(AbstractTask.java:316)
> at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731)
> at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
> at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401)
> at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734)
> at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714)
> at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233)
> at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
> at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
> at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.permute(ExhaustiveCombinationStrategy.java:19)
> at com.scottlogic.deg.generator.generation.databags.MultiplexingDataBagSource.generate(MultiplexingDataBagSource.java:17)
> at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:379)
>
>
> java.lang.OutOfMemoryError: Java heap space
> at java.util.HashMap.resize(HashMap.java:704)
> at java.util.HashMap.putVal(HashMap.java:629)
> at java.util.HashMap.put(HashMap.java:612)
> at com.scottlogic.deg.generator.generation.databags.DataBag$DataBagBuilder.set(DataBag.java:76)
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:42)
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.Iterator.forEachRemaining(Iterator.java:116)
> at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721)
> at java.util.stream.AbstractTask.compute(AbstractTask.java:316)
> at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731)
> at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
> at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401)
> at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734)
> at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714)
> at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233)
> at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
> at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:528)
### Environment:
ExhaustiveCombination branch running in full generation strategy & exhaustive combination strategy on IntelliJ.
### Additional Context:
Related to existing performance issues.
| 1.0 | Cucumber Exhaustive Combination: GC overhead limit errors thrown for small requests - ## Bug Report
Exhaustive combination tests across three fields, each with a set of five data options, were run. This was expected to generate 124 rows of data, not expected to be a large request, but failed with a memory error.
[Test Example 2.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648567/Test.Example.2.txt)
A second test expecting 25 roes from three fields (one set of 1 data option and two sets of five data options) also failed with the same memory error.
[Test Example 1.txt](https://github.com/ScottLogic/data-engineering-generator/files/2648569/Test.Example.1.txt)
### Steps to Reproduce:
1. Run tests from the feature file ExhaustiveCombination.feature
2. See that there is a failure for the test **Scenario: Running an exhaustive combination strategy across three fields with five data options each should be successful**
3. Note that the test is long running and then fails with an out of memory error
### Expected Result:
These tests should generate the expected 124 / 25 rows of data based on the exhaustive combination strategy.
### Actual Result:
The tests fail with a memory error:
> java.lang.OutOfMemoryError: GC overhead limit exceeded
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:35)
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.Iterator.forEachRemaining(Iterator.java:116)
> at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721)
> at java.util.stream.AbstractTask.compute(AbstractTask.java:316)
> at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731)
> at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
> at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401)
> at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734)
> at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714)
> at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233)
> at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
> at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
> at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.permute(ExhaustiveCombinationStrategy.java:19)
> at com.scottlogic.deg.generator.generation.databags.MultiplexingDataBagSource.generate(MultiplexingDataBagSource.java:17)
> at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:379)
>
>
> java.lang.OutOfMemoryError: Java heap space
> at java.util.HashMap.resize(HashMap.java:704)
> at java.util.HashMap.putVal(HashMap.java:629)
> at java.util.HashMap.put(HashMap.java:612)
> at com.scottlogic.deg.generator.generation.databags.DataBag$DataBagBuilder.set(DataBag.java:76)
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller.lambda$generate$0(FieldSpecFulfiller.java:42)
> at com.scottlogic.deg.generator.generation.FieldSpecFulfiller$$Lambda$133/726181440.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.Iterator.forEachRemaining(Iterator.java:116)
> at java.util.Spliterators$IteratorSpliterator.forEachRemaining(Spliterators.java:1801)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.StreamSpliterators$WrappingSpliterator.forEachRemaining(StreamSpliterators.java:312)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:747)
> at java.util.stream.ReduceOps$ReduceTask.doLeaf(ReduceOps.java:721)
> at java.util.stream.AbstractTask.compute(AbstractTask.java:316)
> at java.util.concurrent.CountedCompleter.exec(CountedCompleter.java:731)
> at java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:289)
> at java.util.concurrent.ForkJoinTask.doInvoke(ForkJoinTask.java:401)
> at java.util.concurrent.ForkJoinTask.invoke(ForkJoinTask.java:734)
> at java.util.stream.ReduceOps$ReduceOp.evaluateParallel(ReduceOps.java:714)
> at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:233)
> at java.util.stream.ReferencePipeline.collect(ReferencePipeline.java:499)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy.lambda$permute$0(ExhaustiveCombinationStrategy.java:18)
> at com.scottlogic.deg.generator.generation.combination_strategies.ExhaustiveCombinationStrategy$$Lambda$87/737945227.apply(Unknown Source)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.stream.ReferencePipeline$3$1.accept(ReferencePipeline.java:193)
> at java.util.ArrayList$ArrayListSpliterator.forEachRemaining(ArrayList.java:1382)
> at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:481)
> at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:471)
> at java.util.stream.ReduceOps$ReduceOp.evaluateSequential(ReduceOps.java:708)
> at ✽.the following data should be generated:(C:/DataGenerator/data-engineering-generator/generator/src/test/java/com/scottlogic/deg/generator/cucumber/ExhaustiveCombination.feature:528)
### Environment:
ExhaustiveCombination branch running in full generation strategy & exhaustive combination strategy on IntelliJ.
### Additional Context:
Related to existing performance issues.
| priority | cucumber exhaustive combination gc overhead limit errors thrown for small requests bug report exhaustive combination tests across three fields each with a set of five data options were run this was expected to generate rows of data not expected to be a large request but failed with a memory error a second test expecting roes from three fields one set of data option and two sets of five data options also failed with the same memory error steps to reproduce run tests from the feature file exhaustivecombination feature see that there is a failure for the test scenario running an exhaustive combination strategy across three fields with five data options each should be successful note that the test is long running and then fails with an out of memory error expected result these tests should generate the expected rows of data based on the exhaustive combination strategy actual result the tests fail with a memory error java lang outofmemoryerror gc overhead limit exceeded at com scottlogic deg generator generation fieldspecfulfiller lambda generate fieldspecfulfiller java at com scottlogic deg generator generation fieldspecfulfiller lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util iterator foreachremaining iterator java at java util spliterators iteratorspliterator foreachremaining spliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream streamspliterators wrappingspliterator foreachremaining streamspliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reducetask doleaf reduceops java at java util stream reduceops reducetask doleaf reduceops java at java util stream abstracttask compute abstracttask java at java util concurrent countedcompleter exec countedcompleter java at java util concurrent forkjointask doexec forkjointask java at java util concurrent forkjointask doinvoke forkjointask java at java util concurrent forkjointask invoke forkjointask java at java util stream reduceops reduceop evaluateparallel reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda permute exhaustivecombinationstrategy java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util stream referencepipeline accept referencepipeline java at java util arraylist arraylistspliterator foreachremaining arraylist java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy permute exhaustivecombinationstrategy java at com scottlogic deg generator generation databags multiplexingdatabagsource generate multiplexingdatabagsource java at ✽ the following data should be generated c datagenerator data engineering generator generator src test java com scottlogic deg generator cucumber exhaustivecombination feature java lang outofmemoryerror java heap space at java util hashmap resize hashmap java at java util hashmap putval hashmap java at java util hashmap put hashmap java at com scottlogic deg generator generation databags databag databagbuilder set databag java at com scottlogic deg generator generation fieldspecfulfiller lambda generate fieldspecfulfiller java at com scottlogic deg generator generation fieldspecfulfiller lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util iterator foreachremaining iterator java at java util spliterators iteratorspliterator foreachremaining spliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream streamspliterators wrappingspliterator foreachremaining streamspliterators java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reducetask doleaf reduceops java at java util stream reduceops reducetask doleaf reduceops java at java util stream abstracttask compute abstracttask java at java util concurrent countedcompleter exec countedcompleter java at java util concurrent forkjointask doexec forkjointask java at java util concurrent forkjointask doinvoke forkjointask java at java util concurrent forkjointask invoke forkjointask java at java util stream reduceops reduceop evaluateparallel reduceops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline collect referencepipeline java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda permute exhaustivecombinationstrategy java at com scottlogic deg generator generation combination strategies exhaustivecombinationstrategy lambda apply unknown source at java util stream referencepipeline accept referencepipeline java at java util stream referencepipeline accept referencepipeline java at java util arraylist arraylistspliterator foreachremaining arraylist java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream reduceops reduceop evaluatesequential reduceops java at ✽ the following data should be generated c datagenerator data engineering generator generator src test java com scottlogic deg generator cucumber exhaustivecombination feature environment exhaustivecombination branch running in full generation strategy exhaustive combination strategy on intellij additional context related to existing performance issues | 1 |
143,589 | 5,520,549,714 | IssuesEvent | 2017-03-19 06:36:21 | nim-lang/Nim | https://api.github.com/repos/nim-lang/Nim | closed | A version of httpclient that uses streams in the stdlib | High Priority Stdlib | I decided I would build a script in `nim` that would download some rather large files, as you do. I wanted to use `httpclient` but it bothered me that there were no streams for the body, just strings. This meant to me that the downloaded data is stored into memory. After reading the source of `httpclient`, this is indeed the case. I can see how this would not be a problem for small files but when downloading rather large files - that is files that would easily deplete any RAM - it would be a bit of a problem. So, in my naive attempt, I changed all usage of strings, when downloading the body, into streams; and here is my gist for it [httpclientstreams.nim](https://gist.github.com/NebulaFox/89926ebfa56840ca49a0202cc39d9de3).
I successfully tested my changes on downloading some images and other media with `httpclientstreams`. As a result of using streams, I removed all the `*Content` methods, because as a user, if I just wanted the content, I would provide a `Stream` and forget about the `Response`. As an additional consequence of using streams for the body of the response, I have also removed the `body` field from `Response`.
As this was strictly a naive attempt, there are leftover `proc`s from `httpclient` that should be deleted, but in saying that is `httpclientstreams` an acceptable implementation to be included into the stdlib?
A couple of other things to note is a benefit of using `streams` is that it would make it possible to process the incoming body of data on the fly. And, I can see `httpclient` easily using the streamed-version encapsulating the methods with `StringStream`, however I do understand that `streams` is not supported for JavaScript.
| 1.0 | A version of httpclient that uses streams in the stdlib - I decided I would build a script in `nim` that would download some rather large files, as you do. I wanted to use `httpclient` but it bothered me that there were no streams for the body, just strings. This meant to me that the downloaded data is stored into memory. After reading the source of `httpclient`, this is indeed the case. I can see how this would not be a problem for small files but when downloading rather large files - that is files that would easily deplete any RAM - it would be a bit of a problem. So, in my naive attempt, I changed all usage of strings, when downloading the body, into streams; and here is my gist for it [httpclientstreams.nim](https://gist.github.com/NebulaFox/89926ebfa56840ca49a0202cc39d9de3).
I successfully tested my changes on downloading some images and other media with `httpclientstreams`. As a result of using streams, I removed all the `*Content` methods, because as a user, if I just wanted the content, I would provide a `Stream` and forget about the `Response`. As an additional consequence of using streams for the body of the response, I have also removed the `body` field from `Response`.
As this was strictly a naive attempt, there are leftover `proc`s from `httpclient` that should be deleted, but in saying that is `httpclientstreams` an acceptable implementation to be included into the stdlib?
A couple of other things to note is a benefit of using `streams` is that it would make it possible to process the incoming body of data on the fly. And, I can see `httpclient` easily using the streamed-version encapsulating the methods with `StringStream`, however I do understand that `streams` is not supported for JavaScript.
| priority | a version of httpclient that uses streams in the stdlib i decided i would build a script in nim that would download some rather large files as you do i wanted to use httpclient but it bothered me that there were no streams for the body just strings this meant to me that the downloaded data is stored into memory after reading the source of httpclient this is indeed the case i can see how this would not be a problem for small files but when downloading rather large files that is files that would easily deplete any ram it would be a bit of a problem so in my naive attempt i changed all usage of strings when downloading the body into streams and here is my gist for it i successfully tested my changes on downloading some images and other media with httpclientstreams as a result of using streams i removed all the content methods because as a user if i just wanted the content i would provide a stream and forget about the response as an additional consequence of using streams for the body of the response i have also removed the body field from response as this was strictly a naive attempt there are leftover proc s from httpclient that should be deleted but in saying that is httpclientstreams an acceptable implementation to be included into the stdlib a couple of other things to note is a benefit of using streams is that it would make it possible to process the incoming body of data on the fly and i can see httpclient easily using the streamed version encapsulating the methods with stringstream however i do understand that streams is not supported for javascript | 1 |
235,966 | 7,744,442,565 | IssuesEvent | 2018-05-29 15:24:00 | Gloirin/m2gTest | https://api.github.com/repos/Gloirin/m2gTest | closed | 0004932:
add id and label to filter objects + filter groups | Feature Request Tinebase high priority | **Reported by pschuele on 6 Oct 2011 17:28**
add id and label to filter objects + filter groups
- label should be saved in persistent filter, id not
| 1.0 | 0004932:
add id and label to filter objects + filter groups - **Reported by pschuele on 6 Oct 2011 17:28**
add id and label to filter objects + filter groups
- label should be saved in persistent filter, id not
| priority | add id and label to filter objects filter groups reported by pschuele on oct add id and label to filter objects filter groups label should be saved in persistent filter id not | 1 |
174,236 | 6,538,052,542 | IssuesEvent | 2017-09-01 02:57:32 | enforcer574/smashclub | https://api.github.com/repos/enforcer574/smashclub | opened | Investigate use of special ASCII characters in profile fields | Incident Priority 2 (High) Vulnerability | One account has used a special HTML character in their catch phrase.
This is most likely not an issue, but it warrants an investigation into how the current input sanitization handles these characters and if it presents an injection vulnerability. | 1.0 | Investigate use of special ASCII characters in profile fields - One account has used a special HTML character in their catch phrase.
This is most likely not an issue, but it warrants an investigation into how the current input sanitization handles these characters and if it presents an injection vulnerability. | priority | investigate use of special ascii characters in profile fields one account has used a special html character in their catch phrase this is most likely not an issue but it warrants an investigation into how the current input sanitization handles these characters and if it presents an injection vulnerability | 1 |
328,920 | 10,001,478,950 | IssuesEvent | 2019-07-12 15:43:29 | databio/refgenconf | https://api.github.com/repos/databio/refgenconf | closed | adding genome attributes to config file | enhancement priority-high | related to #7 (but for genomes instead of for assets)
what if we want to add some other attributes about a genome? examples include a description, URL to where it came from, what species it comes from, how long it is, etc.
maybe the config format should introduce an "assets" attribute under genome so they are not right under it? | 1.0 | adding genome attributes to config file - related to #7 (but for genomes instead of for assets)
what if we want to add some other attributes about a genome? examples include a description, URL to where it came from, what species it comes from, how long it is, etc.
maybe the config format should introduce an "assets" attribute under genome so they are not right under it? | priority | adding genome attributes to config file related to but for genomes instead of for assets what if we want to add some other attributes about a genome examples include a description url to where it came from what species it comes from how long it is etc maybe the config format should introduce an assets attribute under genome so they are not right under it | 1 |
787,797 | 27,731,352,788 | IssuesEvent | 2023-03-15 08:17:45 | AY2223S2-CS2103T-W15-1/tp | https://api.github.com/repos/AY2223S2-CS2103T-W15-1/tp | closed | As a user I want to be able to add a new contact with missing telegram | type.Story priority.High | ...so I can add people when I don’t know their telegram | 1.0 | As a user I want to be able to add a new contact with missing telegram - ...so I can add people when I don’t know their telegram | priority | as a user i want to be able to add a new contact with missing telegram so i can add people when i don’t know their telegram | 1 |
248,195 | 7,928,255,472 | IssuesEvent | 2018-07-06 10:56:31 | checkorg1/test | https://api.github.com/repos/checkorg1/test | closed | Token UI to support display of variants and to show meta-content on hover | Highest Priority | _From sardhanari on 16-07-2017 08:19 IST_
<img width="130" alt="screen shot 2017-07-15 at 10 02 36 pm" src="https://user-images.githubusercontent.com/28209236/28243954-0995c710-69aa-11e7-8a07-5a951d112e4a.png">
Selecting the **chevron**(down arrow symbol) shows the list of variants. Selecting a variant replaces the token. So in the compare page, one can check the association strengths against the different variants. Similarly in the neighborhood page, we could cycle through the different variants to see how it affects the neighborhood. In the neighborhood page, the chevron symbol shows up only for the token whose neighborhood is being looked at.
Hovering on the **info icon**(i in a circle) will display the meta-content specific to the token. The info icon should be displayed on any token for which there is meta-content. That way one does not have to go to the neighborhood page explicitly to see the meta-content.
HTML/CSS to render the above button design:
```html
<div class="chip nearest-word ellipsis draggable-token" id="chip-ranking" nearest-word="pd_1"
style="
box-shadow: rgba(200,200,200,1) 1px 0px 0px 0px, rgba(200,200,200,1) -1px 0px 0px 0px, rgba(200,200,200,1) 0px 1px 0px 0px, rgba(200,200,200,1) 0px -1px 0px 0px;
">
<div>pd_1</div>
<i class="fa fa-chevron-down" aria-hidden="true" style="
line-height: inherit;
margin-left: 8px;
"></i>
<div style="
border-left: 1px solid #f7f7f7;
border-right: 1px solid #a2a2a2;
margin-left: 5px;
"></div>
<i class="fa fa-info-circle" aria-hidden="true" style="
line-height: inherit;
font-size: 15px;
margin-left: 9px;
x solid #38546d;
"></i>
</div>
```
**Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/109] | 1.0 | Token UI to support display of variants and to show meta-content on hover - _From sardhanari on 16-07-2017 08:19 IST_
<img width="130" alt="screen shot 2017-07-15 at 10 02 36 pm" src="https://user-images.githubusercontent.com/28209236/28243954-0995c710-69aa-11e7-8a07-5a951d112e4a.png">
Selecting the **chevron**(down arrow symbol) shows the list of variants. Selecting a variant replaces the token. So in the compare page, one can check the association strengths against the different variants. Similarly in the neighborhood page, we could cycle through the different variants to see how it affects the neighborhood. In the neighborhood page, the chevron symbol shows up only for the token whose neighborhood is being looked at.
Hovering on the **info icon**(i in a circle) will display the meta-content specific to the token. The info icon should be displayed on any token for which there is meta-content. That way one does not have to go to the neighborhood page explicitly to see the meta-content.
HTML/CSS to render the above button design:
```html
<div class="chip nearest-word ellipsis draggable-token" id="chip-ranking" nearest-word="pd_1"
style="
box-shadow: rgba(200,200,200,1) 1px 0px 0px 0px, rgba(200,200,200,1) -1px 0px 0px 0px, rgba(200,200,200,1) 0px 1px 0px 0px, rgba(200,200,200,1) 0px -1px 0px 0px;
">
<div>pd_1</div>
<i class="fa fa-chevron-down" aria-hidden="true" style="
line-height: inherit;
margin-left: 8px;
"></i>
<div style="
border-left: 1px solid #f7f7f7;
border-right: 1px solid #a2a2a2;
margin-left: 5px;
"></div>
<i class="fa fa-info-circle" aria-hidden="true" style="
line-height: inherit;
font-size: 15px;
margin-left: 9px;
x solid #38546d;
"></i>
</div>
```
**Copied from original issue**:[https://github.com/lumenbiomics/nfer/issues/109] | priority | token ui to support display of variants and to show meta content on hover from sardhanari on ist img width alt screen shot at pm src selecting the chevron down arrow symbol shows the list of variants selecting a variant replaces the token so in the compare page one can check the association strengths against the different variants similarly in the neighborhood page we could cycle through the different variants to see how it affects the neighborhood in the neighborhood page the chevron symbol shows up only for the token whose neighborhood is being looked at hovering on the info icon i in a circle will display the meta content specific to the token the info icon should be displayed on any token for which there is meta content that way one does not have to go to the neighborhood page explicitly to see the meta content html css to render the above button design html div class chip nearest word ellipsis draggable token id chip ranking nearest word pd style box shadow rgba rgba rgba rgba pd i class fa fa chevron down aria hidden true style line height inherit margin left div style border left solid border right solid margin left i class fa fa info circle aria hidden true style line height inherit font size margin left x solid copied from original issue | 1 |
285,263 | 8,756,713,256 | IssuesEvent | 2018-12-14 18:40:35 | Datapolitan-Training/data-analysis-r | https://api.github.com/repos/Datapolitan-Training/data-analysis-r | closed | Add explicit sort and filter instructions | High Priority enhancement | Add the code to do the sorting with basic `sort()` or `order()` operations | 1.0 | Add explicit sort and filter instructions - Add the code to do the sorting with basic `sort()` or `order()` operations | priority | add explicit sort and filter instructions add the code to do the sorting with basic sort or order operations | 1 |
808,036 | 30,031,256,633 | IssuesEvent | 2023-06-27 09:44:16 | tweag/nixpkgs-graph-explorer | https://api.github.com/repos/tweag/nixpkgs-graph-explorer | closed | Display query error | component:backend component:ui priority:high | **Is your feature request related to a problem? Please describe.**
When a request is done using an invalid query, I want to see the error to fix the query.
For now, it only displays "Error fetching graph data :( Please check your query and try again later".
<img width="704" alt="image" src="https://github.com/tweag/nixpkgs-graph-explorer/assets/1882000/f1a8df02-9feb-49a0-a721-cef717e1df74">
**Describe the solution you'd like**
When the query is invalid, I want to see the query error.
| 1.0 | Display query error - **Is your feature request related to a problem? Please describe.**
When a request is done using an invalid query, I want to see the error to fix the query.
For now, it only displays "Error fetching graph data :( Please check your query and try again later".
<img width="704" alt="image" src="https://github.com/tweag/nixpkgs-graph-explorer/assets/1882000/f1a8df02-9feb-49a0-a721-cef717e1df74">
**Describe the solution you'd like**
When the query is invalid, I want to see the query error.
| priority | display query error is your feature request related to a problem please describe when a request is done using an invalid query i want to see the error to fix the query for now it only displays error fetching graph data please check your query and try again later img width alt image src describe the solution you d like when the query is invalid i want to see the query error | 1 |
724,836 | 24,943,297,227 | IssuesEvent | 2022-10-31 20:56:34 | bounswe/bounswe2022group4 | https://api.github.com/repos/bounswe/bounswe2022group4 | closed | Frontend: Lifting State Up Login Information | Category - To Do Priority - High Status: In Progress whom: individual Difficulty - Hard Language - React.js Team - Frontend | Description:
After user log in, he/she should not view sign in and sign up buttons anymore. In order to prevent that i need to store login information in a global variable.
Steps:
1) Make research about how to lifting a state up in React.
2) Store login information of the user in global variable.
3) Use the variable in navigation bar
4) Hide profile page if user not logged in
5) Hide sign in and sign up button and show sign out button if user logged in
Reviewer: @BeratDamar
Deadline: 31.10.2022 23.59 | 1.0 | Frontend: Lifting State Up Login Information - Description:
After user log in, he/she should not view sign in and sign up buttons anymore. In order to prevent that i need to store login information in a global variable.
Steps:
1) Make research about how to lifting a state up in React.
2) Store login information of the user in global variable.
3) Use the variable in navigation bar
4) Hide profile page if user not logged in
5) Hide sign in and sign up button and show sign out button if user logged in
Reviewer: @BeratDamar
Deadline: 31.10.2022 23.59 | priority | frontend lifting state up login information description after user log in he she should not view sign in and sign up buttons anymore in order to prevent that i need to store login information in a global variable steps make research about how to lifting a state up in react store login information of the user in global variable use the variable in navigation bar hide profile page if user not logged in hide sign in and sign up button and show sign out button if user logged in reviewer beratdamar deadline | 1 |
635,114 | 20,379,366,256 | IssuesEvent | 2022-02-21 19:23:40 | cds-snc/notification-planning | https://api.github.com/repos/cds-snc/notification-planning | opened | Add page to UI when a session times out or user logs out | High Priority | Haute priorité Security | Sécurité UX | _Add page to UI when a session times out or user logs out_
# Description
As a (user), I need to be able to do (X) so that I can achieve (Y) outcome.
WHY are we building?
WHAT are we building?
VALUE created by our solution
# Acceptance Criteria** (Definition of done)
_To be refined through discussion with the team_
Given some context, when (X) action occurs, then (Y) outcome is achieved
If this user story emerged from User Research insights:
- [ ] Link research insight back to hypothesis from Epic or Objective in Airtable
- [ ] Once change/fix/feature is implemented, mark insight as "resolved" in Airtable
- [ ] Once change/fix/feature is implements, link insight to design artifacts (Figma) in Airtable
* A11y
* Bilingualism
* Privacy considerations
* Security controls in place
* Measuring success and metrics
| 1.0 | Add page to UI when a session times out or user logs out - _Add page to UI when a session times out or user logs out_
# Description
As a (user), I need to be able to do (X) so that I can achieve (Y) outcome.
WHY are we building?
WHAT are we building?
VALUE created by our solution
# Acceptance Criteria** (Definition of done)
_To be refined through discussion with the team_
Given some context, when (X) action occurs, then (Y) outcome is achieved
If this user story emerged from User Research insights:
- [ ] Link research insight back to hypothesis from Epic or Objective in Airtable
- [ ] Once change/fix/feature is implemented, mark insight as "resolved" in Airtable
- [ ] Once change/fix/feature is implements, link insight to design artifacts (Figma) in Airtable
* A11y
* Bilingualism
* Privacy considerations
* Security controls in place
* Measuring success and metrics
| priority | add page to ui when a session times out or user logs out add page to ui when a session times out or user logs out description as a user i need to be able to do x so that i can achieve y outcome why are we building what are we building value created by our solution acceptance criteria definition of done to be refined through discussion with the team given some context when x action occurs then y outcome is achieved if this user story emerged from user research insights link research insight back to hypothesis from epic or objective in airtable once change fix feature is implemented mark insight as resolved in airtable once change fix feature is implements link insight to design artifacts figma in airtable bilingualism privacy considerations security controls in place measuring success and metrics | 1 |
339,850 | 10,263,145,722 | IssuesEvent | 2019-08-22 13:49:05 | deep-learning-indaba/Baobab | https://api.github.com/repos/deep-learning-indaba/Baobab | closed | Attendance Admin: User with both normal registration and guest_registration are not indicated as "guests" | High Priority back-end front-end | Attendance Admin: User with both normal registration and guest_registration are not indicated as "Invited Guests"

| 1.0 | Attendance Admin: User with both normal registration and guest_registration are not indicated as "guests" - Attendance Admin: User with both normal registration and guest_registration are not indicated as "Invited Guests"

| priority | attendance admin user with both normal registration and guest registration are not indicated as guests attendance admin user with both normal registration and guest registration are not indicated as invited guests | 1 |
418,607 | 12,200,645,654 | IssuesEvent | 2020-04-30 05:20:34 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | [0.9.0 staging-1531] Strange disconnects | Priority: High Status: Fixed | Can't reproduce by 100%, but today I have it 3 times:
1st - I created laws, 2nd and 3rd I checked work party/contracts.
When disconnect start in log exceptions spam:
```
Exception: Expected to receive type key, since there is no view mapping for ID 4475, but type key was not given
at Eco.Shared.View.ViewManager.GetView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.UnpackageView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Unity.Properties.TypeConversion+Converter`2+ConvertDelegate[TSource,TDestination].Invoke (TSource value) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Utils.ListExtensions.FromBson[T] (System.Collections.Generic.IList`1[T] list, Eco.Shared.Serialization.BSONArray bsonArray) [0x00000] in <00000000000000000000000000000000>:0
at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0
at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0
at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0
at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0
Rethrow as TargetInvocationException: Exception has been thrown by the target of an invocation.
at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0
at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0
at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0
at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0
UnityEngine.Logger:LogException(Exception, Object)
UnityEngine.Debug:LogException(Exception)
FramePlanner.PlannerGroup:OnUpdate()
FramePlanner.FramePlannerSystem:OnUpdate()
Unity.Entities.ComponentSystem:InternalUpdate()
Unity.Entities.ComponentSystemGroup:OnUpdate()
Unity.Entities.ComponentSystem:InternalUpdate()
UnityEngine.Experimental.LowLevel.UpdateFunction:Invoke()
```
https://drive.google.com/file/d/1NYKGOIEf2XpbGXydUDLAT3hN550Se_ot/view?usp=sharing
then I can't connect to server, i need to restart game | 1.0 | [0.9.0 staging-1531] Strange disconnects - Can't reproduce by 100%, but today I have it 3 times:
1st - I created laws, 2nd and 3rd I checked work party/contracts.
When disconnect start in log exceptions spam:
```
Exception: Expected to receive type key, since there is no view mapping for ID 4475, but type key was not given
at Eco.Shared.View.ViewManager.GetView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.UnpackageView (Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Unity.Properties.TypeConversion+Converter`2+ConvertDelegate[TSource,TDestination].Invoke (TSource value) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Utils.ListExtensions.FromBson[T] (System.Collections.Generic.IList`1[T] list, Eco.Shared.Serialization.BSONArray bsonArray) [0x00000] in <00000000000000000000000000000000>:0
at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0
at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0
at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0
at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0
Rethrow as TargetInvocationException: Exception has been thrown by the target of an invocation.
at System.Reflection.MonoMethod.Invoke (System.Object obj, System.Reflection.BindingFlags invokeAttr, System.Reflection.Binder binder, System.Object[] parameters, System.Globalization.CultureInfo culture) [0x00000] in <00000000000000000000000000000000>:0
at System.Reflection.MethodBase.Invoke (System.Object obj, System.Object[] parameters) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIList (System.Type type, Eco.Shared.Serialization.BSONValue value) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBsonIDictionary (System.Type type, Eco.Shared.Serialization.BSONObject obj) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.Serialization.BsonManipulator.FromBson (Eco.Shared.Serialization.BSONValue value, System.Type type) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.View.UpdateProperty (Eco.Shared.Serialization.BSONValue propNameID, Eco.Shared.Serialization.BSONObject bson) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ViewNotify (Eco.Shared.Serialization.BSONObject bson, Eco.Shared.View.View view, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ApplyProperties (Eco.Shared.View.View view, Eco.Shared.Serialization.BSONObject bson, System.Collections.Generic.List`1[T] notifications) [0x00000] in <00000000000000000000000000000000>:0
at Eco.Shared.View.ViewManager.ReceiveViewUpdate (Eco.Shared.Serialization.BSONArray updates) [0x00000] in <00000000000000000000000000000000>:0
at ClientPacketHandler.ReceiveWhile (Eco.Shared.Networking.NetworkClient client, System.Func`1[TResult] condition) [0x00000] in <00000000000000000000000000000000>:0
at System.Action`1[T].Invoke (T obj) [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.PlannerGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at FramePlanner.FramePlannerSystem.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystemGroup.OnUpdate () [0x00000] in <00000000000000000000000000000000>:0
at Unity.Entities.ComponentSystem.InternalUpdate () [0x00000] in <00000000000000000000000000000000>:0
at UnityEngine.Experimental.LowLevel.PlayerLoopSystem+UpdateFunction.Invoke () [0x00000] in <00000000000000000000000000000000>:0
UnityEngine.Logger:LogException(Exception, Object)
UnityEngine.Debug:LogException(Exception)
FramePlanner.PlannerGroup:OnUpdate()
FramePlanner.FramePlannerSystem:OnUpdate()
Unity.Entities.ComponentSystem:InternalUpdate()
Unity.Entities.ComponentSystemGroup:OnUpdate()
Unity.Entities.ComponentSystem:InternalUpdate()
UnityEngine.Experimental.LowLevel.UpdateFunction:Invoke()
```
https://drive.google.com/file/d/1NYKGOIEf2XpbGXydUDLAT3hN550Se_ot/view?usp=sharing
then I can't connect to server, i need to restart game | priority | strange disconnects can t reproduce by but today i have it times i created laws and i checked work party contracts when disconnect start in log exceptions spam exception expected to receive type key since there is no view mapping for id but type key was not given at eco shared view viewmanager getview eco shared serialization bsonobject bson in at eco shared view viewmanager unpackageview eco shared serialization bsonobject bson in at unity properties typeconversion converter convertdelegate invoke tsource value in at eco shared serialization bsonmanipulator frombson eco shared serialization bsonvalue value system type type in at eco shared utils listextensions frombson system collections generic ilist list eco shared serialization bsonarray bsonarray in at system reflection monomethod invoke system object obj system reflection bindingflags invokeattr system reflection binder binder system object parameters system globalization cultureinfo culture in at system reflection methodbase invoke system object obj system object parameters in at eco shared serialization bsonmanipulator frombsonilist system type type eco shared serialization bsonvalue value in at eco shared serialization bsonmanipulator frombsonidictionary system type type eco shared serialization bsonobject obj in at eco shared serialization bsonmanipulator frombson eco shared serialization bsonvalue value system type type in at eco shared view view updateproperty eco shared serialization bsonvalue propnameid eco shared serialization bsonobject bson in at eco shared view viewmanager viewnotify eco shared serialization bsonobject bson eco shared view view view system collections generic list notifications in at eco shared view viewmanager applyproperties eco shared view view view eco shared serialization bsonobject bson system collections generic list notifications in at eco shared view viewmanager receiveviewupdate eco shared serialization bsonarray updates in at clientpackethandler receivewhile eco shared networking networkclient client system func condition in at system action invoke t obj in at frameplanner plannergroup onupdate in at frameplanner frameplannersystem onupdate in at unity entities componentsystem internalupdate in at unity entities componentsystemgroup onupdate in at unity entities componentsystem internalupdate in at unityengine experimental lowlevel playerloopsystem updatefunction invoke in rethrow as targetinvocationexception exception has been thrown by the target of an invocation at system reflection monomethod invoke system object obj system reflection bindingflags invokeattr system reflection binder binder system object parameters system globalization cultureinfo culture in at system reflection methodbase invoke system object obj system object parameters in at eco shared serialization bsonmanipulator frombsonilist system type type eco shared serialization bsonvalue value in at eco shared serialization bsonmanipulator frombsonidictionary system type type eco shared serialization bsonobject obj in at eco shared serialization bsonmanipulator frombson eco shared serialization bsonvalue value system type type in at eco shared view view updateproperty eco shared serialization bsonvalue propnameid eco shared serialization bsonobject bson in at eco shared view viewmanager viewnotify eco shared serialization bsonobject bson eco shared view view view system collections generic list notifications in at eco shared view viewmanager applyproperties eco shared view view view eco shared serialization bsonobject bson system collections generic list notifications in at eco shared view viewmanager receiveviewupdate eco shared serialization bsonarray updates in at clientpackethandler receivewhile eco shared networking networkclient client system func condition in at system action invoke t obj in at frameplanner plannergroup onupdate in at frameplanner frameplannersystem onupdate in at unity entities componentsystem internalupdate in at unity entities componentsystemgroup onupdate in at unity entities componentsystem internalupdate in at unityengine experimental lowlevel playerloopsystem updatefunction invoke in unityengine logger logexception exception object unityengine debug logexception exception frameplanner plannergroup onupdate frameplanner frameplannersystem onupdate unity entities componentsystem internalupdate unity entities componentsystemgroup onupdate unity entities componentsystem internalupdate unityengine experimental lowlevel updatefunction invoke then i can t connect to server i need to restart game | 1 |
640,101 | 20,773,477,543 | IssuesEvent | 2022-03-16 08:13:52 | wso2/product-apim-tooling | https://api.github.com/repos/wso2/product-apim-tooling | closed | apictl certificate has expired | Type/Bug Priority/Highest Affected/4.0.1 | $_apictl import api --file ./PetstoreAPI --environment dev_
Login to APIM in dev
Username:admin
Password:
apictl: Error in connecting. Reason: Post "https://localhost:9443/client-registration/v0.17/register": x509: certificate has expired or is not yet valid: current time 2022-01-26T23:07:52+05:30 is after 2022-01-25T07:30:43Z
Exit status 1
How to get rid of this error? Using apictl-4.0.1 and wso2am-4.0.0 | 1.0 | apictl certificate has expired - $_apictl import api --file ./PetstoreAPI --environment dev_
Login to APIM in dev
Username:admin
Password:
apictl: Error in connecting. Reason: Post "https://localhost:9443/client-registration/v0.17/register": x509: certificate has expired or is not yet valid: current time 2022-01-26T23:07:52+05:30 is after 2022-01-25T07:30:43Z
Exit status 1
How to get rid of this error? Using apictl-4.0.1 and wso2am-4.0.0 | priority | apictl certificate has expired apictl import api file petstoreapi environment dev login to apim in dev username admin password apictl error in connecting reason post certificate has expired or is not yet valid current time is after exit status how to get rid of this error using apictl and | 1 |
279,391 | 8,664,782,218 | IssuesEvent | 2018-11-28 21:13:35 | CzolgIT/servertanksgame | https://api.github.com/repos/CzolgIT/servertanksgame | closed | Receiving and sending UDP data (server side) | enhancement priority: highest | As described above, allow server to send data about the current state of the player. | 1.0 | Receiving and sending UDP data (server side) - As described above, allow server to send data about the current state of the player. | priority | receiving and sending udp data server side as described above allow server to send data about the current state of the player | 1 |
145,509 | 5,576,964,436 | IssuesEvent | 2017-03-28 08:27:01 | onaio/onadata | https://api.github.com/repos/onaio/onadata | closed | BacklogLimitExceeded on csv import | Error Priority: High | ```python
Internal Server Error: /api/v1/forms/[pk]/csv_import.json
BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json
ea3c17c4-373e-446f-9e29-e5af265f9641
Request Method: GET
Request URL: https://api.ona.io/api/v1/forms/[pk]/csv_import.json?job_uuid=ea3c17c4-373e-446f-9e29-e5af265f9641
Django Version: 1.9.5
Python Executable: /..../.virtualenvs/onadata/bin/uwsgi
Python Version: 2.7.6
Python Path: ['/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter', '/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter/cWriterow', '.', '', '/..../.virtualenvs/onadata/src/django-nose', '/..../.virtualenvs/onadata/src/django-digest', '/..../.virtualenvs/onadata/src/j2xclient', '/..../.virtualenvs/onadata/src/cors-headers', '/..../.virtualenvs/onadata/src/pricing', '/..../.virtualenvs/onadata/src/savreaderwriter', '/..../.virtualenvs/onadata/src/google-export', '/..../.virtualenvs/onadata/lib/python2.7', '/..../.virtualenvs/onadata/lib/python2.7/plat-x86_64-linux-gnu', '/..../.virtualenvs/onadata/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/lib/python2.7/lib-old', '/..../.virtualenvs/onadata/lib/python2.7/lib-dynload', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/local/lib/python2.7/site-packages']
Server time: Mon, 6 Feb 2017 09:55:35 -0500
Installed Applications:
('google_export',
'pricing',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.sites',
'django.contrib.messages',
'django.contrib.staticfiles',
'django.contrib.humanize',
'django.contrib.admin',
'django.contrib.admindocs',
'django.contrib.gis',
'registration',
'django_nose',
'django_digest',
'corsheaders',
'oauth2_provider',
'rest_framework',
'rest_framework.authtoken',
'taggit',
'readonly',
'onadata.apps.logger',
'onadata.apps.viewer',
'onadata.apps.main',
'onadata.apps.restservice',
'onadata.apps.api',
'guardian',
'djcelery',
'onadata.apps.sms_support',
'onadata.libs',
'reversion')
Installed Middleware:
('multidb.middleware.PinningRouterMiddleware',
'onadata.libs.profiling.sql.SqlTimingMiddleware',
'django.middleware.http.ConditionalGetMiddleware',
'django.middleware.cache.UpdateCacheMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.cache.FetchFromCacheMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'onadata.libs.utils.middleware.LocaleMiddlewareWithTweaks',
'django.middleware.csrf.CsrfViewMiddleware',
'corsheaders.middleware.CorsMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'onadata.libs.utils.middleware.HTTPResponseNotAllowedMiddleware',
'readonly.middleware.DatabaseReadOnlyMiddleware')
Traceback:
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response
149. response = self.process_exception_by_middleware(e, request)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response
147. response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/views/decorators/csrf.py" in wrapped_view
58. return view_func(*args, **kwargs)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/viewsets.py" in view
87. return self.dispatch(request, *args, **kwargs)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch
466. response = self.handle_exception(exc)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch
463. response = handler(request, *args, **kwargs)
File "./onadata/apps/api/viewsets/xform_viewset.py" in csv_import
549. request.query_params.get('job_uuid')))
File "./onadata/libs/utils/csv_import.py" in get_async_csv_submission_status
303. result = (job.result or job.state)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in result
356. return self._get_task_meta()['result']
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in _get_task_meta
339. return self._maybe_set_cache(self.backend.get_task_meta(self.id))
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/backends/amqp.py" in get_task_meta
180. raise self.BacklogLimitExceeded(task_id)
Exception Type: BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json
Exception Value: ea3c17c4-373e-446f-9e29-e5af265f9641
Request information:
GET:
job_uuid = u'ea3c17c4-373e-446f-9e29-e5af265f9641'
POST: No POST data
FILES: No FILES data
COOKIES: No cookie data
``` | 1.0 | BacklogLimitExceeded on csv import - ```python
Internal Server Error: /api/v1/forms/[pk]/csv_import.json
BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json
ea3c17c4-373e-446f-9e29-e5af265f9641
Request Method: GET
Request URL: https://api.ona.io/api/v1/forms/[pk]/csv_import.json?job_uuid=ea3c17c4-373e-446f-9e29-e5af265f9641
Django Version: 1.9.5
Python Executable: /..../.virtualenvs/onadata/bin/uwsgi
Python Version: 2.7.6
Python Path: ['/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter', '/..../.virtualenvs/onadata/src/savreaderwriter/savReaderWriter/cWriterow', '.', '', '/..../.virtualenvs/onadata/src/django-nose', '/..../.virtualenvs/onadata/src/django-digest', '/..../.virtualenvs/onadata/src/j2xclient', '/..../.virtualenvs/onadata/src/cors-headers', '/..../.virtualenvs/onadata/src/pricing', '/..../.virtualenvs/onadata/src/savreaderwriter', '/..../.virtualenvs/onadata/src/google-export', '/..../.virtualenvs/onadata/lib/python2.7', '/..../.virtualenvs/onadata/lib/python2.7/plat-x86_64-linux-gnu', '/..../.virtualenvs/onadata/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/lib/python2.7/lib-old', '/..../.virtualenvs/onadata/lib/python2.7/lib-dynload', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/..../.virtualenvs/onadata/local/lib/python2.7/site-packages']
Server time: Mon, 6 Feb 2017 09:55:35 -0500
Installed Applications:
('google_export',
'pricing',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.sites',
'django.contrib.messages',
'django.contrib.staticfiles',
'django.contrib.humanize',
'django.contrib.admin',
'django.contrib.admindocs',
'django.contrib.gis',
'registration',
'django_nose',
'django_digest',
'corsheaders',
'oauth2_provider',
'rest_framework',
'rest_framework.authtoken',
'taggit',
'readonly',
'onadata.apps.logger',
'onadata.apps.viewer',
'onadata.apps.main',
'onadata.apps.restservice',
'onadata.apps.api',
'guardian',
'djcelery',
'onadata.apps.sms_support',
'onadata.libs',
'reversion')
Installed Middleware:
('multidb.middleware.PinningRouterMiddleware',
'onadata.libs.profiling.sql.SqlTimingMiddleware',
'django.middleware.http.ConditionalGetMiddleware',
'django.middleware.cache.UpdateCacheMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.cache.FetchFromCacheMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'onadata.libs.utils.middleware.LocaleMiddlewareWithTweaks',
'django.middleware.csrf.CsrfViewMiddleware',
'corsheaders.middleware.CorsMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'onadata.libs.utils.middleware.HTTPResponseNotAllowedMiddleware',
'readonly.middleware.DatabaseReadOnlyMiddleware')
Traceback:
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response
149. response = self.process_exception_by_middleware(e, request)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/core/handlers/base.py" in get_response
147. response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/django/views/decorators/csrf.py" in wrapped_view
58. return view_func(*args, **kwargs)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/viewsets.py" in view
87. return self.dispatch(request, *args, **kwargs)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch
466. response = self.handle_exception(exc)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/rest_framework/views.py" in dispatch
463. response = handler(request, *args, **kwargs)
File "./onadata/apps/api/viewsets/xform_viewset.py" in csv_import
549. request.query_params.get('job_uuid')))
File "./onadata/libs/utils/csv_import.py" in get_async_csv_submission_status
303. result = (job.result or job.state)
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in result
356. return self._get_task_meta()['result']
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/result.py" in _get_task_meta
339. return self._maybe_set_cache(self.backend.get_task_meta(self.id))
File "/..../.virtualenvs/onadata/local/lib/python2.7/site-packages/celery/backends/amqp.py" in get_task_meta
180. raise self.BacklogLimitExceeded(task_id)
Exception Type: BacklogLimitExceeded at /api/v1/forms/[pk]/csv_import.json
Exception Value: ea3c17c4-373e-446f-9e29-e5af265f9641
Request information:
GET:
job_uuid = u'ea3c17c4-373e-446f-9e29-e5af265f9641'
POST: No POST data
FILES: No FILES data
COOKIES: No cookie data
``` | priority | backloglimitexceeded on csv import python internal server error api forms csv import json backloglimitexceeded at api forms csv import json request method get request url csv import json job uuid django version python executable virtualenvs onadata bin uwsgi python version python path server time mon feb installed applications google export pricing django contrib auth django contrib contenttypes django contrib sessions django contrib sites django contrib messages django contrib staticfiles django contrib humanize django contrib admin django contrib admindocs django contrib gis registration django nose django digest corsheaders provider rest framework rest framework authtoken taggit readonly onadata apps logger onadata apps viewer onadata apps main onadata apps restservice onadata apps api guardian djcelery onadata apps sms support onadata libs reversion installed middleware multidb middleware pinningroutermiddleware onadata libs profiling sql sqltimingmiddleware django middleware http conditionalgetmiddleware django middleware cache updatecachemiddleware django middleware common commonmiddleware django middleware cache fetchfromcachemiddleware django contrib sessions middleware sessionmiddleware onadata libs utils middleware localemiddlewarewithtweaks django middleware csrf csrfviewmiddleware corsheaders middleware corsmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware onadata libs utils middleware httpresponsenotallowedmiddleware readonly middleware databasereadonlymiddleware traceback file virtualenvs onadata local lib site packages django core handlers base py in get response response self process exception by middleware e request file virtualenvs onadata local lib site packages django core handlers base py in get response response wrapped callback request callback args callback kwargs file virtualenvs onadata local lib site packages django views decorators csrf py in wrapped view return view func args kwargs file virtualenvs onadata local lib site packages rest framework viewsets py in view return self dispatch request args kwargs file virtualenvs onadata local lib site packages rest framework views py in dispatch response self handle exception exc file virtualenvs onadata local lib site packages rest framework views py in dispatch response handler request args kwargs file onadata apps api viewsets xform viewset py in csv import request query params get job uuid file onadata libs utils csv import py in get async csv submission status result job result or job state file virtualenvs onadata local lib site packages celery result py in result return self get task meta file virtualenvs onadata local lib site packages celery result py in get task meta return self maybe set cache self backend get task meta self id file virtualenvs onadata local lib site packages celery backends amqp py in get task meta raise self backloglimitexceeded task id exception type backloglimitexceeded at api forms csv import json exception value request information get job uuid u post no post data files no files data cookies no cookie data | 1 |
735,980 | 25,451,660,447 | IssuesEvent | 2022-11-24 10:55:50 | owncloud/web | https://api.github.com/repos/owncloud/web | closed | Recipients disappear after re-sharing to the group | Type:Bug Priority:p2-high | ownCloud Web UI 6.0.0-rc.6
Infinite Scale 2.0.0-rc.1+736f513cf Community -local
Steps:
- admin creates file and shares to marie
- marie accepts file and re-shares it to the users and the group
Actual: after re-sharing to the group again, all recipients or public links disappear, but still remain after reloading the page
https://user-images.githubusercontent.com/84779829/201134209-e8df26b8-2a62-4616-a90d-ac0aba9ae550.mov
| 1.0 | Recipients disappear after re-sharing to the group - ownCloud Web UI 6.0.0-rc.6
Infinite Scale 2.0.0-rc.1+736f513cf Community -local
Steps:
- admin creates file and shares to marie
- marie accepts file and re-shares it to the users and the group
Actual: after re-sharing to the group again, all recipients or public links disappear, but still remain after reloading the page
https://user-images.githubusercontent.com/84779829/201134209-e8df26b8-2a62-4616-a90d-ac0aba9ae550.mov
| priority | recipients disappear after re sharing to the group owncloud web ui rc infinite scale rc community local steps admin creates file and shares to marie marie accepts file and re shares it to the users and the group actual after re sharing to the group again all recipients or public links disappear but still remain after reloading the page | 1 |
73,348 | 3,410,993,068 | IssuesEvent | 2015-12-04 22:58:34 | rurban/perl-compiler | https://api.github.com/repos/rurban/perl-compiler | opened | POSIX: Constant subroutine MB_CUR_MAX redefined warning | Priority-High | This is similar to issue #335, this is happening in many binaries
```
> perlcc -e 'use POSIX; print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")'
pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:15052: warning: integer constant is too large for ‘unsigned long’ type
pccOZTKF.c:15053: warning: integer constant is too large for ‘unsigned long’ type
Constant subroutine MB_CUR_MAX redefined at ./a.out line 4294967295.
ok
```
Notice that if ENV is not reset then we have no warnings...
```
> perl -e 'exec("./a.out")'
ok
> ./a.out
ok
```
one workaround is to disable POSIX import
{code}
> perlcc -e 'use POSIX (); print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")'
ok
{code} | 1.0 | POSIX: Constant subroutine MB_CUR_MAX redefined warning - This is similar to issue #335, this is happening in many binaries
```
> perlcc -e 'use POSIX; print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")'
pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14025: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14026: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:14027: warning: this decimal constant is unsigned only in ISO C90
pccOZTKF.c:15052: warning: integer constant is too large for ‘unsigned long’ type
pccOZTKF.c:15053: warning: integer constant is too large for ‘unsigned long’ type
Constant subroutine MB_CUR_MAX redefined at ./a.out line 4294967295.
ok
```
Notice that if ENV is not reset then we have no warnings...
```
> perl -e 'exec("./a.out")'
ok
> ./a.out
ok
```
one workaround is to disable POSIX import
{code}
> perlcc -e 'use POSIX (); print qq/ok\n/' && perl -e 'local %ENV; exec("./a.out")'
ok
{code} | priority | posix constant subroutine mb cur max redefined warning this is similar to issue this is happening in many binaries perlcc e use posix print qq ok n perl e local env exec a out pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning this decimal constant is unsigned only in iso pccoztkf c warning integer constant is too large for ‘unsigned long’ type pccoztkf c warning integer constant is too large for ‘unsigned long’ type constant subroutine mb cur max redefined at a out line ok notice that if env is not reset then we have no warnings perl e exec a out ok a out ok one workaround is to disable posix import code perlcc e use posix print qq ok n perl e local env exec a out ok code | 1 |
212,529 | 7,238,203,450 | IssuesEvent | 2018-02-13 13:53:41 | SELinuxProject/selinux-kernel | https://api.github.com/repos/SELinuxProject/selinux-kernel | opened | kernel softlockup due to too many SIDs/contexts | bug priority/high | As reported by yangjhong1 on selinux list, when too many SIDs/contexts have been allocated (e.g. 300000+ as a result of repeated docker container creations for 2 days), sidtab_search_context becomes very slow and can cause a kernel softlockup warning.
docker randomly selects a category pair for every container creation, so this can occur just from creating containers over time, even if old containers are removed promptly (category set reuse for removed containers will eventually occur but each selection is random). It can also occur from any other activity that allocates SIDs/contexts, even those that simply probe for context validity.
sidtab_search_context() is a reverse lookup in the sidtab and presently just walks the entire hash table.
At a minimum, we need to add a reverse hash table to help mitigate this, possibly using a SELinux hashtab or the core kernel's hashtable.h or rhashtable.h data structures. We might also want a fast check of the context category set to see if it has ever been previously used (i.e. maintain a ebitmap of used categories, and check whether it contains the context's category set) so that we can fail fast on a lookup of a new category set. However, the fact that we might need to support 300000+ SIDs/contexts also suggests that we should likely revisit the sidtab forward hash table since it is too small to efficiently handle that. That too is a candidate to be replaced by e.g. hashtable or rhashtable.
I can see both short term and long term fixes for this bug; short term might just be adding simple reverse hash table and perhaps a category ebitmap test; longer term might be reworking the forward hash and switching over to hashtable or rhashtable structures. | 1.0 | kernel softlockup due to too many SIDs/contexts - As reported by yangjhong1 on selinux list, when too many SIDs/contexts have been allocated (e.g. 300000+ as a result of repeated docker container creations for 2 days), sidtab_search_context becomes very slow and can cause a kernel softlockup warning.
docker randomly selects a category pair for every container creation, so this can occur just from creating containers over time, even if old containers are removed promptly (category set reuse for removed containers will eventually occur but each selection is random). It can also occur from any other activity that allocates SIDs/contexts, even those that simply probe for context validity.
sidtab_search_context() is a reverse lookup in the sidtab and presently just walks the entire hash table.
At a minimum, we need to add a reverse hash table to help mitigate this, possibly using a SELinux hashtab or the core kernel's hashtable.h or rhashtable.h data structures. We might also want a fast check of the context category set to see if it has ever been previously used (i.e. maintain a ebitmap of used categories, and check whether it contains the context's category set) so that we can fail fast on a lookup of a new category set. However, the fact that we might need to support 300000+ SIDs/contexts also suggests that we should likely revisit the sidtab forward hash table since it is too small to efficiently handle that. That too is a candidate to be replaced by e.g. hashtable or rhashtable.
I can see both short term and long term fixes for this bug; short term might just be adding simple reverse hash table and perhaps a category ebitmap test; longer term might be reworking the forward hash and switching over to hashtable or rhashtable structures. | priority | kernel softlockup due to too many sids contexts as reported by on selinux list when too many sids contexts have been allocated e g as a result of repeated docker container creations for days sidtab search context becomes very slow and can cause a kernel softlockup warning docker randomly selects a category pair for every container creation so this can occur just from creating containers over time even if old containers are removed promptly category set reuse for removed containers will eventually occur but each selection is random it can also occur from any other activity that allocates sids contexts even those that simply probe for context validity sidtab search context is a reverse lookup in the sidtab and presently just walks the entire hash table at a minimum we need to add a reverse hash table to help mitigate this possibly using a selinux hashtab or the core kernel s hashtable h or rhashtable h data structures we might also want a fast check of the context category set to see if it has ever been previously used i e maintain a ebitmap of used categories and check whether it contains the context s category set so that we can fail fast on a lookup of a new category set however the fact that we might need to support sids contexts also suggests that we should likely revisit the sidtab forward hash table since it is too small to efficiently handle that that too is a candidate to be replaced by e g hashtable or rhashtable i can see both short term and long term fixes for this bug short term might just be adding simple reverse hash table and perhaps a category ebitmap test longer term might be reworking the forward hash and switching over to hashtable or rhashtable structures | 1 |
629,046 | 20,021,983,715 | IssuesEvent | 2022-02-01 17:12:18 | dtcenter/METplus | https://api.github.com/repos/dtcenter/METplus | closed | Implement various enhancements to climatology settings | type: enhancement component: use case wrapper priority: high requestor: NOAA/EMC alert: NEED ACCOUNT KEY required: FOR OFFICIAL RELEASE METplus: Configuration | Various enhancements are needed to improve the logic for setting climatology variables in the climo_mean and climo_stdev dictionaries.
## Describe the Enhancements ##
- [x] Add support for setting a list of field levels in climo variables
Setting the following:
`GRID_STAT_CLIMO_MEAN_FIELD = {name="UGRD"; level=["P850","P500","P250"];}`
Results in an improperly formatted value:
METPLUS_CLIMO_MEAN_DICT=climo_mean = {file_name = ["/scratch1/NCEPDEV/global/glopara/fix/fix_verif/climo_fi
les/cmean_1d.19591203"];field = [{name="UGRD"; level=["P850",**P500,P250**];}];}
Add a test for setting climo field info and enhance the logic to properly format a list of levels. Make sure the behavior works for all apps that use climatology variables.
- [x] Add support for using forecast or observation field info for climatology
It is often the case that the climatology field info is the same as the forecast field info. If \<APP\>\_CLIMO_[MEAN/STDEV]_FIELD_NAME is not set, then the wrapper should check if \<APP\>\_CLIMO\_[MEAN/STDEV]\_USE\_[FCST/OBS] is set to True. Error out if both are True. If one is True, then use the field info from the appropriate input as the field info for the climo variable
- [x] Add support for setting file_type for climo_[mean/stdev]
Add optional \<APP\>\_CLIMO_[MEAN/STDEV]_FILE_TYPE to set file_type in the appropriate climo dictionary -- this was implemented in PR #1353
### Time Estimate ###
~1 day
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [X] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Enhancement Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | Implement various enhancements to climatology settings - Various enhancements are needed to improve the logic for setting climatology variables in the climo_mean and climo_stdev dictionaries.
## Describe the Enhancements ##
- [x] Add support for setting a list of field levels in climo variables
Setting the following:
`GRID_STAT_CLIMO_MEAN_FIELD = {name="UGRD"; level=["P850","P500","P250"];}`
Results in an improperly formatted value:
METPLUS_CLIMO_MEAN_DICT=climo_mean = {file_name = ["/scratch1/NCEPDEV/global/glopara/fix/fix_verif/climo_fi
les/cmean_1d.19591203"];field = [{name="UGRD"; level=["P850",**P500,P250**];}];}
Add a test for setting climo field info and enhance the logic to properly format a list of levels. Make sure the behavior works for all apps that use climatology variables.
- [x] Add support for using forecast or observation field info for climatology
It is often the case that the climatology field info is the same as the forecast field info. If \<APP\>\_CLIMO_[MEAN/STDEV]_FIELD_NAME is not set, then the wrapper should check if \<APP\>\_CLIMO\_[MEAN/STDEV]\_USE\_[FCST/OBS] is set to True. Error out if both are True. If one is True, then use the field info from the appropriate input as the field info for the climo variable
- [x] Add support for setting file_type for climo_[mean/stdev]
Add optional \<APP\>\_CLIMO_[MEAN/STDEV]_FILE_TYPE to set file_type in the appropriate climo dictionary -- this was implemented in PR #1353
### Time Estimate ###
~1 day
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [X] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
*Define the source of funding and account keys here or state NONE.*
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Enhancement Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | implement various enhancements to climatology settings various enhancements are needed to improve the logic for setting climatology variables in the climo mean and climo stdev dictionaries describe the enhancements add support for setting a list of field levels in climo variables setting the following grid stat climo mean field name ugrd level results in an improperly formatted value metplus climo mean dict climo mean file name ncepdev global glopara fix fix verif climo fi les cmean field add a test for setting climo field info and enhance the logic to properly format a list of levels make sure the behavior works for all apps that use climatology variables add support for using forecast or observation field info for climatology it is often the case that the climatology field info is the same as the forecast field info if climo field name is not set then the wrapper should check if climo use is set to true error out if both are true if one is true then use the field info from the appropriate input as the field info for the climo variable add support for setting file type for climo add optional climo file type to set file type in the appropriate climo dictionary this was implemented in pr time estimate day sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the source of funding and account keys here or state none define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
733,938 | 25,330,966,316 | IssuesEvent | 2022-11-18 13:13:35 | csmith1188/formbar | https://api.github.com/repos/csmith1188/formbar | opened | Major Bug TUTD | bug high priority frontend backend large | Sometimes, a student will cause the formbar alternate adding or removing a selected "thumb" from a student every second. Each time it plays the sound for adding a thumb (even when it is removing the thumb) and updates the formbar. The permissions of the student makes no difference, so they cannot be blocked by being made a guest. This seems to coincide with the logpoint that logs the students while in the /controlpanel. May be related to new websocket features. | 1.0 | Major Bug TUTD - Sometimes, a student will cause the formbar alternate adding or removing a selected "thumb" from a student every second. Each time it plays the sound for adding a thumb (even when it is removing the thumb) and updates the formbar. The permissions of the student makes no difference, so they cannot be blocked by being made a guest. This seems to coincide with the logpoint that logs the students while in the /controlpanel. May be related to new websocket features. | priority | major bug tutd sometimes a student will cause the formbar alternate adding or removing a selected thumb from a student every second each time it plays the sound for adding a thumb even when it is removing the thumb and updates the formbar the permissions of the student makes no difference so they cannot be blocked by being made a guest this seems to coincide with the logpoint that logs the students while in the controlpanel may be related to new websocket features | 1 |
601,030 | 18,363,895,785 | IssuesEvent | 2021-10-09 18:09:50 | alec-roberson/MS-CV2 | https://api.github.com/repos/alec-roberson/MS-CV2 | closed | Get the training process working on the server | bug good first issue high priority | Being a different machine, the server has inevitably run into various problems and bugs when trying to execute the training code for our networks. | 1.0 | Get the training process working on the server - Being a different machine, the server has inevitably run into various problems and bugs when trying to execute the training code for our networks. | priority | get the training process working on the server being a different machine the server has inevitably run into various problems and bugs when trying to execute the training code for our networks | 1 |
681,648 | 23,319,543,446 | IssuesEvent | 2022-08-08 15:11:29 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Performance numbers decrease due to extra DB queries executing in IS 6.0.0 | Priority/Highest Severity/Blocker bug QA-Reported product-performance | **Describe the issue:**
Following bold and italic extra DB queries can be observed at the IS 6.0.0 compared to 5.11.0-GA. Currently this is identified as the major reason for performance decrease compared to previous version.
IS 6.0.0-alpha2 - "CaseInsensitiveUernames" property set to false
- SELECT * FROM IDN_RECOVERY_DATA WHERE LOWER(USER_NAME)=LOWER(?) AND USER_DOMAIN = ? AND TENANT_ID = ? AND SCENARIO NOT IN ('EMAIL_VERIFICATION_ON_UPDATE', 'MOBILE_VERIFICATION_ON_UPDATE')
- _**SELECT WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME,WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID FROM WF_REQUEST,WF_REQUEST_ENTITY_RELATIONSHIP WHERE WF_REQUEST.UUID = WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME = ? AND WF_REQUEST.STATUS = ? AND WF_REQUEST.OPERATION_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.TENANT_ID = ?**_
- SELECT UM_USER_ID, UM_USER_NAME, UM_USER_PASSWORD, UM_SALT_VALUE, UM_REQUIRE_CHANGE, UM_CHANGED_TIME FROM UM_USER WHERE LOWER(UM_USER_NAME)=LOWER(?) AND UM_TENANT_ID=?
- _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_
- _**SELECT DATA_KEY, DATA_VALUE FROM IDN_IDENTITY_USER_DATA WHERE TENANT_ID = ? AND LOWER(USER_NAME) = LOWER(?)**_
- _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_
[1] https://docs.google.com/spreadsheets/d/1I79C69KFB6lFypLh7esmKxeYMJRl5eQQjd1LyFZweRI/edit?usp=sharing - tab "Comparison-Summery-Detail"
**How to reproduce:**
1. Enable correlation logs in IS 6.0.0 and IS 5.11.0-GA product versions - https://is.docs.wso2.com/en/latest/setup/working-with-product-observability/?query=CaseInsensitiveUernames
2. Execute performance test "User authentication for the super tenant" - https://github.com/aaujayasena/performance-is/blob/master/common/jmeter/authenticate/Authenticate_Super_Tenant_User.jmx
3. Compare performance numbers between product versions
4. Compare DB queries filtering correlation ID
**Expected behavior:**
Performance numbers of IS 6.0.0-alpha2 >= 5.11.0-GA
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: IS 6.0.0-alpha2 and IS 5.11.0-GA
- OS: MAC for local testing and ubuntu for aws environment
- Database: MSSQL
- Userstore: JDBC
---
### Optional Fields
**Related issues:**
https://github.com/wso2/product-is/issues/14372
**Suggested labels:**
<!-- Only to be used by non-members -->
| 1.0 | Performance numbers decrease due to extra DB queries executing in IS 6.0.0 - **Describe the issue:**
Following bold and italic extra DB queries can be observed at the IS 6.0.0 compared to 5.11.0-GA. Currently this is identified as the major reason for performance decrease compared to previous version.
IS 6.0.0-alpha2 - "CaseInsensitiveUernames" property set to false
- SELECT * FROM IDN_RECOVERY_DATA WHERE LOWER(USER_NAME)=LOWER(?) AND USER_DOMAIN = ? AND TENANT_ID = ? AND SCENARIO NOT IN ('EMAIL_VERIFICATION_ON_UPDATE', 'MOBILE_VERIFICATION_ON_UPDATE')
- _**SELECT WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME,WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID FROM WF_REQUEST,WF_REQUEST_ENTITY_RELATIONSHIP WHERE WF_REQUEST.UUID = WF_REQUEST_ENTITY_RELATIONSHIP.REQUEST_ID AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.ENTITY_NAME = ? AND WF_REQUEST.STATUS = ? AND WF_REQUEST.OPERATION_TYPE = ? AND WF_REQUEST_ENTITY_RELATIONSHIP.TENANT_ID = ?**_
- SELECT UM_USER_ID, UM_USER_NAME, UM_USER_PASSWORD, UM_SALT_VALUE, UM_REQUIRE_CHANGE, UM_CHANGED_TIME FROM UM_USER WHERE LOWER(UM_USER_NAME)=LOWER(?) AND UM_TENANT_ID=?
- _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_
- _**SELECT DATA_KEY, DATA_VALUE FROM IDN_IDENTITY_USER_DATA WHERE TENANT_ID = ? AND LOWER(USER_NAME) = LOWER(?)**_
- _**SELECT UM_ATTR_NAME, UM_ATTR_VALUE FROM UM_USER_ATTRIBUTE, UM_USER WHERE UM_USER.UM_ID = UM_USER_ATTRIBUTE.UM_USER_ID AND UM_USER.UM_USER_ID=? AND UM_PROFILE_ID=? AND UM_USER_ATTRIBUTE.UM_TENANT_ID=? AND UM_USER.UM_TENANT_ID=?**_
[1] https://docs.google.com/spreadsheets/d/1I79C69KFB6lFypLh7esmKxeYMJRl5eQQjd1LyFZweRI/edit?usp=sharing - tab "Comparison-Summery-Detail"
**How to reproduce:**
1. Enable correlation logs in IS 6.0.0 and IS 5.11.0-GA product versions - https://is.docs.wso2.com/en/latest/setup/working-with-product-observability/?query=CaseInsensitiveUernames
2. Execute performance test "User authentication for the super tenant" - https://github.com/aaujayasena/performance-is/blob/master/common/jmeter/authenticate/Authenticate_Super_Tenant_User.jmx
3. Compare performance numbers between product versions
4. Compare DB queries filtering correlation ID
**Expected behavior:**
Performance numbers of IS 6.0.0-alpha2 >= 5.11.0-GA
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: IS 6.0.0-alpha2 and IS 5.11.0-GA
- OS: MAC for local testing and ubuntu for aws environment
- Database: MSSQL
- Userstore: JDBC
---
### Optional Fields
**Related issues:**
https://github.com/wso2/product-is/issues/14372
**Suggested labels:**
<!-- Only to be used by non-members -->
| priority | performance numbers decrease due to extra db queries executing in is describe the issue following bold and italic extra db queries can be observed at the is compared to ga currently this is identified as the major reason for performance decrease compared to previous version is caseinsensitiveuernames property set to false select from idn recovery data where lower user name lower and user domain and tenant id and scenario not in email verification on update mobile verification on update select wf request entity relationship entity name wf request entity relationship request id from wf request wf request entity relationship where wf request uuid wf request entity relationship request id and wf request entity relationship entity type and wf request entity relationship entity name and wf request status and wf request operation type and wf request entity relationship tenant id select um user id um user name um user password um salt value um require change um changed time from um user where lower um user name lower and um tenant id select um attr name um attr value from um user attribute um user where um user um id um user attribute um user id and um user um user id and um profile id and um user attribute um tenant id and um user um tenant id select data key data value from idn identity user data where tenant id and lower user name lower select um attr name um attr value from um user attribute um user where um user um id um user attribute um user id and um user um user id and um profile id and um user attribute um tenant id and um user um tenant id tab comparison summery detail how to reproduce enable correlation logs in is and is ga product versions execute performance test user authentication for the super tenant compare performance numbers between product versions compare db queries filtering correlation id expected behavior performance numbers of is ga environment information please complete the following information remove any unnecessary fields product version is and is ga os mac for local testing and ubuntu for aws environment database mssql userstore jdbc optional fields related issues suggested labels | 1 |
87,563 | 3,755,754,261 | IssuesEvent | 2016-03-12 21:46:58 | johnnyflowers/schoolcalendars | https://api.github.com/repos/johnnyflowers/schoolcalendars | closed | Google calendar URLs should be formatted consistently | high priority | We need to be sure that for every google calendar we host, the URL is constructed to include:
?showTitle=1&showNav=1&showDate=1
in the actual URL itself. Otherwise, users will not be able to browse into future months.
This is likely something you already did last year, but we have no way of knowing because of the lack of git, heh. | 1.0 | Google calendar URLs should be formatted consistently - We need to be sure that for every google calendar we host, the URL is constructed to include:
?showTitle=1&showNav=1&showDate=1
in the actual URL itself. Otherwise, users will not be able to browse into future months.
This is likely something you already did last year, but we have no way of knowing because of the lack of git, heh. | priority | google calendar urls should be formatted consistently we need to be sure that for every google calendar we host the url is constructed to include showtitle shownav showdate in the actual url itself otherwise users will not be able to browse into future months this is likely something you already did last year but we have no way of knowing because of the lack of git heh | 1 |
829,271 | 31,862,042,017 | IssuesEvent | 2023-09-15 11:40:14 | Automattic/woocommerce-payments | https://api.github.com/repos/Automattic/woocommerce-payments | opened | Enable Deferred Intent UPE for legacy UPE users | type: enhancement priority: high component: checkout category: projects | ### Description
Similar to https://github.com/Automattic/woocommerce-payments/issues/7003, but this time instead of newly onboarded merchants, legacy UPE merchants should have the deferred UPE feature flag enabled by default.
💡 The solution for this task **should be merged between Sep 19 and October 6,** so that it is released as part of WooPayments 6.6.0 | 1.0 | Enable Deferred Intent UPE for legacy UPE users - ### Description
Similar to https://github.com/Automattic/woocommerce-payments/issues/7003, but this time instead of newly onboarded merchants, legacy UPE merchants should have the deferred UPE feature flag enabled by default.
💡 The solution for this task **should be merged between Sep 19 and October 6,** so that it is released as part of WooPayments 6.6.0 | priority | enable deferred intent upe for legacy upe users description similar to but this time instead of newly onboarded merchants legacy upe merchants should have the deferred upe feature flag enabled by default 💡 the solution for this task should be merged between sep and october so that it is released as part of woopayments | 1 |
819,920 | 30,755,433,802 | IssuesEvent | 2023-07-29 02:04:40 | GC-spigot/AdvancedEnchantments | https://api.github.com/repos/GC-spigot/AdvancedEnchantments | closed | Bug on sacred white parchment | Priority: High Bug: Confirmed Status: In development | ### Describe the bug
The keep after death is in 'true', the scroll is removed and the item remains. The test I did resulted in the scroll still remaining on the item even after death.
### How to reproduce
Tests were carried out through death by drowning, still the item was not dropped but the holy white scroll was not removed from the item
### Screenshots / Videos
Test keep after death: false
https://drive.google.com/file/d/1-BOYeVFQepheGxEKvt7lakjlR-LsWG7G/view?usp=drivesdk
Test keep after death: true
https://drive.google.com/file/d/1-YYD5GXm23wlTf24NI-yqid_tKE_-On7/view?usp=drivesdk
### "/ae plinfo" link
https://paste.md-5.net/joharazodo
### Server Log
https://mcpaste.io/2015c6aaba30d058 | 1.0 | Bug on sacred white parchment - ### Describe the bug
The keep after death is in 'true', the scroll is removed and the item remains. The test I did resulted in the scroll still remaining on the item even after death.
### How to reproduce
Tests were carried out through death by drowning, still the item was not dropped but the holy white scroll was not removed from the item
### Screenshots / Videos
Test keep after death: false
https://drive.google.com/file/d/1-BOYeVFQepheGxEKvt7lakjlR-LsWG7G/view?usp=drivesdk
Test keep after death: true
https://drive.google.com/file/d/1-YYD5GXm23wlTf24NI-yqid_tKE_-On7/view?usp=drivesdk
### "/ae plinfo" link
https://paste.md-5.net/joharazodo
### Server Log
https://mcpaste.io/2015c6aaba30d058 | priority | bug on sacred white parchment describe the bug the keep after death is in true the scroll is removed and the item remains the test i did resulted in the scroll still remaining on the item even after death how to reproduce tests were carried out through death by drowning still the item was not dropped but the holy white scroll was not removed from the item screenshots videos test keep after death false test keep after death true ae plinfo link server log | 1 |
488,009 | 14,073,374,306 | IssuesEvent | 2020-11-04 04:38:07 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | SAML back channel SLO not working | Priority/Highest Severity/Critical bug | **Describe the issue:**
1. Configured SAML back channel SLO using the doc [1]
2. Once logout form one application other application doesn't get logged out.
3. Getting following errors in the console.
```
[2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo
[2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error
```
[1] https://is.docs.wso2.com/en/5.11.0/learn/handling-saml-single-logout-requests-from-federated-identity-providers/
**How to reproduce:**
1. Configured SAML back channel SLO using the doc [1]
2. Once logout form one application other application doesn't get logged out.
3. Getting following errors in the console.
```
[2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo
[2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error
```
**Expected behavior:**
First application should automatically logged out
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version:IS-5.11.0-beta2
- OS: Linux
- Database: POSTGRE
- Userstore: JDBC
---
| 1.0 | SAML back channel SLO not working - **Describe the issue:**
1. Configured SAML back channel SLO using the doc [1]
2. Once logout form one application other application doesn't get logged out.
3. Getting following errors in the console.
```
[2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo
[2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error
```
[1] https://is.docs.wso2.com/en/5.11.0/learn/handling-saml-single-logout-requests-from-federated-identity-providers/
**How to reproduce:**
1. Configured SAML back channel SLO using the doc [1]
2. Once logout form one application other application doesn't get logged out.
3. Getting following errors in the console.
```
[2020-10-21 11:16:55,811] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] INFO {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Sending single log out request again with retry count 1 after waiting for 60000 milli seconds to https://localhost:9443/identity/saml/slo
[2020-10-21 11:16:55,818] [24186cc2-8c3c-4ad9-966a-c7ea318b08ad] WARN {org.wso2.carbon.identity.sso.saml.logout.LogoutRequestSender} - Failed single logout response from https://localhost:9443/identity/saml/slo with status code Internal Server Error
```
**Expected behavior:**
First application should automatically logged out
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version:IS-5.11.0-beta2
- OS: Linux
- Database: POSTGRE
- Userstore: JDBC
---
| priority | saml back channel slo not working describe the issue configured saml back channel slo using the doc once logout form one application other application doesn t get logged out getting following errors in the console info org carbon identity sso saml logout logoutrequestsender sending single log out request again with retry count after waiting for milli seconds to warn org carbon identity sso saml logout logoutrequestsender failed single logout response from with status code internal server error how to reproduce configured saml back channel slo using the doc once logout form one application other application doesn t get logged out getting following errors in the console info org carbon identity sso saml logout logoutrequestsender sending single log out request again with retry count after waiting for milli seconds to warn org carbon identity sso saml logout logoutrequestsender failed single logout response from with status code internal server error expected behavior first application should automatically logged out environment information please complete the following information remove any unnecessary fields product version is os linux database postgre userstore jdbc | 1 |
296,766 | 9,125,860,252 | IssuesEvent | 2019-02-24 17:06:39 | pixijs/pixi.js | https://api.github.com/repos/pixijs/pixi.js | closed | TilingSprite ignoring realtime texture updates with Canvas renderer. | Difficulty: Easy Domain: API Plugin: Tiling Sprite Priority: High Renderer: Canvas Resolution: Won't Fix Status: Accepting PRs Type: Bug Version: v4.x | So I'm using [PIXI.extras.TilingSprite](https://pixijs.github.io/docs/PIXI.extras.TilingSprite.html) with the new 4.0.0-RC2. I updated to this version after seeing this bug with 3.0.11 while using PIXI.js within [nwjs](http://nwjs.io/), which is pretty much Chromium, in both OS X 10.11 and Ubuntu 16.04.
I'm using TilingSprites to allow for sprite sheets, which works great if everything is on a single sprite sheet, but when changing sheets (walking, jumping, etc.) the following update seems to be ignored when using the Canvas renderer.
`spriteObj.texture = textureObj;`
The update works perfectly when using the WebGL renderer though. It seems to work when the object is first created and the initial texture is applied, but seems to stop working after the object is attached to a stage.
I'll work on getting more source code together for a better example, but there isn't a lot to do here to reproduce it: just do an assignment as above and see if it's caught in the next render call. Try that with both the WebGL and Canvas renderers. We're working with the WebGL renderer for the time being, but I'm not sure that our target platform will support this renderer (currently a debian os running nwjs).
| 1.0 | TilingSprite ignoring realtime texture updates with Canvas renderer. - So I'm using [PIXI.extras.TilingSprite](https://pixijs.github.io/docs/PIXI.extras.TilingSprite.html) with the new 4.0.0-RC2. I updated to this version after seeing this bug with 3.0.11 while using PIXI.js within [nwjs](http://nwjs.io/), which is pretty much Chromium, in both OS X 10.11 and Ubuntu 16.04.
I'm using TilingSprites to allow for sprite sheets, which works great if everything is on a single sprite sheet, but when changing sheets (walking, jumping, etc.) the following update seems to be ignored when using the Canvas renderer.
`spriteObj.texture = textureObj;`
The update works perfectly when using the WebGL renderer though. It seems to work when the object is first created and the initial texture is applied, but seems to stop working after the object is attached to a stage.
I'll work on getting more source code together for a better example, but there isn't a lot to do here to reproduce it: just do an assignment as above and see if it's caught in the next render call. Try that with both the WebGL and Canvas renderers. We're working with the WebGL renderer for the time being, but I'm not sure that our target platform will support this renderer (currently a debian os running nwjs).
| priority | tilingsprite ignoring realtime texture updates with canvas renderer so i m using with the new i updated to this version after seeing this bug with while using pixi js within which is pretty much chromium in both os x and ubuntu i m using tilingsprites to allow for sprite sheets which works great if everything is on a single sprite sheet but when changing sheets walking jumping etc the following update seems to be ignored when using the canvas renderer spriteobj texture textureobj the update works perfectly when using the webgl renderer though it seems to work when the object is first created and the initial texture is applied but seems to stop working after the object is attached to a stage i ll work on getting more source code together for a better example but there isn t a lot to do here to reproduce it just do an assignment as above and see if it s caught in the next render call try that with both the webgl and canvas renderers we re working with the webgl renderer for the time being but i m not sure that our target platform will support this renderer currently a debian os running nwjs | 1 |
330,164 | 10,035,664,491 | IssuesEvent | 2019-07-18 08:53:06 | netdata/netdata | https://api.github.com/repos/netdata/netdata | closed | Provide an include configuration mechanism | area/daemon feature request priority/high | It would great if one could modify the configuration of netdata without touching the configuration shipped by the distribution. That way, when upgrades come along, those files are seen as not modified by the upgrade process and you get smoother upgrades, no diffs to read and new options as a bonus.
I don't know if there is a formalisation of this, but the include.d mechanism is great for that. That way we can edit (or ship via salt/ansible/puppet/chef/etc.) atomic files for atomic bits of configuration that don't interfere with the default configuration.
Netdata already uses that for python.d health.d charts.d etc. It would be awesome (or is already the case ?) to be able to have the same for the other configurations:
* netdata.conf -> add netdata.d/*.conf
* stream.conf -> add stream.d/*.conf
* fping.conf -> add fping.d/*.conf
* etc.
| 1.0 | Provide an include configuration mechanism - It would great if one could modify the configuration of netdata without touching the configuration shipped by the distribution. That way, when upgrades come along, those files are seen as not modified by the upgrade process and you get smoother upgrades, no diffs to read and new options as a bonus.
I don't know if there is a formalisation of this, but the include.d mechanism is great for that. That way we can edit (or ship via salt/ansible/puppet/chef/etc.) atomic files for atomic bits of configuration that don't interfere with the default configuration.
Netdata already uses that for python.d health.d charts.d etc. It would be awesome (or is already the case ?) to be able to have the same for the other configurations:
* netdata.conf -> add netdata.d/*.conf
* stream.conf -> add stream.d/*.conf
* fping.conf -> add fping.d/*.conf
* etc.
| priority | provide an include configuration mechanism it would great if one could modify the configuration of netdata without touching the configuration shipped by the distribution that way when upgrades come along those files are seen as not modified by the upgrade process and you get smoother upgrades no diffs to read and new options as a bonus i don t know if there is a formalisation of this but the include d mechanism is great for that that way we can edit or ship via salt ansible puppet chef etc atomic files for atomic bits of configuration that don t interfere with the default configuration netdata already uses that for python d health d charts d etc it would be awesome or is already the case to be able to have the same for the other configurations netdata conf add netdata d conf stream conf add stream d conf fping conf add fping d conf etc | 1 |
64,862 | 3,219,042,002 | IssuesEvent | 2015-10-08 07:21:11 | cs2103aug2015-t16-2j/main | https://api.github.com/repos/cs2103aug2015-t16-2j/main | closed | As a user, I want to edit/ update my entries. | priority.high type.story | So I do not have to delete and retype when I want to change details of my scheduled entry. | 1.0 | As a user, I want to edit/ update my entries. - So I do not have to delete and retype when I want to change details of my scheduled entry. | priority | as a user i want to edit update my entries so i do not have to delete and retype when i want to change details of my scheduled entry | 1 |
186,248 | 6,734,648,638 | IssuesEvent | 2017-10-18 18:47:38 | geosolutions-it/eumetsat-EOWS | https://api.github.com/repos/geosolutions-it/eumetsat-EOWS | opened | Improve NetCDF serving performance | geoserver Priority: High user story | A few things come to my mind here:
- [ ] remove intermediate indexes
- [ ] review and speed up code to reuse internal readers
| 1.0 | Improve NetCDF serving performance - A few things come to my mind here:
- [ ] remove intermediate indexes
- [ ] review and speed up code to reuse internal readers
| priority | improve netcdf serving performance a few things come to my mind here remove intermediate indexes review and speed up code to reuse internal readers | 1 |
277,868 | 8,633,694,499 | IssuesEvent | 2018-11-22 14:36:30 | BBC-News/psammead | https://api.github.com/repos/BBC-News/psammead | opened | GEL Typography, GEL Grid, GEL Spacing and Colour utilities | alpha-2 high priority refinement needed | **Is your feature request related to a problem? Please describe.**
We currently don't have utility files in this repository. We should add them.
**Describe the solution you'd like**
We should have a top-level `utilities` directory.
Currently in Simorgh, we have this structure
lib
├── constants
│ ├── styles.js
│ └── typography.js
├── globalStyles.js
├── globalStyles.test.js
└── layoutGrid.js
helpers
├── mediaQueries.js
Importing this structure here into Psammead is not ideal. We should create a clear structure and define what each file should contain.
I propose having these files in the `utilities` directory:
```
gel-breakpoints.js
gel-spacing.js
gel-grid-styled-components.js
gel-typography-styled-components.js
articles-colours.js
articles-fonts.js
articles-svgs.js
articles-amp-boilerplate.js
```
Here is the mapping from existing code to new component files:
`gel-spacing.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L30-L50
`articles-colours.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L4-L17
`articles-fonts.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L19-L28
`articles-svgs.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L83-L89
`articles-amp-boilerplate.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L90-L95
`gel-breakpoints.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L51-L82 & https://github.com/BBC-News/simorgh/blob/latest/src/app/helpers/mediaQueries.js#L7-L13
`gel-grid-styled-components.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/layoutGrid.js#L16-L82
`gel-typography-styled-components.js`: https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/typography.js#L12-L73
```
I would particularly like to draw attention to the last two files - since they export Styled Components, not just constants.
**Describe alternatives you've considered**
N/A
**Testing notes**
The files should have the correct content for the relevant names.
- [x] Initially labelled with ["Refinement needed"](https://github.com/BBC-News/psammead/labels/Refinement%20Needed)
| 1.0 | GEL Typography, GEL Grid, GEL Spacing and Colour utilities - **Is your feature request related to a problem? Please describe.**
We currently don't have utility files in this repository. We should add them.
**Describe the solution you'd like**
We should have a top-level `utilities` directory.
Currently in Simorgh, we have this structure
lib
├── constants
│ ├── styles.js
│ └── typography.js
├── globalStyles.js
├── globalStyles.test.js
└── layoutGrid.js
helpers
├── mediaQueries.js
Importing this structure here into Psammead is not ideal. We should create a clear structure and define what each file should contain.
I propose having these files in the `utilities` directory:
```
gel-breakpoints.js
gel-spacing.js
gel-grid-styled-components.js
gel-typography-styled-components.js
articles-colours.js
articles-fonts.js
articles-svgs.js
articles-amp-boilerplate.js
```
Here is the mapping from existing code to new component files:
`gel-spacing.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L30-L50
`articles-colours.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L4-L17
`articles-fonts.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L19-L28
`articles-svgs.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L83-L89
`articles-amp-boilerplate.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L90-L95
`gel-breakpoints.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/styles.js#L51-L82 & https://github.com/BBC-News/simorgh/blob/latest/src/app/helpers/mediaQueries.js#L7-L13
`gel-grid-styled-components.js` https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/layoutGrid.js#L16-L82
`gel-typography-styled-components.js`: https://github.com/BBC-News/simorgh/blob/latest/src/app/lib/constants/typography.js#L12-L73
```
I would particularly like to draw attention to the last two files - since they export Styled Components, not just constants.
**Describe alternatives you've considered**
N/A
**Testing notes**
The files should have the correct content for the relevant names.
- [x] Initially labelled with ["Refinement needed"](https://github.com/BBC-News/psammead/labels/Refinement%20Needed)
| priority | gel typography gel grid gel spacing and colour utilities is your feature request related to a problem please describe we currently don t have utility files in this repository we should add them describe the solution you d like we should have a top level utilities directory currently in simorgh we have this structure lib ├── constants │ ├── styles js │ └── typography js ├── globalstyles js ├── globalstyles test js └── layoutgrid js helpers ├── mediaqueries js importing this structure here into psammead is not ideal we should create a clear structure and define what each file should contain i propose having these files in the utilities directory gel breakpoints js gel spacing js gel grid styled components js gel typography styled components js articles colours js articles fonts js articles svgs js articles amp boilerplate js here is the mapping from existing code to new component files gel spacing js articles colours js articles fonts js articles svgs js articles amp boilerplate js gel breakpoints js gel grid styled components js gel typography styled components js i would particularly like to draw attention to the last two files since they export styled components not just constants describe alternatives you ve considered n a testing notes the files should have the correct content for the relevant names initially labelled with | 1 |
67,967 | 3,283,354,962 | IssuesEvent | 2015-10-28 12:14:29 | leeensminger/OED_Wetlands | https://api.github.com/repos/leeensminger/OED_Wetlands | closed | Projects - Associated Mitigation Sites do not display in the datagrid | bug - high priority | The Mitigation Sites tab does not display any records. I confirmed within the database that project ID 8cf9ce93-1f20-4a24-b32b-ddb97f19c49f (MD 124 Lane extension) is associated to 2 mitigation sites:
20f37616-5d1d-4d20-a09a-f3f01b8a6b0d
33556824-9b44-46f5-8dc7-8d85a9ed9475
No records are displayed within the mitigation sites tab.

| 1.0 | Projects - Associated Mitigation Sites do not display in the datagrid - The Mitigation Sites tab does not display any records. I confirmed within the database that project ID 8cf9ce93-1f20-4a24-b32b-ddb97f19c49f (MD 124 Lane extension) is associated to 2 mitigation sites:
20f37616-5d1d-4d20-a09a-f3f01b8a6b0d
33556824-9b44-46f5-8dc7-8d85a9ed9475
No records are displayed within the mitigation sites tab.

| priority | projects associated mitigation sites do not display in the datagrid the mitigation sites tab does not display any records i confirmed within the database that project id md lane extension is associated to mitigation sites no records are displayed within the mitigation sites tab | 1 |
336,004 | 10,169,769,939 | IssuesEvent | 2019-08-08 02:07:16 | treehouses/cli | https://api.github.com/repos/treehouses/cli | closed | Treehouses tor external/internal ports are switched | high priority | `treehouses tor add 80 8080` sets external to 80, local to 8080, but displays the opposite | 1.0 | Treehouses tor external/internal ports are switched - `treehouses tor add 80 8080` sets external to 80, local to 8080, but displays the opposite | priority | treehouses tor external internal ports are switched treehouses tor add sets external to local to but displays the opposite | 1 |
211,686 | 7,203,548,571 | IssuesEvent | 2018-02-06 09:35:34 | vmware/vic-product | https://api.github.com/repos/vmware/vic-product | closed | Document 1.3.0 to 1.3.1 upgrade | component/ova kind/user-doc priority/high pub/vsphere | Following Slack chat with @andrewtchin, there are differences between upgrading from 1.2.x to 1.3.x and upgrading from 1.3.0 to 1.3.1. This is due to the migration from 2 disks to 4 disks during upgrade from 1.2.x, meaning that in 1.3.x to 1.3.y upgrades, different disks need to be added/removed from the appliances.
@andrewtchin and @morris-jason can you please point me to any existing issues that provide details of the 1.3.x to 1.3.y procedure?
Thanks! | 1.0 | Document 1.3.0 to 1.3.1 upgrade - Following Slack chat with @andrewtchin, there are differences between upgrading from 1.2.x to 1.3.x and upgrading from 1.3.0 to 1.3.1. This is due to the migration from 2 disks to 4 disks during upgrade from 1.2.x, meaning that in 1.3.x to 1.3.y upgrades, different disks need to be added/removed from the appliances.
@andrewtchin and @morris-jason can you please point me to any existing issues that provide details of the 1.3.x to 1.3.y procedure?
Thanks! | priority | document to upgrade following slack chat with andrewtchin there are differences between upgrading from x to x and upgrading from to this is due to the migration from disks to disks during upgrade from x meaning that in x to y upgrades different disks need to be added removed from the appliances andrewtchin and morris jason can you please point me to any existing issues that provide details of the x to y procedure thanks | 1 |
176,312 | 6,558,208,966 | IssuesEvent | 2017-09-06 20:27:37 | zehro/UAH-Theatre | https://api.github.com/repos/zehro/UAH-Theatre | closed | Set up the SQL Database | Feature Priority - High Team - Database | ## Description
We need a database to hold all the inventory items/data.
## Tasks
- [x] Create an inventory database using a SQL database
- [x] Connect to the database via Python Flask
- [x] Set up basic CRUM operations to interact with the database
## Done Done Criteria
1. Test that the SQL database is live, interactable, and editable | 1.0 | Set up the SQL Database - ## Description
We need a database to hold all the inventory items/data.
## Tasks
- [x] Create an inventory database using a SQL database
- [x] Connect to the database via Python Flask
- [x] Set up basic CRUM operations to interact with the database
## Done Done Criteria
1. Test that the SQL database is live, interactable, and editable | priority | set up the sql database description we need a database to hold all the inventory items data tasks create an inventory database using a sql database connect to the database via python flask set up basic crum operations to interact with the database done done criteria test that the sql database is live interactable and editable | 1 |
470,172 | 13,533,984,629 | IssuesEvent | 2020-09-16 04:31:46 | SchizoDuckie/DuckieTV | https://api.github.com/repos/SchizoDuckie/DuckieTV | opened | Trakt TV updated shows API is now hard limited to 100 results meaning shows aren't updating | bug high priority | Not sure when this change happened but I noticed recently that the calendar hasn't been updating, this issue #1272 is also probably related.
When DuckieTV checks for updated shows it calls `https://api-v2launch.trakt.tv/shows/updates/{date}?limit=10000`, since this was added it there was never a limit on the amount of results that could be returned but it is now capped at 100 which means there is basically 0 chance any of your shows will be getting updated.
Changing the update period also isn't likely to help as the changes we request are only `yyyy-mm-dd` which means you'll only be getting changes for the whole day so even if you check every hour, you'll still only be getting all changes for that day and if there is already over 100 then you're not going to get any updates.
Only way to fix this is the code is going to have to change to load every page of changes while the total results returned is less than 100, unless there is another way to get changes.
| 1.0 | Trakt TV updated shows API is now hard limited to 100 results meaning shows aren't updating - Not sure when this change happened but I noticed recently that the calendar hasn't been updating, this issue #1272 is also probably related.
When DuckieTV checks for updated shows it calls `https://api-v2launch.trakt.tv/shows/updates/{date}?limit=10000`, since this was added it there was never a limit on the amount of results that could be returned but it is now capped at 100 which means there is basically 0 chance any of your shows will be getting updated.
Changing the update period also isn't likely to help as the changes we request are only `yyyy-mm-dd` which means you'll only be getting changes for the whole day so even if you check every hour, you'll still only be getting all changes for that day and if there is already over 100 then you're not going to get any updates.
Only way to fix this is the code is going to have to change to load every page of changes while the total results returned is less than 100, unless there is another way to get changes.
| priority | trakt tv updated shows api is now hard limited to results meaning shows aren t updating not sure when this change happened but i noticed recently that the calendar hasn t been updating this issue is also probably related when duckietv checks for updated shows it calls since this was added it there was never a limit on the amount of results that could be returned but it is now capped at which means there is basically chance any of your shows will be getting updated changing the update period also isn t likely to help as the changes we request are only yyyy mm dd which means you ll only be getting changes for the whole day so even if you check every hour you ll still only be getting all changes for that day and if there is already over then you re not going to get any updates only way to fix this is the code is going to have to change to load every page of changes while the total results returned is less than unless there is another way to get changes | 1 |
407,751 | 11,936,946,385 | IssuesEvent | 2020-04-02 11:15:25 | AY1920S2-CS2103T-F10-1/main | https://api.github.com/repos/AY1920S2-CS2103T-F10-1/main | closed | As a user, I want to generate a pdf version of my resume. | priority.High status.Ongoing | The generated resume needn't be very well formatted.
There should be, however, clear division between each section. | 1.0 | As a user, I want to generate a pdf version of my resume. - The generated resume needn't be very well formatted.
There should be, however, clear division between each section. | priority | as a user i want to generate a pdf version of my resume the generated resume needn t be very well formatted there should be however clear division between each section | 1 |
402,178 | 11,805,960,489 | IssuesEvent | 2020-03-19 08:35:12 | luna/ide | https://api.github.com/repos/luna/ide | closed | After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS | Category: IDE Priority: Highest Type: Bug | <!--
Please ensure that you are using the latest version of BaseGL before reporting
the bug! It may have been fixed since.
-->
### General Summary
<!--
- Please include a high-level description of your bug here.
-->
After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS. It doesn't matter if the text is copied from the editor or from anywhere else. Pasting on other systems works ok. There is no new errors in dev console.
### Steps to Reproduce
<!--
Please list the reproduction steps for your bug.
-->
1. Open Enso Studio and File MAnager Mock on MacOS
2. Write any text
3. Copy part of it
4. Paste it
### Expected Result
<!--
- A description of the results you expected from the reproduction steps.
-->
Changed text and fully working IDE
### Actual Result
<!--
- A description of what actually happens when you perform these steps.
- Please include any error output if relevant.
-->
Navigation with arrows is no longer working
### Luna Version
<!--
- Please include the version of BaseGL you are using here.
-->
| 1.0 | After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS - <!--
Please ensure that you are using the latest version of BaseGL before reporting
the bug! It may have been fixed since.
-->
### General Summary
<!--
- Please include a high-level description of your bug here.
-->
After pasting text to text editor navigation with keyboard (arrows) stops working on MacOS. It doesn't matter if the text is copied from the editor or from anywhere else. Pasting on other systems works ok. There is no new errors in dev console.
### Steps to Reproduce
<!--
Please list the reproduction steps for your bug.
-->
1. Open Enso Studio and File MAnager Mock on MacOS
2. Write any text
3. Copy part of it
4. Paste it
### Expected Result
<!--
- A description of the results you expected from the reproduction steps.
-->
Changed text and fully working IDE
### Actual Result
<!--
- A description of what actually happens when you perform these steps.
- Please include any error output if relevant.
-->
Navigation with arrows is no longer working
### Luna Version
<!--
- Please include the version of BaseGL you are using here.
-->
| priority | after pasting text to text editor navigation with keyboard arrows stops working on macos please ensure that you are using the latest version of basegl before reporting the bug it may have been fixed since general summary please include a high level description of your bug here after pasting text to text editor navigation with keyboard arrows stops working on macos it doesn t matter if the text is copied from the editor or from anywhere else pasting on other systems works ok there is no new errors in dev console steps to reproduce please list the reproduction steps for your bug open enso studio and file manager mock on macos write any text copy part of it paste it expected result a description of the results you expected from the reproduction steps changed text and fully working ide actual result a description of what actually happens when you perform these steps please include any error output if relevant navigation with arrows is no longer working luna version please include the version of basegl you are using here | 1 |
79,080 | 3,520,196,820 | IssuesEvent | 2016-01-12 19:49:20 | Thaerious/NERVE | https://api.github.com/repos/Thaerious/NERVE | reopened | Although logged into CWRC, can't get list of CWRC entity links | bug - can not replicate bug - high priority | I'm having no problems with the Geonames lookup.
When trying to link to names, sometimes I get the message saying that I haven't logged into the CWRC beta site, even though I have, and sometimes I get this message:
java.lang.NullPointerException: TEC not set.
TEC not set.
- Throwable.java:116
- Exception.java:29
- RuntimeException.java:29
- NullPointerException.java:29
- GenericLinkForm.java:40
- GenericLinkForm.java:39
- PopupPanel.java:512
- DropFlowListener.java:256
- CWRCLookup.java:38
- Request.java:250
- RequestBuilder.java:412
- XMLHttpRequest.java:329
- Impl.java:247
- Impl.java:299
- Impl.java:72 | 1.0 | Although logged into CWRC, can't get list of CWRC entity links - I'm having no problems with the Geonames lookup.
When trying to link to names, sometimes I get the message saying that I haven't logged into the CWRC beta site, even though I have, and sometimes I get this message:
java.lang.NullPointerException: TEC not set.
TEC not set.
- Throwable.java:116
- Exception.java:29
- RuntimeException.java:29
- NullPointerException.java:29
- GenericLinkForm.java:40
- GenericLinkForm.java:39
- PopupPanel.java:512
- DropFlowListener.java:256
- CWRCLookup.java:38
- Request.java:250
- RequestBuilder.java:412
- XMLHttpRequest.java:329
- Impl.java:247
- Impl.java:299
- Impl.java:72 | priority | although logged into cwrc can t get list of cwrc entity links i m having no problems with the geonames lookup when trying to link to names sometimes i get the message saying that i haven t logged into the cwrc beta site even though i have and sometimes i get this message java lang nullpointerexception tec not set tec not set throwable java exception java runtimeexception java nullpointerexception java genericlinkform java genericlinkform java popuppanel java dropflowlistener java cwrclookup java request java requestbuilder java xmlhttprequest java impl java impl java impl java | 1 |
28,564 | 2,707,929,279 | IssuesEvent | 2015-04-08 03:42:52 | openfarmcc/OpenFarm | https://api.github.com/repos/openfarmcc/OpenFarm | closed | Subscribe to Newsletter on User creation flow does not work | bug High Priority User Accounts | Mailchimp receives no email address and Active Admin shows that the boolean has not been switched to true. | 1.0 | Subscribe to Newsletter on User creation flow does not work - Mailchimp receives no email address and Active Admin shows that the boolean has not been switched to true. | priority | subscribe to newsletter on user creation flow does not work mailchimp receives no email address and active admin shows that the boolean has not been switched to true | 1 |
519,614 | 15,054,636,425 | IssuesEvent | 2021-02-03 17:41:28 | mredwin1/DivisionManagementSystem | https://api.github.com/repos/mredwin1/DivisionManagementSystem | closed | When generating the attendance document only look at history up to the incident date | Area: Backend Priority: High Type: Bug | Currently the history includes their entire history but should only include thing up to the incident they are receiving encase something gets edited and there is a future point. | 1.0 | When generating the attendance document only look at history up to the incident date - Currently the history includes their entire history but should only include thing up to the incident they are receiving encase something gets edited and there is a future point. | priority | when generating the attendance document only look at history up to the incident date currently the history includes their entire history but should only include thing up to the incident they are receiving encase something gets edited and there is a future point | 1 |
620,604 | 19,565,652,922 | IssuesEvent | 2022-01-03 23:37:01 | bounswe/2021SpringGroup6 | https://api.github.com/repos/bounswe/2021SpringGroup6 | closed | Android - User's created events | Type: Task Status: In Progress Platform: Mobile Priority: High | The events created by the user will be displayed on home page. The user will be able to navigate to the Event Detail pages of his/her created events. On Event Detail Page, the user will be able to reject or accept the interested users of the event. | 1.0 | Android - User's created events - The events created by the user will be displayed on home page. The user will be able to navigate to the Event Detail pages of his/her created events. On Event Detail Page, the user will be able to reject or accept the interested users of the event. | priority | android user s created events the events created by the user will be displayed on home page the user will be able to navigate to the event detail pages of his her created events on event detail page the user will be able to reject or accept the interested users of the event | 1 |
188,132 | 6,773,076,747 | IssuesEvent | 2017-10-27 03:10:52 | xcat2/xcat-core | https://api.github.com/repos/xcat2/xcat-core | closed | [WSC] trying to put xcat common openbmc attributes into group def fails when createing node def | component:coral priority:high sprint2 type:usability | we have the following group definition:
````
[root@c699mgt00 create_cnodes]# lsdef -t group ws_compute -z
# <xCAT data object stanza file>
ws_compute:
objtype=group
cons=openbmc
consoleondemand=1
grouptype=static
mgt=openbmc
netboot=petitboot
nichostnamesuffixes.ib2=-ib2
nichostnamesuffixes.ib0=-ib0
nichostnamesuffixes.ib1=-ib1
nichostnamesuffixes.ib3=-ib3
nicnetworks.ib2=IB02
nicnetworks.ib0=IB00
nicnetworks.ib1=IB01
nicnetworks.ib3=IB03
nictypes.ib2=Infiniband
nictypes.ib0=Infiniband
nictypes.ib1=Infiniband
nictypes.ib3=Infiniband
primarynic=mac
profile=compute
usercomment=[offline] -- from ws_compute group def
````
We then try to do:
````
chdef -z <<EOF
c699c001:
objtype=node
nodetype=mp
bmc=10.3.1.101
groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01
hostnames=c699f01-U04
ip=10.3.1.1
nicips.ib0=10.10.1.1
nicips.ib1=10.11.1.1
nicips.ib2=10.12.1.1
nicips.ib3=10.13.1.1
rack=c699f01
switch=c699tor01
switchport=1
unit=u04
EOF
````
it then tells us:
Cannot set the 'bmc' attribute unless a value is provided for 'mgt'.
mgt => The method to use to do general hardware management of the node. This attribute is used as the default if power or getmac is not set. Valid values: openbmc, ipmi, blade, hmc, ivm, fsp, bpa, kvm, esx, rhevm. See the power attribute for more details.
Error: Cannot set the attr='bmc' attribute unless 'mgt=openbmc'.
1 object definitions have been created or modified.
New object definitions 'c699c001' have been created.
The resultant definition is:
````
[root@c699mgt00 create_cnodes]# lsdef c699c001
Object name: c699c001
cons=openbmc
consoleondemand=1
groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01
hostnames=c699f01-U04
ip=10.3.1.1
mgt=openbmc
netboot=petitboot
nichostnamesuffixes.ib2=-ib2
nichostnamesuffixes.ib0=-ib0
nichostnamesuffixes.ib1=-ib1
nichostnamesuffixes.ib3=-ib3
nicips.ib2=10.12.1.1
nicips.ib0=10.10.1.1
nicips.ib1=10.11.1.1
nicips.ib3=10.13.1.1
nicnetworks.ib2=IB02
nicnetworks.ib0=IB00
nicnetworks.ib1=IB01
nicnetworks.ib3=IB03
nictypes.ib2=Infiniband
nictypes.ib0=Infiniband
nictypes.ib1=Infiniband
nictypes.ib3=Infiniband
nodetype=mp
postbootscripts=otherpkgs
primarynic=mac
profile=compute
rack=c699f01
switch=c699tor01
switchport=1
unit=u04
````
note, If i repeat the command a second time it works:
````
[root@c699mgt00 create_cnodes]# chdef -z <<EOF
c699c001:
objtype=node
nodetype=mp
bmc=10.3.1.101
groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01
hostnames=c699f01-U04
ip=10.3.1.1
nicips.ib0=10.10.1.1
nicips.ib1=10.11.1.1
nicips.ib2=10.12.1.1
nicips.ib3=10.13.1.1
rack=c699f01
switch=c699tor01
switchport=1
unit=u04
EOF
1 object definitions have been created or modified.
```` | 1.0 | [WSC] trying to put xcat common openbmc attributes into group def fails when createing node def - we have the following group definition:
````
[root@c699mgt00 create_cnodes]# lsdef -t group ws_compute -z
# <xCAT data object stanza file>
ws_compute:
objtype=group
cons=openbmc
consoleondemand=1
grouptype=static
mgt=openbmc
netboot=petitboot
nichostnamesuffixes.ib2=-ib2
nichostnamesuffixes.ib0=-ib0
nichostnamesuffixes.ib1=-ib1
nichostnamesuffixes.ib3=-ib3
nicnetworks.ib2=IB02
nicnetworks.ib0=IB00
nicnetworks.ib1=IB01
nicnetworks.ib3=IB03
nictypes.ib2=Infiniband
nictypes.ib0=Infiniband
nictypes.ib1=Infiniband
nictypes.ib3=Infiniband
primarynic=mac
profile=compute
usercomment=[offline] -- from ws_compute group def
````
We then try to do:
````
chdef -z <<EOF
c699c001:
objtype=node
nodetype=mp
bmc=10.3.1.101
groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01
hostnames=c699f01-U04
ip=10.3.1.1
nicips.ib0=10.10.1.1
nicips.ib1=10.11.1.1
nicips.ib2=10.12.1.1
nicips.ib3=10.13.1.1
rack=c699f01
switch=c699tor01
switchport=1
unit=u04
EOF
````
it then tells us:
Cannot set the 'bmc' attribute unless a value is provided for 'mgt'.
mgt => The method to use to do general hardware management of the node. This attribute is used as the default if power or getmac is not set. Valid values: openbmc, ipmi, blade, hmc, ivm, fsp, bpa, kvm, esx, rhevm. See the power attribute for more details.
Error: Cannot set the attr='bmc' attribute unless 'mgt=openbmc'.
1 object definitions have been created or modified.
New object definitions 'c699c001' have been created.
The resultant definition is:
````
[root@c699mgt00 create_cnodes]# lsdef c699c001
Object name: c699c001
cons=openbmc
consoleondemand=1
groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01
hostnames=c699f01-U04
ip=10.3.1.1
mgt=openbmc
netboot=petitboot
nichostnamesuffixes.ib2=-ib2
nichostnamesuffixes.ib0=-ib0
nichostnamesuffixes.ib1=-ib1
nichostnamesuffixes.ib3=-ib3
nicips.ib2=10.12.1.1
nicips.ib0=10.10.1.1
nicips.ib1=10.11.1.1
nicips.ib3=10.13.1.1
nicnetworks.ib2=IB02
nicnetworks.ib0=IB00
nicnetworks.ib1=IB01
nicnetworks.ib3=IB03
nictypes.ib2=Infiniband
nictypes.ib0=Infiniband
nictypes.ib1=Infiniband
nictypes.ib3=Infiniband
nodetype=mp
postbootscripts=otherpkgs
primarynic=mac
profile=compute
rack=c699f01
switch=c699tor01
switchport=1
unit=u04
````
note, If i repeat the command a second time it works:
````
[root@c699mgt00 create_cnodes]# chdef -z <<EOF
c699c001:
objtype=node
nodetype=mp
bmc=10.3.1.101
groups=all,compute,lsf,rh74_compute,witherspoon,ws_compute,ws_sequoiac699f01
hostnames=c699f01-U04
ip=10.3.1.1
nicips.ib0=10.10.1.1
nicips.ib1=10.11.1.1
nicips.ib2=10.12.1.1
nicips.ib3=10.13.1.1
rack=c699f01
switch=c699tor01
switchport=1
unit=u04
EOF
1 object definitions have been created or modified.
```` | priority | trying to put xcat common openbmc attributes into group def fails when createing node def we have the following group definition lsdef t group ws compute z ws compute objtype group cons openbmc consoleondemand grouptype static mgt openbmc netboot petitboot nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nicnetworks nicnetworks nicnetworks nicnetworks nictypes infiniband nictypes infiniband nictypes infiniband nictypes infiniband primarynic mac profile compute usercomment from ws compute group def we then try to do chdef z eof objtype node nodetype mp bmc groups all compute lsf compute witherspoon ws compute ws hostnames ip nicips nicips nicips nicips rack switch switchport unit eof it then tells us cannot set the bmc attribute unless a value is provided for mgt mgt the method to use to do general hardware management of the node this attribute is used as the default if power or getmac is not set valid values openbmc ipmi blade hmc ivm fsp bpa kvm esx rhevm see the power attribute for more details error cannot set the attr bmc attribute unless mgt openbmc object definitions have been created or modified new object definitions have been created the resultant definition is lsdef object name cons openbmc consoleondemand groups all compute lsf compute witherspoon ws compute ws hostnames ip mgt openbmc netboot petitboot nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nichostnamesuffixes nicips nicips nicips nicips nicnetworks nicnetworks nicnetworks nicnetworks nictypes infiniband nictypes infiniband nictypes infiniband nictypes infiniband nodetype mp postbootscripts otherpkgs primarynic mac profile compute rack switch switchport unit note if i repeat the command a second time it works chdef z eof objtype node nodetype mp bmc groups all compute lsf compute witherspoon ws compute ws hostnames ip nicips nicips nicips nicips rack switch switchport unit eof object definitions have been created or modified | 1 |
705,114 | 24,221,961,680 | IssuesEvent | 2022-09-26 11:38:49 | dnd-side-project/dnd-7th-7-frontend | https://api.github.com/repos/dnd-side-project/dnd-7th-7-frontend | closed | [feat] 경로 등록 API form data 가공 및 post 작업 | 🔧 Type: Feature 🔴 Priority: High | ## 🔥 설명 (이슈에 대한 자세한 설명을 적어주세요!)
-
<br>
## 🔥 ETC (기타 사항을 적어주세요!)
-
<br>
>#### types
>- feat : 새로운 기능
>- update : 로직 변경
>- bug : 버그 수정
>- refactor : 파일 및 폴더 리팩토링
>- style : 코드 형식 변경
>- docs : 문서 관련
>- test : 테스트 코드
>- etc : 기타 변경 사항
>
| 1.0 | [feat] 경로 등록 API form data 가공 및 post 작업 - ## 🔥 설명 (이슈에 대한 자세한 설명을 적어주세요!)
-
<br>
## 🔥 ETC (기타 사항을 적어주세요!)
-
<br>
>#### types
>- feat : 새로운 기능
>- update : 로직 변경
>- bug : 버그 수정
>- refactor : 파일 및 폴더 리팩토링
>- style : 코드 형식 변경
>- docs : 문서 관련
>- test : 테스트 코드
>- etc : 기타 변경 사항
>
| priority | 경로 등록 api form data 가공 및 post 작업 🔥 설명 이슈에 대한 자세한 설명을 적어주세요 🔥 etc 기타 사항을 적어주세요 types feat 새로운 기능 update 로직 변경 bug 버그 수정 refactor 파일 및 폴더 리팩토링 style 코드 형식 변경 docs 문서 관련 test 테스트 코드 etc 기타 변경 사항 | 1 |
648,104 | 21,165,233,292 | IssuesEvent | 2022-04-07 13:04:37 | Aam-Digital/ndb-core | https://api.github.com/repos/Aam-Digital/ndb-core | opened | new notes added on child details tab are only displayed after reload | Type: Bug Status: High Priority | **Describe the bug**
Creating a new note from the "Notes" tab of the Child Details view opens a popup form and works correctly, but the new note is not displayed in the table after saving. Only after opening that child's details again is the new note shown in the table.
| 1.0 | new notes added on child details tab are only displayed after reload - **Describe the bug**
Creating a new note from the "Notes" tab of the Child Details view opens a popup form and works correctly, but the new note is not displayed in the table after saving. Only after opening that child's details again is the new note shown in the table.
| priority | new notes added on child details tab are only displayed after reload describe the bug creating a new note from the notes tab of the child details view opens a popup form and works correctly but the new note is not displayed in the table after saving only after opening that child s details again is the new note shown in the table | 1 |
554,889 | 16,441,572,844 | IssuesEvent | 2021-05-20 14:53:41 | hunchat/hunchat-api | https://api.github.com/repos/hunchat/hunchat-api | opened | Make `user.username` the default for `user.name` | priority:high | In the creation of a `User` object, if no `name` is provided, set `username` as `name`.
Note: I might have to set `blank=True` in the model field definition. | 1.0 | Make `user.username` the default for `user.name` - In the creation of a `User` object, if no `name` is provided, set `username` as `name`.
Note: I might have to set `blank=True` in the model field definition. | priority | make user username the default for user name in the creation of a user object if no name is provided set username as name note i might have to set blank true in the model field definition | 1 |
324,639 | 9,906,793,878 | IssuesEvent | 2019-06-27 14:33:07 | infor-design/enterprise | https://api.github.com/repos/infor-design/enterprise | reopened | Dropdown: 'startsWith' filter works by word rather than start of a String | [2] focus: datagrid priority: high type: bug :bug: | <!-- Please be aware that this is a publicly visible bug report. Do not post any credentials, screenshots with proprietary information, or anything you think shouldn't be visible to the world. If private information is required to be shared for a quality bug report, please email one of the [code owners](https://github.com/infor-design/enterprise/blob/master/.github/CODEOWNERS) directly. -->
**Describe the bug**
The 'startsWith' filter currently works on words of a String, rather than start of a String (see below for example). Currently, there is no way to filter on start of a String.
Example list:

If I enter 'C' I expect 'Correction' to be selected. This works

However, if I enter 'A' I expect ''Allocated' to be selected but instead 'Not Allocated' is selected. It seems that the 'startsWith' works on word splits.

**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://design.infor.com/code/ids-enterprise/latest/demo/components/dropdown/test-filter-types.html
2. Enter 'S' into 'StartsWith Filter' dropdown.
3. 'American Samoa' selected by default
**Expected behavior**
Expected 'South Carolina' to be selected by default
**Version**
Found in 4.14.0, exists in 4.15.0
**Platform**
- Device: PC
- OS Version: Windows 10
- Browser Name: Chrome
- Browser Version: Version 72.0.3626.81 (Official Build) (64-bit)
**Additional context**
Discussion had with @tmcconechy and thought process was to change existing 'startsWith' behavior to filter based on start of String, and to add a new filter option such as 'startsWithWord' to filter based on the current behavior of 'startsWith'.
| 1.0 | Dropdown: 'startsWith' filter works by word rather than start of a String - <!-- Please be aware that this is a publicly visible bug report. Do not post any credentials, screenshots with proprietary information, or anything you think shouldn't be visible to the world. If private information is required to be shared for a quality bug report, please email one of the [code owners](https://github.com/infor-design/enterprise/blob/master/.github/CODEOWNERS) directly. -->
**Describe the bug**
The 'startsWith' filter currently works on words of a String, rather than start of a String (see below for example). Currently, there is no way to filter on start of a String.
Example list:

If I enter 'C' I expect 'Correction' to be selected. This works

However, if I enter 'A' I expect ''Allocated' to be selected but instead 'Not Allocated' is selected. It seems that the 'startsWith' works on word splits.

**To Reproduce**
Steps to reproduce the behavior:
1. Go to https://design.infor.com/code/ids-enterprise/latest/demo/components/dropdown/test-filter-types.html
2. Enter 'S' into 'StartsWith Filter' dropdown.
3. 'American Samoa' selected by default
**Expected behavior**
Expected 'South Carolina' to be selected by default
**Version**
Found in 4.14.0, exists in 4.15.0
**Platform**
- Device: PC
- OS Version: Windows 10
- Browser Name: Chrome
- Browser Version: Version 72.0.3626.81 (Official Build) (64-bit)
**Additional context**
Discussion had with @tmcconechy and thought process was to change existing 'startsWith' behavior to filter based on start of String, and to add a new filter option such as 'startsWithWord' to filter based on the current behavior of 'startsWith'.
| priority | dropdown startswith filter works by word rather than start of a string describe the bug the startswith filter currently works on words of a string rather than start of a string see below for example currently there is no way to filter on start of a string example list if i enter c i expect correction to be selected this works however if i enter a i expect allocated to be selected but instead not allocated is selected it seems that the startswith works on word splits to reproduce steps to reproduce the behavior go to enter s into startswith filter dropdown american samoa selected by default expected behavior expected south carolina to be selected by default version found in exists in platform device pc os version windows browser name chrome browser version version official build bit additional context discussion had with tmcconechy and thought process was to change existing startswith behavior to filter based on start of string and to add a new filter option such as startswithword to filter based on the current behavior of startswith | 1 |
396,361 | 11,708,394,350 | IssuesEvent | 2020-03-08 13:03:38 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | opened | Graph failed to attach when adding comment | Priority: High | 
This comment had both a map and a graph, but only one showed. | 1.0 | Graph failed to attach when adding comment - 
This comment had both a map and a graph, but only one showed. | priority | graph failed to attach when adding comment this comment had both a map and a graph but only one showed | 1 |
669,333 | 22,620,275,596 | IssuesEvent | 2022-06-30 05:21:58 | heading1/WYLSBingsu | https://api.github.com/repos/heading1/WYLSBingsu | closed | [FE] 알림 컴포넌트 생성 | 🖥 Frontend ❗️high-priority 🔨 Feature | ## 🔨 기능 설명
알림 컴포넌트 생성
## 📑 완료 조건
알림 컴포넌트 생성
## 💭 관련 백로그
[대분류]-[중분류]-[소분류]
공통 컴포넌트 - 알림 컴포넌트 - 알림컴포넌트 생성
## 💭 예상 작업 시간
1h
| 1.0 | [FE] 알림 컴포넌트 생성 - ## 🔨 기능 설명
알림 컴포넌트 생성
## 📑 완료 조건
알림 컴포넌트 생성
## 💭 관련 백로그
[대분류]-[중분류]-[소분류]
공통 컴포넌트 - 알림 컴포넌트 - 알림컴포넌트 생성
## 💭 예상 작업 시간
1h
| priority | 알림 컴포넌트 생성 🔨 기능 설명 알림 컴포넌트 생성 📑 완료 조건 알림 컴포넌트 생성 💭 관련 백로그 공통 컴포넌트 알림 컴포넌트 알림컴포넌트 생성 💭 예상 작업 시간 | 1 |
114,648 | 4,642,039,090 | IssuesEvent | 2016-09-30 08:08:47 | rndsolutions/hawkcd | https://api.github.com/repos/rndsolutions/hawkcd | closed | Fetch Material task doesn't show all available materials | awaiting verification bug high priority ui | **Steps to reproduce**
1. Log in as Admin
2. Edit an existing Pipeline's tasks
3. Add a Fetch Material task
**Expected behavior**
All materials should be displayed in Material selection of the Fetch Material task.
**Actual behavior**
Only the assigned Material is shown. | 1.0 | Fetch Material task doesn't show all available materials - **Steps to reproduce**
1. Log in as Admin
2. Edit an existing Pipeline's tasks
3. Add a Fetch Material task
**Expected behavior**
All materials should be displayed in Material selection of the Fetch Material task.
**Actual behavior**
Only the assigned Material is shown. | priority | fetch material task doesn t show all available materials steps to reproduce log in as admin edit an existing pipeline s tasks add a fetch material task expected behavior all materials should be displayed in material selection of the fetch material task actual behavior only the assigned material is shown | 1 |
760,322 | 26,636,771,265 | IssuesEvent | 2023-01-24 22:48:12 | yuri2303/portifolio | https://api.github.com/repos/yuri2303/portifolio | closed | criar estrutura inicia do projeto | priority: high weight:2 type: feature | ## Descrição
Seleciona um template pronto e adicionar os arquivos para o projeto. | 1.0 | criar estrutura inicia do projeto - ## Descrição
Seleciona um template pronto e adicionar os arquivos para o projeto. | priority | criar estrutura inicia do projeto descrição seleciona um template pronto e adicionar os arquivos para o projeto | 1 |
660,814 | 22,031,983,042 | IssuesEvent | 2022-05-28 02:09:29 | uwlib-cams/sinopia_maps | https://api.github.com/repos/uwlib-cams/sinopia_maps | closed | RECONFIGURE OUTPUT TO RTs because many things have changed | high priority xslt messy issue | ack
🤢 🤢 🤢
RDF RT structure has changed

😢 😢 😢 | 1.0 | RECONFIGURE OUTPUT TO RTs because many things have changed - ack
🤢 🤢 🤢
RDF RT structure has changed

😢 😢 😢 | priority | reconfigure output to rts because many things have changed ack 🤢 🤢 🤢 rdf rt structure has changed 😢 😢 😢 | 1 |
176,906 | 6,569,079,614 | IssuesEvent | 2017-09-09 02:01:44 | ianroberts131/visual-algorithms | https://api.github.com/repos/ianroberts131/visual-algorithms | closed | Fix Speed Buttons for Merge Sort | Priority: High Type: Bug | # Bug
## What is the current behavior?
When changing the speed for Merge Sort, the algorithm doesn't run properly
## What is the expected behavior?
Should be able to change the speed and have the algorithm function properly
| 1.0 | Fix Speed Buttons for Merge Sort - # Bug
## What is the current behavior?
When changing the speed for Merge Sort, the algorithm doesn't run properly
## What is the expected behavior?
Should be able to change the speed and have the algorithm function properly
| priority | fix speed buttons for merge sort bug what is the current behavior when changing the speed for merge sort the algorithm doesn t run properly what is the expected behavior should be able to change the speed and have the algorithm function properly | 1 |
237,981 | 7,768,810,133 | IssuesEvent | 2018-06-03 22:05:27 | 7547-G2/hoy-como-backend | https://api.github.com/repos/7547-G2/hoy-como-backend | closed | Crear get de extras para un plato en Mobile | Backend High-Priority Sprint 4 | se envia id del plato en el path
se espera recibir jsonArray:
[{"extra_id": 99, "nombre_extra": "string", "precio": 999}, {...}, ...] | 1.0 | Crear get de extras para un plato en Mobile - se envia id del plato en el path
se espera recibir jsonArray:
[{"extra_id": 99, "nombre_extra": "string", "precio": 999}, {...}, ...] | priority | crear get de extras para un plato en mobile se envia id del plato en el path se espera recibir jsonarray | 1 |
384,782 | 11,403,270,934 | IssuesEvent | 2020-01-31 06:38:28 | inverse-inc/packetfence | https://api.github.com/repos/inverse-inc/packetfence | closed | API: unable to create a switch with a MAC address as ID | API Priority: High Type: Bug | **Describe the bug**
When you want to create a switch with MAC address as ID, you can't.
**To Reproduce**
Steps to reproduce the behavior:
```console
# pfperl-api get -M POST /api/v1/config/switches -c '{"id":"ff:ff:ff:ff:ff:ff","description":"test"}' | python -m json.tool
{
"errors": [
{
"field": "id",
"message": "Wrong value"
}
],
"message": "Unable to validate",
"status": 422
}
```
**Expected behavior**
Switch creation should occur.
EDIT:
For this switch:
- `DELETE` is working
- `PATCH` is not working
- `GET` is working | 1.0 | API: unable to create a switch with a MAC address as ID - **Describe the bug**
When you want to create a switch with MAC address as ID, you can't.
**To Reproduce**
Steps to reproduce the behavior:
```console
# pfperl-api get -M POST /api/v1/config/switches -c '{"id":"ff:ff:ff:ff:ff:ff","description":"test"}' | python -m json.tool
{
"errors": [
{
"field": "id",
"message": "Wrong value"
}
],
"message": "Unable to validate",
"status": 422
}
```
**Expected behavior**
Switch creation should occur.
EDIT:
For this switch:
- `DELETE` is working
- `PATCH` is not working
- `GET` is working | priority | api unable to create a switch with a mac address as id describe the bug when you want to create a switch with mac address as id you can t to reproduce steps to reproduce the behavior console pfperl api get m post api config switches c id ff ff ff ff ff ff description test python m json tool errors field id message wrong value message unable to validate status expected behavior switch creation should occur edit for this switch delete is working patch is not working get is working | 1 |
801,997 | 28,564,889,169 | IssuesEvent | 2023-04-21 00:33:19 | WordPress/wporg-two-factor | https://api.github.com/repos/WordPress/wporg-two-factor | closed | Temporarily updating the email address but not savings breaks password view. | bug ui priority: high | ## GIF

**Steps to reproduce**
1. Visit `users/{username}/edit/account/?screen=email`
2. Change the email address to an invalid email address
3. Click "Back"
4. Click "Password" list item
5. Notice the UI is broken.
| 1.0 | Temporarily updating the email address but not savings breaks password view. - ## GIF

**Steps to reproduce**
1. Visit `users/{username}/edit/account/?screen=email`
2. Change the email address to an invalid email address
3. Click "Back"
4. Click "Password" list item
5. Notice the UI is broken.
| priority | temporarily updating the email address but not savings breaks password view gif steps to reproduce visit users username edit account screen email change the email address to an invalid email address click back click password list item notice the ui is broken | 1 |
224,606 | 7,471,942,370 | IssuesEvent | 2018-04-03 10:56:26 | ballerina-lang/composer | https://api.github.com/repos/ballerina-lang/composer | closed | Change the border color of the selected element | 0.94-pre-release Imported Priority/Highest Severity/Major component/Composer | Can we change the color of the border of a code segment? If e have a large code block e.g: a while loop. It is difficult to recognize the boundary of the loop. Can we change the color when the element is selected? | 1.0 | Change the border color of the selected element - Can we change the color of the border of a code segment? If e have a large code block e.g: a while loop. It is difficult to recognize the boundary of the loop. Can we change the color when the element is selected? | priority | change the border color of the selected element can we change the color of the border of a code segment if e have a large code block e g a while loop it is difficult to recognize the boundary of the loop can we change the color when the element is selected | 1 |
167,728 | 6,345,281,142 | IssuesEvent | 2017-07-27 21:52:32 | ualbertalib/DMPonline_v4 | https://api.github.com/repos/ualbertalib/DMPonline_v4 | closed | Guidance is visible even it's set to be not published. | bug Priority:High | Reported by UQAM:
1.

2. Plans created in English and in French

3. Display

| 1.0 | Guidance is visible even it's set to be not published. - Reported by UQAM:
1.

2. Plans created in English and in French

3. Display

| priority | guidance is visible even it s set to be not published reported by uqam plans created in english and in french display | 1 |
630,244 | 20,102,605,431 | IssuesEvent | 2022-02-07 07:00:23 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | reopened | Wrong operation in stft's center attribute | high priority triage review module: correctness (silent) module: fft | ### 🐛 Describe the bug
According to the pytorch's document, if I set `center` to `True` the first frame should be centered in the beginning of stft:
https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L620
And the following code shows that the stft pads by `n_fft//2`
https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L692
This works fine in most cases but if the `win_length` and `n_fft` are different padding `n_fft//2` does not gurantee the first frame be centered. I think `win_length//2` is the right choice for the proper operation.
### Versions
Collecting environment information...
PyTorch version: 1.10.1+cu113
Is debug build: False
CUDA used to build PyTorch: 11.3
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.1 LTS (x86_64)
GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Clang version: Could not collect
CMake version: version 3.16.3
Libc version: glibc-2.31
Python version: 3.8.10 (default, Sep 28 2021, 16:10:42) [GCC 9.3.0] (64-bit runtime)
Python platform: Linux-3.10.0-693.5.2.el7.x86_64-x86_64-with-glibc2.29
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: Tesla P40
GPU 1: Tesla P40
GPU 2: Tesla P40
GPU 3: Tesla P40
GPU 4: Tesla P40
GPU 5: Tesla P40
GPU 6: Tesla P40
GPU 7: Tesla P40
Nvidia driver version: 455.32.00
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.0.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.21.4
[pip3] pytorch-lightning==1.4.9
[pip3] torch==1.10.1+cu113
[pip3] torchaudio==0.10.1+cu113
[pip3] torchmetrics==0.6.1
[pip3] torchvision==0.11.2+cu113
[conda] Could not collect
cc @ezyang @gchanan @zou3519 @anjali411 @dylanbespalko @mruberry @Lezcano @nikitaved @peterbell10 | 1.0 | Wrong operation in stft's center attribute - ### 🐛 Describe the bug
According to the pytorch's document, if I set `center` to `True` the first frame should be centered in the beginning of stft:
https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L620
And the following code shows that the stft pads by `n_fft//2`
https://github.com/pytorch/pytorch/blob/2367face24afb159f73ebf40dc6f23e46132b770/torch/functional.py#L692
This works fine in most cases but if the `win_length` and `n_fft` are different padding `n_fft//2` does not gurantee the first frame be centered. I think `win_length//2` is the right choice for the proper operation.
### Versions
Collecting environment information...
PyTorch version: 1.10.1+cu113
Is debug build: False
CUDA used to build PyTorch: 11.3
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.1 LTS (x86_64)
GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Clang version: Could not collect
CMake version: version 3.16.3
Libc version: glibc-2.31
Python version: 3.8.10 (default, Sep 28 2021, 16:10:42) [GCC 9.3.0] (64-bit runtime)
Python platform: Linux-3.10.0-693.5.2.el7.x86_64-x86_64-with-glibc2.29
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: Tesla P40
GPU 1: Tesla P40
GPU 2: Tesla P40
GPU 3: Tesla P40
GPU 4: Tesla P40
GPU 5: Tesla P40
GPU 6: Tesla P40
GPU 7: Tesla P40
Nvidia driver version: 455.32.00
cuDNN version: Probably one of the following:
/usr/lib/x86_64-linux-gnu/libcudnn.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_adv_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_cnn_train.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_infer.so.8.0.5
/usr/lib/x86_64-linux-gnu/libcudnn_ops_train.so.8.0.5
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] numpy==1.21.4
[pip3] pytorch-lightning==1.4.9
[pip3] torch==1.10.1+cu113
[pip3] torchaudio==0.10.1+cu113
[pip3] torchmetrics==0.6.1
[pip3] torchvision==0.11.2+cu113
[conda] Could not collect
cc @ezyang @gchanan @zou3519 @anjali411 @dylanbespalko @mruberry @Lezcano @nikitaved @peterbell10 | priority | wrong operation in stft s center attribute 🐛 describe the bug according to the pytorch s document if i set center to true the first frame should be centered in the beginning of stft and the following code shows that the stft pads by n fft this works fine in most cases but if the win length and n fft are different padding n fft does not gurantee the first frame be centered i think win length is the right choice for the proper operation versions collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version libc version glibc python version default sep bit runtime python platform linux with is cuda available true cuda runtime version could not collect gpu models and configuration gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla gpu tesla nvidia driver version cudnn version probably one of the following usr lib linux gnu libcudnn so usr lib linux gnu libcudnn adv infer so usr lib linux gnu libcudnn adv train so usr lib linux gnu libcudnn cnn infer so usr lib linux gnu libcudnn cnn train so usr lib linux gnu libcudnn ops infer so usr lib linux gnu libcudnn ops train so hip runtime version n a miopen runtime version n a versions of relevant libraries numpy pytorch lightning torch torchaudio torchmetrics torchvision could not collect cc ezyang gchanan dylanbespalko mruberry lezcano nikitaved | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.