Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
113,115 | 4,543,414,948 | IssuesEvent | 2016-09-10 04:05:59 | codeforamerica/intake | https://api.github.com/repos/codeforamerica/intake | closed | Large PDF requests can break Heroku's 30 second limit | partner-SFC priority-high | This morning a user couldn't access an 18-page generated PDF. When they tried, the request timed out after 30 seconds, a hard stop imposed by Heroku. There is no way to increase this time limit for request timeouts on Heroku. Accessing the same pdf later did not result in an error.
I was able to recreate the error by refreshing the page while it was loading.
Here are options that could prevent this in the future:
* creating a separate background process that allows the web request to complete in a short amount of time.
* create PDFs ahead of time, and serve them as static files
* moving off of Heroku and allowing PDF requests to take longer. | 1.0 | Large PDF requests can break Heroku's 30 second limit - This morning a user couldn't access an 18-page generated PDF. When they tried, the request timed out after 30 seconds, a hard stop imposed by Heroku. There is no way to increase this time limit for request timeouts on Heroku. Accessing the same pdf later did not result in an error.
I was able to recreate the error by refreshing the page while it was loading.
Here are options that could prevent this in the future:
* creating a separate background process that allows the web request to complete in a short amount of time.
* create PDFs ahead of time, and serve them as static files
* moving off of Heroku and allowing PDF requests to take longer. | priority | large pdf requests can break heroku s second limit this morning a user couldn t access an page generated pdf when they tried the request timed out after seconds a hard stop imposed by heroku there is no way to increase this time limit for request timeouts on heroku accessing the same pdf later did not result in an error i was able to recreate the error by refreshing the page while it was loading here are options that could prevent this in the future creating a separate background process that allows the web request to complete in a short amount of time create pdfs ahead of time and serve them as static files moving off of heroku and allowing pdf requests to take longer | 1 |
469,740 | 13,524,751,035 | IssuesEvent | 2020-09-15 12:02:48 | gnosis/conditional-tokens-explorer | https://api.github.com/repos/gnosis/conditional-tokens-explorer | opened | Partition section - Edit partition bugs | High priority bug | Related to #237 , #66 , #67
1. Empty rows are displayed in the section when drag and drop all the outcomes and then click on the reset button (see the video)
https://drive.google.com/file/d/1DJeNOQ5emlvRIEjTfz2HK1-7UDe_p-ta/view
2. Empty rows are displayed in the section when replace an outcome into Outcomes section by **clicking on the outcome circle** and then click on the reset button
3. The system allows to save a partition with only one collection in it: repeat steps 1 or 2 of the current issue, an then, then there is only one outcome and 1 empty line, click on the Save button (see the video). The transaction for positions creating will be finished with failure
https://drive.google.com/file/d/1-hYTJCH9QatnUD1HNiUX7jaDeXPk9PkO/view
4. Numbers do not fit the circle area when a condition contains outcomes more than 100 outcomes (3-digits outcomes)

5. The edited partitioning before login is not saved after a user logs in (see the video) Note: the issue with amount field on the video is described here: #252
https://drive.google.com/file/d/1HscxRWRKrSb8_e1AOgjIt_QYhaMJo04-/view
6. Bun icon is still displayed when drag and drop a position (should not be displayed according to the mock-up (https://zpl.io/aR5Z76K)
7. a Circle in the New collection prevoew field should be in red when hover a mouse on it according to the mock-up (https://zpl.io/aR5Z76K)

8. No horizontal lines should be displayed in the partition when it contains a lot of outcomes, placed one after another according to the mock-up (https://zpl.io/aR5Z76K)
9. 'Move' and 'Bin' icons should be top-aligned when outcomes exceed 1 line (mock-up https://zpl.io/aR5Z76K)

10. New collection preview should contain the lines between the outcomes according to the mock-up https://zpl.io/V039zl9

11. There is no confirmation before removing New Collection Preview. It should be according to the mock-up https://zpl.io/a3WXQZx


| 1.0 | Partition section - Edit partition bugs - Related to #237 , #66 , #67
1. Empty rows are displayed in the section when drag and drop all the outcomes and then click on the reset button (see the video)
https://drive.google.com/file/d/1DJeNOQ5emlvRIEjTfz2HK1-7UDe_p-ta/view
2. Empty rows are displayed in the section when replace an outcome into Outcomes section by **clicking on the outcome circle** and then click on the reset button
3. The system allows to save a partition with only one collection in it: repeat steps 1 or 2 of the current issue, an then, then there is only one outcome and 1 empty line, click on the Save button (see the video). The transaction for positions creating will be finished with failure
https://drive.google.com/file/d/1-hYTJCH9QatnUD1HNiUX7jaDeXPk9PkO/view
4. Numbers do not fit the circle area when a condition contains outcomes more than 100 outcomes (3-digits outcomes)

5. The edited partitioning before login is not saved after a user logs in (see the video) Note: the issue with amount field on the video is described here: #252
https://drive.google.com/file/d/1HscxRWRKrSb8_e1AOgjIt_QYhaMJo04-/view
6. Bun icon is still displayed when drag and drop a position (should not be displayed according to the mock-up (https://zpl.io/aR5Z76K)
7. a Circle in the New collection prevoew field should be in red when hover a mouse on it according to the mock-up (https://zpl.io/aR5Z76K)

8. No horizontal lines should be displayed in the partition when it contains a lot of outcomes, placed one after another according to the mock-up (https://zpl.io/aR5Z76K)
9. 'Move' and 'Bin' icons should be top-aligned when outcomes exceed 1 line (mock-up https://zpl.io/aR5Z76K)

10. New collection preview should contain the lines between the outcomes according to the mock-up https://zpl.io/V039zl9

11. There is no confirmation before removing New Collection Preview. It should be according to the mock-up https://zpl.io/a3WXQZx


| priority | partition section edit partition bugs related to empty rows are displayed in the section when drag and drop all the outcomes and then click on the reset button see the video empty rows are displayed in the section when replace an outcome into outcomes section by clicking on the outcome circle and then click on the reset button the system allows to save a partition with only one collection in it repeat steps or of the current issue an then then there is only one outcome and empty line click on the save button see the video the transaction for positions creating will be finished with failure numbers do not fit the circle area when a condition contains outcomes more than outcomes digits outcomes the edited partitioning before login is not saved after a user logs in see the video note the issue with amount field on the video is described here bun icon is still displayed when drag and drop a position should not be displayed according to the mock up a circle in the new collection prevoew field should be in red when hover a mouse on it according to the mock up no horizontal lines should be displayed in the partition when it contains a lot of outcomes placed one after another according to the mock up move and bin icons should be top aligned when outcomes exceed line mock up new collection preview should contain the lines between the outcomes according to the mock up there is no confirmation before removing new collection preview it should be according to the mock up | 1 |
179,389 | 6,624,619,396 | IssuesEvent | 2017-09-22 12:30:41 | YetiForceCompany/YetiForceCRM | https://api.github.com/repos/YetiForceCompany/YetiForceCRM | closed | Błąd skrzynki pocztowej po włączeniu podglądu wiadomości | Category::Bug Subcategory::HighPriority | Po zaznaczeniu w ustawieniach poczty Pokaż podgląd wiadomości w folderze Odebrane w skrzynce pocztowej nie wyświetlają się żadne wiadomości. Wyłączenie podglądu przywraca widok wiadomości.
Błąd wywołany w YetiForce 4.0.0. w wersji testowej zamieszczonej na stronie yetiforce.com
<!--- Before you create a new issue, please check out our [manual] (https://yetiforce.com/en/github/issues/126-issues.html) --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug. Descriptions can be provided in English or Polish (remember to add [PL] for Polish in the title). -->
#### Actual Behavior
<!--- Describe the result -->
#### Expected Behavior
<!--- Describe what you would want the result to be -->
#### How to trigger the error
<!--- If possible, please make a video using [ScreenToGif] (https://screentogif.codeplex.com/) or any other program used for recording actions from your desktop. -->
1.
2.
3.
#### Your Environment
<!---Describe the environment -->
* YetiForce Version used:
* Browser name and version:
* Environment name and version:
* Operating System and version:
<!--- Please check on your issue from time to time, in case we have questions or need some extra information. --->
| 1.0 | Błąd skrzynki pocztowej po włączeniu podglądu wiadomości - Po zaznaczeniu w ustawieniach poczty Pokaż podgląd wiadomości w folderze Odebrane w skrzynce pocztowej nie wyświetlają się żadne wiadomości. Wyłączenie podglądu przywraca widok wiadomości.
Błąd wywołany w YetiForce 4.0.0. w wersji testowej zamieszczonej na stronie yetiforce.com
<!--- Before you create a new issue, please check out our [manual] (https://yetiforce.com/en/github/issues/126-issues.html) --->
#### Issue
<!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug. Descriptions can be provided in English or Polish (remember to add [PL] for Polish in the title). -->
#### Actual Behavior
<!--- Describe the result -->
#### Expected Behavior
<!--- Describe what you would want the result to be -->
#### How to trigger the error
<!--- If possible, please make a video using [ScreenToGif] (https://screentogif.codeplex.com/) or any other program used for recording actions from your desktop. -->
1.
2.
3.
#### Your Environment
<!---Describe the environment -->
* YetiForce Version used:
* Browser name and version:
* Environment name and version:
* Operating System and version:
<!--- Please check on your issue from time to time, in case we have questions or need some extra information. --->
| priority | błąd skrzynki pocztowej po włączeniu podglądu wiadomości po zaznaczeniu w ustawieniach poczty pokaż podgląd wiadomości w folderze odebrane w skrzynce pocztowej nie wyświetlają się żadne wiadomości wyłączenie podglądu przywraca widok wiadomości błąd wywołany w yetiforce w wersji testowej zamieszczonej na stronie yetiforce com issue actual behavior expected behavior how to trigger the error your environment yetiforce version used browser name and version environment name and version operating system and version | 1 |
707,570 | 24,309,969,361 | IssuesEvent | 2022-09-29 21:10:00 | ArctosDB/arctos | https://api.github.com/repos/ArctosDB/arctos | closed | Counties of Kenya | Priority-High (Needed for work) Function-Locality/Event/Georeferencing | ### First Step: Explain what geography needs created.
It looks like Kenya has undergone some revision. Some may be in Arctos, some not.
[Counties of Kenya](https://en.wikipedia.org/wiki/Counties_of_Kenya) needed for CSULB data migration include [Machakos](https://en.wikipedia.org/wiki/Machakos_County), [Makueni](https://en.wikipedia.org/wiki/Makueni_County) and [Nyeri](https://en.wikipedia.org/wiki/Nyeri_County)
We will respond with a CSV template, or request more information.
| 1.0 | Counties of Kenya - ### First Step: Explain what geography needs created.
It looks like Kenya has undergone some revision. Some may be in Arctos, some not.
[Counties of Kenya](https://en.wikipedia.org/wiki/Counties_of_Kenya) needed for CSULB data migration include [Machakos](https://en.wikipedia.org/wiki/Machakos_County), [Makueni](https://en.wikipedia.org/wiki/Makueni_County) and [Nyeri](https://en.wikipedia.org/wiki/Nyeri_County)
We will respond with a CSV template, or request more information.
| priority | counties of kenya first step explain what geography needs created it looks like kenya has undergone some revision some may be in arctos some not needed for csulb data migration include and we will respond with a csv template or request more information | 1 |
800,502 | 28,368,744,845 | IssuesEvent | 2023-04-12 15:28:06 | ArctosDB/arctos | https://api.github.com/repos/ArctosDB/arctos | closed | search results: add "fields" to FLAT | Priority-High (Needed for work) Enhancement Bug | **Is your feature request related to a problem? Please describe.**
https://github.com/ArctosDB/arctos/issues/6017 is causing pain for users and glitches in the system and I think cannot be ignored any longer even though the identifier discussions have stalled. I have to remove the many and mostly unused identifiers options from results ASAP in the name of stability.
I'd like to compensate by adding the things that are most used, and some loose ends that have been floating around. Here's what I think needs to happen, please let me know ASAP if anything else is regularly necessary in results.
* superorder (for @DerekSikes via https://github.com/ArctosDB/arctos/issues/6035)
* Preparator Number (see https://github.com/ArctosDB/arctos/issues/6031, I think this is stable)
* AF
* NK
above from https://github.com/ArctosDB/arctos/issues/5460 @jebrad did I miss anything?
**Describe what you're trying to accomplish**
Cache whatever's necessary to make Arctos performant, without making any new messes.
**Describe the solution you'd like**
Input from anyone who wants to include stuff in search results. (Perhaps excepting Attributes, which is a bit of a monster and needs a dedicated discussion.)
**Describe alternatives you've considered**
Things that a few users occasionally need can be quickly added dynamically, and there are download options for most of this; not adding something here does not mean it's not available.
**Additional context**
**Priority**
This is already smouldering, I'll make the changes and run the updates this weekend unless someone has a compelling reason not to.
Other things can be added at any time, but (depending on the specifics) it can take several days of CPU to complete; anything that can be done now should be done now for the sake of efficiency.
| 1.0 | search results: add "fields" to FLAT - **Is your feature request related to a problem? Please describe.**
https://github.com/ArctosDB/arctos/issues/6017 is causing pain for users and glitches in the system and I think cannot be ignored any longer even though the identifier discussions have stalled. I have to remove the many and mostly unused identifiers options from results ASAP in the name of stability.
I'd like to compensate by adding the things that are most used, and some loose ends that have been floating around. Here's what I think needs to happen, please let me know ASAP if anything else is regularly necessary in results.
* superorder (for @DerekSikes via https://github.com/ArctosDB/arctos/issues/6035)
* Preparator Number (see https://github.com/ArctosDB/arctos/issues/6031, I think this is stable)
* AF
* NK
above from https://github.com/ArctosDB/arctos/issues/5460 @jebrad did I miss anything?
**Describe what you're trying to accomplish**
Cache whatever's necessary to make Arctos performant, without making any new messes.
**Describe the solution you'd like**
Input from anyone who wants to include stuff in search results. (Perhaps excepting Attributes, which is a bit of a monster and needs a dedicated discussion.)
**Describe alternatives you've considered**
Things that a few users occasionally need can be quickly added dynamically, and there are download options for most of this; not adding something here does not mean it's not available.
**Additional context**
**Priority**
This is already smouldering, I'll make the changes and run the updates this weekend unless someone has a compelling reason not to.
Other things can be added at any time, but (depending on the specifics) it can take several days of CPU to complete; anything that can be done now should be done now for the sake of efficiency.
| priority | search results add fields to flat is your feature request related to a problem please describe is causing pain for users and glitches in the system and i think cannot be ignored any longer even though the identifier discussions have stalled i have to remove the many and mostly unused identifiers options from results asap in the name of stability i d like to compensate by adding the things that are most used and some loose ends that have been floating around here s what i think needs to happen please let me know asap if anything else is regularly necessary in results superorder for dereksikes via preparator number see i think this is stable af nk above from jebrad did i miss anything describe what you re trying to accomplish cache whatever s necessary to make arctos performant without making any new messes describe the solution you d like input from anyone who wants to include stuff in search results perhaps excepting attributes which is a bit of a monster and needs a dedicated discussion describe alternatives you ve considered things that a few users occasionally need can be quickly added dynamically and there are download options for most of this not adding something here does not mean it s not available additional context priority this is already smouldering i ll make the changes and run the updates this weekend unless someone has a compelling reason not to other things can be added at any time but depending on the specifics it can take several days of cpu to complete anything that can be done now should be done now for the sake of efficiency | 1 |
487,021 | 14,017,938,687 | IssuesEvent | 2020-10-29 16:13:35 | wazuh/wazuh-documentation | https://api.github.com/repos/wazuh/wazuh-documentation | opened | Imrpove the OD upgrade section | priority: highest type: refactor | Hello team! This issue aims to improve the upgrade guide section for Open Distro.
We may face the situation where a user has the Elastic or Open Distro repository and this may lead to unwanted upgrades and broken installations.
We should disable those repositories in order to prevent accidental upgrades.
Regards,
David | 1.0 | Imrpove the OD upgrade section - Hello team! This issue aims to improve the upgrade guide section for Open Distro.
We may face the situation where a user has the Elastic or Open Distro repository and this may lead to unwanted upgrades and broken installations.
We should disable those repositories in order to prevent accidental upgrades.
Regards,
David | priority | imrpove the od upgrade section hello team this issue aims to improve the upgrade guide section for open distro we may face the situation where a user has the elastic or open distro repository and this may lead to unwanted upgrades and broken installations we should disable those repositories in order to prevent accidental upgrades regards david | 1 |
392,715 | 11,594,906,818 | IssuesEvent | 2020-02-24 16:05:38 | malaquiasdev/mamae-eu-quero-service | https://api.github.com/repos/malaquiasdev/mamae-eu-quero-service | closed | Some properties of Baby model should use required true | 🙇 good first issue 🚧 Status: WIP 🛠 Type: Minor 🤘 Priority: High | **Is your feature request related to a problem? Please describe.**
The properties below should not be optional.
- [x] motherMail
- [x] sexy
- [x] name
**Describe the solution you'd like**
We need put the schema property like below:
```
name: {
type: String,
required: true
}
```
After this we will need catch the error in the src/functions/client/create.js module.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| 1.0 | Some properties of Baby model should use required true - **Is your feature request related to a problem? Please describe.**
The properties below should not be optional.
- [x] motherMail
- [x] sexy
- [x] name
**Describe the solution you'd like**
We need put the schema property like below:
```
name: {
type: String,
required: true
}
```
After this we will need catch the error in the src/functions/client/create.js module.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
| priority | some properties of baby model should use required true is your feature request related to a problem please describe the properties below should not be optional mothermail sexy name describe the solution you d like we need put the schema property like below name type string required true after this we will need catch the error in the src functions client create js module describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here | 1 |
830,412 | 32,007,185,989 | IssuesEvent | 2023-09-21 15:30:24 | The-Aether-Team/The-Aether | https://api.github.com/repos/The-Aether-Team/The-Aether | opened | Feature: Active Moa Skin textures | priority/high status/in-progress feat/art | Need to produce Moa skin textures for active patreon pledges
- [ ] Ascentan Tier Moa skins
- [ ] Valkyrie Tier Moa skins
| 1.0 | Feature: Active Moa Skin textures - Need to produce Moa skin textures for active patreon pledges
- [ ] Ascentan Tier Moa skins
- [ ] Valkyrie Tier Moa skins
| priority | feature active moa skin textures need to produce moa skin textures for active patreon pledges ascentan tier moa skins valkyrie tier moa skins | 1 |
130,444 | 5,116,064,414 | IssuesEvent | 2017-01-07 00:22:39 | TheTyee/design-article.thetyee.ca | https://api.github.com/repos/TheTyee/design-article.thetyee.ca | closed | Homepage nav broken on ipad | Priority: High Status: Bug Type: CSS & presentation | Here's a screenshot from ipad mini Safair.
I can't reproduce this issue on desktop Safari.

| 1.0 | Homepage nav broken on ipad - Here's a screenshot from ipad mini Safair.
I can't reproduce this issue on desktop Safari.

| priority | homepage nav broken on ipad here s a screenshot from ipad mini safair i can t reproduce this issue on desktop safari | 1 |
783,070 | 27,517,486,341 | IssuesEvent | 2023-03-06 12:59:44 | AY2223S2-CS2113-T12-2/tp | https://api.github.com/repos/AY2223S2-CS2113-T12-2/tp | opened | Load and store component data from JSON files | priority.High | Add DataStorage class which loads and stores all data to be accessed | 1.0 | Load and store component data from JSON files - Add DataStorage class which loads and stores all data to be accessed | priority | load and store component data from json files add datastorage class which loads and stores all data to be accessed | 1 |
289,902 | 8,880,035,223 | IssuesEvent | 2019-01-14 03:05:22 | zulip/zulip | https://api.github.com/repos/zulip/zulip | closed | ldap: Add support for providing the "full name" field split across two attributes | area: authentication in progress priority: high | In some LDAP databases, they only have separate "first name" and "last name" fields; we should offer a way to do that (inspired by comments in https://github.com/zulip/zulip/issues/9710).
I think a reasonable approach for this is to support users setting the `first_name` and `last_name` keys in `AUTH_LDAP_ATTR_MAP`, and then just have some custom code similar to what we have in 5dd646f33fa154ea98df1366b53e8904e5adf762 for combining together the two values in the two `get_or_build_user` functions. | 1.0 | ldap: Add support for providing the "full name" field split across two attributes - In some LDAP databases, they only have separate "first name" and "last name" fields; we should offer a way to do that (inspired by comments in https://github.com/zulip/zulip/issues/9710).
I think a reasonable approach for this is to support users setting the `first_name` and `last_name` keys in `AUTH_LDAP_ATTR_MAP`, and then just have some custom code similar to what we have in 5dd646f33fa154ea98df1366b53e8904e5adf762 for combining together the two values in the two `get_or_build_user` functions. | priority | ldap add support for providing the full name field split across two attributes in some ldap databases they only have separate first name and last name fields we should offer a way to do that inspired by comments in i think a reasonable approach for this is to support users setting the first name and last name keys in auth ldap attr map and then just have some custom code similar to what we have in for combining together the two values in the two get or build user functions | 1 |
425,045 | 12,334,793,745 | IssuesEvent | 2020-05-14 10:49:37 | EBISPOT/goci | https://api.github.com/repos/EBISPOT/goci | opened | Submission template validator should validate that study tags provided by submitter are unique within the submission | Blocker Priority: High Type: Enhancement | Submission template validator should validate that study tags provided by submitter are unique within the submission. As study tags are used to identify the study they must be unique within a particular submission. | 1.0 | Submission template validator should validate that study tags provided by submitter are unique within the submission - Submission template validator should validate that study tags provided by submitter are unique within the submission. As study tags are used to identify the study they must be unique within a particular submission. | priority | submission template validator should validate that study tags provided by submitter are unique within the submission submission template validator should validate that study tags provided by submitter are unique within the submission as study tags are used to identify the study they must be unique within a particular submission | 1 |
107,505 | 4,309,869,194 | IssuesEvent | 2016-07-21 17:22:33 | isawnyu/isaw.web | https://api.github.com/repos/isawnyu/isaw.web | closed | calendar buttons and grid need tweaks (from isaw.theme) | bug deploy high priority style | Copying this issue over from isaw.theme issue #46:
In calendar view, the Month and Week buttons look unfinished (missing a side border), and the calendar grid is misaligned in Week view.
I think this may be fixed, but I can't tell at the moment (can't see events). I've copied this issue here to remind myself to check. | 1.0 | calendar buttons and grid need tweaks (from isaw.theme) - Copying this issue over from isaw.theme issue #46:
In calendar view, the Month and Week buttons look unfinished (missing a side border), and the calendar grid is misaligned in Week view.
I think this may be fixed, but I can't tell at the moment (can't see events). I've copied this issue here to remind myself to check. | priority | calendar buttons and grid need tweaks from isaw theme copying this issue over from isaw theme issue in calendar view the month and week buttons look unfinished missing a side border and the calendar grid is misaligned in week view i think this may be fixed but i can t tell at the moment can t see events i ve copied this issue here to remind myself to check | 1 |
492,259 | 14,199,317,794 | IssuesEvent | 2020-11-16 01:59:36 | nhn/tui.grid | https://api.github.com/repos/nhn/tui.grid | closed | "Cannot read property 'rowSpanMap' of undefined" occurs when deleting rows | Bug Priority: High | **Describe the bug**
Error occurs when the raw data is recalled and the row is deleted by recalling the data back to ajax.
**To Reproduce**
Steps to reproduce the behavior:
1. Create Grid (Initialize Data)
2. Recall data using ajax (approximately 700 results)
3. Select 400th from first row (click + shift used, Scroll down during this process)
4. tui-grid.js:2388 Uncaused TypeError: Unable to read property 'rowSpanMap' for undefined (Occurs when the scroll position is more than half of the total position.)
5. Recall data using ajax (approximately 700 results)
6. tui-grid.js:2777 Uncaught TypeError: Cannot read property 'forEach' of undefined
**Desktop (please complete the following information):**
- OS: Window10
- Browser: Chrome
- Version 85.0.4183.121
| 1.0 | "Cannot read property 'rowSpanMap' of undefined" occurs when deleting rows - **Describe the bug**
Error occurs when the raw data is recalled and the row is deleted by recalling the data back to ajax.
**To Reproduce**
Steps to reproduce the behavior:
1. Create Grid (Initialize Data)
2. Recall data using ajax (approximately 700 results)
3. Select 400th from first row (click + shift used, Scroll down during this process)
4. tui-grid.js:2388 Uncaused TypeError: Unable to read property 'rowSpanMap' for undefined (Occurs when the scroll position is more than half of the total position.)
5. Recall data using ajax (approximately 700 results)
6. tui-grid.js:2777 Uncaught TypeError: Cannot read property 'forEach' of undefined
**Desktop (please complete the following information):**
- OS: Window10
- Browser: Chrome
- Version 85.0.4183.121
| priority | cannot read property rowspanmap of undefined occurs when deleting rows describe the bug error occurs when the raw data is recalled and the row is deleted by recalling the data back to ajax to reproduce steps to reproduce the behavior create grid initialize data recall data using ajax approximately results select from first row click shift used scroll down during this process tui grid js uncaused typeerror unable to read property rowspanmap for undefined occurs when the scroll position is more than half of the total position recall data using ajax approximately results tui grid js uncaught typeerror cannot read property foreach of undefined desktop please complete the following information os browser chrome version | 1 |
221,767 | 7,395,968,817 | IssuesEvent | 2018-03-18 05:53:35 | chimano/SOEN341_SA2 | https://api.github.com/repos/chimano/SOEN341_SA2 | closed | User's should not be able to vote on their answer | bug priority: high | The system should not allow users to vote on their own question and answer. (~30 minutes) | 1.0 | User's should not be able to vote on their answer - The system should not allow users to vote on their own question and answer. (~30 minutes) | priority | user s should not be able to vote on their answer the system should not allow users to vote on their own question and answer minutes | 1 |
492,998 | 14,224,225,541 | IssuesEvent | 2020-11-17 19:21:15 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | opened | [Story Player] Improve player build/layout strategy | P1: High Priority Type: Bug WG: stories | Opening most pages using a player ([example](https://www.greenschemetv.net/gluten-free-cream-cheese-brownies/)) with a very fast computer/network speed, I'd expect the story to be loaded before it gets in the viewport. We should consider making our build/layout strategy more aggressive.
cc @ampproject/wg-stories | 1.0 | [Story Player] Improve player build/layout strategy - Opening most pages using a player ([example](https://www.greenschemetv.net/gluten-free-cream-cheese-brownies/)) with a very fast computer/network speed, I'd expect the story to be loaded before it gets in the viewport. We should consider making our build/layout strategy more aggressive.
cc @ampproject/wg-stories | priority | improve player build layout strategy opening most pages using a player with a very fast computer network speed i d expect the story to be loaded before it gets in the viewport we should consider making our build layout strategy more aggressive cc ampproject wg stories | 1 |
140,601 | 5,412,974,039 | IssuesEvent | 2017-03-01 15:45:00 | gwt-plugins/gwt-eclipse-plugin | https://api.github.com/repos/gwt-plugins/gwt-eclipse-plugin | closed | GWT plug-ins check for updates and do analytics ping whenever a Java build occurs (on a GWT Project) | enhancement High Priority | I do not think it is sensible for the GWT plug-ins to check for GWT plug-in updates on every java compile of any project with a GWT nature. This can be detrimental to performance.
I know that update checking can be turned off in the preferences, but perhaps just checking on Eclipse start-up or something (or once a day or whatever) would be more sensible (just a thought)?
Much more annoying than the updates is the compilation analytics ping request as there doesn't seem to be a way of turning this off, and (again) it happens on every Java build of a GWT natured project!
It should at least respect the general error reporting preferences (e.g.org.eclipse.epp.logging.aeri.ide/sendMode=NEVER) and enable me to opt out of sending information back etc. Also, it does of course fill up my logs with (what to me at least is) not useful information (although it is not alone in that of course).
`/**
* A compilation participant that is used to trigger an update check of the GWT Plugin's feature
* whenever a Java build is triggered on a GWT project.
*/
public class UpdateTriggerCompilationParticipant extends CompilationParticipant {
@Override
public boolean isActive(IJavaProject project) {
if (!project.exists()) {
return false;
}
if (GWTNature.isGWTProject(project.getProject())) {
GdtExtPlugin.getFeatureUpdateManager().checkForUpdates();
GdtExtPlugin.getAnalyticsPingManager().sendCompilationPing();
return true;
} else {
return false;
}
}
}
` | 1.0 | GWT plug-ins check for updates and do analytics ping whenever a Java build occurs (on a GWT Project) - I do not think it is sensible for the GWT plug-ins to check for GWT plug-in updates on every java compile of any project with a GWT nature. This can be detrimental to performance.
I know that update checking can be turned off in the preferences, but perhaps just checking on Eclipse start-up or something (or once a day or whatever) would be more sensible (just a thought)?
Much more annoying than the updates is the compilation analytics ping request as there doesn't seem to be a way of turning this off, and (again) it happens on every Java build of a GWT natured project!
It should at least respect the general error reporting preferences (e.g.org.eclipse.epp.logging.aeri.ide/sendMode=NEVER) and enable me to opt out of sending information back etc. Also, it does of course fill up my logs with (what to me at least is) not useful information (although it is not alone in that of course).
`/**
* A compilation participant that is used to trigger an update check of the GWT Plugin's feature
* whenever a Java build is triggered on a GWT project.
*/
public class UpdateTriggerCompilationParticipant extends CompilationParticipant {
@Override
public boolean isActive(IJavaProject project) {
if (!project.exists()) {
return false;
}
if (GWTNature.isGWTProject(project.getProject())) {
GdtExtPlugin.getFeatureUpdateManager().checkForUpdates();
GdtExtPlugin.getAnalyticsPingManager().sendCompilationPing();
return true;
} else {
return false;
}
}
}
` | priority | gwt plug ins check for updates and do analytics ping whenever a java build occurs on a gwt project i do not think it is sensible for the gwt plug ins to check for gwt plug in updates on every java compile of any project with a gwt nature this can be detrimental to performance i know that update checking can be turned off in the preferences but perhaps just checking on eclipse start up or something or once a day or whatever would be more sensible just a thought much more annoying than the updates is the compilation analytics ping request as there doesn t seem to be a way of turning this off and again it happens on every java build of a gwt natured project it should at least respect the general error reporting preferences e g org eclipse epp logging aeri ide sendmode never and enable me to opt out of sending information back etc also it does of course fill up my logs with what to me at least is not useful information although it is not alone in that of course a compilation participant that is used to trigger an update check of the gwt plugin s feature whenever a java build is triggered on a gwt project public class updatetriggercompilationparticipant extends compilationparticipant override public boolean isactive ijavaproject project if project exists return false if gwtnature isgwtproject project getproject gdtextplugin getfeatureupdatemanager checkforupdates gdtextplugin getanalyticspingmanager sendcompilationping return true else return false | 1 |
118,186 | 4,732,888,962 | IssuesEvent | 2016-10-19 09:21:56 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | closed | Simpler way to remove existing layers from the layer tree | enhancement pending review Priority: High task | The standard way for removing layers from the layer tree is too complex, see image below:

I would like to have something similar to what we have done for SIRA, se below:


| 1.0 | Simpler way to remove existing layers from the layer tree - The standard way for removing layers from the layer tree is too complex, see image below:

I would like to have something similar to what we have done for SIRA, se below:


| priority | simpler way to remove existing layers from the layer tree the standard way for removing layers from the layer tree is too complex see image below i would like to have something similar to what we have done for sira se below | 1 |
519,095 | 15,044,078,814 | IssuesEvent | 2021-02-03 02:08:37 | Psychoanalytic-Electronic-Publishing/OpenPubArchive-Content-Server | https://api.github.com/repos/Psychoanalytic-Electronic-Publishing/OpenPubArchive-Content-Server | opened | DevOps Build Process Optimizations | optimization priority-high | Today may have been a bit unusual, with an error from solr seeming to repeatedly happen, while meanwhile the Development solr database appears to be updated, as does the Development RDS database. Maybe I'm missing something.
It's not acceptable that a full build is still running, 20 hours later. I ran the entire process here in about 4 hours on my PC. But the worst part of this is that if I were really building a production set with a deadline, this would really extend the time...especially if I had to run it more than once. Even incremental builds don't solve the problem...I don't have timing data, but as I recall the last time I manually ran one, the incremental process takes many hours as well...when in fact, it may only be a handful of files being loaded.
We need to do better.
Some ideas for improving build performance:
1) In a rebuild scenario, once we see the crossing point of the forward and reverse runs, we should stop or at least let then ext step (build images) proceed...that takes far too long on AWS to finish even though the work is actually done. And the fact that the process can't "complete" until each direction has looked at all the files, makes me wait hours more.
So I've added a flag: --halfway which can be used for both the forward and reverse processes--and they will quit after each processing half of the files. Hopefully, that will help.
2) We could add more processes to the full build...my 4 hour timing is running forward, reverse, and targeting some of the bigger journals separately (only works when resetting the cores, so it's not set to "all" but rather checks if each file is in the database, so overlap with two main processes isn't a problem, and they just zip past when they are already processed (of course it doesn't seem to zip on AWS). I sometimes run as many as 6 or so processes here on my Intel I9 9900K 8-core PC to optimize the timing.
3) For smaller changes it may be a good idea to make another processing option...where I can make more targeted changes to development solr and development RDS quickly...and then generate what's needed for stage. I can run my builds remotely on those, and for smaller updates, I think it may run more quickly than the current update process. Then we'd have a script to create what's needed to create the AMI's to push to stage and eventually production (after testing). This could even run off of stage, as an option...so after checking stage, if there are some issues, it can be dealt with there, then the AMIs generated and pushed to production.
Though we can discuss why incremental also seems to takes so long before going down that road. One thing that's odd--I noticed when running incremental builds, it's falsely detecting changes and processing some items, where it shouldn't have to. Not sure why that happens--it doesn't happen when run locally. Perhaps something about comparing times between S3 and Solr?
I do have another ace up my sleeve to make it faster...but I'm hesitant because it involves a loss of an interesting search feature (even though the client isn't using it) and requires changes to the server as well. Still I'm considering it if we can't speed things up. I'm open to ideas. I am ok with a 12 hour total for a full build--one I can start in the evening and have it ready in the morning. I'd prefer faster but I can live with that. Incremental builds should take very little time, anywhere from 15 minutes at most for a few new files to an hour or two, for example, to rerun the whole PEPCurrent folder (<7000 files).
But I don't think a full build that takes over 20 hours, like the current one is taking, works for starters. Maybe that's an anomaly, but it's certainly holding me up today...and I don't want to stop it to try again, because it seems to take 14 or more hours to rebuild in any case.
| 1.0 | DevOps Build Process Optimizations - Today may have been a bit unusual, with an error from solr seeming to repeatedly happen, while meanwhile the Development solr database appears to be updated, as does the Development RDS database. Maybe I'm missing something.
It's not acceptable that a full build is still running, 20 hours later. I ran the entire process here in about 4 hours on my PC. But the worst part of this is that if I were really building a production set with a deadline, this would really extend the time...especially if I had to run it more than once. Even incremental builds don't solve the problem...I don't have timing data, but as I recall the last time I manually ran one, the incremental process takes many hours as well...when in fact, it may only be a handful of files being loaded.
We need to do better.
Some ideas for improving build performance:
1) In a rebuild scenario, once we see the crossing point of the forward and reverse runs, we should stop or at least let then ext step (build images) proceed...that takes far too long on AWS to finish even though the work is actually done. And the fact that the process can't "complete" until each direction has looked at all the files, makes me wait hours more.
So I've added a flag: --halfway which can be used for both the forward and reverse processes--and they will quit after each processing half of the files. Hopefully, that will help.
2) We could add more processes to the full build...my 4 hour timing is running forward, reverse, and targeting some of the bigger journals separately (only works when resetting the cores, so it's not set to "all" but rather checks if each file is in the database, so overlap with two main processes isn't a problem, and they just zip past when they are already processed (of course it doesn't seem to zip on AWS). I sometimes run as many as 6 or so processes here on my Intel I9 9900K 8-core PC to optimize the timing.
3) For smaller changes it may be a good idea to make another processing option...where I can make more targeted changes to development solr and development RDS quickly...and then generate what's needed for stage. I can run my builds remotely on those, and for smaller updates, I think it may run more quickly than the current update process. Then we'd have a script to create what's needed to create the AMI's to push to stage and eventually production (after testing). This could even run off of stage, as an option...so after checking stage, if there are some issues, it can be dealt with there, then the AMIs generated and pushed to production.
Though we can discuss why incremental also seems to takes so long before going down that road. One thing that's odd--I noticed when running incremental builds, it's falsely detecting changes and processing some items, where it shouldn't have to. Not sure why that happens--it doesn't happen when run locally. Perhaps something about comparing times between S3 and Solr?
I do have another ace up my sleeve to make it faster...but I'm hesitant because it involves a loss of an interesting search feature (even though the client isn't using it) and requires changes to the server as well. Still I'm considering it if we can't speed things up. I'm open to ideas. I am ok with a 12 hour total for a full build--one I can start in the evening and have it ready in the morning. I'd prefer faster but I can live with that. Incremental builds should take very little time, anywhere from 15 minutes at most for a few new files to an hour or two, for example, to rerun the whole PEPCurrent folder (<7000 files).
But I don't think a full build that takes over 20 hours, like the current one is taking, works for starters. Maybe that's an anomaly, but it's certainly holding me up today...and I don't want to stop it to try again, because it seems to take 14 or more hours to rebuild in any case.
| priority | devops build process optimizations today may have been a bit unusual with an error from solr seeming to repeatedly happen while meanwhile the development solr database appears to be updated as does the development rds database maybe i m missing something it s not acceptable that a full build is still running hours later i ran the entire process here in about hours on my pc but the worst part of this is that if i were really building a production set with a deadline this would really extend the time especially if i had to run it more than once even incremental builds don t solve the problem i don t have timing data but as i recall the last time i manually ran one the incremental process takes many hours as well when in fact it may only be a handful of files being loaded we need to do better some ideas for improving build performance in a rebuild scenario once we see the crossing point of the forward and reverse runs we should stop or at least let then ext step build images proceed that takes far too long on aws to finish even though the work is actually done and the fact that the process can t complete until each direction has looked at all the files makes me wait hours more so i ve added a flag halfway which can be used for both the forward and reverse processes and they will quit after each processing half of the files hopefully that will help we could add more processes to the full build my hour timing is running forward reverse and targeting some of the bigger journals separately only works when resetting the cores so it s not set to all but rather checks if each file is in the database so overlap with two main processes isn t a problem and they just zip past when they are already processed of course it doesn t seem to zip on aws i sometimes run as many as or so processes here on my intel core pc to optimize the timing for smaller changes it may be a good idea to make another processing option where i can make more targeted changes to development solr and development rds quickly and then generate what s needed for stage i can run my builds remotely on those and for smaller updates i think it may run more quickly than the current update process then we d have a script to create what s needed to create the ami s to push to stage and eventually production after testing this could even run off of stage as an option so after checking stage if there are some issues it can be dealt with there then the amis generated and pushed to production though we can discuss why incremental also seems to takes so long before going down that road one thing that s odd i noticed when running incremental builds it s falsely detecting changes and processing some items where it shouldn t have to not sure why that happens it doesn t happen when run locally perhaps something about comparing times between and solr i do have another ace up my sleeve to make it faster but i m hesitant because it involves a loss of an interesting search feature even though the client isn t using it and requires changes to the server as well still i m considering it if we can t speed things up i m open to ideas i am ok with a hour total for a full build one i can start in the evening and have it ready in the morning i d prefer faster but i can live with that incremental builds should take very little time anywhere from minutes at most for a few new files to an hour or two for example to rerun the whole pepcurrent folder files but i don t think a full build that takes over hours like the current one is taking works for starters maybe that s an anomaly but it s certainly holding me up today and i don t want to stop it to try again because it seems to take or more hours to rebuild in any case | 1 |
280,948 | 8,688,557,861 | IssuesEvent | 2018-12-03 16:23:46 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | opened | Indirect - ConvFit crash when selecting Fit from different buttons | Component: Indirect Inelastic Misc: Bug Priority: High | **Found by:** Unscripted testing
### Expected behavior
No crash - the relevant buttons should be disabled
### Steps to reproduce the behavior
1. `Interfaces`->`Indirect`->`Data Analysis`->`ConvFit` Tab
3. Load the data below
3. Fit Type is `Teixeira Water`
4. Click the 'Fit' combobox above the FitPropertyBrowser and then click 'Sequential Fit'
5. Quickly click 'Run' while it is still fitting
crash!
**Data Files**
[irs26176_graphite002_red.zip](https://github.com/mantidproject/mantid/files/2640253/irs26176_graphite002_red.zip)
[irs26173_graphite002_res.zip](https://github.com/mantidproject/mantid/files/2640254/irs26173_graphite002_res.zip)
### Platforms affected
All | 1.0 | Indirect - ConvFit crash when selecting Fit from different buttons - **Found by:** Unscripted testing
### Expected behavior
No crash - the relevant buttons should be disabled
### Steps to reproduce the behavior
1. `Interfaces`->`Indirect`->`Data Analysis`->`ConvFit` Tab
3. Load the data below
3. Fit Type is `Teixeira Water`
4. Click the 'Fit' combobox above the FitPropertyBrowser and then click 'Sequential Fit'
5. Quickly click 'Run' while it is still fitting
crash!
**Data Files**
[irs26176_graphite002_red.zip](https://github.com/mantidproject/mantid/files/2640253/irs26176_graphite002_red.zip)
[irs26173_graphite002_res.zip](https://github.com/mantidproject/mantid/files/2640254/irs26173_graphite002_res.zip)
### Platforms affected
All | priority | indirect convfit crash when selecting fit from different buttons found by unscripted testing expected behavior no crash the relevant buttons should be disabled steps to reproduce the behavior interfaces indirect data analysis convfit tab load the data below fit type is teixeira water click the fit combobox above the fitpropertybrowser and then click sequential fit quickly click run while it is still fitting crash data files platforms affected all | 1 |
707,147 | 24,297,091,096 | IssuesEvent | 2022-09-29 10:59:00 | bryntum/support | https://api.github.com/repos/bryntum/support | closed | Task editor in Scheduler Pro gets empty when opened too fast after closing | bug resolved high-priority OEM | Reproducible on timeline demo:
1. select event
2. press Enter to open the editor
3. press Escape to close and immediately Enter once again
Editor fields get empty. This is much easier to reproduce on slower environnments like Salesforce

| 1.0 | Task editor in Scheduler Pro gets empty when opened too fast after closing - Reproducible on timeline demo:
1. select event
2. press Enter to open the editor
3. press Escape to close and immediately Enter once again
Editor fields get empty. This is much easier to reproduce on slower environnments like Salesforce

| priority | task editor in scheduler pro gets empty when opened too fast after closing reproducible on timeline demo select event press enter to open the editor press escape to close and immediately enter once again editor fields get empty this is much easier to reproduce on slower environnments like salesforce | 1 |
201,479 | 7,032,063,673 | IssuesEvent | 2017-12-26 23:31:33 | woborschilde/fluentlogin | https://api.github.com/repos/woborschilde/fluentlogin | closed | User App Login | kind: feature request priority: high status: done time span: long-term | Parts:
- Login
- Logout
- Register
- Password Lost
_This issue is archived and was originally created on 03 Aug 2017._ | 1.0 | User App Login - Parts:
- Login
- Logout
- Register
- Password Lost
_This issue is archived and was originally created on 03 Aug 2017._ | priority | user app login parts login logout register password lost this issue is archived and was originally created on aug | 1 |
748,677 | 26,132,664,411 | IssuesEvent | 2022-12-29 07:55:55 | NomicFoundation/hardhat | https://api.github.com/repos/NomicFoundation/hardhat | closed | `estimateGas` issue with hardhat network when autoMine is off | type:bug priority:high | We are having some issue with `estimateGas` when autoMine is off.
Hardhat version: version "2.6.1"
hardhat.config.js
```
localhost: {
hardfork: "istanbul",
},
hardhat: {
hardfork: "istanbul",
},
```
The `gasLimit` of the `hardhat` network is greatly larger than the `localhost` one, causing the transaction being dropped by the next block due to the block gasLimit constraint. The next block being mined can only hold 2 transactions.
The transaction data when sending to a `localhost` network:
```
{
hash: '0x72d6362e91291c311012d9cef61b9c9f3ddcaa52b8706e2f0286550cc3a51aa6',
type: null,
accessList: null,
blockHash: null,
blockNumber: null,
transactionIndex: null,
confirmations: 0,
from: '0x3C44CdDdB6a900fa2b585dd299e03d12FA4293BC',
gasPrice: BigNumber { _hex: '0x01dcd65000', _isBigNumber: true },
gasLimit: BigNumber { _hex: '0x0ea7fd', _isBigNumber: true },
to: '0x9A9f2CCfdE556A7E9Ff0848998Aa4a0CFD8863AE',
value: BigNumber { _hex: '0x00', _isBigNumber: true },
nonce: 12,
data: '0x42e77284000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000010000000000000000000000003c44cdddb6a900fa2b585dd299e03d12fa4293bc',
r: '0xf24fb81b34cffd9463e22786632605125630dae1d9042d6b6def669353c9ea15',
s: '0x0c77da00a7191abe556898036b6fb1c849107e39be70e2244943e5239f19730f',
v: 62710,
creates: null,
chainId: 31337,
wait: [Function (anonymous)]
}
```
The transaction data when sending to the default `hardhat` network:
```
{
hash: '0x5dce93e43d33274b720399bb9e659b042990b4556980db79c0686affd5607bb8',
type: null,
accessList: null,
blockHash: null,
blockNumber: null,
transactionIndex: null,
confirmations: 0,
from: '0x3C44CdDdB6a900fa2b585dd299e03d12FA4293BC',
gasPrice: BigNumber { _hex: '0x01dcd65000', _isBigNumber: true },
gasLimit: BigNumber { _hex: '0x01badb18', _isBigNumber: true },
to: '0x9A9f2CCfdE556A7E9Ff0848998Aa4a0CFD8863AE',
value: BigNumber { _hex: '0x00', _isBigNumber: true },
nonce: 12,
data: '0x42e77284000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000010000000000000000000000003c44cdddb6a900fa2b585dd299e03d12fa4293bc',
r: '0xeb2a0c24521571805b031ef169984919de500112daee1a5008f05192580c7286',
s: '0x7d3f542d5ca41c628f4e807fa1bc3b148db62a32b8e9933ead16a9a00f340034',
v: 62710,
creates: null,
chainId: 31337,
wait: [Function (anonymous)]
}
```
| 1.0 | `estimateGas` issue with hardhat network when autoMine is off - We are having some issue with `estimateGas` when autoMine is off.
Hardhat version: version "2.6.1"
hardhat.config.js
```
localhost: {
hardfork: "istanbul",
},
hardhat: {
hardfork: "istanbul",
},
```
The `gasLimit` of the `hardhat` network is greatly larger than the `localhost` one, causing the transaction being dropped by the next block due to the block gasLimit constraint. The next block being mined can only hold 2 transactions.
The transaction data when sending to a `localhost` network:
```
{
hash: '0x72d6362e91291c311012d9cef61b9c9f3ddcaa52b8706e2f0286550cc3a51aa6',
type: null,
accessList: null,
blockHash: null,
blockNumber: null,
transactionIndex: null,
confirmations: 0,
from: '0x3C44CdDdB6a900fa2b585dd299e03d12FA4293BC',
gasPrice: BigNumber { _hex: '0x01dcd65000', _isBigNumber: true },
gasLimit: BigNumber { _hex: '0x0ea7fd', _isBigNumber: true },
to: '0x9A9f2CCfdE556A7E9Ff0848998Aa4a0CFD8863AE',
value: BigNumber { _hex: '0x00', _isBigNumber: true },
nonce: 12,
data: '0x42e77284000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000010000000000000000000000003c44cdddb6a900fa2b585dd299e03d12fa4293bc',
r: '0xf24fb81b34cffd9463e22786632605125630dae1d9042d6b6def669353c9ea15',
s: '0x0c77da00a7191abe556898036b6fb1c849107e39be70e2244943e5239f19730f',
v: 62710,
creates: null,
chainId: 31337,
wait: [Function (anonymous)]
}
```
The transaction data when sending to the default `hardhat` network:
```
{
hash: '0x5dce93e43d33274b720399bb9e659b042990b4556980db79c0686affd5607bb8',
type: null,
accessList: null,
blockHash: null,
blockNumber: null,
transactionIndex: null,
confirmations: 0,
from: '0x3C44CdDdB6a900fa2b585dd299e03d12FA4293BC',
gasPrice: BigNumber { _hex: '0x01dcd65000', _isBigNumber: true },
gasLimit: BigNumber { _hex: '0x01badb18', _isBigNumber: true },
to: '0x9A9f2CCfdE556A7E9Ff0848998Aa4a0CFD8863AE',
value: BigNumber { _hex: '0x00', _isBigNumber: true },
nonce: 12,
data: '0x42e77284000000000000000000000000000000000000000000000000000000000000002000000000000000000000000000000000000000000000000000000000000000010000000000000000000000003c44cdddb6a900fa2b585dd299e03d12fa4293bc',
r: '0xeb2a0c24521571805b031ef169984919de500112daee1a5008f05192580c7286',
s: '0x7d3f542d5ca41c628f4e807fa1bc3b148db62a32b8e9933ead16a9a00f340034',
v: 62710,
creates: null,
chainId: 31337,
wait: [Function (anonymous)]
}
```
| priority | estimategas issue with hardhat network when automine is off we are having some issue with estimategas when automine is off hardhat version version hardhat config js localhost hardfork istanbul hardhat hardfork istanbul the gaslimit of the hardhat network is greatly larger than the localhost one causing the transaction being dropped by the next block due to the block gaslimit constraint the next block being mined can only hold transactions the transaction data when sending to a localhost network hash type null accesslist null blockhash null blocknumber null transactionindex null confirmations from gasprice bignumber hex isbignumber true gaslimit bignumber hex isbignumber true to value bignumber hex isbignumber true nonce data r s v creates null chainid wait the transaction data when sending to the default hardhat network hash type null accesslist null blockhash null blocknumber null transactionindex null confirmations from gasprice bignumber hex isbignumber true gaslimit bignumber hex isbignumber true to value bignumber hex isbignumber true nonce data r s v creates null chainid wait | 1 |
266,619 | 8,372,759,590 | IssuesEvent | 2018-10-05 08:12:32 | IBM/watson-assistant-workbench | https://api.github.com/repos/IBM/watson-assistant-workbench | closed | Importing nodes reverse their order | Priority: high XML bug | Due to changes in #130 the order of imported nodes is reverted. | 1.0 | Importing nodes reverse their order - Due to changes in #130 the order of imported nodes is reverted. | priority | importing nodes reverse their order due to changes in the order of imported nodes is reverted | 1 |
561,861 | 16,626,085,632 | IssuesEvent | 2021-06-03 09:43:41 | nhost/hasura-backend-plus | https://api.github.com/repos/nhost/hasura-backend-plus | closed | Is there a way to dynamically pass a redirect_url to an auth provider? | Priority: High Scope: Authentication Type: Feature Request | As we have several frontends with different domains and only one hasura-backend-plus instance, we have different redirect URL depending on the calling frontend app.
Would there be a way to dynamically pass a redirect_url to an auth provider if given, and fallback to the standard environment variable `PROVIDER_SUCCESS_REDIRECT` otherwise?
Example: //hasura-backend-plus/auth/providers/google?redirect_url_success=http://my-front-end-1/home | 1.0 | Is there a way to dynamically pass a redirect_url to an auth provider? - As we have several frontends with different domains and only one hasura-backend-plus instance, we have different redirect URL depending on the calling frontend app.
Would there be a way to dynamically pass a redirect_url to an auth provider if given, and fallback to the standard environment variable `PROVIDER_SUCCESS_REDIRECT` otherwise?
Example: //hasura-backend-plus/auth/providers/google?redirect_url_success=http://my-front-end-1/home | priority | is there a way to dynamically pass a redirect url to an auth provider as we have several frontends with different domains and only one hasura backend plus instance we have different redirect url depending on the calling frontend app would there be a way to dynamically pass a redirect url to an auth provider if given and fallback to the standard environment variable provider success redirect otherwise example hasura backend plus auth providers google redirect url success | 1 |
583,768 | 17,398,055,062 | IssuesEvent | 2021-08-02 15:42:14 | Systems-Learning-and-Development-Lab/MMM | https://api.github.com/repos/Systems-Learning-and-Development-Lab/MMM | closed | Error | priority-high | 
, אני יודעת שאמרת שזה לא קשור אבל זה קרה אחרי שהוספתי אוכלוסייה שלישית
בעצם זה קורה די הרבה, גם כשמריצה מולקולות מתנגשות זו בזו | 1.0 | Error - 
, אני יודעת שאמרת שזה לא קשור אבל זה קרה אחרי שהוספתי אוכלוסייה שלישית
בעצם זה קורה די הרבה, גם כשמריצה מולקולות מתנגשות זו בזו | priority | error אני יודעת שאמרת שזה לא קשור אבל זה קרה אחרי שהוספתי אוכלוסייה שלישית בעצם זה קורה די הרבה גם כשמריצה מולקולות מתנגשות זו בזו | 1 |
405,830 | 11,883,187,968 | IssuesEvent | 2020-03-27 15:33:32 | redhat-developer/vscode-tekton | https://api.github.com/repos/redhat-developer/vscode-tekton | closed | Add support for conditions | enhancement in progress priority/high | [Tekton conditions](https://github.com/tektoncd/pipeline/blob/master/docs/conditions.md) enable conditional execution of tasks in a pipeline. `Condition`s should be added in the Tekton window in the tree to allow user to view and edit them. | 1.0 | Add support for conditions - [Tekton conditions](https://github.com/tektoncd/pipeline/blob/master/docs/conditions.md) enable conditional execution of tasks in a pipeline. `Condition`s should be added in the Tekton window in the tree to allow user to view and edit them. | priority | add support for conditions enable conditional execution of tasks in a pipeline condition s should be added in the tekton window in the tree to allow user to view and edit them | 1 |
737,985 | 25,540,320,531 | IssuesEvent | 2022-11-29 14:53:57 | opendatahub-io/odh-dashboard | https://api.github.com/repos/opendatahub-io/odh-dashboard | closed | [Bug]: Dashboard cannot launch 300 user notebooks with 1s or 2s delay | kind/bug feature/notebook-controller priority/high | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Follow up of #627
## Details
Perform a scale test of notebook creation at user-level
* 300 users
* 1s or 2s delay between each user
* 2 consecutive runs
[300 users, 1s delay](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/512/pull-ci-openshift-psap-ci-artifacts-master-ods-nb-ux-on-ocp/1574135534509363200/artifacts/nb-ux-on-ocp/test/artifacts/test_run_1/plotting/report_00_report:_error_report.html) 0/300 successes
[300 users, 2s delay](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/512/pull-ci-openshift-psap-ci-artifacts-master-ods-plot-nb-ux-on-ocp/1574305203211997184/artifacts/plot-nb-ux-on-ocp/test/artifacts/report_00_report:_error_report.html) 31/300 successes
And similar results with m5a.2xlarge masters instead of m6a.xlarge masters:
[300 users, 1s delay, m5a.2xlarge masters](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/515/pull-ci-openshift-psap-ci-artifacts-master-ods-nb-ux-on-ocp/1575154323908726784/artifacts/nb-ux-on-ocp/test/artifacts/000_prepare/001_test_run_1/plotting/report_00_report:_error_report.html) 3/300 successes (first run)
[second run](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/515/pull-ci-openshift-psap-ci-artifacts-master-ods-nb-ux-on-ocp/1575154323908726784/artifacts/nb-ux-on-ocp/test/artifacts/000_prepare/002_test_run_2/plotting/report_00_report:_error_report.html) 215/300 successes
We got that some of these performance issues are related to `LIST` k8s calls. We have already addressed the `nb-events` endpoint in #627 and now we need to find all the bottlenecks left and fix them.
### Expected Behavior
No issues with multiple users spawning notebooks at the same time.
### Steps To Reproduce
Perform a scale test of notebook creation at user-level
* 300 users
* 1s or 2s delay between each user
* 2 consecutive runs
### Workaround (if any)
_No response_
### OpenShift Infrastructure Version
_No response_
### Openshift Version
_No response_
### What browsers are you seeing the problem on?
_No response_
### Open Data Hub Version
_No response_
### Relevant log output
_No response_ | 1.0 | [Bug]: Dashboard cannot launch 300 user notebooks with 1s or 2s delay - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Follow up of #627
## Details
Perform a scale test of notebook creation at user-level
* 300 users
* 1s or 2s delay between each user
* 2 consecutive runs
[300 users, 1s delay](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/512/pull-ci-openshift-psap-ci-artifacts-master-ods-nb-ux-on-ocp/1574135534509363200/artifacts/nb-ux-on-ocp/test/artifacts/test_run_1/plotting/report_00_report:_error_report.html) 0/300 successes
[300 users, 2s delay](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/512/pull-ci-openshift-psap-ci-artifacts-master-ods-plot-nb-ux-on-ocp/1574305203211997184/artifacts/plot-nb-ux-on-ocp/test/artifacts/report_00_report:_error_report.html) 31/300 successes
And similar results with m5a.2xlarge masters instead of m6a.xlarge masters:
[300 users, 1s delay, m5a.2xlarge masters](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/515/pull-ci-openshift-psap-ci-artifacts-master-ods-nb-ux-on-ocp/1575154323908726784/artifacts/nb-ux-on-ocp/test/artifacts/000_prepare/001_test_run_1/plotting/report_00_report:_error_report.html) 3/300 successes (first run)
[second run](https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/origin-ci-test/pr-logs/pull/openshift-psap_ci-artifacts/515/pull-ci-openshift-psap-ci-artifacts-master-ods-nb-ux-on-ocp/1575154323908726784/artifacts/nb-ux-on-ocp/test/artifacts/000_prepare/002_test_run_2/plotting/report_00_report:_error_report.html) 215/300 successes
We got that some of these performance issues are related to `LIST` k8s calls. We have already addressed the `nb-events` endpoint in #627 and now we need to find all the bottlenecks left and fix them.
### Expected Behavior
No issues with multiple users spawning notebooks at the same time.
### Steps To Reproduce
Perform a scale test of notebook creation at user-level
* 300 users
* 1s or 2s delay between each user
* 2 consecutive runs
### Workaround (if any)
_No response_
### OpenShift Infrastructure Version
_No response_
### Openshift Version
_No response_
### What browsers are you seeing the problem on?
_No response_
### Open Data Hub Version
_No response_
### Relevant log output
_No response_ | priority | dashboard cannot launch user notebooks with or delay is there an existing issue for this i have searched the existing issues current behavior follow up of details perform a scale test of notebook creation at user level users or delay between each user consecutive runs successes successes and similar results with masters instead of xlarge masters successes first run successes we got that some of these performance issues are related to list calls we have already addressed the nb events endpoint in and now we need to find all the bottlenecks left and fix them expected behavior no issues with multiple users spawning notebooks at the same time steps to reproduce perform a scale test of notebook creation at user level users or delay between each user consecutive runs workaround if any no response openshift infrastructure version no response openshift version no response what browsers are you seeing the problem on no response open data hub version no response relevant log output no response | 1 |
193,237 | 6,882,849,599 | IssuesEvent | 2017-11-21 06:45:38 | ppy/osu-framework | https://api.github.com/repos/ppy/osu-framework | opened | Fix cursor not hiding | high priority | Code is currently removed (https://github.com/smoogipoo/osu-framework/blob/netstandard/osu.Framework/Platform/GameWindow.cs#L81-L98), because it caused low-level exception on Windows. This needs to be investigated at an OpenTK level. | 1.0 | Fix cursor not hiding - Code is currently removed (https://github.com/smoogipoo/osu-framework/blob/netstandard/osu.Framework/Platform/GameWindow.cs#L81-L98), because it caused low-level exception on Windows. This needs to be investigated at an OpenTK level. | priority | fix cursor not hiding code is currently removed because it caused low level exception on windows this needs to be investigated at an opentk level | 1 |
514,177 | 14,934,904,565 | IssuesEvent | 2021-01-25 11:10:05 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Allow overriding default serializers with a property | Module: Serialization Priority: High Source: Internal Team: Client Team: Core Type: Enhancement good first issue | We have added default serializers to the 4.2 series here:
https://github.com/hazelcast/hazelcast/pull/17934
There is a problem with backward compatibility. If a user had CustomSerializer for optional in 4.1, in 4.2 there is no way to use their serializers, and Hazelcast will throw `java.lang.IllegalArgumentException: [class java.util.Optional] serializer cannot be overridden`
Users can basically remove the Optional serializer and continue but this does not play well with Rolling Upgrade. For Rolling Upgrade to work, the user should be able to continue to use the same serializer that is used in the old version.
The proposal is to add a way to override default serializers to cover this scenario.
1. It should be enabled with a property explicitly. Something like `hazelcast.serialization.allowOverrideDefaultSerializers` in ClusterProperty with default value `false`.
2. When the instance does not start and throws the `IllegalArgumentException`, we should mention the new property and the implications on a log.
3. This new property should be documented in release notes and reference manual.
| 1.0 | Allow overriding default serializers with a property - We have added default serializers to the 4.2 series here:
https://github.com/hazelcast/hazelcast/pull/17934
There is a problem with backward compatibility. If a user had CustomSerializer for optional in 4.1, in 4.2 there is no way to use their serializers, and Hazelcast will throw `java.lang.IllegalArgumentException: [class java.util.Optional] serializer cannot be overridden`
Users can basically remove the Optional serializer and continue but this does not play well with Rolling Upgrade. For Rolling Upgrade to work, the user should be able to continue to use the same serializer that is used in the old version.
The proposal is to add a way to override default serializers to cover this scenario.
1. It should be enabled with a property explicitly. Something like `hazelcast.serialization.allowOverrideDefaultSerializers` in ClusterProperty with default value `false`.
2. When the instance does not start and throws the `IllegalArgumentException`, we should mention the new property and the implications on a log.
3. This new property should be documented in release notes and reference manual.
| priority | allow overriding default serializers with a property we have added default serializers to the series here there is a problem with backward compatibility if a user had customserializer for optional in in there is no way to use their serializers and hazelcast will throw java lang illegalargumentexception serializer cannot be overridden users can basically remove the optional serializer and continue but this does not play well with rolling upgrade for rolling upgrade to work the user should be able to continue to use the same serializer that is used in the old version the proposal is to add a way to override default serializers to cover this scenario it should be enabled with a property explicitly something like hazelcast serialization allowoverridedefaultserializers in clusterproperty with default value false when the instance does not start and throws the illegalargumentexception we should mention the new property and the implications on a log this new property should be documented in release notes and reference manual | 1 |
28,366 | 2,701,180,998 | IssuesEvent | 2015-04-05 00:41:32 | TypeStrong/atom-typescript | https://api.github.com/repos/TypeStrong/atom-typescript | opened | External modules Dependency Diagram | priority:high | This is one of the things I promised when I suggest that you *must* use external modules.
This is a problem for which people have asked for a solution before. Specifically doing cyclic checks is also useful. | 1.0 | External modules Dependency Diagram - This is one of the things I promised when I suggest that you *must* use external modules.
This is a problem for which people have asked for a solution before. Specifically doing cyclic checks is also useful. | priority | external modules dependency diagram this is one of the things i promised when i suggest that you must use external modules this is a problem for which people have asked for a solution before specifically doing cyclic checks is also useful | 1 |
680,605 | 23,279,558,593 | IssuesEvent | 2022-08-05 10:37:23 | chakra-ui/chakra-ui | https://api.github.com/repos/chakra-ui/chakra-ui | closed | vite:dep-pre-bundle cannot resolve entry for some Chakra UI packages | Priority: High 🚨 | ### Description
I have a Vite SPA using React and Typescript template from Vite. After installing dependencies and running "yarn dev" which starts the development server, errors like this pop up: "[plugin vite:dep-pre-bundle] Failed to resolve entry for package "@chakra-ui/react-utils". The package may have incorrect main/module/exports specified in its package.json."
### Link to Reproduction
https://github.com/frle10/chakra-ui-package-versions-bug
### Steps to reproduce
1. Clone repo from link to reproduction
2. Install dependencies using yarn (yarn berry is set up in the repo, make sure you're installing with it)
3. Run yarn dev
4. See errors in logs from yarn dev that I described
### Chakra UI Version
2.2.4
### Browser
Google Chrome 103.0.5060.134
### Operating System
- [ ] macOS
- [ ] Windows
- [X] Linux
### Additional Information
Some error messages you should see:
```
[plugin vite:dep-scan] Failed to resolve entry for package "@chakra-ui/react".
The package may have incorrect main/module/exports specified in its package.json.
```
```
The plugin "vite:dep-scan" was triggered by this import
src/main.tsx:4:31:
4 │ import { ChakraProvider } from '@chakra-ui/react';
```
Interesting note, if you use yarn 1.22.19 and not yarn berry, everything works normally. | 1.0 | vite:dep-pre-bundle cannot resolve entry for some Chakra UI packages - ### Description
I have a Vite SPA using React and Typescript template from Vite. After installing dependencies and running "yarn dev" which starts the development server, errors like this pop up: "[plugin vite:dep-pre-bundle] Failed to resolve entry for package "@chakra-ui/react-utils". The package may have incorrect main/module/exports specified in its package.json."
### Link to Reproduction
https://github.com/frle10/chakra-ui-package-versions-bug
### Steps to reproduce
1. Clone repo from link to reproduction
2. Install dependencies using yarn (yarn berry is set up in the repo, make sure you're installing with it)
3. Run yarn dev
4. See errors in logs from yarn dev that I described
### Chakra UI Version
2.2.4
### Browser
Google Chrome 103.0.5060.134
### Operating System
- [ ] macOS
- [ ] Windows
- [X] Linux
### Additional Information
Some error messages you should see:
```
[plugin vite:dep-scan] Failed to resolve entry for package "@chakra-ui/react".
The package may have incorrect main/module/exports specified in its package.json.
```
```
The plugin "vite:dep-scan" was triggered by this import
src/main.tsx:4:31:
4 │ import { ChakraProvider } from '@chakra-ui/react';
```
Interesting note, if you use yarn 1.22.19 and not yarn berry, everything works normally. | priority | vite dep pre bundle cannot resolve entry for some chakra ui packages description i have a vite spa using react and typescript template from vite after installing dependencies and running yarn dev which starts the development server errors like this pop up failed to resolve entry for package chakra ui react utils the package may have incorrect main module exports specified in its package json link to reproduction steps to reproduce clone repo from link to reproduction install dependencies using yarn yarn berry is set up in the repo make sure you re installing with it run yarn dev see errors in logs from yarn dev that i described chakra ui version browser google chrome operating system macos windows linux additional information some error messages you should see failed to resolve entry for package chakra ui react the package may have incorrect main module exports specified in its package json the plugin vite dep scan was triggered by this import src main tsx │ import chakraprovider from chakra ui react interesting note if you use yarn and not yarn berry everything works normally | 1 |
287,830 | 8,821,348,643 | IssuesEvent | 2019-01-02 00:38:00 | Veil-Project/veil | https://api.github.com/repos/Veil-Project/veil | closed | Fix GUI for Bech32 Stealth Address type | GUI high priority wallet | When you create an address from the address book tab, it still creates basecoin addresses. | 1.0 | Fix GUI for Bech32 Stealth Address type - When you create an address from the address book tab, it still creates basecoin addresses. | priority | fix gui for stealth address type when you create an address from the address book tab it still creates basecoin addresses | 1 |
781,535 | 27,441,187,978 | IssuesEvent | 2023-03-02 11:07:01 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Gradle task for `:ballerina-cli:test` does not build all tests locally | Type/Bug Priority/High Area/CLI Team/DevTools | **Description:**
When building the `:ballerina-cli` locally, the test task executes tests inconsistently. RunCommandTest doesn't get executed along with a few others. On some occasions, tests don't execute at all.
This makes it difficult to write new tests and verify that the feature works properly. Running with 'no build cache' or 'rerunning all tasks' does not rectify this issue.
**Steps to reproduce:**
Run gw build on ballerina-cli
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Gradle task for `:ballerina-cli:test` does not build all tests locally - **Description:**
When building the `:ballerina-cli` locally, the test task executes tests inconsistently. RunCommandTest doesn't get executed along with a few others. On some occasions, tests don't execute at all.
This makes it difficult to write new tests and verify that the feature works properly. Running with 'no build cache' or 'rerunning all tasks' does not rectify this issue.
**Steps to reproduce:**
Run gw build on ballerina-cli
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| priority | gradle task for ballerina cli test does not build all tests locally description when building the ballerina cli locally the test task executes tests inconsistently runcommandtest doesn t get executed along with a few others on some occasions tests don t execute at all this makes it difficult to write new tests and verify that the feature works properly running with no build cache or rerunning all tasks does not rectify this issue steps to reproduce run gw build on ballerina cli affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional | 1 |
506,309 | 14,662,258,241 | IssuesEvent | 2020-12-29 06:43:13 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | addons.mozilla.org - site is not usable | browser-firefox-mobile engine-gecko ml-needsdiagnosis-false ml-probability-high priority-important | <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64535 -->
**URL**: https://addons.mozilla.org/pl/android/addon/ublock-origin/?utm_source=addons.mozilla.org&utm_medium=referral&utm_content=collection
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Button to install addon is greyed out on mobile. Works in desktop mode just fine, so it's definitely problem with the mobile version of the site
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/30211375-502b-4d84-98e0-d0ad93f00f85.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/1956830e-060b-4693-b10b-bcb6cf755b45)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | addons.mozilla.org - site is not usable - <!-- @browser: Firefox Mobile 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64535 -->
**URL**: https://addons.mozilla.org/pl/android/addon/ublock-origin/?utm_source=addons.mozilla.org&utm_medium=referral&utm_content=collection
**Browser / Version**: Firefox Mobile 68.0
**Operating System**: Android
**Tested Another Browser**: No
**Problem type**: Site is not usable
**Description**: Buttons or links not working
**Steps to Reproduce**:
Button to install addon is greyed out on mobile. Works in desktop mode just fine, so it's definitely problem with the mobile version of the site
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/30211375-502b-4d84-98e0-d0ad93f00f85.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200827194101</li><li>channel: default</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/1956830e-060b-4693-b10b-bcb6cf755b45)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | addons mozilla org site is not usable url browser version firefox mobile operating system android tested another browser no problem type site is not usable description buttons or links not working steps to reproduce button to install addon is greyed out on mobile works in desktop mode just fine so it s definitely problem with the mobile version of the site view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel default hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
330,226 | 10,037,050,009 | IssuesEvent | 2019-07-18 12:16:38 | nf-core/tools | https://api.github.com/repos/nf-core/tools | closed | Travis fails build w/ Nextflow v18.11.0-edge | bug high-priority | **EDIT: Not sure how necessary this is as the `18.11.0-edge` is not a stable release and older versions of Nextflow are fine for the pipeline I was working on**
Problems are because up until now Nextflow versions have always consisted of numbers and full stops.
## Issue 1: nf-core bump-version regex
When using `nf-core bump-version . 18.11.0-edge --nextflow` a regex is used to select the version and so the `-edge` is dropped. This can be seen in the [Travis test](https://travis-ci.com/nf-core/deepvariant/jobs/158240899) which fails because Nextflow cannot be installed:

## Issue 2: 'manifest.nextflowVersion' must start with '>='
After correcting the nextflow version `nf-core lint` fails. As shown by this [travis test](https://travis-ci.com/nf-core/deepvariant/jobs/158268212):

This is despite the Nextflow versions being the same in the pipeline manifest and the `.travis.yml`. I think this is caused by this line in `nextflow.config` `nextflowVersion = '>=18.11.0-edge'` however if I remove the `>` I then get this error:
```
ERROR: Test Failures:
http://nf-co.re/errors#4: Config variable 'manifest.nextflowVersion' did not start with '>=' : ''=18.11.0-edge
```
| 1.0 | Travis fails build w/ Nextflow v18.11.0-edge - **EDIT: Not sure how necessary this is as the `18.11.0-edge` is not a stable release and older versions of Nextflow are fine for the pipeline I was working on**
Problems are because up until now Nextflow versions have always consisted of numbers and full stops.
## Issue 1: nf-core bump-version regex
When using `nf-core bump-version . 18.11.0-edge --nextflow` a regex is used to select the version and so the `-edge` is dropped. This can be seen in the [Travis test](https://travis-ci.com/nf-core/deepvariant/jobs/158240899) which fails because Nextflow cannot be installed:

## Issue 2: 'manifest.nextflowVersion' must start with '>='
After correcting the nextflow version `nf-core lint` fails. As shown by this [travis test](https://travis-ci.com/nf-core/deepvariant/jobs/158268212):

This is despite the Nextflow versions being the same in the pipeline manifest and the `.travis.yml`. I think this is caused by this line in `nextflow.config` `nextflowVersion = '>=18.11.0-edge'` however if I remove the `>` I then get this error:
```
ERROR: Test Failures:
http://nf-co.re/errors#4: Config variable 'manifest.nextflowVersion' did not start with '>=' : ''=18.11.0-edge
```
| priority | travis fails build w nextflow edge edit not sure how necessary this is as the edge is not a stable release and older versions of nextflow are fine for the pipeline i was working on problems are because up until now nextflow versions have always consisted of numbers and full stops issue nf core bump version regex when using nf core bump version edge nextflow a regex is used to select the version and so the edge is dropped this can be seen in the which fails because nextflow cannot be installed issue manifest nextflowversion must start with after correcting the nextflow version nf core lint fails as shown by this this is despite the nextflow versions being the same in the pipeline manifest and the travis yml i think this is caused by this line in nextflow config nextflowversion edge however if i remove the i then get this error error test failures config variable manifest nextflowversion did not start with edge | 1 |
714,805 | 24,576,044,204 | IssuesEvent | 2022-10-13 12:26:19 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | [Bug]: Compiler crash with possible invalid Ballerina code | Type/Bug Priority/High Team/CompilerFE Area/BIR Crash Reason/EngineeringMistake userCategory/Compilation | ### Description
The program below that might be invalid Ballerina program (type declaration mismatch):
- doesn't report the problem in the VS Code
- compiler crashes
### Steps to Reproduce
```
$ cat issue.bal
public function main() {
record {|int x;|} x = table [];
}
```
```
$ bal run issue.bal
Compiling source
issue.bal
ballerina: Oh no, something really went wrong. Bad. Sad.
We appreciate it if you can report the code that broke Ballerina in
https://github.com/ballerina-platform/ballerina-lang/issues with the
log you get below and your sample code.
We thank you for helping make us better.
[2022-10-10 11:49:42,848] SEVERE {b7a.log.crash} - class org.wso2.ballerinalang.compiler.semantics.model.types.BType cannot be cast to class org.wso2.ballerinalang.compiler.semantics.model.types.BTableType (org.wso2.ballerinalang.compiler.semantics.model.types.BType and org.wso2.ballerinalang.compiler.semantics.model.types.BTableType are in unnamed module of loader 'app')
java.lang.ClassCastException: class org.wso2.ballerinalang.compiler.semantics.model.types.BType cannot be cast to class org.wso2.ballerinalang.compiler.semantics.model.types.BTableType (org.wso2.ballerinalang.compiler.semantics.model.types.BType and org.wso2.ballerinalang.compiler.semantics.model.types.BTableType are in unnamed module of loader 'app')
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:2177)
at org.wso2.ballerinalang.compiler.tree.expressions.BLangTableConstructorExpr.accept(BLangTableConstructorExpr.java:70)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:1569)
at org.wso2.ballerinalang.compiler.tree.expressions.BLangTypeConversionExpr.accept(BLangTypeConversionExpr.java:79)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:1063)
at org.wso2.ballerinalang.compiler.tree.statements.BLangSimpleVariableDef.accept(BLangSimpleVariableDef.java:50)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:732)
at org.wso2.ballerinalang.compiler.tree.BLangBlockFunctionBody.accept(BLangBlockFunctionBody.java:60)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:692)
at org.wso2.ballerinalang.compiler.tree.BLangFunction.accept(BLangFunction.java:76)
at org.wso2.ballerinalang.compiler.bir.BIRGen.lambda$visit$5(BIRGen.java:331)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:331)
at org.wso2.ballerinalang.compiler.tree.BLangPackage.accept(BLangPackage.java:167)
at org.wso2.ballerinalang.compiler.bir.BIRGen.genBIR(BIRGen.java:271)
at io.ballerina.projects.internal.CompilerPhaseRunner.birGen(CompilerPhaseRunner.java:220)
at io.ballerina.projects.internal.CompilerPhaseRunner.performBirGenPhases(CompilerPhaseRunner.java:157)
at io.ballerina.projects.ModuleContext.generateCodeInternal(ModuleContext.java:447)
at io.ballerina.projects.ModuleCompilationState$4.generatePlatformSpecificCode(ModuleCompilationState.java:132)
at io.ballerina.projects.ModuleContext.generatePlatformSpecificCode(ModuleContext.java:381)
at io.ballerina.projects.JBallerinaBackend.performCodeGen(JBallerinaBackend.java:165)
at io.ballerina.projects.JBallerinaBackend.<init>(JBallerinaBackend.java:134)
at io.ballerina.projects.JBallerinaBackend.lambda$from$0(JBallerinaBackend.java:112)
at java.base/java.util.HashMap.computeIfAbsent(HashMap.java:1133)
at io.ballerina.projects.PackageCompilation.getCompilerBackend(PackageCompilation.java:168)
at io.ballerina.projects.JBallerinaBackend.from(JBallerinaBackend.java:111)
at io.ballerina.cli.task.CompileTask.execute(CompileTask.java:151)
at io.ballerina.cli.TaskExecutor.executeTasks(TaskExecutor.java:40)
at io.ballerina.cli.cmd.RunCommand.execute(RunCommand.java:203)
at java.base/java.util.Optional.ifPresent(Optional.java:183)
at io.ballerina.cli.launcher.Main.main(Main.java:51)
ERROR [.:(1:1,1:1)] Compilation failed due to: class org.wso2.ballerinalang.compiler.semantics.model.types.BType cannot be cast to class org.wso2.ballerinalang.compiler.semantics.model.types.BTableType (org.wso2.ballerinalang.compiler.semantics.model.types.BType and org.wso2.ballerinalang.compiler.semantics.model.types.BTableType are in unnamed module of loader 'app')
WARNING [issue.bal:(2:5,2:36)] unused variable 'x'
error: compilation contains errors
```
### Affected Version(s)
```
$ bal version
Ballerina 2201.2.0 (Swan Lake Update 2)
Language specification 2022R3
Update Tool 1.3.10
```
### OS, DB, other environment details and versions
_No response_
### Related area
-> Compilation
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_ | 1.0 | [Bug]: Compiler crash with possible invalid Ballerina code - ### Description
The program below that might be invalid Ballerina program (type declaration mismatch):
- doesn't report the problem in the VS Code
- compiler crashes
### Steps to Reproduce
```
$ cat issue.bal
public function main() {
record {|int x;|} x = table [];
}
```
```
$ bal run issue.bal
Compiling source
issue.bal
ballerina: Oh no, something really went wrong. Bad. Sad.
We appreciate it if you can report the code that broke Ballerina in
https://github.com/ballerina-platform/ballerina-lang/issues with the
log you get below and your sample code.
We thank you for helping make us better.
[2022-10-10 11:49:42,848] SEVERE {b7a.log.crash} - class org.wso2.ballerinalang.compiler.semantics.model.types.BType cannot be cast to class org.wso2.ballerinalang.compiler.semantics.model.types.BTableType (org.wso2.ballerinalang.compiler.semantics.model.types.BType and org.wso2.ballerinalang.compiler.semantics.model.types.BTableType are in unnamed module of loader 'app')
java.lang.ClassCastException: class org.wso2.ballerinalang.compiler.semantics.model.types.BType cannot be cast to class org.wso2.ballerinalang.compiler.semantics.model.types.BTableType (org.wso2.ballerinalang.compiler.semantics.model.types.BType and org.wso2.ballerinalang.compiler.semantics.model.types.BTableType are in unnamed module of loader 'app')
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:2177)
at org.wso2.ballerinalang.compiler.tree.expressions.BLangTableConstructorExpr.accept(BLangTableConstructorExpr.java:70)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:1569)
at org.wso2.ballerinalang.compiler.tree.expressions.BLangTypeConversionExpr.accept(BLangTypeConversionExpr.java:79)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:1063)
at org.wso2.ballerinalang.compiler.tree.statements.BLangSimpleVariableDef.accept(BLangSimpleVariableDef.java:50)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:732)
at org.wso2.ballerinalang.compiler.tree.BLangBlockFunctionBody.accept(BLangBlockFunctionBody.java:60)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:692)
at org.wso2.ballerinalang.compiler.tree.BLangFunction.accept(BLangFunction.java:76)
at org.wso2.ballerinalang.compiler.bir.BIRGen.lambda$visit$5(BIRGen.java:331)
at java.base/java.util.ArrayList.forEach(ArrayList.java:1541)
at org.wso2.ballerinalang.compiler.bir.BIRGen.visit(BIRGen.java:331)
at org.wso2.ballerinalang.compiler.tree.BLangPackage.accept(BLangPackage.java:167)
at org.wso2.ballerinalang.compiler.bir.BIRGen.genBIR(BIRGen.java:271)
at io.ballerina.projects.internal.CompilerPhaseRunner.birGen(CompilerPhaseRunner.java:220)
at io.ballerina.projects.internal.CompilerPhaseRunner.performBirGenPhases(CompilerPhaseRunner.java:157)
at io.ballerina.projects.ModuleContext.generateCodeInternal(ModuleContext.java:447)
at io.ballerina.projects.ModuleCompilationState$4.generatePlatformSpecificCode(ModuleCompilationState.java:132)
at io.ballerina.projects.ModuleContext.generatePlatformSpecificCode(ModuleContext.java:381)
at io.ballerina.projects.JBallerinaBackend.performCodeGen(JBallerinaBackend.java:165)
at io.ballerina.projects.JBallerinaBackend.<init>(JBallerinaBackend.java:134)
at io.ballerina.projects.JBallerinaBackend.lambda$from$0(JBallerinaBackend.java:112)
at java.base/java.util.HashMap.computeIfAbsent(HashMap.java:1133)
at io.ballerina.projects.PackageCompilation.getCompilerBackend(PackageCompilation.java:168)
at io.ballerina.projects.JBallerinaBackend.from(JBallerinaBackend.java:111)
at io.ballerina.cli.task.CompileTask.execute(CompileTask.java:151)
at io.ballerina.cli.TaskExecutor.executeTasks(TaskExecutor.java:40)
at io.ballerina.cli.cmd.RunCommand.execute(RunCommand.java:203)
at java.base/java.util.Optional.ifPresent(Optional.java:183)
at io.ballerina.cli.launcher.Main.main(Main.java:51)
ERROR [.:(1:1,1:1)] Compilation failed due to: class org.wso2.ballerinalang.compiler.semantics.model.types.BType cannot be cast to class org.wso2.ballerinalang.compiler.semantics.model.types.BTableType (org.wso2.ballerinalang.compiler.semantics.model.types.BType and org.wso2.ballerinalang.compiler.semantics.model.types.BTableType are in unnamed module of loader 'app')
WARNING [issue.bal:(2:5,2:36)] unused variable 'x'
error: compilation contains errors
```
### Affected Version(s)
```
$ bal version
Ballerina 2201.2.0 (Swan Lake Update 2)
Language specification 2022R3
Update Tool 1.3.10
```
### OS, DB, other environment details and versions
_No response_
### Related area
-> Compilation
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_ | priority | compiler crash with possible invalid ballerina code description the program below that might be invalid ballerina program type declaration mismatch doesn t report the problem in the vs code compiler crashes steps to reproduce cat issue bal public function main record int x x table bal run issue bal compiling source issue bal ballerina oh no something really went wrong bad sad we appreciate it if you can report the code that broke ballerina in with the log you get below and your sample code we thank you for helping make us better severe log crash class org ballerinalang compiler semantics model types btype cannot be cast to class org ballerinalang compiler semantics model types btabletype org ballerinalang compiler semantics model types btype and org ballerinalang compiler semantics model types btabletype are in unnamed module of loader app java lang classcastexception class org ballerinalang compiler semantics model types btype cannot be cast to class org ballerinalang compiler semantics model types btabletype org ballerinalang compiler semantics model types btype and org ballerinalang compiler semantics model types btabletype are in unnamed module of loader app at org ballerinalang compiler bir birgen visit birgen java at org ballerinalang compiler tree expressions blangtableconstructorexpr accept blangtableconstructorexpr java at org ballerinalang compiler bir birgen visit birgen java at org ballerinalang compiler tree expressions blangtypeconversionexpr accept blangtypeconversionexpr java at org ballerinalang compiler bir birgen visit birgen java at org ballerinalang compiler tree statements blangsimplevariabledef accept blangsimplevariabledef java at org ballerinalang compiler bir birgen visit birgen java at org ballerinalang compiler tree blangblockfunctionbody accept blangblockfunctionbody java at org ballerinalang compiler bir birgen visit birgen java at org ballerinalang compiler tree blangfunction accept blangfunction java at org ballerinalang compiler bir birgen lambda visit birgen java at java base java util arraylist foreach arraylist java at org ballerinalang compiler bir birgen visit birgen java at org ballerinalang compiler tree blangpackage accept blangpackage java at org ballerinalang compiler bir birgen genbir birgen java at io ballerina projects internal compilerphaserunner birgen compilerphaserunner java at io ballerina projects internal compilerphaserunner performbirgenphases compilerphaserunner java at io ballerina projects modulecontext generatecodeinternal modulecontext java at io ballerina projects modulecompilationstate generateplatformspecificcode modulecompilationstate java at io ballerina projects modulecontext generateplatformspecificcode modulecontext java at io ballerina projects jballerinabackend performcodegen jballerinabackend java at io ballerina projects jballerinabackend jballerinabackend java at io ballerina projects jballerinabackend lambda from jballerinabackend java at java base java util hashmap computeifabsent hashmap java at io ballerina projects packagecompilation getcompilerbackend packagecompilation java at io ballerina projects jballerinabackend from jballerinabackend java at io ballerina cli task compiletask execute compiletask java at io ballerina cli taskexecutor executetasks taskexecutor java at io ballerina cli cmd runcommand execute runcommand java at java base java util optional ifpresent optional java at io ballerina cli launcher main main main java error compilation failed due to class org ballerinalang compiler semantics model types btype cannot be cast to class org ballerinalang compiler semantics model types btabletype org ballerinalang compiler semantics model types btype and org ballerinalang compiler semantics model types btabletype are in unnamed module of loader app warning unused variable x error compilation contains errors affected version s bal version ballerina swan lake update language specification update tool os db other environment details and versions no response related area compilation related issue s optional no response suggested label s optional no response suggested assignee s optional no response | 1 |
308,183 | 9,435,832,707 | IssuesEvent | 2019-04-13 00:38:08 | sugarlabs/musicblocks | https://api.github.com/repos/sugarlabs/musicblocks | closed | Collision with action name after merging two projects | bug high priority | I found this issue this week when showing the merge feature to a student.
I have the original files, but I also have successfully reproduced the issue which is described below.
1. Use just the action called "action" from crab canon plot (in examples folder)

2. Merge [remix-of-twinkle-twinkle.html.zip](https://github.com/sugarlabs/musicblocks/files/2798553/remix-of-twinkle-twinkle.html.zip)

3. You will notice two actions, both with the same name

4. When you change one of the actions, it changes the name of the other -- meaning you cannot call them individually.

This is the example, resultant [Corrupted-File.html.zip](https://github.com/sugarlabs/musicblocks/files/2798554/Corrupted-File.html.zip)
| 1.0 | Collision with action name after merging two projects - I found this issue this week when showing the merge feature to a student.
I have the original files, but I also have successfully reproduced the issue which is described below.
1. Use just the action called "action" from crab canon plot (in examples folder)

2. Merge [remix-of-twinkle-twinkle.html.zip](https://github.com/sugarlabs/musicblocks/files/2798553/remix-of-twinkle-twinkle.html.zip)

3. You will notice two actions, both with the same name

4. When you change one of the actions, it changes the name of the other -- meaning you cannot call them individually.

This is the example, resultant [Corrupted-File.html.zip](https://github.com/sugarlabs/musicblocks/files/2798554/Corrupted-File.html.zip)
| priority | collision with action name after merging two projects i found this issue this week when showing the merge feature to a student i have the original files but i also have successfully reproduced the issue which is described below use just the action called action from crab canon plot in examples folder merge you will notice two actions both with the same name when you change one of the actions it changes the name of the other meaning you cannot call them individually this is the example resultant | 1 |
262,264 | 8,257,844,078 | IssuesEvent | 2018-09-13 07:11:47 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | opened | Build failed when CONFIG_STM32_ARM_MPU_ENABLE=y | STM32 bug priority: high | Following merge of #9776, building STM32 platform is failed when STM32_ARM_MPU_ENABLE is enabled | 1.0 | Build failed when CONFIG_STM32_ARM_MPU_ENABLE=y - Following merge of #9776, building STM32 platform is failed when STM32_ARM_MPU_ENABLE is enabled | priority | build failed when config arm mpu enable y following merge of building platform is failed when arm mpu enable is enabled | 1 |
694,355 | 23,810,810,262 | IssuesEvent | 2022-09-04 18:32:54 | pokt-network/pocket | https://api.github.com/repos/pokt-network/pocket | closed | [Persistence] Merkling a State Hash | core persistence priority:high | # Objective
Make the persistence **"a real blockchain"**.
Expand on the existing SQL-based persistence implementation to create a verifiable authenticator (i.e. the state hash) which can be used to guarantee the integrity/immutability of the blockchain and verify various parts of the state (i.e. Merkle Proofs).
# Origin Document
[Pocket protocol persistence specification](https://github.com/pokt-network/pocket-network-protocol/tree/main/persistence) defines the persistence specification.
The current persistence interface can be found [here](https://github.com/pokt-network/pocket/tree/main/persistence) with the shared interface available [here](https://github.com/pokt-network/pocket/blob/main/shared/modules/persistence_module.go).
[This pull request](https://github.com/pokt-network/pocket-network-protocol/pull/8) to the persistence specification contains additional details and a spec coverage checklist that this will aim to implement.
# Goals / Deliverables
- Research & Specification
- [ ] Prepare a presentation explaining how Merkle Trees & Key-Value stores operate in blockchains
- [ ] Update the persistence spec from v1.0.3 (the PR mentioned above) to v1.0.4
- Integration
- [ ] Update the primary Persistence module interface
- [ ] Hook up the KV Store, Merkle Tree and Persistence Module to function together
- Merkle Tree
- [ ] Design an interface to interact with a Sparse Merkle Tree
- [ ] Select an open-source and/or implement a Sparse Merkle Tree for Pocket based on an existing specification
- Key-Value Store
- [ ] Design an interface to interact with a Key-Value store via the Merkle Tree
- [ ] Select an open-source Key-Value store for Pocket based on an existing specification
- CLI
- [ ] If needed, update the existing pocket client CLI to test, develop and drive the KV store (similar to Tendermint's [abci-cli]
- [ ] Testing
- [ ] Test each new component / interface in isolation via:
- [ ] Unit tests
- [ ] Fuzzing tests
- [ ] Determine if there is some sort of testing framework that can be implemented (i.e. DSL-like)
- [ ] Determine the degree to which end-to-end tests can be implemented, or simply documented, and do so
## General milestone checklist
- [ ] Update all the relevant CHANGELOGs
- [ ] Update all the relevant READMEs
- [ ] Update the source code tree explanation
- [ ] Add or update a state, sequence or flowchart diagram using [mermaid](https://mermaid-js.github.io/mermaid/)
- [ ] Create followup milestones + issues
- [ ] Document small TODO along the way
# Non-goals
- End-to-end benchmarking of the key-value store and database engine
- Functional state sync
# Testing Methodology
- **All tests**: `make test_all`
- **LocalNet**: verify a `LocalNet` is still functioning correctly by following the instructions at [docs/development/README.md](https://github.com/pokt-network/pocket/tree/main/docs/development)
Remove
---
Creator: @Olshansk
Co-Owners: ???
| 1.0 | [Persistence] Merkling a State Hash - # Objective
Make the persistence **"a real blockchain"**.
Expand on the existing SQL-based persistence implementation to create a verifiable authenticator (i.e. the state hash) which can be used to guarantee the integrity/immutability of the blockchain and verify various parts of the state (i.e. Merkle Proofs).
# Origin Document
[Pocket protocol persistence specification](https://github.com/pokt-network/pocket-network-protocol/tree/main/persistence) defines the persistence specification.
The current persistence interface can be found [here](https://github.com/pokt-network/pocket/tree/main/persistence) with the shared interface available [here](https://github.com/pokt-network/pocket/blob/main/shared/modules/persistence_module.go).
[This pull request](https://github.com/pokt-network/pocket-network-protocol/pull/8) to the persistence specification contains additional details and a spec coverage checklist that this will aim to implement.
# Goals / Deliverables
- Research & Specification
- [ ] Prepare a presentation explaining how Merkle Trees & Key-Value stores operate in blockchains
- [ ] Update the persistence spec from v1.0.3 (the PR mentioned above) to v1.0.4
- Integration
- [ ] Update the primary Persistence module interface
- [ ] Hook up the KV Store, Merkle Tree and Persistence Module to function together
- Merkle Tree
- [ ] Design an interface to interact with a Sparse Merkle Tree
- [ ] Select an open-source and/or implement a Sparse Merkle Tree for Pocket based on an existing specification
- Key-Value Store
- [ ] Design an interface to interact with a Key-Value store via the Merkle Tree
- [ ] Select an open-source Key-Value store for Pocket based on an existing specification
- CLI
- [ ] If needed, update the existing pocket client CLI to test, develop and drive the KV store (similar to Tendermint's [abci-cli]
- [ ] Testing
- [ ] Test each new component / interface in isolation via:
- [ ] Unit tests
- [ ] Fuzzing tests
- [ ] Determine if there is some sort of testing framework that can be implemented (i.e. DSL-like)
- [ ] Determine the degree to which end-to-end tests can be implemented, or simply documented, and do so
## General milestone checklist
- [ ] Update all the relevant CHANGELOGs
- [ ] Update all the relevant READMEs
- [ ] Update the source code tree explanation
- [ ] Add or update a state, sequence or flowchart diagram using [mermaid](https://mermaid-js.github.io/mermaid/)
- [ ] Create followup milestones + issues
- [ ] Document small TODO along the way
# Non-goals
- End-to-end benchmarking of the key-value store and database engine
- Functional state sync
# Testing Methodology
- **All tests**: `make test_all`
- **LocalNet**: verify a `LocalNet` is still functioning correctly by following the instructions at [docs/development/README.md](https://github.com/pokt-network/pocket/tree/main/docs/development)
Remove
---
Creator: @Olshansk
Co-Owners: ???
| priority | merkling a state hash objective make the persistence a real blockchain expand on the existing sql based persistence implementation to create a verifiable authenticator i e the state hash which can be used to guarantee the integrity immutability of the blockchain and verify various parts of the state i e merkle proofs origin document defines the persistence specification the current persistence interface can be found with the shared interface available to the persistence specification contains additional details and a spec coverage checklist that this will aim to implement goals deliverables research specification prepare a presentation explaining how merkle trees key value stores operate in blockchains update the persistence spec from the pr mentioned above to integration update the primary persistence module interface hook up the kv store merkle tree and persistence module to function together merkle tree design an interface to interact with a sparse merkle tree select an open source and or implement a sparse merkle tree for pocket based on an existing specification key value store design an interface to interact with a key value store via the merkle tree select an open source key value store for pocket based on an existing specification cli if needed update the existing pocket client cli to test develop and drive the kv store similar to tendermint s testing test each new component interface in isolation via unit tests fuzzing tests determine if there is some sort of testing framework that can be implemented i e dsl like determine the degree to which end to end tests can be implemented or simply documented and do so general milestone checklist update all the relevant changelogs update all the relevant readmes update the source code tree explanation add or update a state sequence or flowchart diagram using create followup milestones issues document small todo along the way non goals end to end benchmarking of the key value store and database engine functional state sync testing methodology all tests make test all localnet verify a localnet is still functioning correctly by following the instructions at remove creator olshansk co owners | 1 |
469,957 | 13,528,621,998 | IssuesEvent | 2020-09-15 16:59:35 | MaibornWolff/codecharta | https://api.github.com/repos/MaibornWolff/codecharta | opened | Show number of files in the attribute-side-bar when selecting a folder on the map | difficulty:medium pr-visualization priority:high | # Feature request
## Description
As a user, I want see how many files are inside a folder after I selected the folder on the map.
## Acceptance criteria
- Remove UNARY-metric entirely
- Number of files in folder should be displayed in the attribute-side-bar | 1.0 | Show number of files in the attribute-side-bar when selecting a folder on the map - # Feature request
## Description
As a user, I want see how many files are inside a folder after I selected the folder on the map.
## Acceptance criteria
- Remove UNARY-metric entirely
- Number of files in folder should be displayed in the attribute-side-bar | priority | show number of files in the attribute side bar when selecting a folder on the map feature request description as a user i want see how many files are inside a folder after i selected the folder on the map acceptance criteria remove unary metric entirely number of files in folder should be displayed in the attribute side bar | 1 |
174,489 | 6,540,556,223 | IssuesEvent | 2017-09-01 15:53:19 | krisgardner/surveyDevelopment | https://api.github.com/repos/krisgardner/surveyDevelopment | closed | Multi logo - setting script | High Priority | need to have one script as there seems to be 2 setting scripts in use, one in our library and one in html5 tools
| 1.0 | Multi logo - setting script - need to have one script as there seems to be 2 setting scripts in use, one in our library and one in html5 tools
| priority | multi logo setting script need to have one script as there seems to be setting scripts in use one in our library and one in tools | 1 |
187,810 | 6,761,315,051 | IssuesEvent | 2017-10-25 00:55:51 | CS2103AUG2017-W15-B2/main | https://api.github.com/repos/CS2103AUG2017-W15-B2/main | closed | As an user, I want to list all tags | new feature Priority: HIGH | ... so that I can see the tags I have added to users in case I've forgotten some of the tags.
### Tasks:
- Implement ListAllTags command. (Done)
- Teach Parser to accept ListAllTags command. (Done)
- Teach EventsCenter to receive and respond to ListAllTags event. (Done)
- Teach UI to display the tag list. (Done)
- Teach UI to remove the tag list and replace it with original Browser Panel when user is done viewing tag list.
- Update Developer Guide to reflect the feature implemented. (WIP)
#14
| 1.0 | As an user, I want to list all tags - ... so that I can see the tags I have added to users in case I've forgotten some of the tags.
### Tasks:
- Implement ListAllTags command. (Done)
- Teach Parser to accept ListAllTags command. (Done)
- Teach EventsCenter to receive and respond to ListAllTags event. (Done)
- Teach UI to display the tag list. (Done)
- Teach UI to remove the tag list and replace it with original Browser Panel when user is done viewing tag list.
- Update Developer Guide to reflect the feature implemented. (WIP)
#14
| priority | as an user i want to list all tags so that i can see the tags i have added to users in case i ve forgotten some of the tags tasks implement listalltags command done teach parser to accept listalltags command done teach eventscenter to receive and respond to listalltags event done teach ui to display the tag list done teach ui to remove the tag list and replace it with original browser panel when user is done viewing tag list update developer guide to reflect the feature implemented wip | 1 |
723,537 | 24,900,450,129 | IssuesEvent | 2022-10-28 20:15:37 | bcgov/digital-journeys | https://api.github.com/repos/bcgov/digital-journeys | reopened | Unexpected error when authenticating with identity provider - KEYCLOAK/DUPLICATE USER | bug Epic Telework Agreement high priority | [Uploading... Ticket653_DuplicateUser_KEYCLOAK.msg]Background: a few users are unable to get past the login/authentication page, whether selecting the link generated by the employee's submission or from the Telework web page. Approx count so far: 3-4 users affected.
Steps to Reproduce: 1) Read, review a telework agreement in the supervisor's task list. 2) Create a new telework agreement via the Telework's gov't website: [Guide to submitting a BC Public Service digital telework agreement - Province of British Columbia (gov.bc.ca)](url)
Actual Behavior: Telework Website/Employee generated Task list Link: 1) Login with IDIR Username and Password, select: Continue. Error Page loads: PSA-Digital-Journey. We are sorry.... Unexpected error when authenticating with identity provider
Expected Behaviour: User is able to access the task list and also create their own telework agreement. Both environments load: task list and the telework form depending on their role as an approver or as an employee.
Environment: Occurring in Production
Tracking Cases: User: Director, Access: needed for both telework and approval link, Date: Thur Oct, 20/22
| 1.0 | Unexpected error when authenticating with identity provider - KEYCLOAK/DUPLICATE USER - [Uploading... Ticket653_DuplicateUser_KEYCLOAK.msg]Background: a few users are unable to get past the login/authentication page, whether selecting the link generated by the employee's submission or from the Telework web page. Approx count so far: 3-4 users affected.
Steps to Reproduce: 1) Read, review a telework agreement in the supervisor's task list. 2) Create a new telework agreement via the Telework's gov't website: [Guide to submitting a BC Public Service digital telework agreement - Province of British Columbia (gov.bc.ca)](url)
Actual Behavior: Telework Website/Employee generated Task list Link: 1) Login with IDIR Username and Password, select: Continue. Error Page loads: PSA-Digital-Journey. We are sorry.... Unexpected error when authenticating with identity provider
Expected Behaviour: User is able to access the task list and also create their own telework agreement. Both environments load: task list and the telework form depending on their role as an approver or as an employee.
Environment: Occurring in Production
Tracking Cases: User: Director, Access: needed for both telework and approval link, Date: Thur Oct, 20/22
| priority | unexpected error when authenticating with identity provider keycloak duplicate user background a few users are unable to get past the login authentication page whether selecting the link generated by the employee s submission or from the telework web page approx count so far users affected steps to reproduce read review a telework agreement in the supervisor s task list create a new telework agreement via the telework s gov t website url actual behavior telework website employee generated task list link login with idir username and password select continue error page loads psa digital journey we are sorry unexpected error when authenticating with identity provider expected behaviour user is able to access the task list and also create their own telework agreement both environments load task list and the telework form depending on their role as an approver or as an employee environment occurring in production tracking cases user director access needed for both telework and approval link date thur oct | 1 |
339,390 | 10,253,622,348 | IssuesEvent | 2019-08-21 11:46:16 | FundacionParaguaya/MentorApp | https://api.github.com/repos/FundacionParaguaya/MentorApp | closed | App crashes on families screen | bug high priority | This is most likely cause on fresh installation, without accepted fine location permission. | 1.0 | App crashes on families screen - This is most likely cause on fresh installation, without accepted fine location permission. | priority | app crashes on families screen this is most likely cause on fresh installation without accepted fine location permission | 1 |
263,169 | 8,279,190,107 | IssuesEvent | 2018-09-18 01:42:55 | nnsuite/TAOS-CI | https://api.github.com/repos/nnsuite/TAOS-CI | opened | [Indent] Indent result is toggled. | bug priority-high | 
as the picture, indent result is toggled so CI make error. | 1.0 | [Indent] Indent result is toggled. - 
as the picture, indent result is toggled so CI make error. | priority | indent result is toggled as the picture indent result is toggled so ci make error | 1 |
319,648 | 9,747,672,758 | IssuesEvent | 2019-06-03 14:51:12 | gamificalostudio/Tankerfield | https://api.github.com/repos/gamificalostudio/Tankerfield | closed | F7 god mode does not work | Frequency: High Priority: Minor fixed | **Bug description**
F7 god mode does not work
**Steps to reproduce**
**Expected behaviour**
**Actual behaviour**
**Game version**
Tankerfield 0.8
**Screenshots**
**Additional context**
| 1.0 | F7 god mode does not work - **Bug description**
F7 god mode does not work
**Steps to reproduce**
**Expected behaviour**
**Actual behaviour**
**Game version**
Tankerfield 0.8
**Screenshots**
**Additional context**
| priority | god mode does not work bug description god mode does not work steps to reproduce expected behaviour actual behaviour game version tankerfield screenshots additional context | 1 |
259,088 | 8,183,364,511 | IssuesEvent | 2018-08-29 08:50:41 | ISISScientificComputing/autoreduce | https://api.github.com/repos/ISISScientificComputing/autoreduce | opened | Development permission denied to archive | Bug High Priority develop | When attempting to configure a new run on the WebApp
* Open WebApp
* `Configure New Run`
See that there is an error caused. Looking at `WebApp/autoreduction_webapp/autoreduction.log` reveals that this appears to be a permissions error when attempting to read `reduce_vars.py`
| 1.0 | Development permission denied to archive - When attempting to configure a new run on the WebApp
* Open WebApp
* `Configure New Run`
See that there is an error caused. Looking at `WebApp/autoreduction_webapp/autoreduction.log` reveals that this appears to be a permissions error when attempting to read `reduce_vars.py`
| priority | development permission denied to archive when attempting to configure a new run on the webapp open webapp configure new run see that there is an error caused looking at webapp autoreduction webapp autoreduction log reveals that this appears to be a permissions error when attempting to read reduce vars py | 1 |
795,923 | 28,092,348,206 | IssuesEvent | 2023-03-30 13:49:07 | status-im/status-desktop | https://api.github.com/repos/status-im/status-desktop | closed | Missing unread message badges on community and channels | bug priority 1: high E:Bugfixes S:3 messenger-team | # Bug Report
## Description
In the 0.11RC1, I opened the app after a night's sleep. I saw that I have 3 AC notifications, 2 of them on the community, but the community didn't have a dot, nor a badge.
If I manually go in the channel, I can see that there are new messages. Some other channels that didn't have mentions for me also had messages and didn't have the unread message dot.
#### Expected behavior
All channels that have unread messages that are not muted should have the dot. If those are mentions or replies, have the badge with the appropriate number.
### Additional Information
- Status desktop version: 0.11RC1
| 1.0 | Missing unread message badges on community and channels - # Bug Report
## Description
In the 0.11RC1, I opened the app after a night's sleep. I saw that I have 3 AC notifications, 2 of them on the community, but the community didn't have a dot, nor a badge.
If I manually go in the channel, I can see that there are new messages. Some other channels that didn't have mentions for me also had messages and didn't have the unread message dot.
#### Expected behavior
All channels that have unread messages that are not muted should have the dot. If those are mentions or replies, have the badge with the appropriate number.
### Additional Information
- Status desktop version: 0.11RC1
| priority | missing unread message badges on community and channels bug report description in the i opened the app after a night s sleep i saw that i have ac notifications of them on the community but the community didn t have a dot nor a badge if i manually go in the channel i can see that there are new messages some other channels that didn t have mentions for me also had messages and didn t have the unread message dot expected behavior all channels that have unread messages that are not muted should have the dot if those are mentions or replies have the badge with the appropriate number additional information status desktop version | 1 |
702,839 | 24,138,181,357 | IssuesEvent | 2022-09-21 13:01:49 | DroidKaigi/conference-app-2022 | https://api.github.com/repos/DroidKaigi/conference-app-2022 | closed | Implement announcement screen | welcome contribute high priority | **Idea Description**
apply notification screen design
**Reference images and links**
coming soon | 1.0 | Implement announcement screen - **Idea Description**
apply notification screen design
**Reference images and links**
coming soon | priority | implement announcement screen idea description apply notification screen design reference images and links coming soon | 1 |
545,714 | 15,955,647,845 | IssuesEvent | 2021-04-15 14:49:16 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Compiling from src got AssertionError: No matching output for forward derivative in native_functions.yaml _embedding_bag_dense_backward | high priority module: embedding module: nn triage review | ## ❓ Questions and Help
When I try to install as in the instructions I got an error that I nailed from the logs to this command:
```
$ cd /home/tyoc213/Documents/github/pytorch && /home/tyoc213/miniconda3/envs/xla/bin/python tools/setup_helpers/generate_code.py --declarations-path /home/tyoc213/Documents/github/pytorch/build/aten/src/ATen/Declarations.yaml --native-functions-path aten/src/ATen/native/native_functions.yaml --nn-path aten/src --force_schema_registration
Traceback (most recent call last):
File "tools/setup_helpers/generate_code.py", line 183, in <module>
main()
File "tools/setup_helpers/generate_code.py", line 168, in main
generate_code(
File "tools/setup_helpers/generate_code.py", line 60, in generate_code
gen_autograd_python(
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/gen_autograd.py", line 83, in gen_autograd_python
differentiability_infos = load_derivatives(
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 42, in load_derivatives
infos = [
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 43, in <listcomp>
create_differentiability_info(defn, functions_by_signature, functions_by_schema)
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 397, in create_differentiability_info
derivatives, forward_derivatives, args_with_derivatives, non_differentiable_arg_names = set_up_derivatives(canonical)
File "/home/tyoc213/Documents/github/pytorch/tools/codegen/context.py", line 36, in wrapper
return func(f)
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 334, in set_up_derivatives
forward_derivatives.append(create_forward_derivative(f, formula, names))
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 117, in create_forward_derivative
assert var_type is not None, "No matching output for forward derivative definition"
AssertionError: No matching output for forward derivative definition
in aten/src/ATen/native/native_functions.yaml:1490:
_embedding_bag_dense_backward(Tensor grad, Tensor indices, Tensor offset2bag, Tensor bag_size, Tensor maximum_indices, int num_weights, bool scale_grad_by_freq, int mode, Tensor? per_sample_weights, int padding_idx=-1) -> Tensor
```
If it serves something `tools/autograd/load_derivatives.py`

prints:
```
NativeFunction(func=FunctionSchema(name=OperatorName(name=BaseOperatorName(base='_embedding_bag_dense_backward', inplace=False, dunder_method=False), overload_name=''), arguments=Arguments(pre_self_positional=(), self_arg=None, post_self_positional=(Argument(name='grad', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='indices', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='offset2bag', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='bag_size', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='maximum_indices', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='num_weights', type=BaseType(name=<BaseTy.int: 4>), default=None, annotation=None), Argument(name='scale_grad_by_freq', type=BaseType(name=<BaseTy.bool: 8>), default=None, annotation=None), Argument(name='mode', type=BaseType(name=<BaseTy.int: 4>), default=None, annotation=None), Argument(name='per_sample_weights', type=OptionalType(elem=BaseType(name=<BaseTy.Tensor: 3>)), default=None, annotation=None), Argument(name='padding_idx', type=BaseType(name=<BaseTy.int: 4>), default='-1', annotation=None)), pre_tensor_options_kwarg_only=(), tensor_options=None, post_tensor_options_kwarg_only=(), out=()), returns=(Return(name=None, type=BaseType(name=<BaseTy.Tensor: 3>), annotation=None),)), device_guard=True, python_module=None, category_override=None, variants={<Variant.function: 1>}, manual_kernel_registration=False, manual_cpp_binding=False, dispatch={<DispatchKey.CPU: 1>: '_embedding_bag_dense_backward_cpu', <DispatchKey.CUDA: 2>: '_embedding_bag_dense_backward_cuda'}, loc=Location(file='aten/src/ATen/native/native_functions.yaml', line=1490), structured=False, structured_delegate=None, structured_inherits=None, cpp_no_default_args=set())
------
non_differentiable
------
('offsets',)
+++++++++++++++++++++
```
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @albanD @mruberry | 1.0 | Compiling from src got AssertionError: No matching output for forward derivative in native_functions.yaml _embedding_bag_dense_backward - ## ❓ Questions and Help
When I try to install as in the instructions I got an error that I nailed from the logs to this command:
```
$ cd /home/tyoc213/Documents/github/pytorch && /home/tyoc213/miniconda3/envs/xla/bin/python tools/setup_helpers/generate_code.py --declarations-path /home/tyoc213/Documents/github/pytorch/build/aten/src/ATen/Declarations.yaml --native-functions-path aten/src/ATen/native/native_functions.yaml --nn-path aten/src --force_schema_registration
Traceback (most recent call last):
File "tools/setup_helpers/generate_code.py", line 183, in <module>
main()
File "tools/setup_helpers/generate_code.py", line 168, in main
generate_code(
File "tools/setup_helpers/generate_code.py", line 60, in generate_code
gen_autograd_python(
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/gen_autograd.py", line 83, in gen_autograd_python
differentiability_infos = load_derivatives(
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 42, in load_derivatives
infos = [
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 43, in <listcomp>
create_differentiability_info(defn, functions_by_signature, functions_by_schema)
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 397, in create_differentiability_info
derivatives, forward_derivatives, args_with_derivatives, non_differentiable_arg_names = set_up_derivatives(canonical)
File "/home/tyoc213/Documents/github/pytorch/tools/codegen/context.py", line 36, in wrapper
return func(f)
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 334, in set_up_derivatives
forward_derivatives.append(create_forward_derivative(f, formula, names))
File "/home/tyoc213/Documents/github/pytorch/tools/autograd/load_derivatives.py", line 117, in create_forward_derivative
assert var_type is not None, "No matching output for forward derivative definition"
AssertionError: No matching output for forward derivative definition
in aten/src/ATen/native/native_functions.yaml:1490:
_embedding_bag_dense_backward(Tensor grad, Tensor indices, Tensor offset2bag, Tensor bag_size, Tensor maximum_indices, int num_weights, bool scale_grad_by_freq, int mode, Tensor? per_sample_weights, int padding_idx=-1) -> Tensor
```
If it serves something `tools/autograd/load_derivatives.py`

prints:
```
NativeFunction(func=FunctionSchema(name=OperatorName(name=BaseOperatorName(base='_embedding_bag_dense_backward', inplace=False, dunder_method=False), overload_name=''), arguments=Arguments(pre_self_positional=(), self_arg=None, post_self_positional=(Argument(name='grad', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='indices', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='offset2bag', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='bag_size', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='maximum_indices', type=BaseType(name=<BaseTy.Tensor: 3>), default=None, annotation=None), Argument(name='num_weights', type=BaseType(name=<BaseTy.int: 4>), default=None, annotation=None), Argument(name='scale_grad_by_freq', type=BaseType(name=<BaseTy.bool: 8>), default=None, annotation=None), Argument(name='mode', type=BaseType(name=<BaseTy.int: 4>), default=None, annotation=None), Argument(name='per_sample_weights', type=OptionalType(elem=BaseType(name=<BaseTy.Tensor: 3>)), default=None, annotation=None), Argument(name='padding_idx', type=BaseType(name=<BaseTy.int: 4>), default='-1', annotation=None)), pre_tensor_options_kwarg_only=(), tensor_options=None, post_tensor_options_kwarg_only=(), out=()), returns=(Return(name=None, type=BaseType(name=<BaseTy.Tensor: 3>), annotation=None),)), device_guard=True, python_module=None, category_override=None, variants={<Variant.function: 1>}, manual_kernel_registration=False, manual_cpp_binding=False, dispatch={<DispatchKey.CPU: 1>: '_embedding_bag_dense_backward_cpu', <DispatchKey.CUDA: 2>: '_embedding_bag_dense_backward_cuda'}, loc=Location(file='aten/src/ATen/native/native_functions.yaml', line=1490), structured=False, structured_delegate=None, structured_inherits=None, cpp_no_default_args=set())
------
non_differentiable
------
('offsets',)
+++++++++++++++++++++
```
cc @ezyang @gchanan @zou3519 @bdhirsh @jbschlosser @anjali411 @albanD @mruberry | priority | compiling from src got assertionerror no matching output for forward derivative in native functions yaml embedding bag dense backward ❓ questions and help when i try to install as in the instructions i got an error that i nailed from the logs to this command cd home documents github pytorch home envs xla bin python tools setup helpers generate code py declarations path home documents github pytorch build aten src aten declarations yaml native functions path aten src aten native native functions yaml nn path aten src force schema registration traceback most recent call last file tools setup helpers generate code py line in main file tools setup helpers generate code py line in main generate code file tools setup helpers generate code py line in generate code gen autograd python file home documents github pytorch tools autograd gen autograd py line in gen autograd python differentiability infos load derivatives file home documents github pytorch tools autograd load derivatives py line in load derivatives infos file home documents github pytorch tools autograd load derivatives py line in create differentiability info defn functions by signature functions by schema file home documents github pytorch tools autograd load derivatives py line in create differentiability info derivatives forward derivatives args with derivatives non differentiable arg names set up derivatives canonical file home documents github pytorch tools codegen context py line in wrapper return func f file home documents github pytorch tools autograd load derivatives py line in set up derivatives forward derivatives append create forward derivative f formula names file home documents github pytorch tools autograd load derivatives py line in create forward derivative assert var type is not none no matching output for forward derivative definition assertionerror no matching output for forward derivative definition in aten src aten native native functions yaml embedding bag dense backward tensor grad tensor indices tensor tensor bag size tensor maximum indices int num weights bool scale grad by freq int mode tensor per sample weights int padding idx tensor if it serves something tools autograd load derivatives py prints nativefunction func functionschema name operatorname name baseoperatorname base embedding bag dense backward inplace false dunder method false overload name arguments arguments pre self positional self arg none post self positional argument name grad type basetype name default none annotation none argument name indices type basetype name default none annotation none argument name type basetype name default none annotation none argument name bag size type basetype name default none annotation none argument name maximum indices type basetype name default none annotation none argument name num weights type basetype name default none annotation none argument name scale grad by freq type basetype name default none annotation none argument name mode type basetype name default none annotation none argument name per sample weights type optionaltype elem basetype name default none annotation none argument name padding idx type basetype name default annotation none pre tensor options kwarg only tensor options none post tensor options kwarg only out returns return name none type basetype name annotation none device guard true python module none category override none variants manual kernel registration false manual cpp binding false dispatch embedding bag dense backward cpu embedding bag dense backward cuda loc location file aten src aten native native functions yaml line structured false structured delegate none structured inherits none cpp no default args set non differentiable offsets cc ezyang gchanan bdhirsh jbschlosser alband mruberry | 1 |
95,599 | 3,954,339,967 | IssuesEvent | 2016-04-29 16:37:22 | larsiusprime/tdrpg-bugs | https://api.github.com/repos/larsiusprime/tdrpg-bugs | closed | Null Object Reference crash when Azra dies | 1 bug DQ CORE Priority HIGH | 0.5.7win7x64

Different trace than #665:
```Called from ApplicationMain::main ApplicationMain.hx line 159
Called from ApplicationMain::create ApplicationMain.hx line 71
Called from lime.app.Application::exec lime/app/Application.hx line 222
Called from lime._backend.native.NativeApplication::exec lime/_backend/native/Na
tiveApplication.hx line 136
Called from lime._backend.native.NativeApplication::handleRenderEvent lime/_back
end/native/NativeApplication.hx line 368
Called from lime.app.Event_Void_Void::dispatch lime/app/Event.hx line 145
Called from *::_Function_2_1 lime/app/Application.hx line 110
Called from lime.app.Application::render lime/app/Application.hx line 762
Called from openfl.display.Stage::render openfl/display/Stage.hx line 677
Called from openfl.display.DisplayObjectContainer::__broadcast openfl/display/Di
splayObjectContainer.hx line 355
Called from openfl.display.DisplayObject::__broadcast openfl/display/DisplayObje
ct.hx line 251
Called from openfl.events.EventDispatcher::__dispatchEvent openfl/events/EventDi
spatcher.hx line 237
Called from flixel.FlxGame::onEnterFrame flixel/FlxGame.hx line 529
Called from flixel.FlxGame::step flixel/FlxGame.hx line 662
Called from com.leveluplabs.tdrpg.Main::update com/leveluplabs/tdrpg/Main.hx lin
e 1481
Called from flixel.FlxGame::update flixel/FlxGame.hx line 751
Called from flixel.FlxState::tryUpdate flixel/FlxState.hx line 168
Called from com.leveluplabs.tdrpg.State_Battle::update com/leveluplabs/tdrpg/Sta
te_Battle.hx line 880
Called from com.leveluplabs.tdrpg.State_Battle::battleUpdate com/leveluplabs/tdr
pg/State_Battle.hx line 995
Called from com.leveluplabs.tdrpg.BattleModel::update com/leveluplabs/tdrpg/Batt
leModel.hx line 1412
Called from com.leveluplabs.tdrpg.BattleModel::doUpdate com/leveluplabs/tdrpg/Ba
ttleModel.hx line 1816
Called from com.leveluplabs.tdrpg.BattleModel::updateEnemies com/leveluplabs/tdr
pg/BattleModel.hx line 3204
Called from com.leveluplabs.tdrpg.EnemySprite::updateFromModel com/leveluplabs/t
drpg/EnemySprite.hx line 510
Called from com.leveluplabs.tdrpg.EnemySprite::updateWalk com/leveluplabs/tdrpg/
EnemySprite.hx line 2155
Called from com.leveluplabs.tdrpg.EnemySprite::getNextWalk com/leveluplabs/tdrpg
/EnemySprite.hx line 1125
Called from flixel.util._FlxSignal.FlxSignal2::dispatch2 flixel/util/FlxSignal.h
x line 292
Called from com.leveluplabs.tdrpg.BattleModel::onEnemySignal com/leveluplabs/tdr
pg/BattleModel.hx line 2670
Called from com.leveluplabs.tdrpg.BattleModel::hitMcguffin com/leveluplabs/tdrpg
/BattleModel.hx line 2174
Called from com.leveluplabs.tdrpg.BattleModel::damageMcguffin com/leveluplabs/td
rpg/BattleModel.hx line 2202``` | 1.0 | Null Object Reference crash when Azra dies - 0.5.7win7x64

Different trace than #665:
```Called from ApplicationMain::main ApplicationMain.hx line 159
Called from ApplicationMain::create ApplicationMain.hx line 71
Called from lime.app.Application::exec lime/app/Application.hx line 222
Called from lime._backend.native.NativeApplication::exec lime/_backend/native/Na
tiveApplication.hx line 136
Called from lime._backend.native.NativeApplication::handleRenderEvent lime/_back
end/native/NativeApplication.hx line 368
Called from lime.app.Event_Void_Void::dispatch lime/app/Event.hx line 145
Called from *::_Function_2_1 lime/app/Application.hx line 110
Called from lime.app.Application::render lime/app/Application.hx line 762
Called from openfl.display.Stage::render openfl/display/Stage.hx line 677
Called from openfl.display.DisplayObjectContainer::__broadcast openfl/display/Di
splayObjectContainer.hx line 355
Called from openfl.display.DisplayObject::__broadcast openfl/display/DisplayObje
ct.hx line 251
Called from openfl.events.EventDispatcher::__dispatchEvent openfl/events/EventDi
spatcher.hx line 237
Called from flixel.FlxGame::onEnterFrame flixel/FlxGame.hx line 529
Called from flixel.FlxGame::step flixel/FlxGame.hx line 662
Called from com.leveluplabs.tdrpg.Main::update com/leveluplabs/tdrpg/Main.hx lin
e 1481
Called from flixel.FlxGame::update flixel/FlxGame.hx line 751
Called from flixel.FlxState::tryUpdate flixel/FlxState.hx line 168
Called from com.leveluplabs.tdrpg.State_Battle::update com/leveluplabs/tdrpg/Sta
te_Battle.hx line 880
Called from com.leveluplabs.tdrpg.State_Battle::battleUpdate com/leveluplabs/tdr
pg/State_Battle.hx line 995
Called from com.leveluplabs.tdrpg.BattleModel::update com/leveluplabs/tdrpg/Batt
leModel.hx line 1412
Called from com.leveluplabs.tdrpg.BattleModel::doUpdate com/leveluplabs/tdrpg/Ba
ttleModel.hx line 1816
Called from com.leveluplabs.tdrpg.BattleModel::updateEnemies com/leveluplabs/tdr
pg/BattleModel.hx line 3204
Called from com.leveluplabs.tdrpg.EnemySprite::updateFromModel com/leveluplabs/t
drpg/EnemySprite.hx line 510
Called from com.leveluplabs.tdrpg.EnemySprite::updateWalk com/leveluplabs/tdrpg/
EnemySprite.hx line 2155
Called from com.leveluplabs.tdrpg.EnemySprite::getNextWalk com/leveluplabs/tdrpg
/EnemySprite.hx line 1125
Called from flixel.util._FlxSignal.FlxSignal2::dispatch2 flixel/util/FlxSignal.h
x line 292
Called from com.leveluplabs.tdrpg.BattleModel::onEnemySignal com/leveluplabs/tdr
pg/BattleModel.hx line 2670
Called from com.leveluplabs.tdrpg.BattleModel::hitMcguffin com/leveluplabs/tdrpg
/BattleModel.hx line 2174
Called from com.leveluplabs.tdrpg.BattleModel::damageMcguffin com/leveluplabs/td
rpg/BattleModel.hx line 2202``` | priority | null object reference crash when azra dies different trace than called from applicationmain main applicationmain hx line called from applicationmain create applicationmain hx line called from lime app application exec lime app application hx line called from lime backend native nativeapplication exec lime backend native na tiveapplication hx line called from lime backend native nativeapplication handlerenderevent lime back end native nativeapplication hx line called from lime app event void void dispatch lime app event hx line called from function lime app application hx line called from lime app application render lime app application hx line called from openfl display stage render openfl display stage hx line called from openfl display displayobjectcontainer broadcast openfl display di splayobjectcontainer hx line called from openfl display displayobject broadcast openfl display displayobje ct hx line called from openfl events eventdispatcher dispatchevent openfl events eventdi spatcher hx line called from flixel flxgame onenterframe flixel flxgame hx line called from flixel flxgame step flixel flxgame hx line called from com leveluplabs tdrpg main update com leveluplabs tdrpg main hx lin e called from flixel flxgame update flixel flxgame hx line called from flixel flxstate tryupdate flixel flxstate hx line called from com leveluplabs tdrpg state battle update com leveluplabs tdrpg sta te battle hx line called from com leveluplabs tdrpg state battle battleupdate com leveluplabs tdr pg state battle hx line called from com leveluplabs tdrpg battlemodel update com leveluplabs tdrpg batt lemodel hx line called from com leveluplabs tdrpg battlemodel doupdate com leveluplabs tdrpg ba ttlemodel hx line called from com leveluplabs tdrpg battlemodel updateenemies com leveluplabs tdr pg battlemodel hx line called from com leveluplabs tdrpg enemysprite updatefrommodel com leveluplabs t drpg enemysprite hx line called from com leveluplabs tdrpg enemysprite updatewalk com leveluplabs tdrpg enemysprite hx line called from com leveluplabs tdrpg enemysprite getnextwalk com leveluplabs tdrpg enemysprite hx line called from flixel util flxsignal flixel util flxsignal h x line called from com leveluplabs tdrpg battlemodel onenemysignal com leveluplabs tdr pg battlemodel hx line called from com leveluplabs tdrpg battlemodel hitmcguffin com leveluplabs tdrpg battlemodel hx line called from com leveluplabs tdrpg battlemodel damagemcguffin com leveluplabs td rpg battlemodel hx line | 1 |
583,967 | 17,402,090,572 | IssuesEvent | 2021-08-02 21:18:37 | oasis-dev/oasis-new-website | https://api.github.com/repos/oasis-dev/oasis-new-website | closed | Check WP dashboard for historical versions of all key policies | Content High priority OASIS | Issue #74 and #75 suggest that not all versioned policies were successfully migrated. Check each key policy to make sure the expected historical versions are present on the system. | 1.0 | Check WP dashboard for historical versions of all key policies - Issue #74 and #75 suggest that not all versioned policies were successfully migrated. Check each key policy to make sure the expected historical versions are present on the system. | priority | check wp dashboard for historical versions of all key policies issue and suggest that not all versioned policies were successfully migrated check each key policy to make sure the expected historical versions are present on the system | 1 |
136,141 | 5,272,046,796 | IssuesEvent | 2017-02-06 11:35:49 | BinPar/PRM | https://api.github.com/repos/BinPar/PRM | opened | INFORME PROMOCION: NO TIENE EN CUENTA LAS NUEVAS ZONAS DE NOEMÍ Y ALEX | Priority: High | A la hora de obtener el informe de promoción, solo refleja las promociones realizadas en las antiguas zonas.
No aparece ninguna acción realizada en provincias como Barcelona, Granada, Valencia, Málaga.
No he probado con el histórico, pero imagino que pasará lo mismo.
@CristianBinpar | 1.0 | INFORME PROMOCION: NO TIENE EN CUENTA LAS NUEVAS ZONAS DE NOEMÍ Y ALEX - A la hora de obtener el informe de promoción, solo refleja las promociones realizadas en las antiguas zonas.
No aparece ninguna acción realizada en provincias como Barcelona, Granada, Valencia, Málaga.
No he probado con el histórico, pero imagino que pasará lo mismo.
@CristianBinpar | priority | informe promocion no tiene en cuenta las nuevas zonas de noemí y alex a la hora de obtener el informe de promoción solo refleja las promociones realizadas en las antiguas zonas no aparece ninguna acción realizada en provincias como barcelona granada valencia málaga no he probado con el histórico pero imagino que pasará lo mismo cristianbinpar | 1 |
777,582 | 27,286,656,096 | IssuesEvent | 2023-02-23 13:58:16 | kartoza/ckanext-dalrrd-emc-dcpr | https://api.github.com/repos/kartoza/ckanext-dalrrd-emc-dcpr | closed | Number of records associated with SASDI themes is not accurate. | type:bug priority:high | When there are 0 records associated with a SASDI theme, the site still displays an incorrect number.
E.g.:
- The 'Administrative Boundaries 1' theme shows 7 associated records

- However, when the theme is selected 0 records are found

When a SASDI theme actually has associated records, the number reflected on the site appears to be correct. | 1.0 | Number of records associated with SASDI themes is not accurate. - When there are 0 records associated with a SASDI theme, the site still displays an incorrect number.
E.g.:
- The 'Administrative Boundaries 1' theme shows 7 associated records

- However, when the theme is selected 0 records are found

When a SASDI theme actually has associated records, the number reflected on the site appears to be correct. | priority | number of records associated with sasdi themes is not accurate when there are records associated with a sasdi theme the site still displays an incorrect number e g the administrative boundaries theme shows associated records however when the theme is selected records are found when a sasdi theme actually has associated records the number reflected on the site appears to be correct | 1 |
631,316 | 20,150,118,175 | IssuesEvent | 2022-02-09 11:31:07 | ita-social-projects/horondi_client_fe | https://api.github.com/repos/ita-social-projects/horondi_client_fe | closed | [Products Page. Sort by] Inconsistent items are displayed when sorting items by 'popularity' | bug data base priority: high severity: major Functional | **Environment:** Windows 10 Pro 64bit, Firefox 89.0 64bit
**Reproducible:** Always
**Pre-conditions:**
1. Go to https://horondi-front-staging.azurewebsites.net/
2. Open the 'Sort by' drop-down list
**Description:**
**Steps to reproduce:**
1. Choose 'popularity' from drop-down list
**Actual result:**
Inconsistent items are showed
**Expected result:**
The items are sorted by 'popularity'
[TC_STEP#1](https://jira.softserve.academy/browse/LVHRB-212) | 1.0 | [Products Page. Sort by] Inconsistent items are displayed when sorting items by 'popularity' - **Environment:** Windows 10 Pro 64bit, Firefox 89.0 64bit
**Reproducible:** Always
**Pre-conditions:**
1. Go to https://horondi-front-staging.azurewebsites.net/
2. Open the 'Sort by' drop-down list
**Description:**
**Steps to reproduce:**
1. Choose 'popularity' from drop-down list
**Actual result:**
Inconsistent items are showed
**Expected result:**
The items are sorted by 'popularity'
[TC_STEP#1](https://jira.softserve.academy/browse/LVHRB-212) | priority | inconsistent items are displayed when sorting items by popularity environment windows pro firefox reproducible always pre conditions go to open the sort by drop down list description steps to reproduce choose popularity from drop down list actual result inconsistent items are showed expected result the items are sorted by popularity | 1 |
308,367 | 9,438,463,676 | IssuesEvent | 2019-04-14 00:04:35 | x13pixels/remedybg-issues | https://api.github.com/repos/x13pixels/remedybg-issues | opened | Memory window alignment | Component: Memory Window Priority: 7 (Med-High) Status: Accepted Type: Bug | [direbroom, simon, casey, et al.] when entering an address in the mem window, either start the window text at that address (not aligned), or highlight the byte of the address jumped to | 1.0 | Memory window alignment - [direbroom, simon, casey, et al.] when entering an address in the mem window, either start the window text at that address (not aligned), or highlight the byte of the address jumped to | priority | memory window alignment when entering an address in the mem window either start the window text at that address not aligned or highlight the byte of the address jumped to | 1 |
1,329 | 2,511,803,736 | IssuesEvent | 2015-01-14 11:26:01 | kbauman/Capital-Sisters | https://api.github.com/repos/kbauman/Capital-Sisters | closed | finalizing map template | high priority styling | can we add a little padding under the map so it doesn't crowd the text below? right now it's running into the content header | 1.0 | finalizing map template - can we add a little padding under the map so it doesn't crowd the text below? right now it's running into the content header | priority | finalizing map template can we add a little padding under the map so it doesn t crowd the text below right now it s running into the content header | 1 |
249,485 | 7,962,449,657 | IssuesEvent | 2018-07-13 14:21:04 | poanetwork/token-wizard | https://api.github.com/repos/poanetwork/token-wizard | opened | (Feature) add source code of Proxy smart-contract to Publish page | enhancement high priority migration to auth-os to do | **Problem**: a part of #1035
**Solution**: The source code of Proxy smart-contract could be added in the textarea below all other properties of crowdsale, as it is done in `master` branch of Token Wizard (live link: https://wizard.poa.network/) | 1.0 | (Feature) add source code of Proxy smart-contract to Publish page - **Problem**: a part of #1035
**Solution**: The source code of Proxy smart-contract could be added in the textarea below all other properties of crowdsale, as it is done in `master` branch of Token Wizard (live link: https://wizard.poa.network/) | priority | feature add source code of proxy smart contract to publish page problem a part of solution the source code of proxy smart contract could be added in the textarea below all other properties of crowdsale as it is done in master branch of token wizard live link | 1 |
578,730 | 17,150,861,080 | IssuesEvent | 2021-07-13 20:24:09 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | closed | Render thesauri category when showing metadata (13 pts) | Priority: High Sprint | Thesaurus can have the values grouped in categories, as users we want the category rendered with the values so there is more context for those values.
<img width="392" alt="Screenshot 2021-05-03 at 10 26 50" src="https://user-images.githubusercontent.com/948117/116855736-2c662f00-abfa-11eb-81f1-28c602ef9d31.png">
<img width="395" alt="Screenshot 2021-05-03 at 10 26 58" src="https://user-images.githubusercontent.com/948117/116855741-2f611f80-abfa-11eb-93b1-075f10c8aa9e.png">
⚠️
Right now we are denormalising the labels of the values selected, categories will require to be denormalised too
```
[
{value: dj28djh, label: 'Art1', parent: {value: 344sd, label: Laws Chapter 2}},
{value: d2as8d, label: 'Art2', parent: {value: 321sf, label: Laws Chapter 1}},
{value: d2as8d, label: 'Art3', parent: null}
]
```
Updating the thesauri should update the denormalised value, actions like:
- Rename a group | 1.0 | Render thesauri category when showing metadata (13 pts) - Thesaurus can have the values grouped in categories, as users we want the category rendered with the values so there is more context for those values.
<img width="392" alt="Screenshot 2021-05-03 at 10 26 50" src="https://user-images.githubusercontent.com/948117/116855736-2c662f00-abfa-11eb-81f1-28c602ef9d31.png">
<img width="395" alt="Screenshot 2021-05-03 at 10 26 58" src="https://user-images.githubusercontent.com/948117/116855741-2f611f80-abfa-11eb-93b1-075f10c8aa9e.png">
⚠️
Right now we are denormalising the labels of the values selected, categories will require to be denormalised too
```
[
{value: dj28djh, label: 'Art1', parent: {value: 344sd, label: Laws Chapter 2}},
{value: d2as8d, label: 'Art2', parent: {value: 321sf, label: Laws Chapter 1}},
{value: d2as8d, label: 'Art3', parent: null}
]
```
Updating the thesauri should update the denormalised value, actions like:
- Rename a group | priority | render thesauri category when showing metadata pts thesaurus can have the values grouped in categories as users we want the category rendered with the values so there is more context for those values img width alt screenshot at src img width alt screenshot at src ⚠️ right now we are denormalising the labels of the values selected categories will require to be denormalised too value label parent value label laws chapter value label parent value label laws chapter value label parent null updating the thesauri should update the denormalised value actions like rename a group | 1 |
431,925 | 12,486,443,054 | IssuesEvent | 2020-05-31 02:10:30 | ItsDrike/CommandBot | https://api.github.com/repos/ItsDrike/CommandBot | closed | Setup Travis-CI | priority: 1 - high status: WIP type: feature | We should setup **Travis-CI** utility to run all the unit tests automatically.
This will save us a lot of time when debugging. | 1.0 | Setup Travis-CI - We should setup **Travis-CI** utility to run all the unit tests automatically.
This will save us a lot of time when debugging. | priority | setup travis ci we should setup travis ci utility to run all the unit tests automatically this will save us a lot of time when debugging | 1 |
137,796 | 5,316,523,365 | IssuesEvent | 2017-02-13 20:06:12 | influxdata/docs.influxdata.com | https://api.github.com/repos/influxdata/docs.influxdata.com | closed | Download links of Chronograph, InfluxDB and Kapacitor products are faulty. | Priority: high Type: bug |
###### URL for relevant page?
https://portal.influxdata.com/downloads
**Problem Description:**
All download links are faulty for Chronograph, InfluxDB and Kapacitor products (but not Telegraf) and result in a 404 Not Found.
**Faulty Example:**
wget https://dl.influxdata.com/telegraf/releases/influxdb_1.2.0_amd64.deb
All product links are prefixed with the telegraf url which results into all download links for all other products being faulty except for the telegraf.
**Example Solution:**
Convert
wget https://dl.influxdata.com/telegraf/releases/influxdb_1.2.0_amd64.deb
To
wget https://dl.influxdata.com/influxdb/releases/influxdb_1.2.0_amd64.deb
Probably by introducing a variable in the docs dependent on the product the user selected.
| 1.0 | Download links of Chronograph, InfluxDB and Kapacitor products are faulty. -
###### URL for relevant page?
https://portal.influxdata.com/downloads
**Problem Description:**
All download links are faulty for Chronograph, InfluxDB and Kapacitor products (but not Telegraf) and result in a 404 Not Found.
**Faulty Example:**
wget https://dl.influxdata.com/telegraf/releases/influxdb_1.2.0_amd64.deb
All product links are prefixed with the telegraf url which results into all download links for all other products being faulty except for the telegraf.
**Example Solution:**
Convert
wget https://dl.influxdata.com/telegraf/releases/influxdb_1.2.0_amd64.deb
To
wget https://dl.influxdata.com/influxdb/releases/influxdb_1.2.0_amd64.deb
Probably by introducing a variable in the docs dependent on the product the user selected.
| priority | download links of chronograph influxdb and kapacitor products are faulty url for relevant page problem description all download links are faulty for chronograph influxdb and kapacitor products but not telegraf and result in a not found faulty example wget all product links are prefixed with the telegraf url which results into all download links for all other products being faulty except for the telegraf example solution convert wget to wget probably by introducing a variable in the docs dependent on the product the user selected | 1 |
212,951 | 7,244,323,900 | IssuesEvent | 2018-02-14 14:48:44 | Rubicon-Bot/Rubicon | https://api.github.com/repos/Rubicon-Bot/Rubicon | closed | Improve Mute | Bug High Priority | <p><strong>Bugreport</strong><br><br><strong>Bug report by ForYaSee#9353 </strong><br><br><strong>Description</strong><br><br></p>Move the mute role to the top of the role Hierarchie | 1.0 | Improve Mute - <p><strong>Bugreport</strong><br><br><strong>Bug report by ForYaSee#9353 </strong><br><br><strong>Description</strong><br><br></p>Move the mute role to the top of the role Hierarchie | priority | improve mute bugreport bug report by foryasee description move the mute role to the top of the role hierarchie | 1 |
148,424 | 5,681,710,819 | IssuesEvent | 2017-04-13 07:31:16 | leo-project/leofs | https://api.github.com/repos/leo-project/leofs | opened | [leo_storage][mq] Possibility of MQ's data loss | Bug Priority-HIGH v1.3 _leo_storage | While I've been fixing #702, and coincidentally found that the current version of LeoStorage is a possibility of MQ's data loss.
When a message with the same name is registered in a very short time, it may not be reregistered into the MQ. We need to register a message with a unique key. | 1.0 | [leo_storage][mq] Possibility of MQ's data loss - While I've been fixing #702, and coincidentally found that the current version of LeoStorage is a possibility of MQ's data loss.
When a message with the same name is registered in a very short time, it may not be reregistered into the MQ. We need to register a message with a unique key. | priority | possibility of mq s data loss while i ve been fixing and coincidentally found that the current version of leostorage is a possibility of mq s data loss when a message with the same name is registered in a very short time it may not be reregistered into the mq we need to register a message with a unique key | 1 |
161,895 | 6,137,997,402 | IssuesEvent | 2017-06-26 13:42:36 | bedita/bedita | https://api.github.com/repos/bedita/bedita | closed | Error on saving artwork | Priority - High Topic - Core Type - Bug | Saving an artwork, I get
http://(domain)/artworks/save
```
Notice (8): Undefined index: description [APP/models/behaviors/notify.php, line 163]
Warning (2): Cannot modify header information - headers already sent by (output started at /home/bedita3/iebook/cake/libs/debugger.php:686) [CORE/cake/libs/controller/controller.php, line 742]
```
what I did:
In an artwork, I add a multimedia document; the document is without description
Then, when I save the artwork, I got the error
contatct me for the data example
| 1.0 | Error on saving artwork - Saving an artwork, I get
http://(domain)/artworks/save
```
Notice (8): Undefined index: description [APP/models/behaviors/notify.php, line 163]
Warning (2): Cannot modify header information - headers already sent by (output started at /home/bedita3/iebook/cake/libs/debugger.php:686) [CORE/cake/libs/controller/controller.php, line 742]
```
what I did:
In an artwork, I add a multimedia document; the document is without description
Then, when I save the artwork, I got the error
contatct me for the data example
| priority | error on saving artwork saving an artwork i get notice undefined index description warning cannot modify header information headers already sent by output started at home iebook cake libs debugger php what i did in an artwork i add a multimedia document the document is without description then when i save the artwork i got the error contatct me for the data example | 1 |
737,510 | 25,519,170,848 | IssuesEvent | 2022-11-28 18:55:02 | Cheos137/ArmorpointsPlusplus | https://api.github.com/repos/Cheos137/ArmorpointsPlusplus | closed | [Feature Request]: add option to render multiple absorption bars over health bar. | feature request minecraft: all supported future update priority: high | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
Recently I switched from this mod https://www.curseforge.com/minecraft/mc-mods/health-overlay to using that one.
This one is much more better. But I miss one feature still.
I need ability to stack absorption bar render over my health bar.
### Solution
Like that.

I also like that you can edit texture sheet instead of tinting with hex color. Will be also useful for cases when effect is reaching 20 health points.
### Alternatives
_No response_
### Additional Context
_No response_ | 1.0 | [Feature Request]: add option to render multiple absorption bars over health bar. - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Description
Recently I switched from this mod https://www.curseforge.com/minecraft/mc-mods/health-overlay to using that one.
This one is much more better. But I miss one feature still.
I need ability to stack absorption bar render over my health bar.
### Solution
Like that.

I also like that you can edit texture sheet instead of tinting with hex color. Will be also useful for cases when effect is reaching 20 health points.
### Alternatives
_No response_
### Additional Context
_No response_ | priority | add option to render multiple absorption bars over health bar is there an existing issue for this i have searched the existing issues description recently i switched from this mod to using that one this one is much more better but i miss one feature still i need ability to stack absorption bar render over my health bar solution like that i also like that you can edit texture sheet instead of tinting with hex color will be also useful for cases when effect is reaching health points alternatives no response additional context no response | 1 |
828,414 | 31,827,213,524 | IssuesEvent | 2023-09-14 08:19:15 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | Block Hooks: Incorrect placement in the parent container for hooked blocks | [Type] Bug [Priority] High [Feature] Block API [Status] In Progress | - Part of https://github.com/WordPress/gutenberg/issues/53987.
### Description
Every block that has inner blocks also has some wrapping HTML element. The way hooked blocks are currently implemented is that it puts the injected block at exactly first or last place in the parsed block, which happens to be outside of the wrapping element. A simplified example based on REST API response that shows the issue:
```html
<!-- wp:group {"tagName":"main","style":{"spacing":{"margin":{"top":"var:preset|spacing|50","bottom":"var:preset|spacing|70"}}},"layout":{"type":"constrained"}} -->
<main class="wp-block-group" style="margin-top:var(--wp--preset--spacing--50);margin-bottom:var(--wp--preset--spacing--70)">
<!-- wp:pattern {"slug":"twentytwentythree/cta"} /-->
</main>
<!-- wp:ockham/like-button /-->
<!-- /wp:group /-->
```
It should be withing the `main` tag instead:
```html
<!-- wp:group {"tagName":"main","style":{"spacing":{"margin":{"top":"var:preset|spacing|50","bottom":"var:preset|spacing|70"}}},"layout":{"type":"constrained"}} -->
<main class="wp-block-group" style="margin-top:var(--wp--preset--spacing--50);margin-bottom:var(--wp--preset--spacing--70)">
<!-- wp:pattern {"slug":"twentytwentythree/cta"} /-->
<!-- wp:ockham/like-button /-->
</main>
<!-- /wp:group /-->
```
### Step-by-step reproduction instructions
1. Install and activate the plugin with the hooked block. Example block can be downloaded from https://github.com/ockham/like-button/releases/tag/v0.3.1.
2. Activate Twenty Twenty-Three theme or any other theme that uses `core/comment-template` block.
3. Go to the single post page and see where the Like Button gets injected. Ensure that all templates and template parts used on that page don't have any customizations applied.
4. Check the source of the page and notice that the Like button is after the closing tag for the comment row wrapper.
### Screenshots, screen recording, code snippet
The configuration used for the Like Button:
<img width="372" alt="Screenshot 2023-09-07 at 17 10 48" src="https://github.com/WordPress/gutenberg/assets/699132/9b257871-cb46-4ec7-9d64-ac9a5fadf388">
The same issue exists for `firstChild` and `lastChild`. Everything works correctly for `before` and `after` positions.
The issue was hard to discover because the way it's presented in the block editor is perfectly valid:
<img width="345" alt="Screenshot 2023-09-07 at 17 12 04" src="https://github.com/WordPress/gutenberg/assets/699132/fba3cb85-3e00-4d19-bd9a-c6aeb5abf70f">
The REST API endpoint contains the hooked block as expected, at least when you don't check it with enough attention:
<img width="1553" alt="Screenshot 2023-09-07 at 17 12 51" src="https://github.com/WordPress/gutenberg/assets/699132/39f8309f-64a4-4343-b517-2be42908d7ae">
The issue is easier to spot when checking the source of the webpage because the block gets printed after (or before when using `firstChild`) the closing tag for the target parent block:
<img width="1340" alt="Screenshot 2023-09-07 at 17 13 34" src="https://github.com/WordPress/gutenberg/assets/699132/72634e20-ee4a-49cc-987f-c0837e02fd31">
It's hard to tell by looking at the page:
<img width="785" alt="Screenshot 2023-09-08 at 17 47 29" src="https://github.com/WordPress/gutenberg/assets/699132/373c2961-e631-4112-a7fe-8862911ddfad">
Well, until you use source inspector:
<img width="847" alt="Screenshot 2023-09-08 at 17 49 17" src="https://github.com/WordPress/gutenberg/assets/699132/d5d24293-b8f1-4584-9bdb-466de52186f0">
### Environment info
The latest Gutenberg version.
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes | 1.0 | Block Hooks: Incorrect placement in the parent container for hooked blocks - - Part of https://github.com/WordPress/gutenberg/issues/53987.
### Description
Every block that has inner blocks also has some wrapping HTML element. The way hooked blocks are currently implemented is that it puts the injected block at exactly first or last place in the parsed block, which happens to be outside of the wrapping element. A simplified example based on REST API response that shows the issue:
```html
<!-- wp:group {"tagName":"main","style":{"spacing":{"margin":{"top":"var:preset|spacing|50","bottom":"var:preset|spacing|70"}}},"layout":{"type":"constrained"}} -->
<main class="wp-block-group" style="margin-top:var(--wp--preset--spacing--50);margin-bottom:var(--wp--preset--spacing--70)">
<!-- wp:pattern {"slug":"twentytwentythree/cta"} /-->
</main>
<!-- wp:ockham/like-button /-->
<!-- /wp:group /-->
```
It should be withing the `main` tag instead:
```html
<!-- wp:group {"tagName":"main","style":{"spacing":{"margin":{"top":"var:preset|spacing|50","bottom":"var:preset|spacing|70"}}},"layout":{"type":"constrained"}} -->
<main class="wp-block-group" style="margin-top:var(--wp--preset--spacing--50);margin-bottom:var(--wp--preset--spacing--70)">
<!-- wp:pattern {"slug":"twentytwentythree/cta"} /-->
<!-- wp:ockham/like-button /-->
</main>
<!-- /wp:group /-->
```
### Step-by-step reproduction instructions
1. Install and activate the plugin with the hooked block. Example block can be downloaded from https://github.com/ockham/like-button/releases/tag/v0.3.1.
2. Activate Twenty Twenty-Three theme or any other theme that uses `core/comment-template` block.
3. Go to the single post page and see where the Like Button gets injected. Ensure that all templates and template parts used on that page don't have any customizations applied.
4. Check the source of the page and notice that the Like button is after the closing tag for the comment row wrapper.
### Screenshots, screen recording, code snippet
The configuration used for the Like Button:
<img width="372" alt="Screenshot 2023-09-07 at 17 10 48" src="https://github.com/WordPress/gutenberg/assets/699132/9b257871-cb46-4ec7-9d64-ac9a5fadf388">
The same issue exists for `firstChild` and `lastChild`. Everything works correctly for `before` and `after` positions.
The issue was hard to discover because the way it's presented in the block editor is perfectly valid:
<img width="345" alt="Screenshot 2023-09-07 at 17 12 04" src="https://github.com/WordPress/gutenberg/assets/699132/fba3cb85-3e00-4d19-bd9a-c6aeb5abf70f">
The REST API endpoint contains the hooked block as expected, at least when you don't check it with enough attention:
<img width="1553" alt="Screenshot 2023-09-07 at 17 12 51" src="https://github.com/WordPress/gutenberg/assets/699132/39f8309f-64a4-4343-b517-2be42908d7ae">
The issue is easier to spot when checking the source of the webpage because the block gets printed after (or before when using `firstChild`) the closing tag for the target parent block:
<img width="1340" alt="Screenshot 2023-09-07 at 17 13 34" src="https://github.com/WordPress/gutenberg/assets/699132/72634e20-ee4a-49cc-987f-c0837e02fd31">
It's hard to tell by looking at the page:
<img width="785" alt="Screenshot 2023-09-08 at 17 47 29" src="https://github.com/WordPress/gutenberg/assets/699132/373c2961-e631-4112-a7fe-8862911ddfad">
Well, until you use source inspector:
<img width="847" alt="Screenshot 2023-09-08 at 17 49 17" src="https://github.com/WordPress/gutenberg/assets/699132/d5d24293-b8f1-4584-9bdb-466de52186f0">
### Environment info
The latest Gutenberg version.
### Please confirm that you have searched existing issues in the repo.
Yes
### Please confirm that you have tested with all plugins deactivated except Gutenberg.
Yes | priority | block hooks incorrect placement in the parent container for hooked blocks part of description every block that has inner blocks also has some wrapping html element the way hooked blocks are currently implemented is that it puts the injected block at exactly first or last place in the parsed block which happens to be outside of the wrapping element a simplified example based on rest api response that shows the issue html it should be withing the main tag instead html step by step reproduction instructions install and activate the plugin with the hooked block example block can be downloaded from activate twenty twenty three theme or any other theme that uses core comment template block go to the single post page and see where the like button gets injected ensure that all templates and template parts used on that page don t have any customizations applied check the source of the page and notice that the like button is after the closing tag for the comment row wrapper screenshots screen recording code snippet the configuration used for the like button img width alt screenshot at src the same issue exists for firstchild and lastchild everything works correctly for before and after positions the issue was hard to discover because the way it s presented in the block editor is perfectly valid img width alt screenshot at src the rest api endpoint contains the hooked block as expected at least when you don t check it with enough attention img width alt screenshot at src the issue is easier to spot when checking the source of the webpage because the block gets printed after or before when using firstchild the closing tag for the target parent block img width alt screenshot at src it s hard to tell by looking at the page img width alt screenshot at src well until you use source inspector img width alt screenshot at src environment info the latest gutenberg version please confirm that you have searched existing issues in the repo yes please confirm that you have tested with all plugins deactivated except gutenberg yes | 1 |
786,943 | 27,699,441,050 | IssuesEvent | 2023-03-14 06:36:07 | rpm-software-management/dnf5 | https://api.github.com/repos/rpm-software-management/dnf5 | closed | Missing logic around installroot | Priority: HIGH | DNF configuration uses several paths that can be related to intallroot or not. As an example lets use `reposdir` configuration option with value `/etc/yum.repos.d/`. If dnf is used in installroot `/tmp/test`, the path is expected to be `/tmp/test/etc/yum.repos.d/`
In DNF4 there is an auto-detection that decide whether path is supposed be modify with `installroot` prefix or not. The mechanism is well described in DNF man pages (option `--installroot`) - see https://dnf.readthedocs.io/en/latest/command_ref.html.
The auto-detection in DNF4 has two weaknesses - it does not work correctly on RHEL and there is no way in API how to enable the auto-detection. Therefore we discussed to replace the auto-detection by two options.
- [ ] Design possibility how to set DNF5 (from commandline) to take configurations from root or installroot path
Solution examples:
- conflicting options `--installroot=<path>`, `--root=<path>`
- `--installroot=<path>`, `--use_root`
- [ ] Design API interface to modify paths in configuration class
Example of possibilities
- Method in base
- Optional argument of Base::setup()
- Configuration (bool)
- [ ] Document the new behavior also as a change to DNF4
- [ ] Resolves https://github.com/rpm-software-management/dnf5/issues/248
- [ ] Enable tests in our CI for installroot
| 1.0 | Missing logic around installroot - DNF configuration uses several paths that can be related to intallroot or not. As an example lets use `reposdir` configuration option with value `/etc/yum.repos.d/`. If dnf is used in installroot `/tmp/test`, the path is expected to be `/tmp/test/etc/yum.repos.d/`
In DNF4 there is an auto-detection that decide whether path is supposed be modify with `installroot` prefix or not. The mechanism is well described in DNF man pages (option `--installroot`) - see https://dnf.readthedocs.io/en/latest/command_ref.html.
The auto-detection in DNF4 has two weaknesses - it does not work correctly on RHEL and there is no way in API how to enable the auto-detection. Therefore we discussed to replace the auto-detection by two options.
- [ ] Design possibility how to set DNF5 (from commandline) to take configurations from root or installroot path
Solution examples:
- conflicting options `--installroot=<path>`, `--root=<path>`
- `--installroot=<path>`, `--use_root`
- [ ] Design API interface to modify paths in configuration class
Example of possibilities
- Method in base
- Optional argument of Base::setup()
- Configuration (bool)
- [ ] Document the new behavior also as a change to DNF4
- [ ] Resolves https://github.com/rpm-software-management/dnf5/issues/248
- [ ] Enable tests in our CI for installroot
| priority | missing logic around installroot dnf configuration uses several paths that can be related to intallroot or not as an example lets use reposdir configuration option with value etc yum repos d if dnf is used in installroot tmp test the path is expected to be tmp test etc yum repos d in there is an auto detection that decide whether path is supposed be modify with installroot prefix or not the mechanism is well described in dnf man pages option installroot see the auto detection in has two weaknesses it does not work correctly on rhel and there is no way in api how to enable the auto detection therefore we discussed to replace the auto detection by two options design possibility how to set from commandline to take configurations from root or installroot path solution examples conflicting options installroot root installroot use root design api interface to modify paths in configuration class example of possibilities method in base optional argument of base setup configuration bool document the new behavior also as a change to resolves enable tests in our ci for installroot | 1 |
235,952 | 7,744,430,634 | IssuesEvent | 2018-05-29 15:22:17 | Gloirin/m2gTest | https://api.github.com/repos/Gloirin/m2gTest | closed | 0004838:
remove invalid chars in pdf export | CRM bug high priority | **Reported by pschuele on 19 Sep 2011 11:36**
**Version:** Maischa (2011-05-3)
remove invalid chars in pdf export
use iconv to remove the invalid chars
**Additional information:** xception 'Tinebase_Exception_UnexpectedValue' with message 'Detected an illegal character in input string: [...] f?...' in /opt/local/tine/2011-05-3/htdocs/Tinebase/Export/Pdf.php:478
| 1.0 | 0004838:
remove invalid chars in pdf export - **Reported by pschuele on 19 Sep 2011 11:36**
**Version:** Maischa (2011-05-3)
remove invalid chars in pdf export
use iconv to remove the invalid chars
**Additional information:** xception 'Tinebase_Exception_UnexpectedValue' with message 'Detected an illegal character in input string: [...] f?...' in /opt/local/tine/2011-05-3/htdocs/Tinebase/Export/Pdf.php:478
| priority | remove invalid chars in pdf export reported by pschuele on sep version maischa remove invalid chars in pdf export use iconv to remove the invalid chars additional information xception tinebase exception unexpectedvalue with message detected an illegal character in input string f in opt local tine htdocs tinebase export pdf php | 1 |
799,796 | 28,314,168,068 | IssuesEvent | 2023-04-10 18:03:58 | AY2223S2-CS2103T-F12-2/tp | https://api.github.com/repos/AY2223S2-CS2103T-F12-2/tp | closed | Adapt all documentations and features to fit CoDoc | priority.High | There are still many traces of the parent program [AddressBook](https://github.com/nus-cs2103-AY2223S2/tp) left in many areas.
Need to fix them along development before the final delivery of the program (v1.4). | 1.0 | Adapt all documentations and features to fit CoDoc - There are still many traces of the parent program [AddressBook](https://github.com/nus-cs2103-AY2223S2/tp) left in many areas.
Need to fix them along development before the final delivery of the program (v1.4). | priority | adapt all documentations and features to fit codoc there are still many traces of the parent program left in many areas need to fix them along development before the final delivery of the program | 1 |
4,895 | 2,565,200,839 | IssuesEvent | 2015-02-07 04:02:46 | roundware/roundware-server | https://api.github.com/repos/roundware/roundware-server | closed | overlapping initial assets [REPRISE] | bug high priority | Improvements were made with the previous fix for initial overlapping assets (#154), but the issue is rearing its ugly head again. I'm not sure what changed to bring it back, but I can vouch for the fact that my testing using the Listen Map of the previous fix did show that it was fixed and the same test now shows an issue.
I have narrowed down the issue to being related to the timing of the initial ```request_stream``` call followed by a ```move_listener``` call. If ```move_listener``` happens too quickly after ```request_stream``` returns a stream, initial overlaps happen, but if ```move_listener``` happens further removed from ```request_stream```, all is OK. I doubt these values are consistent, but using the Listen Map on the FAM prod server, a wait of 1s causes the issue whereas a wait of 2s doesn't.
I’ve investigated the logs and as best I can tell the most interesting difference between the overlap and non-overlap scenarios is in these few log lines:
Overlaps DON’T happen:
```
[17/Oct/2014 15:09:57] DEBUG <roundwared.stream.get_message:229> Announcing 1 is playing.
[17/Oct/2014 15:09:58] "GET /api/1/?operation=request_stream&session_id=1&tags=3%2C4%2C5%2C8%2C9%2C22&latitude=1&longitude=1 HTTP/1.1" 200 57
[17/Oct/2014 15:10:00] DEBUG <roundwared.recording_collection.get_recording:72> Getting a recording from the bucket.
[17/Oct/2014 15:10:00] DEBUG <roundwared.recording_collection.get_recording:76> We have 0 unplayed recordings.
[17/Oct/2014 15:10:00] "GET /api/1?operation=move_listener&session_id=1&latitude=1&longitude=1 HTTP/1.1" 301 0
[17/Oct/2014 15:10:00] DEBUG <roundwared.stream.move_listener:145> stream: move_listener: recvd lat and long, moving...
[17/Oct/2014 15:10:00] DEBUG <roundwared.stream.move_listener:150> move_listener(1.0,1.0)
```
Overlaps DO happen:
```
[17/Oct/2014 14:57:14] DEBUG <roundwared.stream.get_message:229> Announcing 1 is playing.
[17/Oct/2014 14:57:14] "GET /api/1/?operation=request_stream&session_id=1&tags=3%2C4%2C5%2C8%2C9%2C22&latitude=1&longitude=1 HTTP/1.1" 200 57
[17/Oct/2014 14:57:16] "GET /api/1?operation=move_listener&session_id=1&latitude=1&longitude=1 HTTP/1.1" 301 0
[17/Oct/2014 14:57:16] DEBUG <roundwared.stream.move_listener:145> stream: move_listener: recvd lat and long, moving...
[17/Oct/2014 14:57:16] DEBUG <roundwared.stream.move_listener:150> move_listener(1.0,1.0)
```
So for some reason, in order for the overlaps to not happen, the ```get_recording``` code that spits out the additional lines above needs to run; if it doesn’t run, the first two recordings start immediately and therefore overlap. I do not yet know why that function would be skipped, but it appears that if the move_listener call happens too soon (or otherwise at just the right time) then it is skipped. | 1.0 | overlapping initial assets [REPRISE] - Improvements were made with the previous fix for initial overlapping assets (#154), but the issue is rearing its ugly head again. I'm not sure what changed to bring it back, but I can vouch for the fact that my testing using the Listen Map of the previous fix did show that it was fixed and the same test now shows an issue.
I have narrowed down the issue to being related to the timing of the initial ```request_stream``` call followed by a ```move_listener``` call. If ```move_listener``` happens too quickly after ```request_stream``` returns a stream, initial overlaps happen, but if ```move_listener``` happens further removed from ```request_stream```, all is OK. I doubt these values are consistent, but using the Listen Map on the FAM prod server, a wait of 1s causes the issue whereas a wait of 2s doesn't.
I’ve investigated the logs and as best I can tell the most interesting difference between the overlap and non-overlap scenarios is in these few log lines:
Overlaps DON’T happen:
```
[17/Oct/2014 15:09:57] DEBUG <roundwared.stream.get_message:229> Announcing 1 is playing.
[17/Oct/2014 15:09:58] "GET /api/1/?operation=request_stream&session_id=1&tags=3%2C4%2C5%2C8%2C9%2C22&latitude=1&longitude=1 HTTP/1.1" 200 57
[17/Oct/2014 15:10:00] DEBUG <roundwared.recording_collection.get_recording:72> Getting a recording from the bucket.
[17/Oct/2014 15:10:00] DEBUG <roundwared.recording_collection.get_recording:76> We have 0 unplayed recordings.
[17/Oct/2014 15:10:00] "GET /api/1?operation=move_listener&session_id=1&latitude=1&longitude=1 HTTP/1.1" 301 0
[17/Oct/2014 15:10:00] DEBUG <roundwared.stream.move_listener:145> stream: move_listener: recvd lat and long, moving...
[17/Oct/2014 15:10:00] DEBUG <roundwared.stream.move_listener:150> move_listener(1.0,1.0)
```
Overlaps DO happen:
```
[17/Oct/2014 14:57:14] DEBUG <roundwared.stream.get_message:229> Announcing 1 is playing.
[17/Oct/2014 14:57:14] "GET /api/1/?operation=request_stream&session_id=1&tags=3%2C4%2C5%2C8%2C9%2C22&latitude=1&longitude=1 HTTP/1.1" 200 57
[17/Oct/2014 14:57:16] "GET /api/1?operation=move_listener&session_id=1&latitude=1&longitude=1 HTTP/1.1" 301 0
[17/Oct/2014 14:57:16] DEBUG <roundwared.stream.move_listener:145> stream: move_listener: recvd lat and long, moving...
[17/Oct/2014 14:57:16] DEBUG <roundwared.stream.move_listener:150> move_listener(1.0,1.0)
```
So for some reason, in order for the overlaps to not happen, the ```get_recording``` code that spits out the additional lines above needs to run; if it doesn’t run, the first two recordings start immediately and therefore overlap. I do not yet know why that function would be skipped, but it appears that if the move_listener call happens too soon (or otherwise at just the right time) then it is skipped. | priority | overlapping initial assets improvements were made with the previous fix for initial overlapping assets but the issue is rearing its ugly head again i m not sure what changed to bring it back but i can vouch for the fact that my testing using the listen map of the previous fix did show that it was fixed and the same test now shows an issue i have narrowed down the issue to being related to the timing of the initial request stream call followed by a move listener call if move listener happens too quickly after request stream returns a stream initial overlaps happen but if move listener happens further removed from request stream all is ok i doubt these values are consistent but using the listen map on the fam prod server a wait of causes the issue whereas a wait of doesn t i’ve investigated the logs and as best i can tell the most interesting difference between the overlap and non overlap scenarios is in these few log lines overlaps don’t happen debug announcing is playing get api operation request stream session id tags latitude longitude http debug getting a recording from the bucket debug we have unplayed recordings get api operation move listener session id latitude longitude http debug stream move listener recvd lat and long moving debug move listener overlaps do happen debug announcing is playing get api operation request stream session id tags latitude longitude http get api operation move listener session id latitude longitude http debug stream move listener recvd lat and long moving debug move listener so for some reason in order for the overlaps to not happen the get recording code that spits out the additional lines above needs to run if it doesn’t run the first two recordings start immediately and therefore overlap i do not yet know why that function would be skipped but it appears that if the move listener call happens too soon or otherwise at just the right time then it is skipped | 1 |
276,961 | 8,614,914,944 | IssuesEvent | 2018-11-19 18:56:14 | TimSusa/midi-sliders | https://api.github.com/repos/TimSusa/midi-sliders | closed | [BUG] surface malfunction | BLOCKER Bug High Priority | Big problem in Web App (equates Electron App 0.9.0), filmed here:
[https://we.tl/t-vr3egggnRF](url)
**Description**
After reassigning all drivers in my preset:
- sliders won't move, **but** send out MIDI data correctly
- no visual feedback on the buttons
- buttons toggle ON works fine, toggle OFF not working
During filming the following debugger log was created:
[midi-sliders.herokuapp.com-1542620266081.log](https://github.com/TimSusa/midi-sliders/files/2594644/midi-sliders.herokuapp.com-1542620266081.log)
Pls let me know if you need more data. | 1.0 | [BUG] surface malfunction - Big problem in Web App (equates Electron App 0.9.0), filmed here:
[https://we.tl/t-vr3egggnRF](url)
**Description**
After reassigning all drivers in my preset:
- sliders won't move, **but** send out MIDI data correctly
- no visual feedback on the buttons
- buttons toggle ON works fine, toggle OFF not working
During filming the following debugger log was created:
[midi-sliders.herokuapp.com-1542620266081.log](https://github.com/TimSusa/midi-sliders/files/2594644/midi-sliders.herokuapp.com-1542620266081.log)
Pls let me know if you need more data. | priority | surface malfunction big problem in web app equates electron app filmed here url description after reassigning all drivers in my preset sliders won t move but send out midi data correctly no visual feedback on the buttons buttons toggle on works fine toggle off not working during filming the following debugger log was created pls let me know if you need more data | 1 |
458,719 | 13,180,367,135 | IssuesEvent | 2020-08-12 12:41:07 | zowe/api-layer | https://api.github.com/repos/zowe/api-layer | closed | x509: Design composite APIML & SAF authentication | 20PI3 Objective Priority: High Security enhancement in progress | As an APIML contributor, i want to complete and present a design proposal for x509 certificate authentication that will validate the development direction and delivery of this new security enhancement for Zowe.
Deliverables:
Create a design doc that covers:
General architecture for x509
Expectations around token validity
Expectations around certificate setup on customer site
Expectations around reconfiguring the certs on apiml
Design presented and agreed with Zowe architects at weekly Zowe meeting | 1.0 | x509: Design composite APIML & SAF authentication - As an APIML contributor, i want to complete and present a design proposal for x509 certificate authentication that will validate the development direction and delivery of this new security enhancement for Zowe.
Deliverables:
Create a design doc that covers:
General architecture for x509
Expectations around token validity
Expectations around certificate setup on customer site
Expectations around reconfiguring the certs on apiml
Design presented and agreed with Zowe architects at weekly Zowe meeting | priority | design composite apiml saf authentication as an apiml contributor i want to complete and present a design proposal for certificate authentication that will validate the development direction and delivery of this new security enhancement for zowe deliverables create a design doc that covers general architecture for expectations around token validity expectations around certificate setup on customer site expectations around reconfiguring the certs on apiml design presented and agreed with zowe architects at weekly zowe meeting | 1 |
273,327 | 8,529,295,507 | IssuesEvent | 2018-11-03 10:29:44 | CS2103-AY1819S1-F10-4/main | https://api.github.com/repos/CS2103-AY1819S1-F10-4/main | opened | Annotate code | priority.high severity.high type.bug | Due to the major refactor I did in #21 and #186 , it caused RepoSense to think that almost all code is written by me.
To fix this, one way is to configure RepoSense which we already did, and one more thing is to annotate our code. So individually, please open a PR with all your code annotated correctly.
For more info on how to annotate, visit: https://nus-cs2103-ay1819s1.github.io/cs2103-website/admin/project-w12-mid-v14.html
**Please complete this as soon as possible.**
| 1.0 | Annotate code - Due to the major refactor I did in #21 and #186 , it caused RepoSense to think that almost all code is written by me.
To fix this, one way is to configure RepoSense which we already did, and one more thing is to annotate our code. So individually, please open a PR with all your code annotated correctly.
For more info on how to annotate, visit: https://nus-cs2103-ay1819s1.github.io/cs2103-website/admin/project-w12-mid-v14.html
**Please complete this as soon as possible.**
| priority | annotate code due to the major refactor i did in and it caused reposense to think that almost all code is written by me to fix this one way is to configure reposense which we already did and one more thing is to annotate our code so individually please open a pr with all your code annotated correctly for more info on how to annotate visit please complete this as soon as possible | 1 |
667,423 | 22,472,305,824 | IssuesEvent | 2022-06-22 09:10:04 | geneontology/go-annotation | https://api.github.com/repos/geneontology/go-annotation | closed | PTN001625584 PEROXIREDOXIN (PTHR10430) | FlyBase PAINT annotation high priority | * **PTHR ID & PTN node:**
PTN001625584 PEROXIREDOXIN (PTHR10430)
* **Type of Issue: Erroneous source or erroneous propagation, or other issue**
Erroneous sources
This node is annotated with "thioredoxin peroxidase activity" (GO:0008379).
But I think it should use the new term thioredoxin-dependent peroxiredoxin activity (GO:0140824) = EC:1.11.1.24
Source annotations are:
PTN000046566 (DROME|FlyBase=FBgn0038570|UniProtKB=Q960M4) **<- I've done this one**
PTN000046576 (YEAST|SGD=S000004099|UniProtKB=P38013) @srengel
PTN001625588 (EMENI|Gene=AN8692|UniProtKB=Q5ASN8)
PTN000046599 (ARATH|TAIR=locus=2009709|UniProtKB=Q9SRZ4) @tberardini
PTN002470679 (HUMAN|HGNC=9355|UniProtKB=P30044) | 1.0 | PTN001625584 PEROXIREDOXIN (PTHR10430) - * **PTHR ID & PTN node:**
PTN001625584 PEROXIREDOXIN (PTHR10430)
* **Type of Issue: Erroneous source or erroneous propagation, or other issue**
Erroneous sources
This node is annotated with "thioredoxin peroxidase activity" (GO:0008379).
But I think it should use the new term thioredoxin-dependent peroxiredoxin activity (GO:0140824) = EC:1.11.1.24
Source annotations are:
PTN000046566 (DROME|FlyBase=FBgn0038570|UniProtKB=Q960M4) **<- I've done this one**
PTN000046576 (YEAST|SGD=S000004099|UniProtKB=P38013) @srengel
PTN001625588 (EMENI|Gene=AN8692|UniProtKB=Q5ASN8)
PTN000046599 (ARATH|TAIR=locus=2009709|UniProtKB=Q9SRZ4) @tberardini
PTN002470679 (HUMAN|HGNC=9355|UniProtKB=P30044) | priority | peroxiredoxin pthr id ptn node peroxiredoxin type of issue erroneous source or erroneous propagation or other issue erroneous sources this node is annotated with thioredoxin peroxidase activity go but i think it should use the new term thioredoxin dependent peroxiredoxin activity go ec source annotations are drome flybase uniprotkb i ve done this one yeast sgd uniprotkb srengel emeni gene uniprotkb arath tair locus uniprotkb tberardini human hgnc uniprotkb | 1 |
566,217 | 16,815,578,375 | IssuesEvent | 2021-06-17 06:56:03 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Jar files required for execution contains duplicate names | Area/ProjectAPI Priority/High Team/DevTools Type/Improvement | **Description:**
At the c2c module, we are executing the thin executable jar file with dependency jars copied to the Docker image.
When coping the dependency jars to the folder, we noticed that jar files have duplicate names. Therefore some dependency jar files are getting overridden. Can we generate the jar files with unique names?
**Steps to reproduce:**
**Affected Versions:**
Swan-Lake Beta 1
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | Jar files required for execution contains duplicate names - **Description:**
At the c2c module, we are executing the thin executable jar file with dependency jars copied to the Docker image.
When coping the dependency jars to the folder, we noticed that jar files have duplicate names. Therefore some dependency jar files are getting overridden. Can we generate the jar files with unique names?
**Steps to reproduce:**
**Affected Versions:**
Swan-Lake Beta 1
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| priority | jar files required for execution contains duplicate names description at the module we are executing the thin executable jar file with dependency jars copied to the docker image when coping the dependency jars to the folder we noticed that jar files have duplicate names therefore some dependency jar files are getting overridden can we generate the jar files with unique names steps to reproduce affected versions swan lake beta os db other environment details and versions related issues optional suggested labels optional suggested assignees optional | 1 |
592,974 | 17,935,183,130 | IssuesEvent | 2021-09-10 14:29:37 | vaticle/typedb | https://api.github.com/repos/vaticle/typedb | closed | Thing.isInferred: always returns 'false' for Attributes | type: bug priority: high | ## Description
Thing.isInferred is always false for Attributes, even when they are inferred (i.e: they are returned by a query with reasoning on, but are not returned by a query with reasoning disabled)
## Environment
1. OS (where TypeDB server runs): Mac OS 10
2. TypeDB version (and platform):TypeDB 2.3.0
3. TypeDB client: client-java
## Reproducible Steps
Steps to create the smallest reproducible scenario:
Create an entity type `person`, an attribute type `name`, and a rule:
```tql
define
rule steve-rule: when { $x isa person; } then { $x has name "Steve"; };
```
Then retrieve the attribute "Steve" in a query with reasoning enabled.
## Expected Output
`isInferred` should be true
## Actual Output
`isInferred` is false
## Additional information
This is important because new Studio will rely on the flag. | 1.0 | Thing.isInferred: always returns 'false' for Attributes - ## Description
Thing.isInferred is always false for Attributes, even when they are inferred (i.e: they are returned by a query with reasoning on, but are not returned by a query with reasoning disabled)
## Environment
1. OS (where TypeDB server runs): Mac OS 10
2. TypeDB version (and platform):TypeDB 2.3.0
3. TypeDB client: client-java
## Reproducible Steps
Steps to create the smallest reproducible scenario:
Create an entity type `person`, an attribute type `name`, and a rule:
```tql
define
rule steve-rule: when { $x isa person; } then { $x has name "Steve"; };
```
Then retrieve the attribute "Steve" in a query with reasoning enabled.
## Expected Output
`isInferred` should be true
## Actual Output
`isInferred` is false
## Additional information
This is important because new Studio will rely on the flag. | priority | thing isinferred always returns false for attributes description thing isinferred is always false for attributes even when they are inferred i e they are returned by a query with reasoning on but are not returned by a query with reasoning disabled environment os where typedb server runs mac os typedb version and platform typedb typedb client client java reproducible steps steps to create the smallest reproducible scenario create an entity type person an attribute type name and a rule tql define rule steve rule when x isa person then x has name steve then retrieve the attribute steve in a query with reasoning enabled expected output isinferred should be true actual output isinferred is false additional information this is important because new studio will rely on the flag | 1 |
568,809 | 16,989,108,320 | IssuesEvent | 2021-06-30 17:56:40 | cagov/cannabis.ca.gov | https://api.github.com/repos/cagov/cannabis.ca.gov | opened | Content bug: Email address contains extra @ for grant recipients of the Contact Us page | 1 pt Hard Launch Deliverable High Priority - Must QA | see screenshot

| 1.0 | Content bug: Email address contains extra @ for grant recipients of the Contact Us page - see screenshot

| priority | content bug email address contains extra for grant recipients of the contact us page see screenshot | 1 |
32,571 | 2,755,981,568 | IssuesEvent | 2015-04-27 02:42:16 | IIsi50MHz/chromey-calculator | https://api.github.com/repos/IIsi50MHz/chromey-calculator | closed | Move wiki content to correct location | Component-Docs Priority-High Type-Task Usability | Need to move all wiki content that was exported from Google Code, since it exported to a different branch than Github needs requires to display properly. | 1.0 | Move wiki content to correct location - Need to move all wiki content that was exported from Google Code, since it exported to a different branch than Github needs requires to display properly. | priority | move wiki content to correct location need to move all wiki content that was exported from google code since it exported to a different branch than github needs requires to display properly | 1 |
762,332 | 26,715,487,941 | IssuesEvent | 2023-01-28 12:57:49 | kitabisa/teler | https://api.github.com/repos/kitabisa/teler | closed | [BUG] Error! open internal/alert/template/telegram.tmpl: no such file or directory | Type: Bug Priority: High | **Describe the bug**
Not send slert to telegram
**To Reproduce**
Just run teler with docker
```sh
tail -f $PWD/access.log | sudo docker run -i --rm -v $PWD/teler.yaml:/teler.yaml -e TELER_CONFIG=/teler.yaml kitabisa/teler
```
or binary
```sh
tail -f ./access.log | teler -c ./teler.yaml
```
then got error:
> [ERR] Error! open internal/alert/template/telegram.tmpl: no such file or director
Steps to reproduce the behavior:
```yaml
log_format: |
<format>
....
....
alert:
active: true
provider: "telegram"
notifications:
telegram:
token: "<token>"
chat_id: "<chat_id>"
```
**Expected behavior**
I think it should send alert to telegram
**Screenshots**

**Environment (please complete the following information):**
- OS: Ubuntu 22.04
- OS version: Linux pg-laptop 5.15.0-58-generic <span>#</span>64-Ubuntu SMP Thu Jan 5 11:43:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
- teler Version v2.0.0-dev.2
| 1.0 | [BUG] Error! open internal/alert/template/telegram.tmpl: no such file or directory - **Describe the bug**
Not send slert to telegram
**To Reproduce**
Just run teler with docker
```sh
tail -f $PWD/access.log | sudo docker run -i --rm -v $PWD/teler.yaml:/teler.yaml -e TELER_CONFIG=/teler.yaml kitabisa/teler
```
or binary
```sh
tail -f ./access.log | teler -c ./teler.yaml
```
then got error:
> [ERR] Error! open internal/alert/template/telegram.tmpl: no such file or director
Steps to reproduce the behavior:
```yaml
log_format: |
<format>
....
....
alert:
active: true
provider: "telegram"
notifications:
telegram:
token: "<token>"
chat_id: "<chat_id>"
```
**Expected behavior**
I think it should send alert to telegram
**Screenshots**

**Environment (please complete the following information):**
- OS: Ubuntu 22.04
- OS version: Linux pg-laptop 5.15.0-58-generic <span>#</span>64-Ubuntu SMP Thu Jan 5 11:43:13 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux
- teler Version v2.0.0-dev.2
| priority | error open internal alert template telegram tmpl no such file or directory describe the bug not send slert to telegram to reproduce just run teler with docker sh tail f pwd access log sudo docker run i rm v pwd teler yaml teler yaml e teler config teler yaml kitabisa teler or binary sh tail f access log teler c teler yaml then got error error open internal alert template telegram tmpl no such file or director steps to reproduce the behavior yaml log format alert active true provider telegram notifications telegram token chat id expected behavior i think it should send alert to telegram screenshots environment please complete the following information os ubuntu os version linux pg laptop generic ubuntu smp thu jan utc gnu linux teler version dev | 1 |
16,249 | 2,612,845,434 | IssuesEvent | 2015-02-27 17:03:19 | nextgis/nextgislogger | https://api.github.com/repos/nextgis/nextgislogger | closed | Запись в лог координат GPS | enhancement priority:high | Была бы полезна возможность включать в настройках программы запись координат GPS.
Иногда необходимо мониторить GSM на поверхности. | 1.0 | Запись в лог координат GPS - Была бы полезна возможность включать в настройках программы запись координат GPS.
Иногда необходимо мониторить GSM на поверхности. | priority | запись в лог координат gps была бы полезна возможность включать в настройках программы запись координат gps иногда необходимо мониторить gsm на поверхности | 1 |
519,610 | 15,054,597,756 | IssuesEvent | 2021-02-03 17:38:24 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | closed | hints.leave_on_load is True, in contrary to what the changelog stated | priority: 0 - high | **Version info**:
qutebrowser v2.0.1
Git commit:
Backend: QtWebEngine (Chromium 83.0.4103.122)
Qt: 5.15.2
**Does the bug happen if you start with `--temp-basedir`?**:
Yes.
**Description**
`hints.leave_on_load` is not `False` by default, as it was stated in the changelog for the 2.0 release.

| 1.0 | hints.leave_on_load is True, in contrary to what the changelog stated - **Version info**:
qutebrowser v2.0.1
Git commit:
Backend: QtWebEngine (Chromium 83.0.4103.122)
Qt: 5.15.2
**Does the bug happen if you start with `--temp-basedir`?**:
Yes.
**Description**
`hints.leave_on_load` is not `False` by default, as it was stated in the changelog for the 2.0 release.

| priority | hints leave on load is true in contrary to what the changelog stated version info qutebrowser git commit backend qtwebengine chromium qt does the bug happen if you start with temp basedir yes description hints leave on load is not false by default as it was stated in the changelog for the release | 1 |
365,069 | 10,775,257,195 | IssuesEvent | 2019-11-03 13:10:13 | AY1920S1-CS2103T-T09-1/main | https://api.github.com/repos/AY1920S1-CS2103T-T09-1/main | closed | Fix autoschedule bugs | priority.High type.Bug | There might be wrong intepretation for autoschedule if for e.g, t/CS 2103 could be wrongly intepreted as a `tag` named `CS` and time `2103`, where instead user could meant a `tag` named that is named `CS 2103` | 1.0 | Fix autoschedule bugs - There might be wrong intepretation for autoschedule if for e.g, t/CS 2103 could be wrongly intepreted as a `tag` named `CS` and time `2103`, where instead user could meant a `tag` named that is named `CS 2103` | priority | fix autoschedule bugs there might be wrong intepretation for autoschedule if for e g t cs could be wrongly intepreted as a tag named cs and time where instead user could meant a tag named that is named cs | 1 |
741,342 | 25,790,173,431 | IssuesEvent | 2022-12-10 02:43:03 | dtcenter/METplus | https://api.github.com/repos/dtcenter/METplus | closed | Clean up METplus use case configuration files | type: enhancement priority: high requestor: NCAR reporting: DTC NOAA BASE requestor: METplus Team required: FOR DEVELOPMENT RELEASE METplus: Configuration | *Replace italics below with details for this issue.*
## Describe the Enhancement ##
*Provide a description of the enhancement request here.*
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
2792541
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Enhancement Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| 1.0 | Clean up METplus use case configuration files - *Replace italics below with details for this issue.*
## Describe the Enhancement ##
*Provide a description of the enhancement request here.*
### Time Estimate ###
*Estimate the amount of work required here.*
*Issues should represent approximately 1 to 3 days of work.*
### Sub-Issues ###
Consider breaking the enhancement down into sub-issues.
- [ ] *Add a checkbox for each sub-issue here.*
### Relevant Deadlines ###
*List relevant project deadlines here or state NONE.*
### Funding Source ###
2792541
## Define the Metadata ##
### Assignee ###
- [ ] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [ ] Select **component(s)**
- [ ] Select **priority**
- [ ] Select **requestor(s)**
### Projects and Milestone ###
- [ ] Select **Repository** and/or **Organization** level **Project(s)** or add **alert: NEED PROJECT ASSIGNMENT** label
- [ ] Select **Milestone** as the next official version or **Future Versions**
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [ ] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdatadb](https://github.com/dtcenter/METdatadb/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Enhancement Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **develop**.
Branch name: `feature_<Issue Number>_<Description>`
- [ ] Complete the development and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **develop**.
Pull request: `feature <Issue Number> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Close this issue.
| priority | clean up metplus use case configuration files replace italics below with details for this issue describe the enhancement provide a description of the enhancement request here time estimate estimate the amount of work required here issues should represent approximately to days of work sub issues consider breaking the enhancement down into sub issues add a checkbox for each sub issue here relevant deadlines list relevant project deadlines here or state none funding source define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select repository and or organization level project s or add alert need project assignment label select milestone as the next official version or future versions define related issue s consider the impact to the other metplus components enhancement checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of develop branch name feature complete the development and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into develop pull request feature define the pull request metadata as permissions allow select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version iterate until the reviewer s accept and merge your changes delete your fork or branch close this issue | 1 |
93,292 | 3,898,702,761 | IssuesEvent | 2016-04-17 08:31:41 | WhitestormJS/whitestorm.js | https://api.github.com/repos/WhitestormJS/whitestorm.js | opened | [Improve] Adding helpers [Light, Shape, World] | High [priority] in progress js r9 | We are going to add helpers such as light helper, shadowmap helper, grid, etc.
Such helpers will allow users to simply debug their 3d objects. It is very useful.

- [ ] WHS.Light
- [ ] WHS.Shape
- [ ] WHS.World | 1.0 | [Improve] Adding helpers [Light, Shape, World] - We are going to add helpers such as light helper, shadowmap helper, grid, etc.
Such helpers will allow users to simply debug their 3d objects. It is very useful.

- [ ] WHS.Light
- [ ] WHS.Shape
- [ ] WHS.World | priority | adding helpers we are going to add helpers such as light helper shadowmap helper grid etc such helpers will allow users to simply debug their objects it is very useful whs light whs shape whs world | 1 |
781,556 | 27,441,911,672 | IssuesEvent | 2023-03-02 11:38:19 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | Re-enable support for using mapping constructors with record binding patterns | Type/Improvement Priority/High Type/SpecDeviation Team/CompilerFE SwanLakeDump Area/TypeChecker Lang/BindingPatterns Deferred | Support for using record binding patterns with mapping constructors was removed with https://github.com/ballerina-platform/ballerina-lang/pull/11053/commits/009046794e35f5ea8ab3c986176b2248734be762
Need to re-enable this. I had a quick look at this, with the check added here to disable the support removed. It seems to be working. Need to test more and re-enable. | 1.0 | Re-enable support for using mapping constructors with record binding patterns - Support for using record binding patterns with mapping constructors was removed with https://github.com/ballerina-platform/ballerina-lang/pull/11053/commits/009046794e35f5ea8ab3c986176b2248734be762
Need to re-enable this. I had a quick look at this, with the check added here to disable the support removed. It seems to be working. Need to test more and re-enable. | priority | re enable support for using mapping constructors with record binding patterns support for using record binding patterns with mapping constructors was removed with need to re enable this i had a quick look at this with the check added here to disable the support removed it seems to be working need to test more and re enable | 1 |
371,660 | 10,975,475,025 | IssuesEvent | 2019-11-29 11:34:23 | conan-io/conan | https://api.github.com/repos/conan-io/conan | closed | [bug] conan download pkg/version@user/channel#randomrev downloads latest | complex: medium priority: high stage: review type: bug | Latest 1.20, Windows 10, python 3.7
With revisions disable this doesn't fail (it should). And worse, it creates an entry with revision ``randomrev`` in the cache.
Also, it seems that it is not possible to ``conan download`` a package with a given package revision. | 1.0 | [bug] conan download pkg/version@user/channel#randomrev downloads latest - Latest 1.20, Windows 10, python 3.7
With revisions disable this doesn't fail (it should). And worse, it creates an entry with revision ``randomrev`` in the cache.
Also, it seems that it is not possible to ``conan download`` a package with a given package revision. | priority | conan download pkg version user channel randomrev downloads latest latest windows python with revisions disable this doesn t fail it should and worse it creates an entry with revision randomrev in the cache also it seems that it is not possible to conan download a package with a given package revision | 1 |
588,345 | 17,658,684,498 | IssuesEvent | 2021-08-21 03:42:46 | ghostly-developer/OpenPong | https://api.github.com/repos/ghostly-developer/OpenPong | closed | Powerup | high priority | - [x] Increase Ball Size
- [x] Increase Paddle Length
- [x] Decrease Paddle Length
- [x] Butterfingers (Increase time to finish decay)
- [x] Pong Smash (inrease velocity limit on key press)
- [x] Make a sprite for all powerups | 1.0 | Powerup - - [x] Increase Ball Size
- [x] Increase Paddle Length
- [x] Decrease Paddle Length
- [x] Butterfingers (Increase time to finish decay)
- [x] Pong Smash (inrease velocity limit on key press)
- [x] Make a sprite for all powerups | priority | powerup increase ball size increase paddle length decrease paddle length butterfingers increase time to finish decay pong smash inrease velocity limit on key press make a sprite for all powerups | 1 |
708,015 | 24,327,264,241 | IssuesEvent | 2022-09-30 15:50:12 | UC-Davis-molecular-computing/scadnano | https://api.github.com/repos/UC-Davis-molecular-computing/scadnano | closed | ensure DialogType is specified each time constructor is called | high priority | I found a bug when right-clicking on a domain and clicking "set domain name". It was due to the parameter `type` not being specified when calling the `Dialog` constructor. (https://github.com/UC-Davis-molecular-computing/scadnano/pull/816/files)
This may be the only instance, but we should check all calls to the `Dialog` constructor to ensure that `type` is specified. | 1.0 | ensure DialogType is specified each time constructor is called - I found a bug when right-clicking on a domain and clicking "set domain name". It was due to the parameter `type` not being specified when calling the `Dialog` constructor. (https://github.com/UC-Davis-molecular-computing/scadnano/pull/816/files)
This may be the only instance, but we should check all calls to the `Dialog` constructor to ensure that `type` is specified. | priority | ensure dialogtype is specified each time constructor is called i found a bug when right clicking on a domain and clicking set domain name it was due to the parameter type not being specified when calling the dialog constructor this may be the only instance but we should check all calls to the dialog constructor to ensure that type is specified | 1 |
53,902 | 3,052,182,141 | IssuesEvent | 2015-08-12 13:30:04 | handsontable/handsontable | https://api.github.com/repos/handsontable/handsontable | closed | CustomBorders - handsontbale becomes too slow | Bug Guess: a day or more Priority: high | Hi guys,
in our ASP.NET MVC project we use handsontable with custom borders, which are stored in the array that comes from controller. Seems like this works fine only for small tables, for scrollable ones with ~20 rows it gets very slow. As soon as I remove borders at all table gets as quick as that in scrollable tables example.
How can I fix it? Unfortunately, we can't waive custom borders.
Thank you for your support! | 1.0 | CustomBorders - handsontbale becomes too slow - Hi guys,
in our ASP.NET MVC project we use handsontable with custom borders, which are stored in the array that comes from controller. Seems like this works fine only for small tables, for scrollable ones with ~20 rows it gets very slow. As soon as I remove borders at all table gets as quick as that in scrollable tables example.
How can I fix it? Unfortunately, we can't waive custom borders.
Thank you for your support! | priority | customborders handsontbale becomes too slow hi guys in our asp net mvc project we use handsontable with custom borders which are stored in the array that comes from controller seems like this works fine only for small tables for scrollable ones with rows it gets very slow as soon as i remove borders at all table gets as quick as that in scrollable tables example how can i fix it unfortunately we can t waive custom borders thank you for your support | 1 |
603,212 | 18,532,721,076 | IssuesEvent | 2021-10-21 08:09:46 | ballerina-platform/ballerina-standard-library | https://api.github.com/repos/ballerina-platform/ballerina-standard-library | closed | Spec Deviation in sql:ParameterizedQuery and sql:ParameterizedCallQuery | Points/1.5 Priority/High Type/Bug module/sql SwanLakeDump Team/DIU | **Description:**
As per the spec, RawTemplate should support Backtick(\`) within the Query. This can be supported by introducing the following interpolation, ${"`"}.
Currently sql:ParameterizedQuery and sql:ParameterizedCallQuery does not support ${"`"} intropolation. | 1.0 | Spec Deviation in sql:ParameterizedQuery and sql:ParameterizedCallQuery - **Description:**
As per the spec, RawTemplate should support Backtick(\`) within the Query. This can be supported by introducing the following interpolation, ${"`"}.
Currently sql:ParameterizedQuery and sql:ParameterizedCallQuery does not support ${"`"} intropolation. | priority | spec deviation in sql parameterizedquery and sql parameterizedcallquery description as per the spec rawtemplate should support backtick within the query this can be supported by introducing the following interpolation currently sql parameterizedquery and sql parameterizedcallquery does not support intropolation | 1 |
111,367 | 4,469,365,727 | IssuesEvent | 2016-08-25 12:52:17 | dotKom/onlineweb4 | https://api.github.com/repos/dotKom/onlineweb4 | opened | Inventory add/edit fails with a validation error even though fields are filled out | Priority: High Status: Available Type: Bug | Inputting this data

produces this result

| 1.0 | Inventory add/edit fails with a validation error even though fields are filled out - Inputting this data

produces this result

| priority | inventory add edit fails with a validation error even though fields are filled out inputting this data produces this result | 1 |
266,268 | 8,364,932,464 | IssuesEvent | 2018-10-04 01:53:41 | OpenBazaar/openbazaar-go | https://api.github.com/repos/OpenBazaar/openbazaar-go | opened | Bug: /ob/estimatetotal triggers a server error [multiwallet] | bug ❗❗❗high priority | Steps to reproduce:
- Be on the latest commit of the `multiwallet` branch
- Make an `POST /ob/estimatetotal` call with the following payload:
```JSON
{
"items": [
{
"listingHash": "zb2rhaJGKkQkQdPXL84teyCwKG5Mk8ukWRhgfM95fZbwUHRWd",
"quantity": 1,
"options": []
}
],
"moderator": "",
"paymentCoin": "BCH"
}
```
The server errors with the following:

| 1.0 | Bug: /ob/estimatetotal triggers a server error [multiwallet] - Steps to reproduce:
- Be on the latest commit of the `multiwallet` branch
- Make an `POST /ob/estimatetotal` call with the following payload:
```JSON
{
"items": [
{
"listingHash": "zb2rhaJGKkQkQdPXL84teyCwKG5Mk8ukWRhgfM95fZbwUHRWd",
"quantity": 1,
"options": []
}
],
"moderator": "",
"paymentCoin": "BCH"
}
```
The server errors with the following:

| priority | bug ob estimatetotal triggers a server error steps to reproduce be on the latest commit of the multiwallet branch make an post ob estimatetotal call with the following payload json items listinghash quantity options moderator paymentcoin bch the server errors with the following | 1 |
290,629 | 8,901,589,264 | IssuesEvent | 2019-01-17 03:09:59 | phetsims/sun | https://api.github.com/repos/phetsims/sun | closed | maxWidth for Dialog | priority:2-high | In https://github.com/phetsims/joist/issues/546, we used maxWidth on KeyboardHelpDialog. The change used Dialog's optional layoutStrategy:
```js
layoutStrategy: function( dialog, simBounds, screenBounds, scale ) {
// empirically determined to have some margin relative to max width
dialog.maxWidth = ( simBounds.width / scale ) * 0.90;
// this is the default centering strategy in Dialog
dialog.center = simBounds.center.times( 1.0 / scale );
}
```
But `simBounds` is the global bounds of the simulation so the dialog can extend outside of dev bounds if there is space in wider screens. In https://github.com/phetsims/friction/issues/158 it was mentioned that this was odd because most dialogs are constrained to the dev bounds. Questions for developer meeting:
1) Is it OK for the KeyboardHelpDialog to extend outside of dev bounds when there is space on wider screens?
2) If not, is there a good general way we can restrict the dialog to dev bounds?
3) Is this something that all Dialogs should do? Most PhET dialogs are within dev bounds because content or Text in the Dialog have maxWidth set individually. | 1.0 | maxWidth for Dialog - In https://github.com/phetsims/joist/issues/546, we used maxWidth on KeyboardHelpDialog. The change used Dialog's optional layoutStrategy:
```js
layoutStrategy: function( dialog, simBounds, screenBounds, scale ) {
// empirically determined to have some margin relative to max width
dialog.maxWidth = ( simBounds.width / scale ) * 0.90;
// this is the default centering strategy in Dialog
dialog.center = simBounds.center.times( 1.0 / scale );
}
```
But `simBounds` is the global bounds of the simulation so the dialog can extend outside of dev bounds if there is space in wider screens. In https://github.com/phetsims/friction/issues/158 it was mentioned that this was odd because most dialogs are constrained to the dev bounds. Questions for developer meeting:
1) Is it OK for the KeyboardHelpDialog to extend outside of dev bounds when there is space on wider screens?
2) If not, is there a good general way we can restrict the dialog to dev bounds?
3) Is this something that all Dialogs should do? Most PhET dialogs are within dev bounds because content or Text in the Dialog have maxWidth set individually. | priority | maxwidth for dialog in we used maxwidth on keyboardhelpdialog the change used dialog s optional layoutstrategy js layoutstrategy function dialog simbounds screenbounds scale empirically determined to have some margin relative to max width dialog maxwidth simbounds width scale this is the default centering strategy in dialog dialog center simbounds center times scale but simbounds is the global bounds of the simulation so the dialog can extend outside of dev bounds if there is space in wider screens in it was mentioned that this was odd because most dialogs are constrained to the dev bounds questions for developer meeting is it ok for the keyboardhelpdialog to extend outside of dev bounds when there is space on wider screens if not is there a good general way we can restrict the dialog to dev bounds is this something that all dialogs should do most phet dialogs are within dev bounds because content or text in the dialog have maxwidth set individually | 1 |
176,555 | 6,560,945,032 | IssuesEvent | 2017-09-07 11:20:02 | rathena/rathena | https://api.github.com/repos/rathena/rathena | closed | mapserver crash with getitem3 | bug:script mode:prerenewal mode:renewal priority:high server:map status:confirmed | * **rAthena Hash**: 60d5912
* **Client Date**: 20160203
* **Server Mode**: PRE-RENEWAL
* **Description of Issue**:
triggering getitem3 will cause mapserver to crash.
| 1.0 | mapserver crash with getitem3 - * **rAthena Hash**: 60d5912
* **Client Date**: 20160203
* **Server Mode**: PRE-RENEWAL
* **Description of Issue**:
triggering getitem3 will cause mapserver to crash.
| priority | mapserver crash with rathena hash client date server mode pre renewal description of issue triggering will cause mapserver to crash | 1 |
231,329 | 7,626,104,784 | IssuesEvent | 2018-05-04 01:01:45 | hackoregon/civic-devops | https://api.github.com/repos/hackoregon/civic-devops | opened | Create ECS services for Civic-2017 and Civic-2018 | Priority: high | These will be the two ECS services for the web frontend. Both are very similar to the django services in that they are a containerized web service with the need to expose a single port.
The port for these two services is 3000.
The resources requirements is unknown, but it ought to be extremely low considered how thing the server-side component of the websites are. | 1.0 | Create ECS services for Civic-2017 and Civic-2018 - These will be the two ECS services for the web frontend. Both are very similar to the django services in that they are a containerized web service with the need to expose a single port.
The port for these two services is 3000.
The resources requirements is unknown, but it ought to be extremely low considered how thing the server-side component of the websites are. | priority | create ecs services for civic and civic these will be the two ecs services for the web frontend both are very similar to the django services in that they are a containerized web service with the need to expose a single port the port for these two services is the resources requirements is unknown but it ought to be extremely low considered how thing the server side component of the websites are | 1 |
222,625 | 7,434,594,401 | IssuesEvent | 2018-03-26 11:36:59 | alakajam-team/alakajam | https://api.github.com/repos/alakajam-team/alakajam | closed | [SQLite] Deleting an entry makes the server hang forever | bug high priority | 1. Launch dev server with clean database.
2. Log in as `administrator`.
3. Head to http://localhost:8000/2nd-alakajam/2/super-game/edit.
4. Click `Delete`.
For me, the server just stops responding. Eventually I get this in the console:
```
10:55:20.655 ERROR (/core/middleware.js:310) Knex: Timeout acquiring a connection. The pool is probably full. Are you missing a .transacting(trx) call?
TimeoutError: Knex: Timeout acquiring a connection. The pool is probably full. Are you missing a .transacting(trx) call?
at Timeout._onTimeout (/home/thomas/alakajam/node_modules/knex/lib/client.js:287:18)
at ontimeout (timers.js:460:11)
at tryOnTimeout (timers.js:298:5)
at Timer.listOnTimeout (timers.js:261:5)
From previous event:
at Client_SQLite3.acquireConnection (/home/thomas/alakajam/node_modules/knex/lib/client.js:280:12)
at /home/thomas/alakajam/node_modules/knex/lib/runner.js:208:30
From previous event:
at /home/thomas/alakajam/node_modules/knex/lib/runner.js:206:35
From previous event:
at Runner.ensureConnection (/home/thomas/alakajam/node_modules/knex/lib/runner.js:205:34)
at Runner.run (/home/thomas/alakajam/node_modules/knex/lib/runner.js:51:42)
at Builder.Target.then (/home/thomas/alakajam/node_modules/knex/lib/interface.js:35:43)
at runCallback (timers.js:757:18)
at tryOnImmediate (timers.js:718:5)
at processImmediate [as _immediateCallback] (timers.js:698:5)
``` | 1.0 | [SQLite] Deleting an entry makes the server hang forever - 1. Launch dev server with clean database.
2. Log in as `administrator`.
3. Head to http://localhost:8000/2nd-alakajam/2/super-game/edit.
4. Click `Delete`.
For me, the server just stops responding. Eventually I get this in the console:
```
10:55:20.655 ERROR (/core/middleware.js:310) Knex: Timeout acquiring a connection. The pool is probably full. Are you missing a .transacting(trx) call?
TimeoutError: Knex: Timeout acquiring a connection. The pool is probably full. Are you missing a .transacting(trx) call?
at Timeout._onTimeout (/home/thomas/alakajam/node_modules/knex/lib/client.js:287:18)
at ontimeout (timers.js:460:11)
at tryOnTimeout (timers.js:298:5)
at Timer.listOnTimeout (timers.js:261:5)
From previous event:
at Client_SQLite3.acquireConnection (/home/thomas/alakajam/node_modules/knex/lib/client.js:280:12)
at /home/thomas/alakajam/node_modules/knex/lib/runner.js:208:30
From previous event:
at /home/thomas/alakajam/node_modules/knex/lib/runner.js:206:35
From previous event:
at Runner.ensureConnection (/home/thomas/alakajam/node_modules/knex/lib/runner.js:205:34)
at Runner.run (/home/thomas/alakajam/node_modules/knex/lib/runner.js:51:42)
at Builder.Target.then (/home/thomas/alakajam/node_modules/knex/lib/interface.js:35:43)
at runCallback (timers.js:757:18)
at tryOnImmediate (timers.js:718:5)
at processImmediate [as _immediateCallback] (timers.js:698:5)
``` | priority | deleting an entry makes the server hang forever launch dev server with clean database log in as administrator head to click delete for me the server just stops responding eventually i get this in the console error core middleware js knex timeout acquiring a connection the pool is probably full are you missing a transacting trx call timeouterror knex timeout acquiring a connection the pool is probably full are you missing a transacting trx call at timeout ontimeout home thomas alakajam node modules knex lib client js at ontimeout timers js at tryontimeout timers js at timer listontimeout timers js from previous event at client acquireconnection home thomas alakajam node modules knex lib client js at home thomas alakajam node modules knex lib runner js from previous event at home thomas alakajam node modules knex lib runner js from previous event at runner ensureconnection home thomas alakajam node modules knex lib runner js at runner run home thomas alakajam node modules knex lib runner js at builder target then home thomas alakajam node modules knex lib interface js at runcallback timers js at tryonimmediate timers js at processimmediate timers js | 1 |
462,473 | 13,247,873,898 | IssuesEvent | 2020-08-19 17:59:08 | RobotLocomotion/drake | https://api.github.com/repos/RobotLocomotion/drake | closed | Port examples/rod2d to SceneGraph | component: geometry illustration priority: high team: dynamics type: onramp | Currently, `examples/rod2d` uses the attic-based drake visualizer glue. We should port it to use `SceneGraph` instead.
Relates #12158.
https://github.com/RobotLocomotion/drake/blob/4ca6218ebff1aed9bf096ea01589899e9804015c/examples/rod2d/rod2d_sim.cc#L88-L91 | 1.0 | Port examples/rod2d to SceneGraph - Currently, `examples/rod2d` uses the attic-based drake visualizer glue. We should port it to use `SceneGraph` instead.
Relates #12158.
https://github.com/RobotLocomotion/drake/blob/4ca6218ebff1aed9bf096ea01589899e9804015c/examples/rod2d/rod2d_sim.cc#L88-L91 | priority | port examples to scenegraph currently examples uses the attic based drake visualizer glue we should port it to use scenegraph instead relates | 1 |
311,086 | 9,528,211,723 | IssuesEvent | 2019-04-29 07:44:42 | wazuh/wazuh-splunk | https://api.github.com/repos/wazuh/wazuh-splunk | closed | Add SCA checks | priority/high | Tasks:
- [x] Add SCA regex fields in the checks table.
- [x] Add description field in the checks table.
- [x] Add policy checksum.
- [x] Add the policy ID to the table.
- [x] Add filter bar.
- [x] Add stats in title bar. | 1.0 | Add SCA checks - Tasks:
- [x] Add SCA regex fields in the checks table.
- [x] Add description field in the checks table.
- [x] Add policy checksum.
- [x] Add the policy ID to the table.
- [x] Add filter bar.
- [x] Add stats in title bar. | priority | add sca checks tasks add sca regex fields in the checks table add description field in the checks table add policy checksum add the policy id to the table add filter bar add stats in title bar | 1 |
217,356 | 7,320,926,834 | IssuesEvent | 2018-03-02 09:33:31 | status-im/status-react | https://api.github.com/repos/status-im/status-react | opened | Error: No item 1 in vector of length 1 | bug high-priority | ### Description
*Type*: Bug
*Summary*: Error when relogin
`1:37 I/ReactNativeJS: NAME: "Error"
01:37 I/ReactNativeJS: MESSAGE: Error: No item 1 in vector of length 1
01:37 I/ReactNativeJS:
01:37 I/ReactNativeJS: This error is located at:
01:37 I/ReactNativeJS: in home
01:37 I/ReactNativeJS: in reagent21
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in reagent20
01:37 I/ReactNativeJS: in with-activity-indicator
01:37 I/ReactNativeJS: in reagent18
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in main-tabs
01:37 I/ReactNativeJS: in reagent17
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in main
01:37 I/ReactNativeJS: in root
01:37 I/ReactNativeJS: in reagent1
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in t`
#### Actual behavior

### Reproduction
- Open Status
- Logout
- Login back and wait for some time
It happens also when deleting chat and loggin back
### Additional Information
* Status version: latest nightly 260
* Operating System: Android, iOS
#### Logs
https://app.testfairy.com/projects/4803622-status/builds/7812289/sessions/3/?accessToken=t8sODtp2IvtLTkVG4nOrqNGFttY | 1.0 | Error: No item 1 in vector of length 1 - ### Description
*Type*: Bug
*Summary*: Error when relogin
`1:37 I/ReactNativeJS: NAME: "Error"
01:37 I/ReactNativeJS: MESSAGE: Error: No item 1 in vector of length 1
01:37 I/ReactNativeJS:
01:37 I/ReactNativeJS: This error is located at:
01:37 I/ReactNativeJS: in home
01:37 I/ReactNativeJS: in reagent21
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in reagent20
01:37 I/ReactNativeJS: in with-activity-indicator
01:37 I/ReactNativeJS: in reagent18
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in main-tabs
01:37 I/ReactNativeJS: in reagent17
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in main
01:37 I/ReactNativeJS: in root
01:37 I/ReactNativeJS: in reagent1
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in RCTView
01:37 I/ReactNativeJS: in t`
#### Actual behavior

### Reproduction
- Open Status
- Logout
- Login back and wait for some time
It happens also when deleting chat and loggin back
### Additional Information
* Status version: latest nightly 260
* Operating System: Android, iOS
#### Logs
https://app.testfairy.com/projects/4803622-status/builds/7812289/sessions/3/?accessToken=t8sODtp2IvtLTkVG4nOrqNGFttY | priority | error no item in vector of length description type bug summary error when relogin i reactnativejs name error i reactnativejs message error no item in vector of length i reactnativejs i reactnativejs this error is located at i reactnativejs in home i reactnativejs in i reactnativejs in rctview i reactnativejs in i reactnativejs in with activity indicator i reactnativejs in i reactnativejs in rctview i reactnativejs in rctview i reactnativejs in main tabs i reactnativejs in i reactnativejs in rctview i reactnativejs in main i reactnativejs in root i reactnativejs in i reactnativejs in rctview i reactnativejs in rctview i reactnativejs in t actual behavior reproduction open status logout login back and wait for some time it happens also when deleting chat and loggin back additional information status version latest nightly operating system android ios logs | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.