Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 855 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 13 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 240k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
210,625 | 7,191,620,426 | IssuesEvent | 2018-02-02 21:47:01 | RhoInc/safety-results-over-time | https://api.github.com/repos/RhoInc/safety-results-over-time | closed | Add ordering variable setting to settings.time_settings and use it to order the x-axis domain. | high priority | # Test notes
* Visits are sorted according to the corresponding numeric visit variable specified in `time_settings.order_col` if it exists.
* Visits are sorted alphanumerically otherwise. | 1.0 | Add ordering variable setting to settings.time_settings and use it to order the x-axis domain. - # Test notes
* Visits are sorted according to the corresponding numeric visit variable specified in `time_settings.order_col` if it exists.
* Visits are sorted alphanumerically otherwise. | priority | add ordering variable setting to settings time settings and use it to order the x axis domain test notes visits are sorted according to the corresponding numeric visit variable specified in time settings order col if it exists visits are sorted alphanumerically otherwise | 1 |
136,060 | 5,270,579,327 | IssuesEvent | 2017-02-06 05:54:30 | mantidproject/mantid | https://api.github.com/repos/mantidproject/mantid | closed | PyChop bugfixes | Component: Direct Inelastic Misc: Bug Priority: High | Instrument scientists on LET and MERLIN have identified several necessary changes to PyChop to fix bugs or for useability reasons:
1. On Merlin, PyChop currently outputs all possible Ei reps, even where the incident neutron flux means that such reps would be unobservable. A restriction that Ei < 7 meV should be applied to not display low energy reps which cannot be seen. LET should have a similar restriction for Ei > 30 meV.
2. On LET, the pulse removal chopper frequency should be less than 150 Hz (physical limit of the chopper).
3. On LET, in "Expert mode", the chopper 2 phase box disappears. Also "Expert Mode" should show another tab which displays the moderator and chopper time components of the overall resolution.
4. On LET, the frequency of chopper 4 in high resolution mode should be the same as the pulse removal chopper (chopper 3) frequency - not the resolution chopper frequency divided by 2 as at present (for other modes, it should stay as the resolution chopper frequency / 2).
5. LET has new resolution chopper disk, with slots of size 31, 20 and 15mm, so for high resolution mode, chopper 5 slot_width should now be 15mm. | 1.0 | PyChop bugfixes - Instrument scientists on LET and MERLIN have identified several necessary changes to PyChop to fix bugs or for useability reasons:
1. On Merlin, PyChop currently outputs all possible Ei reps, even where the incident neutron flux means that such reps would be unobservable. A restriction that Ei < 7 meV should be applied to not display low energy reps which cannot be seen. LET should have a similar restriction for Ei > 30 meV.
2. On LET, the pulse removal chopper frequency should be less than 150 Hz (physical limit of the chopper).
3. On LET, in "Expert mode", the chopper 2 phase box disappears. Also "Expert Mode" should show another tab which displays the moderator and chopper time components of the overall resolution.
4. On LET, the frequency of chopper 4 in high resolution mode should be the same as the pulse removal chopper (chopper 3) frequency - not the resolution chopper frequency divided by 2 as at present (for other modes, it should stay as the resolution chopper frequency / 2).
5. LET has new resolution chopper disk, with slots of size 31, 20 and 15mm, so for high resolution mode, chopper 5 slot_width should now be 15mm. | priority | pychop bugfixes instrument scientists on let and merlin have identified several necessary changes to pychop to fix bugs or for useability reasons on merlin pychop currently outputs all possible ei reps even where the incident neutron flux means that such reps would be unobservable a restriction that ei mev on let the pulse removal chopper frequency should be less than hz physical limit of the chopper on let in expert mode the chopper phase box disappears also expert mode should show another tab which displays the moderator and chopper time components of the overall resolution on let the frequency of chopper in high resolution mode should be the same as the pulse removal chopper chopper frequency not the resolution chopper frequency divided by as at present for other modes it should stay as the resolution chopper frequency let has new resolution chopper disk with slots of size and so for high resolution mode chopper slot width should now be | 1 |
461,281 | 13,227,998,442 | IssuesEvent | 2020-08-18 05:00:39 | Bruce773/lyric-keeper | https://api.github.com/repos/Bruce773/lyric-keeper | closed | Auto-scroll on lyrics pages is broken | Bug TIMEFRAME: High-priority | #### Description (1):
- User can decrease scroll below "-500"
- When scroll is decreased past "-500", the site breaks
#### Description (2):
- If the user does not stop auto-scroll before clicking "BACK", the homepage scrolls
#### Reproduce (1):
1. Go to any lyric page
2. Click the decrease button several times
3. Watch the site break!
#### Reproduce (2):
1. Go to any lyric page
2. Turn on auto-scroll
3. Click the back button
4. Watch the homepage scroll on it's own! | 1.0 | Auto-scroll on lyrics pages is broken - #### Description (1):
- User can decrease scroll below "-500"
- When scroll is decreased past "-500", the site breaks
#### Description (2):
- If the user does not stop auto-scroll before clicking "BACK", the homepage scrolls
#### Reproduce (1):
1. Go to any lyric page
2. Click the decrease button several times
3. Watch the site break!
#### Reproduce (2):
1. Go to any lyric page
2. Turn on auto-scroll
3. Click the back button
4. Watch the homepage scroll on it's own! | priority | auto scroll on lyrics pages is broken description user can decrease scroll below when scroll is decreased past the site breaks description if the user does not stop auto scroll before clicking back the homepage scrolls reproduce go to any lyric page click the decrease button several times watch the site break reproduce go to any lyric page turn on auto scroll click the back button watch the homepage scroll on it s own | 1 |
720,622 | 24,799,222,492 | IssuesEvent | 2022-10-24 20:05:41 | red-hat-storage/ocs-ci | https://api.github.com/repos/red-hat-storage/ocs-ci | closed | Ensure Add OSDs happen before or after upgrade | High Priority lifecycle/stale | Based on the PG imbalance issues and bz and the huge timeout it may need, we need to separate out the add osd logic out of upgrade suite to keep it clean for regression runs. | 1.0 | Ensure Add OSDs happen before or after upgrade - Based on the PG imbalance issues and bz and the huge timeout it may need, we need to separate out the add osd logic out of upgrade suite to keep it clean for regression runs. | priority | ensure add osds happen before or after upgrade based on the pg imbalance issues and bz and the huge timeout it may need we need to separate out the add osd logic out of upgrade suite to keep it clean for regression runs | 1 |
795,351 | 28,070,492,248 | IssuesEvent | 2023-03-29 18:40:39 | QuiltMC/quiltflower | https://api.github.com/repos/QuiltMC/quiltflower | closed | Decompiler produces a strange unneeded cast but it's `(<unknown>)` | bug Priority: High Subsystem: Variables | Source:
```java
public void test() {
boolean vvv1 = true, vvv2 = false;
try {
if (vvv2) {
throw new RuntimeException();
}
vvv2 = vvv1;
} finally {
char vvv31;
}
vvv1 = vvv2;
Object vvv34 = null;
while (vvv34 != null) {
vvv2 = vvv1;
System.out.println(vvv34);
vvv1 = vvv2;
}
}
```
Result:
```java
public void test() {
boolean vvv1 = true;
boolean vvv2 = false;
try {
if (vvv2) {
throw new RuntimeException();
}
vvv2 = vvv1;
} finally {
;
}
boolean var8 = (<unknown>)vvv2; // what?
for(Object vvv34 = null; vvv34 != null; var8 = var8) {
System.out.println(vvv34);
}
}
``` | 1.0 | Decompiler produces a strange unneeded cast but it's `(<unknown>)` - Source:
```java
public void test() {
boolean vvv1 = true, vvv2 = false;
try {
if (vvv2) {
throw new RuntimeException();
}
vvv2 = vvv1;
} finally {
char vvv31;
}
vvv1 = vvv2;
Object vvv34 = null;
while (vvv34 != null) {
vvv2 = vvv1;
System.out.println(vvv34);
vvv1 = vvv2;
}
}
```
Result:
```java
public void test() {
boolean vvv1 = true;
boolean vvv2 = false;
try {
if (vvv2) {
throw new RuntimeException();
}
vvv2 = vvv1;
} finally {
;
}
boolean var8 = (<unknown>)vvv2; // what?
for(Object vvv34 = null; vvv34 != null; var8 = var8) {
System.out.println(vvv34);
}
}
``` | priority | decompiler produces a strange unneeded cast but it s source java public void test boolean true false try if throw new runtimeexception finally char object null while null system out println result java public void test boolean true boolean false try if throw new runtimeexception finally boolean what for object null null system out println | 1 |
345,474 | 10,368,113,329 | IssuesEvent | 2019-09-07 14:22:51 | ahmedkaludi/accelerated-mobile-pages | https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages | closed | Change Internal Links to AMP | NEXT UPDATE [Priority: HIGH] bug | Hi guys,
The option **Change Internal Links to AMP** is changing my links to subdomains as well.
example: pebmed.com.br/amp has an external link to events.pebmed.com.br in which is added /amp. This generates a broken link as bellow:
events.pebmed.com.br/amp | 1.0 | Change Internal Links to AMP - Hi guys,
The option **Change Internal Links to AMP** is changing my links to subdomains as well.
example: pebmed.com.br/amp has an external link to events.pebmed.com.br in which is added /amp. This generates a broken link as bellow:
events.pebmed.com.br/amp | priority | change internal links to amp hi guys the option change internal links to amp is changing my links to subdomains as well example pebmed com br amp has an external link to events pebmed com br in which is added amp this generates a broken link as bellow events pebmed com br amp | 1 |
358,666 | 10,619,153,590 | IssuesEvent | 2019-10-13 11:09:41 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Plasteel Explodes (only a little), causing R-Walls to spam explosion logs when blown up | Oversight Priority: High | Blow up the supermatter while as an admin and see the fun for yourself. This is so severe that if it hits enough r-walls it can cause client crashing. | 1.0 | Plasteel Explodes (only a little), causing R-Walls to spam explosion logs when blown up - Blow up the supermatter while as an admin and see the fun for yourself. This is so severe that if it hits enough r-walls it can cause client crashing. | priority | plasteel explodes only a little causing r walls to spam explosion logs when blown up blow up the supermatter while as an admin and see the fun for yourself this is so severe that if it hits enough r walls it can cause client crashing | 1 |
483,480 | 13,925,415,161 | IssuesEvent | 2020-10-21 16:48:55 | eventespresso/barista | https://api.github.com/repos/eventespresso/barista | closed | Block Ticket Price Updates for Tickets with Registrations | C: data systems 🗑 D: EDTR ✏️ P2: HIGH priority 😮 T: bug 🐞 | From Garth:
> I'm able to edit ticket prices after registrations are submitted (pending or approved). That used to not be allowed.
&&
> i was able to BULK EDIT a ticket price after registrations were approved.
>
> You'll notice that on this event, the first ticket is called "Free Ticket". I have two attendees that registered and were approved, then, I BULK EDITED the tickets together and changed the prices to a non-zero dollar amount. Now the free ticket is no longer free. In master, we prohibited people from changing the price of tickets once a registration had been complete as to not confuse the accounting of registrations and transactions: https://barista.eventespresso.com/wp-admin/admin.php?page=espresso_events&action=edit&post=40
For single tickets we need to disable both the Ticket Price inline edit input as well as the Ticket Price Calculator for any ticket with registrations, and display an appropriate notice in the tooltip (for the TPC icon button) and on the Ticket Price form step.
For bulk editing we can't simply disable the selector checkbox for any tickets with registrations because then you would be unable to edit any other details of the ticket which should still be possible.
So let's do the following:
- detect tickets with registrations (sold > 0) and if any are selected (either individually or via the "select all" checkbox):
- disable the "edit ticket prices" option in the bulk actions dropdown
- display an error between the table and the bulk actions dropdown
Here's a suggestion for the appropriate notice text:
> Ticket price modifications are blocked for Tickets that have already been sold to registrants, because doing so would negatively affect internal accounting for the event. If you still need to modify ticket prices, then create a copy of those tickets, edit the prices for the new tickets, and then archive the old tickets.
That's a quick fix that puts the responsibility for that process on the user, whereas the legacy editor would automatically copy and archive a ticket when the price was modified. The legacy editor was not capable of bulk editing however. A more complex solution would be to add an option to perform the copy and archive process for users automatically via bulk editing, but that might be tricky.
| 1.0 | Block Ticket Price Updates for Tickets with Registrations - From Garth:
> I'm able to edit ticket prices after registrations are submitted (pending or approved). That used to not be allowed.
&&
> i was able to BULK EDIT a ticket price after registrations were approved.
>
> You'll notice that on this event, the first ticket is called "Free Ticket". I have two attendees that registered and were approved, then, I BULK EDITED the tickets together and changed the prices to a non-zero dollar amount. Now the free ticket is no longer free. In master, we prohibited people from changing the price of tickets once a registration had been complete as to not confuse the accounting of registrations and transactions: https://barista.eventespresso.com/wp-admin/admin.php?page=espresso_events&action=edit&post=40
For single tickets we need to disable both the Ticket Price inline edit input as well as the Ticket Price Calculator for any ticket with registrations, and display an appropriate notice in the tooltip (for the TPC icon button) and on the Ticket Price form step.
For bulk editing we can't simply disable the selector checkbox for any tickets with registrations because then you would be unable to edit any other details of the ticket which should still be possible.
So let's do the following:
- detect tickets with registrations (sold > 0) and if any are selected (either individually or via the "select all" checkbox):
- disable the "edit ticket prices" option in the bulk actions dropdown
- display an error between the table and the bulk actions dropdown
Here's a suggestion for the appropriate notice text:
> Ticket price modifications are blocked for Tickets that have already been sold to registrants, because doing so would negatively affect internal accounting for the event. If you still need to modify ticket prices, then create a copy of those tickets, edit the prices for the new tickets, and then archive the old tickets.
That's a quick fix that puts the responsibility for that process on the user, whereas the legacy editor would automatically copy and archive a ticket when the price was modified. The legacy editor was not capable of bulk editing however. A more complex solution would be to add an option to perform the copy and archive process for users automatically via bulk editing, but that might be tricky.
| priority | block ticket price updates for tickets with registrations from garth i m able to edit ticket prices after registrations are submitted pending or approved that used to not be allowed i was able to bulk edit a ticket price after registrations were approved you ll notice that on this event the first ticket is called free ticket i have two attendees that registered and were approved then i bulk edited the tickets together and changed the prices to a non zero dollar amount now the free ticket is no longer free in master we prohibited people from changing the price of tickets once a registration had been complete as to not confuse the accounting of registrations and transactions for single tickets we need to disable both the ticket price inline edit input as well as the ticket price calculator for any ticket with registrations and display an appropriate notice in the tooltip for the tpc icon button and on the ticket price form step for bulk editing we can t simply disable the selector checkbox for any tickets with registrations because then you would be unable to edit any other details of the ticket which should still be possible so let s do the following detect tickets with registrations sold and if any are selected either individually or via the select all checkbox disable the edit ticket prices option in the bulk actions dropdown display an error between the table and the bulk actions dropdown here s a suggestion for the appropriate notice text ticket price modifications are blocked for tickets that have already been sold to registrants because doing so would negatively affect internal accounting for the event if you still need to modify ticket prices then create a copy of those tickets edit the prices for the new tickets and then archive the old tickets that s a quick fix that puts the responsibility for that process on the user whereas the legacy editor would automatically copy and archive a ticket when the price was modified the legacy editor was not capable of bulk editing however a more complex solution would be to add an option to perform the copy and archive process for users automatically via bulk editing but that might be tricky | 1 |
613,025 | 19,071,219,140 | IssuesEvent | 2021-11-27 00:08:58 | HEPData/hepdata | https://api.github.com/repos/HEPData/hepdata | closed | records: broken download link on resource file landing page and special treatment needed for .html files | type: bug priority: high complexity: medium | Some records migrated from the old HepData site, for example, [ins1208923](https://www.hepdata.net/record/ins1208923) ([old HepData](http://hepdata.cedar.ac.uk/view/ins1208923)) have resources like:
```yaml
additional_resources:
- {description: Rivet analysis, location: 'http://rivet.hepforge.org/analyses#CMS_2013_I1208923'}
- {description: Link to correlation matrices for the Inclusive Jet data, location: /resource/9901922/inclusive_corr.html}
- {description: Link to fastNLO v2.1 table (inclusive jets), location: /resource/1208923/fnl2332d_I1208923.tab.gz}
- {description: Link to correlation matrices for the Dijet data, location: /resource/9901922/dijet_corr.html}
- {description: Link to fastNLO v2.1 table (dijet mass), location: /resource/1208923/fnl2412e_I1208923.tab.gz}
- {description: Link to the detailed systematics for the Inclusive Jet data, location: /resource/9901922/inclusive_sys.html}
- {description: Link to the detailed systematics for the DiJet data, location: /resource/9901922/dijet_sys.html}
```
The resource files are stored in a subdirectory of the main data directory (`/opt/hepdata/var/data/94/66887/resources`). The last six resources are correctly given a DOI, but the "Download" links are broken from the resource file landing pages, although they work correctly from the "Additional Publication Resources" widget.
Moreover, the landing page for the `.html` files just downloads the `.html` file rather than return the landing page. Presumably this is because the media type is `text/html`, so we might have to treat this as a special case, e.g. if the media type is `text/html` only return the content directly instead of the landing page if the `flask.request.accept_mimetypes` list has length 1, since [most web browsers](https://developer.mozilla.org/en-US/docs/Web/HTTP/Content_negotiation/List_of_default_Accept_values) will send an `Accept` header with multiple media types.
Could you please address these issues, @alisonrclarke? | 1.0 | records: broken download link on resource file landing page and special treatment needed for .html files - Some records migrated from the old HepData site, for example, [ins1208923](https://www.hepdata.net/record/ins1208923) ([old HepData](http://hepdata.cedar.ac.uk/view/ins1208923)) have resources like:
```yaml
additional_resources:
- {description: Rivet analysis, location: 'http://rivet.hepforge.org/analyses#CMS_2013_I1208923'}
- {description: Link to correlation matrices for the Inclusive Jet data, location: /resource/9901922/inclusive_corr.html}
- {description: Link to fastNLO v2.1 table (inclusive jets), location: /resource/1208923/fnl2332d_I1208923.tab.gz}
- {description: Link to correlation matrices for the Dijet data, location: /resource/9901922/dijet_corr.html}
- {description: Link to fastNLO v2.1 table (dijet mass), location: /resource/1208923/fnl2412e_I1208923.tab.gz}
- {description: Link to the detailed systematics for the Inclusive Jet data, location: /resource/9901922/inclusive_sys.html}
- {description: Link to the detailed systematics for the DiJet data, location: /resource/9901922/dijet_sys.html}
```
The resource files are stored in a subdirectory of the main data directory (`/opt/hepdata/var/data/94/66887/resources`). The last six resources are correctly given a DOI, but the "Download" links are broken from the resource file landing pages, although they work correctly from the "Additional Publication Resources" widget.
Moreover, the landing page for the `.html` files just downloads the `.html` file rather than return the landing page. Presumably this is because the media type is `text/html`, so we might have to treat this as a special case, e.g. if the media type is `text/html` only return the content directly instead of the landing page if the `flask.request.accept_mimetypes` list has length 1, since [most web browsers](https://developer.mozilla.org/en-US/docs/Web/HTTP/Content_negotiation/List_of_default_Accept_values) will send an `Accept` header with multiple media types.
Could you please address these issues, @alisonrclarke? | priority | records broken download link on resource file landing page and special treatment needed for html files some records migrated from the old hepdata site for example have resources like yaml additional resources description rivet analysis location description link to correlation matrices for the inclusive jet data location resource inclusive corr html description link to fastnlo table inclusive jets location resource tab gz description link to correlation matrices for the dijet data location resource dijet corr html description link to fastnlo table dijet mass location resource tab gz description link to the detailed systematics for the inclusive jet data location resource inclusive sys html description link to the detailed systematics for the dijet data location resource dijet sys html the resource files are stored in a subdirectory of the main data directory opt hepdata var data resources the last six resources are correctly given a doi but the download links are broken from the resource file landing pages although they work correctly from the additional publication resources widget moreover the landing page for the html files just downloads the html file rather than return the landing page presumably this is because the media type is text html so we might have to treat this as a special case e g if the media type is text html only return the content directly instead of the landing page if the flask request accept mimetypes list has length since will send an accept header with multiple media types could you please address these issues alisonrclarke | 1 |
327,172 | 9,967,648,434 | IssuesEvent | 2019-07-08 14:02:17 | VandyHacks/vaken | https://api.github.com/repos/VandyHacks/vaken | closed | Don't generate OAuth users with a hardcoded password | Priority: High Topic: Security Type: Bug | This is a potential security vulnerability.
in `src/server/auth.ts` | 1.0 | Don't generate OAuth users with a hardcoded password - This is a potential security vulnerability.
in `src/server/auth.ts` | priority | don t generate oauth users with a hardcoded password this is a potential security vulnerability in src server auth ts | 1 |
751,739 | 26,255,654,549 | IssuesEvent | 2023-01-06 00:17:42 | apache/incubator-devlake | https://api.github.com/repos/apache/incubator-devlake | closed | [Feature][frame] transformationRules has set default value. | type/feature-request priority/high Stale | ### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/incubator-devlake/issues?q=is%3Aissue) and found no similar feature requirement.
### Description
when users use devlake, such as the github plugin, the user doesn't set the transformationRules, but the final data results have converted by the default transformationRules, Is this reasonable? Should the backend remove these default handling?
The main focus I think is that the values displayed in the front-end interface are for user reference or are they intended to be used as default values?
1. Front-end transformationRules show:

2. Backend transformationRules show:

### Use case
_No response_
### Related issues
_No response_
### Are you willing to submit a PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
| 1.0 | [Feature][frame] transformationRules has set default value. - ### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/incubator-devlake/issues?q=is%3Aissue) and found no similar feature requirement.
### Description
when users use devlake, such as the github plugin, the user doesn't set the transformationRules, but the final data results have converted by the default transformationRules, Is this reasonable? Should the backend remove these default handling?
The main focus I think is that the values displayed in the front-end interface are for user reference or are they intended to be used as default values?
1. Front-end transformationRules show:

2. Backend transformationRules show:

### Use case
_No response_
### Related issues
_No response_
### Are you willing to submit a PR?
- [ ] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
| priority | transformationrules has set default value search before asking i had searched in the and found no similar feature requirement description when users use devlake such as the github plugin the user doesn t set the transformationrules but the final data results have converted by the default transformationrules is this reasonable should the backend remove these default handling? the main focus i think is that the values displayed in the front end interface are for user reference or are they intended to be used as default values front end transformationrules show backend transformationrules show use case no response related issues no response are you willing to submit a pr yes i am willing to submit a pr code of conduct i agree to follow this project s | 1 |
469,362 | 13,507,359,988 | IssuesEvent | 2020-09-14 05:46:46 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | apps.facebook.com - site is not usable | browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical | <!-- @browser: Firefox 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/58151 -->
**URL**: https://apps.facebook.com/doubleucasino/?force_login=1&casinoidx=&gameidx=&subtype=&adflag=
**Browser / Version**: Firefox 81.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
dont work
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200910180444</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/9/d66f84c4-3a80-4358-9511-17e5841f196d)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | apps.facebook.com - site is not usable - <!-- @browser: Firefox 81.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/58151 -->
**URL**: https://apps.facebook.com/doubleucasino/?force_login=1&casinoidx=&gameidx=&subtype=&adflag=
**Browser / Version**: Firefox 81.0
**Operating System**: Windows 7
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
dont work
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200910180444</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/9/d66f84c4-3a80-4358-9511-17e5841f196d)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | apps facebook com site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce dont work browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 1 |
199,979 | 6,996,774,488 | IssuesEvent | 2017-12-16 05:11:31 | dart-lang/pub | https://api.github.com/repos/dart-lang/pub | reopened | Pub should handle pub server errors gracefully | enhancement high priority | The pub server will sometimes return a 500 (internal error), if there is a temporary failure on the server.
pub get should handle these results, and retry the fetch of the object from the server. If it doesn't retry, then it should return with a distinctive error code that indicates temporary failure. Maybe this should happen for other network failures when running pub get as well.
This would make the testing of pub get less flaky, and allow us to test it on the buildbot.
This seems related to issue #1534 and pr https://github.com/dart-lang/pub/pull/1552
@mkustermann | 1.0 | Pub should handle pub server errors gracefully - The pub server will sometimes return a 500 (internal error), if there is a temporary failure on the server.
pub get should handle these results, and retry the fetch of the object from the server. If it doesn't retry, then it should return with a distinctive error code that indicates temporary failure. Maybe this should happen for other network failures when running pub get as well.
This would make the testing of pub get less flaky, and allow us to test it on the buildbot.
This seems related to issue #1534 and pr https://github.com/dart-lang/pub/pull/1552
@mkustermann | priority | pub should handle pub server errors gracefully the pub server will sometimes return a internal error if there is a temporary failure on the server pub get should handle these results and retry the fetch of the object from the server if it doesn t retry then it should return with a distinctive error code that indicates temporary failure maybe this should happen for other network failures when running pub get as well this would make the testing of pub get less flaky and allow us to test it on the buildbot this seems related to issue and pr mkustermann | 1 |
502,591 | 14,562,494,558 | IssuesEvent | 2020-12-17 00:13:44 | Pocket-PC-Planet/PPCP-AppManager | https://api.github.com/repos/Pocket-PC-Planet/PPCP-AppManager | closed | Link handler not working | Core High Priority bug | Nothing happens, no messages appear when enabled in dev mode when link clicked. Tested on WM5 HP iPAQ rx1955. | 1.0 | Link handler not working - Nothing happens, no messages appear when enabled in dev mode when link clicked. Tested on WM5 HP iPAQ rx1955. | priority | link handler not working nothing happens no messages appear when enabled in dev mode when link clicked tested on hp ipaq | 1 |
693,165 | 23,765,382,293 | IssuesEvent | 2022-09-01 12:23:56 | HumanExposure/factotum | https://api.github.com/repos/HumanExposure/factotum | closed | Bump elasticsearch | bug Priority :: High | Deployments are showing that the main factotum container is failing to start up because elasticsearch is failing to start. The logs aren't very informative either. On the other hand, API container is up and running - likely because API container lacks the `depends_on` relationship with elasticsearch the way the main factotum container does.
Based on trial and error, bumping elasticsearch to 7.17.6 has shown to be a possible solution, at least locally, for reasons unknown (documentation could not be found).
Acceptance criteria:
- bump elasticsearch from 7.16.1 to 7.17.6
- confirm successful application startup and completion of logstash pipeline on deployments
logs:
```
{"log":"\u0009... 1 more\n","stream":"stderr","time":"2022-09-01T02:05:03.627891882Z"}
{"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/Level\n","stream":"stderr","time":"2022-09-01T02:06:05.972415467Z"}
{"log":"\u0009at org.elasticsearch.cli.CommandLoggingConfigurator.configureLoggingWithoutConfig(CommandLoggingConfigurator.java:27)\n","stream":"stderr","time":"2022-09-01T02:06:05.972526009Z"}
--
{"log":"\u0009... 3 more\n","stream":"stderr","time":"2022-09-01T02:06:05.973176009Z"}
{"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/status/StatusListener\n","stream":"stderr","time":"2022-09-01T02:06:07.197484171Z"}
{"log":"\u0009at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:78)\n","stream":"stderr","time":"2022-09-01T02:06:07.197583131Z"}
``` | 1.0 | Bump elasticsearch - Deployments are showing that the main factotum container is failing to start up because elasticsearch is failing to start. The logs aren't very informative either. On the other hand, API container is up and running - likely because API container lacks the `depends_on` relationship with elasticsearch the way the main factotum container does.
Based on trial and error, bumping elasticsearch to 7.17.6 has shown to be a possible solution, at least locally, for reasons unknown (documentation could not be found).
Acceptance criteria:
- bump elasticsearch from 7.16.1 to 7.17.6
- confirm successful application startup and completion of logstash pipeline on deployments
logs:
```
{"log":"\u0009... 1 more\n","stream":"stderr","time":"2022-09-01T02:05:03.627891882Z"}
{"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/Level\n","stream":"stderr","time":"2022-09-01T02:06:05.972415467Z"}
{"log":"\u0009at org.elasticsearch.cli.CommandLoggingConfigurator.configureLoggingWithoutConfig(CommandLoggingConfigurator.java:27)\n","stream":"stderr","time":"2022-09-01T02:06:05.972526009Z"}
--
{"log":"\u0009... 3 more\n","stream":"stderr","time":"2022-09-01T02:06:05.973176009Z"}
{"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/status/StatusListener\n","stream":"stderr","time":"2022-09-01T02:06:07.197484171Z"}
{"log":"\u0009at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:78)\n","stream":"stderr","time":"2022-09-01T02:06:07.197583131Z"}
``` | priority | bump elasticsearch deployments are showing that the main factotum container is failing to start up because elasticsearch is failing to start the logs aren t very informative either on the other hand api container is up and running likely because api container lacks the depends on relationship with elasticsearch the way the main factotum container does based on trial and error bumping elasticsearch to has shown to be a possible solution at least locally for reasons unknown documentation could not be found acceptance criteria bump elasticsearch from to confirm successful application startup and completion of logstash pipeline on deployments logs log more n stream stderr time log exception in thread main java lang noclassdeffounderror org apache logging level n stream stderr time log org elasticsearch cli commandloggingconfigurator configureloggingwithoutconfig commandloggingconfigurator java n stream stderr time log more n stream stderr time log exception in thread main java lang noclassdeffounderror org apache logging status statuslistener n stream stderr time log org elasticsearch bootstrap elasticsearch main elasticsearch java n stream stderr time | 1 |
360,776 | 10,697,258,735 | IssuesEvent | 2019-10-23 16:09:21 | wso2/analytics-apim | https://api.github.com/repos/wso2/analytics-apim | closed | Subscriptions Over time table count isn't getting populated | Priority/Highest Severity/Critical | **Description:**
Subscriptions Over time table count isn't getting populated.
**Affected Product Version:**
wso2am-analytics-3.0.0
**OS, DB, other environment details and versions:**
Chrome
<img width="764" alt="Screen Shot 2019-10-23 at 11 49 21 AM" src="https://user-images.githubusercontent.com/19324135/67363633-74c77a00-f58b-11e9-8fe2-fd60ba79fb35.png">
| 1.0 | Subscriptions Over time table count isn't getting populated - **Description:**
Subscriptions Over time table count isn't getting populated.
**Affected Product Version:**
wso2am-analytics-3.0.0
**OS, DB, other environment details and versions:**
Chrome
<img width="764" alt="Screen Shot 2019-10-23 at 11 49 21 AM" src="https://user-images.githubusercontent.com/19324135/67363633-74c77a00-f58b-11e9-8fe2-fd60ba79fb35.png">
| priority | subscriptions over time table count isn t getting populated description subscriptions over time table count isn t getting populated affected product version analytics os db other environment details and versions chrome img width alt screen shot at am src | 1 |
48,806 | 3,000,030,036 | IssuesEvent | 2015-07-23 22:11:49 | GRIS-UdeM/ZirkOSC | https://api.github.com/repos/GRIS-UdeM/ZirkOSC | opened | Source elevation à 90° change l'azimuth à 180° | bug high priority | Lorsque l'on change l'élévation d'une source avec le slider et qu'on atteint 90°, son azimut est automatiquement changé à 180°. Ensuite de cela, lorsque l'on redescend l'élévation, l'azimut alterne entre 180° et -180°.
<https://dl.dropboxusercontent.com/u/2295576/GRIS/ZirkOSC_3_0_0_Circular02.mov> | 1.0 | Source elevation à 90° change l'azimuth à 180° - Lorsque l'on change l'élévation d'une source avec le slider et qu'on atteint 90°, son azimut est automatiquement changé à 180°. Ensuite de cela, lorsque l'on redescend l'élévation, l'azimut alterne entre 180° et -180°.
<https://dl.dropboxusercontent.com/u/2295576/GRIS/ZirkOSC_3_0_0_Circular02.mov> | priority | source elevation à ° change l azimuth à ° lorsque l on change l élévation d une source avec le slider et qu on atteint ° son azimut est automatiquement changé à ° ensuite de cela lorsque l on redescend l élévation l azimut alterne entre ° et ° | 1 |
471,803 | 13,610,811,623 | IssuesEvent | 2020-09-23 07:57:41 | wso2/product-apim-tooling | https://api.github.com/repos/wso2/product-apim-tooling | closed | Cannot do PUBLISHED -> CREATED transition by updating api.yaml | Affected/3.1.0 Priority/High Type/Bug | **Description:**
PUBLISHED -> CREATED transition is a valid API Lifecycle transition which is normally done by "Demote to Created" lifecycle action.
By editing the api.yaml of an exported API, we should be able to to that transition but it is not possible and an error is thrown in APIM.
**Affected Product Version:**
3.1.0, 3.2.0-RC2
**Steps to reproduce:**
1. Create an publish an API
2. Export the API using APICTL
3. Extract the export API zip file
4. Update the api.yaml of the extracted ZIP
`status: CREATED` from `status: PUBLISHED`
5. Import the API with --update
```
$ apictl import-api -f Hello-1.0/ -e 310 --update
Existing API found, attempting to update it...
API ID: 3381cfcc-e9b4-47d4-8036-8e9143fb6fa1
Error importing API.
Status: 500
Response: {"code":500,"message":"Internal server error","description":"Error while importing API","moreInfo":"","error":[]}
apictl: Error importing API Reason: 500
Exit status 1
```
Below error is thrown in the APIM side.
```
[2020-08-10 21:50:18,281] ERROR - ImportApiServiceImpl Error while importing API
org.wso2.carbon.apimgt.impl.importexport.APIImportExportException: Error while importing API: Hello version: 1.0
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:356) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI(APIImportUtil.java:115) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive_aroundBody8(APIImportExportManager.java:154) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive(APIImportExportManager.java:144) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.rest.api.admin.v1.impl.ImportApiServiceImpl.importApiPost(ImportApiServiceImpl.java:103) [classes/:?]
at org.wso2.carbon.apimgt.rest.api.admin.v1.ImportApi.importApiPost(ImportApi.java:58) [classes/:?]
at sun.reflect.GeneratedMethodAccessor273.invoke(Unknown Source) ~[?:?]
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_231]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_231]
at org.apache.cxf.service.invoker.AbstractInvoker.performInvocation(AbstractInvoker.java:179) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.service.invoker.AbstractInvoker.invoke(AbstractInvoker.java:96) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:193) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:103) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor$1.run(ServiceInvokerInterceptor.java:59) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor.handleMessage(ServiceInvokerInterceptor.java:96) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.phase.PhaseInterceptorChain.doIntercept(PhaseInterceptorChain.java:308) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.ChainInitiationObserver.onMessage(ChainInitiationObserver.java:121) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.http.AbstractHTTPDestination.invoke(AbstractHTTPDestination.java:267) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.ServletController.invokeDestination(ServletController.java:234) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:208) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:160) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.CXFNonSpringServlet.invoke(CXFNonSpringServlet.java:216) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.handleRequest(AbstractHTTPServlet.java:301) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.doPost(AbstractHTTPServlet.java:220) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:660) [tomcat-servlet-api_9.0.31.wso2v1.jar:?]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.service(AbstractHTTPServlet.java:276) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:541) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92) [tomcat_9.0.31.wso2v1.jar:?]
at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:86) [org.wso2.carbon.identity.context.rewrite.valve_1.4.0.jar:?]
at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:110) [org.wso2.carbon.identity.authz.valve_1.4.0.jar:?]
at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:75) [org.wso2.carbon.identity.auth.valve_1.4.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:49) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:145) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:688) [tomcat_9.0.31.wso2v1.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve.invoke(RequestCorrelationIdValve.java:119) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:367) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:65) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:868) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1639) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat_9.0.31.wso2v1.jar:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_231]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_231]
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.31.wso2v1.jar:?]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_231]
Caused by: org.wso2.carbon.apimgt.api.APIManagementException: Error while setting registry lifecycle checklist items for the API: Hello
at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException_aroundBody132(AbstractAPIManager.java:1869) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException(AbstractAPIManager.java:1868) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5777) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
... 56 more
Caused by: org.wso2.carbon.governance.api.exception.GovernanceException: Invalid check list item.
at org.wso2.carbon.governance.api.common.dataobjects.GovernanceArtifactImpl.isLCItemChecked(GovernanceArtifactImpl.java:897) ~[org.wso2.carbon.governance.api_4.8.19.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5769) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
... 56 more
``` | 1.0 | Cannot do PUBLISHED -> CREATED transition by updating api.yaml - **Description:**
PUBLISHED -> CREATED transition is a valid API Lifecycle transition which is normally done by "Demote to Created" lifecycle action.
By editing the api.yaml of an exported API, we should be able to to that transition but it is not possible and an error is thrown in APIM.
**Affected Product Version:**
3.1.0, 3.2.0-RC2
**Steps to reproduce:**
1. Create an publish an API
2. Export the API using APICTL
3. Extract the export API zip file
4. Update the api.yaml of the extracted ZIP
`status: CREATED` from `status: PUBLISHED`
5. Import the API with --update
```
$ apictl import-api -f Hello-1.0/ -e 310 --update
Existing API found, attempting to update it...
API ID: 3381cfcc-e9b4-47d4-8036-8e9143fb6fa1
Error importing API.
Status: 500
Response: {"code":500,"message":"Internal server error","description":"Error while importing API","moreInfo":"","error":[]}
apictl: Error importing API Reason: 500
Exit status 1
```
Below error is thrown in the APIM side.
```
[2020-08-10 21:50:18,281] ERROR - ImportApiServiceImpl Error while importing API
org.wso2.carbon.apimgt.impl.importexport.APIImportExportException: Error while importing API: Hello version: 1.0
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:356) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI(APIImportUtil.java:115) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive_aroundBody8(APIImportExportManager.java:154) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive(APIImportExportManager.java:144) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.rest.api.admin.v1.impl.ImportApiServiceImpl.importApiPost(ImportApiServiceImpl.java:103) [classes/:?]
at org.wso2.carbon.apimgt.rest.api.admin.v1.ImportApi.importApiPost(ImportApi.java:58) [classes/:?]
at sun.reflect.GeneratedMethodAccessor273.invoke(Unknown Source) ~[?:?]
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_231]
at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_231]
at org.apache.cxf.service.invoker.AbstractInvoker.performInvocation(AbstractInvoker.java:179) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.service.invoker.AbstractInvoker.invoke(AbstractInvoker.java:96) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:193) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8]
at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:103) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor$1.run(ServiceInvokerInterceptor.java:59) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.interceptor.ServiceInvokerInterceptor.handleMessage(ServiceInvokerInterceptor.java:96) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.phase.PhaseInterceptorChain.doIntercept(PhaseInterceptorChain.java:308) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.ChainInitiationObserver.onMessage(ChainInitiationObserver.java:121) [cxf-core-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.http.AbstractHTTPDestination.invoke(AbstractHTTPDestination.java:267) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.ServletController.invokeDestination(ServletController.java:234) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:208) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:160) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.CXFNonSpringServlet.invoke(CXFNonSpringServlet.java:216) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.handleRequest(AbstractHTTPServlet.java:301) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.doPost(AbstractHTTPServlet.java:220) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at javax.servlet.http.HttpServlet.service(HttpServlet.java:660) [tomcat-servlet-api_9.0.31.wso2v1.jar:?]
at org.apache.cxf.transport.servlet.AbstractHTTPServlet.service(AbstractHTTPServlet.java:276) [cxf-rt-transports-http-3.2.8.jar:3.2.8]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:541) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92) [tomcat_9.0.31.wso2v1.jar:?]
at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:86) [org.wso2.carbon.identity.context.rewrite.valve_1.4.0.jar:?]
at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:110) [org.wso2.carbon.identity.authz.valve_1.4.0.jar:?]
at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:75) [org.wso2.carbon.identity.auth.valve_1.4.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:49) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:145) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:688) [tomcat_9.0.31.wso2v1.jar:?]
at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve.invoke(RequestCorrelationIdValve.java:119) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?]
at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:367) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:65) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:868) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1639) [tomcat_9.0.31.wso2v1.jar:?]
at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat_9.0.31.wso2v1.jar:?]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_231]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_231]
at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.31.wso2v1.jar:?]
at java.lang.Thread.run(Thread.java:748) [?:1.8.0_231]
Caused by: org.wso2.carbon.apimgt.api.APIManagementException: Error while setting registry lifecycle checklist items for the API: Hello
at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException_aroundBody132(AbstractAPIManager.java:1869) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException(AbstractAPIManager.java:1868) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5777) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
... 56 more
Caused by: org.wso2.carbon.governance.api.exception.GovernanceException: Invalid check list item.
at org.wso2.carbon.governance.api.common.dataobjects.GovernanceArtifactImpl.isLCItemChecked(GovernanceArtifactImpl.java:897) ~[org.wso2.carbon.governance.api_4.8.19.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5769) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?]
... 56 more
``` | priority | cannot do published created transition by updating api yaml description published created transition is a valid api lifecycle transition which is normally done by demote to created lifecycle action by editing the api yaml of an exported api we should be able to to that transition but it is not possible and an error is thrown in apim affected product version steps to reproduce create an publish an api export the api using apictl extract the export api zip file update the api yaml of the extracted zip status created from status published import the api with update apictl import api f hello e update existing api found attempting to update it api id error importing api status response code message internal server error description error while importing api moreinfo error apictl error importing api reason exit status below error is thrown in the apim side error importapiserviceimpl error while importing api org carbon apimgt impl importexport apiimportexportexception error while importing api hello version at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java at org carbon apimgt impl importexport apiimportexportmanager importapiarchive apiimportexportmanager java at org carbon apimgt impl importexport apiimportexportmanager importapiarchive apiimportexportmanager java at org carbon apimgt rest api admin impl importapiserviceimpl importapipost importapiserviceimpl java at org carbon apimgt rest api admin importapi importapipost importapi java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache cxf service invoker abstractinvoker performinvocation abstractinvoker java at org apache cxf service invoker abstractinvoker invoke abstractinvoker java at org apache cxf jaxrs jaxrsinvoker invoke jaxrsinvoker java at org apache cxf jaxrs jaxrsinvoker invoke jaxrsinvoker java at org apache cxf interceptor serviceinvokerinterceptor run serviceinvokerinterceptor java at org apache cxf interceptor serviceinvokerinterceptor handlemessage serviceinvokerinterceptor java at org apache cxf phase phaseinterceptorchain dointercept phaseinterceptorchain java at org apache cxf transport chaininitiationobserver onmessage chaininitiationobserver java at org apache cxf transport http abstracthttpdestination invoke abstracthttpdestination java at org apache cxf transport servlet servletcontroller invokedestination servletcontroller java at org apache cxf transport servlet servletcontroller invoke servletcontroller java at org apache cxf transport servlet servletcontroller invoke servletcontroller java at org apache cxf transport servlet cxfnonspringservlet invoke cxfnonspringservlet java at org apache cxf transport servlet abstracthttpservlet handlerequest abstracthttpservlet java at org apache cxf transport servlet abstracthttpservlet dopost abstracthttpservlet java at javax servlet http httpservlet service httpservlet java at org apache cxf transport servlet abstracthttpservlet service abstracthttpservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org carbon identity context rewrite valve tenantcontextrewritevalve invoke tenantcontextrewritevalve java at org carbon identity authz valve authorizationvalve invoke authorizationvalve java at org carbon identity auth valve authenticationvalve invoke authenticationvalve java at org carbon tomcat ext valves compositevalve continueinvocation compositevalve java at org carbon tomcat ext valves tomcatvalvecontainer invokevalves tomcatvalvecontainer java at org carbon tomcat ext valves compositevalve invoke compositevalve java at org carbon tomcat ext valves carbonstuckthreaddetectionvalve invoke carbonstuckthreaddetectionvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org carbon tomcat ext valves carboncontextcreatorvalve invoke carboncontextcreatorvalve java at org carbon tomcat ext valves requestcorrelationidvalve invoke requestcorrelationidvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote service java at org apache coyote abstractprocessorlight process abstractprocessorlight java at org apache coyote abstractprotocol connectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net socketprocessorbase run socketprocessorbase java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java caused by org carbon apimgt api apimanagementexception error while setting registry lifecycle checklist items for the api hello at org carbon apimgt impl abstractapimanager handleexception abstractapimanager java at org carbon apimgt impl abstractapimanager handleexception abstractapimanager java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java more caused by org carbon governance api exception governanceexception invalid check list item at org carbon governance api common dataobjects governanceartifactimpl islcitemchecked governanceartifactimpl java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java more | 1 |
27,508 | 2,693,529,725 | IssuesEvent | 2015-04-01 15:12:52 | IQSS/dataverse | https://api.github.com/repos/IQSS/dataverse | opened | Listing Dataverses Widget is Displaying Unexpected Datasets and Dataverses | Component: Search/Browse Component: UX & Upgrade Priority: High Status: Dev | I am seeing unexpected dataverse and dataset results when I use a the Listing Dataverses Widget for a specific Dataverse:
**Example: Testing Widgets Dataverse**
For this Dataverse I should only see the following 2 datasets listed:

But for this script:
```
<script src="https://dvn-build.hmdc.harvard.edu/resources/js/widgets.js?alias=test-dv&dvUrl=https://dvn-build.hmdc.harvard.edu&widget=iframe&heightPx=500">
```
I am seeing other Dataverses and Datasets that dont belong with this Dataverse (even Drafts and unpublished datasets/dataverses, which I thought you would only see published things since its meant to be public)

| 1.0 | Listing Dataverses Widget is Displaying Unexpected Datasets and Dataverses - I am seeing unexpected dataverse and dataset results when I use a the Listing Dataverses Widget for a specific Dataverse:
**Example: Testing Widgets Dataverse**
For this Dataverse I should only see the following 2 datasets listed:

But for this script:
```
<script src="https://dvn-build.hmdc.harvard.edu/resources/js/widgets.js?alias=test-dv&dvUrl=https://dvn-build.hmdc.harvard.edu&widget=iframe&heightPx=500">
```
I am seeing other Dataverses and Datasets that dont belong with this Dataverse (even Drafts and unpublished datasets/dataverses, which I thought you would only see published things since its meant to be public)

| priority | listing dataverses widget is displaying unexpected datasets and dataverses i am seeing unexpected dataverse and dataset results when i use a the listing dataverses widget for a specific dataverse example testing widgets dataverse for this dataverse i should only see the following datasets listed but for this script script src i am seeing other dataverses and datasets that dont belong with this dataverse even drafts and unpublished datasets dataverses which i thought you would only see published things since its meant to be public | 1 |
388,728 | 11,491,646,646 | IssuesEvent | 2020-02-11 19:24:01 | ClinGen/clincoded | https://api.github.com/repos/ClinGen/clincoded | closed | Transfer of GDMs from UNC BCC to Hereditary Cancer GCEP | EP request GCI curation blocker priority: high | Hi @wrightmw we are in need of transferring the following records from the UNC Biocuration Core affiliation to the Hereditary Cancer GCEP.
This is a high priority need, as several of these GDMs are cardinal genes disease records that require publishing on the website. Also, there publishing would help to remove erroneous data previously published using the Gene Curation Express.
The URLS for the records are as follows:
1) MEN1: Multiple Endocrine Neoplasia: https://curation.clinicalgenome.org/curation-central/?gdm=f4aa4e77-0d27-44a5-a92a-1af9fc4c7068&pmid=9103196
2) RET: multiple endocrine neoplasia type 2A: https://curation.clinicalgenome.org/curation-central/?gdm=438681cb-7912-41b7-b3d7-571db9dfe1ce&pmid=8103403
3) RET: multiple endocrine neoplasia type 2B: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368
4) TP53: Li Fraumeni: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368
Please also check with @xiluo if these are correct as well. | 1.0 | Transfer of GDMs from UNC BCC to Hereditary Cancer GCEP - Hi @wrightmw we are in need of transferring the following records from the UNC Biocuration Core affiliation to the Hereditary Cancer GCEP.
This is a high priority need, as several of these GDMs are cardinal genes disease records that require publishing on the website. Also, there publishing would help to remove erroneous data previously published using the Gene Curation Express.
The URLS for the records are as follows:
1) MEN1: Multiple Endocrine Neoplasia: https://curation.clinicalgenome.org/curation-central/?gdm=f4aa4e77-0d27-44a5-a92a-1af9fc4c7068&pmid=9103196
2) RET: multiple endocrine neoplasia type 2A: https://curation.clinicalgenome.org/curation-central/?gdm=438681cb-7912-41b7-b3d7-571db9dfe1ce&pmid=8103403
3) RET: multiple endocrine neoplasia type 2B: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368
4) TP53: Li Fraumeni: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368
Please also check with @xiluo if these are correct as well. | priority | transfer of gdms from unc bcc to hereditary cancer gcep hi wrightmw we are in need of transferring the following records from the unc biocuration core affiliation to the hereditary cancer gcep this is a high priority need as several of these gdms are cardinal genes disease records that require publishing on the website also there publishing would help to remove erroneous data previously published using the gene curation express the urls for the records are as follows multiple endocrine neoplasia ret multiple endocrine neoplasia type ret multiple endocrine neoplasia type li fraumeni please also check with xiluo if these are correct as well | 1 |
29,180 | 2,714,007,142 | IssuesEvent | 2015-04-09 22:50:39 | nickpaventi/culligan-diy | https://api.github.com/repos/nickpaventi/culligan-diy | opened | Product Category [Mobile]: Feature #1 and 2 need style updates | High Priority | **Feature 1**
- [ ] Add consistent padding to left and right of feature content, currently touches left edge
- [ ] Add padding between product name and 'perfect solution...' title
**Feature 2**
- [ ] Image and content should flow like Feature 1
- [ ] Colors may need to change for tags, content and CTA to work in this new flow
Let's discuss real time if needed...

| 1.0 | Product Category [Mobile]: Feature #1 and 2 need style updates - **Feature 1**
- [ ] Add consistent padding to left and right of feature content, currently touches left edge
- [ ] Add padding between product name and 'perfect solution...' title
**Feature 2**
- [ ] Image and content should flow like Feature 1
- [ ] Colors may need to change for tags, content and CTA to work in this new flow
Let's discuss real time if needed...

| priority | product category feature and need style updates feature add consistent padding to left and right of feature content currently touches left edge add padding between product name and perfect solution title feature image and content should flow like feature colors may need to change for tags content and cta to work in this new flow let s discuss real time if needed | 1 |
598,267 | 18,241,681,087 | IssuesEvent | 2021-10-01 13:38:07 | g1moon/hello-waterski | https://api.github.com/repos/g1moon/hello-waterski | opened | 줄서기 등록 서비스 | priority high | - 줄서기 시스템에 대기를 등록하는 모달을 만든다.
- [ ] /line/:[spotId] 에 줄서기 버튼을 만들고, 클릭하면 등록하는 모달을 만든다.
- [ ] 모달에는 ridingType, boatType을 입력 받는다.
- [ ] 제출이 완료되면 data/line.json에 추가하고, allLineData에 새로운 정보를 추가한다.
- [ ] 새롭게 oneSpot 데이터도 갱신한다(다시 그릴지, 상태에 추가만 할지 고민).
| 1.0 | 줄서기 등록 서비스 - - 줄서기 시스템에 대기를 등록하는 모달을 만든다.
- [ ] /line/:[spotId] 에 줄서기 버튼을 만들고, 클릭하면 등록하는 모달을 만든다.
- [ ] 모달에는 ridingType, boatType을 입력 받는다.
- [ ] 제출이 완료되면 data/line.json에 추가하고, allLineData에 새로운 정보를 추가한다.
- [ ] 새롭게 oneSpot 데이터도 갱신한다(다시 그릴지, 상태에 추가만 할지 고민).
| priority | 줄서기 등록 서비스 줄서기 시스템에 대기를 등록하는 모달을 만든다 line 에 줄서기 버튼을 만들고 클릭하면 등록하는 모달을 만든다 모달에는 ridingtype boattype을 입력 받는다 제출이 완료되면 data line json에 추가하고 alllinedata에 새로운 정보를 추가한다 새롭게 onespot 데이터도 갱신한다 다시 그릴지 상태에 추가만 할지 고민 | 1 |
234,351 | 7,720,017,744 | IssuesEvent | 2018-05-23 21:22:21 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | Reset sequence broken on nRF5x chips | bug priority: high | We seem to go on an infite loop during the reset sequence with the introduciton of this commit:
72c7ded5612870c20a3efc3f36a9688ff85a4418
| 1.0 | Reset sequence broken on nRF5x chips - We seem to go on an infite loop during the reset sequence with the introduciton of this commit:
72c7ded5612870c20a3efc3f36a9688ff85a4418
| priority | reset sequence broken on chips we seem to go on an infite loop during the reset sequence with the introduciton of this commit | 1 |
384,181 | 11,384,842,486 | IssuesEvent | 2020-01-29 09:52:42 | unitystation/unitystation | https://api.github.com/repos/unitystation/unitystation | opened | Comms Consoles now print out empty Centcomm reports | Bug High Priority | Just experienced now on build 3989
The reports look like this when read:

| 1.0 | Comms Consoles now print out empty Centcomm reports - Just experienced now on build 3989
The reports look like this when read:

| priority | comms consoles now print out empty centcomm reports just experienced now on build the reports look like this when read | 1 |
718,440 | 24,717,273,830 | IssuesEvent | 2022-10-20 08:01:15 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | DHCP (IPv4) seemingly dies by trying to assign an IP of 0.0.0.0 | bug priority: high area: Networking | **Describe the bug**
This strange issue leads to an ethernet device on a network being completely dead until it is restarted. A device gets a fixed IP from a DHCP server for 20 seconds, the server is changed to give a different fixed IP and the zephyr board renews the IP. Seemingly at this time, the DHCP client seems to think it has been assigned the IP 0.0.0.0 for whatever reason, this IP clearly is not valid, and the DHCP client fails to remove/add it, it then stays in this state of constantly trying to add the invalid IP. Rebooting the module allows it to get a DHCP lease.
**To Reproduce**
Needs a DHCP server - I used dhcpd, with this configuration:
```
default-lease-time 20;
max-lease-time 20;
authoritative;
server-name name;
host zephyr {
hardware ethernet 68:27:19:CF:09:DD;
fixed-address 192.168.1.209;
}
subnet 192.168.1.0 netmask 255.255.255.0 {
range 192.168.1.202 192.168.1.208;
}
```
Then just run the server, load the zephyr DHCP sample on a board and let it get an IP. Once it gets an IP, change the fixed address of the zephyr device to something else, e.g. .210, restart the dhcpd server, then watch as the DHCP client falls over.
**Expected behavior**
DHCP to continue working
**Impact**
Showstopper, device is dead, complexity to denial of service a device is 0
**Logs and console output**
Log: https://pastebin.com/WeHvD1pN
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 0.15.0
- Commit SHA or Version used: 9018338863c0c5e7039d80df5f0d463075a0a31a | 1.0 | DHCP (IPv4) seemingly dies by trying to assign an IP of 0.0.0.0 - **Describe the bug**
This strange issue leads to an ethernet device on a network being completely dead until it is restarted. A device gets a fixed IP from a DHCP server for 20 seconds, the server is changed to give a different fixed IP and the zephyr board renews the IP. Seemingly at this time, the DHCP client seems to think it has been assigned the IP 0.0.0.0 for whatever reason, this IP clearly is not valid, and the DHCP client fails to remove/add it, it then stays in this state of constantly trying to add the invalid IP. Rebooting the module allows it to get a DHCP lease.
**To Reproduce**
Needs a DHCP server - I used dhcpd, with this configuration:
```
default-lease-time 20;
max-lease-time 20;
authoritative;
server-name name;
host zephyr {
hardware ethernet 68:27:19:CF:09:DD;
fixed-address 192.168.1.209;
}
subnet 192.168.1.0 netmask 255.255.255.0 {
range 192.168.1.202 192.168.1.208;
}
```
Then just run the server, load the zephyr DHCP sample on a board and let it get an IP. Once it gets an IP, change the fixed address of the zephyr device to something else, e.g. .210, restart the dhcpd server, then watch as the DHCP client falls over.
**Expected behavior**
DHCP to continue working
**Impact**
Showstopper, device is dead, complexity to denial of service a device is 0
**Logs and console output**
Log: https://pastebin.com/WeHvD1pN
**Environment (please complete the following information):**
- OS: Linux
- Toolchain: Zephyr SDK 0.15.0
- Commit SHA or Version used: 9018338863c0c5e7039d80df5f0d463075a0a31a | priority | dhcp seemingly dies by trying to assign an ip of describe the bug this strange issue leads to an ethernet device on a network being completely dead until it is restarted a device gets a fixed ip from a dhcp server for seconds the server is changed to give a different fixed ip and the zephyr board renews the ip seemingly at this time the dhcp client seems to think it has been assigned the ip for whatever reason this ip clearly is not valid and the dhcp client fails to remove add it it then stays in this state of constantly trying to add the invalid ip rebooting the module allows it to get a dhcp lease to reproduce needs a dhcp server i used dhcpd with this configuration default lease time max lease time authoritative server name name host zephyr hardware ethernet cf dd fixed address subnet netmask range then just run the server load the zephyr dhcp sample on a board and let it get an ip once it gets an ip change the fixed address of the zephyr device to something else e g restart the dhcpd server then watch as the dhcp client falls over expected behavior dhcp to continue working impact showstopper device is dead complexity to denial of service a device is logs and console output log environment please complete the following information os linux toolchain zephyr sdk commit sha or version used | 1 |
212,875 | 7,243,582,785 | IssuesEvent | 2018-02-14 12:17:34 | jrantamaki/supertimemachine | https://api.github.com/repos/jrantamaki/supertimemachine | closed | Bug: Calculation of elapsed time is wrong | bug frontend priority: high | Used Duration does not work properly when timestamps are for different dates. | 1.0 | Bug: Calculation of elapsed time is wrong - Used Duration does not work properly when timestamps are for different dates. | priority | bug calculation of elapsed time is wrong used duration does not work properly when timestamps are for different dates | 1 |
787,973 | 27,737,955,055 | IssuesEvent | 2023-03-15 12:33:47 | ITISFoundation/osparc-simcore | https://api.github.com/repos/ITISFoundation/osparc-simcore | opened | Garbage collector not working on aws-prod | bug High Priority | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Garbage collector for some days only shows error, there are no more "regular" logs and garbage collection seems to not happen. The errors are of this kind:
```
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886863c0>: '843fbe7b-2e50-56b3-9ad9-752de771bf21'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188e67ac0>: '41d7bcb2-af42-5104-b662-5c66e747bbf4'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57c40>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1885ce4c0>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '843fbe7b-2e50-56b3-9ad9-752de771bf21'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18879eb40>: '41d7bcb2-af42-5104-b662-5c66e747bbf4'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886073c0>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886372c0>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18872e940>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86'
```
The graylog queries that can be used to check if this happens are:
- `container_name:/.*collector.*/ AND NOT "Error in"` --> Display all non-error loglines, referring to "real" garbage collection
- `container_name:/.*collector.*/ AND "Error in"` --> Display the errors mentioned
Further evidence of garbage collection not working is that in prometheus one can see a s4-lite service running for many days, to observe this use the PromQL querry:
`container_memory_usage_bytes{image=~"^.*[.osparc.io].*/simcore/services/dynamic/s4l-core-lite.*$",name=~"dy-sidecar-b57f4e59-13d0-476d-9954-9855adf657b7.*"}`
Comparison with the redis keys, that correspond to open browser-tabs or sessions, show that there was no session key for the user that owns the project containing this s4l for some days, so the garbage collector should have kicked in:
`redis_key_value{key=~"^user_id=2:.*$"}`
### Expected Behavior
Garbage collection works
### Steps To Reproduce
The GC does not work on aws-prod
### Anything else?
This affects production and may cause it to not run smooth if services accumulate. From my feelings, I would put this on high urgency. | 1.0 | Garbage collector not working on aws-prod - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
Garbage collector for some days only shows error, there are no more "regular" logs and garbage collection seems to not happen. The errors are of this kind:
```
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886863c0>: '843fbe7b-2e50-56b3-9ad9-752de771bf21'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188e67ac0>: '41d7bcb2-af42-5104-b662-5c66e747bbf4'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57c40>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26'
WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1885ce4c0>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '843fbe7b-2e50-56b3-9ad9-752de771bf21'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18879eb40>: '41d7bcb2-af42-5104-b662-5c66e747bbf4'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886073c0>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886372c0>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26'
WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18872e940>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86'
```
The graylog queries that can be used to check if this happens are:
- `container_name:/.*collector.*/ AND NOT "Error in"` --> Display all non-error loglines, referring to "real" garbage collection
- `container_name:/.*collector.*/ AND "Error in"` --> Display the errors mentioned
Further evidence of garbage collection not working is that in prometheus one can see a s4-lite service running for many days, to observe this use the PromQL querry:
`container_memory_usage_bytes{image=~"^.*[.osparc.io].*/simcore/services/dynamic/s4l-core-lite.*$",name=~"dy-sidecar-b57f4e59-13d0-476d-9954-9855adf657b7.*"}`
Comparison with the redis keys, that correspond to open browser-tabs or sessions, show that there was no session key for the user that owns the project containing this s4l for some days, so the garbage collector should have kicked in:
`redis_key_value{key=~"^user_id=2:.*$"}`
### Expected Behavior
Garbage collection works
### Steps To Reproduce
The GC does not work on aws-prod
### Anything else?
This affects production and may cause it to not run smooth if services accumulate. From my feelings, I would put this on high urgency. | priority | garbage collector not working on aws prod is there an existing issue for this i have searched the existing issues current behavior garbage collector for some days only shows error there are no more regular logs and garbage collection seems to not happen the errors are of this kind warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task the graylog queries that can be used to check if this happens are container name collector and not error in display all non error loglines referring to real garbage collection container name collector and error in display the errors mentioned further evidence of garbage collection not working is that in prometheus one can see a lite service running for many days to observe this use the promql querry container memory usage bytes image simcore services dynamic core lite name dy sidecar comparison with the redis keys that correspond to open browser tabs or sessions show that there was no session key for the user that owns the project containing this for some days so the garbage collector should have kicked in redis key value key user id expected behavior garbage collection works steps to reproduce the gc does not work on aws prod anything else this affects production and may cause it to not run smooth if services accumulate from my feelings i would put this on high urgency | 1 |
699,236 | 24,009,797,449 | IssuesEvent | 2022-09-14 17:44:19 | opendatahub-io/odh-dashboard | https://api.github.com/repos/opendatahub-io/odh-dashboard | closed | [Bug]: Stability Fixes | kind/bug feature/notebook-controller priority/high | ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
There are a couple stability issues with the KFNBC right now.
1. If you do not have an existing PVC, you cannot spawn a notebook
2. If you are on a high loaded cluster, the endpoint call to create a Notebook can timeout with a 504 Gateway Timeout
### Expected Behavior
1. Fix the issue with the PVC, it needs to be created when we create a Notebook
2. Move the backend logic for route fetching out of the endpoint to create a Notebook -- this is the major slowdown, move it to the frontend and lazy fetch it to avoid slowness on cluster
### Steps To Reproduce
For both issues - Create a Notebook
1. Should make PVC calls in the network when you create
2. The endpoint for Notebook create should not annotate the route onto the Notebook
### Workaround (if any)
_No response_
### OpenShift Infrastructure Version
_No response_
### Openshift Version
_No response_
### What browsers are you seeing the problem on?
_No response_
### Open Data Hub Version
```yml
2.2
```
### Relevant log output
_No response_ | 1.0 | [Bug]: Stability Fixes - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
There are a couple stability issues with the KFNBC right now.
1. If you do not have an existing PVC, you cannot spawn a notebook
2. If you are on a high loaded cluster, the endpoint call to create a Notebook can timeout with a 504 Gateway Timeout
### Expected Behavior
1. Fix the issue with the PVC, it needs to be created when we create a Notebook
2. Move the backend logic for route fetching out of the endpoint to create a Notebook -- this is the major slowdown, move it to the frontend and lazy fetch it to avoid slowness on cluster
### Steps To Reproduce
For both issues - Create a Notebook
1. Should make PVC calls in the network when you create
2. The endpoint for Notebook create should not annotate the route onto the Notebook
### Workaround (if any)
_No response_
### OpenShift Infrastructure Version
_No response_
### Openshift Version
_No response_
### What browsers are you seeing the problem on?
_No response_
### Open Data Hub Version
```yml
2.2
```
### Relevant log output
_No response_ | priority | stability fixes is there an existing issue for this i have searched the existing issues current behavior there are a couple stability issues with the kfnbc right now if you do not have an existing pvc you cannot spawn a notebook if you are on a high loaded cluster the endpoint call to create a notebook can timeout with a gateway timeout expected behavior fix the issue with the pvc it needs to be created when we create a notebook move the backend logic for route fetching out of the endpoint to create a notebook this is the major slowdown move it to the frontend and lazy fetch it to avoid slowness on cluster steps to reproduce for both issues create a notebook should make pvc calls in the network when you create the endpoint for notebook create should not annotate the route onto the notebook workaround if any no response openshift infrastructure version no response openshift version no response what browsers are you seeing the problem on no response open data hub version yml relevant log output no response | 1 |
773,850 | 27,173,997,780 | IssuesEvent | 2023-02-17 22:31:01 | opendatahub-io/odh-dashboard | https://api.github.com/repos/opendatahub-io/odh-dashboard | opened | [A11y] Address A11y issues on Data Science Projects | kind/bug priority/high feature/dsg a11y | This ticket will look to address issues of an accessibility (a11y) nature in the Data Science Projects (DS Projects) section of the main nav.
> Get the [Chrome Dev Tools Axe Extension](https://chrome.google.com/webstore/detail/axe-devtools-web-accessib/lhdoppojpmngadmnindnejefpokejbdd)
> Running this extension on the page first loaded and as you use the page will help reveal what needs to be fixed
## Areas covered by this ticket
Main nav: Data Science Projects
> Note this covers all pages within' as well. See data setup for more information how to reveal the sub pages.
Key areas to make sure to test:
* Empty DS Projects page
* 1-n projects added
* Project details of one of the pages
* Empty all sections (aka, new project)
* Each section filled with data
* Workbenches, Cluster Storage, Data Connections, Model Serving
* Most sections have an expandable row for each row of the table -- more content shows up when you expand it
* Each row should have a kebab -- those typically have followup actions
* There are several modals; most are associated to the kebab actions or the secondary buttons along the page
* Stopping a Workbench has a modal -- once a Workbench is created, there is a toggle switch in the row -- click off
## Data setup for testing
* Creating a project is done from the main nav page -- you should have a button to create a project and it should turn the empty state into a table -- after 10 items, it becomes paginated
* Inside a project
* Creating a workbench => name can be anything, image selection is needed, typically the first one is fine, everything else should have a default
* There are expandable items and hidden items, click on the various links and select things in the dropdowns -- you'll likely need to run the axe test on each stage for coverage
* Create Cluster Storage => It's a modal, just need a name
* A Workbench can have multiple storages on different paths -- a storage cannot be associated with multiple workbenches
* A storage can be added to a workbench via the expanded workbench row -- it also requires the storage to be unassociated to any workbench (another modal)
* Create Data Connection => The data in here is not important, but the name, id, and key will need to be filled out before it can be submitted
* Data Connections can be connected to multiple workbenches, but workbenches cannot have more than one data connection
* Models and model servers
* Configuring a server is pretty straight forward, there are additional fields to expand using the checkboxes
* Once configured (only one can be at this time), you'll be able to deploy several models
* Models need a Data Connection & the details of this field is not important for axe testing | 1.0 | [A11y] Address A11y issues on Data Science Projects - This ticket will look to address issues of an accessibility (a11y) nature in the Data Science Projects (DS Projects) section of the main nav.
> Get the [Chrome Dev Tools Axe Extension](https://chrome.google.com/webstore/detail/axe-devtools-web-accessib/lhdoppojpmngadmnindnejefpokejbdd)
> Running this extension on the page first loaded and as you use the page will help reveal what needs to be fixed
## Areas covered by this ticket
Main nav: Data Science Projects
> Note this covers all pages within' as well. See data setup for more information how to reveal the sub pages.
Key areas to make sure to test:
* Empty DS Projects page
* 1-n projects added
* Project details of one of the pages
* Empty all sections (aka, new project)
* Each section filled with data
* Workbenches, Cluster Storage, Data Connections, Model Serving
* Most sections have an expandable row for each row of the table -- more content shows up when you expand it
* Each row should have a kebab -- those typically have followup actions
* There are several modals; most are associated to the kebab actions or the secondary buttons along the page
* Stopping a Workbench has a modal -- once a Workbench is created, there is a toggle switch in the row -- click off
## Data setup for testing
* Creating a project is done from the main nav page -- you should have a button to create a project and it should turn the empty state into a table -- after 10 items, it becomes paginated
* Inside a project
* Creating a workbench => name can be anything, image selection is needed, typically the first one is fine, everything else should have a default
* There are expandable items and hidden items, click on the various links and select things in the dropdowns -- you'll likely need to run the axe test on each stage for coverage
* Create Cluster Storage => It's a modal, just need a name
* A Workbench can have multiple storages on different paths -- a storage cannot be associated with multiple workbenches
* A storage can be added to a workbench via the expanded workbench row -- it also requires the storage to be unassociated to any workbench (another modal)
* Create Data Connection => The data in here is not important, but the name, id, and key will need to be filled out before it can be submitted
* Data Connections can be connected to multiple workbenches, but workbenches cannot have more than one data connection
* Models and model servers
* Configuring a server is pretty straight forward, there are additional fields to expand using the checkboxes
* Once configured (only one can be at this time), you'll be able to deploy several models
* Models need a Data Connection & the details of this field is not important for axe testing | priority | address issues on data science projects this ticket will look to address issues of an accessibility nature in the data science projects ds projects section of the main nav get the running this extension on the page first loaded and as you use the page will help reveal what needs to be fixed areas covered by this ticket main nav data science projects note this covers all pages within as well see data setup for more information how to reveal the sub pages key areas to make sure to test empty ds projects page n projects added project details of one of the pages empty all sections aka new project each section filled with data workbenches cluster storage data connections model serving most sections have an expandable row for each row of the table more content shows up when you expand it each row should have a kebab those typically have followup actions there are several modals most are associated to the kebab actions or the secondary buttons along the page stopping a workbench has a modal once a workbench is created there is a toggle switch in the row click off data setup for testing creating a project is done from the main nav page you should have a button to create a project and it should turn the empty state into a table after items it becomes paginated inside a project creating a workbench name can be anything image selection is needed typically the first one is fine everything else should have a default there are expandable items and hidden items click on the various links and select things in the dropdowns you ll likely need to run the axe test on each stage for coverage create cluster storage it s a modal just need a name a workbench can have multiple storages on different paths a storage cannot be associated with multiple workbenches a storage can be added to a workbench via the expanded workbench row it also requires the storage to be unassociated to any workbench another modal create data connection the data in here is not important but the name id and key will need to be filled out before it can be submitted data connections can be connected to multiple workbenches but workbenches cannot have more than one data connection models and model servers configuring a server is pretty straight forward there are additional fields to expand using the checkboxes once configured only one can be at this time you ll be able to deploy several models models need a data connection the details of this field is not important for axe testing | 1 |
2,273 | 2,524,996,072 | IssuesEvent | 2015-01-20 21:29:55 | graybeal/ont | https://api.github.com/repos/graybeal/ont | closed | feature to create ontology and use created ontology to replace existing version | 1 star enhancement imported Milestone-Beta1 Priority-High watchdog | _From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on November 09, 2009 14:51:48_
What capability do you want added or improved? Shouldn't we add some capability to avoid the issue that we had with CF--
where we had an XML file of a vocabulary (not yet an ontology) and we
needed, within the portal, to: a) develop a new ontology, and b) upload
that new ontology to replace an existing ontology. In other words, we were
trying to create ontology and then replace one with this new version. If
I'm not mistaken, portal users can only create an ontology and then
register it as a new ontology (not as an updated version) in the portal. Where do you want this capability to be accessible? from the voc2rdf component What sort of input/command mechanism do you want? button for registering as an updated version What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=217_ | 1.0 | feature to create ontology and use created ontology to replace existing version - _From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on November 09, 2009 14:51:48_
What capability do you want added or improved? Shouldn't we add some capability to avoid the issue that we had with CF--
where we had an XML file of a vocabulary (not yet an ontology) and we
needed, within the portal, to: a) develop a new ontology, and b) upload
that new ontology to replace an existing ontology. In other words, we were
trying to create ontology and then replace one with this new version. If
I'm not mistaken, portal users can only create an ontology and then
register it as a new ontology (not as an updated version) in the portal. Where do you want this capability to be accessible? from the voc2rdf component What sort of input/command mechanism do you want? button for registering as an updated version What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.)
_Original issue: http://code.google.com/p/mmisw/issues/detail?id=217_ | priority | feature to create ontology and use created ontology to replace existing version from on november what capability do you want added or improved shouldn t we add some capability to avoid the issue that we had with cf where we had an xml file of a vocabulary not yet an ontology and we needed within the portal to a develop a new ontology and b upload that new ontology to replace an existing ontology in other words we were trying to create ontology and then replace one with this new version if i m not mistaken portal users can only create an ontology and then register it as a new ontology not as an updated version in the portal where do you want this capability to be accessible from the component what sort of input command mechanism do you want button for registering as an updated version what is the desired output content format location other details of your desired capability what version of the product are you using please provide any additional information below particular ontology ies text contents of vocabulary operating system browser version firefox safari chrome ie etc screenshot etc original issue | 1 |
114,305 | 4,628,545,978 | IssuesEvent | 2016-09-28 05:14:32 | Aplokodika/MachineLearning | https://api.github.com/repos/Aplokodika/MachineLearning | closed | Need to add proper mechanisms for initializing the learning-rate and momentum values. | High priority task To Do | As of now, there isn't a proper mechanism for initializing the learning rate and the momentum. In this architecture, each neuron contains its own learning rate and momentum. | 1.0 | Need to add proper mechanisms for initializing the learning-rate and momentum values. - As of now, there isn't a proper mechanism for initializing the learning rate and the momentum. In this architecture, each neuron contains its own learning rate and momentum. | priority | need to add proper mechanisms for initializing the learning rate and momentum values as of now there isn t a proper mechanism for initializing the learning rate and the momentum in this architecture each neuron contains its own learning rate and momentum | 1 |
673,377 | 22,960,184,719 | IssuesEvent | 2022-07-19 14:47:16 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Cannot start the server with installServer.bat in windows | Priority/Highest Severity/Major bug Affected-6.0.0 QA-Reported JDK-17 | **How to reproduce:**
1. Get pack from https://wso2.org/jenkins/job/products/job/product-is/4152/
2. Follow running the server windows as a service https://is.docs.wso2.com/en/latest/setup/installing-as-a-windows-service/
3. Set carbon_home and JAVA_HOME as env variables
4. Copy the IS-pack wrapper.conf to YAJSW13.03/conf. wrapper.conf file was edited as per the comment in https://github.com/wso2/product-is/issues/14148#issuecomment-1178955108
5. Navigate to YAJSW13.03-home/bat
6. Run
`runConsole.bat
`
9. Server will start successfully
Was able to get the server started successfully with the above command. When try to start the server with below commands was unable to get it started successfully. Reporting with a less severity as server can be started with step 6 command.
10. Then try the other option of running windows as a service with below commands
`installService.bat
`
`startService.bat
`
`stopService.bat
Server is not starting succesfully. Did we try these commands on windows 10 after we did the fixes for earlier fixed issues on windows. @Thumimku @ThumulaPerera
Attaching used wrapper.conf
[wrapper.zip](https://github.com/wso2/product-is/files/9120483/wrapper.zip)
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
H2 default
windows 11
YAJSW 13.03 | 1.0 | Cannot start the server with installServer.bat in windows - **How to reproduce:**
1. Get pack from https://wso2.org/jenkins/job/products/job/product-is/4152/
2. Follow running the server windows as a service https://is.docs.wso2.com/en/latest/setup/installing-as-a-windows-service/
3. Set carbon_home and JAVA_HOME as env variables
4. Copy the IS-pack wrapper.conf to YAJSW13.03/conf. wrapper.conf file was edited as per the comment in https://github.com/wso2/product-is/issues/14148#issuecomment-1178955108
5. Navigate to YAJSW13.03-home/bat
6. Run
`runConsole.bat
`
9. Server will start successfully
Was able to get the server started successfully with the above command. When try to start the server with below commands was unable to get it started successfully. Reporting with a less severity as server can be started with step 6 command.
10. Then try the other option of running windows as a service with below commands
`installService.bat
`
`startService.bat
`
`stopService.bat
Server is not starting succesfully. Did we try these commands on windows 10 after we did the fixes for earlier fixed issues on windows. @Thumimku @ThumulaPerera
Attaching used wrapper.conf
[wrapper.zip](https://github.com/wso2/product-is/files/9120483/wrapper.zip)
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
H2 default
windows 11
YAJSW 13.03 | priority | cannot start the server with installserver bat in windows how to reproduce get pack from follow running the server windows as a service set carbon home and java home as env variables copy the is pack wrapper conf to conf wrapper conf file was edited as per the comment in navigate to home bat run runconsole bat server will start successfully was able to get the server started successfully with the above command when try to start the server with below commands was unable to get it started successfully reporting with a less severity as server can be started with step command then try the other option of running windows as a service with below commands installservice bat startservice bat stopservice bat server is not starting succesfully did we try these commands on windows after we did the fixes for earlier fixed issues on windows thumimku thumulaperera attaching used wrapper conf environment information please complete the following information remove any unnecessary fields default windows yajsw | 1 |
219,382 | 7,341,693,376 | IssuesEvent | 2018-03-07 03:30:34 | hackoregon/civic-devops | https://api.github.com/repos/hackoregon/civic-devops | closed | Get AWS accounts for Tony and Khuong | Priority: high | Tony and Khuong require accounts to be created in the AWS environment. | 1.0 | Get AWS accounts for Tony and Khuong - Tony and Khuong require accounts to be created in the AWS environment. | priority | get aws accounts for tony and khuong tony and khuong require accounts to be created in the aws environment | 1 |
179,527 | 6,625,938,498 | IssuesEvent | 2017-09-22 17:25:49 | RepreZen/KaiZen-OpenAPI-Editor | https://api.github.com/repos/RepreZen/KaiZen-OpenAPI-Editor | closed | Code assist should support typeahead with dynamic filtering | High Priority Pending Review | The code assist feature in other Eclipse editors supports two useful behaviors that we should add to SwagEdit:
1. After invoking code assist, you can start typing, and the list of suggestions will remain visible, filtered to show the suggestions that start with the substring you entered.
2. If you invoke code assist with one or more characters already typed, immediately to the left of the insertion point, and then invoke code assist, the suggestion list will be pre-filtered to show matching suggestions.
| 1.0 | Code assist should support typeahead with dynamic filtering - The code assist feature in other Eclipse editors supports two useful behaviors that we should add to SwagEdit:
1. After invoking code assist, you can start typing, and the list of suggestions will remain visible, filtered to show the suggestions that start with the substring you entered.
2. If you invoke code assist with one or more characters already typed, immediately to the left of the insertion point, and then invoke code assist, the suggestion list will be pre-filtered to show matching suggestions.
| priority | code assist should support typeahead with dynamic filtering the code assist feature in other eclipse editors supports two useful behaviors that we should add to swagedit after invoking code assist you can start typing and the list of suggestions will remain visible filtered to show the suggestions that start with the substring you entered if you invoke code assist with one or more characters already typed immediately to the left of the insertion point and then invoke code assist the suggestion list will be pre filtered to show matching suggestions | 1 |
131,139 | 5,143,788,456 | IssuesEvent | 2017-01-12 16:51:02 | Esri/visibility-addin-dotnet | https://api.github.com/repos/Esri/visibility-addin-dotnet | closed | Arcmap and Pro LLOS, and RLOS does not detect point outside of an elevation Mosaic dataset extent that is not rectangular. | 2 - In Progress B - Bug priority - high | @jmccausland @dhagens @topowright @mfunk
In Arcmap and Pro 1.3, RLOS does not detect user entered point outside of an elevation Mosaic Data Set extent that is not rectangular in an area that would be in the minumum bounding rectangle.

| 1.0 | Arcmap and Pro LLOS, and RLOS does not detect point outside of an elevation Mosaic dataset extent that is not rectangular. - @jmccausland @dhagens @topowright @mfunk
In Arcmap and Pro 1.3, RLOS does not detect user entered point outside of an elevation Mosaic Data Set extent that is not rectangular in an area that would be in the minumum bounding rectangle.

| priority | arcmap and pro llos and rlos does not detect point outside of an elevation mosaic dataset extent that is not rectangular jmccausland dhagens topowright mfunk in arcmap and pro rlos does not detect user entered point outside of an elevation mosaic data set extent that is not rectangular in an area that would be in the minumum bounding rectangle | 1 |
367,117 | 10,840,704,210 | IssuesEvent | 2019-11-12 08:56:49 | ProtonMail/proton-contacts | https://api.github.com/repos/ProtonMail/proton-contacts | closed | V4 Beta feedback - something happened to the name field on create/edit contact | Priority: high bug | The text input is misaligned with the label, misaligned with the other text inputs and smaller size than the other text inputs.
<img width="703" alt="Capture d’écran 2019-11-11 à 15 12 13" src="https://user-images.githubusercontent.com/40168570/68594557-190c5480-0498-11ea-90c5-96a61ef46f65.png">
| 1.0 | V4 Beta feedback - something happened to the name field on create/edit contact - The text input is misaligned with the label, misaligned with the other text inputs and smaller size than the other text inputs.
<img width="703" alt="Capture d’écran 2019-11-11 à 15 12 13" src="https://user-images.githubusercontent.com/40168570/68594557-190c5480-0498-11ea-90c5-96a61ef46f65.png">
| priority | beta feedback something happened to the name field on create edit contact the text input is misaligned with the label misaligned with the other text inputs and smaller size than the other text inputs img width alt capture d’écran à src | 1 |
618,737 | 19,486,122,001 | IssuesEvent | 2021-12-26 12:08:57 | covidpass-org/covidpass | https://api.github.com/repos/covidpass-org/covidpass | closed | Not working with third dose health pass | bug priority:high | Hello,
It seems that it does not work with a 3rd dose health pass. I have the message : Invalid QR code (for French health pass, Safari on iOS 15).
Thanks. | 1.0 | Not working with third dose health pass - Hello,
It seems that it does not work with a 3rd dose health pass. I have the message : Invalid QR code (for French health pass, Safari on iOS 15).
Thanks. | priority | not working with third dose health pass hello it seems that it does not work with a dose health pass i have the message invalid qr code for french health pass safari on ios thanks | 1 |
473,914 | 13,649,309,760 | IssuesEvent | 2020-09-26 13:51:09 | ahmedkaludi/accelerated-mobile-pages | https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages | closed | Internal feedback Improvement for Longtime Users | NEXT UPDATE [Priority: HIGH] enhancement | Internal feedback Improvement for Longtime Users | 1.0 | Internal feedback Improvement for Longtime Users - Internal feedback Improvement for Longtime Users | priority | internal feedback improvement for longtime users internal feedback improvement for longtime users | 1 |
684,428 | 23,417,823,865 | IssuesEvent | 2022-08-13 08:01:12 | IntelCompH2020/topicmodeler | https://api.github.com/repos/IntelCompH2020/topicmodeler | opened | Study stability / coherence of topics versus training set size | High Priority Experimental Work / Validation | We have a few huge corpus, in the order of tens or millions of documents. Training is costly. The question here is:
Do we really need to train with the whole corpus ? Are topics much better than if training with say a maximum of 2 M documents ??? This should be studied because if no improvement is gained when training with very large corpora we could sample the training set, and then carry out inference on the whole set when calculating the indicators.
| 1.0 | Study stability / coherence of topics versus training set size - We have a few huge corpus, in the order of tens or millions of documents. Training is costly. The question here is:
Do we really need to train with the whole corpus ? Are topics much better than if training with say a maximum of 2 M documents ??? This should be studied because if no improvement is gained when training with very large corpora we could sample the training set, and then carry out inference on the whole set when calculating the indicators.
| priority | study stability coherence of topics versus training set size we have a few huge corpus in the order of tens or millions of documents training is costly the question here is do we really need to train with the whole corpus are topics much better than if training with say a maximum of m documents this should be studied because if no improvement is gained when training with very large corpora we could sample the training set and then carry out inference on the whole set when calculating the indicators | 1 |
154,253 | 5,916,657,908 | IssuesEvent | 2017-05-22 11:07:45 | DOAJ/doaj | https://api.github.com/repos/DOAJ/doaj | closed | Libraries IP information | feedback high priority | Dear CL
Do we have this info? where can I get it?
Hello. I am the Director at the Tom Green County Library and we have had a change in IP address. Please provide the information you currently have on file for our IPs. Thank you. | 1.0 | Libraries IP information - Dear CL
Do we have this info? where can I get it?
Hello. I am the Director at the Tom Green County Library and we have had a change in IP address. Please provide the information you currently have on file for our IPs. Thank you. | priority | libraries ip information dear cl do we have this info where can i get it hello i am the director at the tom green county library and we have had a change in ip address please provide the information you currently have on file for our ips thank you | 1 |
464,576 | 13,327,231,136 | IssuesEvent | 2020-08-27 12:53:29 | zeebe-io/zeebe | https://api.github.com/repos/zeebe-io/zeebe | closed | Do not log RESOURCE_EXHAUSTED exceptions in job worker (GO) | Priority: High Scope: clients/go Status: Needs Review Type: Maintenance | **Description**
Related to https://github.com/zeebe-io/zeebe/issues/5187 we have seen recently an huge increase of log statements, which are related to `RESOURCE_EXHAUSTED` errors. The job worker in the java or go client, log errors when they happen during activating. These kind of "errors" are not really errors and are handled gracefully internally by the job worker, via retry so it is not necessary to log them. The user can't do much in this case. The log statements spread more confusion and seem to be not really helpful.
| 1.0 | Do not log RESOURCE_EXHAUSTED exceptions in job worker (GO) - **Description**
Related to https://github.com/zeebe-io/zeebe/issues/5187 we have seen recently an huge increase of log statements, which are related to `RESOURCE_EXHAUSTED` errors. The job worker in the java or go client, log errors when they happen during activating. These kind of "errors" are not really errors and are handled gracefully internally by the job worker, via retry so it is not necessary to log them. The user can't do much in this case. The log statements spread more confusion and seem to be not really helpful.
| priority | do not log resource exhausted exceptions in job worker go description related to we have seen recently an huge increase of log statements which are related to resource exhausted errors the job worker in the java or go client log errors when they happen during activating these kind of errors are not really errors and are handled gracefully internally by the job worker via retry so it is not necessary to log them the user can t do much in this case the log statements spread more confusion and seem to be not really helpful | 1 |
526,950 | 15,305,412,669 | IssuesEvent | 2021-02-24 18:05:16 | ansible/awx | https://api.github.com/repos/ansible/awx | opened | Prevent users from selecting job templates that are missing a project or inventory when creating workflow node | component:ui priority:high state:in_progress type:feature | ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
In the old UI we would disable rows that were missing a project/inv and didn't prompt for inventory:
<img width="1673" alt="Screen Shot 2021-02-24 at 1 00 46 PM" src="https://user-images.githubusercontent.com/9889020/109044677-ad57f600-76a0-11eb-90cc-b89735c86d58.png">
We should do something similar in the new UI since selecting a JT like this will result in the API returning an error when we attempt to save the workflow.
It would be nice if we could also add a tooltip or something like that to the row explaining why it can't be selected.
For existing nodes, the row will be selected on edit but it will still be disabled:
<img width="1674" alt="Screen Shot 2021-02-24 at 1 05 00 PM" src="https://user-images.githubusercontent.com/9889020/109044872-ebedb080-76a0-11eb-959e-b8d9c037e91f.png">
| 1.0 | Prevent users from selecting job templates that are missing a project or inventory when creating workflow node - ##### ISSUE TYPE
- Feature Idea
##### SUMMARY
In the old UI we would disable rows that were missing a project/inv and didn't prompt for inventory:
<img width="1673" alt="Screen Shot 2021-02-24 at 1 00 46 PM" src="https://user-images.githubusercontent.com/9889020/109044677-ad57f600-76a0-11eb-90cc-b89735c86d58.png">
We should do something similar in the new UI since selecting a JT like this will result in the API returning an error when we attempt to save the workflow.
It would be nice if we could also add a tooltip or something like that to the row explaining why it can't be selected.
For existing nodes, the row will be selected on edit but it will still be disabled:
<img width="1674" alt="Screen Shot 2021-02-24 at 1 05 00 PM" src="https://user-images.githubusercontent.com/9889020/109044872-ebedb080-76a0-11eb-959e-b8d9c037e91f.png">
| priority | prevent users from selecting job templates that are missing a project or inventory when creating workflow node issue type feature idea summary in the old ui we would disable rows that were missing a project inv and didn t prompt for inventory img width alt screen shot at pm src we should do something similar in the new ui since selecting a jt like this will result in the api returning an error when we attempt to save the workflow it would be nice if we could also add a tooltip or something like that to the row explaining why it can t be selected for existing nodes the row will be selected on edit but it will still be disabled img width alt screen shot at pm src | 1 |
712,357 | 24,492,468,892 | IssuesEvent | 2022-10-10 04:35:04 | IAmTamal/Milan | https://api.github.com/repos/IAmTamal/Milan | closed | Join us part still shown after logging in. | 🟧 priority: high 🕹 aspect: interface 🛠 goal: fix 🛠 status : under development hacktoberfest | ### Description
The join us part with login and register option is still shown to user which is already logged in. This should not happen,
### Screenshots

### Additional information
_No response_
### 🥦 Browser
Mozilla Firefox
### 👀 Have you checked if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Contributing Guidelines?
- [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md)
### Are you willing to work on this issue ?
Yes I am willing to submit a PR! | 1.0 | Join us part still shown after logging in. - ### Description
The join us part with login and register option is still shown to user which is already logged in. This should not happen,
### Screenshots

### Additional information
_No response_
### 🥦 Browser
Mozilla Firefox
### 👀 Have you checked if this issue has been raised before?
- [X] I checked and didn't find similar issue
### 🏢 Have you read the Contributing Guidelines?
- [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md)
### Are you willing to work on this issue ?
Yes I am willing to submit a PR! | priority | join us part still shown after logging in description the join us part with login and register option is still shown to user which is already logged in this should not happen screenshots additional information no response 🥦 browser mozilla firefox 👀 have you checked if this issue has been raised before i checked and didn t find similar issue 🏢 have you read the contributing guidelines i have read the are you willing to work on this issue yes i am willing to submit a pr | 1 |
664,477 | 22,272,264,418 | IssuesEvent | 2022-06-10 13:28:27 | opencrvs/opencrvs-core | https://api.github.com/repos/opencrvs/opencrvs-core | closed | Certificate Config: Add data tag to images | 👹Bug Priority: high | **Describe the bug**
- difficult to work out correct order for images so that the signature properly displays
- suggest to use a data tag for images. So that the signature is appropriately tagged and doesn't matter where it appears in the code.
| 1.0 | Certificate Config: Add data tag to images - **Describe the bug**
- difficult to work out correct order for images so that the signature properly displays
- suggest to use a data tag for images. So that the signature is appropriately tagged and doesn't matter where it appears in the code.
| priority | certificate config add data tag to images describe the bug difficult to work out correct order for images so that the signature properly displays suggest to use a data tag for images so that the signature is appropriately tagged and doesn t matter where it appears in the code | 1 |
133,817 | 5,215,316,115 | IssuesEvent | 2017-01-26 04:11:53 | imrogues/angularjs | https://api.github.com/repos/imrogues/angularjs | opened | Watching Collections | [priority] high [status] accepted [type] feature | ### Description
Add the third and final dirty–checking mechanism to our implementation of `Scope`: shallow collection–watching.
The `$watchCollection` function is not simple, but that’s mostly because it provides an important, non-trivial facility: We can watch for changes in large arrays and objects much more efficiently than we could with just deep–watching.
---
### Issue Checklist
- [ ] How `$watchCollection` can be used with arrays, objects, and other values.
- [ ] What `$watchCollection` does with arrays.
- [ ] What `$watchCollection` does with objects.
- [ ] Array–like objects and their role in `$watchCollection`.
All issues in milestone: [2 Scopes](https://github.com/imrogues/angularjs/milestone/2)
---
### Assignees
- [ ] Final assign @imrogues | 1.0 | Watching Collections - ### Description
Add the third and final dirty–checking mechanism to our implementation of `Scope`: shallow collection–watching.
The `$watchCollection` function is not simple, but that’s mostly because it provides an important, non-trivial facility: We can watch for changes in large arrays and objects much more efficiently than we could with just deep–watching.
---
### Issue Checklist
- [ ] How `$watchCollection` can be used with arrays, objects, and other values.
- [ ] What `$watchCollection` does with arrays.
- [ ] What `$watchCollection` does with objects.
- [ ] Array–like objects and their role in `$watchCollection`.
All issues in milestone: [2 Scopes](https://github.com/imrogues/angularjs/milestone/2)
---
### Assignees
- [ ] Final assign @imrogues | priority | watching collections description add the third and final dirty–checking mechanism to our implementation of scope shallow collection–watching the watchcollection function is not simple but that’s mostly because it provides an important non trivial facility we can watch for changes in large arrays and objects much more efficiently than we could with just deep–watching issue checklist how watchcollection can be used with arrays objects and other values what watchcollection does with arrays what watchcollection does with objects array–like objects and their role in watchcollection all issues in milestone assignees final assign imrogues | 1 |
523,593 | 15,185,961,645 | IssuesEvent | 2021-02-15 11:40:33 | percipioglobal/craft | https://api.github.com/repos/percipioglobal/craft | closed | Too much "searchable" fields. | performance 🏃♀️ priority: high | In the current field setup, too many fields are still "searchable"
eg. the image optimise fields | 1.0 | Too much "searchable" fields. - In the current field setup, too many fields are still "searchable"
eg. the image optimise fields | priority | too much searchable fields in the current field setup too many fields are still searchable eg the image optimise fields | 1 |
695,166 | 23,847,344,636 | IssuesEvent | 2022-09-06 14:56:45 | WordPress/openverse-frontend | https://api.github.com/repos/WordPress/openverse-frontend | closed | The link to report at the source is incorrectly linking to the direct URL | good first issue help wanted 🟧 priority: high 🛠 goal: fix 🕹 aspect: interface | ## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
After submitting a report for an item on Openverse, the user is encourage to submit the same report at the source of the media item. The link should be to the `foreign_landing_url` of the image which has the "Report it" button, not the direct URL (that opens the original image).
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
<img width="436" alt="Screen Shot 2022-09-05 at 4 22 24 PM" src="https://user-images.githubusercontent.com/15233243/188460269-d9471bab-bb02-403d-b111-9396ce3fa7d6.png">
## Additional context
<!-- Add any other context about the problem here; or delete the section entirely. -->
This line should have `media.foreign_landing_url` instead of `media.url`:
https://github.com/WordPress/openverse-frontend/blob/bf0c72e5ea660354e20ffa33a4caf0abd65a5db9/src/components/VContentReport/VContentReportForm.vue#L13
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in resolving this bug.
| 1.0 | The link to report at the source is incorrectly linking to the direct URL - ## Description
<!-- Concisely describe the bug. Compare your experience with what you expected to happen. -->
<!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." -->
After submitting a report for an item on Openverse, the user is encourage to submit the same report at the source of the media item. The link should be to the `foreign_landing_url` of the image which has the "Report it" button, not the direct URL (that opens the original image).
## Screenshots
<!-- Add screenshots to show the problem; or delete the section entirely. -->
<img width="436" alt="Screen Shot 2022-09-05 at 4 22 24 PM" src="https://user-images.githubusercontent.com/15233243/188460269-d9471bab-bb02-403d-b111-9396ce3fa7d6.png">
## Additional context
<!-- Add any other context about the problem here; or delete the section entirely. -->
This line should have `media.foreign_landing_url` instead of `media.url`:
https://github.com/WordPress/openverse-frontend/blob/bf0c72e5ea660354e20ffa33a4caf0abd65a5db9/src/components/VContentReport/VContentReportForm.vue#L13
## Resolution
<!-- Replace the [ ] with [x] to check the box. -->
- [ ] 🙋 I would be interested in resolving this bug.
| priority | the link to report at the source is incorrectly linking to the direct url description after submitting a report for an item on openverse the user is encourage to submit the same report at the source of the media item the link should be to the foreign landing url of the image which has the report it button not the direct url that opens the original image screenshots img width alt screen shot at pm src additional context this line should have media foreign landing url instead of media url resolution 🙋 i would be interested in resolving this bug | 1 |
677,841 | 23,177,467,416 | IssuesEvent | 2022-07-31 16:29:18 | projectdiscovery/httpx | https://api.github.com/repos/projectdiscovery/httpx | closed | Ports Flag is not working with scheme | Priority: High Status: Completed Type: Bug | <!--
1. Please search to see if an issue already exists for the bug you encountered.
2. For support requests, FAQs or "How to" questions, please use the GitHub Discussions section instead - https://github.com/projectdiscovery/httpx/discussions or
3. Join our discord server at https://discord.gg/projectdiscovery and post the question on the #httpx channel.
-->
<!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. -->
### httpx version:
1.2.3
<!-- You can find current version of httpx with "httpx -version" -->
<!-- We only accept issues that are reproducible on the latest version of httpx. -->
<!-- You can find the latest version of project at https://github.com/projectdiscovery/httpx/releases/ -->
### Current Behavior:
When the ports flag is used within scheme (-p https:443) it is breaking.
<!-- A concise description of what you're experiencing. -->
### Expected Behavior:
It should probe https on port 443.
<!-- A concise description of what you expected to happen. -->
### Steps To Reproduce:
<!--
Example: steps to reproduce the behavior:
1. Run 'echo example.com | httpx -ports https:443'
-->
1. Run 'echo example.com | httpx -ports https:443'
### Anything else:
Response:
'[FTL] Could not cast port to integer from your value: https:443. Resulting error: strconv.Atoi: parsing "https:443": invalid syntax.'
<!-- Links? References? Screnshots? Anything that will give us more context about the issue that you are encountering! --> | 1.0 | Ports Flag is not working with scheme - <!--
1. Please search to see if an issue already exists for the bug you encountered.
2. For support requests, FAQs or "How to" questions, please use the GitHub Discussions section instead - https://github.com/projectdiscovery/httpx/discussions or
3. Join our discord server at https://discord.gg/projectdiscovery and post the question on the #httpx channel.
-->
<!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. -->
### httpx version:
1.2.3
<!-- You can find current version of httpx with "httpx -version" -->
<!-- We only accept issues that are reproducible on the latest version of httpx. -->
<!-- You can find the latest version of project at https://github.com/projectdiscovery/httpx/releases/ -->
### Current Behavior:
When the ports flag is used within scheme (-p https:443) it is breaking.
<!-- A concise description of what you're experiencing. -->
### Expected Behavior:
It should probe https on port 443.
<!-- A concise description of what you expected to happen. -->
### Steps To Reproduce:
<!--
Example: steps to reproduce the behavior:
1. Run 'echo example.com | httpx -ports https:443'
-->
1. Run 'echo example.com | httpx -ports https:443'
### Anything else:
Response:
'[FTL] Could not cast port to integer from your value: https:443. Resulting error: strconv.Atoi: parsing "https:443": invalid syntax.'
<!-- Links? References? Screnshots? Anything that will give us more context about the issue that you are encountering! --> | priority | ports flag is not working with scheme please search to see if an issue already exists for the bug you encountered for support requests faqs or how to questions please use the github discussions section instead or join our discord server at and post the question on the httpx channel httpx version current behavior when the ports flag is used within scheme p https it is breaking expected behavior it should probe https on port steps to reproduce example steps to reproduce the behavior run echo example com httpx ports https run echo example com httpx ports https anything else response could not cast port to integer from your value https resulting error strconv atoi parsing https invalid syntax | 1 |
529,036 | 15,379,318,099 | IssuesEvent | 2021-03-02 19:30:06 | cch5ng/job_tracker | https://api.github.com/repos/cch5ng/job_tracker | closed | form validation | enhancement high priority | definitely need FE form validation
should have min required fields (Job, Event)
* this should affect content being saved for Alerts and also the form submit click handler
then should sync up with BE form validation | 1.0 | form validation - definitely need FE form validation
should have min required fields (Job, Event)
* this should affect content being saved for Alerts and also the form submit click handler
then should sync up with BE form validation | priority | form validation definitely need fe form validation should have min required fields job event this should affect content being saved for alerts and also the form submit click handler then should sync up with be form validation | 1 |
468,844 | 13,491,652,572 | IssuesEvent | 2020-09-11 16:50:10 | bloom-housing/bloom | https://api.github.com/repos/bloom-housing/bloom | opened | Jones: Remove Preference Section from Eligibility if no Preferences | High Priority | Since there are no preferences this section of eligibility shouldn't display.

| 1.0 | Jones: Remove Preference Section from Eligibility if no Preferences - Since there are no preferences this section of eligibility shouldn't display.

| priority | jones remove preference section from eligibility if no preferences since there are no preferences this section of eligibility shouldn t display | 1 |
761,515 | 26,684,365,984 | IssuesEvent | 2023-01-26 20:30:20 | gamefreedomgit/Maelstrom | https://api.github.com/repos/gamefreedomgit/Maelstrom | closed | [Paladin][Ret] Guardian Of Ancient Kings / Judgements of the Bold / Long Arm of the Law issue | Class: Paladin Pet Spell Priority: High Status: Confirmed | Two more bugs with Ret Paladin:
- Guardian of Ancient Kings has an assist function, but not an attack function - essentially, he just stands around looking pretty.
- Judgements of the Bold and Long Arm of the Law do not proc on Absorb effects. | 1.0 | [Paladin][Ret] Guardian Of Ancient Kings / Judgements of the Bold / Long Arm of the Law issue - Two more bugs with Ret Paladin:
- Guardian of Ancient Kings has an assist function, but not an attack function - essentially, he just stands around looking pretty.
- Judgements of the Bold and Long Arm of the Law do not proc on Absorb effects. | priority | guardian of ancient kings judgements of the bold long arm of the law issue two more bugs with ret paladin guardian of ancient kings has an assist function but not an attack function essentially he just stands around looking pretty judgements of the bold and long arm of the law do not proc on absorb effects | 1 |
180,174 | 6,647,048,086 | IssuesEvent | 2017-09-28 01:23:15 | CrowdRescueHQ/CrowdRescue | https://api.github.com/repos/CrowdRescueHQ/CrowdRescue | closed | Incident Updates/Suggested Edits | discussion priority-high | We need to fill the need of having a "suggested edits" queue, where one group of volunteers can fill out an update form and other volunteers can process it. This could work like suggested edits or it could be a form which feeds a queue. Both would be similar, but the latter would require a new table. We need to get more feedback about this.
Each edit should get logged and possible have an attached note or reason. (Could be built through case notes, could be separate). | 1.0 | Incident Updates/Suggested Edits - We need to fill the need of having a "suggested edits" queue, where one group of volunteers can fill out an update form and other volunteers can process it. This could work like suggested edits or it could be a form which feeds a queue. Both would be similar, but the latter would require a new table. We need to get more feedback about this.
Each edit should get logged and possible have an attached note or reason. (Could be built through case notes, could be separate). | priority | incident updates suggested edits we need to fill the need of having a suggested edits queue where one group of volunteers can fill out an update form and other volunteers can process it this could work like suggested edits or it could be a form which feeds a queue both would be similar but the latter would require a new table we need to get more feedback about this each edit should get logged and possible have an attached note or reason could be built through case notes could be separate | 1 |
818,287 | 30,681,878,093 | IssuesEvent | 2023-07-26 09:41:22 | bryntum/support | https://api.github.com/repos/bryntum/support | closed | [VUE] An exception when use workingTime config in calendar timeline view | bug resolved high-priority premium forum large-account OEM | [Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=25269&p=125948#p125948)
Not able to repro this in vanilla example. Please read forum topic, `workingTime` that throws an error depends on current time.
See app attached below
https://github.com/bryntum/support/assets/7203098/aff90109-b9fe-4033-b7f7-f5138c104b64
[bryntum-calendar-vite.zip](https://github.com/bryntum/support/files/11738789/bryntum-calendar-vite.zip)
| 1.0 | [VUE] An exception when use workingTime config in calendar timeline view - [Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=25269&p=125948#p125948)
Not able to repro this in vanilla example. Please read forum topic, `workingTime` that throws an error depends on current time.
See app attached below
https://github.com/bryntum/support/assets/7203098/aff90109-b9fe-4033-b7f7-f5138c104b64
[bryntum-calendar-vite.zip](https://github.com/bryntum/support/files/11738789/bryntum-calendar-vite.zip)
| priority | an exception when use workingtime config in calendar timeline view not able to repro this in vanilla example please read forum topic workingtime that throws an error depends on current time see app attached below | 1 |
269,970 | 8,444,984,599 | IssuesEvent | 2018-10-18 20:05:17 | semperfiwebdesign/all-in-one-seo-pack | https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack | closed | noindex/nofollow for no SEO on post type only works together | Bug Priority | High | For #1797 and #1943 we made it so noindex/nofollow works when SEO for that post type is disabled, however this only works if you're using both noindex and nofollow. If you're trying to use just one it doesn't work.
To make the logic easier, it may be best to pull this out into its own function(s). Eventually it should be in its own file, rather than junking up that big class file. | 1.0 | noindex/nofollow for no SEO on post type only works together - For #1797 and #1943 we made it so noindex/nofollow works when SEO for that post type is disabled, however this only works if you're using both noindex and nofollow. If you're trying to use just one it doesn't work.
To make the logic easier, it may be best to pull this out into its own function(s). Eventually it should be in its own file, rather than junking up that big class file. | priority | noindex nofollow for no seo on post type only works together for and we made it so noindex nofollow works when seo for that post type is disabled however this only works if you re using both noindex and nofollow if you re trying to use just one it doesn t work to make the logic easier it may be best to pull this out into its own function s eventually it should be in its own file rather than junking up that big class file | 1 |
688,433 | 23,581,834,381 | IssuesEvent | 2022-08-23 08:20:12 | Public-Health-Scotland/source-linkage-files | https://api.github.com/repos/Public-Health-Scotland/source-linkage-files | closed | Bug with `calculate_stay` | bug Priority: High | There is an issue with `calculate_stay`. If a `sc_qtr` is provided it will always ignore the `end_date`.
For example, the below should both return 1.
```R
> calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02"))
[1] 1
> calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02"), "2019Q4")
[1] 91
```
This was an oversight, as these cases aren't tested for either.
Tagging @Jennit07 as you wrote the function, please assign yourself if you're working on it, otherwise I will probably come back to it soon! | 1.0 | Bug with `calculate_stay` - There is an issue with `calculate_stay`. If a `sc_qtr` is provided it will always ignore the `end_date`.
For example, the below should both return 1.
```R
> calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02"))
[1] 1
> calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02"), "2019Q4")
[1] 91
```
This was an oversight, as these cases aren't tested for either.
Tagging @Jennit07 as you wrote the function, please assign yourself if you're working on it, otherwise I will probably come back to it soon! | priority | bug with calculate stay there is an issue with calculate stay if a sc qtr is provided it will always ignore the end date for example the below should both return r calculate stay as date as date calculate stay as date as date this was an oversight as these cases aren t tested for either tagging as you wrote the function please assign yourself if you re working on it otherwise i will probably come back to it soon | 1 |
93,485 | 3,900,865,156 | IssuesEvent | 2016-04-18 08:28:20 | DanGrew/JenkinsTestTracker | https://api.github.com/repos/DanGrew/JenkinsTestTracker | opened | Window resize moves divider | High Priority | The divider seems to be reset to 50% when resized or dragged to another screen. | 1.0 | Window resize moves divider - The divider seems to be reset to 50% when resized or dragged to another screen. | priority | window resize moves divider the divider seems to be reset to when resized or dragged to another screen | 1 |
200,745 | 7,011,252,618 | IssuesEvent | 2017-12-20 04:36:50 | PATRIC3/patric3_website | https://api.github.com/repos/PATRIC3/patric3_website | closed | ID Mapping: showing unwanted results | High Priority Tool: ID Mapping | I tried to search using a set of refseq locus tags below to find corresponding PATRIC ids.
BB0457
BB0836
BB0839
BB0344
BB0467
The result table show multiple matches for each of the ids, mapping to PATRIC and refseq features, and CDS and non CDS features, like shown in the attached image. This is very confusing for users for several reasons.
1. When I ask for a match to patric ids, the results should not show anything that doesn't have PATRIC ids.
Solution: : match for only annotation = PATRIC or PATRIC ID not null.
2. The table shows matches to multiple annotation sources and feature types. However, those attributes are not shown in the table. Making it difficult to understand what they are.
Solution: make the id mapping result table look similar to the feature table + one additional column for user provided ids. It will be good to have the facet filter above the table as well, to provide easy filtering.
<img width="1126" alt="screen shot 2017-12-07 at 6 35 46 am" src="https://user-images.githubusercontent.com/1236463/33716197-e1289bbc-db1a-11e7-925b-1731363fe552.png">
| 1.0 | ID Mapping: showing unwanted results - I tried to search using a set of refseq locus tags below to find corresponding PATRIC ids.
BB0457
BB0836
BB0839
BB0344
BB0467
The result table show multiple matches for each of the ids, mapping to PATRIC and refseq features, and CDS and non CDS features, like shown in the attached image. This is very confusing for users for several reasons.
1. When I ask for a match to patric ids, the results should not show anything that doesn't have PATRIC ids.
Solution: : match for only annotation = PATRIC or PATRIC ID not null.
2. The table shows matches to multiple annotation sources and feature types. However, those attributes are not shown in the table. Making it difficult to understand what they are.
Solution: make the id mapping result table look similar to the feature table + one additional column for user provided ids. It will be good to have the facet filter above the table as well, to provide easy filtering.
<img width="1126" alt="screen shot 2017-12-07 at 6 35 46 am" src="https://user-images.githubusercontent.com/1236463/33716197-e1289bbc-db1a-11e7-925b-1731363fe552.png">
| priority | id mapping showing unwanted results i tried to search using a set of refseq locus tags below to find corresponding patric ids the result table show multiple matches for each of the ids mapping to patric and refseq features and cds and non cds features like shown in the attached image this is very confusing for users for several reasons when i ask for a match to patric ids the results should not show anything that doesn t have patric ids solution match for only annotation patric or patric id not null the table shows matches to multiple annotation sources and feature types however those attributes are not shown in the table making it difficult to understand what they are solution make the id mapping result table look similar to the feature table one additional column for user provided ids it will be good to have the facet filter above the table as well to provide easy filtering img width alt screen shot at am src | 1 |
764,695 | 26,812,466,222 | IssuesEvent | 2023-02-01 23:56:41 | zulip/zulip | https://api.github.com/repos/zulip/zulip | opened | Add setup tips to user invite modal | area: onboarding priority: high area: invitations | As [discussed on CZO](https://chat.zulip.org/#narrow/stream/101-design/topic/setup.20tips.20in.20user.20invite.20modal/near/1490259), we should add some setup tips to the user invite modal.
General notes:
- Each of these would only be shown if the user has the permissions needed to take the relevant action.
- Include links to the appropriate settings panels throughout (noted below without their URLs).
## At the top of the modal
### Org profile configuration
- If org description is missing: "You may want to [configure]() your organization's login page prior to inviting users."
- If there is a description, but the org profile picture is missing: "You may want to [upload a profile picture]() for your organization prior to inviting users."
### User configs
- If default new user settings have not been changed, and custom profile fields have not been added: "You may want to configure [default new user settings]() and [custom profile fields]() prior to inviting users.
If only one of the above holds, just show the relevant part of the warning:
- You may want to configure [default new user settings]() prior to inviting users.
- You may want to configure [custom profile fields]() prior to inviting users.
## Just above the "Streams they should join" section
- If no streams have been created: "You may want to create streams prior to inviting users."
---
## Blockers:
- [ ] #23466 (?) | 1.0 | Add setup tips to user invite modal - As [discussed on CZO](https://chat.zulip.org/#narrow/stream/101-design/topic/setup.20tips.20in.20user.20invite.20modal/near/1490259), we should add some setup tips to the user invite modal.
General notes:
- Each of these would only be shown if the user has the permissions needed to take the relevant action.
- Include links to the appropriate settings panels throughout (noted below without their URLs).
## At the top of the modal
### Org profile configuration
- If org description is missing: "You may want to [configure]() your organization's login page prior to inviting users."
- If there is a description, but the org profile picture is missing: "You may want to [upload a profile picture]() for your organization prior to inviting users."
### User configs
- If default new user settings have not been changed, and custom profile fields have not been added: "You may want to configure [default new user settings]() and [custom profile fields]() prior to inviting users.
If only one of the above holds, just show the relevant part of the warning:
- You may want to configure [default new user settings]() prior to inviting users.
- You may want to configure [custom profile fields]() prior to inviting users.
## Just above the "Streams they should join" section
- If no streams have been created: "You may want to create streams prior to inviting users."
---
## Blockers:
- [ ] #23466 (?) | priority | add setup tips to user invite modal as we should add some setup tips to the user invite modal general notes each of these would only be shown if the user has the permissions needed to take the relevant action include links to the appropriate settings panels throughout noted below without their urls at the top of the modal org profile configuration if org description is missing you may want to your organization s login page prior to inviting users if there is a description but the org profile picture is missing you may want to for your organization prior to inviting users user configs if default new user settings have not been changed and custom profile fields have not been added you may want to configure and prior to inviting users if only one of the above holds just show the relevant part of the warning you may want to configure prior to inviting users you may want to configure prior to inviting users just above the streams they should join section if no streams have been created you may want to create streams prior to inviting users blockers | 1 |
230,653 | 7,612,618,402 | IssuesEvent | 2018-05-01 18:12:32 | AmatCoder/mednaffe | https://api.github.com/repos/AmatCoder/mednaffe | closed | Mednaffe won't accept up/right D-Pad inputs in controller setup | Priority-High bug | Using Mednaffe 0.8.7, trying to setup my controller for the PlayStation core. I'm using a DualShock 4 with DS4Windows, however this issue also occurs if I turn DS4Win off and use the DS4's native DirectInput.
When configuring my inputs, I can configure up/down/left/right just fine:

However, if I change tabs to 'Port 2', close the input config window, or do _essentially anything else_ that involves leaving the 'Port 1' input config screen, the config changes itself to this:

Sadly, playing games with only two directions available can be a _tad_ difficult ;P. Any idea what could be causing this? | 1.0 | Mednaffe won't accept up/right D-Pad inputs in controller setup - Using Mednaffe 0.8.7, trying to setup my controller for the PlayStation core. I'm using a DualShock 4 with DS4Windows, however this issue also occurs if I turn DS4Win off and use the DS4's native DirectInput.
When configuring my inputs, I can configure up/down/left/right just fine:

However, if I change tabs to 'Port 2', close the input config window, or do _essentially anything else_ that involves leaving the 'Port 1' input config screen, the config changes itself to this:

Sadly, playing games with only two directions available can be a _tad_ difficult ;P. Any idea what could be causing this? | priority | mednaffe won t accept up right d pad inputs in controller setup using mednaffe trying to setup my controller for the playstation core i m using a dualshock with however this issue also occurs if i turn off and use the s native directinput when configuring my inputs i can configure up down left right just fine however if i change tabs to port close the input config window or do essentially anything else that involves leaving the port input config screen the config changes itself to this sadly playing games with only two directions available can be a tad difficult p any idea what could be causing this | 1 |
279,858 | 8,674,601,015 | IssuesEvent | 2018-11-30 08:16:31 | FundacionParaguaya/MentorApp | https://api.github.com/repos/FundacionParaguaya/MentorApp | closed | Touch Responsiveness on Buttons (Globally through the app) | bug high priority question | The touch area of buttons is not responsive to a large enough area to be "usable".
This can be most clearly seen on the hamburger icon and on the skip question functionality.
We are visibly seeing users struggling when interacting with these buttons.
Please can you review and enlarge the area responsive to touch (should this be the issue) | 1.0 | Touch Responsiveness on Buttons (Globally through the app) - The touch area of buttons is not responsive to a large enough area to be "usable".
This can be most clearly seen on the hamburger icon and on the skip question functionality.
We are visibly seeing users struggling when interacting with these buttons.
Please can you review and enlarge the area responsive to touch (should this be the issue) | priority | touch responsiveness on buttons globally through the app the touch area of buttons is not responsive to a large enough area to be usable this can be most clearly seen on the hamburger icon and on the skip question functionality we are visibly seeing users struggling when interacting with these buttons please can you review and enlarge the area responsive to touch should this be the issue | 1 |
52,591 | 3,024,302,084 | IssuesEvent | 2015-08-02 13:35:50 | MinetestForFun/minetest-minetestforfun-server | https://api.github.com/repos/MinetestForFun/minetest-minetestforfun-server | opened | Too many boats | Modding ➤ Mapgen Priority: High | The sea_modpack include boats and uboots, they are too many, it's a problem.... And it makes the mapgen heavier !
- [ ] Divided per 4 or 5 the boats/uboots spawn rate
| 1.0 | Too many boats - The sea_modpack include boats and uboots, they are too many, it's a problem.... And it makes the mapgen heavier !
- [ ] Divided per 4 or 5 the boats/uboots spawn rate
| priority | too many boats the sea modpack include boats and uboots they are too many it s a problem and it makes the mapgen heavier divided per or the boats uboots spawn rate | 1 |
533,143 | 15,577,645,533 | IssuesEvent | 2021-03-17 13:45:41 | AY2021S2-CS2103T-W12-3/tp | https://api.github.com/repos/AY2021S2-CS2103T-W12-3/tp | closed | As an insurance agent, I can add, edit and view cruicial personal and insurance information of each client, | priority.High type.Epic | ... so that I can more easily keep track of my clients and their needs. | 1.0 | As an insurance agent, I can add, edit and view cruicial personal and insurance information of each client, - ... so that I can more easily keep track of my clients and their needs. | priority | as an insurance agent i can add edit and view cruicial personal and insurance information of each client so that i can more easily keep track of my clients and their needs | 1 |
127,295 | 5,028,194,746 | IssuesEvent | 2016-12-15 17:29:12 | odalic/sti | https://api.github.com/repos/odalic/sti | opened | UI - results - Labels for concept classifying columns not depicted? | bug priority: High | _From @tomas-knap on December 13, 2016 19:51_
When I run Odalic agains ADEQUATe KB, in case of disambiguation, UI properly shows labels (see below, URI contains UUID, but there is nice label). But in case of classificiation, label is not obtained and UUID is used. Why?

Solution: It should also take the label of the concept being suggested for classification.
@KataBoku did you encounter that as well?
_Copied from original issue: odalic/odalic-ui#190_ | 1.0 | UI - results - Labels for concept classifying columns not depicted? - _From @tomas-knap on December 13, 2016 19:51_
When I run Odalic agains ADEQUATe KB, in case of disambiguation, UI properly shows labels (see below, URI contains UUID, but there is nice label). But in case of classificiation, label is not obtained and UUID is used. Why?

Solution: It should also take the label of the concept being suggested for classification.
@KataBoku did you encounter that as well?
_Copied from original issue: odalic/odalic-ui#190_ | priority | ui results labels for concept classifying columns not depicted from tomas knap on december when i run odalic agains adequate kb in case of disambiguation ui properly shows labels see below uri contains uuid but there is nice label but in case of classificiation label is not obtained and uuid is used why solution it should also take the label of the concept being suggested for classification kataboku did you encounter that as well copied from original issue odalic odalic ui | 1 |
447,746 | 12,892,683,034 | IssuesEvent | 2020-07-13 20:06:17 | zulip/zulip | https://api.github.com/repos/zulip/zulip | closed | Improve virtualenv-clone hashing logic | area: production area: tooling bug in progress priority: high | Our `virtualenv` management logic is designed to have a few properties:
* After running `provision` or the equivalent production operation, you always have a correct virtualenv as though you'd built it fresh
* We avoid the performance pain of redownloading and/or rebuilding all of our packages fresh for the common operations of "adding a new package" or "upgrading a minor version of a package", both in the development environment and in production.
* https://zulip.readthedocs.io/en/latest/subsystems/dependencies.html is a useful reference.
There are a few issues we need to fix with the current implementation:
* We don't currently compare the `Python` version when deciding whether we need to build a new virtualenv from scratch.
* Our index of packages included in the virtualenv doesn't have package versions; we should include those and only allow package upgrades to use the `virtualenv-clone` plus cache approach (downgrades are often not tested upstream).
We should be able to test whether the implementation is working manually through inspection of `/srv/zulip-venv-cache/` before and after adding a new python dependency (see the docs for how to do this using `update-locked-requirements`), downgrading in `dev.txt`, and upgrading/downgrading the Python version itself. We don't currently have unit tests of any this logic, though it wouldn't be a bad idea to write some simple tests of the logic functions from `scripts/lib/setup_venv.py` in `tools/tests/` somewhere (`tools/test-tools` is the runner for those).
@andersk FYI; we might want to fix this before we merge https://github.com/zulip/zulip/pull/12837. | 1.0 | Improve virtualenv-clone hashing logic - Our `virtualenv` management logic is designed to have a few properties:
* After running `provision` or the equivalent production operation, you always have a correct virtualenv as though you'd built it fresh
* We avoid the performance pain of redownloading and/or rebuilding all of our packages fresh for the common operations of "adding a new package" or "upgrading a minor version of a package", both in the development environment and in production.
* https://zulip.readthedocs.io/en/latest/subsystems/dependencies.html is a useful reference.
There are a few issues we need to fix with the current implementation:
* We don't currently compare the `Python` version when deciding whether we need to build a new virtualenv from scratch.
* Our index of packages included in the virtualenv doesn't have package versions; we should include those and only allow package upgrades to use the `virtualenv-clone` plus cache approach (downgrades are often not tested upstream).
We should be able to test whether the implementation is working manually through inspection of `/srv/zulip-venv-cache/` before and after adding a new python dependency (see the docs for how to do this using `update-locked-requirements`), downgrading in `dev.txt`, and upgrading/downgrading the Python version itself. We don't currently have unit tests of any this logic, though it wouldn't be a bad idea to write some simple tests of the logic functions from `scripts/lib/setup_venv.py` in `tools/tests/` somewhere (`tools/test-tools` is the runner for those).
@andersk FYI; we might want to fix this before we merge https://github.com/zulip/zulip/pull/12837. | priority | improve virtualenv clone hashing logic our virtualenv management logic is designed to have a few properties after running provision or the equivalent production operation you always have a correct virtualenv as though you d built it fresh we avoid the performance pain of redownloading and or rebuilding all of our packages fresh for the common operations of adding a new package or upgrading a minor version of a package both in the development environment and in production is a useful reference there are a few issues we need to fix with the current implementation we don t currently compare the python version when deciding whether we need to build a new virtualenv from scratch our index of packages included in the virtualenv doesn t have package versions we should include those and only allow package upgrades to use the virtualenv clone plus cache approach downgrades are often not tested upstream we should be able to test whether the implementation is working manually through inspection of srv zulip venv cache before and after adding a new python dependency see the docs for how to do this using update locked requirements downgrading in dev txt and upgrading downgrading the python version itself we don t currently have unit tests of any this logic though it wouldn t be a bad idea to write some simple tests of the logic functions from scripts lib setup venv py in tools tests somewhere tools test tools is the runner for those andersk fyi we might want to fix this before we merge | 1 |
21,188 | 2,636,478,661 | IssuesEvent | 2015-03-10 04:01:07 | cs2103jan2015-w09-1j/main | https://api.github.com/repos/cs2103jan2015-w09-1j/main | closed | A user can create floating tasks | priority.high type.story type.task | ... so that the user can schedule them if he/she needs.
add -t{task} -des{task description} | 1.0 | A user can create floating tasks - ... so that the user can schedule them if he/she needs.
add -t{task} -des{task description} | priority | a user can create floating tasks so that the user can schedule them if he she needs add t task des task description | 1 |
142,818 | 5,477,616,537 | IssuesEvent | 2017-03-12 10:17:45 | fossasia/open-event-orga-server | https://api.github.com/repos/fossasia/open-event-orga-server | closed | Schedule Calendar View: Show calendar only 15 minutes after last session | bug enhancement Priority: High | As the session info is not showing over the surrounding areas, please reduce the time of the schedule shown after the last session.
Only show the schedule until "15 minutes after the last session". | 1.0 | Schedule Calendar View: Show calendar only 15 minutes after last session - As the session info is not showing over the surrounding areas, please reduce the time of the schedule shown after the last session.
Only show the schedule until "15 minutes after the last session". | priority | schedule calendar view show calendar only minutes after last session as the session info is not showing over the surrounding areas please reduce the time of the schedule shown after the last session only show the schedule until minutes after the last session | 1 |
359,986 | 10,683,166,116 | IssuesEvent | 2019-10-22 07:43:45 | xpressengine/xpressengine | https://api.github.com/repos/xpressengine/xpressengine | closed | 사이트 기본 설정을 변경 했을 때 로그인 설정이 변경되는 문제 | priority/high type/bug | 사이트 기본 설정을 변경하면 로그인 관련 설정이 초기화 돼서 로그인 불가 상태가 되고
로그인 관련 설정이 초기값으로 지정되고 있음 | 1.0 | 사이트 기본 설정을 변경 했을 때 로그인 설정이 변경되는 문제 - 사이트 기본 설정을 변경하면 로그인 관련 설정이 초기화 돼서 로그인 불가 상태가 되고
로그인 관련 설정이 초기값으로 지정되고 있음 | priority | 사이트 기본 설정을 변경 했을 때 로그인 설정이 변경되는 문제 사이트 기본 설정을 변경하면 로그인 관련 설정이 초기화 돼서 로그인 불가 상태가 되고 로그인 관련 설정이 초기값으로 지정되고 있음 | 1 |
601,788 | 18,432,318,728 | IssuesEvent | 2021-10-14 09:05:31 | haskell/cabal | https://api.github.com/repos/haskell/cabal | closed | `cabal install` rebuilds all dependencies after a `cabal build` | type: bug priority: high :fire: cabal-install: cmd/install | **Describe the bug**
`cabal install` rebuilds all dependencies after a `cabal build` in windows
**To Reproduce**
Steps to reproduce the behavior:
* A simple cabal project with only one dependency `say`
* First we build the project using a fresh store dir and print the store contents and the cabal-plan fingerprint
```
D:\ws\haskell\cabal-test>cabal clean
D:\ws\haskell\cabal-test>rmdir sr /s /q
D:\ws\haskell\cabal-test>cabal --store-dir .\sr build
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (first run)
- cabal-test-0.1.0.0 (exe:cabal-test) (first run)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
.....................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-13304\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4\lib
Configuring library for cabal-test-0.1.0.0..
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
............
Configuring executable 'cabal-test' for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
.......................
Linking D:\ws\haskell\cabal-test\dist-newstyle\build\x86_64-windows\ghc-8.6.5\cabal-test-0.1.0.0\x\cabal-test\build\cabal-test\cabal-test.exe ...
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
incoming
package.db
say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4
D:\ws\haskell\cabal-test>cabal-plan fingerprint
________________________________________________________________ B array-0.5.3.0
________________________________________________________________ B base-4.12.0.0
________________________________________________________________ B binary-0.8.6.0
________________________________________________________________ B bytestring-0.10.8.2
________________________________________________________________ L cabal-test-0.1.0.0
________________________________________________________________ B containers-0.6.0.1
________________________________________________________________ B deepseq-1.4.4.0
________________________________________________________________ B ghc-prim-0.5.3
________________________________________________________________ B integer-gmp-1.0.2.0
________________________________________________________________ B rts-1.0
f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1
________________________________________________________________ B text-1.2.3.1
________________________________________________________________ B transformers-0.5.6.2
```
* Then we install the project using the same store dir
* Another bug: you can set a relative dir for the store using `cabal build` but no using `cabal install`
```
D:\ws\haskell\cabal-test>cabal --store-dir .\sr install --builddir dist2
Wrote tarball sdist to
D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (requires build)
- cabal-test-0.1.0.0 (exe:cabal-test) (requires build)
Configuring library for say-0.1.0.1..
cabal-3.6.2.0.exe: expected an absolute directory name for --prefix:
.\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
cabal-3.6.2.0.exe: Failed to build say-0.1.0.1 (which is required by
exe:cabal-test from cabal-test-0.1.0.0). The failure occurred during the
configure step.
D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install -
-builddir dist2
Wrote tarball sdist to
D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (requires build)
- cabal-test-0.1.0.0 (exe:cabal-test) (requires build)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5
e3ea47\lib
Configuring library for cabal-test-0.1.0.0..
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
..............
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88\lib
Configuring executable 'cabal-test' for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
...................
Linking dist\build\cabal-test\cabal-test.exe ...
Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9
d05e08950e4e5de83d054a50961\bin
Warning: The directory
D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin
is not in the system search path.
Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe'
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961
cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88
incoming
package.db
say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4
say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=dist2
________________________________________________________________ B array-0.5.3.0
________________________________________________________________ B base-4.12.0.0
________________________________________________________________ B binary-0.8.6.0
________________________________________________________________ B bytestring-0.10.8.2
________________________________________________________________ L cabal-test-0.1.0.0
________________________________________________________________ B containers-0.6.0.1
________________________________________________________________ B deepseq-1.4.4.0
________________________________________________________________ B ghc-prim-0.5.3
________________________________________________________________ B integer-gmp-1.0.2.0
________________________________________________________________ B rts-1.0
f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1
________________________________________________________________ B text-1.2.3.1
________________________________________________________________ B transformers-0.5.6.2
D:\ws\haskell\cabal-test>cabal-plan diff --builddir=dist-newstyle --builddir=dist2
(No output, lovely)
```
* As you can check `cabal-plan diff` does not output anything, from other tests it seems it is broken in windows, including in the msys2 shell
* After deleting the store we do the same in reverse order, this time preserving the temp build dir used by `cabal install`
* There is no flag to preserve the directory, we have to copy the tmp dir quickly :facepalm:
```
D:\ws\haskell\cabal-test> rmdir sr /s /q
D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install -builddir dist2
Wrote tarball sdist to
D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (requires build)
- cabal-test-0.1.0.0 (exe:cabal-test) (requires build)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
............................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7de3ea47\lib
Configuring library for cabal-test-0.1.0.0..
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
..................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce9bbdaaf168f88\lib
Configuring executable 'cabal-test' for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
..................
Linking dist\build\cabal-test\cabal-test.exe ...
Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incomng\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfedd05e08950e4e5de83d054a50961\bin
Warning: The directory
D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\r\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin is not in the system search path.
Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe'
cabal-3.6.2.0.exe: Path 'D:\cabal\bin\cabal-test' already exists. Use --overwrite-policy=always to overwrite.
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961
cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88
incoming
package.db
say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
```
* the fingerprint for `say` is the same in the temp build dir
```
D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=tmp2\cabal-install.-1
0896\dist2
________________________________________________________________ B array-0.5.3.0
________________________________________________________________ B base-4.12.0.0
________________________________________________________________ B binary-0.8.6.0
________________________________________________________________ B bytestring-0.10.8.2
14aa96bfe9d34856cf85a09bc4b314aac137e5e325f3f4beeb77b70ebc57a4a6 G cabal-test-0.1.0.0
________________________________________________________________ B containers-0.6.0.1
________________________________________________________________ B deepseq-1.4.4.0
________________________________________________________________ B ghc-prim-0.5.3
________________________________________________________________ B integer-gmp-1.0.2.0
________________________________________________________________ B rts-1.0
f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1
________________________________________________________________ B text-1.2.3.1
________________________________________________________________ B transformers-0.5.6.2
```
* we do the build after the install, deps are rebuilt again:
```
D:\ws\haskell\cabal-test>cabal --store-dir .\sr build
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (dependency rebuilt)
- cabal-test-0.1.0.0 (exe:cabal-test) (dependency rebuilt)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
[1 of 2] Compiling Paths_say ( dist\build\autogen\Paths_say.hs, dist\buil
d\Paths_say.o )
[2 of 2] Compiling Say ( src\Say.hs, dist\build\Say.o )
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-9200\ws
\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c
596e4\lib
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961
cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88
incoming
package.db
say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4
say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
```
* output of `cabal info` for all the three build dirs involved:
<details> <summary> cabal info </summary>
```
# cabal-plan info --builddir=dist-newstyle
Tree
~~~~
cabal-test-0.1.0.0
[cabal-test-0.1.0.0:exe:"cabal-test"]
├─ base-4.12.0.0
│ ├─ ghc-prim-0.5.3
│ │ └─ rts-1.0
│ ├─ integer-gmp-1.0.2.0
│ │ └─ ghc-prim-0.5.3 ┄┄
│ └─ rts-1.0 ┄┄
├─ bytestring-0.10.8.2
│ ├─ base-4.12.0.0 ┄┄
│ ├─ deepseq-1.4.4.0
│ │ ├─ array-0.5.3.0
│ │ │ └─ base-4.12.0.0 ┄┄
│ │ └─ base-4.12.0.0 ┄┄
│ ├─ ghc-prim-0.5.3 ┄┄
│ └─ integer-gmp-1.0.2.0 ┄┄
└─ cabal-test-0.1.0.0
├─ base-4.12.0.0 ┄┄
├─ bytestring-0.10.8.2 ┄┄
├─ say-0.1.0.1
│ ├─ base-4.12.0.0 ┄┄
│ ├─ bytestring-0.10.8.2 ┄┄
│ ├─ text-1.2.3.1
│ │ ├─ array-0.5.3.0 ┄┄
│ │ ├─ base-4.12.0.0 ┄┄
│ │ ├─ binary-0.8.6.0
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ │ └─ containers-0.6.0.1
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ │ └─ ghc-prim-0.5.3 ┄┄
│ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ ├─ ghc-prim-0.5.3 ┄┄
│ │ └─ integer-gmp-1.0.2.0 ┄┄
│ └─ transformers-0.5.6.2
│ └─ base-4.12.0.0 ┄┄
└─ text-1.2.3.1 ┄┄
Top-sorted
~~~~~~~~~~
UnitId "rts"
UnitId "ghc-prim-0.5.3"
UnitId "integer-gmp-1.0.2.0"
UnitId "base-4.12.0.0"
UnitId "array-0.5.3.0"
UnitId "deepseq-1.4.4.0"
UnitId "bytestring-0.10.8.2"
UnitId "containers-0.6.0.1"
UnitId "binary-0.8.6.0"
UnitId "text-1.2.3.1"
UnitId "transformers-0.5.6.2"
UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4"
UnitId "cabal-test-0.1.0.0-inplace"
UnitId "cabal-test-0.1.0.0-inplace-cabal-test"
Direct deps
~~~~~~~~~~~
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameLib
base-4.12.0.0
bytestring-0.10.8.2
say-0.1.0.1
text-1.2.3.1
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameExe "cabal-test"
base-4.12.0.0
bytestring-0.10.8.2
cabal-test-0.1.0.0
```
```
# cabal-plan info --builddir=dist2
Tree
~~~~
cabal-test-0.1.0.0
[cabal-test-0.1.0.0:exe:"cabal-test"]
├─ base-4.12.0.0
│ ├─ ghc-prim-0.5.3
│ │ └─ rts-1.0
│ ├─ integer-gmp-1.0.2.0
│ │ └─ ghc-prim-0.5.3 ┄┄
│ └─ rts-1.0 ┄┄
├─ bytestring-0.10.8.2
│ ├─ base-4.12.0.0 ┄┄
│ ├─ deepseq-1.4.4.0
│ │ ├─ array-0.5.3.0
│ │ │ └─ base-4.12.0.0 ┄┄
│ │ └─ base-4.12.0.0 ┄┄
│ ├─ ghc-prim-0.5.3 ┄┄
│ └─ integer-gmp-1.0.2.0 ┄┄
└─ cabal-test-0.1.0.0
├─ base-4.12.0.0 ┄┄
├─ bytestring-0.10.8.2 ┄┄
├─ say-0.1.0.1
│ ├─ base-4.12.0.0 ┄┄
│ ├─ bytestring-0.10.8.2 ┄┄
│ ├─ text-1.2.3.1
│ │ ├─ array-0.5.3.0 ┄┄
│ │ ├─ base-4.12.0.0 ┄┄
│ │ ├─ binary-0.8.6.0
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ │ └─ containers-0.6.0.1
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ │ └─ ghc-prim-0.5.3 ┄┄
│ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ ├─ ghc-prim-0.5.3 ┄┄
│ │ └─ integer-gmp-1.0.2.0 ┄┄
│ └─ transformers-0.5.6.2
│ └─ base-4.12.0.0 ┄┄
└─ text-1.2.3.1 ┄┄
Top-sorted
~~~~~~~~~~
UnitId "rts"
UnitId "ghc-prim-0.5.3"
UnitId "integer-gmp-1.0.2.0"
UnitId "base-4.12.0.0"
UnitId "array-0.5.3.0"
UnitId "deepseq-1.4.4.0"
UnitId "bytestring-0.10.8.2"
UnitId "containers-0.6.0.1"
UnitId "binary-0.8.6.0"
UnitId "text-1.2.3.1"
UnitId "transformers-0.5.6.2"
UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4"
UnitId "cabal-test-0.1.0.0-inplace"
UnitId "cabal-test-0.1.0.0-inplace-cabal-test"
Direct deps
~~~~~~~~~~~
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameLib
base-4.12.0.0
bytestring-0.10.8.2
say-0.1.0.1
text-1.2.3.1
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameExe "cabal-test"
base-4.12.0.0
bytestring-0.10.8.2
cabal-test-0.1.0.0
```
```
# cabal-plan info --builddir ./tmp2/cabal-install.-10896/dist2/
Tree
~~~~
cabal-test-0.1.0.0
[cabal-test-0.1.0.0:exe:"cabal-test"]
├─ base-4.12.0.0
│ ├─ ghc-prim-0.5.3
│ │ └─ rts-1.0
│ ├─ integer-gmp-1.0.2.0
│ │ └─ ghc-prim-0.5.3 ┄┄
│ └─ rts-1.0 ┄┄
├─ bytestring-0.10.8.2
│ ├─ base-4.12.0.0 ┄┄
│ ├─ deepseq-1.4.4.0
│ │ ├─ array-0.5.3.0
│ │ │ └─ base-4.12.0.0 ┄┄
│ │ └─ base-4.12.0.0 ┄┄
│ ├─ ghc-prim-0.5.3 ┄┄
│ └─ integer-gmp-1.0.2.0 ┄┄
└─ cabal-test-0.1.0.0
├─ base-4.12.0.0 ┄┄
├─ bytestring-0.10.8.2 ┄┄
├─ say-0.1.0.1
│ ├─ base-4.12.0.0 ┄┄
│ ├─ bytestring-0.10.8.2 ┄┄
│ ├─ text-1.2.3.1
│ │ ├─ array-0.5.3.0 ┄┄
│ │ ├─ base-4.12.0.0 ┄┄
│ │ ├─ binary-0.8.6.0
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ │ └─ containers-0.6.0.1
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ │ └─ ghc-prim-0.5.3 ┄┄
│ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ ├─ ghc-prim-0.5.3 ┄┄
│ │ └─ integer-gmp-1.0.2.0 ┄┄
│ └─ transformers-0.5.6.2
│ └─ base-4.12.0.0 ┄┄
└─ text-1.2.3.1 ┄┄
Top-sorted
~~~~~~~~~~
UnitId "rts"
UnitId "ghc-prim-0.5.3"
UnitId "integer-gmp-1.0.2.0"
UnitId "base-4.12.0.0"
UnitId "array-0.5.3.0"
UnitId "deepseq-1.4.4.0"
UnitId "bytestring-0.10.8.2"
UnitId "containers-0.6.0.1"
UnitId "binary-0.8.6.0"
UnitId "text-1.2.3.1"
UnitId "transformers-0.5.6.2"
UnitId "say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47"
UnitId "cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88"
UnitId "cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961"
Direct deps
~~~~~~~~~~~
```
</details>
* Summary: the unit id for say (`f56dd8dcb828e676e2fe979752cedce7d5e3ea47` from the temp build dir is different from the other two (`9d85f3e60b43742f757835053a7ea996c1c596e4`). Both are installed in the store dir.
* `cabal-plan diff` does not output anything (and afaik it should at least in the first case)
```
# cabal-plan diff --builddir=dist-newstyle --builddir ./tmp2/cabal-install.-10896/dist2
# cabal-plan diff --builddir=./dist-newstyle --builddir=./dist2
#
```
**Expected behavior**
`cabal install` should only rebuild local packages like in linux
**System information**
- Operating system: windows 10 and 7
- `cabal`: 3.6.2.0 but is reproduced for sure with 3.6.0.0 and 3.4.0.0 and afair with any cabal version i tried since 2.4.0.0
- `ghc` versions: 8.6.5 and 8.10.7, but probably with any
**Additional context**
* I decide to open a new issue as #7693 is about another different issue and #6919 talks about rebuild the local packages in any os
* See another reproduction case with haskell-language-server in https://github.com/haskell/cabal/issues/7693#issuecomment-940472595
@fgaz @gbaz i hope there is enough info to at least demonstrate the behaviour, let me know if more info or tests are needed | 1.0 | `cabal install` rebuilds all dependencies after a `cabal build` - **Describe the bug**
`cabal install` rebuilds all dependencies after a `cabal build` in windows
**To Reproduce**
Steps to reproduce the behavior:
* A simple cabal project with only one dependency `say`
* First we build the project using a fresh store dir and print the store contents and the cabal-plan fingerprint
```
D:\ws\haskell\cabal-test>cabal clean
D:\ws\haskell\cabal-test>rmdir sr /s /q
D:\ws\haskell\cabal-test>cabal --store-dir .\sr build
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (first run)
- cabal-test-0.1.0.0 (exe:cabal-test) (first run)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
.....................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-13304\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4\lib
Configuring library for cabal-test-0.1.0.0..
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
............
Configuring executable 'cabal-test' for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
.......................
Linking D:\ws\haskell\cabal-test\dist-newstyle\build\x86_64-windows\ghc-8.6.5\cabal-test-0.1.0.0\x\cabal-test\build\cabal-test\cabal-test.exe ...
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
incoming
package.db
say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4
D:\ws\haskell\cabal-test>cabal-plan fingerprint
________________________________________________________________ B array-0.5.3.0
________________________________________________________________ B base-4.12.0.0
________________________________________________________________ B binary-0.8.6.0
________________________________________________________________ B bytestring-0.10.8.2
________________________________________________________________ L cabal-test-0.1.0.0
________________________________________________________________ B containers-0.6.0.1
________________________________________________________________ B deepseq-1.4.4.0
________________________________________________________________ B ghc-prim-0.5.3
________________________________________________________________ B integer-gmp-1.0.2.0
________________________________________________________________ B rts-1.0
f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1
________________________________________________________________ B text-1.2.3.1
________________________________________________________________ B transformers-0.5.6.2
```
* Then we install the project using the same store dir
* Another bug: you can set a relative dir for the store using `cabal build` but no using `cabal install`
```
D:\ws\haskell\cabal-test>cabal --store-dir .\sr install --builddir dist2
Wrote tarball sdist to
D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (requires build)
- cabal-test-0.1.0.0 (exe:cabal-test) (requires build)
Configuring library for say-0.1.0.1..
cabal-3.6.2.0.exe: expected an absolute directory name for --prefix:
.\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
cabal-3.6.2.0.exe: Failed to build say-0.1.0.1 (which is required by
exe:cabal-test from cabal-test-0.1.0.0). The failure occurred during the
configure step.
D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install -
-builddir dist2
Wrote tarball sdist to
D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (requires build)
- cabal-test-0.1.0.0 (exe:cabal-test) (requires build)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5
e3ea47\lib
Configuring library for cabal-test-0.1.0.0..
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
..............
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88\lib
Configuring executable 'cabal-test' for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
...................
Linking dist\build\cabal-test\cabal-test.exe ...
Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9
d05e08950e4e5de83d054a50961\bin
Warning: The directory
D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin
is not in the system search path.
Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe'
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961
cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88
incoming
package.db
say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4
say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=dist2
________________________________________________________________ B array-0.5.3.0
________________________________________________________________ B base-4.12.0.0
________________________________________________________________ B binary-0.8.6.0
________________________________________________________________ B bytestring-0.10.8.2
________________________________________________________________ L cabal-test-0.1.0.0
________________________________________________________________ B containers-0.6.0.1
________________________________________________________________ B deepseq-1.4.4.0
________________________________________________________________ B ghc-prim-0.5.3
________________________________________________________________ B integer-gmp-1.0.2.0
________________________________________________________________ B rts-1.0
f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1
________________________________________________________________ B text-1.2.3.1
________________________________________________________________ B transformers-0.5.6.2
D:\ws\haskell\cabal-test>cabal-plan diff --builddir=dist-newstyle --builddir=dist2
(No output, lovely)
```
* As you can check `cabal-plan diff` does not output anything, from other tests it seems it is broken in windows, including in the msys2 shell
* After deleting the store we do the same in reverse order, this time preserving the temp build dir used by `cabal install`
* There is no flag to preserve the directory, we have to copy the tmp dir quickly :facepalm:
```
D:\ws\haskell\cabal-test> rmdir sr /s /q
D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install -builddir dist2
Wrote tarball sdist to
D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz
Resolving dependencies...
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (requires build)
- cabal-test-0.1.0.0 (exe:cabal-test) (requires build)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
............................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7de3ea47\lib
Configuring library for cabal-test-0.1.0.0..
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
..................
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce9bbdaaf168f88\lib
Configuring executable 'cabal-test' for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
..................
Linking dist\build\cabal-test\cabal-test.exe ...
Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incomng\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfedd05e08950e4e5de83d054a50961\bin
Warning: The directory
D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\r\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin is not in the system search path.
Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe'
cabal-3.6.2.0.exe: Path 'D:\cabal\bin\cabal-test' already exists. Use --overwrite-policy=always to overwrite.
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961
cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88
incoming
package.db
say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
```
* the fingerprint for `say` is the same in the temp build dir
```
D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=tmp2\cabal-install.-1
0896\dist2
________________________________________________________________ B array-0.5.3.0
________________________________________________________________ B base-4.12.0.0
________________________________________________________________ B binary-0.8.6.0
________________________________________________________________ B bytestring-0.10.8.2
14aa96bfe9d34856cf85a09bc4b314aac137e5e325f3f4beeb77b70ebc57a4a6 G cabal-test-0.1.0.0
________________________________________________________________ B containers-0.6.0.1
________________________________________________________________ B deepseq-1.4.4.0
________________________________________________________________ B ghc-prim-0.5.3
________________________________________________________________ B integer-gmp-1.0.2.0
________________________________________________________________ B rts-1.0
f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1
________________________________________________________________ B text-1.2.3.1
________________________________________________________________ B transformers-0.5.6.2
```
* we do the build after the install, deps are rebuilt again:
```
D:\ws\haskell\cabal-test>cabal --store-dir .\sr build
Build profile: -w ghc-8.6.5 -O1
In order, the following will be built (use -v for more details):
- say-0.1.0.1 (lib) (requires build)
- cabal-test-0.1.0.0 (lib) (dependency rebuilt)
- cabal-test-0.1.0.0 (exe:cabal-test) (dependency rebuilt)
Configuring library for say-0.1.0.1..
Preprocessing library for say-0.1.0.1..
Building library for say-0.1.0.1..
[1 of 2] Compiling Paths_say ( dist\build\autogen\Paths_say.hs, dist\buil
d\Paths_say.o )
[2 of 2] Compiling Say ( src\Say.hs, dist\build\Say.o )
Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-9200\ws
\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c
596e4\lib
Preprocessing library for cabal-test-0.1.0.0..
Building library for cabal-test-0.1.0.0..
Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0..
Building executable 'cabal-test' for cabal-test-0.1.0.0..
D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b
cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961
cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88
incoming
package.db
say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4
say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47
```
* output of `cabal info` for all the three build dirs involved:
<details> <summary> cabal info </summary>
```
# cabal-plan info --builddir=dist-newstyle
Tree
~~~~
cabal-test-0.1.0.0
[cabal-test-0.1.0.0:exe:"cabal-test"]
├─ base-4.12.0.0
│ ├─ ghc-prim-0.5.3
│ │ └─ rts-1.0
│ ├─ integer-gmp-1.0.2.0
│ │ └─ ghc-prim-0.5.3 ┄┄
│ └─ rts-1.0 ┄┄
├─ bytestring-0.10.8.2
│ ├─ base-4.12.0.0 ┄┄
│ ├─ deepseq-1.4.4.0
│ │ ├─ array-0.5.3.0
│ │ │ └─ base-4.12.0.0 ┄┄
│ │ └─ base-4.12.0.0 ┄┄
│ ├─ ghc-prim-0.5.3 ┄┄
│ └─ integer-gmp-1.0.2.0 ┄┄
└─ cabal-test-0.1.0.0
├─ base-4.12.0.0 ┄┄
├─ bytestring-0.10.8.2 ┄┄
├─ say-0.1.0.1
│ ├─ base-4.12.0.0 ┄┄
│ ├─ bytestring-0.10.8.2 ┄┄
│ ├─ text-1.2.3.1
│ │ ├─ array-0.5.3.0 ┄┄
│ │ ├─ base-4.12.0.0 ┄┄
│ │ ├─ binary-0.8.6.0
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ │ └─ containers-0.6.0.1
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ │ └─ ghc-prim-0.5.3 ┄┄
│ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ ├─ ghc-prim-0.5.3 ┄┄
│ │ └─ integer-gmp-1.0.2.0 ┄┄
│ └─ transformers-0.5.6.2
│ └─ base-4.12.0.0 ┄┄
└─ text-1.2.3.1 ┄┄
Top-sorted
~~~~~~~~~~
UnitId "rts"
UnitId "ghc-prim-0.5.3"
UnitId "integer-gmp-1.0.2.0"
UnitId "base-4.12.0.0"
UnitId "array-0.5.3.0"
UnitId "deepseq-1.4.4.0"
UnitId "bytestring-0.10.8.2"
UnitId "containers-0.6.0.1"
UnitId "binary-0.8.6.0"
UnitId "text-1.2.3.1"
UnitId "transformers-0.5.6.2"
UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4"
UnitId "cabal-test-0.1.0.0-inplace"
UnitId "cabal-test-0.1.0.0-inplace-cabal-test"
Direct deps
~~~~~~~~~~~
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameLib
base-4.12.0.0
bytestring-0.10.8.2
say-0.1.0.1
text-1.2.3.1
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameExe "cabal-test"
base-4.12.0.0
bytestring-0.10.8.2
cabal-test-0.1.0.0
```
```
# cabal-plan info --builddir=dist2
Tree
~~~~
cabal-test-0.1.0.0
[cabal-test-0.1.0.0:exe:"cabal-test"]
├─ base-4.12.0.0
│ ├─ ghc-prim-0.5.3
│ │ └─ rts-1.0
│ ├─ integer-gmp-1.0.2.0
│ │ └─ ghc-prim-0.5.3 ┄┄
│ └─ rts-1.0 ┄┄
├─ bytestring-0.10.8.2
│ ├─ base-4.12.0.0 ┄┄
│ ├─ deepseq-1.4.4.0
│ │ ├─ array-0.5.3.0
│ │ │ └─ base-4.12.0.0 ┄┄
│ │ └─ base-4.12.0.0 ┄┄
│ ├─ ghc-prim-0.5.3 ┄┄
│ └─ integer-gmp-1.0.2.0 ┄┄
└─ cabal-test-0.1.0.0
├─ base-4.12.0.0 ┄┄
├─ bytestring-0.10.8.2 ┄┄
├─ say-0.1.0.1
│ ├─ base-4.12.0.0 ┄┄
│ ├─ bytestring-0.10.8.2 ┄┄
│ ├─ text-1.2.3.1
│ │ ├─ array-0.5.3.0 ┄┄
│ │ ├─ base-4.12.0.0 ┄┄
│ │ ├─ binary-0.8.6.0
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ │ └─ containers-0.6.0.1
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ │ └─ ghc-prim-0.5.3 ┄┄
│ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ ├─ ghc-prim-0.5.3 ┄┄
│ │ └─ integer-gmp-1.0.2.0 ┄┄
│ └─ transformers-0.5.6.2
│ └─ base-4.12.0.0 ┄┄
└─ text-1.2.3.1 ┄┄
Top-sorted
~~~~~~~~~~
UnitId "rts"
UnitId "ghc-prim-0.5.3"
UnitId "integer-gmp-1.0.2.0"
UnitId "base-4.12.0.0"
UnitId "array-0.5.3.0"
UnitId "deepseq-1.4.4.0"
UnitId "bytestring-0.10.8.2"
UnitId "containers-0.6.0.1"
UnitId "binary-0.8.6.0"
UnitId "text-1.2.3.1"
UnitId "transformers-0.5.6.2"
UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4"
UnitId "cabal-test-0.1.0.0-inplace"
UnitId "cabal-test-0.1.0.0-inplace-cabal-test"
Direct deps
~~~~~~~~~~~
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameLib
base-4.12.0.0
bytestring-0.10.8.2
say-0.1.0.1
text-1.2.3.1
PkgId (PkgName "cabal-test") (Ver [0,1,0,0])
CompNameExe "cabal-test"
base-4.12.0.0
bytestring-0.10.8.2
cabal-test-0.1.0.0
```
```
# cabal-plan info --builddir ./tmp2/cabal-install.-10896/dist2/
Tree
~~~~
cabal-test-0.1.0.0
[cabal-test-0.1.0.0:exe:"cabal-test"]
├─ base-4.12.0.0
│ ├─ ghc-prim-0.5.3
│ │ └─ rts-1.0
│ ├─ integer-gmp-1.0.2.0
│ │ └─ ghc-prim-0.5.3 ┄┄
│ └─ rts-1.0 ┄┄
├─ bytestring-0.10.8.2
│ ├─ base-4.12.0.0 ┄┄
│ ├─ deepseq-1.4.4.0
│ │ ├─ array-0.5.3.0
│ │ │ └─ base-4.12.0.0 ┄┄
│ │ └─ base-4.12.0.0 ┄┄
│ ├─ ghc-prim-0.5.3 ┄┄
│ └─ integer-gmp-1.0.2.0 ┄┄
└─ cabal-test-0.1.0.0
├─ base-4.12.0.0 ┄┄
├─ bytestring-0.10.8.2 ┄┄
├─ say-0.1.0.1
│ ├─ base-4.12.0.0 ┄┄
│ ├─ bytestring-0.10.8.2 ┄┄
│ ├─ text-1.2.3.1
│ │ ├─ array-0.5.3.0 ┄┄
│ │ ├─ base-4.12.0.0 ┄┄
│ │ ├─ binary-0.8.6.0
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ │ └─ containers-0.6.0.1
│ │ │ ├─ array-0.5.3.0 ┄┄
│ │ │ ├─ base-4.12.0.0 ┄┄
│ │ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ │ └─ ghc-prim-0.5.3 ┄┄
│ │ ├─ bytestring-0.10.8.2 ┄┄
│ │ ├─ deepseq-1.4.4.0 ┄┄
│ │ ├─ ghc-prim-0.5.3 ┄┄
│ │ └─ integer-gmp-1.0.2.0 ┄┄
│ └─ transformers-0.5.6.2
│ └─ base-4.12.0.0 ┄┄
└─ text-1.2.3.1 ┄┄
Top-sorted
~~~~~~~~~~
UnitId "rts"
UnitId "ghc-prim-0.5.3"
UnitId "integer-gmp-1.0.2.0"
UnitId "base-4.12.0.0"
UnitId "array-0.5.3.0"
UnitId "deepseq-1.4.4.0"
UnitId "bytestring-0.10.8.2"
UnitId "containers-0.6.0.1"
UnitId "binary-0.8.6.0"
UnitId "text-1.2.3.1"
UnitId "transformers-0.5.6.2"
UnitId "say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47"
UnitId "cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88"
UnitId "cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961"
Direct deps
~~~~~~~~~~~
```
</details>
* Summary: the unit id for say (`f56dd8dcb828e676e2fe979752cedce7d5e3ea47` from the temp build dir is different from the other two (`9d85f3e60b43742f757835053a7ea996c1c596e4`). Both are installed in the store dir.
* `cabal-plan diff` does not output anything (and afaik it should at least in the first case)
```
# cabal-plan diff --builddir=dist-newstyle --builddir ./tmp2/cabal-install.-10896/dist2
# cabal-plan diff --builddir=./dist-newstyle --builddir=./dist2
#
```
**Expected behavior**
`cabal install` should only rebuild local packages like in linux
**System information**
- Operating system: windows 10 and 7
- `cabal`: 3.6.2.0 but is reproduced for sure with 3.6.0.0 and 3.4.0.0 and afair with any cabal version i tried since 2.4.0.0
- `ghc` versions: 8.6.5 and 8.10.7, but probably with any
**Additional context**
* I decide to open a new issue as #7693 is about another different issue and #6919 talks about rebuild the local packages in any os
* See another reproduction case with haskell-language-server in https://github.com/haskell/cabal/issues/7693#issuecomment-940472595
@fgaz @gbaz i hope there is enough info to at least demonstrate the behaviour, let me know if more info or tests are needed | priority | cabal install rebuilds all dependencies after a cabal build describe the bug cabal install rebuilds all dependencies after a cabal build in windows to reproduce steps to reproduce the behavior a simple cabal project with only one dependency say first we build the project using a fresh store dir and print the store contents and the cabal plan fingerprint d ws haskell cabal test cabal clean d ws haskell cabal test rmdir sr s q d ws haskell cabal test cabal store dir sr build resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib first run cabal test exe cabal test first run configuring library for say preprocessing library for say building library for say installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc say lib configuring library for cabal test preprocessing library for cabal test building library for cabal test configuring executable cabal test for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test linking d ws haskell cabal test dist newstyle build windows ghc cabal test x cabal test build cabal test cabal test exe d ws haskell cabal test dir sr ghc b incoming package db say d ws haskell cabal test cabal plan fingerprint b array b base b binary b bytestring l cabal test b containers b deepseq b ghc prim b integer gmp b rts g say b text b transformers then we install the project using the same store dir another bug you can set a relative dir for the store using cabal build but no using cabal install d ws haskell cabal test cabal store dir sr install builddir wrote tarball sdist to d ws haskell cabal test sdist cabal test tar gz resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib requires build cabal test exe cabal test requires build configuring library for say cabal exe expected an absolute directory name for prefix sr ghc say cabal exe failed to build say which is required by exe cabal test from cabal test the failure occurred during the configure step d ws haskell cabal test cabal store dir d ws haskell cabal test sr install builddir wrote tarball sdist to d ws haskell cabal test sdist cabal test tar gz resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib requires build cabal test exe cabal test requires build configuring library for say preprocessing library for say building library for say installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc say lib configuring library for cabal test preprocessing library for cabal test building library for cabal test installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc cabal test lib configuring executable cabal test for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test linking dist build cabal test cabal test exe installing executable cabal test in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc cabal test bin warning the directory d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc cabal test bin is not in the system search path symlinking cabal test exe to d cabal bin cabal test exe d ws haskell cabal test dir sr ghc b cabal test cabal test incoming package db say say d ws haskell cabal test cabal plan fingerprint builddir b array b base b binary b bytestring l cabal test b containers b deepseq b ghc prim b integer gmp b rts g say b text b transformers d ws haskell cabal test cabal plan diff builddir dist newstyle builddir no output lovely as you can check cabal plan diff does not output anything from other tests it seems it is broken in windows including in the shell after deleting the store we do the same in reverse order this time preserving the temp build dir used by cabal install there is no flag to preserve the directory we have to copy the tmp dir quickly facepalm d ws haskell cabal test rmdir sr s q d ws haskell cabal test cabal store dir d ws haskell cabal test sr install builddir wrote tarball sdist to d ws haskell cabal test sdist cabal test tar gz resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib requires build cabal test exe cabal test requires build configuring library for say preprocessing library for say building library for say installing library in d ws haskell cabal test sr ghc incoming new s haskell cabal test sr ghc say lib configuring library for cabal test preprocessing library for cabal test building library for cabal test installing library in d ws haskell cabal test sr ghc incoming new s haskell cabal test sr ghc cabal test lib configuring executable cabal test for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test linking dist build cabal test cabal test exe installing executable cabal test in d ws haskell cabal test sr ghc incomng new ws haskell cabal test sr ghc cabal test bin warning the directory d ws haskell cabal test sr ghc incoming new ws haskell cabal test r ghc cabal test bin is not in the system search path symlinking cabal test exe to d cabal bin cabal test exe cabal exe path d cabal bin cabal test already exists use overwrite policy always to overwrite d ws haskell cabal test dir sr ghc b cabal test cabal test incoming package db say the fingerprint for say is the same in the temp build dir d ws haskell cabal test cabal plan fingerprint builddir cabal install b array b base b binary b bytestring g cabal test b containers b deepseq b ghc prim b integer gmp b rts g say b text b transformers we do the build after the install deps are rebuilt again d ws haskell cabal test cabal store dir sr build build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib dependency rebuilt cabal test exe cabal test dependency rebuilt configuring library for say preprocessing library for say building library for say compiling paths say dist build autogen paths say hs dist buil d paths say o compiling say src say hs dist build say o installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc say lib preprocessing library for cabal test building library for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test d ws haskell cabal test dir sr ghc b cabal test cabal test incoming package db say say output of cabal info for all the three build dirs involved cabal info cabal plan info builddir dist newstyle tree cabal test ├─ base │ ├─ ghc prim │ │ └─ rts │ ├─ integer gmp │ │ └─ ghc prim ┄┄ │ └─ rts ┄┄ ├─ bytestring │ ├─ base ┄┄ │ ├─ deepseq │ │ ├─ array │ │ │ └─ base ┄┄ │ │ └─ base ┄┄ │ ├─ ghc prim ┄┄ │ └─ integer gmp ┄┄ └─ cabal test ├─ base ┄┄ ├─ bytestring ┄┄ ├─ say │ ├─ base ┄┄ │ ├─ bytestring ┄┄ │ ├─ text │ │ ├─ array ┄┄ │ │ ├─ base ┄┄ │ │ ├─ binary │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ bytestring ┄┄ │ │ │ └─ containers │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ deepseq ┄┄ │ │ │ └─ ghc prim ┄┄ │ │ ├─ bytestring ┄┄ │ │ ├─ deepseq ┄┄ │ │ ├─ ghc prim ┄┄ │ │ └─ integer gmp ┄┄ │ └─ transformers │ └─ base ┄┄ └─ text ┄┄ top sorted unitid rts unitid ghc prim unitid integer gmp unitid base unitid array unitid deepseq unitid bytestring unitid containers unitid binary unitid text unitid transformers unitid say unitid cabal test inplace unitid cabal test inplace cabal test direct deps pkgid pkgname cabal test ver compnamelib base bytestring say text pkgid pkgname cabal test ver compnameexe cabal test base bytestring cabal test cabal plan info builddir tree cabal test ├─ base │ ├─ ghc prim │ │ └─ rts │ ├─ integer gmp │ │ └─ ghc prim ┄┄ │ └─ rts ┄┄ ├─ bytestring │ ├─ base ┄┄ │ ├─ deepseq │ │ ├─ array │ │ │ └─ base ┄┄ │ │ └─ base ┄┄ │ ├─ ghc prim ┄┄ │ └─ integer gmp ┄┄ └─ cabal test ├─ base ┄┄ ├─ bytestring ┄┄ ├─ say │ ├─ base ┄┄ │ ├─ bytestring ┄┄ │ ├─ text │ │ ├─ array ┄┄ │ │ ├─ base ┄┄ │ │ ├─ binary │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ bytestring ┄┄ │ │ │ └─ containers │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ deepseq ┄┄ │ │ │ └─ ghc prim ┄┄ │ │ ├─ bytestring ┄┄ │ │ ├─ deepseq ┄┄ │ │ ├─ ghc prim ┄┄ │ │ └─ integer gmp ┄┄ │ └─ transformers │ └─ base ┄┄ └─ text ┄┄ top sorted unitid rts unitid ghc prim unitid integer gmp unitid base unitid array unitid deepseq unitid bytestring unitid containers unitid binary unitid text unitid transformers unitid say unitid cabal test inplace unitid cabal test inplace cabal test direct deps pkgid pkgname cabal test ver compnamelib base bytestring say text pkgid pkgname cabal test ver compnameexe cabal test base bytestring cabal test cabal plan info builddir cabal install tree cabal test ├─ base │ ├─ ghc prim │ │ └─ rts │ ├─ integer gmp │ │ └─ ghc prim ┄┄ │ └─ rts ┄┄ ├─ bytestring │ ├─ base ┄┄ │ ├─ deepseq │ │ ├─ array │ │ │ └─ base ┄┄ │ │ └─ base ┄┄ │ ├─ ghc prim ┄┄ │ └─ integer gmp ┄┄ └─ cabal test ├─ base ┄┄ ├─ bytestring ┄┄ ├─ say │ ├─ base ┄┄ │ ├─ bytestring ┄┄ │ ├─ text │ │ ├─ array ┄┄ │ │ ├─ base ┄┄ │ │ ├─ binary │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ bytestring ┄┄ │ │ │ └─ containers │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ deepseq ┄┄ │ │ │ └─ ghc prim ┄┄ │ │ ├─ bytestring ┄┄ │ │ ├─ deepseq ┄┄ │ │ ├─ ghc prim ┄┄ │ │ └─ integer gmp ┄┄ │ └─ transformers │ └─ base ┄┄ └─ text ┄┄ top sorted unitid rts unitid ghc prim unitid integer gmp unitid base unitid array unitid deepseq unitid bytestring unitid containers unitid binary unitid text unitid transformers unitid say unitid cabal test unitid cabal test direct deps summary the unit id for say from the temp build dir is different from the other two both are installed in the store dir cabal plan diff does not output anything and afaik it should at least in the first case cabal plan diff builddir dist newstyle builddir cabal install cabal plan diff builddir dist newstyle builddir expected behavior cabal install should only rebuild local packages like in linux system information operating system windows and cabal but is reproduced for sure with and and afair with any cabal version i tried since ghc versions and but probably with any additional context i decide to open a new issue as is about another different issue and talks about rebuild the local packages in any os see another reproduction case with haskell language server in fgaz gbaz i hope there is enough info to at least demonstrate the behaviour let me know if more info or tests are needed | 1 |
787,687 | 27,727,039,493 | IssuesEvent | 2023-03-15 03:41:14 | jkachhad/Stormhalter | https://api.github.com/repos/jkachhad/Stormhalter | closed | Creature List Movement | Enhancement High Priority UX | Creature list becomes unusable when there are more than a page or two of creatures on screen. It jumps around and reorders multiple times a round even if nothing moves on the screen. [Reported by Zathras] | 1.0 | Creature List Movement - Creature list becomes unusable when there are more than a page or two of creatures on screen. It jumps around and reorders multiple times a round even if nothing moves on the screen. [Reported by Zathras] | priority | creature list movement creature list becomes unusable when there are more than a page or two of creatures on screen it jumps around and reorders multiple times a round even if nothing moves on the screen | 1 |
388,933 | 11,495,240,690 | IssuesEvent | 2020-02-12 04:07:28 | wso2/product-is | https://api.github.com/repos/wso2/product-is | opened | Handling Last login time as identity claim | Affected/5.10.0-beta Complexity/Low Component/Identity Mgt Component/Migration Priority/High Type/Docs | With the fix \[1] we are handling the Last login time(which claim is used in account suspension) as identity claim in identity server. Before the fix we have handled the last login claim as default claim. Since this is a behavioural change we have introduced a new property to disable this behaviour.
If you want to get the previous behaviour please add the following property in deployment.toml file
```
[account.suspension]
lastlogin_as_default_claim=true
```
\[1] https://github.com/wso2-extensions/identity-governance/pull/358 | 1.0 | Handling Last login time as identity claim - With the fix \[1] we are handling the Last login time(which claim is used in account suspension) as identity claim in identity server. Before the fix we have handled the last login claim as default claim. Since this is a behavioural change we have introduced a new property to disable this behaviour.
If you want to get the previous behaviour please add the following property in deployment.toml file
```
[account.suspension]
lastlogin_as_default_claim=true
```
\[1] https://github.com/wso2-extensions/identity-governance/pull/358 | priority | handling last login time as identity claim with the fix we are handling the last login time which claim is used in account suspension as identity claim in identity server before the fix we have handled the last login claim as default claim since this is a behavioural change we have introduced a new property to disable this behaviour if you want to get the previous behaviour please add the following property in deployment toml file lastlogin as default claim true | 1 |
94,373 | 3,925,006,801 | IssuesEvent | 2016-04-22 17:16:23 | ualbertalib/HydraNorth | https://api.github.com/repos/ualbertalib/HydraNorth | opened | Show release version in public interface | priority:high size:small | To make it easier for us to know what features are supposed to be in the current public interface, let's include the release tag or commit hash, either in the footer or in a comment. | 1.0 | Show release version in public interface - To make it easier for us to know what features are supposed to be in the current public interface, let's include the release tag or commit hash, either in the footer or in a comment. | priority | show release version in public interface to make it easier for us to know what features are supposed to be in the current public interface let s include the release tag or commit hash either in the footer or in a comment | 1 |
402,785 | 11,824,837,491 | IssuesEvent | 2020-03-21 09:10:45 | bryntum/support | https://api.github.com/repos/bryntum/support | opened | Should round percentDone value for tasks in task editor | bug high-priority | <img width="648" alt="Screenshot 2020-03-21 at 10 09 59" src="https://user-images.githubusercontent.com/218570/77223368-28915d80-6b5c-11ea-9239-f2972fa54a9f.png">
| 1.0 | Should round percentDone value for tasks in task editor - <img width="648" alt="Screenshot 2020-03-21 at 10 09 59" src="https://user-images.githubusercontent.com/218570/77223368-28915d80-6b5c-11ea-9239-f2972fa54a9f.png">
| priority | should round percentdone value for tasks in task editor img width alt screenshot at src | 1 |
422,414 | 12,278,024,805 | IssuesEvent | 2020-05-08 09:07:31 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | opened | Contents are truncated in deployment.toml by running secure vault | Affected/3.0.0 Priority/High Type/Bug WUM | ### Description:
Deployment.toml configurations - all after [secret] section are loosing by running secure vault.
### Steps to reproduce:
1. Configure secret aliases in a middle position of the toml file. (Ex:- next to the keystore configs)
2. Run ciphertool.sh -Dconfigure
3. This will remove all the contents after secret alias section which we newly configured in the toml file.
### Affected Product Version:
3.0.0 , wum - 1586442818040
### Environment details (with versions):
- OS: MacOS
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members--> | 1.0 | Contents are truncated in deployment.toml by running secure vault - ### Description:
Deployment.toml configurations - all after [secret] section are loosing by running secure vault.
### Steps to reproduce:
1. Configure secret aliases in a middle position of the toml file. (Ex:- next to the keystore configs)
2. Run ciphertool.sh -Dconfigure
3. This will remove all the contents after secret alias section which we newly configured in the toml file.
### Affected Product Version:
3.0.0 , wum - 1586442818040
### Environment details (with versions):
- OS: MacOS
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members--> | priority | contents are truncated in deployment toml by running secure vault description deployment toml configurations all after section are loosing by running secure vault steps to reproduce configure secret aliases in a middle position of the toml file ex next to the keystore configs run ciphertool sh dconfigure this will remove all the contents after secret alias section which we newly configured in the toml file affected product version wum environment details with versions os macos optional fields related issues suggested labels suggested assignees | 1 |
513,938 | 14,928,695,764 | IssuesEvent | 2021-01-24 20:14:33 | bounswe/bounswe2020group4 | https://api.github.com/repos/bounswe/bounswe2020group4 | closed | Admin : Ban Vendor & Customer , Unban Vendor & Customer | Backend Coding Effort: Medium Priority: High Status: Needs Review | Deadline: 24.01.2021
Admin should be able to ban vendor and customer.
If it is necessary , admin should be able to unban them again. | 1.0 | Admin : Ban Vendor & Customer , Unban Vendor & Customer - Deadline: 24.01.2021
Admin should be able to ban vendor and customer.
If it is necessary , admin should be able to unban them again. | priority | admin ban vendor customer unban vendor customer deadline admin should be able to ban vendor and customer if it is necessary admin should be able to unban them again | 1 |
269,135 | 8,432,376,450 | IssuesEvent | 2018-10-17 01:42:40 | LakeEffectRobotics/LakeEffectScoutingApp | https://api.github.com/repos/LakeEffectRobotics/LakeEffectScoutingApp | closed | Add packet splitting fix to server receiving data as well | HIGH PRIORITY | It's probably causing the weird data issues where only half of the data gets sent, then another half, then all of it. | 1.0 | Add packet splitting fix to server receiving data as well - It's probably causing the weird data issues where only half of the data gets sent, then another half, then all of it. | priority | add packet splitting fix to server receiving data as well it s probably causing the weird data issues where only half of the data gets sent then another half then all of it | 1 |
622,093 | 19,606,730,335 | IssuesEvent | 2022-01-06 10:25:07 | wso2/ballerina-plugin-vscode | https://api.github.com/repos/wso2/ballerina-plugin-vscode | opened | Tryit Code Action for the services are shown even when there are syntax errors | bug Type/Bug Priority/High | **Description:**
Consider the following capture,
<img width="620" alt="Screenshot 2022-01-06 at 15 51 03" src="https://user-images.githubusercontent.com/1329674/148368168-054915de-6a0e-478c-b226-5e41e791f108.png">
Here the source is having a syntax error and even in this case, we show the try it code action. We should not encourage the user/ navigate the user to such actions if the result of the action is ending in an error scenario.
Shall we allow this only in scenarios where semantic/ syntax errors are not available
| 1.0 | Tryit Code Action for the services are shown even when there are syntax errors - **Description:**
Consider the following capture,
<img width="620" alt="Screenshot 2022-01-06 at 15 51 03" src="https://user-images.githubusercontent.com/1329674/148368168-054915de-6a0e-478c-b226-5e41e791f108.png">
Here the source is having a syntax error and even in this case, we show the try it code action. We should not encourage the user/ navigate the user to such actions if the result of the action is ending in an error scenario.
Shall we allow this only in scenarios where semantic/ syntax errors are not available
| priority | tryit code action for the services are shown even when there are syntax errors description consider the following capture img width alt screenshot at src here the source is having a syntax error and even in this case we show the try it code action we should not encourage the user navigate the user to such actions if the result of the action is ending in an error scenario shall we allow this only in scenarios where semantic syntax errors are not available | 1 |
637,622 | 20,673,407,997 | IssuesEvent | 2022-03-10 06:22:00 | yukiHaga/regex-hunting | https://api.github.com/repos/yukiHaga/regex-hunting | opened | ランクアップした時に経験値を繰り越しさせる | Priority: high | ## 概要
ランクアップした時に経験値を繰り越しさせる
繰り越しされてないから、トータルの経験値と合わない部分がある。
## やること
- [ ] ランクアップした時に経験値を繰り越しさせる。
## 受け入れ条件
- [ ] ランクアップした時に、経験値が反映されている。
## 参考記事
特になし。 | 1.0 | ランクアップした時に経験値を繰り越しさせる - ## 概要
ランクアップした時に経験値を繰り越しさせる
繰り越しされてないから、トータルの経験値と合わない部分がある。
## やること
- [ ] ランクアップした時に経験値を繰り越しさせる。
## 受け入れ条件
- [ ] ランクアップした時に、経験値が反映されている。
## 参考記事
特になし。 | priority | ランクアップした時に経験値を繰り越しさせる 概要 ランクアップした時に経験値を繰り越しさせる 繰り越しされてないから、トータルの経験値と合わない部分がある。 やること ランクアップした時に経験値を繰り越しさせる。 受け入れ条件 ランクアップした時に、経験値が反映されている。 参考記事 特になし。 | 1 |
258,813 | 8,179,918,391 | IssuesEvent | 2018-08-28 17:50:22 | workcraft/workcraft | https://api.github.com/repos/workcraft/workcraft | closed | Recognise MUTEX elements on Verilog import | enhancement priority:high status:confirmed tag:model:circuit | Improve Verilog import, so it recognises MUTEX elements. The definition of MUTEX is already in the _Digital Circuit -> Mutex name and request-grant pair_ global property. | 1.0 | Recognise MUTEX elements on Verilog import - Improve Verilog import, so it recognises MUTEX elements. The definition of MUTEX is already in the _Digital Circuit -> Mutex name and request-grant pair_ global property. | priority | recognise mutex elements on verilog import improve verilog import so it recognises mutex elements the definition of mutex is already in the digital circuit mutex name and request grant pair global property | 1 |
358,156 | 10,618,494,502 | IssuesEvent | 2019-10-13 05:06:12 | AY1920S1-CS2103T-F13-2/main | https://api.github.com/repos/AY1920S1-CS2103T-F13-2/main | closed | As a user, I would like to sort items based on ascending or descending fields | priority.High type.Story | So that I can have more flexibility in what items I want to view at the top. | 1.0 | As a user, I would like to sort items based on ascending or descending fields - So that I can have more flexibility in what items I want to view at the top. | priority | as a user i would like to sort items based on ascending or descending fields so that i can have more flexibility in what items i want to view at the top | 1 |
104,792 | 4,221,241,320 | IssuesEvent | 2016-07-01 03:56:35 | fossasia/open-event-orga-server | https://api.github.com/repos/fossasia/open-event-orga-server | closed | Don't show trashed items in API | Priority: High Rest-API | Parent issue : #1220
Trash has been implemented using a bool column in each table (in_trash) which is True if the item has been deleted. Currently API makes no distinction between trashed and active items.
As trash is only for admins and not for the end-user, the API should not return trashed items and treat them as deleted. | 1.0 | Don't show trashed items in API - Parent issue : #1220
Trash has been implemented using a bool column in each table (in_trash) which is True if the item has been deleted. Currently API makes no distinction between trashed and active items.
As trash is only for admins and not for the end-user, the API should not return trashed items and treat them as deleted. | priority | don t show trashed items in api parent issue trash has been implemented using a bool column in each table in trash which is true if the item has been deleted currently api makes no distinction between trashed and active items as trash is only for admins and not for the end user the api should not return trashed items and treat them as deleted | 1 |
264,521 | 8,316,105,312 | IssuesEvent | 2018-09-25 08:03:41 | jsheroes/jsheroes.io | https://api.github.com/repos/jsheroes/jsheroes.io | closed | Cookie popup not closing on iPhone 8 | bug high-priority | This bug was reported recently but I can't seem to reproduce it on iPhone SE (was thinking it's a problem with Safari)
We should check this out asap. | 1.0 | Cookie popup not closing on iPhone 8 - This bug was reported recently but I can't seem to reproduce it on iPhone SE (was thinking it's a problem with Safari)
We should check this out asap. | priority | cookie popup not closing on iphone this bug was reported recently but i can t seem to reproduce it on iphone se was thinking it s a problem with safari we should check this out asap | 1 |
473,207 | 13,638,656,604 | IssuesEvent | 2020-09-25 09:42:33 | status-im/status-react | https://api.github.com/repos/status-im/status-react | closed | Add aUSDC token and optimize search for Aave (LEND) (aka Ethlend) | high-priority | # Problem
See [comment](https://github.com/status-im/status-react/issues/11185#issuecomment-697948911) by @j-zerah. We have an event coming up with AAVE. Given the size of their community being introduced to Status we want to make sure they can find their respective tokens.
- [aUSDC ](https://etherscan.io/token/0x9bA00D6856a4eDF4665BcA2C2309936572473B7E)(they'll issue this) https://www.coingecko.com/en/coins/aave-usdc
- [AAVE LEND](https://etherscan.io/token/0x80fB784B7eD66730e8b1DBd9820aFD29931aab03). EthLend is currently available, but can not be found when searching for AAVE (nice to have if this is a quick change)
## Implementation
## Acceptance Criteria
- Can select aUSDC in manage assets
- Can find LEND when searching for AAVE (nice to have)
## Notes
## Future Steps
| 1.0 | Add aUSDC token and optimize search for Aave (LEND) (aka Ethlend) - # Problem
See [comment](https://github.com/status-im/status-react/issues/11185#issuecomment-697948911) by @j-zerah. We have an event coming up with AAVE. Given the size of their community being introduced to Status we want to make sure they can find their respective tokens.
- [aUSDC ](https://etherscan.io/token/0x9bA00D6856a4eDF4665BcA2C2309936572473B7E)(they'll issue this) https://www.coingecko.com/en/coins/aave-usdc
- [AAVE LEND](https://etherscan.io/token/0x80fB784B7eD66730e8b1DBd9820aFD29931aab03). EthLend is currently available, but can not be found when searching for AAVE (nice to have if this is a quick change)
## Implementation
## Acceptance Criteria
- Can select aUSDC in manage assets
- Can find LEND when searching for AAVE (nice to have)
## Notes
## Future Steps
| priority | add ausdc token and optimize search for aave lend aka ethlend problem see by j zerah we have an event coming up with aave given the size of their community being introduced to status we want to make sure they can find their respective tokens issue this ethlend is currently available but can not be found when searching for aave nice to have if this is a quick change implementation acceptance criteria can select ausdc in manage assets can find lend when searching for aave nice to have notes future steps | 1 |
769,924 | 27,022,568,353 | IssuesEvent | 2023-02-11 06:53:01 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | MacOS M1 CPU/MPS tests are backlogged due to HW availability | high priority triaged ci: sev | ## Current Status
*ongoing*
## Error looks like
Long queue times for M1 runners, which results in longer than usual merges/timeouts
## Incident timeline (all times pacific)
*Include when the incident began, when it was detected, mitigated, root caused, and finally closed.*
<img width="652" alt="image" src="https://user-images.githubusercontent.com/2453524/218155131-48628732-6574-4081-b56b-a3a5a2e40979.png">
## User impact
*How does this affect users of PyTorch CI?*
## Root cause
Looks like we only have 16 runners for the entire org.
## Mitigation
*How did we mitigate the issue?*
## Prevention/followups
*How do we prevent issues like this in the future?*
cc @ezyang @gchanan @zou3519 | 1.0 | MacOS M1 CPU/MPS tests are backlogged due to HW availability - ## Current Status
*ongoing*
## Error looks like
Long queue times for M1 runners, which results in longer than usual merges/timeouts
## Incident timeline (all times pacific)
*Include when the incident began, when it was detected, mitigated, root caused, and finally closed.*
<img width="652" alt="image" src="https://user-images.githubusercontent.com/2453524/218155131-48628732-6574-4081-b56b-a3a5a2e40979.png">
## User impact
*How does this affect users of PyTorch CI?*
## Root cause
Looks like we only have 16 runners for the entire org.
## Mitigation
*How did we mitigate the issue?*
## Prevention/followups
*How do we prevent issues like this in the future?*
cc @ezyang @gchanan @zou3519 | priority | macos cpu mps tests are backlogged due to hw availability current status ongoing error looks like long queue times for runners which results in longer than usual merges timeouts incident timeline all times pacific include when the incident began when it was detected mitigated root caused and finally closed img width alt image src user impact how does this affect users of pytorch ci root cause looks like we only have runners for the entire org mitigation how did we mitigate the issue prevention followups how do we prevent issues like this in the future cc ezyang gchanan | 1 |
283,574 | 8,719,960,669 | IssuesEvent | 2018-12-08 06:43:47 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | make install / make package links for osmesa to libGL | bug likelihood medium priority reviewed severity high | To make the dynamic lib switch work.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1549
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: make install / make package links for osmesa to libGL
Assigned to: Eric Brugger
Category:
Target version: 2.7
Author: Cyrus Harrison
Start: 07/30/2013
Due date:
% Done: 0
Estimated time:
Created: 07/30/2013 11:51 am
Updated: 08/30/2013 02:25 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: trunk
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
To make the dynamic lib switch work.
Comments:
Hi Everyone,For Linux make install / make package with osmesa, I added the proper symlinks that allow us to use OSMesa as libGL at runtime.Trunk commit:edge84@cyrush:svn commit -m "linux / osmesa: install symlinks so osmesa can be used as libGL dynamically" FindVisItMesa.cmakeSending FindVisItMesa.cmakeTransmitting file data .Committed revision r21734.It took quite a while b/c :INSTALL(CODE....using: \$ENV{DESTDIR}\${CMAKE_INSTALL_PREFIX}, not just ${CMAKE_INSTALL_PREFIX})-Cyrus
| 1.0 | make install / make package links for osmesa to libGL - To make the dynamic lib switch work.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 1549
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: make install / make package links for osmesa to libGL
Assigned to: Eric Brugger
Category:
Target version: 2.7
Author: Cyrus Harrison
Start: 07/30/2013
Due date:
% Done: 0
Estimated time:
Created: 07/30/2013 11:51 am
Updated: 08/30/2013 02:25 pm
Likelihood: 3 - Occasional
Severity: 5 - Very Serious
Found in version: trunk
Impact:
Expected Use:
OS: All
Support Group: Any
Description:
To make the dynamic lib switch work.
Comments:
Hi Everyone,For Linux make install / make package with osmesa, I added the proper symlinks that allow us to use OSMesa as libGL at runtime.Trunk commit:edge84@cyrush:svn commit -m "linux / osmesa: install symlinks so osmesa can be used as libGL dynamically" FindVisItMesa.cmakeSending FindVisItMesa.cmakeTransmitting file data .Committed revision r21734.It took quite a while b/c :INSTALL(CODE....using: \$ENV{DESTDIR}\${CMAKE_INSTALL_PREFIX}, not just ${CMAKE_INSTALL_PREFIX})-Cyrus
| priority | make install make package links for osmesa to libgl to make the dynamic lib switch work redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject make install make package links for osmesa to libgl assigned to eric brugger category target version author cyrus harrison start due date done estimated time created am updated pm likelihood occasional severity very serious found in version trunk impact expected use os all support group any description to make the dynamic lib switch work comments hi everyone for linux make install make package with osmesa i added the proper symlinks that allow us to use osmesa as libgl at runtime trunk commit cyrush svn commit m linux osmesa install symlinks so osmesa can be used as libgl dynamically findvisitmesa cmakesending findvisitmesa cmaketransmitting file data committed revision it took quite a while b c install code using env destdir cmake install prefix not just cmake install prefix cyrus | 1 |
197,160 | 6,952,990,273 | IssuesEvent | 2017-12-06 19:23:59 | Kademi/kademi-dev | https://api.github.com/repos/Kademi/kademi-dev | closed | Email sending failed | bug High priority | http://vlad32-test.admin.kademi-ci.co/emails/?q=&status=&job=
With error : No MimeMessage found
I've post to real and mailinator addresses - sit are the same. | 1.0 | Email sending failed - http://vlad32-test.admin.kademi-ci.co/emails/?q=&status=&job=
With error : No MimeMessage found
I've post to real and mailinator addresses - sit are the same. | priority | email sending failed with error no mimemessage found i ve post to real and mailinator addresses sit are the same | 1 |
220,878 | 7,371,800,126 | IssuesEvent | 2018-03-13 13:01:19 | OpenSRP/opensrp-server | https://api.github.com/repos/OpenSRP/opensrp-server | closed | Create /add Shipment end-point in OpenSRP server | High Priority | 1. NiFi sends a POST request to the /add shipment end-point on OpenSRP server with a single Shipment object as defined in the docs linked below
2. The server attempts to create a shipment object and stores it in couchDB
3. The server returns a 201 status if successful and 500 status if it fails
More at: https://openlmis.atlassian.net/wiki/spaces/OP/pages/237666309/Proof+of+Concept+Create+Nifi+API+Interactions | 1.0 | Create /add Shipment end-point in OpenSRP server - 1. NiFi sends a POST request to the /add shipment end-point on OpenSRP server with a single Shipment object as defined in the docs linked below
2. The server attempts to create a shipment object and stores it in couchDB
3. The server returns a 201 status if successful and 500 status if it fails
More at: https://openlmis.atlassian.net/wiki/spaces/OP/pages/237666309/Proof+of+Concept+Create+Nifi+API+Interactions | priority | create add shipment end point in opensrp server nifi sends a post request to the add shipment end point on opensrp server with a single shipment object as defined in the docs linked below the server attempts to create a shipment object and stores it in couchdb the server returns a status if successful and status if it fails more at | 1 |
52,761 | 3,028,942,695 | IssuesEvent | 2015-08-04 09:26:06 | pombase/pombase-chado | https://api.github.com/repos/pombase/pombase-chado | closed | Use FYPO from Git/GitHub in Chado load | high priority next-load | The Chado load scripts are still pointing at the subversion repo on SourceForge | 1.0 | Use FYPO from Git/GitHub in Chado load - The Chado load scripts are still pointing at the subversion repo on SourceForge | priority | use fypo from git github in chado load the chado load scripts are still pointing at the subversion repo on sourceforge | 1 |
357,698 | 10,616,948,544 | IssuesEvent | 2019-10-12 15:33:08 | bounswe/bounswe2019group8 | https://api.github.com/repos/bounswe/bounswe2019group8 | opened | Auto-deploy backend app using Jenkins in a Docker container | Backend Effort: High Priority: Low Type: Enhancement | **Actions:**
1. When a push is made to a specific branch, GitHub webhook will trigger the Jenkins pipeline.
1. Jenkins will clone the repo, build image from Dockerfile, push the image to the Docker Hub and run the the app in a container from the pushed image.
**Deadline:** 19.10.2019 - 23.59
| 1.0 | Auto-deploy backend app using Jenkins in a Docker container - **Actions:**
1. When a push is made to a specific branch, GitHub webhook will trigger the Jenkins pipeline.
1. Jenkins will clone the repo, build image from Dockerfile, push the image to the Docker Hub and run the the app in a container from the pushed image.
**Deadline:** 19.10.2019 - 23.59
| priority | auto deploy backend app using jenkins in a docker container actions when a push is made to a specific branch github webhook will trigger the jenkins pipeline jenkins will clone the repo build image from dockerfile push the image to the docker hub and run the the app in a container from the pushed image deadline | 1 |
131,857 | 5,166,425,158 | IssuesEvent | 2017-01-17 16:12:11 | snaiperskaya96/test-import-repo | https://api.github.com/repos/snaiperskaya96/test-import-repo | opened | Add 'primary colour' to brand settings page | Accepted Enhancement High Priority | https://trello.com/c/gK8xRfFF/279-add-primary-colour-to-brand-settings-page
Move brand specific CSS into the HTML skeleton view. | 1.0 | Add 'primary colour' to brand settings page - https://trello.com/c/gK8xRfFF/279-add-primary-colour-to-brand-settings-page
Move brand specific CSS into the HTML skeleton view. | priority | add primary colour to brand settings page move brand specific css into the html skeleton view | 1 |
496,498 | 14,348,697,796 | IssuesEvent | 2020-11-29 13:36:18 | xournalpp/xournalpp | https://api.github.com/repos/xournalpp/xournalpp | closed | Zooming looses the spot in the document | bug confirmed priority::high | **Affects versions :**
- Xubuntu 18.04
- Desktop: Xfce
- version of libgtk 3.22.30
- Version of Xournal++: 1.0.19
- Installation method: PPA
**Describe the bug**
Changing from 100% zoom to fit page zoom, jumps the view around in the document.
**To Reproduce**
Steps to reproduce the behaviour:
Open a multi page document.
Do a zoom to 100%, scroll a few pages.
Do a zoom to fit.
The view of the document jumps back to where you started scrolling.
**Expected behavior**
doing the above should show the same bit of document just in different zoom depths.
| 1.0 | Zooming looses the spot in the document - **Affects versions :**
- Xubuntu 18.04
- Desktop: Xfce
- version of libgtk 3.22.30
- Version of Xournal++: 1.0.19
- Installation method: PPA
**Describe the bug**
Changing from 100% zoom to fit page zoom, jumps the view around in the document.
**To Reproduce**
Steps to reproduce the behaviour:
Open a multi page document.
Do a zoom to 100%, scroll a few pages.
Do a zoom to fit.
The view of the document jumps back to where you started scrolling.
**Expected behavior**
doing the above should show the same bit of document just in different zoom depths.
| priority | zooming looses the spot in the document affects versions xubuntu desktop xfce version of libgtk version of xournal installation method ppa describe the bug changing from zoom to fit page zoom jumps the view around in the document to reproduce steps to reproduce the behaviour open a multi page document do a zoom to scroll a few pages do a zoom to fit the view of the document jumps back to where you started scrolling expected behavior doing the above should show the same bit of document just in different zoom depths | 1 |
748,414 | 26,122,093,773 | IssuesEvent | 2022-12-28 13:51:22 | abpframework/abp | https://api.github.com/repos/abpframework/abp | closed | Angular - Settings are not visible in the menu. | bug priority:high ui-angular | **App:** Angular `7.0.0-rc.5`
Settings are not visible in the menu, although the settings authority is available.

| 1.0 | Angular - Settings are not visible in the menu. - **App:** Angular `7.0.0-rc.5`
Settings are not visible in the menu, although the settings authority is available.

| priority | angular settings are not visible in the menu app angular rc settings are not visible in the menu although the settings authority is available | 1 |
807,258 | 29,985,661,485 | IssuesEvent | 2023-06-25 07:39:17 | Unity-Technologies/com.unity.netcode.gameobjects | https://api.github.com/repos/Unity-Technologies/com.unity.netcode.gameobjects | closed | Shutdown on client side cause an error | type:bug stat:awaiting triage priority:high stat:imported | ### Description
A clear and concise description of what the bug is.
### Reproduce Steps
1. When a client has been connected
2. Call NetworkManager.Shutdown on Client side.
3. Will cause an error blew:
4.
NullReferenceException: Object reference not set to an instance of an object
Unity.Netcode.NetworkManager.Shutdown (System.Boolean discardMessageQueue) (at Library/PackageCache/com.unity.netcode.gameobjects@1.5.1/Runtime/Core/NetworkManager.cs:945)
### Environment
- OS: Win11
- Unity Version: 2022.3.19f1c1
- Netcode Version: 1.51
- Netcode Commit: [e.g. https://github.com/Unity-Technologies/com.unity.netcode.gameobjects/commit/ba418fa5b600ad9eb61fab0575f12fbecc2c6520]
### Additional Context
Add any other context about the problem here. Logs, code snippets would be useful here but please also consider attaching a minimal Unity project that reproduces the issue. | 1.0 | Shutdown on client side cause an error - ### Description
A clear and concise description of what the bug is.
### Reproduce Steps
1. When a client has been connected
2. Call NetworkManager.Shutdown on Client side.
3. Will cause an error blew:
4.
NullReferenceException: Object reference not set to an instance of an object
Unity.Netcode.NetworkManager.Shutdown (System.Boolean discardMessageQueue) (at Library/PackageCache/com.unity.netcode.gameobjects@1.5.1/Runtime/Core/NetworkManager.cs:945)
### Environment
- OS: Win11
- Unity Version: 2022.3.19f1c1
- Netcode Version: 1.51
- Netcode Commit: [e.g. https://github.com/Unity-Technologies/com.unity.netcode.gameobjects/commit/ba418fa5b600ad9eb61fab0575f12fbecc2c6520]
### Additional Context
Add any other context about the problem here. Logs, code snippets would be useful here but please also consider attaching a minimal Unity project that reproduces the issue. | priority | shutdown on client side cause an error description a clear and concise description of what the bug is reproduce steps when a client has been connected call networkmanager shutdown on client side will cause an error blew nullreferenceexception object reference not set to an instance of an object unity netcode networkmanager shutdown system boolean discardmessagequeue at library packagecache com unity netcode gameobjects runtime core networkmanager cs environment os unity version netcode version netcode commit additional context add any other context about the problem here logs code snippets would be useful here but please also consider attaching a minimal unity project that reproduces the issue | 1 |
292,552 | 8,959,510,827 | IssuesEvent | 2019-01-27 23:10:34 | aa-software2112/SOEN390_SimpleCamera | https://api.github.com/repos/aa-software2112/SOEN390_SimpleCamera | closed | Setting up Travis CI | Setup CI priority: high risk: high story | As a developer, I would like to set up Travis CI for continuous integration.
- [x] Integrating Travis CI to the repo #14 [1 hour] @samgo1
- [x] Write an initial passing script #15 [1 hour] @samgo1
- [x] Choose and integrate a testing framework #11 [2 hours] @KrishnaPatel1 @Lussier115
- [x] Write 3 unit tests to ensure it is functioning in the Travis build #12 [2 hours] @aa-software2112 @lkadian @anthonyle17
- [x] Choose and integrate an emulator for integration and acceptance testing in Travis #13 [2 hours] @aa-software2112 @Lussier115
- [x] Remove "clean" and "release" build within Travis @KrishnaPatel1
Please refer to the Wiki for more information on Testing.
[Story points: 3]
[Ideal hours: 8] | 1.0 | Setting up Travis CI - As a developer, I would like to set up Travis CI for continuous integration.
- [x] Integrating Travis CI to the repo #14 [1 hour] @samgo1
- [x] Write an initial passing script #15 [1 hour] @samgo1
- [x] Choose and integrate a testing framework #11 [2 hours] @KrishnaPatel1 @Lussier115
- [x] Write 3 unit tests to ensure it is functioning in the Travis build #12 [2 hours] @aa-software2112 @lkadian @anthonyle17
- [x] Choose and integrate an emulator for integration and acceptance testing in Travis #13 [2 hours] @aa-software2112 @Lussier115
- [x] Remove "clean" and "release" build within Travis @KrishnaPatel1
Please refer to the Wiki for more information on Testing.
[Story points: 3]
[Ideal hours: 8] | priority | setting up travis ci as a developer i would like to set up travis ci for continuous integration integrating travis ci to the repo write an initial passing script choose and integrate a testing framework write unit tests to ensure it is functioning in the travis build aa lkadian choose and integrate an emulator for integration and acceptance testing in travis aa remove clean and release build within travis please refer to the wiki for more information on testing | 1 |
320,037 | 9,763,956,167 | IssuesEvent | 2019-06-05 14:49:57 | wherebyus/general-tasks | https://api.github.com/repos/wherebyus/general-tasks | closed | My Account: "Subscriptions" should be called "My Membership" | Feasibility: High Priority: Low Product: Membership Product: Site checked | 
## DOD
- [x] "Subscriptions" in the My Account left navigation is now called "My Membership"
- [ ] The URL endpoint is now called `my-membership` (Note this may be dependent on #486)
- [ ] The "Subscription" table column header is now called "ID"
- [ ] Any "Subscription" references in the order receipts are references to Memberships instead | 1.0 | My Account: "Subscriptions" should be called "My Membership" - 
## DOD
- [x] "Subscriptions" in the My Account left navigation is now called "My Membership"
- [ ] The URL endpoint is now called `my-membership` (Note this may be dependent on #486)
- [ ] The "Subscription" table column header is now called "ID"
- [ ] Any "Subscription" references in the order receipts are references to Memberships instead | priority | my account subscriptions should be called my membership dod subscriptions in the my account left navigation is now called my membership the url endpoint is now called my membership note this may be dependent on the subscription table column header is now called id any subscription references in the order receipts are references to memberships instead | 1 |
789,576 | 27,796,541,309 | IssuesEvent | 2023-03-17 12:59:04 | bryntum/support | https://api.github.com/repos/bryntum/support | closed | Uncaught error in populateEventMenu when dropping an event on the first resource | bug example resolved high-priority premium forum large-account OEM | [Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=23631&p=117002#p117002)
"Hello,
When I drop an event on the first resource of the scheduler, the event plays a "bouncing" animation, and the event context menu fails to open on right click.
The issue cannot be reproduced when dropping an event on other resources than the first one.

The attached video demonstrates that:
Drag&drop of event 1 onto resource B -> working context menu on event 1
Drag&drop of event 2 onto resource A -> populateEventMenu fails on event 2
I reproduced this issue on the inline data demo, using the following code and data:
```
import { SchedulerPro } from '../../build/schedulerpro.module.js?465031';
import shared from '../_shared/shared.module.js?465031';
export const data = {
"eventsData": [
{
"id": 1,
"name": "1",
"startDate": "2022-10-02T09:12:00",
"endDate": "2022-10-02T12:48:00",
},
{
"id": 2,
"name": "2",
"startDate": "2022-10-02T09:48:00",
"endDate": "2022-10-02T16:24:00",
}
],
"resourcesData": [
{
"id": 0,
"name": "A",
"image": false
},
{
"id": 1,
"name": "B",
"image": false
}
],
"assignmentsData": [
{
"id": 0,
"resourceId": 0,
"eventId": 1
},
{
"id": 1,
"resourceId": 1,
"eventId": 2
}
],
};
const
schedulerPro = new SchedulerPro({
appendTo : 'container',
startDate : '2022-10-02',
endDate : '2022-10-02',
viewPreset : 'hourAndDay',
forceFit : true,
columns : [
{
type : 'resourceInfo',
text : 'Name',
field : 'name',
showEventCount : true,
width : 150
}
],
resources : data.resourcesData,
events : data.eventsData,
assignments : data.assignmentsData,
dependencies : data.dependenciesData
});
```
I tried disabling the event menu to workaround this error, without success.
Is there an error in my data or scheduler configuration?"
https://user-images.githubusercontent.com/16693227/214033933-d4dd0bd1-33e2-4d9f-893e-6d88e38647ce.mp4
| 1.0 | Uncaught error in populateEventMenu when dropping an event on the first resource - [Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=23631&p=117002#p117002)
"Hello,
When I drop an event on the first resource of the scheduler, the event plays a "bouncing" animation, and the event context menu fails to open on right click.
The issue cannot be reproduced when dropping an event on other resources than the first one.

The attached video demonstrates that:
Drag&drop of event 1 onto resource B -> working context menu on event 1
Drag&drop of event 2 onto resource A -> populateEventMenu fails on event 2
I reproduced this issue on the inline data demo, using the following code and data:
```
import { SchedulerPro } from '../../build/schedulerpro.module.js?465031';
import shared from '../_shared/shared.module.js?465031';
export const data = {
"eventsData": [
{
"id": 1,
"name": "1",
"startDate": "2022-10-02T09:12:00",
"endDate": "2022-10-02T12:48:00",
},
{
"id": 2,
"name": "2",
"startDate": "2022-10-02T09:48:00",
"endDate": "2022-10-02T16:24:00",
}
],
"resourcesData": [
{
"id": 0,
"name": "A",
"image": false
},
{
"id": 1,
"name": "B",
"image": false
}
],
"assignmentsData": [
{
"id": 0,
"resourceId": 0,
"eventId": 1
},
{
"id": 1,
"resourceId": 1,
"eventId": 2
}
],
};
const
schedulerPro = new SchedulerPro({
appendTo : 'container',
startDate : '2022-10-02',
endDate : '2022-10-02',
viewPreset : 'hourAndDay',
forceFit : true,
columns : [
{
type : 'resourceInfo',
text : 'Name',
field : 'name',
showEventCount : true,
width : 150
}
],
resources : data.resourcesData,
events : data.eventsData,
assignments : data.assignmentsData,
dependencies : data.dependenciesData
});
```
I tried disabling the event menu to workaround this error, without success.
Is there an error in my data or scheduler configuration?"
https://user-images.githubusercontent.com/16693227/214033933-d4dd0bd1-33e2-4d9f-893e-6d88e38647ce.mp4
| priority | uncaught error in populateeventmenu when dropping an event on the first resource hello when i drop an event on the first resource of the scheduler the event plays a bouncing animation and the event context menu fails to open on right click the issue cannot be reproduced when dropping an event on other resources than the first one the attached video demonstrates that drag drop of event onto resource b working context menu on event drag drop of event onto resource a populateeventmenu fails on event i reproduced this issue on the inline data demo using the following code and data import schedulerpro from build schedulerpro module js import shared from shared shared module js export const data eventsdata id name startdate enddate id name startdate enddate resourcesdata id name a image false id name b image false assignmentsdata id resourceid eventid id resourceid eventid const schedulerpro new schedulerpro appendto container startdate enddate viewpreset hourandday forcefit true columns type resourceinfo text name field name showeventcount true width resources data resourcesdata events data eventsdata assignments data assignmentsdata dependencies data dependenciesdata i tried disabling the event menu to workaround this error without success is there an error in my data or scheduler configuration | 1 |
623,791 | 19,679,288,427 | IssuesEvent | 2022-01-11 15:20:10 | openghg/openghg | https://api.github.com/repos/openghg/openghg | closed | `search` doesn't return data if ranking data is set for dates that don't overlap data | bug high-priority | I've just encountered this issue
```
ranking datrange = '2015-01-01-00:00:00+00:00_2019-01-01-00:00:00+00:00'
```
And the datasource data keys
```
data_keys = {'2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00'
'2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00'
'2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00'}
```
As it looks for data keys that overlap the ranking data, if the data doesn't overlap then it breaks. Really it should just return the data it finds instead.
| 1.0 | `search` doesn't return data if ranking data is set for dates that don't overlap data - I've just encountered this issue
```
ranking datrange = '2015-01-01-00:00:00+00:00_2019-01-01-00:00:00+00:00'
```
And the datasource data keys
```
data_keys = {'2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00'
'2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00'
'2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00'}
```
As it looks for data keys that overlap the ranking data, if the data doesn't overlap then it breaks. Really it should just return the data it finds instead.
| priority | search doesn t return data if ranking data is set for dates that don t overlap data i ve just encountered this issue ranking datrange and the datasource data keys data keys data uuid data uuid data uuid as it looks for data keys that overlap the ranking data if the data doesn t overlap then it breaks really it should just return the data it finds instead | 1 |
762,171 | 26,710,658,901 | IssuesEvent | 2023-01-27 23:19:16 | huridocs/uwazi | https://api.github.com/repos/huridocs/uwazi | closed | Public forms attachments not working on firefox. | Bug :lady_beetle: Priority: High | **Describe the bug**
On Firefox, attachments or documents dropzone doesn’t open the upload file dialog on click. The dropzone itself works and files can be attached via drag and drop, but when clicking on the zone to open the upload file dialog, it doesn’t.
In Chrome it works fine.
**To Reproduce**
Steps to reproduce the behavior:
1. Open Firefox
2. Go to https://demo.uwazi.io/en/page/hevpal3191/submit-a-document or any public form
3. Try to upload a document by clicking on 'Select files on your device'.
4. Nothing happens.
**Expected behavior**
Same as in Chrome, a file dialog should open.
**Device (please select all that apply)**
- Desktop
**Browser**
- Firefox
**Additional context**
This is on a Linux OS. Maybe for Windows or Mac it's different. On mobile Android it works fine in Firefox.
| 1.0 | Public forms attachments not working on firefox. - **Describe the bug**
On Firefox, attachments or documents dropzone doesn’t open the upload file dialog on click. The dropzone itself works and files can be attached via drag and drop, but when clicking on the zone to open the upload file dialog, it doesn’t.
In Chrome it works fine.
**To Reproduce**
Steps to reproduce the behavior:
1. Open Firefox
2. Go to https://demo.uwazi.io/en/page/hevpal3191/submit-a-document or any public form
3. Try to upload a document by clicking on 'Select files on your device'.
4. Nothing happens.
**Expected behavior**
Same as in Chrome, a file dialog should open.
**Device (please select all that apply)**
- Desktop
**Browser**
- Firefox
**Additional context**
This is on a Linux OS. Maybe for Windows or Mac it's different. On mobile Android it works fine in Firefox.
| priority | public forms attachments not working on firefox describe the bug on firefox attachments or documents dropzone doesn’t open the upload file dialog on click the dropzone itself works and files can be attached via drag and drop but when clicking on the zone to open the upload file dialog it doesn’t in chrome it works fine to reproduce steps to reproduce the behavior open firefox go to or any public form try to upload a document by clicking on select files on your device nothing happens expected behavior same as in chrome a file dialog should open device please select all that apply desktop browser firefox additional context this is on a linux os maybe for windows or mac it s different on mobile android it works fine in firefox | 1 |
421,160 | 12,254,605,569 | IssuesEvent | 2020-05-06 08:46:54 | GiftForGood/website | https://api.github.com/repos/GiftForGood/website | opened | Reset Password for NPO | c.UserStory m.MVP priority.High | # User Story
<!--
https://github.com/GiftForGood/website/issues?q=is%3Aissue+label%3Ac.UserStory
-->
## Describe the user story in detail.
As a NPO, I want to reset my password so that I can change my password if I misplaced my previous password.
| 1.0 | Reset Password for NPO - # User Story
<!--
https://github.com/GiftForGood/website/issues?q=is%3Aissue+label%3Ac.UserStory
-->
## Describe the user story in detail.
As a NPO, I want to reset my password so that I can change my password if I misplaced my previous password.
| priority | reset password for npo user story describe the user story in detail as a npo i want to reset my password so that i can change my password if i misplaced my previous password | 1 |
386,415 | 11,438,600,697 | IssuesEvent | 2020-02-05 04:12:04 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | closed | Error while parsing OpenAPIDefinition when updating API definition with the same swagger | 3.1.0 Priority/Highest Severity/Major Type/Bug | 1. Create an API using the below customer-info-api.yaml (using the same swagger in https://github.com/wso2/product-apim/issues/7068).
2. Update the swagger, go to API Definition and click on IMPORT DEFINITION.
3. Upload the same swagger.
[customer-info-api.zip](https://github.com/wso2/product-apim/files/3925652/customer-info-api.zip)
`[2019-12-05 13:25:46,362] ERROR - ApisApiServiceImpl Error while parsing OpenAPI definition:attribute components.securitySchemes.default.scopes is missing
`

| 1.0 | Error while parsing OpenAPIDefinition when updating API definition with the same swagger - 1. Create an API using the below customer-info-api.yaml (using the same swagger in https://github.com/wso2/product-apim/issues/7068).
2. Update the swagger, go to API Definition and click on IMPORT DEFINITION.
3. Upload the same swagger.
[customer-info-api.zip](https://github.com/wso2/product-apim/files/3925652/customer-info-api.zip)
`[2019-12-05 13:25:46,362] ERROR - ApisApiServiceImpl Error while parsing OpenAPI definition:attribute components.securitySchemes.default.scopes is missing
`

| priority | error while parsing openapidefinition when updating api definition with the same swagger create an api using the below customer info api yaml using the same swagger in update the swagger go to api definition and click on import definition upload the same swagger error apisapiserviceimpl error while parsing openapi definition attribute components securityschemes default scopes is missing | 1 |
235,777 | 7,742,974,690 | IssuesEvent | 2018-05-29 11:18:35 | wso2/product-apim | https://api.github.com/repos/wso2/product-apim | opened | Add swagger validation before saving the API | 2.5.0 Priority/Highest Type/Improvement | **Description:**
API Manager allows to save swagger file with syntax errors. For example, if we remove the HTTP method in the swagger file using the in-built swagger editor, we could save the API without resources.
Refer image
<img width="1323" alt="screen shot 2018-05-29 at 4 46 26 pm" src="https://user-images.githubusercontent.com/4861150/40655731-e90a0d8e-635f-11e8-89eb-97260e92508a.png">
Publishing the API (with errors) is possible after that. Add a validation to the swagger definition before continuing the flow
**Affected Product Version:**
APIM 2.1.0
| 1.0 | Add swagger validation before saving the API - **Description:**
API Manager allows to save swagger file with syntax errors. For example, if we remove the HTTP method in the swagger file using the in-built swagger editor, we could save the API without resources.
Refer image
<img width="1323" alt="screen shot 2018-05-29 at 4 46 26 pm" src="https://user-images.githubusercontent.com/4861150/40655731-e90a0d8e-635f-11e8-89eb-97260e92508a.png">
Publishing the API (with errors) is possible after that. Add a validation to the swagger definition before continuing the flow
**Affected Product Version:**
APIM 2.1.0
| priority | add swagger validation before saving the api description api manager allows to save swagger file with syntax errors for example if we remove the http method in the swagger file using the in built swagger editor we could save the api without resources refer image img width alt screen shot at pm src publishing the api with errors is possible after that add a validation to the swagger definition before continuing the flow affected product version apim | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.