Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
210,625
7,191,620,426
IssuesEvent
2018-02-02 21:47:01
RhoInc/safety-results-over-time
https://api.github.com/repos/RhoInc/safety-results-over-time
closed
Add ordering variable setting to settings.time_settings and use it to order the x-axis domain.
high priority
# Test notes * Visits are sorted according to the corresponding numeric visit variable specified in `time_settings.order_col` if it exists. * Visits are sorted alphanumerically otherwise.
1.0
Add ordering variable setting to settings.time_settings and use it to order the x-axis domain. - # Test notes * Visits are sorted according to the corresponding numeric visit variable specified in `time_settings.order_col` if it exists. * Visits are sorted alphanumerically otherwise.
priority
add ordering variable setting to settings time settings and use it to order the x axis domain test notes visits are sorted according to the corresponding numeric visit variable specified in time settings order col if it exists visits are sorted alphanumerically otherwise
1
136,060
5,270,579,327
IssuesEvent
2017-02-06 05:54:30
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
PyChop bugfixes
Component: Direct Inelastic Misc: Bug Priority: High
Instrument scientists on LET and MERLIN have identified several necessary changes to PyChop to fix bugs or for useability reasons: 1. On Merlin, PyChop currently outputs all possible Ei reps, even where the incident neutron flux means that such reps would be unobservable. A restriction that Ei < 7 meV should be applied to not display low energy reps which cannot be seen. LET should have a similar restriction for Ei > 30 meV. 2. On LET, the pulse removal chopper frequency should be less than 150 Hz (physical limit of the chopper). 3. On LET, in "Expert mode", the chopper 2 phase box disappears. Also "Expert Mode" should show another tab which displays the moderator and chopper time components of the overall resolution. 4. On LET, the frequency of chopper 4 in high resolution mode should be the same as the pulse removal chopper (chopper 3) frequency - not the resolution chopper frequency divided by 2 as at present (for other modes, it should stay as the resolution chopper frequency / 2). 5. LET has new resolution chopper disk, with slots of size 31, 20 and 15mm, so for high resolution mode, chopper 5 slot_width should now be 15mm.
1.0
PyChop bugfixes - Instrument scientists on LET and MERLIN have identified several necessary changes to PyChop to fix bugs or for useability reasons: 1. On Merlin, PyChop currently outputs all possible Ei reps, even where the incident neutron flux means that such reps would be unobservable. A restriction that Ei < 7 meV should be applied to not display low energy reps which cannot be seen. LET should have a similar restriction for Ei > 30 meV. 2. On LET, the pulse removal chopper frequency should be less than 150 Hz (physical limit of the chopper). 3. On LET, in "Expert mode", the chopper 2 phase box disappears. Also "Expert Mode" should show another tab which displays the moderator and chopper time components of the overall resolution. 4. On LET, the frequency of chopper 4 in high resolution mode should be the same as the pulse removal chopper (chopper 3) frequency - not the resolution chopper frequency divided by 2 as at present (for other modes, it should stay as the resolution chopper frequency / 2). 5. LET has new resolution chopper disk, with slots of size 31, 20 and 15mm, so for high resolution mode, chopper 5 slot_width should now be 15mm.
priority
pychop bugfixes instrument scientists on let and merlin have identified several necessary changes to pychop to fix bugs or for useability reasons on merlin pychop currently outputs all possible ei reps even where the incident neutron flux means that such reps would be unobservable a restriction that ei mev on let the pulse removal chopper frequency should be less than hz physical limit of the chopper on let in expert mode the chopper phase box disappears also expert mode should show another tab which displays the moderator and chopper time components of the overall resolution on let the frequency of chopper in high resolution mode should be the same as the pulse removal chopper chopper frequency not the resolution chopper frequency divided by as at present for other modes it should stay as the resolution chopper frequency let has new resolution chopper disk with slots of size and so for high resolution mode chopper slot width should now be
1
461,281
13,227,998,442
IssuesEvent
2020-08-18 05:00:39
Bruce773/lyric-keeper
https://api.github.com/repos/Bruce773/lyric-keeper
closed
Auto-scroll on lyrics pages is broken
Bug TIMEFRAME: High-priority
#### Description (1): - User can decrease scroll below "-500" - When scroll is decreased past "-500", the site breaks #### Description (2): - If the user does not stop auto-scroll before clicking "BACK", the homepage scrolls #### Reproduce (1): 1. Go to any lyric page 2. Click the decrease button several times 3. Watch the site break! #### Reproduce (2): 1. Go to any lyric page 2. Turn on auto-scroll 3. Click the back button 4. Watch the homepage scroll on it's own!
1.0
Auto-scroll on lyrics pages is broken - #### Description (1): - User can decrease scroll below "-500" - When scroll is decreased past "-500", the site breaks #### Description (2): - If the user does not stop auto-scroll before clicking "BACK", the homepage scrolls #### Reproduce (1): 1. Go to any lyric page 2. Click the decrease button several times 3. Watch the site break! #### Reproduce (2): 1. Go to any lyric page 2. Turn on auto-scroll 3. Click the back button 4. Watch the homepage scroll on it's own!
priority
auto scroll on lyrics pages is broken description user can decrease scroll below when scroll is decreased past the site breaks description if the user does not stop auto scroll before clicking back the homepage scrolls reproduce go to any lyric page click the decrease button several times watch the site break reproduce go to any lyric page turn on auto scroll click the back button watch the homepage scroll on it s own
1
720,622
24,799,222,492
IssuesEvent
2022-10-24 20:05:41
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
closed
Ensure Add OSDs happen before or after upgrade
High Priority lifecycle/stale
Based on the PG imbalance issues and bz and the huge timeout it may need, we need to separate out the add osd logic out of upgrade suite to keep it clean for regression runs.
1.0
Ensure Add OSDs happen before or after upgrade - Based on the PG imbalance issues and bz and the huge timeout it may need, we need to separate out the add osd logic out of upgrade suite to keep it clean for regression runs.
priority
ensure add osds happen before or after upgrade based on the pg imbalance issues and bz and the huge timeout it may need we need to separate out the add osd logic out of upgrade suite to keep it clean for regression runs
1
795,351
28,070,492,248
IssuesEvent
2023-03-29 18:40:39
QuiltMC/quiltflower
https://api.github.com/repos/QuiltMC/quiltflower
closed
Decompiler produces a strange unneeded cast but it's `(<unknown>)`
bug Priority: High Subsystem: Variables
Source: ```java public void test() { boolean vvv1 = true, vvv2 = false; try { if (vvv2) { throw new RuntimeException(); } vvv2 = vvv1; } finally { char vvv31; } vvv1 = vvv2; Object vvv34 = null; while (vvv34 != null) { vvv2 = vvv1; System.out.println(vvv34); vvv1 = vvv2; } } ``` Result: ```java public void test() { boolean vvv1 = true; boolean vvv2 = false; try { if (vvv2) { throw new RuntimeException(); } vvv2 = vvv1; } finally { ; } boolean var8 = (<unknown>)vvv2; // what? for(Object vvv34 = null; vvv34 != null; var8 = var8) { System.out.println(vvv34); } } ```
1.0
Decompiler produces a strange unneeded cast but it's `(<unknown>)` - Source: ```java public void test() { boolean vvv1 = true, vvv2 = false; try { if (vvv2) { throw new RuntimeException(); } vvv2 = vvv1; } finally { char vvv31; } vvv1 = vvv2; Object vvv34 = null; while (vvv34 != null) { vvv2 = vvv1; System.out.println(vvv34); vvv1 = vvv2; } } ``` Result: ```java public void test() { boolean vvv1 = true; boolean vvv2 = false; try { if (vvv2) { throw new RuntimeException(); } vvv2 = vvv1; } finally { ; } boolean var8 = (<unknown>)vvv2; // what? for(Object vvv34 = null; vvv34 != null; var8 = var8) { System.out.println(vvv34); } } ```
priority
decompiler produces a strange unneeded cast but it s source java public void test boolean true false try if throw new runtimeexception finally char object null while null system out println result java public void test boolean true boolean false try if throw new runtimeexception finally boolean what for object null null system out println
1
345,474
10,368,113,329
IssuesEvent
2019-09-07 14:22:51
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
Change Internal Links to AMP
NEXT UPDATE [Priority: HIGH] bug
Hi guys, The option **Change Internal Links to AMP** is changing my links to subdomains as well. example: pebmed.com.br/amp has an external link to events.pebmed.com.br in which is added /amp. This generates a broken link as bellow: events.pebmed.com.br/amp
1.0
Change Internal Links to AMP - Hi guys, The option **Change Internal Links to AMP** is changing my links to subdomains as well. example: pebmed.com.br/amp has an external link to events.pebmed.com.br in which is added /amp. This generates a broken link as bellow: events.pebmed.com.br/amp
priority
change internal links to amp hi guys the option change internal links to amp is changing my links to subdomains as well example pebmed com br amp has an external link to events pebmed com br in which is added amp this generates a broken link as bellow events pebmed com br amp
1
358,666
10,619,153,590
IssuesEvent
2019-10-13 11:09:41
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
Plasteel Explodes (only a little), causing R-Walls to spam explosion logs when blown up
Oversight Priority: High
Blow up the supermatter while as an admin and see the fun for yourself. This is so severe that if it hits enough r-walls it can cause client crashing.
1.0
Plasteel Explodes (only a little), causing R-Walls to spam explosion logs when blown up - Blow up the supermatter while as an admin and see the fun for yourself. This is so severe that if it hits enough r-walls it can cause client crashing.
priority
plasteel explodes only a little causing r walls to spam explosion logs when blown up blow up the supermatter while as an admin and see the fun for yourself this is so severe that if it hits enough r walls it can cause client crashing
1
483,480
13,925,415,161
IssuesEvent
2020-10-21 16:48:55
eventespresso/barista
https://api.github.com/repos/eventespresso/barista
closed
Block Ticket Price Updates for Tickets with Registrations
C: data systems 🗑 D: EDTR ✏️ P2: HIGH priority 😮 T: bug 🐞
From Garth: > I'm able to edit ticket prices after registrations are submitted (pending or approved). That used to not be allowed. && > i was able to BULK EDIT a ticket price after registrations were approved. > > You'll notice that on this event, the first ticket is called "Free Ticket". I have two attendees that registered and were approved, then, I BULK EDITED the tickets together and changed the prices to a non-zero dollar amount. Now the free ticket is no longer free. In master, we prohibited people from changing the price of tickets once a registration had been complete as to not confuse the accounting of registrations and transactions: https://barista.eventespresso.com/wp-admin/admin.php?page=espresso_events&action=edit&post=40 For single tickets we need to disable both the Ticket Price inline edit input as well as the Ticket Price Calculator for any ticket with registrations, and display an appropriate notice in the tooltip (for the TPC icon button) and on the Ticket Price form step. For bulk editing we can't simply disable the selector checkbox for any tickets with registrations because then you would be unable to edit any other details of the ticket which should still be possible. So let's do the following: - detect tickets with registrations (sold > 0) and if any are selected (either individually or via the "select all" checkbox): - disable the "edit ticket prices" option in the bulk actions dropdown - display an error between the table and the bulk actions dropdown Here's a suggestion for the appropriate notice text: > Ticket price modifications are blocked for Tickets that have already been sold to registrants, because doing so would negatively affect internal accounting for the event. If you still need to modify ticket prices, then create a copy of those tickets, edit the prices for the new tickets, and then archive the old tickets. That's a quick fix that puts the responsibility for that process on the user, whereas the legacy editor would automatically copy and archive a ticket when the price was modified. The legacy editor was not capable of bulk editing however. A more complex solution would be to add an option to perform the copy and archive process for users automatically via bulk editing, but that might be tricky.
1.0
Block Ticket Price Updates for Tickets with Registrations - From Garth: > I'm able to edit ticket prices after registrations are submitted (pending or approved). That used to not be allowed. && > i was able to BULK EDIT a ticket price after registrations were approved. > > You'll notice that on this event, the first ticket is called "Free Ticket". I have two attendees that registered and were approved, then, I BULK EDITED the tickets together and changed the prices to a non-zero dollar amount. Now the free ticket is no longer free. In master, we prohibited people from changing the price of tickets once a registration had been complete as to not confuse the accounting of registrations and transactions: https://barista.eventespresso.com/wp-admin/admin.php?page=espresso_events&action=edit&post=40 For single tickets we need to disable both the Ticket Price inline edit input as well as the Ticket Price Calculator for any ticket with registrations, and display an appropriate notice in the tooltip (for the TPC icon button) and on the Ticket Price form step. For bulk editing we can't simply disable the selector checkbox for any tickets with registrations because then you would be unable to edit any other details of the ticket which should still be possible. So let's do the following: - detect tickets with registrations (sold > 0) and if any are selected (either individually or via the "select all" checkbox): - disable the "edit ticket prices" option in the bulk actions dropdown - display an error between the table and the bulk actions dropdown Here's a suggestion for the appropriate notice text: > Ticket price modifications are blocked for Tickets that have already been sold to registrants, because doing so would negatively affect internal accounting for the event. If you still need to modify ticket prices, then create a copy of those tickets, edit the prices for the new tickets, and then archive the old tickets. That's a quick fix that puts the responsibility for that process on the user, whereas the legacy editor would automatically copy and archive a ticket when the price was modified. The legacy editor was not capable of bulk editing however. A more complex solution would be to add an option to perform the copy and archive process for users automatically via bulk editing, but that might be tricky.
priority
block ticket price updates for tickets with registrations from garth i m able to edit ticket prices after registrations are submitted pending or approved that used to not be allowed i was able to bulk edit a ticket price after registrations were approved you ll notice that on this event the first ticket is called free ticket i have two attendees that registered and were approved then i bulk edited the tickets together and changed the prices to a non zero dollar amount now the free ticket is no longer free in master we prohibited people from changing the price of tickets once a registration had been complete as to not confuse the accounting of registrations and transactions for single tickets we need to disable both the ticket price inline edit input as well as the ticket price calculator for any ticket with registrations and display an appropriate notice in the tooltip for the tpc icon button and on the ticket price form step for bulk editing we can t simply disable the selector checkbox for any tickets with registrations because then you would be unable to edit any other details of the ticket which should still be possible so let s do the following detect tickets with registrations sold and if any are selected either individually or via the select all checkbox disable the edit ticket prices option in the bulk actions dropdown display an error between the table and the bulk actions dropdown here s a suggestion for the appropriate notice text ticket price modifications are blocked for tickets that have already been sold to registrants because doing so would negatively affect internal accounting for the event if you still need to modify ticket prices then create a copy of those tickets edit the prices for the new tickets and then archive the old tickets that s a quick fix that puts the responsibility for that process on the user whereas the legacy editor would automatically copy and archive a ticket when the price was modified the legacy editor was not capable of bulk editing however a more complex solution would be to add an option to perform the copy and archive process for users automatically via bulk editing but that might be tricky
1
613,025
19,071,219,140
IssuesEvent
2021-11-27 00:08:58
HEPData/hepdata
https://api.github.com/repos/HEPData/hepdata
closed
records: broken download link on resource file landing page and special treatment needed for .html files
type: bug priority: high complexity: medium
Some records migrated from the old HepData site, for example, [ins1208923](https://www.hepdata.net/record/ins1208923) ([old HepData](http://hepdata.cedar.ac.uk/view/ins1208923)) have resources like: ```yaml additional_resources: - {description: Rivet analysis, location: 'http://rivet.hepforge.org/analyses#CMS_2013_I1208923'} - {description: Link to correlation matrices for the Inclusive Jet data, location: /resource/9901922/inclusive_corr.html} - {description: Link to fastNLO v2.1 table (inclusive jets), location: /resource/1208923/fnl2332d_I1208923.tab.gz} - {description: Link to correlation matrices for the Dijet data, location: /resource/9901922/dijet_corr.html} - {description: Link to fastNLO v2.1 table (dijet mass), location: /resource/1208923/fnl2412e_I1208923.tab.gz} - {description: Link to the detailed systematics for the Inclusive Jet data, location: /resource/9901922/inclusive_sys.html} - {description: Link to the detailed systematics for the DiJet data, location: /resource/9901922/dijet_sys.html} ``` The resource files are stored in a subdirectory of the main data directory (`/opt/hepdata/var/data/94/66887/resources`). The last six resources are correctly given a DOI, but the "Download" links are broken from the resource file landing pages, although they work correctly from the "Additional Publication Resources" widget. Moreover, the landing page for the `.html` files just downloads the `.html` file rather than return the landing page. Presumably this is because the media type is `text/html`, so we might have to treat this as a special case, e.g. if the media type is `text/html` only return the content directly instead of the landing page if the `flask.request.accept_mimetypes` list has length 1, since [most web browsers](https://developer.mozilla.org/en-US/docs/Web/HTTP/Content_negotiation/List_of_default_Accept_values) will send an `Accept` header with multiple media types. Could you please address these issues, @alisonrclarke?
1.0
records: broken download link on resource file landing page and special treatment needed for .html files - Some records migrated from the old HepData site, for example, [ins1208923](https://www.hepdata.net/record/ins1208923) ([old HepData](http://hepdata.cedar.ac.uk/view/ins1208923)) have resources like: ```yaml additional_resources: - {description: Rivet analysis, location: 'http://rivet.hepforge.org/analyses#CMS_2013_I1208923'} - {description: Link to correlation matrices for the Inclusive Jet data, location: /resource/9901922/inclusive_corr.html} - {description: Link to fastNLO v2.1 table (inclusive jets), location: /resource/1208923/fnl2332d_I1208923.tab.gz} - {description: Link to correlation matrices for the Dijet data, location: /resource/9901922/dijet_corr.html} - {description: Link to fastNLO v2.1 table (dijet mass), location: /resource/1208923/fnl2412e_I1208923.tab.gz} - {description: Link to the detailed systematics for the Inclusive Jet data, location: /resource/9901922/inclusive_sys.html} - {description: Link to the detailed systematics for the DiJet data, location: /resource/9901922/dijet_sys.html} ``` The resource files are stored in a subdirectory of the main data directory (`/opt/hepdata/var/data/94/66887/resources`). The last six resources are correctly given a DOI, but the "Download" links are broken from the resource file landing pages, although they work correctly from the "Additional Publication Resources" widget. Moreover, the landing page for the `.html` files just downloads the `.html` file rather than return the landing page. Presumably this is because the media type is `text/html`, so we might have to treat this as a special case, e.g. if the media type is `text/html` only return the content directly instead of the landing page if the `flask.request.accept_mimetypes` list has length 1, since [most web browsers](https://developer.mozilla.org/en-US/docs/Web/HTTP/Content_negotiation/List_of_default_Accept_values) will send an `Accept` header with multiple media types. Could you please address these issues, @alisonrclarke?
priority
records broken download link on resource file landing page and special treatment needed for html files some records migrated from the old hepdata site for example have resources like yaml additional resources description rivet analysis location description link to correlation matrices for the inclusive jet data location resource inclusive corr html description link to fastnlo table inclusive jets location resource tab gz description link to correlation matrices for the dijet data location resource dijet corr html description link to fastnlo table dijet mass location resource tab gz description link to the detailed systematics for the inclusive jet data location resource inclusive sys html description link to the detailed systematics for the dijet data location resource dijet sys html the resource files are stored in a subdirectory of the main data directory opt hepdata var data resources the last six resources are correctly given a doi but the download links are broken from the resource file landing pages although they work correctly from the additional publication resources widget moreover the landing page for the html files just downloads the html file rather than return the landing page presumably this is because the media type is text html so we might have to treat this as a special case e g if the media type is text html only return the content directly instead of the landing page if the flask request accept mimetypes list has length since will send an accept header with multiple media types could you please address these issues alisonrclarke
1
327,172
9,967,648,434
IssuesEvent
2019-07-08 14:02:17
VandyHacks/vaken
https://api.github.com/repos/VandyHacks/vaken
closed
Don't generate OAuth users with a hardcoded password
Priority: High Topic: Security Type: Bug
This is a potential security vulnerability. in `src/server/auth.ts`
1.0
Don't generate OAuth users with a hardcoded password - This is a potential security vulnerability. in `src/server/auth.ts`
priority
don t generate oauth users with a hardcoded password this is a potential security vulnerability in src server auth ts
1
751,739
26,255,654,549
IssuesEvent
2023-01-06 00:17:42
apache/incubator-devlake
https://api.github.com/repos/apache/incubator-devlake
closed
[Feature][frame] transformationRules has set default value.
type/feature-request priority/high Stale
### Search before asking - [X] I had searched in the [issues](https://github.com/apache/incubator-devlake/issues?q=is%3Aissue) and found no similar feature requirement. ### Description when users use devlake, such as the github plugin, the user doesn't set the transformationRules, but the final data results have converted by the default transformationRules, Is this reasonable? Should the backend remove these default handling? The main focus I think is that the values displayed in the front-end interface are for user reference or are they intended to be used as default values? 1. Front-end transformationRules show: ![image](https://user-images.githubusercontent.com/101256042/191277354-7fe96edb-02a5-4440-ad01-4a15c522e1b1.png) 2. Backend transformationRules show: ![image](https://user-images.githubusercontent.com/101256042/191277491-45313dba-933f-4d0e-90a6-6af4b12deef7.png) ### Use case _No response_ ### Related issues _No response_ ### Are you willing to submit a PR? - [ ] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
1.0
[Feature][frame] transformationRules has set default value. - ### Search before asking - [X] I had searched in the [issues](https://github.com/apache/incubator-devlake/issues?q=is%3Aissue) and found no similar feature requirement. ### Description when users use devlake, such as the github plugin, the user doesn't set the transformationRules, but the final data results have converted by the default transformationRules, Is this reasonable? Should the backend remove these default handling? The main focus I think is that the values displayed in the front-end interface are for user reference or are they intended to be used as default values? 1. Front-end transformationRules show: ![image](https://user-images.githubusercontent.com/101256042/191277354-7fe96edb-02a5-4440-ad01-4a15c522e1b1.png) 2. Backend transformationRules show: ![image](https://user-images.githubusercontent.com/101256042/191277491-45313dba-933f-4d0e-90a6-6af4b12deef7.png) ### Use case _No response_ ### Related issues _No response_ ### Are you willing to submit a PR? - [ ] Yes I am willing to submit a PR! ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
priority
transformationrules has set default value search before asking i had searched in the and found no similar feature requirement description when users use devlake such as the github plugin the user doesn t set the transformationrules but the final data results have converted by the default transformationrules is this reasonable should the backend remove these default handling? the main focus i think is that the values displayed in the front end interface are for user reference or are they intended to be used as default values front end transformationrules show backend transformationrules show use case no response related issues no response are you willing to submit a pr yes i am willing to submit a pr code of conduct i agree to follow this project s
1
469,362
13,507,359,988
IssuesEvent
2020-09-14 05:46:46
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
apps.facebook.com - site is not usable
browser-firefox engine-gecko ml-needsdiagnosis-false ml-probability-high priority-critical
<!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/58151 --> **URL**: https://apps.facebook.com/doubleucasino/?force_login=1&casinoidx=&gameidx=&subtype=&adflag= **Browser / Version**: Firefox 81.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: dont work <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200910180444</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/d66f84c4-3a80-4358-9511-17e5841f196d) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
apps.facebook.com - site is not usable - <!-- @browser: Firefox 81.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 6.1; Win64; x64; rv:81.0) Gecko/20100101 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/58151 --> **URL**: https://apps.facebook.com/doubleucasino/?force_login=1&casinoidx=&gameidx=&subtype=&adflag= **Browser / Version**: Firefox 81.0 **Operating System**: Windows 7 **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Page not loading correctly **Steps to Reproduce**: dont work <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200910180444</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/9/d66f84c4-3a80-4358-9511-17e5841f196d) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
apps facebook com site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce dont work browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
199,979
6,996,774,488
IssuesEvent
2017-12-16 05:11:31
dart-lang/pub
https://api.github.com/repos/dart-lang/pub
reopened
Pub should handle pub server errors gracefully
enhancement high priority
The pub server will sometimes return a 500 (internal error), if there is a temporary failure on the server. pub get should handle these results, and retry the fetch of the object from the server. If it doesn't retry, then it should return with a distinctive error code that indicates temporary failure. Maybe this should happen for other network failures when running pub get as well. This would make the testing of pub get less flaky, and allow us to test it on the buildbot. This seems related to issue #1534 and pr https://github.com/dart-lang/pub/pull/1552 @mkustermann
1.0
Pub should handle pub server errors gracefully - The pub server will sometimes return a 500 (internal error), if there is a temporary failure on the server. pub get should handle these results, and retry the fetch of the object from the server. If it doesn't retry, then it should return with a distinctive error code that indicates temporary failure. Maybe this should happen for other network failures when running pub get as well. This would make the testing of pub get less flaky, and allow us to test it on the buildbot. This seems related to issue #1534 and pr https://github.com/dart-lang/pub/pull/1552 @mkustermann
priority
pub should handle pub server errors gracefully the pub server will sometimes return a internal error if there is a temporary failure on the server pub get should handle these results and retry the fetch of the object from the server if it doesn t retry then it should return with a distinctive error code that indicates temporary failure maybe this should happen for other network failures when running pub get as well this would make the testing of pub get less flaky and allow us to test it on the buildbot this seems related to issue and pr mkustermann
1
502,591
14,562,494,558
IssuesEvent
2020-12-17 00:13:44
Pocket-PC-Planet/PPCP-AppManager
https://api.github.com/repos/Pocket-PC-Planet/PPCP-AppManager
closed
Link handler not working
Core High Priority bug
Nothing happens, no messages appear when enabled in dev mode when link clicked. Tested on WM5 HP iPAQ rx1955.
1.0
Link handler not working - Nothing happens, no messages appear when enabled in dev mode when link clicked. Tested on WM5 HP iPAQ rx1955.
priority
link handler not working nothing happens no messages appear when enabled in dev mode when link clicked tested on hp ipaq
1
693,165
23,765,382,293
IssuesEvent
2022-09-01 12:23:56
HumanExposure/factotum
https://api.github.com/repos/HumanExposure/factotum
closed
Bump elasticsearch
bug Priority :: High
Deployments are showing that the main factotum container is failing to start up because elasticsearch is failing to start. The logs aren't very informative either. On the other hand, API container is up and running - likely because API container lacks the `depends_on` relationship with elasticsearch the way the main factotum container does. Based on trial and error, bumping elasticsearch to 7.17.6 has shown to be a possible solution, at least locally, for reasons unknown (documentation could not be found). Acceptance criteria: - bump elasticsearch from 7.16.1 to 7.17.6 - confirm successful application startup and completion of logstash pipeline on deployments logs: ``` {"log":"\u0009... 1 more\n","stream":"stderr","time":"2022-09-01T02:05:03.627891882Z"} {"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/Level\n","stream":"stderr","time":"2022-09-01T02:06:05.972415467Z"} {"log":"\u0009at org.elasticsearch.cli.CommandLoggingConfigurator.configureLoggingWithoutConfig(CommandLoggingConfigurator.java:27)\n","stream":"stderr","time":"2022-09-01T02:06:05.972526009Z"} -- {"log":"\u0009... 3 more\n","stream":"stderr","time":"2022-09-01T02:06:05.973176009Z"} {"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/status/StatusListener\n","stream":"stderr","time":"2022-09-01T02:06:07.197484171Z"} {"log":"\u0009at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:78)\n","stream":"stderr","time":"2022-09-01T02:06:07.197583131Z"} ```
1.0
Bump elasticsearch - Deployments are showing that the main factotum container is failing to start up because elasticsearch is failing to start. The logs aren't very informative either. On the other hand, API container is up and running - likely because API container lacks the `depends_on` relationship with elasticsearch the way the main factotum container does. Based on trial and error, bumping elasticsearch to 7.17.6 has shown to be a possible solution, at least locally, for reasons unknown (documentation could not be found). Acceptance criteria: - bump elasticsearch from 7.16.1 to 7.17.6 - confirm successful application startup and completion of logstash pipeline on deployments logs: ``` {"log":"\u0009... 1 more\n","stream":"stderr","time":"2022-09-01T02:05:03.627891882Z"} {"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/Level\n","stream":"stderr","time":"2022-09-01T02:06:05.972415467Z"} {"log":"\u0009at org.elasticsearch.cli.CommandLoggingConfigurator.configureLoggingWithoutConfig(CommandLoggingConfigurator.java:27)\n","stream":"stderr","time":"2022-09-01T02:06:05.972526009Z"} -- {"log":"\u0009... 3 more\n","stream":"stderr","time":"2022-09-01T02:06:05.973176009Z"} {"log":"Exception in thread \"main\" java.lang.NoClassDefFoundError: org/apache/logging/log4j/status/StatusListener\n","stream":"stderr","time":"2022-09-01T02:06:07.197484171Z"} {"log":"\u0009at org.elasticsearch.bootstrap.Elasticsearch.main(Elasticsearch.java:78)\n","stream":"stderr","time":"2022-09-01T02:06:07.197583131Z"} ```
priority
bump elasticsearch deployments are showing that the main factotum container is failing to start up because elasticsearch is failing to start the logs aren t very informative either on the other hand api container is up and running likely because api container lacks the depends on relationship with elasticsearch the way the main factotum container does based on trial and error bumping elasticsearch to has shown to be a possible solution at least locally for reasons unknown documentation could not be found acceptance criteria bump elasticsearch from to confirm successful application startup and completion of logstash pipeline on deployments logs log more n stream stderr time log exception in thread main java lang noclassdeffounderror org apache logging level n stream stderr time log org elasticsearch cli commandloggingconfigurator configureloggingwithoutconfig commandloggingconfigurator java n stream stderr time log more n stream stderr time log exception in thread main java lang noclassdeffounderror org apache logging status statuslistener n stream stderr time log org elasticsearch bootstrap elasticsearch main elasticsearch java n stream stderr time
1
360,776
10,697,258,735
IssuesEvent
2019-10-23 16:09:21
wso2/analytics-apim
https://api.github.com/repos/wso2/analytics-apim
closed
Subscriptions Over time table count isn't getting populated
Priority/Highest Severity/Critical
**Description:** Subscriptions Over time table count isn't getting populated. **Affected Product Version:** wso2am-analytics-3.0.0 **OS, DB, other environment details and versions:** Chrome <img width="764" alt="Screen Shot 2019-10-23 at 11 49 21 AM" src="https://user-images.githubusercontent.com/19324135/67363633-74c77a00-f58b-11e9-8fe2-fd60ba79fb35.png">
1.0
Subscriptions Over time table count isn't getting populated - **Description:** Subscriptions Over time table count isn't getting populated. **Affected Product Version:** wso2am-analytics-3.0.0 **OS, DB, other environment details and versions:** Chrome <img width="764" alt="Screen Shot 2019-10-23 at 11 49 21 AM" src="https://user-images.githubusercontent.com/19324135/67363633-74c77a00-f58b-11e9-8fe2-fd60ba79fb35.png">
priority
subscriptions over time table count isn t getting populated description subscriptions over time table count isn t getting populated affected product version analytics os db other environment details and versions chrome img width alt screen shot at am src
1
48,806
3,000,030,036
IssuesEvent
2015-07-23 22:11:49
GRIS-UdeM/ZirkOSC
https://api.github.com/repos/GRIS-UdeM/ZirkOSC
opened
Source elevation à 90° change l'azimuth à 180°
bug high priority
Lorsque l'on change l'élévation d'une source avec le slider et qu'on atteint 90°, son azimut est automatiquement changé à 180°. Ensuite de cela, lorsque l'on redescend l'élévation, l'azimut alterne entre 180° et -180°. <https://dl.dropboxusercontent.com/u/2295576/GRIS/ZirkOSC_3_0_0_Circular02.mov>
1.0
Source elevation à 90° change l'azimuth à 180° - Lorsque l'on change l'élévation d'une source avec le slider et qu'on atteint 90°, son azimut est automatiquement changé à 180°. Ensuite de cela, lorsque l'on redescend l'élévation, l'azimut alterne entre 180° et -180°. <https://dl.dropboxusercontent.com/u/2295576/GRIS/ZirkOSC_3_0_0_Circular02.mov>
priority
source elevation à ° change l azimuth à ° lorsque l on change l élévation d une source avec le slider et qu on atteint ° son azimut est automatiquement changé à ° ensuite de cela lorsque l on redescend l élévation l azimut alterne entre ° et °
1
471,803
13,610,811,623
IssuesEvent
2020-09-23 07:57:41
wso2/product-apim-tooling
https://api.github.com/repos/wso2/product-apim-tooling
closed
Cannot do PUBLISHED -> CREATED transition by updating api.yaml
Affected/3.1.0 Priority/High Type/Bug
**Description:** PUBLISHED -> CREATED transition is a valid API Lifecycle transition which is normally done by "Demote to Created" lifecycle action. By editing the api.yaml of an exported API, we should be able to to that transition but it is not possible and an error is thrown in APIM. **Affected Product Version:** 3.1.0, 3.2.0-RC2 **Steps to reproduce:** 1. Create an publish an API 2. Export the API using APICTL 3. Extract the export API zip file 4. Update the api.yaml of the extracted ZIP `status: CREATED` from `status: PUBLISHED` 5. Import the API with --update ``` $ apictl import-api -f Hello-1.0/ -e 310 --update Existing API found, attempting to update it... API ID: 3381cfcc-e9b4-47d4-8036-8e9143fb6fa1 Error importing API. Status: 500 Response: {"code":500,"message":"Internal server error","description":"Error while importing API","moreInfo":"","error":[]} apictl: Error importing API Reason: 500 Exit status 1 ``` Below error is thrown in the APIM side. ``` [2020-08-10 21:50:18,281] ERROR - ImportApiServiceImpl Error while importing API org.wso2.carbon.apimgt.impl.importexport.APIImportExportException: Error while importing API: Hello version: 1.0 at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:356) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI(APIImportUtil.java:115) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive_aroundBody8(APIImportExportManager.java:154) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive(APIImportExportManager.java:144) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.rest.api.admin.v1.impl.ImportApiServiceImpl.importApiPost(ImportApiServiceImpl.java:103) [classes/:?] at org.wso2.carbon.apimgt.rest.api.admin.v1.ImportApi.importApiPost(ImportApi.java:58) [classes/:?] at sun.reflect.GeneratedMethodAccessor273.invoke(Unknown Source) ~[?:?] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_231] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_231] at org.apache.cxf.service.invoker.AbstractInvoker.performInvocation(AbstractInvoker.java:179) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.service.invoker.AbstractInvoker.invoke(AbstractInvoker.java:96) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:193) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8] at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:103) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8] at org.apache.cxf.interceptor.ServiceInvokerInterceptor$1.run(ServiceInvokerInterceptor.java:59) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.interceptor.ServiceInvokerInterceptor.handleMessage(ServiceInvokerInterceptor.java:96) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.phase.PhaseInterceptorChain.doIntercept(PhaseInterceptorChain.java:308) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.transport.ChainInitiationObserver.onMessage(ChainInitiationObserver.java:121) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.transport.http.AbstractHTTPDestination.invoke(AbstractHTTPDestination.java:267) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.ServletController.invokeDestination(ServletController.java:234) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:208) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:160) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.CXFNonSpringServlet.invoke(CXFNonSpringServlet.java:216) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.AbstractHTTPServlet.handleRequest(AbstractHTTPServlet.java:301) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.AbstractHTTPServlet.doPost(AbstractHTTPServlet.java:220) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at javax.servlet.http.HttpServlet.service(HttpServlet.java:660) [tomcat-servlet-api_9.0.31.wso2v1.jar:?] at org.apache.cxf.transport.servlet.AbstractHTTPServlet.service(AbstractHTTPServlet.java:276) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:541) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92) [tomcat_9.0.31.wso2v1.jar:?] at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:86) [org.wso2.carbon.identity.context.rewrite.valve_1.4.0.jar:?] at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:110) [org.wso2.carbon.identity.authz.valve_1.4.0.jar:?] at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:75) [org.wso2.carbon.identity.auth.valve_1.4.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:49) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:145) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:688) [tomcat_9.0.31.wso2v1.jar:?] at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve.invoke(RequestCorrelationIdValve.java:119) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:367) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:65) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:868) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1639) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat_9.0.31.wso2v1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_231] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_231] at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.31.wso2v1.jar:?] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_231] Caused by: org.wso2.carbon.apimgt.api.APIManagementException: Error while setting registry lifecycle checklist items for the API: Hello at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException_aroundBody132(AbstractAPIManager.java:1869) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException(AbstractAPIManager.java:1868) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5777) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] ... 56 more Caused by: org.wso2.carbon.governance.api.exception.GovernanceException: Invalid check list item. at org.wso2.carbon.governance.api.common.dataobjects.GovernanceArtifactImpl.isLCItemChecked(GovernanceArtifactImpl.java:897) ~[org.wso2.carbon.governance.api_4.8.19.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5769) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] ... 56 more ```
1.0
Cannot do PUBLISHED -> CREATED transition by updating api.yaml - **Description:** PUBLISHED -> CREATED transition is a valid API Lifecycle transition which is normally done by "Demote to Created" lifecycle action. By editing the api.yaml of an exported API, we should be able to to that transition but it is not possible and an error is thrown in APIM. **Affected Product Version:** 3.1.0, 3.2.0-RC2 **Steps to reproduce:** 1. Create an publish an API 2. Export the API using APICTL 3. Extract the export API zip file 4. Update the api.yaml of the extracted ZIP `status: CREATED` from `status: PUBLISHED` 5. Import the API with --update ``` $ apictl import-api -f Hello-1.0/ -e 310 --update Existing API found, attempting to update it... API ID: 3381cfcc-e9b4-47d4-8036-8e9143fb6fa1 Error importing API. Status: 500 Response: {"code":500,"message":"Internal server error","description":"Error while importing API","moreInfo":"","error":[]} apictl: Error importing API Reason: 500 Exit status 1 ``` Below error is thrown in the APIM side. ``` [2020-08-10 21:50:18,281] ERROR - ImportApiServiceImpl Error while importing API org.wso2.carbon.apimgt.impl.importexport.APIImportExportException: Error while importing API: Hello version: 1.0 at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:356) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI(APIImportUtil.java:115) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive_aroundBody8(APIImportExportManager.java:154) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.APIImportExportManager.importAPIArchive(APIImportExportManager.java:144) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.rest.api.admin.v1.impl.ImportApiServiceImpl.importApiPost(ImportApiServiceImpl.java:103) [classes/:?] at org.wso2.carbon.apimgt.rest.api.admin.v1.ImportApi.importApiPost(ImportApi.java:58) [classes/:?] at sun.reflect.GeneratedMethodAccessor273.invoke(Unknown Source) ~[?:?] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) ~[?:1.8.0_231] at java.lang.reflect.Method.invoke(Method.java:498) ~[?:1.8.0_231] at org.apache.cxf.service.invoker.AbstractInvoker.performInvocation(AbstractInvoker.java:179) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.service.invoker.AbstractInvoker.invoke(AbstractInvoker.java:96) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:193) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8] at org.apache.cxf.jaxrs.JAXRSInvoker.invoke(JAXRSInvoker.java:103) [cxf-rt-frontend-jaxrs-3.2.8.jar:3.2.8] at org.apache.cxf.interceptor.ServiceInvokerInterceptor$1.run(ServiceInvokerInterceptor.java:59) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.interceptor.ServiceInvokerInterceptor.handleMessage(ServiceInvokerInterceptor.java:96) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.phase.PhaseInterceptorChain.doIntercept(PhaseInterceptorChain.java:308) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.transport.ChainInitiationObserver.onMessage(ChainInitiationObserver.java:121) [cxf-core-3.2.8.jar:3.2.8] at org.apache.cxf.transport.http.AbstractHTTPDestination.invoke(AbstractHTTPDestination.java:267) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.ServletController.invokeDestination(ServletController.java:234) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:208) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.ServletController.invoke(ServletController.java:160) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.CXFNonSpringServlet.invoke(CXFNonSpringServlet.java:216) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.AbstractHTTPServlet.handleRequest(AbstractHTTPServlet.java:301) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.cxf.transport.servlet.AbstractHTTPServlet.doPost(AbstractHTTPServlet.java:220) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at javax.servlet.http.HttpServlet.service(HttpServlet.java:660) [tomcat-servlet-api_9.0.31.wso2v1.jar:?] at org.apache.cxf.transport.servlet.AbstractHTTPServlet.service(AbstractHTTPServlet.java:276) [cxf-rt-transports-http-3.2.8.jar:3.2.8] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:231) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:53) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:193) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:166) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:202) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:96) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:541) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:139) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:92) [tomcat_9.0.31.wso2v1.jar:?] at org.wso2.carbon.identity.context.rewrite.valve.TenantContextRewriteValve.invoke(TenantContextRewriteValve.java:86) [org.wso2.carbon.identity.context.rewrite.valve_1.4.0.jar:?] at org.wso2.carbon.identity.authz.valve.AuthorizationValve.invoke(AuthorizationValve.java:110) [org.wso2.carbon.identity.authz.valve_1.4.0.jar:?] at org.wso2.carbon.identity.auth.valve.AuthenticationValve.invoke(AuthenticationValve.java:75) [org.wso2.carbon.identity.auth.valve_1.4.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.CompositeValve.continueInvocation(CompositeValve.java:99) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.TomcatValveContainer.invokeValves(TomcatValveContainer.java:49) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.CompositeValve.invoke(CompositeValve.java:62) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.CarbonStuckThreadDetectionValve.invoke(CarbonStuckThreadDetectionValve.java:145) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.apache.catalina.valves.AbstractAccessLogValve.invoke(AbstractAccessLogValve.java:688) [tomcat_9.0.31.wso2v1.jar:?] at org.wso2.carbon.tomcat.ext.valves.CarbonContextCreatorValve.invoke(CarbonContextCreatorValve.java:57) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.wso2.carbon.tomcat.ext.valves.RequestCorrelationIdValve.invoke(RequestCorrelationIdValve.java:119) [org.wso2.carbon.tomcat.ext_4.6.0.jar:?] at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:74) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:343) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.coyote.http11.Http11Processor.service(Http11Processor.java:367) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.coyote.AbstractProcessorLight.process(AbstractProcessorLight.java:65) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.coyote.AbstractProtocol$ConnectionHandler.process(AbstractProtocol.java:868) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.doRun(NioEndpoint.java:1639) [tomcat_9.0.31.wso2v1.jar:?] at org.apache.tomcat.util.net.SocketProcessorBase.run(SocketProcessorBase.java:49) [tomcat_9.0.31.wso2v1.jar:?] at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) [?:1.8.0_231] at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) [?:1.8.0_231] at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) [tomcat_9.0.31.wso2v1.jar:?] at java.lang.Thread.run(Thread.java:748) [?:1.8.0_231] Caused by: org.wso2.carbon.apimgt.api.APIManagementException: Error while setting registry lifecycle checklist items for the API: Hello at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException_aroundBody132(AbstractAPIManager.java:1869) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.AbstractAPIManager.handleException(AbstractAPIManager.java:1868) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5777) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] ... 56 more Caused by: org.wso2.carbon.governance.api.exception.GovernanceException: Invalid check list item. at org.wso2.carbon.governance.api.common.dataobjects.GovernanceArtifactImpl.isLCItemChecked(GovernanceArtifactImpl.java:897) ~[org.wso2.carbon.governance.api_4.8.19.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems_aroundBody274(APIProviderImpl.java:5769) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.APIProviderImpl.changeAPILCCheckListItems(APIProviderImpl.java:5750) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems_aroundBody78(UserAwareAPIProvider.java:372) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.UserAwareAPIProvider.changeAPILCCheckListItems(UserAwareAPIProvider.java:368) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] at org.wso2.carbon.apimgt.impl.importexport.utils.APIImportUtil.importAPI_aroundBody2(APIImportUtil.java:334) ~[org.wso2.carbon.apimgt.impl_6.7.181.jar:?] ... 56 more ```
priority
cannot do published created transition by updating api yaml description published created transition is a valid api lifecycle transition which is normally done by demote to created lifecycle action by editing the api yaml of an exported api we should be able to to that transition but it is not possible and an error is thrown in apim affected product version steps to reproduce create an publish an api export the api using apictl extract the export api zip file update the api yaml of the extracted zip status created from status published import the api with update apictl import api f hello e update existing api found attempting to update it api id error importing api status response code message internal server error description error while importing api moreinfo error apictl error importing api reason exit status below error is thrown in the apim side error importapiserviceimpl error while importing api org carbon apimgt impl importexport apiimportexportexception error while importing api hello version at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java at org carbon apimgt impl importexport apiimportexportmanager importapiarchive apiimportexportmanager java at org carbon apimgt impl importexport apiimportexportmanager importapiarchive apiimportexportmanager java at org carbon apimgt rest api admin impl importapiserviceimpl importapipost importapiserviceimpl java at org carbon apimgt rest api admin importapi importapipost importapi java at sun reflect invoke unknown source at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org apache cxf service invoker abstractinvoker performinvocation abstractinvoker java at org apache cxf service invoker abstractinvoker invoke abstractinvoker java at org apache cxf jaxrs jaxrsinvoker invoke jaxrsinvoker java at org apache cxf jaxrs jaxrsinvoker invoke jaxrsinvoker java at org apache cxf interceptor serviceinvokerinterceptor run serviceinvokerinterceptor java at org apache cxf interceptor serviceinvokerinterceptor handlemessage serviceinvokerinterceptor java at org apache cxf phase phaseinterceptorchain dointercept phaseinterceptorchain java at org apache cxf transport chaininitiationobserver onmessage chaininitiationobserver java at org apache cxf transport http abstracthttpdestination invoke abstracthttpdestination java at org apache cxf transport servlet servletcontroller invokedestination servletcontroller java at org apache cxf transport servlet servletcontroller invoke servletcontroller java at org apache cxf transport servlet servletcontroller invoke servletcontroller java at org apache cxf transport servlet cxfnonspringservlet invoke cxfnonspringservlet java at org apache cxf transport servlet abstracthttpservlet handlerequest abstracthttpservlet java at org apache cxf transport servlet abstracthttpservlet dopost abstracthttpservlet java at javax servlet http httpservlet service httpservlet java at org apache cxf transport servlet abstracthttpservlet service abstracthttpservlet java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache tomcat websocket server wsfilter dofilter wsfilter java at org apache catalina core applicationfilterchain internaldofilter applicationfilterchain java at org apache catalina core applicationfilterchain dofilter applicationfilterchain java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina authenticator authenticatorbase invoke authenticatorbase java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina valves errorreportvalve invoke errorreportvalve java at org carbon identity context rewrite valve tenantcontextrewritevalve invoke tenantcontextrewritevalve java at org carbon identity authz valve authorizationvalve invoke authorizationvalve java at org carbon identity auth valve authenticationvalve invoke authenticationvalve java at org carbon tomcat ext valves compositevalve continueinvocation compositevalve java at org carbon tomcat ext valves tomcatvalvecontainer invokevalves tomcatvalvecontainer java at org carbon tomcat ext valves compositevalve invoke compositevalve java at org carbon tomcat ext valves carbonstuckthreaddetectionvalve invoke carbonstuckthreaddetectionvalve java at org apache catalina valves abstractaccesslogvalve invoke abstractaccesslogvalve java at org carbon tomcat ext valves carboncontextcreatorvalve invoke carboncontextcreatorvalve java at org carbon tomcat ext valves requestcorrelationidvalve invoke requestcorrelationidvalve java at org apache catalina core standardenginevalve invoke standardenginevalve java at org apache catalina connector coyoteadapter service coyoteadapter java at org apache coyote service java at org apache coyote abstractprocessorlight process abstractprocessorlight java at org apache coyote abstractprotocol connectionhandler process abstractprotocol java at org apache tomcat util net nioendpoint socketprocessor dorun nioendpoint java at org apache tomcat util net socketprocessorbase run socketprocessorbase java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at org apache tomcat util threads taskthread wrappingrunnable run taskthread java at java lang thread run thread java caused by org carbon apimgt api apimanagementexception error while setting registry lifecycle checklist items for the api hello at org carbon apimgt impl abstractapimanager handleexception abstractapimanager java at org carbon apimgt impl abstractapimanager handleexception abstractapimanager java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java more caused by org carbon governance api exception governanceexception invalid check list item at org carbon governance api common dataobjects governanceartifactimpl islcitemchecked governanceartifactimpl java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl apiproviderimpl changeapilcchecklistitems apiproviderimpl java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl userawareapiprovider changeapilcchecklistitems userawareapiprovider java at org carbon apimgt impl importexport utils apiimportutil importapi apiimportutil java more
1
27,508
2,693,529,725
IssuesEvent
2015-04-01 15:12:52
IQSS/dataverse
https://api.github.com/repos/IQSS/dataverse
opened
Listing Dataverses Widget is Displaying Unexpected Datasets and Dataverses
Component: Search/Browse Component: UX & Upgrade Priority: High Status: Dev
I am seeing unexpected dataverse and dataset results when I use a the Listing Dataverses Widget for a specific Dataverse: **Example: Testing Widgets Dataverse** For this Dataverse I should only see the following 2 datasets listed: ![image](https://cloud.githubusercontent.com/assets/5013837/6943968/4367f9e2-d85f-11e4-9630-1b86a9d662a9.png) But for this script: ``` <script src="https://dvn-build.hmdc.harvard.edu/resources/js/widgets.js?alias=test-dv&amp;dvUrl=https://dvn-build.hmdc.harvard.edu&amp;widget=iframe&amp;heightPx=500"> ``` I am seeing other Dataverses and Datasets that dont belong with this Dataverse (even Drafts and unpublished datasets/dataverses, which I thought you would only see published things since its meant to be public) ![image](https://cloud.githubusercontent.com/assets/5013837/6944006/7a76c986-d85f-11e4-99f5-8eb202d0f56b.png)
1.0
Listing Dataverses Widget is Displaying Unexpected Datasets and Dataverses - I am seeing unexpected dataverse and dataset results when I use a the Listing Dataverses Widget for a specific Dataverse: **Example: Testing Widgets Dataverse** For this Dataverse I should only see the following 2 datasets listed: ![image](https://cloud.githubusercontent.com/assets/5013837/6943968/4367f9e2-d85f-11e4-9630-1b86a9d662a9.png) But for this script: ``` <script src="https://dvn-build.hmdc.harvard.edu/resources/js/widgets.js?alias=test-dv&amp;dvUrl=https://dvn-build.hmdc.harvard.edu&amp;widget=iframe&amp;heightPx=500"> ``` I am seeing other Dataverses and Datasets that dont belong with this Dataverse (even Drafts and unpublished datasets/dataverses, which I thought you would only see published things since its meant to be public) ![image](https://cloud.githubusercontent.com/assets/5013837/6944006/7a76c986-d85f-11e4-99f5-8eb202d0f56b.png)
priority
listing dataverses widget is displaying unexpected datasets and dataverses i am seeing unexpected dataverse and dataset results when i use a the listing dataverses widget for a specific dataverse example testing widgets dataverse for this dataverse i should only see the following datasets listed but for this script script src i am seeing other dataverses and datasets that dont belong with this dataverse even drafts and unpublished datasets dataverses which i thought you would only see published things since its meant to be public
1
388,728
11,491,646,646
IssuesEvent
2020-02-11 19:24:01
ClinGen/clincoded
https://api.github.com/repos/ClinGen/clincoded
closed
Transfer of GDMs from UNC BCC to Hereditary Cancer GCEP
EP request GCI curation blocker priority: high
Hi @wrightmw we are in need of transferring the following records from the UNC Biocuration Core affiliation to the Hereditary Cancer GCEP. This is a high priority need, as several of these GDMs are cardinal genes disease records that require publishing on the website. Also, there publishing would help to remove erroneous data previously published using the Gene Curation Express. The URLS for the records are as follows: 1) MEN1: Multiple Endocrine Neoplasia: https://curation.clinicalgenome.org/curation-central/?gdm=f4aa4e77-0d27-44a5-a92a-1af9fc4c7068&pmid=9103196 2) RET: multiple endocrine neoplasia type 2A: https://curation.clinicalgenome.org/curation-central/?gdm=438681cb-7912-41b7-b3d7-571db9dfe1ce&pmid=8103403 3) RET: multiple endocrine neoplasia type 2B: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368 4) TP53: Li Fraumeni: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368 Please also check with @xiluo if these are correct as well.
1.0
Transfer of GDMs from UNC BCC to Hereditary Cancer GCEP - Hi @wrightmw we are in need of transferring the following records from the UNC Biocuration Core affiliation to the Hereditary Cancer GCEP. This is a high priority need, as several of these GDMs are cardinal genes disease records that require publishing on the website. Also, there publishing would help to remove erroneous data previously published using the Gene Curation Express. The URLS for the records are as follows: 1) MEN1: Multiple Endocrine Neoplasia: https://curation.clinicalgenome.org/curation-central/?gdm=f4aa4e77-0d27-44a5-a92a-1af9fc4c7068&pmid=9103196 2) RET: multiple endocrine neoplasia type 2A: https://curation.clinicalgenome.org/curation-central/?gdm=438681cb-7912-41b7-b3d7-571db9dfe1ce&pmid=8103403 3) RET: multiple endocrine neoplasia type 2B: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368 4) TP53: Li Fraumeni: https://curation.clinicalgenome.org/curation-central/?gdm=45d58554-8fee-4204-b5e7-2117004c1c5f&pmid=15381368 Please also check with @xiluo if these are correct as well.
priority
transfer of gdms from unc bcc to hereditary cancer gcep hi wrightmw we are in need of transferring the following records from the unc biocuration core affiliation to the hereditary cancer gcep this is a high priority need as several of these gdms are cardinal genes disease records that require publishing on the website also there publishing would help to remove erroneous data previously published using the gene curation express the urls for the records are as follows multiple endocrine neoplasia ret multiple endocrine neoplasia type ret multiple endocrine neoplasia type li fraumeni please also check with xiluo if these are correct as well
1
29,180
2,714,007,142
IssuesEvent
2015-04-09 22:50:39
nickpaventi/culligan-diy
https://api.github.com/repos/nickpaventi/culligan-diy
opened
Product Category [Mobile]: Feature #1 and 2 need style updates
High Priority
**Feature 1** - [ ] Add consistent padding to left and right of feature content, currently touches left edge - [ ] Add padding between product name and 'perfect solution...' title **Feature 2** - [ ] Image and content should flow like Feature 1 - [ ] Colors may need to change for tags, content and CTA to work in this new flow Let's discuss real time if needed... ![features-mobile](https://cloud.githubusercontent.com/assets/10550484/7078666/285134e4-ded0-11e4-9eb3-64a5fecc0a27.png)
1.0
Product Category [Mobile]: Feature #1 and 2 need style updates - **Feature 1** - [ ] Add consistent padding to left and right of feature content, currently touches left edge - [ ] Add padding between product name and 'perfect solution...' title **Feature 2** - [ ] Image and content should flow like Feature 1 - [ ] Colors may need to change for tags, content and CTA to work in this new flow Let's discuss real time if needed... ![features-mobile](https://cloud.githubusercontent.com/assets/10550484/7078666/285134e4-ded0-11e4-9eb3-64a5fecc0a27.png)
priority
product category feature and need style updates feature add consistent padding to left and right of feature content currently touches left edge add padding between product name and perfect solution title feature image and content should flow like feature colors may need to change for tags content and cta to work in this new flow let s discuss real time if needed
1
598,267
18,241,681,087
IssuesEvent
2021-10-01 13:38:07
g1moon/hello-waterski
https://api.github.com/repos/g1moon/hello-waterski
opened
줄서기 등록 서비스
priority high
- 줄서기 시스템에 대기를 등록하는 모달을 만든다. - [ ] /line/:[spotId] 에 줄서기 버튼을 만들고, 클릭하면 등록하는 모달을 만든다. - [ ] 모달에는 ridingType, boatType을 입력 받는다. - [ ] 제출이 완료되면 data/line.json에 추가하고, allLineData에 새로운 정보를 추가한다. - [ ] 새롭게 oneSpot 데이터도 갱신한다(다시 그릴지, 상태에 추가만 할지 고민).
1.0
줄서기 등록 서비스 - - 줄서기 시스템에 대기를 등록하는 모달을 만든다. - [ ] /line/:[spotId] 에 줄서기 버튼을 만들고, 클릭하면 등록하는 모달을 만든다. - [ ] 모달에는 ridingType, boatType을 입력 받는다. - [ ] 제출이 완료되면 data/line.json에 추가하고, allLineData에 새로운 정보를 추가한다. - [ ] 새롭게 oneSpot 데이터도 갱신한다(다시 그릴지, 상태에 추가만 할지 고민).
priority
줄서기 등록 서비스 줄서기 시스템에 대기를 등록하는 모달을 만든다 line 에 줄서기 버튼을 만들고 클릭하면 등록하는 모달을 만든다 모달에는 ridingtype boattype을 입력 받는다 제출이 완료되면 data line json에 추가하고 alllinedata에 새로운 정보를 추가한다 새롭게 onespot 데이터도 갱신한다 다시 그릴지 상태에 추가만 할지 고민
1
234,351
7,720,017,744
IssuesEvent
2018-05-23 21:22:21
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
Reset sequence broken on nRF5x chips
bug priority: high
We seem to go on an infite loop during the reset sequence with the introduciton of this commit: 72c7ded5612870c20a3efc3f36a9688ff85a4418
1.0
Reset sequence broken on nRF5x chips - We seem to go on an infite loop during the reset sequence with the introduciton of this commit: 72c7ded5612870c20a3efc3f36a9688ff85a4418
priority
reset sequence broken on chips we seem to go on an infite loop during the reset sequence with the introduciton of this commit
1
384,181
11,384,842,486
IssuesEvent
2020-01-29 09:52:42
unitystation/unitystation
https://api.github.com/repos/unitystation/unitystation
opened
Comms Consoles now print out empty Centcomm reports
Bug High Priority
Just experienced now on build 3989 The reports look like this when read: ![image](https://user-images.githubusercontent.com/20813925/73346019-d9078180-42d0-11ea-9119-4130c9ce75e9.png)
1.0
Comms Consoles now print out empty Centcomm reports - Just experienced now on build 3989 The reports look like this when read: ![image](https://user-images.githubusercontent.com/20813925/73346019-d9078180-42d0-11ea-9119-4130c9ce75e9.png)
priority
comms consoles now print out empty centcomm reports just experienced now on build the reports look like this when read
1
718,440
24,717,273,830
IssuesEvent
2022-10-20 08:01:15
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
DHCP (IPv4) seemingly dies by trying to assign an IP of 0.0.0.0
bug priority: high area: Networking
**Describe the bug** This strange issue leads to an ethernet device on a network being completely dead until it is restarted. A device gets a fixed IP from a DHCP server for 20 seconds, the server is changed to give a different fixed IP and the zephyr board renews the IP. Seemingly at this time, the DHCP client seems to think it has been assigned the IP 0.0.0.0 for whatever reason, this IP clearly is not valid, and the DHCP client fails to remove/add it, it then stays in this state of constantly trying to add the invalid IP. Rebooting the module allows it to get a DHCP lease. **To Reproduce** Needs a DHCP server - I used dhcpd, with this configuration: ``` default-lease-time 20; max-lease-time 20; authoritative; server-name name; host zephyr { hardware ethernet 68:27:19:CF:09:DD; fixed-address 192.168.1.209; } subnet 192.168.1.0 netmask 255.255.255.0 { range 192.168.1.202 192.168.1.208; } ``` Then just run the server, load the zephyr DHCP sample on a board and let it get an IP. Once it gets an IP, change the fixed address of the zephyr device to something else, e.g. .210, restart the dhcpd server, then watch as the DHCP client falls over. **Expected behavior** DHCP to continue working **Impact** Showstopper, device is dead, complexity to denial of service a device is 0 **Logs and console output** Log: https://pastebin.com/WeHvD1pN **Environment (please complete the following information):** - OS: Linux - Toolchain: Zephyr SDK 0.15.0 - Commit SHA or Version used: 9018338863c0c5e7039d80df5f0d463075a0a31a
1.0
DHCP (IPv4) seemingly dies by trying to assign an IP of 0.0.0.0 - **Describe the bug** This strange issue leads to an ethernet device on a network being completely dead until it is restarted. A device gets a fixed IP from a DHCP server for 20 seconds, the server is changed to give a different fixed IP and the zephyr board renews the IP. Seemingly at this time, the DHCP client seems to think it has been assigned the IP 0.0.0.0 for whatever reason, this IP clearly is not valid, and the DHCP client fails to remove/add it, it then stays in this state of constantly trying to add the invalid IP. Rebooting the module allows it to get a DHCP lease. **To Reproduce** Needs a DHCP server - I used dhcpd, with this configuration: ``` default-lease-time 20; max-lease-time 20; authoritative; server-name name; host zephyr { hardware ethernet 68:27:19:CF:09:DD; fixed-address 192.168.1.209; } subnet 192.168.1.0 netmask 255.255.255.0 { range 192.168.1.202 192.168.1.208; } ``` Then just run the server, load the zephyr DHCP sample on a board and let it get an IP. Once it gets an IP, change the fixed address of the zephyr device to something else, e.g. .210, restart the dhcpd server, then watch as the DHCP client falls over. **Expected behavior** DHCP to continue working **Impact** Showstopper, device is dead, complexity to denial of service a device is 0 **Logs and console output** Log: https://pastebin.com/WeHvD1pN **Environment (please complete the following information):** - OS: Linux - Toolchain: Zephyr SDK 0.15.0 - Commit SHA or Version used: 9018338863c0c5e7039d80df5f0d463075a0a31a
priority
dhcp seemingly dies by trying to assign an ip of describe the bug this strange issue leads to an ethernet device on a network being completely dead until it is restarted a device gets a fixed ip from a dhcp server for seconds the server is changed to give a different fixed ip and the zephyr board renews the ip seemingly at this time the dhcp client seems to think it has been assigned the ip for whatever reason this ip clearly is not valid and the dhcp client fails to remove add it it then stays in this state of constantly trying to add the invalid ip rebooting the module allows it to get a dhcp lease to reproduce needs a dhcp server i used dhcpd with this configuration default lease time max lease time authoritative server name name host zephyr hardware ethernet cf dd fixed address subnet netmask range then just run the server load the zephyr dhcp sample on a board and let it get an ip once it gets an ip change the fixed address of the zephyr device to something else e g restart the dhcpd server then watch as the dhcp client falls over expected behavior dhcp to continue working impact showstopper device is dead complexity to denial of service a device is logs and console output log environment please complete the following information os linux toolchain zephyr sdk commit sha or version used
1
212,875
7,243,582,785
IssuesEvent
2018-02-14 12:17:34
jrantamaki/supertimemachine
https://api.github.com/repos/jrantamaki/supertimemachine
closed
Bug: Calculation of elapsed time is wrong
bug frontend priority: high
Used Duration does not work properly when timestamps are for different dates.
1.0
Bug: Calculation of elapsed time is wrong - Used Duration does not work properly when timestamps are for different dates.
priority
bug calculation of elapsed time is wrong used duration does not work properly when timestamps are for different dates
1
787,973
27,737,955,055
IssuesEvent
2023-03-15 12:33:47
ITISFoundation/osparc-simcore
https://api.github.com/repos/ITISFoundation/osparc-simcore
opened
Garbage collector not working on aws-prod
bug High Priority
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Garbage collector for some days only shows error, there are no more "regular" logs and garbage collection seems to not happen. The errors are of this kind: ``` WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886863c0>: '843fbe7b-2e50-56b3-9ad9-752de771bf21' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188e67ac0>: '41d7bcb2-af42-5104-b662-5c66e747bbf4' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57c40>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1885ce4c0>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '843fbe7b-2e50-56b3-9ad9-752de771bf21' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18879eb40>: '41d7bcb2-af42-5104-b662-5c66e747bbf4' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886073c0>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886372c0>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18872e940>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86' ``` The graylog queries that can be used to check if this happens are: - `container_name:/.*collector.*/ AND NOT "Error in"` --> Display all non-error loglines, referring to "real" garbage collection - `container_name:/.*collector.*/ AND "Error in"` --> Display the errors mentioned Further evidence of garbage collection not working is that in prometheus one can see a s4-lite service running for many days, to observe this use the PromQL querry: `container_memory_usage_bytes{image=~"^.*[.osparc.io].*/simcore/services/dynamic/s4l-core-lite.*$",name=~"dy-sidecar-b57f4e59-13d0-476d-9954-9855adf657b7.*"}` Comparison with the redis keys, that correspond to open browser-tabs or sessions, show that there was no session key for the user that owns the project containing this s4l for some days, so the garbage collector should have kicked in: `redis_key_value{key=~"^user_id=2:.*$"}` ### Expected Behavior Garbage collection works ### Steps To Reproduce The GC does not work on aws-prod ### Anything else? This affects production and may cause it to not run smooth if services accumulate. From my feelings, I would put this on high urgency.
1.0
Garbage collector not working on aws-prod - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Garbage collector for some days only shows error, there are no more "regular" logs and garbage collection seems to not happen. The errors are of this kind: ``` WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886863c0>: '843fbe7b-2e50-56b3-9ad9-752de771bf21' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188e67ac0>: '41d7bcb2-af42-5104-b662-5c66e747bbf4' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57c40>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26' WARNING: [2023-03-15 12:26:49,577/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1885ce4c0>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 1-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887d8ec0>: 'f2d26379-e6fc-50dd-956a-3f4f67d2542c' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 2-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa188b57e40>: '843fbe7b-2e50-56b3-9ad9-752de771bf21' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 3-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18879eb40>: '41d7bcb2-af42-5104-b662-5c66e747bbf4' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 4-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de8c0>: '67c34fc6-fa9f-5eaf-bc0d-8012117707cc' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 5-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1887de0c0>: 'b57f4e59-13d0-476d-9954-9855adf657b7' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 6-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886073c0>: 'fd123ae9-3242-5eb1-bf02-c04b942f2992' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 7-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa1886372c0>: '7e135c19-c89d-5081-bb90-d07ee9d3dc26' WARNING: [2023-03-15 12:27:21,783/MainProcess] [servicelib.utils:logged_gather(122)] - Error in 14-th concurrent task <coroutine object _remove_single_orphaned_service at 0x7fa18872e940>: '0c417ffb-8d03-4b68-9ead-dbef12a4af86' ``` The graylog queries that can be used to check if this happens are: - `container_name:/.*collector.*/ AND NOT "Error in"` --> Display all non-error loglines, referring to "real" garbage collection - `container_name:/.*collector.*/ AND "Error in"` --> Display the errors mentioned Further evidence of garbage collection not working is that in prometheus one can see a s4-lite service running for many days, to observe this use the PromQL querry: `container_memory_usage_bytes{image=~"^.*[.osparc.io].*/simcore/services/dynamic/s4l-core-lite.*$",name=~"dy-sidecar-b57f4e59-13d0-476d-9954-9855adf657b7.*"}` Comparison with the redis keys, that correspond to open browser-tabs or sessions, show that there was no session key for the user that owns the project containing this s4l for some days, so the garbage collector should have kicked in: `redis_key_value{key=~"^user_id=2:.*$"}` ### Expected Behavior Garbage collection works ### Steps To Reproduce The GC does not work on aws-prod ### Anything else? This affects production and may cause it to not run smooth if services accumulate. From my feelings, I would put this on high urgency.
priority
garbage collector not working on aws prod is there an existing issue for this i have searched the existing issues current behavior garbage collector for some days only shows error there are no more regular logs and garbage collection seems to not happen the errors are of this kind warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task warning error in th concurrent task the graylog queries that can be used to check if this happens are container name collector and not error in display all non error loglines referring to real garbage collection container name collector and error in display the errors mentioned further evidence of garbage collection not working is that in prometheus one can see a lite service running for many days to observe this use the promql querry container memory usage bytes image simcore services dynamic core lite name dy sidecar comparison with the redis keys that correspond to open browser tabs or sessions show that there was no session key for the user that owns the project containing this for some days so the garbage collector should have kicked in redis key value key user id expected behavior garbage collection works steps to reproduce the gc does not work on aws prod anything else this affects production and may cause it to not run smooth if services accumulate from my feelings i would put this on high urgency
1
699,236
24,009,797,449
IssuesEvent
2022-09-14 17:44:19
opendatahub-io/odh-dashboard
https://api.github.com/repos/opendatahub-io/odh-dashboard
closed
[Bug]: Stability Fixes
kind/bug feature/notebook-controller priority/high
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior There are a couple stability issues with the KFNBC right now. 1. If you do not have an existing PVC, you cannot spawn a notebook 2. If you are on a high loaded cluster, the endpoint call to create a Notebook can timeout with a 504 Gateway Timeout ### Expected Behavior 1. Fix the issue with the PVC, it needs to be created when we create a Notebook 2. Move the backend logic for route fetching out of the endpoint to create a Notebook -- this is the major slowdown, move it to the frontend and lazy fetch it to avoid slowness on cluster ### Steps To Reproduce For both issues - Create a Notebook 1. Should make PVC calls in the network when you create 2. The endpoint for Notebook create should not annotate the route onto the Notebook ### Workaround (if any) _No response_ ### OpenShift Infrastructure Version _No response_ ### Openshift Version _No response_ ### What browsers are you seeing the problem on? _No response_ ### Open Data Hub Version ```yml 2.2 ``` ### Relevant log output _No response_
1.0
[Bug]: Stability Fixes - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior There are a couple stability issues with the KFNBC right now. 1. If you do not have an existing PVC, you cannot spawn a notebook 2. If you are on a high loaded cluster, the endpoint call to create a Notebook can timeout with a 504 Gateway Timeout ### Expected Behavior 1. Fix the issue with the PVC, it needs to be created when we create a Notebook 2. Move the backend logic for route fetching out of the endpoint to create a Notebook -- this is the major slowdown, move it to the frontend and lazy fetch it to avoid slowness on cluster ### Steps To Reproduce For both issues - Create a Notebook 1. Should make PVC calls in the network when you create 2. The endpoint for Notebook create should not annotate the route onto the Notebook ### Workaround (if any) _No response_ ### OpenShift Infrastructure Version _No response_ ### Openshift Version _No response_ ### What browsers are you seeing the problem on? _No response_ ### Open Data Hub Version ```yml 2.2 ``` ### Relevant log output _No response_
priority
stability fixes is there an existing issue for this i have searched the existing issues current behavior there are a couple stability issues with the kfnbc right now if you do not have an existing pvc you cannot spawn a notebook if you are on a high loaded cluster the endpoint call to create a notebook can timeout with a gateway timeout expected behavior fix the issue with the pvc it needs to be created when we create a notebook move the backend logic for route fetching out of the endpoint to create a notebook this is the major slowdown move it to the frontend and lazy fetch it to avoid slowness on cluster steps to reproduce for both issues create a notebook should make pvc calls in the network when you create the endpoint for notebook create should not annotate the route onto the notebook workaround if any no response openshift infrastructure version no response openshift version no response what browsers are you seeing the problem on no response open data hub version yml relevant log output no response
1
773,850
27,173,997,780
IssuesEvent
2023-02-17 22:31:01
opendatahub-io/odh-dashboard
https://api.github.com/repos/opendatahub-io/odh-dashboard
opened
[A11y] Address A11y issues on Data Science Projects
kind/bug priority/high feature/dsg a11y
This ticket will look to address issues of an accessibility (a11y) nature in the Data Science Projects (DS Projects) section of the main nav. > Get the [Chrome Dev Tools Axe Extension](https://chrome.google.com/webstore/detail/axe-devtools-web-accessib/lhdoppojpmngadmnindnejefpokejbdd) > Running this extension on the page first loaded and as you use the page will help reveal what needs to be fixed ## Areas covered by this ticket Main nav: Data Science Projects > Note this covers all pages within' as well. See data setup for more information how to reveal the sub pages. Key areas to make sure to test: * Empty DS Projects page * 1-n projects added * Project details of one of the pages * Empty all sections (aka, new project) * Each section filled with data * Workbenches, Cluster Storage, Data Connections, Model Serving * Most sections have an expandable row for each row of the table -- more content shows up when you expand it * Each row should have a kebab -- those typically have followup actions * There are several modals; most are associated to the kebab actions or the secondary buttons along the page * Stopping a Workbench has a modal -- once a Workbench is created, there is a toggle switch in the row -- click off ## Data setup for testing * Creating a project is done from the main nav page -- you should have a button to create a project and it should turn the empty state into a table -- after 10 items, it becomes paginated * Inside a project * Creating a workbench => name can be anything, image selection is needed, typically the first one is fine, everything else should have a default * There are expandable items and hidden items, click on the various links and select things in the dropdowns -- you'll likely need to run the axe test on each stage for coverage * Create Cluster Storage => It's a modal, just need a name * A Workbench can have multiple storages on different paths -- a storage cannot be associated with multiple workbenches * A storage can be added to a workbench via the expanded workbench row -- it also requires the storage to be unassociated to any workbench (another modal) * Create Data Connection => The data in here is not important, but the name, id, and key will need to be filled out before it can be submitted * Data Connections can be connected to multiple workbenches, but workbenches cannot have more than one data connection * Models and model servers * Configuring a server is pretty straight forward, there are additional fields to expand using the checkboxes * Once configured (only one can be at this time), you'll be able to deploy several models * Models need a Data Connection & the details of this field is not important for axe testing
1.0
[A11y] Address A11y issues on Data Science Projects - This ticket will look to address issues of an accessibility (a11y) nature in the Data Science Projects (DS Projects) section of the main nav. > Get the [Chrome Dev Tools Axe Extension](https://chrome.google.com/webstore/detail/axe-devtools-web-accessib/lhdoppojpmngadmnindnejefpokejbdd) > Running this extension on the page first loaded and as you use the page will help reveal what needs to be fixed ## Areas covered by this ticket Main nav: Data Science Projects > Note this covers all pages within' as well. See data setup for more information how to reveal the sub pages. Key areas to make sure to test: * Empty DS Projects page * 1-n projects added * Project details of one of the pages * Empty all sections (aka, new project) * Each section filled with data * Workbenches, Cluster Storage, Data Connections, Model Serving * Most sections have an expandable row for each row of the table -- more content shows up when you expand it * Each row should have a kebab -- those typically have followup actions * There are several modals; most are associated to the kebab actions or the secondary buttons along the page * Stopping a Workbench has a modal -- once a Workbench is created, there is a toggle switch in the row -- click off ## Data setup for testing * Creating a project is done from the main nav page -- you should have a button to create a project and it should turn the empty state into a table -- after 10 items, it becomes paginated * Inside a project * Creating a workbench => name can be anything, image selection is needed, typically the first one is fine, everything else should have a default * There are expandable items and hidden items, click on the various links and select things in the dropdowns -- you'll likely need to run the axe test on each stage for coverage * Create Cluster Storage => It's a modal, just need a name * A Workbench can have multiple storages on different paths -- a storage cannot be associated with multiple workbenches * A storage can be added to a workbench via the expanded workbench row -- it also requires the storage to be unassociated to any workbench (another modal) * Create Data Connection => The data in here is not important, but the name, id, and key will need to be filled out before it can be submitted * Data Connections can be connected to multiple workbenches, but workbenches cannot have more than one data connection * Models and model servers * Configuring a server is pretty straight forward, there are additional fields to expand using the checkboxes * Once configured (only one can be at this time), you'll be able to deploy several models * Models need a Data Connection & the details of this field is not important for axe testing
priority
address issues on data science projects this ticket will look to address issues of an accessibility nature in the data science projects ds projects section of the main nav get the running this extension on the page first loaded and as you use the page will help reveal what needs to be fixed areas covered by this ticket main nav data science projects note this covers all pages within as well see data setup for more information how to reveal the sub pages key areas to make sure to test empty ds projects page n projects added project details of one of the pages empty all sections aka new project each section filled with data workbenches cluster storage data connections model serving most sections have an expandable row for each row of the table more content shows up when you expand it each row should have a kebab those typically have followup actions there are several modals most are associated to the kebab actions or the secondary buttons along the page stopping a workbench has a modal once a workbench is created there is a toggle switch in the row click off data setup for testing creating a project is done from the main nav page you should have a button to create a project and it should turn the empty state into a table after items it becomes paginated inside a project creating a workbench name can be anything image selection is needed typically the first one is fine everything else should have a default there are expandable items and hidden items click on the various links and select things in the dropdowns you ll likely need to run the axe test on each stage for coverage create cluster storage it s a modal just need a name a workbench can have multiple storages on different paths a storage cannot be associated with multiple workbenches a storage can be added to a workbench via the expanded workbench row it also requires the storage to be unassociated to any workbench another modal create data connection the data in here is not important but the name id and key will need to be filled out before it can be submitted data connections can be connected to multiple workbenches but workbenches cannot have more than one data connection models and model servers configuring a server is pretty straight forward there are additional fields to expand using the checkboxes once configured only one can be at this time you ll be able to deploy several models models need a data connection the details of this field is not important for axe testing
1
2,273
2,524,996,072
IssuesEvent
2015-01-20 21:29:55
graybeal/ont
https://api.github.com/repos/graybeal/ont
closed
feature to create ontology and use created ontology to replace existing version
1 star enhancement imported Milestone-Beta1 Priority-High watchdog
_From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on November 09, 2009 14:51:48_ What capability do you want added or improved? Shouldn't we add some capability to avoid the issue that we had with CF-- where we had an XML file of a vocabulary (not yet an ontology) and we needed, within the portal, to: a) develop a new ontology, and b) upload that new ontology to replace an existing ontology. In other words, we were trying to create ontology and then replace one with this new version. If I'm not mistaken, portal users can only create an ontology and then register it as a new ontology (not as an updated version) in the portal. Where do you want this capability to be accessible? from the voc2rdf component What sort of input/command mechanism do you want? button for registering as an updated version What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.) _Original issue: http://code.google.com/p/mmisw/issues/detail?id=217_
1.0
feature to create ontology and use created ontology to replace existing version - _From [steph_wa...@consolidated.net](https://code.google.com/u/112846428158176258467/) on November 09, 2009 14:51:48_ What capability do you want added or improved? Shouldn't we add some capability to avoid the issue that we had with CF-- where we had an XML file of a vocabulary (not yet an ontology) and we needed, within the portal, to: a) develop a new ontology, and b) upload that new ontology to replace an existing ontology. In other words, we were trying to create ontology and then replace one with this new version. If I'm not mistaken, portal users can only create an ontology and then register it as a new ontology (not as an updated version) in the portal. Where do you want this capability to be accessible? from the voc2rdf component What sort of input/command mechanism do you want? button for registering as an updated version What is the desired output (content, format, location)? Other details of your desired capability? What version of the product are you using? Please provide any additional information below (particular ontology/ies, text contents of vocabulary (voc2rdf), operating system, browser/version (Firefox, Safari, Chrome, IE, etc.), screenshot, etc.) _Original issue: http://code.google.com/p/mmisw/issues/detail?id=217_
priority
feature to create ontology and use created ontology to replace existing version from on november what capability do you want added or improved shouldn t we add some capability to avoid the issue that we had with cf where we had an xml file of a vocabulary not yet an ontology and we needed within the portal to a develop a new ontology and b upload that new ontology to replace an existing ontology in other words we were trying to create ontology and then replace one with this new version if i m not mistaken portal users can only create an ontology and then register it as a new ontology not as an updated version in the portal where do you want this capability to be accessible from the component what sort of input command mechanism do you want button for registering as an updated version what is the desired output content format location other details of your desired capability what version of the product are you using please provide any additional information below particular ontology ies text contents of vocabulary operating system browser version firefox safari chrome ie etc screenshot etc original issue
1
114,305
4,628,545,978
IssuesEvent
2016-09-28 05:14:32
Aplokodika/MachineLearning
https://api.github.com/repos/Aplokodika/MachineLearning
closed
Need to add proper mechanisms for initializing the learning-rate and momentum values.
High priority task To Do
As of now, there isn't a proper mechanism for initializing the learning rate and the momentum. In this architecture, each neuron contains its own learning rate and momentum.
1.0
Need to add proper mechanisms for initializing the learning-rate and momentum values. - As of now, there isn't a proper mechanism for initializing the learning rate and the momentum. In this architecture, each neuron contains its own learning rate and momentum.
priority
need to add proper mechanisms for initializing the learning rate and momentum values as of now there isn t a proper mechanism for initializing the learning rate and the momentum in this architecture each neuron contains its own learning rate and momentum
1
673,377
22,960,184,719
IssuesEvent
2022-07-19 14:47:16
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Cannot start the server with installServer.bat in windows
Priority/Highest Severity/Major bug Affected-6.0.0 QA-Reported JDK-17
**How to reproduce:** 1. Get pack from https://wso2.org/jenkins/job/products/job/product-is/4152/ 2. Follow running the server windows as a service https://is.docs.wso2.com/en/latest/setup/installing-as-a-windows-service/ 3. Set carbon_home and JAVA_HOME as env variables 4. Copy the IS-pack wrapper.conf to YAJSW13.03/conf. wrapper.conf file was edited as per the comment in https://github.com/wso2/product-is/issues/14148#issuecomment-1178955108 5. Navigate to YAJSW13.03-home/bat 6. Run `runConsole.bat ` 9. Server will start successfully Was able to get the server started successfully with the above command. When try to start the server with below commands was unable to get it started successfully. Reporting with a less severity as server can be started with step 6 command. 10. Then try the other option of running windows as a service with below commands `installService.bat ` `startService.bat ` `stopService.bat Server is not starting succesfully. Did we try these commands on windows 10 after we did the fixes for earlier fixed issues on windows. @Thumimku @ThumulaPerera Attaching used wrapper.conf [wrapper.zip](https://github.com/wso2/product-is/files/9120483/wrapper.zip) **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** H2 default windows 11 YAJSW 13.03
1.0
Cannot start the server with installServer.bat in windows - **How to reproduce:** 1. Get pack from https://wso2.org/jenkins/job/products/job/product-is/4152/ 2. Follow running the server windows as a service https://is.docs.wso2.com/en/latest/setup/installing-as-a-windows-service/ 3. Set carbon_home and JAVA_HOME as env variables 4. Copy the IS-pack wrapper.conf to YAJSW13.03/conf. wrapper.conf file was edited as per the comment in https://github.com/wso2/product-is/issues/14148#issuecomment-1178955108 5. Navigate to YAJSW13.03-home/bat 6. Run `runConsole.bat ` 9. Server will start successfully Was able to get the server started successfully with the above command. When try to start the server with below commands was unable to get it started successfully. Reporting with a less severity as server can be started with step 6 command. 10. Then try the other option of running windows as a service with below commands `installService.bat ` `startService.bat ` `stopService.bat Server is not starting succesfully. Did we try these commands on windows 10 after we did the fixes for earlier fixed issues on windows. @Thumimku @ThumulaPerera Attaching used wrapper.conf [wrapper.zip](https://github.com/wso2/product-is/files/9120483/wrapper.zip) **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** H2 default windows 11 YAJSW 13.03
priority
cannot start the server with installserver bat in windows how to reproduce get pack from follow running the server windows as a service set carbon home and java home as env variables copy the is pack wrapper conf to conf wrapper conf file was edited as per the comment in navigate to home bat run runconsole bat server will start successfully was able to get the server started successfully with the above command when try to start the server with below commands was unable to get it started successfully reporting with a less severity as server can be started with step command then try the other option of running windows as a service with below commands installservice bat startservice bat stopservice bat server is not starting succesfully did we try these commands on windows after we did the fixes for earlier fixed issues on windows thumimku thumulaperera attaching used wrapper conf environment information please complete the following information remove any unnecessary fields default windows yajsw
1
219,382
7,341,693,376
IssuesEvent
2018-03-07 03:30:34
hackoregon/civic-devops
https://api.github.com/repos/hackoregon/civic-devops
closed
Get AWS accounts for Tony and Khuong
Priority: high
Tony and Khuong require accounts to be created in the AWS environment.
1.0
Get AWS accounts for Tony and Khuong - Tony and Khuong require accounts to be created in the AWS environment.
priority
get aws accounts for tony and khuong tony and khuong require accounts to be created in the aws environment
1
179,527
6,625,938,498
IssuesEvent
2017-09-22 17:25:49
RepreZen/KaiZen-OpenAPI-Editor
https://api.github.com/repos/RepreZen/KaiZen-OpenAPI-Editor
closed
Code assist should support typeahead with dynamic filtering
High Priority Pending Review
The code assist feature in other Eclipse editors supports two useful behaviors that we should add to SwagEdit: 1. After invoking code assist, you can start typing, and the list of suggestions will remain visible, filtered to show the suggestions that start with the substring you entered. 2. If you invoke code assist with one or more characters already typed, immediately to the left of the insertion point, and then invoke code assist, the suggestion list will be pre-filtered to show matching suggestions.
1.0
Code assist should support typeahead with dynamic filtering - The code assist feature in other Eclipse editors supports two useful behaviors that we should add to SwagEdit: 1. After invoking code assist, you can start typing, and the list of suggestions will remain visible, filtered to show the suggestions that start with the substring you entered. 2. If you invoke code assist with one or more characters already typed, immediately to the left of the insertion point, and then invoke code assist, the suggestion list will be pre-filtered to show matching suggestions.
priority
code assist should support typeahead with dynamic filtering the code assist feature in other eclipse editors supports two useful behaviors that we should add to swagedit after invoking code assist you can start typing and the list of suggestions will remain visible filtered to show the suggestions that start with the substring you entered if you invoke code assist with one or more characters already typed immediately to the left of the insertion point and then invoke code assist the suggestion list will be pre filtered to show matching suggestions
1
131,139
5,143,788,456
IssuesEvent
2017-01-12 16:51:02
Esri/visibility-addin-dotnet
https://api.github.com/repos/Esri/visibility-addin-dotnet
closed
Arcmap and Pro LLOS, and RLOS does not detect point outside of an elevation Mosaic dataset extent that is not rectangular.
2 - In Progress B - Bug priority - high
@jmccausland @dhagens @topowright @mfunk In Arcmap and Pro 1.3, RLOS does not detect user entered point outside of an elevation Mosaic Data Set extent that is not rectangular in an area that would be in the minumum bounding rectangle. ![image](https://cloud.githubusercontent.com/assets/6842145/16739361/b48f0566-4767-11e6-9d69-e05b0713322b.png)
1.0
Arcmap and Pro LLOS, and RLOS does not detect point outside of an elevation Mosaic dataset extent that is not rectangular. - @jmccausland @dhagens @topowright @mfunk In Arcmap and Pro 1.3, RLOS does not detect user entered point outside of an elevation Mosaic Data Set extent that is not rectangular in an area that would be in the minumum bounding rectangle. ![image](https://cloud.githubusercontent.com/assets/6842145/16739361/b48f0566-4767-11e6-9d69-e05b0713322b.png)
priority
arcmap and pro llos and rlos does not detect point outside of an elevation mosaic dataset extent that is not rectangular jmccausland dhagens topowright mfunk in arcmap and pro rlos does not detect user entered point outside of an elevation mosaic data set extent that is not rectangular in an area that would be in the minumum bounding rectangle
1
367,117
10,840,704,210
IssuesEvent
2019-11-12 08:56:49
ProtonMail/proton-contacts
https://api.github.com/repos/ProtonMail/proton-contacts
closed
V4 Beta feedback - something happened to the name field on create/edit contact
Priority: high bug
The text input is misaligned with the label, misaligned with the other text inputs and smaller size than the other text inputs. <img width="703" alt="Capture d’écran 2019-11-11 à 15 12 13" src="https://user-images.githubusercontent.com/40168570/68594557-190c5480-0498-11ea-90c5-96a61ef46f65.png">
1.0
V4 Beta feedback - something happened to the name field on create/edit contact - The text input is misaligned with the label, misaligned with the other text inputs and smaller size than the other text inputs. <img width="703" alt="Capture d’écran 2019-11-11 à 15 12 13" src="https://user-images.githubusercontent.com/40168570/68594557-190c5480-0498-11ea-90c5-96a61ef46f65.png">
priority
beta feedback something happened to the name field on create edit contact the text input is misaligned with the label misaligned with the other text inputs and smaller size than the other text inputs img width alt capture d’écran à src
1
618,737
19,486,122,001
IssuesEvent
2021-12-26 12:08:57
covidpass-org/covidpass
https://api.github.com/repos/covidpass-org/covidpass
closed
Not working with third dose health pass
bug priority:high
Hello, It seems that it does not work with a 3rd dose health pass. I have the message : Invalid QR code (for French health pass, Safari on iOS 15). Thanks.
1.0
Not working with third dose health pass - Hello, It seems that it does not work with a 3rd dose health pass. I have the message : Invalid QR code (for French health pass, Safari on iOS 15). Thanks.
priority
not working with third dose health pass hello it seems that it does not work with a dose health pass i have the message invalid qr code for french health pass safari on ios thanks
1
473,914
13,649,309,760
IssuesEvent
2020-09-26 13:51:09
ahmedkaludi/accelerated-mobile-pages
https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages
closed
Internal feedback Improvement for Longtime Users
NEXT UPDATE [Priority: HIGH] enhancement
Internal feedback Improvement for Longtime Users
1.0
Internal feedback Improvement for Longtime Users - Internal feedback Improvement for Longtime Users
priority
internal feedback improvement for longtime users internal feedback improvement for longtime users
1
684,428
23,417,823,865
IssuesEvent
2022-08-13 08:01:12
IntelCompH2020/topicmodeler
https://api.github.com/repos/IntelCompH2020/topicmodeler
opened
Study stability / coherence of topics versus training set size
High Priority Experimental Work / Validation
We have a few huge corpus, in the order of tens or millions of documents. Training is costly. The question here is: Do we really need to train with the whole corpus ? Are topics much better than if training with say a maximum of 2 M documents ??? This should be studied because if no improvement is gained when training with very large corpora we could sample the training set, and then carry out inference on the whole set when calculating the indicators.
1.0
Study stability / coherence of topics versus training set size - We have a few huge corpus, in the order of tens or millions of documents. Training is costly. The question here is: Do we really need to train with the whole corpus ? Are topics much better than if training with say a maximum of 2 M documents ??? This should be studied because if no improvement is gained when training with very large corpora we could sample the training set, and then carry out inference on the whole set when calculating the indicators.
priority
study stability coherence of topics versus training set size we have a few huge corpus in the order of tens or millions of documents training is costly the question here is do we really need to train with the whole corpus are topics much better than if training with say a maximum of m documents this should be studied because if no improvement is gained when training with very large corpora we could sample the training set and then carry out inference on the whole set when calculating the indicators
1
154,253
5,916,657,908
IssuesEvent
2017-05-22 11:07:45
DOAJ/doaj
https://api.github.com/repos/DOAJ/doaj
closed
Libraries IP information
feedback high priority
Dear CL Do we have this info? where can I get it? Hello. I am the Director at the Tom Green County Library and we have had a change in IP address. Please provide the information you currently have on file for our IPs. Thank you.
1.0
Libraries IP information - Dear CL Do we have this info? where can I get it? Hello. I am the Director at the Tom Green County Library and we have had a change in IP address. Please provide the information you currently have on file for our IPs. Thank you.
priority
libraries ip information dear cl do we have this info where can i get it hello i am the director at the tom green county library and we have had a change in ip address please provide the information you currently have on file for our ips thank you
1
464,576
13,327,231,136
IssuesEvent
2020-08-27 12:53:29
zeebe-io/zeebe
https://api.github.com/repos/zeebe-io/zeebe
closed
Do not log RESOURCE_EXHAUSTED exceptions in job worker (GO)
Priority: High Scope: clients/go Status: Needs Review Type: Maintenance
**Description** Related to https://github.com/zeebe-io/zeebe/issues/5187 we have seen recently an huge increase of log statements, which are related to `RESOURCE_EXHAUSTED` errors. The job worker in the java or go client, log errors when they happen during activating. These kind of "errors" are not really errors and are handled gracefully internally by the job worker, via retry so it is not necessary to log them. The user can't do much in this case. The log statements spread more confusion and seem to be not really helpful.
1.0
Do not log RESOURCE_EXHAUSTED exceptions in job worker (GO) - **Description** Related to https://github.com/zeebe-io/zeebe/issues/5187 we have seen recently an huge increase of log statements, which are related to `RESOURCE_EXHAUSTED` errors. The job worker in the java or go client, log errors when they happen during activating. These kind of "errors" are not really errors and are handled gracefully internally by the job worker, via retry so it is not necessary to log them. The user can't do much in this case. The log statements spread more confusion and seem to be not really helpful.
priority
do not log resource exhausted exceptions in job worker go description related to we have seen recently an huge increase of log statements which are related to resource exhausted errors the job worker in the java or go client log errors when they happen during activating these kind of errors are not really errors and are handled gracefully internally by the job worker via retry so it is not necessary to log them the user can t do much in this case the log statements spread more confusion and seem to be not really helpful
1
526,950
15,305,412,669
IssuesEvent
2021-02-24 18:05:16
ansible/awx
https://api.github.com/repos/ansible/awx
opened
Prevent users from selecting job templates that are missing a project or inventory when creating workflow node
component:ui priority:high state:in_progress type:feature
##### ISSUE TYPE - Feature Idea ##### SUMMARY In the old UI we would disable rows that were missing a project/inv and didn't prompt for inventory: <img width="1673" alt="Screen Shot 2021-02-24 at 1 00 46 PM" src="https://user-images.githubusercontent.com/9889020/109044677-ad57f600-76a0-11eb-90cc-b89735c86d58.png"> We should do something similar in the new UI since selecting a JT like this will result in the API returning an error when we attempt to save the workflow. It would be nice if we could also add a tooltip or something like that to the row explaining why it can't be selected. For existing nodes, the row will be selected on edit but it will still be disabled: <img width="1674" alt="Screen Shot 2021-02-24 at 1 05 00 PM" src="https://user-images.githubusercontent.com/9889020/109044872-ebedb080-76a0-11eb-959e-b8d9c037e91f.png">
1.0
Prevent users from selecting job templates that are missing a project or inventory when creating workflow node - ##### ISSUE TYPE - Feature Idea ##### SUMMARY In the old UI we would disable rows that were missing a project/inv and didn't prompt for inventory: <img width="1673" alt="Screen Shot 2021-02-24 at 1 00 46 PM" src="https://user-images.githubusercontent.com/9889020/109044677-ad57f600-76a0-11eb-90cc-b89735c86d58.png"> We should do something similar in the new UI since selecting a JT like this will result in the API returning an error when we attempt to save the workflow. It would be nice if we could also add a tooltip or something like that to the row explaining why it can't be selected. For existing nodes, the row will be selected on edit but it will still be disabled: <img width="1674" alt="Screen Shot 2021-02-24 at 1 05 00 PM" src="https://user-images.githubusercontent.com/9889020/109044872-ebedb080-76a0-11eb-959e-b8d9c037e91f.png">
priority
prevent users from selecting job templates that are missing a project or inventory when creating workflow node issue type feature idea summary in the old ui we would disable rows that were missing a project inv and didn t prompt for inventory img width alt screen shot at pm src we should do something similar in the new ui since selecting a jt like this will result in the api returning an error when we attempt to save the workflow it would be nice if we could also add a tooltip or something like that to the row explaining why it can t be selected for existing nodes the row will be selected on edit but it will still be disabled img width alt screen shot at pm src
1
712,357
24,492,468,892
IssuesEvent
2022-10-10 04:35:04
IAmTamal/Milan
https://api.github.com/repos/IAmTamal/Milan
closed
Join us part still shown after logging in.
🟧 priority: high 🕹 aspect: interface 🛠 goal: fix 🛠 status : under development hacktoberfest
### Description The join us part with login and register option is still shown to user which is already logged in. This should not happen, ### Screenshots ![Screenshot from 2022-10-07 19-12-43](https://user-images.githubusercontent.com/88667329/194568518-691ecb21-5b2a-4205-985c-740f3462406f.png) ### Additional information _No response_ ### 🥦 Browser Mozilla Firefox ### 👀 Have you checked if this issue has been raised before? - [X] I checked and didn't find similar issue ### 🏢 Have you read the Contributing Guidelines? - [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md) ### Are you willing to work on this issue ? Yes I am willing to submit a PR!
1.0
Join us part still shown after logging in. - ### Description The join us part with login and register option is still shown to user which is already logged in. This should not happen, ### Screenshots ![Screenshot from 2022-10-07 19-12-43](https://user-images.githubusercontent.com/88667329/194568518-691ecb21-5b2a-4205-985c-740f3462406f.png) ### Additional information _No response_ ### 🥦 Browser Mozilla Firefox ### 👀 Have you checked if this issue has been raised before? - [X] I checked and didn't find similar issue ### 🏢 Have you read the Contributing Guidelines? - [X] I have read the [Contributing Guidelines](https://github.com/IAmTamal/Milan/blob/main/CONTRIBUTING.md) ### Are you willing to work on this issue ? Yes I am willing to submit a PR!
priority
join us part still shown after logging in description the join us part with login and register option is still shown to user which is already logged in this should not happen screenshots additional information no response 🥦 browser mozilla firefox 👀 have you checked if this issue has been raised before i checked and didn t find similar issue 🏢 have you read the contributing guidelines i have read the are you willing to work on this issue yes i am willing to submit a pr
1
664,477
22,272,264,418
IssuesEvent
2022-06-10 13:28:27
opencrvs/opencrvs-core
https://api.github.com/repos/opencrvs/opencrvs-core
closed
Certificate Config: Add data tag to images
👹Bug Priority: high
**Describe the bug** - difficult to work out correct order for images so that the signature properly displays - suggest to use a data tag for images. So that the signature is appropriately tagged and doesn't matter where it appears in the code.
1.0
Certificate Config: Add data tag to images - **Describe the bug** - difficult to work out correct order for images so that the signature properly displays - suggest to use a data tag for images. So that the signature is appropriately tagged and doesn't matter where it appears in the code.
priority
certificate config add data tag to images describe the bug difficult to work out correct order for images so that the signature properly displays suggest to use a data tag for images so that the signature is appropriately tagged and doesn t matter where it appears in the code
1
133,817
5,215,316,115
IssuesEvent
2017-01-26 04:11:53
imrogues/angularjs
https://api.github.com/repos/imrogues/angularjs
opened
Watching Collections
[priority] high [status] accepted [type] feature
### Description Add the third and final dirty–checking mechanism to our implementation of `Scope`: shallow collection–watching. The `$watchCollection` function is not simple, but that’s mostly because it provides an important, non-trivial facility: We can watch for changes in large arrays and objects much more efficiently than we could with just deep–watching. --- ### Issue Checklist - [ ] How `$watchCollection` can be used with arrays, objects, and other values. - [ ] What `$watchCollection` does with arrays. - [ ] What `$watchCollection` does with objects. - [ ] Array–like objects and their role in `$watchCollection`. All issues in milestone: [2 Scopes](https://github.com/imrogues/angularjs/milestone/2) --- ### Assignees - [ ] Final assign @imrogues
1.0
Watching Collections - ### Description Add the third and final dirty–checking mechanism to our implementation of `Scope`: shallow collection–watching. The `$watchCollection` function is not simple, but that’s mostly because it provides an important, non-trivial facility: We can watch for changes in large arrays and objects much more efficiently than we could with just deep–watching. --- ### Issue Checklist - [ ] How `$watchCollection` can be used with arrays, objects, and other values. - [ ] What `$watchCollection` does with arrays. - [ ] What `$watchCollection` does with objects. - [ ] Array–like objects and their role in `$watchCollection`. All issues in milestone: [2 Scopes](https://github.com/imrogues/angularjs/milestone/2) --- ### Assignees - [ ] Final assign @imrogues
priority
watching collections description add the third and final dirty–checking mechanism to our implementation of scope shallow collection–watching the watchcollection function is not simple but that’s mostly because it provides an important non trivial facility we can watch for changes in large arrays and objects much more efficiently than we could with just deep–watching issue checklist how watchcollection can be used with arrays objects and other values what watchcollection does with arrays what watchcollection does with objects array–like objects and their role in watchcollection all issues in milestone assignees final assign imrogues
1
523,593
15,185,961,645
IssuesEvent
2021-02-15 11:40:33
percipioglobal/craft
https://api.github.com/repos/percipioglobal/craft
closed
Too much "searchable" fields.
performance 🏃‍♀️ priority: high
In the current field setup, too many fields are still "searchable" eg. the image optimise fields
1.0
Too much "searchable" fields. - In the current field setup, too many fields are still "searchable" eg. the image optimise fields
priority
too much searchable fields in the current field setup too many fields are still searchable eg the image optimise fields
1
695,166
23,847,344,636
IssuesEvent
2022-09-06 14:56:45
WordPress/openverse-frontend
https://api.github.com/repos/WordPress/openverse-frontend
closed
The link to report at the source is incorrectly linking to the direct URL
good first issue help wanted 🟧 priority: high 🛠 goal: fix 🕹 aspect: interface
## Description <!-- Concisely describe the bug. Compare your experience with what you expected to happen. --> <!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." --> After submitting a report for an item on Openverse, the user is encourage to submit the same report at the source of the media item. The link should be to the `foreign_landing_url` of the image which has the "Report it" button, not the direct URL (that opens the original image). ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> <img width="436" alt="Screen Shot 2022-09-05 at 4 22 24 PM" src="https://user-images.githubusercontent.com/15233243/188460269-d9471bab-bb02-403d-b111-9396ce3fa7d6.png"> ## Additional context <!-- Add any other context about the problem here; or delete the section entirely. --> This line should have `media.foreign_landing_url` instead of `media.url`: https://github.com/WordPress/openverse-frontend/blob/bf0c72e5ea660354e20ffa33a4caf0abd65a5db9/src/components/VContentReport/VContentReportForm.vue#L13 ## Resolution <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in resolving this bug.
1.0
The link to report at the source is incorrectly linking to the direct URL - ## Description <!-- Concisely describe the bug. Compare your experience with what you expected to happen. --> <!-- For example: "I clicked the 'submit' button and instead of seeing a thank you message, I saw a blank page." --> After submitting a report for an item on Openverse, the user is encourage to submit the same report at the source of the media item. The link should be to the `foreign_landing_url` of the image which has the "Report it" button, not the direct URL (that opens the original image). ## Screenshots <!-- Add screenshots to show the problem; or delete the section entirely. --> <img width="436" alt="Screen Shot 2022-09-05 at 4 22 24 PM" src="https://user-images.githubusercontent.com/15233243/188460269-d9471bab-bb02-403d-b111-9396ce3fa7d6.png"> ## Additional context <!-- Add any other context about the problem here; or delete the section entirely. --> This line should have `media.foreign_landing_url` instead of `media.url`: https://github.com/WordPress/openverse-frontend/blob/bf0c72e5ea660354e20ffa33a4caf0abd65a5db9/src/components/VContentReport/VContentReportForm.vue#L13 ## Resolution <!-- Replace the [ ] with [x] to check the box. --> - [ ] 🙋 I would be interested in resolving this bug.
priority
the link to report at the source is incorrectly linking to the direct url description after submitting a report for an item on openverse the user is encourage to submit the same report at the source of the media item the link should be to the foreign landing url of the image which has the report it button not the direct url that opens the original image screenshots img width alt screen shot at pm src additional context this line should have media foreign landing url instead of media url resolution 🙋 i would be interested in resolving this bug
1
677,841
23,177,467,416
IssuesEvent
2022-07-31 16:29:18
projectdiscovery/httpx
https://api.github.com/repos/projectdiscovery/httpx
closed
Ports Flag is not working with scheme
Priority: High Status: Completed Type: Bug
<!-- 1. Please search to see if an issue already exists for the bug you encountered. 2. For support requests, FAQs or "How to" questions, please use the GitHub Discussions section instead - https://github.com/projectdiscovery/httpx/discussions or 3. Join our discord server at https://discord.gg/projectdiscovery and post the question on the #httpx channel. --> <!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> ### httpx version: 1.2.3 <!-- You can find current version of httpx with "httpx -version" --> <!-- We only accept issues that are reproducible on the latest version of httpx. --> <!-- You can find the latest version of project at https://github.com/projectdiscovery/httpx/releases/ --> ### Current Behavior: When the ports flag is used within scheme (-p https:443) it is breaking. <!-- A concise description of what you're experiencing. --> ### Expected Behavior: It should probe https on port 443. <!-- A concise description of what you expected to happen. --> ### Steps To Reproduce: <!-- Example: steps to reproduce the behavior: 1. Run 'echo example.com | httpx -ports https:443' --> 1. Run 'echo example.com | httpx -ports https:443' ### Anything else: Response: '[FTL] Could not cast port to integer from your value: https:443. Resulting error: strconv.Atoi: parsing "https:443": invalid syntax.' <!-- Links? References? Screnshots? Anything that will give us more context about the issue that you are encountering! -->
1.0
Ports Flag is not working with scheme - <!-- 1. Please search to see if an issue already exists for the bug you encountered. 2. For support requests, FAQs or "How to" questions, please use the GitHub Discussions section instead - https://github.com/projectdiscovery/httpx/discussions or 3. Join our discord server at https://discord.gg/projectdiscovery and post the question on the #httpx channel. --> <!-- ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION. --> ### httpx version: 1.2.3 <!-- You can find current version of httpx with "httpx -version" --> <!-- We only accept issues that are reproducible on the latest version of httpx. --> <!-- You can find the latest version of project at https://github.com/projectdiscovery/httpx/releases/ --> ### Current Behavior: When the ports flag is used within scheme (-p https:443) it is breaking. <!-- A concise description of what you're experiencing. --> ### Expected Behavior: It should probe https on port 443. <!-- A concise description of what you expected to happen. --> ### Steps To Reproduce: <!-- Example: steps to reproduce the behavior: 1. Run 'echo example.com | httpx -ports https:443' --> 1. Run 'echo example.com | httpx -ports https:443' ### Anything else: Response: '[FTL] Could not cast port to integer from your value: https:443. Resulting error: strconv.Atoi: parsing "https:443": invalid syntax.' <!-- Links? References? Screnshots? Anything that will give us more context about the issue that you are encountering! -->
priority
ports flag is not working with scheme please search to see if an issue already exists for the bug you encountered for support requests faqs or how to questions please use the github discussions section instead or join our discord server at and post the question on the httpx channel httpx version current behavior when the ports flag is used within scheme p https it is breaking expected behavior it should probe https on port steps to reproduce example steps to reproduce the behavior run echo example com httpx ports https run echo example com httpx ports https anything else response could not cast port to integer from your value https resulting error strconv atoi parsing https invalid syntax
1
529,036
15,379,318,099
IssuesEvent
2021-03-02 19:30:06
cch5ng/job_tracker
https://api.github.com/repos/cch5ng/job_tracker
closed
form validation
enhancement high priority
definitely need FE form validation should have min required fields (Job, Event) * this should affect content being saved for Alerts and also the form submit click handler then should sync up with BE form validation
1.0
form validation - definitely need FE form validation should have min required fields (Job, Event) * this should affect content being saved for Alerts and also the form submit click handler then should sync up with BE form validation
priority
form validation definitely need fe form validation should have min required fields job event this should affect content being saved for alerts and also the form submit click handler then should sync up with be form validation
1
468,844
13,491,652,572
IssuesEvent
2020-09-11 16:50:10
bloom-housing/bloom
https://api.github.com/repos/bloom-housing/bloom
opened
Jones: Remove Preference Section from Eligibility if no Preferences
High Priority
Since there are no preferences this section of eligibility shouldn't display. ![image](https://user-images.githubusercontent.com/13915/92951958-0e05b600-f414-11ea-84aa-76ad2a822318.png)
1.0
Jones: Remove Preference Section from Eligibility if no Preferences - Since there are no preferences this section of eligibility shouldn't display. ![image](https://user-images.githubusercontent.com/13915/92951958-0e05b600-f414-11ea-84aa-76ad2a822318.png)
priority
jones remove preference section from eligibility if no preferences since there are no preferences this section of eligibility shouldn t display
1
761,515
26,684,365,984
IssuesEvent
2023-01-26 20:30:20
gamefreedomgit/Maelstrom
https://api.github.com/repos/gamefreedomgit/Maelstrom
closed
[Paladin][Ret] Guardian Of Ancient Kings / Judgements of the Bold / Long Arm of the Law issue
Class: Paladin Pet Spell Priority: High Status: Confirmed
Two more bugs with Ret Paladin: - Guardian of Ancient Kings has an assist function, but not an attack function - essentially, he just stands around looking pretty. - Judgements of the Bold and Long Arm of the Law do not proc on Absorb effects.
1.0
[Paladin][Ret] Guardian Of Ancient Kings / Judgements of the Bold / Long Arm of the Law issue - Two more bugs with Ret Paladin: - Guardian of Ancient Kings has an assist function, but not an attack function - essentially, he just stands around looking pretty. - Judgements of the Bold and Long Arm of the Law do not proc on Absorb effects.
priority
guardian of ancient kings judgements of the bold long arm of the law issue two more bugs with ret paladin guardian of ancient kings has an assist function but not an attack function essentially he just stands around looking pretty judgements of the bold and long arm of the law do not proc on absorb effects
1
180,174
6,647,048,086
IssuesEvent
2017-09-28 01:23:15
CrowdRescueHQ/CrowdRescue
https://api.github.com/repos/CrowdRescueHQ/CrowdRescue
closed
Incident Updates/Suggested Edits
discussion priority-high
We need to fill the need of having a "suggested edits" queue, where one group of volunteers can fill out an update form and other volunteers can process it. This could work like suggested edits or it could be a form which feeds a queue. Both would be similar, but the latter would require a new table. We need to get more feedback about this. Each edit should get logged and possible have an attached note or reason. (Could be built through case notes, could be separate).
1.0
Incident Updates/Suggested Edits - We need to fill the need of having a "suggested edits" queue, where one group of volunteers can fill out an update form and other volunteers can process it. This could work like suggested edits or it could be a form which feeds a queue. Both would be similar, but the latter would require a new table. We need to get more feedback about this. Each edit should get logged and possible have an attached note or reason. (Could be built through case notes, could be separate).
priority
incident updates suggested edits we need to fill the need of having a suggested edits queue where one group of volunteers can fill out an update form and other volunteers can process it this could work like suggested edits or it could be a form which feeds a queue both would be similar but the latter would require a new table we need to get more feedback about this each edit should get logged and possible have an attached note or reason could be built through case notes could be separate
1
818,287
30,681,878,093
IssuesEvent
2023-07-26 09:41:22
bryntum/support
https://api.github.com/repos/bryntum/support
closed
[VUE] An exception when use workingTime config in calendar timeline view
bug resolved high-priority premium forum large-account OEM
[Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=25269&p=125948#p125948) Not able to repro this in vanilla example. Please read forum topic, `workingTime` that throws an error depends on current time. See app attached below https://github.com/bryntum/support/assets/7203098/aff90109-b9fe-4033-b7f7-f5138c104b64 [bryntum-calendar-vite.zip](https://github.com/bryntum/support/files/11738789/bryntum-calendar-vite.zip)
1.0
[VUE] An exception when use workingTime config in calendar timeline view - [Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=25269&p=125948#p125948) Not able to repro this in vanilla example. Please read forum topic, `workingTime` that throws an error depends on current time. See app attached below https://github.com/bryntum/support/assets/7203098/aff90109-b9fe-4033-b7f7-f5138c104b64 [bryntum-calendar-vite.zip](https://github.com/bryntum/support/files/11738789/bryntum-calendar-vite.zip)
priority
an exception when use workingtime config in calendar timeline view not able to repro this in vanilla example please read forum topic workingtime that throws an error depends on current time see app attached below
1
269,970
8,444,984,599
IssuesEvent
2018-10-18 20:05:17
semperfiwebdesign/all-in-one-seo-pack
https://api.github.com/repos/semperfiwebdesign/all-in-one-seo-pack
closed
noindex/nofollow for no SEO on post type only works together
Bug Priority | High
For #1797 and #1943 we made it so noindex/nofollow works when SEO for that post type is disabled, however this only works if you're using both noindex and nofollow. If you're trying to use just one it doesn't work. To make the logic easier, it may be best to pull this out into its own function(s). Eventually it should be in its own file, rather than junking up that big class file.
1.0
noindex/nofollow for no SEO on post type only works together - For #1797 and #1943 we made it so noindex/nofollow works when SEO for that post type is disabled, however this only works if you're using both noindex and nofollow. If you're trying to use just one it doesn't work. To make the logic easier, it may be best to pull this out into its own function(s). Eventually it should be in its own file, rather than junking up that big class file.
priority
noindex nofollow for no seo on post type only works together for and we made it so noindex nofollow works when seo for that post type is disabled however this only works if you re using both noindex and nofollow if you re trying to use just one it doesn t work to make the logic easier it may be best to pull this out into its own function s eventually it should be in its own file rather than junking up that big class file
1
688,433
23,581,834,381
IssuesEvent
2022-08-23 08:20:12
Public-Health-Scotland/source-linkage-files
https://api.github.com/repos/Public-Health-Scotland/source-linkage-files
closed
Bug with `calculate_stay`
bug Priority: High
There is an issue with `calculate_stay`. If a `sc_qtr` is provided it will always ignore the `end_date`. For example, the below should both return 1. ```R > calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02")) [1] 1 > calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02"), "2019Q4") [1] 91 ``` This was an oversight, as these cases aren't tested for either. Tagging @Jennit07 as you wrote the function, please assign yourself if you're working on it, otherwise I will probably come back to it soon!
1.0
Bug with `calculate_stay` - There is an issue with `calculate_stay`. If a `sc_qtr` is provided it will always ignore the `end_date`. For example, the below should both return 1. ```R > calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02")) [1] 1 > calculate_stay("1920", as.Date("2020-01-01"), as.Date("2020-01-02"), "2019Q4") [1] 91 ``` This was an oversight, as these cases aren't tested for either. Tagging @Jennit07 as you wrote the function, please assign yourself if you're working on it, otherwise I will probably come back to it soon!
priority
bug with calculate stay there is an issue with calculate stay if a sc qtr is provided it will always ignore the end date for example the below should both return r calculate stay as date as date calculate stay as date as date this was an oversight as these cases aren t tested for either tagging as you wrote the function please assign yourself if you re working on it otherwise i will probably come back to it soon
1
93,485
3,900,865,156
IssuesEvent
2016-04-18 08:28:20
DanGrew/JenkinsTestTracker
https://api.github.com/repos/DanGrew/JenkinsTestTracker
opened
Window resize moves divider
High Priority
The divider seems to be reset to 50% when resized or dragged to another screen.
1.0
Window resize moves divider - The divider seems to be reset to 50% when resized or dragged to another screen.
priority
window resize moves divider the divider seems to be reset to when resized or dragged to another screen
1
200,745
7,011,252,618
IssuesEvent
2017-12-20 04:36:50
PATRIC3/patric3_website
https://api.github.com/repos/PATRIC3/patric3_website
closed
ID Mapping: showing unwanted results
High Priority Tool: ID Mapping
I tried to search using a set of refseq locus tags below to find corresponding PATRIC ids. BB0457 BB0836 BB0839 BB0344 BB0467 The result table show multiple matches for each of the ids, mapping to PATRIC and refseq features, and CDS and non CDS features, like shown in the attached image. This is very confusing for users for several reasons. 1. When I ask for a match to patric ids, the results should not show anything that doesn't have PATRIC ids. Solution: : match for only annotation = PATRIC or PATRIC ID not null. 2. The table shows matches to multiple annotation sources and feature types. However, those attributes are not shown in the table. Making it difficult to understand what they are. Solution: make the id mapping result table look similar to the feature table + one additional column for user provided ids. It will be good to have the facet filter above the table as well, to provide easy filtering. <img width="1126" alt="screen shot 2017-12-07 at 6 35 46 am" src="https://user-images.githubusercontent.com/1236463/33716197-e1289bbc-db1a-11e7-925b-1731363fe552.png">
1.0
ID Mapping: showing unwanted results - I tried to search using a set of refseq locus tags below to find corresponding PATRIC ids. BB0457 BB0836 BB0839 BB0344 BB0467 The result table show multiple matches for each of the ids, mapping to PATRIC and refseq features, and CDS and non CDS features, like shown in the attached image. This is very confusing for users for several reasons. 1. When I ask for a match to patric ids, the results should not show anything that doesn't have PATRIC ids. Solution: : match for only annotation = PATRIC or PATRIC ID not null. 2. The table shows matches to multiple annotation sources and feature types. However, those attributes are not shown in the table. Making it difficult to understand what they are. Solution: make the id mapping result table look similar to the feature table + one additional column for user provided ids. It will be good to have the facet filter above the table as well, to provide easy filtering. <img width="1126" alt="screen shot 2017-12-07 at 6 35 46 am" src="https://user-images.githubusercontent.com/1236463/33716197-e1289bbc-db1a-11e7-925b-1731363fe552.png">
priority
id mapping showing unwanted results i tried to search using a set of refseq locus tags below to find corresponding patric ids the result table show multiple matches for each of the ids mapping to patric and refseq features and cds and non cds features like shown in the attached image this is very confusing for users for several reasons when i ask for a match to patric ids the results should not show anything that doesn t have patric ids solution match for only annotation patric or patric id not null the table shows matches to multiple annotation sources and feature types however those attributes are not shown in the table making it difficult to understand what they are solution make the id mapping result table look similar to the feature table one additional column for user provided ids it will be good to have the facet filter above the table as well to provide easy filtering img width alt screen shot at am src
1
764,695
26,812,466,222
IssuesEvent
2023-02-01 23:56:41
zulip/zulip
https://api.github.com/repos/zulip/zulip
opened
Add setup tips to user invite modal
area: onboarding priority: high area: invitations
As [discussed on CZO](https://chat.zulip.org/#narrow/stream/101-design/topic/setup.20tips.20in.20user.20invite.20modal/near/1490259), we should add some setup tips to the user invite modal. General notes: - Each of these would only be shown if the user has the permissions needed to take the relevant action. - Include links to the appropriate settings panels throughout (noted below without their URLs). ## At the top of the modal ### Org profile configuration - If org description is missing: "You may want to [configure]() your organization's login page prior to inviting users." - If there is a description, but the org profile picture is missing: "You may want to [upload a profile picture]() for your organization prior to inviting users." ### User configs - If default new user settings have not been changed, and custom profile fields have not been added: "You may want to configure [default new user settings]() and [custom profile fields]() prior to inviting users. If only one of the above holds, just show the relevant part of the warning: - You may want to configure [default new user settings]() prior to inviting users. - You may want to configure [custom profile fields]() prior to inviting users. ## Just above the "Streams they should join" section - If no streams have been created: "You may want to create streams prior to inviting users." --- ## Blockers: - [ ] #23466 (?)
1.0
Add setup tips to user invite modal - As [discussed on CZO](https://chat.zulip.org/#narrow/stream/101-design/topic/setup.20tips.20in.20user.20invite.20modal/near/1490259), we should add some setup tips to the user invite modal. General notes: - Each of these would only be shown if the user has the permissions needed to take the relevant action. - Include links to the appropriate settings panels throughout (noted below without their URLs). ## At the top of the modal ### Org profile configuration - If org description is missing: "You may want to [configure]() your organization's login page prior to inviting users." - If there is a description, but the org profile picture is missing: "You may want to [upload a profile picture]() for your organization prior to inviting users." ### User configs - If default new user settings have not been changed, and custom profile fields have not been added: "You may want to configure [default new user settings]() and [custom profile fields]() prior to inviting users. If only one of the above holds, just show the relevant part of the warning: - You may want to configure [default new user settings]() prior to inviting users. - You may want to configure [custom profile fields]() prior to inviting users. ## Just above the "Streams they should join" section - If no streams have been created: "You may want to create streams prior to inviting users." --- ## Blockers: - [ ] #23466 (?)
priority
add setup tips to user invite modal as we should add some setup tips to the user invite modal general notes each of these would only be shown if the user has the permissions needed to take the relevant action include links to the appropriate settings panels throughout noted below without their urls at the top of the modal org profile configuration if org description is missing you may want to your organization s login page prior to inviting users if there is a description but the org profile picture is missing you may want to for your organization prior to inviting users user configs if default new user settings have not been changed and custom profile fields have not been added you may want to configure and prior to inviting users if only one of the above holds just show the relevant part of the warning you may want to configure prior to inviting users you may want to configure prior to inviting users just above the streams they should join section if no streams have been created you may want to create streams prior to inviting users blockers
1
230,653
7,612,618,402
IssuesEvent
2018-05-01 18:12:32
AmatCoder/mednaffe
https://api.github.com/repos/AmatCoder/mednaffe
closed
Mednaffe won't accept up/right D-Pad inputs in controller setup
Priority-High bug
Using Mednaffe 0.8.7, trying to setup my controller for the PlayStation core. I'm using a DualShock 4 with DS4Windows, however this issue also occurs if I turn DS4Win off and use the DS4's native DirectInput. When configuring my inputs, I can configure up/down/left/right just fine: ![image](https://user-images.githubusercontent.com/9293417/39260256-db01e64a-48b0-11e8-8395-e684602fced4.png) However, if I change tabs to 'Port 2', close the input config window, or do _essentially anything else_ that involves leaving the 'Port 1' input config screen, the config changes itself to this: ![image](https://user-images.githubusercontent.com/9293417/39260315-fcfbdaf8-48b0-11e8-8168-c62a9514830a.png) Sadly, playing games with only two directions available can be a _tad_ difficult ;P. Any idea what could be causing this?
1.0
Mednaffe won't accept up/right D-Pad inputs in controller setup - Using Mednaffe 0.8.7, trying to setup my controller for the PlayStation core. I'm using a DualShock 4 with DS4Windows, however this issue also occurs if I turn DS4Win off and use the DS4's native DirectInput. When configuring my inputs, I can configure up/down/left/right just fine: ![image](https://user-images.githubusercontent.com/9293417/39260256-db01e64a-48b0-11e8-8395-e684602fced4.png) However, if I change tabs to 'Port 2', close the input config window, or do _essentially anything else_ that involves leaving the 'Port 1' input config screen, the config changes itself to this: ![image](https://user-images.githubusercontent.com/9293417/39260315-fcfbdaf8-48b0-11e8-8168-c62a9514830a.png) Sadly, playing games with only two directions available can be a _tad_ difficult ;P. Any idea what could be causing this?
priority
mednaffe won t accept up right d pad inputs in controller setup using mednaffe trying to setup my controller for the playstation core i m using a dualshock with however this issue also occurs if i turn off and use the s native directinput when configuring my inputs i can configure up down left right just fine however if i change tabs to port close the input config window or do essentially anything else that involves leaving the port input config screen the config changes itself to this sadly playing games with only two directions available can be a tad difficult p any idea what could be causing this
1
279,858
8,674,601,015
IssuesEvent
2018-11-30 08:16:31
FundacionParaguaya/MentorApp
https://api.github.com/repos/FundacionParaguaya/MentorApp
closed
Touch Responsiveness on Buttons (Globally through the app)
bug high priority question
The touch area of buttons is not responsive to a large enough area to be "usable". This can be most clearly seen on the hamburger icon and on the skip question functionality. We are visibly seeing users struggling when interacting with these buttons. Please can you review and enlarge the area responsive to touch (should this be the issue)
1.0
Touch Responsiveness on Buttons (Globally through the app) - The touch area of buttons is not responsive to a large enough area to be "usable". This can be most clearly seen on the hamburger icon and on the skip question functionality. We are visibly seeing users struggling when interacting with these buttons. Please can you review and enlarge the area responsive to touch (should this be the issue)
priority
touch responsiveness on buttons globally through the app the touch area of buttons is not responsive to a large enough area to be usable this can be most clearly seen on the hamburger icon and on the skip question functionality we are visibly seeing users struggling when interacting with these buttons please can you review and enlarge the area responsive to touch should this be the issue
1
52,591
3,024,302,084
IssuesEvent
2015-08-02 13:35:50
MinetestForFun/minetest-minetestforfun-server
https://api.github.com/repos/MinetestForFun/minetest-minetestforfun-server
opened
Too many boats
Modding ➤ Mapgen Priority: High
The sea_modpack include boats and uboots, they are too many, it's a problem.... And it makes the mapgen heavier ! - [ ] Divided per 4 or 5 the boats/uboots spawn rate
1.0
Too many boats - The sea_modpack include boats and uboots, they are too many, it's a problem.... And it makes the mapgen heavier ! - [ ] Divided per 4 or 5 the boats/uboots spawn rate
priority
too many boats the sea modpack include boats and uboots they are too many it s a problem and it makes the mapgen heavier divided per or the boats uboots spawn rate
1
533,143
15,577,645,533
IssuesEvent
2021-03-17 13:45:41
AY2021S2-CS2103T-W12-3/tp
https://api.github.com/repos/AY2021S2-CS2103T-W12-3/tp
closed
As an insurance agent, I can add, edit and view cruicial personal and insurance information of each client,
priority.High type.Epic
... so that I can more easily keep track of my clients and their needs.
1.0
As an insurance agent, I can add, edit and view cruicial personal and insurance information of each client, - ... so that I can more easily keep track of my clients and their needs.
priority
as an insurance agent i can add edit and view cruicial personal and insurance information of each client so that i can more easily keep track of my clients and their needs
1
127,295
5,028,194,746
IssuesEvent
2016-12-15 17:29:12
odalic/sti
https://api.github.com/repos/odalic/sti
opened
UI - results - Labels for concept classifying columns not depicted?
bug priority: High
_From @tomas-knap on December 13, 2016 19:51_ When I run Odalic agains ADEQUATe KB, in case of disambiguation, UI properly shows labels (see below, URI contains UUID, but there is nice label). But in case of classificiation, label is not obtained and UUID is used. Why? ![screen shot 2016-12-13 at 20 41 07](https://cloud.githubusercontent.com/assets/3014917/21156307/b291a880-c175-11e6-9df9-a70d2f2ae7d0.png) Solution: It should also take the label of the concept being suggested for classification. @KataBoku did you encounter that as well? _Copied from original issue: odalic/odalic-ui#190_
1.0
UI - results - Labels for concept classifying columns not depicted? - _From @tomas-knap on December 13, 2016 19:51_ When I run Odalic agains ADEQUATe KB, in case of disambiguation, UI properly shows labels (see below, URI contains UUID, but there is nice label). But in case of classificiation, label is not obtained and UUID is used. Why? ![screen shot 2016-12-13 at 20 41 07](https://cloud.githubusercontent.com/assets/3014917/21156307/b291a880-c175-11e6-9df9-a70d2f2ae7d0.png) Solution: It should also take the label of the concept being suggested for classification. @KataBoku did you encounter that as well? _Copied from original issue: odalic/odalic-ui#190_
priority
ui results labels for concept classifying columns not depicted from tomas knap on december when i run odalic agains adequate kb in case of disambiguation ui properly shows labels see below uri contains uuid but there is nice label but in case of classificiation label is not obtained and uuid is used why solution it should also take the label of the concept being suggested for classification kataboku did you encounter that as well copied from original issue odalic odalic ui
1
447,746
12,892,683,034
IssuesEvent
2020-07-13 20:06:17
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
Improve virtualenv-clone hashing logic
area: production area: tooling bug in progress priority: high
Our `virtualenv` management logic is designed to have a few properties: * After running `provision` or the equivalent production operation, you always have a correct virtualenv as though you'd built it fresh * We avoid the performance pain of redownloading and/or rebuilding all of our packages fresh for the common operations of "adding a new package" or "upgrading a minor version of a package", both in the development environment and in production. * https://zulip.readthedocs.io/en/latest/subsystems/dependencies.html is a useful reference. There are a few issues we need to fix with the current implementation: * We don't currently compare the `Python` version when deciding whether we need to build a new virtualenv from scratch. * Our index of packages included in the virtualenv doesn't have package versions; we should include those and only allow package upgrades to use the `virtualenv-clone` plus cache approach (downgrades are often not tested upstream). We should be able to test whether the implementation is working manually through inspection of `/srv/zulip-venv-cache/` before and after adding a new python dependency (see the docs for how to do this using `update-locked-requirements`), downgrading in `dev.txt`, and upgrading/downgrading the Python version itself. We don't currently have unit tests of any this logic, though it wouldn't be a bad idea to write some simple tests of the logic functions from `scripts/lib/setup_venv.py` in `tools/tests/` somewhere (`tools/test-tools` is the runner for those). @andersk FYI; we might want to fix this before we merge https://github.com/zulip/zulip/pull/12837.
1.0
Improve virtualenv-clone hashing logic - Our `virtualenv` management logic is designed to have a few properties: * After running `provision` or the equivalent production operation, you always have a correct virtualenv as though you'd built it fresh * We avoid the performance pain of redownloading and/or rebuilding all of our packages fresh for the common operations of "adding a new package" or "upgrading a minor version of a package", both in the development environment and in production. * https://zulip.readthedocs.io/en/latest/subsystems/dependencies.html is a useful reference. There are a few issues we need to fix with the current implementation: * We don't currently compare the `Python` version when deciding whether we need to build a new virtualenv from scratch. * Our index of packages included in the virtualenv doesn't have package versions; we should include those and only allow package upgrades to use the `virtualenv-clone` plus cache approach (downgrades are often not tested upstream). We should be able to test whether the implementation is working manually through inspection of `/srv/zulip-venv-cache/` before and after adding a new python dependency (see the docs for how to do this using `update-locked-requirements`), downgrading in `dev.txt`, and upgrading/downgrading the Python version itself. We don't currently have unit tests of any this logic, though it wouldn't be a bad idea to write some simple tests of the logic functions from `scripts/lib/setup_venv.py` in `tools/tests/` somewhere (`tools/test-tools` is the runner for those). @andersk FYI; we might want to fix this before we merge https://github.com/zulip/zulip/pull/12837.
priority
improve virtualenv clone hashing logic our virtualenv management logic is designed to have a few properties after running provision or the equivalent production operation you always have a correct virtualenv as though you d built it fresh we avoid the performance pain of redownloading and or rebuilding all of our packages fresh for the common operations of adding a new package or upgrading a minor version of a package both in the development environment and in production is a useful reference there are a few issues we need to fix with the current implementation we don t currently compare the python version when deciding whether we need to build a new virtualenv from scratch our index of packages included in the virtualenv doesn t have package versions we should include those and only allow package upgrades to use the virtualenv clone plus cache approach downgrades are often not tested upstream we should be able to test whether the implementation is working manually through inspection of srv zulip venv cache before and after adding a new python dependency see the docs for how to do this using update locked requirements downgrading in dev txt and upgrading downgrading the python version itself we don t currently have unit tests of any this logic though it wouldn t be a bad idea to write some simple tests of the logic functions from scripts lib setup venv py in tools tests somewhere tools test tools is the runner for those andersk fyi we might want to fix this before we merge
1
21,188
2,636,478,661
IssuesEvent
2015-03-10 04:01:07
cs2103jan2015-w09-1j/main
https://api.github.com/repos/cs2103jan2015-w09-1j/main
closed
A user can create floating tasks
priority.high type.story type.task
... so that the user can schedule them if he/she needs. add -t{task} -des{task description}
1.0
A user can create floating tasks - ... so that the user can schedule them if he/she needs. add -t{task} -des{task description}
priority
a user can create floating tasks so that the user can schedule them if he she needs add t task des task description
1
142,818
5,477,616,537
IssuesEvent
2017-03-12 10:17:45
fossasia/open-event-orga-server
https://api.github.com/repos/fossasia/open-event-orga-server
closed
Schedule Calendar View: Show calendar only 15 minutes after last session
bug enhancement Priority: High
As the session info is not showing over the surrounding areas, please reduce the time of the schedule shown after the last session. Only show the schedule until "15 minutes after the last session".
1.0
Schedule Calendar View: Show calendar only 15 minutes after last session - As the session info is not showing over the surrounding areas, please reduce the time of the schedule shown after the last session. Only show the schedule until "15 minutes after the last session".
priority
schedule calendar view show calendar only minutes after last session as the session info is not showing over the surrounding areas please reduce the time of the schedule shown after the last session only show the schedule until minutes after the last session
1
359,986
10,683,166,116
IssuesEvent
2019-10-22 07:43:45
xpressengine/xpressengine
https://api.github.com/repos/xpressengine/xpressengine
closed
사이트 기본 설정을 변경 했을 때 로그인 설정이 변경되는 문제
priority/high type/bug
사이트 기본 설정을 변경하면 로그인 관련 설정이 초기화 돼서 로그인 불가 상태가 되고 로그인 관련 설정이 초기값으로 지정되고 있음
1.0
사이트 기본 설정을 변경 했을 때 로그인 설정이 변경되는 문제 - 사이트 기본 설정을 변경하면 로그인 관련 설정이 초기화 돼서 로그인 불가 상태가 되고 로그인 관련 설정이 초기값으로 지정되고 있음
priority
사이트 기본 설정을 변경 했을 때 로그인 설정이 변경되는 문제 사이트 기본 설정을 변경하면 로그인 관련 설정이 초기화 돼서 로그인 불가 상태가 되고 로그인 관련 설정이 초기값으로 지정되고 있음
1
601,788
18,432,318,728
IssuesEvent
2021-10-14 09:05:31
haskell/cabal
https://api.github.com/repos/haskell/cabal
closed
`cabal install` rebuilds all dependencies after a `cabal build`
type: bug priority: high :fire: cabal-install: cmd/install
**Describe the bug** `cabal install` rebuilds all dependencies after a `cabal build` in windows **To Reproduce** Steps to reproduce the behavior: * A simple cabal project with only one dependency `say` * First we build the project using a fresh store dir and print the store contents and the cabal-plan fingerprint ``` D:\ws\haskell\cabal-test>cabal clean D:\ws\haskell\cabal-test>rmdir sr /s /q D:\ws\haskell\cabal-test>cabal --store-dir .\sr build Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (first run) - cabal-test-0.1.0.0 (exe:cabal-test) (first run) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. ..................... Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-13304\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4\lib Configuring library for cabal-test-0.1.0.0.. Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. ............ Configuring executable 'cabal-test' for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. ....................... Linking D:\ws\haskell\cabal-test\dist-newstyle\build\x86_64-windows\ghc-8.6.5\cabal-test-0.1.0.0\x\cabal-test\build\cabal-test\cabal-test.exe ... D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b incoming package.db say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4 D:\ws\haskell\cabal-test>cabal-plan fingerprint ________________________________________________________________ B array-0.5.3.0 ________________________________________________________________ B base-4.12.0.0 ________________________________________________________________ B binary-0.8.6.0 ________________________________________________________________ B bytestring-0.10.8.2 ________________________________________________________________ L cabal-test-0.1.0.0 ________________________________________________________________ B containers-0.6.0.1 ________________________________________________________________ B deepseq-1.4.4.0 ________________________________________________________________ B ghc-prim-0.5.3 ________________________________________________________________ B integer-gmp-1.0.2.0 ________________________________________________________________ B rts-1.0 f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1 ________________________________________________________________ B text-1.2.3.1 ________________________________________________________________ B transformers-0.5.6.2 ``` * Then we install the project using the same store dir * Another bug: you can set a relative dir for the store using `cabal build` but no using `cabal install` ``` D:\ws\haskell\cabal-test>cabal --store-dir .\sr install --builddir dist2 Wrote tarball sdist to D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (requires build) - cabal-test-0.1.0.0 (exe:cabal-test) (requires build) Configuring library for say-0.1.0.1.. cabal-3.6.2.0.exe: expected an absolute directory name for --prefix: .\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 cabal-3.6.2.0.exe: Failed to build say-0.1.0.1 (which is required by exe:cabal-test from cabal-test-0.1.0.0). The failure occurred during the configure step. D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install - -builddir dist2 Wrote tarball sdist to D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (requires build) - cabal-test-0.1.0.0 (exe:cabal-test) (requires build) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. ................ Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5 e3ea47\lib Configuring library for cabal-test-0.1.0.0.. Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. .............. Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88\lib Configuring executable 'cabal-test' for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. ................... Linking dist\build\cabal-test\cabal-test.exe ... Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9 d05e08950e4e5de83d054a50961\bin Warning: The directory D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin is not in the system search path. Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe' D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961 cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88 incoming package.db say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4 say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=dist2 ________________________________________________________________ B array-0.5.3.0 ________________________________________________________________ B base-4.12.0.0 ________________________________________________________________ B binary-0.8.6.0 ________________________________________________________________ B bytestring-0.10.8.2 ________________________________________________________________ L cabal-test-0.1.0.0 ________________________________________________________________ B containers-0.6.0.1 ________________________________________________________________ B deepseq-1.4.4.0 ________________________________________________________________ B ghc-prim-0.5.3 ________________________________________________________________ B integer-gmp-1.0.2.0 ________________________________________________________________ B rts-1.0 f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1 ________________________________________________________________ B text-1.2.3.1 ________________________________________________________________ B transformers-0.5.6.2 D:\ws\haskell\cabal-test>cabal-plan diff --builddir=dist-newstyle --builddir=dist2 (No output, lovely) ``` * As you can check `cabal-plan diff` does not output anything, from other tests it seems it is broken in windows, including in the msys2 shell * After deleting the store we do the same in reverse order, this time preserving the temp build dir used by `cabal install` * There is no flag to preserve the directory, we have to copy the tmp dir quickly :facepalm: ``` D:\ws\haskell\cabal-test> rmdir sr /s /q D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install -builddir dist2 Wrote tarball sdist to D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (requires build) - cabal-test-0.1.0.0 (exe:cabal-test) (requires build) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. ............................ Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7de3ea47\lib Configuring library for cabal-test-0.1.0.0.. Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. .................. Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce9bbdaaf168f88\lib Configuring executable 'cabal-test' for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. .................. Linking dist\build\cabal-test\cabal-test.exe ... Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incomng\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfedd05e08950e4e5de83d054a50961\bin Warning: The directory D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\r\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin is not in the system search path. Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe' cabal-3.6.2.0.exe: Path 'D:\cabal\bin\cabal-test' already exists. Use --overwrite-policy=always to overwrite. D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961 cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88 incoming package.db say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 ``` * the fingerprint for `say` is the same in the temp build dir ``` D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=tmp2\cabal-install.-1 0896\dist2 ________________________________________________________________ B array-0.5.3.0 ________________________________________________________________ B base-4.12.0.0 ________________________________________________________________ B binary-0.8.6.0 ________________________________________________________________ B bytestring-0.10.8.2 14aa96bfe9d34856cf85a09bc4b314aac137e5e325f3f4beeb77b70ebc57a4a6 G cabal-test-0.1.0.0 ________________________________________________________________ B containers-0.6.0.1 ________________________________________________________________ B deepseq-1.4.4.0 ________________________________________________________________ B ghc-prim-0.5.3 ________________________________________________________________ B integer-gmp-1.0.2.0 ________________________________________________________________ B rts-1.0 f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1 ________________________________________________________________ B text-1.2.3.1 ________________________________________________________________ B transformers-0.5.6.2 ``` * we do the build after the install, deps are rebuilt again: ``` D:\ws\haskell\cabal-test>cabal --store-dir .\sr build Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (dependency rebuilt) - cabal-test-0.1.0.0 (exe:cabal-test) (dependency rebuilt) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. [1 of 2] Compiling Paths_say ( dist\build\autogen\Paths_say.hs, dist\buil d\Paths_say.o ) [2 of 2] Compiling Say ( src\Say.hs, dist\build\Say.o ) Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-9200\ws \haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c 596e4\lib Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961 cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88 incoming package.db say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4 say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 ``` * output of `cabal info` for all the three build dirs involved: <details> <summary> cabal info </summary> ``` # cabal-plan info --builddir=dist-newstyle Tree ~~~~ cabal-test-0.1.0.0 [cabal-test-0.1.0.0:exe:"cabal-test"] ├─ base-4.12.0.0 │ ├─ ghc-prim-0.5.3 │ │ └─ rts-1.0 │ ├─ integer-gmp-1.0.2.0 │ │ └─ ghc-prim-0.5.3 ┄┄ │ └─ rts-1.0 ┄┄ ├─ bytestring-0.10.8.2 │ ├─ base-4.12.0.0 ┄┄ │ ├─ deepseq-1.4.4.0 │ │ ├─ array-0.5.3.0 │ │ │ └─ base-4.12.0.0 ┄┄ │ │ └─ base-4.12.0.0 ┄┄ │ ├─ ghc-prim-0.5.3 ┄┄ │ └─ integer-gmp-1.0.2.0 ┄┄ └─ cabal-test-0.1.0.0 ├─ base-4.12.0.0 ┄┄ ├─ bytestring-0.10.8.2 ┄┄ ├─ say-0.1.0.1 │ ├─ base-4.12.0.0 ┄┄ │ ├─ bytestring-0.10.8.2 ┄┄ │ ├─ text-1.2.3.1 │ │ ├─ array-0.5.3.0 ┄┄ │ │ ├─ base-4.12.0.0 ┄┄ │ │ ├─ binary-0.8.6.0 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ │ └─ containers-0.6.0.1 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ │ └─ ghc-prim-0.5.3 ┄┄ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ ├─ ghc-prim-0.5.3 ┄┄ │ │ └─ integer-gmp-1.0.2.0 ┄┄ │ └─ transformers-0.5.6.2 │ └─ base-4.12.0.0 ┄┄ └─ text-1.2.3.1 ┄┄ Top-sorted ~~~~~~~~~~ UnitId "rts" UnitId "ghc-prim-0.5.3" UnitId "integer-gmp-1.0.2.0" UnitId "base-4.12.0.0" UnitId "array-0.5.3.0" UnitId "deepseq-1.4.4.0" UnitId "bytestring-0.10.8.2" UnitId "containers-0.6.0.1" UnitId "binary-0.8.6.0" UnitId "text-1.2.3.1" UnitId "transformers-0.5.6.2" UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4" UnitId "cabal-test-0.1.0.0-inplace" UnitId "cabal-test-0.1.0.0-inplace-cabal-test" Direct deps ~~~~~~~~~~~ PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameLib base-4.12.0.0 bytestring-0.10.8.2 say-0.1.0.1 text-1.2.3.1 PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameExe "cabal-test" base-4.12.0.0 bytestring-0.10.8.2 cabal-test-0.1.0.0 ``` ``` # cabal-plan info --builddir=dist2 Tree ~~~~ cabal-test-0.1.0.0 [cabal-test-0.1.0.0:exe:"cabal-test"] ├─ base-4.12.0.0 │ ├─ ghc-prim-0.5.3 │ │ └─ rts-1.0 │ ├─ integer-gmp-1.0.2.0 │ │ └─ ghc-prim-0.5.3 ┄┄ │ └─ rts-1.0 ┄┄ ├─ bytestring-0.10.8.2 │ ├─ base-4.12.0.0 ┄┄ │ ├─ deepseq-1.4.4.0 │ │ ├─ array-0.5.3.0 │ │ │ └─ base-4.12.0.0 ┄┄ │ │ └─ base-4.12.0.0 ┄┄ │ ├─ ghc-prim-0.5.3 ┄┄ │ └─ integer-gmp-1.0.2.0 ┄┄ └─ cabal-test-0.1.0.0 ├─ base-4.12.0.0 ┄┄ ├─ bytestring-0.10.8.2 ┄┄ ├─ say-0.1.0.1 │ ├─ base-4.12.0.0 ┄┄ │ ├─ bytestring-0.10.8.2 ┄┄ │ ├─ text-1.2.3.1 │ │ ├─ array-0.5.3.0 ┄┄ │ │ ├─ base-4.12.0.0 ┄┄ │ │ ├─ binary-0.8.6.0 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ │ └─ containers-0.6.0.1 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ │ └─ ghc-prim-0.5.3 ┄┄ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ ├─ ghc-prim-0.5.3 ┄┄ │ │ └─ integer-gmp-1.0.2.0 ┄┄ │ └─ transformers-0.5.6.2 │ └─ base-4.12.0.0 ┄┄ └─ text-1.2.3.1 ┄┄ Top-sorted ~~~~~~~~~~ UnitId "rts" UnitId "ghc-prim-0.5.3" UnitId "integer-gmp-1.0.2.0" UnitId "base-4.12.0.0" UnitId "array-0.5.3.0" UnitId "deepseq-1.4.4.0" UnitId "bytestring-0.10.8.2" UnitId "containers-0.6.0.1" UnitId "binary-0.8.6.0" UnitId "text-1.2.3.1" UnitId "transformers-0.5.6.2" UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4" UnitId "cabal-test-0.1.0.0-inplace" UnitId "cabal-test-0.1.0.0-inplace-cabal-test" Direct deps ~~~~~~~~~~~ PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameLib base-4.12.0.0 bytestring-0.10.8.2 say-0.1.0.1 text-1.2.3.1 PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameExe "cabal-test" base-4.12.0.0 bytestring-0.10.8.2 cabal-test-0.1.0.0 ``` ``` # cabal-plan info --builddir ./tmp2/cabal-install.-10896/dist2/ Tree ~~~~ cabal-test-0.1.0.0 [cabal-test-0.1.0.0:exe:"cabal-test"] ├─ base-4.12.0.0 │ ├─ ghc-prim-0.5.3 │ │ └─ rts-1.0 │ ├─ integer-gmp-1.0.2.0 │ │ └─ ghc-prim-0.5.3 ┄┄ │ └─ rts-1.0 ┄┄ ├─ bytestring-0.10.8.2 │ ├─ base-4.12.0.0 ┄┄ │ ├─ deepseq-1.4.4.0 │ │ ├─ array-0.5.3.0 │ │ │ └─ base-4.12.0.0 ┄┄ │ │ └─ base-4.12.0.0 ┄┄ │ ├─ ghc-prim-0.5.3 ┄┄ │ └─ integer-gmp-1.0.2.0 ┄┄ └─ cabal-test-0.1.0.0 ├─ base-4.12.0.0 ┄┄ ├─ bytestring-0.10.8.2 ┄┄ ├─ say-0.1.0.1 │ ├─ base-4.12.0.0 ┄┄ │ ├─ bytestring-0.10.8.2 ┄┄ │ ├─ text-1.2.3.1 │ │ ├─ array-0.5.3.0 ┄┄ │ │ ├─ base-4.12.0.0 ┄┄ │ │ ├─ binary-0.8.6.0 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ │ └─ containers-0.6.0.1 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ │ └─ ghc-prim-0.5.3 ┄┄ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ ├─ ghc-prim-0.5.3 ┄┄ │ │ └─ integer-gmp-1.0.2.0 ┄┄ │ └─ transformers-0.5.6.2 │ └─ base-4.12.0.0 ┄┄ └─ text-1.2.3.1 ┄┄ Top-sorted ~~~~~~~~~~ UnitId "rts" UnitId "ghc-prim-0.5.3" UnitId "integer-gmp-1.0.2.0" UnitId "base-4.12.0.0" UnitId "array-0.5.3.0" UnitId "deepseq-1.4.4.0" UnitId "bytestring-0.10.8.2" UnitId "containers-0.6.0.1" UnitId "binary-0.8.6.0" UnitId "text-1.2.3.1" UnitId "transformers-0.5.6.2" UnitId "say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47" UnitId "cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88" UnitId "cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961" Direct deps ~~~~~~~~~~~ ``` </details> * Summary: the unit id for say (`f56dd8dcb828e676e2fe979752cedce7d5e3ea47` from the temp build dir is different from the other two (`9d85f3e60b43742f757835053a7ea996c1c596e4`). Both are installed in the store dir. * `cabal-plan diff` does not output anything (and afaik it should at least in the first case) ``` # cabal-plan diff --builddir=dist-newstyle --builddir ./tmp2/cabal-install.-10896/dist2 # cabal-plan diff --builddir=./dist-newstyle --builddir=./dist2 # ``` **Expected behavior** `cabal install` should only rebuild local packages like in linux **System information** - Operating system: windows 10 and 7 - `cabal`: 3.6.2.0 but is reproduced for sure with 3.6.0.0 and 3.4.0.0 and afair with any cabal version i tried since 2.4.0.0 - `ghc` versions: 8.6.5 and 8.10.7, but probably with any **Additional context** * I decide to open a new issue as #7693 is about another different issue and #6919 talks about rebuild the local packages in any os * See another reproduction case with haskell-language-server in https://github.com/haskell/cabal/issues/7693#issuecomment-940472595 @fgaz @gbaz i hope there is enough info to at least demonstrate the behaviour, let me know if more info or tests are needed
1.0
`cabal install` rebuilds all dependencies after a `cabal build` - **Describe the bug** `cabal install` rebuilds all dependencies after a `cabal build` in windows **To Reproduce** Steps to reproduce the behavior: * A simple cabal project with only one dependency `say` * First we build the project using a fresh store dir and print the store contents and the cabal-plan fingerprint ``` D:\ws\haskell\cabal-test>cabal clean D:\ws\haskell\cabal-test>rmdir sr /s /q D:\ws\haskell\cabal-test>cabal --store-dir .\sr build Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (first run) - cabal-test-0.1.0.0 (exe:cabal-test) (first run) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. ..................... Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-13304\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4\lib Configuring library for cabal-test-0.1.0.0.. Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. ............ Configuring executable 'cabal-test' for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. ....................... Linking D:\ws\haskell\cabal-test\dist-newstyle\build\x86_64-windows\ghc-8.6.5\cabal-test-0.1.0.0\x\cabal-test\build\cabal-test\cabal-test.exe ... D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b incoming package.db say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4 D:\ws\haskell\cabal-test>cabal-plan fingerprint ________________________________________________________________ B array-0.5.3.0 ________________________________________________________________ B base-4.12.0.0 ________________________________________________________________ B binary-0.8.6.0 ________________________________________________________________ B bytestring-0.10.8.2 ________________________________________________________________ L cabal-test-0.1.0.0 ________________________________________________________________ B containers-0.6.0.1 ________________________________________________________________ B deepseq-1.4.4.0 ________________________________________________________________ B ghc-prim-0.5.3 ________________________________________________________________ B integer-gmp-1.0.2.0 ________________________________________________________________ B rts-1.0 f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1 ________________________________________________________________ B text-1.2.3.1 ________________________________________________________________ B transformers-0.5.6.2 ``` * Then we install the project using the same store dir * Another bug: you can set a relative dir for the store using `cabal build` but no using `cabal install` ``` D:\ws\haskell\cabal-test>cabal --store-dir .\sr install --builddir dist2 Wrote tarball sdist to D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (requires build) - cabal-test-0.1.0.0 (exe:cabal-test) (requires build) Configuring library for say-0.1.0.1.. cabal-3.6.2.0.exe: expected an absolute directory name for --prefix: .\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 cabal-3.6.2.0.exe: Failed to build say-0.1.0.1 (which is required by exe:cabal-test from cabal-test-0.1.0.0). The failure occurred during the configure step. D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install - -builddir dist2 Wrote tarball sdist to D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (requires build) - cabal-test-0.1.0.0 (exe:cabal-test) (requires build) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. ................ Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5 e3ea47\lib Configuring library for cabal-test-0.1.0.0.. Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. .............. Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88\lib Configuring executable 'cabal-test' for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. ................... Linking dist\build\cabal-test\cabal-test.exe ... Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9 d05e08950e4e5de83d054a50961\bin Warning: The directory D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin is not in the system search path. Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe' D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961 cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88 incoming package.db say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4 say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=dist2 ________________________________________________________________ B array-0.5.3.0 ________________________________________________________________ B base-4.12.0.0 ________________________________________________________________ B binary-0.8.6.0 ________________________________________________________________ B bytestring-0.10.8.2 ________________________________________________________________ L cabal-test-0.1.0.0 ________________________________________________________________ B containers-0.6.0.1 ________________________________________________________________ B deepseq-1.4.4.0 ________________________________________________________________ B ghc-prim-0.5.3 ________________________________________________________________ B integer-gmp-1.0.2.0 ________________________________________________________________ B rts-1.0 f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1 ________________________________________________________________ B text-1.2.3.1 ________________________________________________________________ B transformers-0.5.6.2 D:\ws\haskell\cabal-test>cabal-plan diff --builddir=dist-newstyle --builddir=dist2 (No output, lovely) ``` * As you can check `cabal-plan diff` does not output anything, from other tests it seems it is broken in windows, including in the msys2 shell * After deleting the store we do the same in reverse order, this time preserving the temp build dir used by `cabal install` * There is no flag to preserve the directory, we have to copy the tmp dir quickly :facepalm: ``` D:\ws\haskell\cabal-test> rmdir sr /s /q D:\ws\haskell\cabal-test>cabal --store-dir D:\ws\haskell\cabal-test\sr install -builddir dist2 Wrote tarball sdist to D:\ws\haskell\cabal-test\dist2\sdist\cabal-test-0.1.0.0.tar.gz Resolving dependencies... Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (requires build) - cabal-test-0.1.0.0 (exe:cabal-test) (requires build) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. ............................ Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7de3ea47\lib Configuring library for cabal-test-0.1.0.0.. Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. .................. Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\s\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce9bbdaaf168f88\lib Configuring executable 'cabal-test' for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. .................. Linking dist\build\cabal-test\cabal-test.exe ... Installing executable cabal-test in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incomng\new-10896\ws\haskell\cabal-test\sr\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfedd05e08950e4e5de83d054a50961\bin Warning: The directory D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-10896\ws\haskell\cabal-test\r\ghc-8.6.5\cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961\bin is not in the system search path. Symlinking 'cabal-test.exe' to 'D:\cabal\bin\cabal-test.exe' cabal-3.6.2.0.exe: Path 'D:\cabal\bin\cabal-test' already exists. Use --overwrite-policy=always to overwrite. D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961 cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88 incoming package.db say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 ``` * the fingerprint for `say` is the same in the temp build dir ``` D:\ws\haskell\cabal-test>cabal-plan fingerprint --builddir=tmp2\cabal-install.-1 0896\dist2 ________________________________________________________________ B array-0.5.3.0 ________________________________________________________________ B base-4.12.0.0 ________________________________________________________________ B binary-0.8.6.0 ________________________________________________________________ B bytestring-0.10.8.2 14aa96bfe9d34856cf85a09bc4b314aac137e5e325f3f4beeb77b70ebc57a4a6 G cabal-test-0.1.0.0 ________________________________________________________________ B containers-0.6.0.1 ________________________________________________________________ B deepseq-1.4.4.0 ________________________________________________________________ B ghc-prim-0.5.3 ________________________________________________________________ B integer-gmp-1.0.2.0 ________________________________________________________________ B rts-1.0 f639656fc21925c45f3f55769b9fb7a90699e943376a725e215a5deea473b3e4 G say-0.1.0.1 ________________________________________________________________ B text-1.2.3.1 ________________________________________________________________ B transformers-0.5.6.2 ``` * we do the build after the install, deps are rebuilt again: ``` D:\ws\haskell\cabal-test>cabal --store-dir .\sr build Build profile: -w ghc-8.6.5 -O1 In order, the following will be built (use -v for more details): - say-0.1.0.1 (lib) (requires build) - cabal-test-0.1.0.0 (lib) (dependency rebuilt) - cabal-test-0.1.0.0 (exe:cabal-test) (dependency rebuilt) Configuring library for say-0.1.0.1.. Preprocessing library for say-0.1.0.1.. Building library for say-0.1.0.1.. [1 of 2] Compiling Paths_say ( dist\build\autogen\Paths_say.hs, dist\buil d\Paths_say.o ) [2 of 2] Compiling Say ( src\Say.hs, dist\build\Say.o ) Installing library in D:\ws\haskell\cabal-test\sr\ghc-8.6.5\incoming\new-9200\ws \haskell\cabal-test\sr\ghc-8.6.5\say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c 596e4\lib Preprocessing library for cabal-test-0.1.0.0.. Building library for cabal-test-0.1.0.0.. Preprocessing executable 'cabal-test' for cabal-test-0.1.0.0.. Building executable 'cabal-test' for cabal-test-0.1.0.0.. D:\ws\haskell\cabal-test>dir sr\ghc-8.6.5 /b cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961 cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88 incoming package.db say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4 say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47 ``` * output of `cabal info` for all the three build dirs involved: <details> <summary> cabal info </summary> ``` # cabal-plan info --builddir=dist-newstyle Tree ~~~~ cabal-test-0.1.0.0 [cabal-test-0.1.0.0:exe:"cabal-test"] ├─ base-4.12.0.0 │ ├─ ghc-prim-0.5.3 │ │ └─ rts-1.0 │ ├─ integer-gmp-1.0.2.0 │ │ └─ ghc-prim-0.5.3 ┄┄ │ └─ rts-1.0 ┄┄ ├─ bytestring-0.10.8.2 │ ├─ base-4.12.0.0 ┄┄ │ ├─ deepseq-1.4.4.0 │ │ ├─ array-0.5.3.0 │ │ │ └─ base-4.12.0.0 ┄┄ │ │ └─ base-4.12.0.0 ┄┄ │ ├─ ghc-prim-0.5.3 ┄┄ │ └─ integer-gmp-1.0.2.0 ┄┄ └─ cabal-test-0.1.0.0 ├─ base-4.12.0.0 ┄┄ ├─ bytestring-0.10.8.2 ┄┄ ├─ say-0.1.0.1 │ ├─ base-4.12.0.0 ┄┄ │ ├─ bytestring-0.10.8.2 ┄┄ │ ├─ text-1.2.3.1 │ │ ├─ array-0.5.3.0 ┄┄ │ │ ├─ base-4.12.0.0 ┄┄ │ │ ├─ binary-0.8.6.0 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ │ └─ containers-0.6.0.1 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ │ └─ ghc-prim-0.5.3 ┄┄ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ ├─ ghc-prim-0.5.3 ┄┄ │ │ └─ integer-gmp-1.0.2.0 ┄┄ │ └─ transformers-0.5.6.2 │ └─ base-4.12.0.0 ┄┄ └─ text-1.2.3.1 ┄┄ Top-sorted ~~~~~~~~~~ UnitId "rts" UnitId "ghc-prim-0.5.3" UnitId "integer-gmp-1.0.2.0" UnitId "base-4.12.0.0" UnitId "array-0.5.3.0" UnitId "deepseq-1.4.4.0" UnitId "bytestring-0.10.8.2" UnitId "containers-0.6.0.1" UnitId "binary-0.8.6.0" UnitId "text-1.2.3.1" UnitId "transformers-0.5.6.2" UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4" UnitId "cabal-test-0.1.0.0-inplace" UnitId "cabal-test-0.1.0.0-inplace-cabal-test" Direct deps ~~~~~~~~~~~ PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameLib base-4.12.0.0 bytestring-0.10.8.2 say-0.1.0.1 text-1.2.3.1 PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameExe "cabal-test" base-4.12.0.0 bytestring-0.10.8.2 cabal-test-0.1.0.0 ``` ``` # cabal-plan info --builddir=dist2 Tree ~~~~ cabal-test-0.1.0.0 [cabal-test-0.1.0.0:exe:"cabal-test"] ├─ base-4.12.0.0 │ ├─ ghc-prim-0.5.3 │ │ └─ rts-1.0 │ ├─ integer-gmp-1.0.2.0 │ │ └─ ghc-prim-0.5.3 ┄┄ │ └─ rts-1.0 ┄┄ ├─ bytestring-0.10.8.2 │ ├─ base-4.12.0.0 ┄┄ │ ├─ deepseq-1.4.4.0 │ │ ├─ array-0.5.3.0 │ │ │ └─ base-4.12.0.0 ┄┄ │ │ └─ base-4.12.0.0 ┄┄ │ ├─ ghc-prim-0.5.3 ┄┄ │ └─ integer-gmp-1.0.2.0 ┄┄ └─ cabal-test-0.1.0.0 ├─ base-4.12.0.0 ┄┄ ├─ bytestring-0.10.8.2 ┄┄ ├─ say-0.1.0.1 │ ├─ base-4.12.0.0 ┄┄ │ ├─ bytestring-0.10.8.2 ┄┄ │ ├─ text-1.2.3.1 │ │ ├─ array-0.5.3.0 ┄┄ │ │ ├─ base-4.12.0.0 ┄┄ │ │ ├─ binary-0.8.6.0 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ │ └─ containers-0.6.0.1 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ │ └─ ghc-prim-0.5.3 ┄┄ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ ├─ ghc-prim-0.5.3 ┄┄ │ │ └─ integer-gmp-1.0.2.0 ┄┄ │ └─ transformers-0.5.6.2 │ └─ base-4.12.0.0 ┄┄ └─ text-1.2.3.1 ┄┄ Top-sorted ~~~~~~~~~~ UnitId "rts" UnitId "ghc-prim-0.5.3" UnitId "integer-gmp-1.0.2.0" UnitId "base-4.12.0.0" UnitId "array-0.5.3.0" UnitId "deepseq-1.4.4.0" UnitId "bytestring-0.10.8.2" UnitId "containers-0.6.0.1" UnitId "binary-0.8.6.0" UnitId "text-1.2.3.1" UnitId "transformers-0.5.6.2" UnitId "say-0.1.0.1-9d85f3e60b43742f757835053a7ea996c1c596e4" UnitId "cabal-test-0.1.0.0-inplace" UnitId "cabal-test-0.1.0.0-inplace-cabal-test" Direct deps ~~~~~~~~~~~ PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameLib base-4.12.0.0 bytestring-0.10.8.2 say-0.1.0.1 text-1.2.3.1 PkgId (PkgName "cabal-test") (Ver [0,1,0,0]) CompNameExe "cabal-test" base-4.12.0.0 bytestring-0.10.8.2 cabal-test-0.1.0.0 ``` ``` # cabal-plan info --builddir ./tmp2/cabal-install.-10896/dist2/ Tree ~~~~ cabal-test-0.1.0.0 [cabal-test-0.1.0.0:exe:"cabal-test"] ├─ base-4.12.0.0 │ ├─ ghc-prim-0.5.3 │ │ └─ rts-1.0 │ ├─ integer-gmp-1.0.2.0 │ │ └─ ghc-prim-0.5.3 ┄┄ │ └─ rts-1.0 ┄┄ ├─ bytestring-0.10.8.2 │ ├─ base-4.12.0.0 ┄┄ │ ├─ deepseq-1.4.4.0 │ │ ├─ array-0.5.3.0 │ │ │ └─ base-4.12.0.0 ┄┄ │ │ └─ base-4.12.0.0 ┄┄ │ ├─ ghc-prim-0.5.3 ┄┄ │ └─ integer-gmp-1.0.2.0 ┄┄ └─ cabal-test-0.1.0.0 ├─ base-4.12.0.0 ┄┄ ├─ bytestring-0.10.8.2 ┄┄ ├─ say-0.1.0.1 │ ├─ base-4.12.0.0 ┄┄ │ ├─ bytestring-0.10.8.2 ┄┄ │ ├─ text-1.2.3.1 │ │ ├─ array-0.5.3.0 ┄┄ │ │ ├─ base-4.12.0.0 ┄┄ │ │ ├─ binary-0.8.6.0 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ │ └─ containers-0.6.0.1 │ │ │ ├─ array-0.5.3.0 ┄┄ │ │ │ ├─ base-4.12.0.0 ┄┄ │ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ │ └─ ghc-prim-0.5.3 ┄┄ │ │ ├─ bytestring-0.10.8.2 ┄┄ │ │ ├─ deepseq-1.4.4.0 ┄┄ │ │ ├─ ghc-prim-0.5.3 ┄┄ │ │ └─ integer-gmp-1.0.2.0 ┄┄ │ └─ transformers-0.5.6.2 │ └─ base-4.12.0.0 ┄┄ └─ text-1.2.3.1 ┄┄ Top-sorted ~~~~~~~~~~ UnitId "rts" UnitId "ghc-prim-0.5.3" UnitId "integer-gmp-1.0.2.0" UnitId "base-4.12.0.0" UnitId "array-0.5.3.0" UnitId "deepseq-1.4.4.0" UnitId "bytestring-0.10.8.2" UnitId "containers-0.6.0.1" UnitId "binary-0.8.6.0" UnitId "text-1.2.3.1" UnitId "transformers-0.5.6.2" UnitId "say-0.1.0.1-f56dd8dcb828e676e2fe979752cedce7d5e3ea47" UnitId "cabal-test-0.1.0.0-cfcb93c3d8595a655927c100ce79bbdaaf168f88" UnitId "cabal-test-0.1.0.0-9848e717cfed9d05e08950e4e5de83d054a50961" Direct deps ~~~~~~~~~~~ ``` </details> * Summary: the unit id for say (`f56dd8dcb828e676e2fe979752cedce7d5e3ea47` from the temp build dir is different from the other two (`9d85f3e60b43742f757835053a7ea996c1c596e4`). Both are installed in the store dir. * `cabal-plan diff` does not output anything (and afaik it should at least in the first case) ``` # cabal-plan diff --builddir=dist-newstyle --builddir ./tmp2/cabal-install.-10896/dist2 # cabal-plan diff --builddir=./dist-newstyle --builddir=./dist2 # ``` **Expected behavior** `cabal install` should only rebuild local packages like in linux **System information** - Operating system: windows 10 and 7 - `cabal`: 3.6.2.0 but is reproduced for sure with 3.6.0.0 and 3.4.0.0 and afair with any cabal version i tried since 2.4.0.0 - `ghc` versions: 8.6.5 and 8.10.7, but probably with any **Additional context** * I decide to open a new issue as #7693 is about another different issue and #6919 talks about rebuild the local packages in any os * See another reproduction case with haskell-language-server in https://github.com/haskell/cabal/issues/7693#issuecomment-940472595 @fgaz @gbaz i hope there is enough info to at least demonstrate the behaviour, let me know if more info or tests are needed
priority
cabal install rebuilds all dependencies after a cabal build describe the bug cabal install rebuilds all dependencies after a cabal build in windows to reproduce steps to reproduce the behavior a simple cabal project with only one dependency say first we build the project using a fresh store dir and print the store contents and the cabal plan fingerprint d ws haskell cabal test cabal clean d ws haskell cabal test rmdir sr s q d ws haskell cabal test cabal store dir sr build resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib first run cabal test exe cabal test first run configuring library for say preprocessing library for say building library for say installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc say lib configuring library for cabal test preprocessing library for cabal test building library for cabal test configuring executable cabal test for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test linking d ws haskell cabal test dist newstyle build windows ghc cabal test x cabal test build cabal test cabal test exe d ws haskell cabal test dir sr ghc b incoming package db say d ws haskell cabal test cabal plan fingerprint b array b base b binary b bytestring l cabal test b containers b deepseq b ghc prim b integer gmp b rts g say b text b transformers then we install the project using the same store dir another bug you can set a relative dir for the store using cabal build but no using cabal install d ws haskell cabal test cabal store dir sr install builddir wrote tarball sdist to d ws haskell cabal test sdist cabal test tar gz resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib requires build cabal test exe cabal test requires build configuring library for say cabal exe expected an absolute directory name for prefix sr ghc say cabal exe failed to build say which is required by exe cabal test from cabal test the failure occurred during the configure step d ws haskell cabal test cabal store dir d ws haskell cabal test sr install builddir wrote tarball sdist to d ws haskell cabal test sdist cabal test tar gz resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib requires build cabal test exe cabal test requires build configuring library for say preprocessing library for say building library for say installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc say lib configuring library for cabal test preprocessing library for cabal test building library for cabal test installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc cabal test lib configuring executable cabal test for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test linking dist build cabal test cabal test exe installing executable cabal test in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc cabal test bin warning the directory d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc cabal test bin is not in the system search path symlinking cabal test exe to d cabal bin cabal test exe d ws haskell cabal test dir sr ghc b cabal test cabal test incoming package db say say d ws haskell cabal test cabal plan fingerprint builddir b array b base b binary b bytestring l cabal test b containers b deepseq b ghc prim b integer gmp b rts g say b text b transformers d ws haskell cabal test cabal plan diff builddir dist newstyle builddir no output lovely as you can check cabal plan diff does not output anything from other tests it seems it is broken in windows including in the shell after deleting the store we do the same in reverse order this time preserving the temp build dir used by cabal install there is no flag to preserve the directory we have to copy the tmp dir quickly facepalm d ws haskell cabal test rmdir sr s q d ws haskell cabal test cabal store dir d ws haskell cabal test sr install builddir wrote tarball sdist to d ws haskell cabal test sdist cabal test tar gz resolving dependencies build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib requires build cabal test exe cabal test requires build configuring library for say preprocessing library for say building library for say installing library in d ws haskell cabal test sr ghc incoming new s haskell cabal test sr ghc say lib configuring library for cabal test preprocessing library for cabal test building library for cabal test installing library in d ws haskell cabal test sr ghc incoming new s haskell cabal test sr ghc cabal test lib configuring executable cabal test for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test linking dist build cabal test cabal test exe installing executable cabal test in d ws haskell cabal test sr ghc incomng new ws haskell cabal test sr ghc cabal test bin warning the directory d ws haskell cabal test sr ghc incoming new ws haskell cabal test r ghc cabal test bin is not in the system search path symlinking cabal test exe to d cabal bin cabal test exe cabal exe path d cabal bin cabal test already exists use overwrite policy always to overwrite d ws haskell cabal test dir sr ghc b cabal test cabal test incoming package db say the fingerprint for say is the same in the temp build dir d ws haskell cabal test cabal plan fingerprint builddir cabal install b array b base b binary b bytestring g cabal test b containers b deepseq b ghc prim b integer gmp b rts g say b text b transformers we do the build after the install deps are rebuilt again d ws haskell cabal test cabal store dir sr build build profile w ghc in order the following will be built use v for more details say lib requires build cabal test lib dependency rebuilt cabal test exe cabal test dependency rebuilt configuring library for say preprocessing library for say building library for say compiling paths say dist build autogen paths say hs dist buil d paths say o compiling say src say hs dist build say o installing library in d ws haskell cabal test sr ghc incoming new ws haskell cabal test sr ghc say lib preprocessing library for cabal test building library for cabal test preprocessing executable cabal test for cabal test building executable cabal test for cabal test d ws haskell cabal test dir sr ghc b cabal test cabal test incoming package db say say output of cabal info for all the three build dirs involved cabal info cabal plan info builddir dist newstyle tree cabal test ├─ base │ ├─ ghc prim │ │ └─ rts │ ├─ integer gmp │ │ └─ ghc prim ┄┄ │ └─ rts ┄┄ ├─ bytestring │ ├─ base ┄┄ │ ├─ deepseq │ │ ├─ array │ │ │ └─ base ┄┄ │ │ └─ base ┄┄ │ ├─ ghc prim ┄┄ │ └─ integer gmp ┄┄ └─ cabal test ├─ base ┄┄ ├─ bytestring ┄┄ ├─ say │ ├─ base ┄┄ │ ├─ bytestring ┄┄ │ ├─ text │ │ ├─ array ┄┄ │ │ ├─ base ┄┄ │ │ ├─ binary │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ bytestring ┄┄ │ │ │ └─ containers │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ deepseq ┄┄ │ │ │ └─ ghc prim ┄┄ │ │ ├─ bytestring ┄┄ │ │ ├─ deepseq ┄┄ │ │ ├─ ghc prim ┄┄ │ │ └─ integer gmp ┄┄ │ └─ transformers │ └─ base ┄┄ └─ text ┄┄ top sorted unitid rts unitid ghc prim unitid integer gmp unitid base unitid array unitid deepseq unitid bytestring unitid containers unitid binary unitid text unitid transformers unitid say unitid cabal test inplace unitid cabal test inplace cabal test direct deps pkgid pkgname cabal test ver compnamelib base bytestring say text pkgid pkgname cabal test ver compnameexe cabal test base bytestring cabal test cabal plan info builddir tree cabal test ├─ base │ ├─ ghc prim │ │ └─ rts │ ├─ integer gmp │ │ └─ ghc prim ┄┄ │ └─ rts ┄┄ ├─ bytestring │ ├─ base ┄┄ │ ├─ deepseq │ │ ├─ array │ │ │ └─ base ┄┄ │ │ └─ base ┄┄ │ ├─ ghc prim ┄┄ │ └─ integer gmp ┄┄ └─ cabal test ├─ base ┄┄ ├─ bytestring ┄┄ ├─ say │ ├─ base ┄┄ │ ├─ bytestring ┄┄ │ ├─ text │ │ ├─ array ┄┄ │ │ ├─ base ┄┄ │ │ ├─ binary │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ bytestring ┄┄ │ │ │ └─ containers │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ deepseq ┄┄ │ │ │ └─ ghc prim ┄┄ │ │ ├─ bytestring ┄┄ │ │ ├─ deepseq ┄┄ │ │ ├─ ghc prim ┄┄ │ │ └─ integer gmp ┄┄ │ └─ transformers │ └─ base ┄┄ └─ text ┄┄ top sorted unitid rts unitid ghc prim unitid integer gmp unitid base unitid array unitid deepseq unitid bytestring unitid containers unitid binary unitid text unitid transformers unitid say unitid cabal test inplace unitid cabal test inplace cabal test direct deps pkgid pkgname cabal test ver compnamelib base bytestring say text pkgid pkgname cabal test ver compnameexe cabal test base bytestring cabal test cabal plan info builddir cabal install tree cabal test ├─ base │ ├─ ghc prim │ │ └─ rts │ ├─ integer gmp │ │ └─ ghc prim ┄┄ │ └─ rts ┄┄ ├─ bytestring │ ├─ base ┄┄ │ ├─ deepseq │ │ ├─ array │ │ │ └─ base ┄┄ │ │ └─ base ┄┄ │ ├─ ghc prim ┄┄ │ └─ integer gmp ┄┄ └─ cabal test ├─ base ┄┄ ├─ bytestring ┄┄ ├─ say │ ├─ base ┄┄ │ ├─ bytestring ┄┄ │ ├─ text │ │ ├─ array ┄┄ │ │ ├─ base ┄┄ │ │ ├─ binary │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ bytestring ┄┄ │ │ │ └─ containers │ │ │ ├─ array ┄┄ │ │ │ ├─ base ┄┄ │ │ │ ├─ deepseq ┄┄ │ │ │ └─ ghc prim ┄┄ │ │ ├─ bytestring ┄┄ │ │ ├─ deepseq ┄┄ │ │ ├─ ghc prim ┄┄ │ │ └─ integer gmp ┄┄ │ └─ transformers │ └─ base ┄┄ └─ text ┄┄ top sorted unitid rts unitid ghc prim unitid integer gmp unitid base unitid array unitid deepseq unitid bytestring unitid containers unitid binary unitid text unitid transformers unitid say unitid cabal test unitid cabal test direct deps summary the unit id for say from the temp build dir is different from the other two both are installed in the store dir cabal plan diff does not output anything and afaik it should at least in the first case cabal plan diff builddir dist newstyle builddir cabal install cabal plan diff builddir dist newstyle builddir expected behavior cabal install should only rebuild local packages like in linux system information operating system windows and cabal but is reproduced for sure with and and afair with any cabal version i tried since ghc versions and but probably with any additional context i decide to open a new issue as is about another different issue and talks about rebuild the local packages in any os see another reproduction case with haskell language server in fgaz gbaz i hope there is enough info to at least demonstrate the behaviour let me know if more info or tests are needed
1
787,687
27,727,039,493
IssuesEvent
2023-03-15 03:41:14
jkachhad/Stormhalter
https://api.github.com/repos/jkachhad/Stormhalter
closed
Creature List Movement
Enhancement High Priority UX
Creature list becomes unusable when there are more than a page or two of creatures on screen. It jumps around and reorders multiple times a round even if nothing moves on the screen. [Reported by Zathras]
1.0
Creature List Movement - Creature list becomes unusable when there are more than a page or two of creatures on screen. It jumps around and reorders multiple times a round even if nothing moves on the screen. [Reported by Zathras]
priority
creature list movement creature list becomes unusable when there are more than a page or two of creatures on screen it jumps around and reorders multiple times a round even if nothing moves on the screen
1
388,933
11,495,240,690
IssuesEvent
2020-02-12 04:07:28
wso2/product-is
https://api.github.com/repos/wso2/product-is
opened
Handling Last login time as identity claim
Affected/5.10.0-beta Complexity/Low Component/Identity Mgt Component/Migration Priority/High Type/Docs
With the fix \[1] we are handling the Last login time(which claim is used in account suspension) as identity claim in identity server. Before the fix we have handled the last login claim as default claim. Since this is a behavioural change we have introduced a new property to disable this behaviour. If you want to get the previous behaviour please add the following property in deployment.toml file ``` [account.suspension] lastlogin_as_default_claim=true ``` \[1] https://github.com/wso2-extensions/identity-governance/pull/358
1.0
Handling Last login time as identity claim - With the fix \[1] we are handling the Last login time(which claim is used in account suspension) as identity claim in identity server. Before the fix we have handled the last login claim as default claim. Since this is a behavioural change we have introduced a new property to disable this behaviour. If you want to get the previous behaviour please add the following property in deployment.toml file ``` [account.suspension] lastlogin_as_default_claim=true ``` \[1] https://github.com/wso2-extensions/identity-governance/pull/358
priority
handling last login time as identity claim with the fix we are handling the last login time which claim is used in account suspension as identity claim in identity server before the fix we have handled the last login claim as default claim since this is a behavioural change we have introduced a new property to disable this behaviour if you want to get the previous behaviour please add the following property in deployment toml file lastlogin as default claim true
1
94,373
3,925,006,801
IssuesEvent
2016-04-22 17:16:23
ualbertalib/HydraNorth
https://api.github.com/repos/ualbertalib/HydraNorth
opened
Show release version in public interface
priority:high size:small
To make it easier for us to know what features are supposed to be in the current public interface, let's include the release tag or commit hash, either in the footer or in a comment.
1.0
Show release version in public interface - To make it easier for us to know what features are supposed to be in the current public interface, let's include the release tag or commit hash, either in the footer or in a comment.
priority
show release version in public interface to make it easier for us to know what features are supposed to be in the current public interface let s include the release tag or commit hash either in the footer or in a comment
1
402,785
11,824,837,491
IssuesEvent
2020-03-21 09:10:45
bryntum/support
https://api.github.com/repos/bryntum/support
opened
Should round percentDone value for tasks in task editor
bug high-priority
<img width="648" alt="Screenshot 2020-03-21 at 10 09 59" src="https://user-images.githubusercontent.com/218570/77223368-28915d80-6b5c-11ea-9239-f2972fa54a9f.png">
1.0
Should round percentDone value for tasks in task editor - <img width="648" alt="Screenshot 2020-03-21 at 10 09 59" src="https://user-images.githubusercontent.com/218570/77223368-28915d80-6b5c-11ea-9239-f2972fa54a9f.png">
priority
should round percentdone value for tasks in task editor img width alt screenshot at src
1
422,414
12,278,024,805
IssuesEvent
2020-05-08 09:07:31
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
opened
Contents are truncated in deployment.toml by running secure vault
Affected/3.0.0 Priority/High Type/Bug WUM
### Description: Deployment.toml configurations - all after [secret] section are loosing by running secure vault. ### Steps to reproduce: 1. Configure secret aliases in a middle position of the toml file. (Ex:- next to the keystore configs) 2. Run ciphertool.sh -Dconfigure 3. This will remove all the contents after secret alias section which we newly configured in the toml file. ### Affected Product Version: 3.0.0 , wum - 1586442818040 ### Environment details (with versions): - OS: MacOS --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
1.0
Contents are truncated in deployment.toml by running secure vault - ### Description: Deployment.toml configurations - all after [secret] section are loosing by running secure vault. ### Steps to reproduce: 1. Configure secret aliases in a middle position of the toml file. (Ex:- next to the keystore configs) 2. Run ciphertool.sh -Dconfigure 3. This will remove all the contents after secret alias section which we newly configured in the toml file. ### Affected Product Version: 3.0.0 , wum - 1586442818040 ### Environment details (with versions): - OS: MacOS --- ### Optional Fields #### Related Issues: <!-- Any related issues from this/other repositories--> #### Suggested Labels: <!--Only to be used by non-members--> #### Suggested Assignees: <!--Only to be used by non-members-->
priority
contents are truncated in deployment toml by running secure vault description deployment toml configurations all after section are loosing by running secure vault steps to reproduce configure secret aliases in a middle position of the toml file ex next to the keystore configs run ciphertool sh dconfigure this will remove all the contents after secret alias section which we newly configured in the toml file affected product version wum environment details with versions os macos optional fields related issues suggested labels suggested assignees
1
513,938
14,928,695,764
IssuesEvent
2021-01-24 20:14:33
bounswe/bounswe2020group4
https://api.github.com/repos/bounswe/bounswe2020group4
closed
Admin : Ban Vendor & Customer , Unban Vendor & Customer
Backend Coding Effort: Medium Priority: High Status: Needs Review
Deadline: 24.01.2021 Admin should be able to ban vendor and customer. If it is necessary , admin should be able to unban them again.
1.0
Admin : Ban Vendor & Customer , Unban Vendor & Customer - Deadline: 24.01.2021 Admin should be able to ban vendor and customer. If it is necessary , admin should be able to unban them again.
priority
admin ban vendor customer unban vendor customer deadline admin should be able to ban vendor and customer if it is necessary admin should be able to unban them again
1
269,135
8,432,376,450
IssuesEvent
2018-10-17 01:42:40
LakeEffectRobotics/LakeEffectScoutingApp
https://api.github.com/repos/LakeEffectRobotics/LakeEffectScoutingApp
closed
Add packet splitting fix to server receiving data as well
HIGH PRIORITY
It's probably causing the weird data issues where only half of the data gets sent, then another half, then all of it.
1.0
Add packet splitting fix to server receiving data as well - It's probably causing the weird data issues where only half of the data gets sent, then another half, then all of it.
priority
add packet splitting fix to server receiving data as well it s probably causing the weird data issues where only half of the data gets sent then another half then all of it
1
622,093
19,606,730,335
IssuesEvent
2022-01-06 10:25:07
wso2/ballerina-plugin-vscode
https://api.github.com/repos/wso2/ballerina-plugin-vscode
opened
Tryit Code Action for the services are shown even when there are syntax errors
bug Type/Bug Priority/High
**Description:** Consider the following capture, <img width="620" alt="Screenshot 2022-01-06 at 15 51 03" src="https://user-images.githubusercontent.com/1329674/148368168-054915de-6a0e-478c-b226-5e41e791f108.png"> Here the source is having a syntax error and even in this case, we show the try it code action. We should not encourage the user/ navigate the user to such actions if the result of the action is ending in an error scenario. Shall we allow this only in scenarios where semantic/ syntax errors are not available
1.0
Tryit Code Action for the services are shown even when there are syntax errors - **Description:** Consider the following capture, <img width="620" alt="Screenshot 2022-01-06 at 15 51 03" src="https://user-images.githubusercontent.com/1329674/148368168-054915de-6a0e-478c-b226-5e41e791f108.png"> Here the source is having a syntax error and even in this case, we show the try it code action. We should not encourage the user/ navigate the user to such actions if the result of the action is ending in an error scenario. Shall we allow this only in scenarios where semantic/ syntax errors are not available
priority
tryit code action for the services are shown even when there are syntax errors description consider the following capture img width alt screenshot at src here the source is having a syntax error and even in this case we show the try it code action we should not encourage the user navigate the user to such actions if the result of the action is ending in an error scenario shall we allow this only in scenarios where semantic syntax errors are not available
1
637,622
20,673,407,997
IssuesEvent
2022-03-10 06:22:00
yukiHaga/regex-hunting
https://api.github.com/repos/yukiHaga/regex-hunting
opened
ランクアップした時に経験値を繰り越しさせる
Priority: high
## 概要 ランクアップした時に経験値を繰り越しさせる 繰り越しされてないから、トータルの経験値と合わない部分がある。 ## やること - [ ] ランクアップした時に経験値を繰り越しさせる。 ## 受け入れ条件 - [ ] ランクアップした時に、経験値が反映されている。 ## 参考記事 特になし。
1.0
ランクアップした時に経験値を繰り越しさせる - ## 概要 ランクアップした時に経験値を繰り越しさせる 繰り越しされてないから、トータルの経験値と合わない部分がある。 ## やること - [ ] ランクアップした時に経験値を繰り越しさせる。 ## 受け入れ条件 - [ ] ランクアップした時に、経験値が反映されている。 ## 参考記事 特になし。
priority
ランクアップした時に経験値を繰り越しさせる 概要 ランクアップした時に経験値を繰り越しさせる 繰り越しされてないから、トータルの経験値と合わない部分がある。 やること ランクアップした時に経験値を繰り越しさせる。 受け入れ条件 ランクアップした時に、経験値が反映されている。 参考記事 特になし。
1
258,813
8,179,918,391
IssuesEvent
2018-08-28 17:50:22
workcraft/workcraft
https://api.github.com/repos/workcraft/workcraft
closed
Recognise MUTEX elements on Verilog import
enhancement priority:high status:confirmed tag:model:circuit
Improve Verilog import, so it recognises MUTEX elements. The definition of MUTEX is already in the _Digital Circuit -> Mutex name and request-grant pair_ global property.
1.0
Recognise MUTEX elements on Verilog import - Improve Verilog import, so it recognises MUTEX elements. The definition of MUTEX is already in the _Digital Circuit -> Mutex name and request-grant pair_ global property.
priority
recognise mutex elements on verilog import improve verilog import so it recognises mutex elements the definition of mutex is already in the digital circuit mutex name and request grant pair global property
1
358,156
10,618,494,502
IssuesEvent
2019-10-13 05:06:12
AY1920S1-CS2103T-F13-2/main
https://api.github.com/repos/AY1920S1-CS2103T-F13-2/main
closed
As a user, I would like to sort items based on ascending or descending fields
priority.High type.Story
So that I can have more flexibility in what items I want to view at the top.
1.0
As a user, I would like to sort items based on ascending or descending fields - So that I can have more flexibility in what items I want to view at the top.
priority
as a user i would like to sort items based on ascending or descending fields so that i can have more flexibility in what items i want to view at the top
1
104,792
4,221,241,320
IssuesEvent
2016-07-01 03:56:35
fossasia/open-event-orga-server
https://api.github.com/repos/fossasia/open-event-orga-server
closed
Don't show trashed items in API
Priority: High Rest-API
Parent issue : #1220 Trash has been implemented using a bool column in each table (in_trash) which is True if the item has been deleted. Currently API makes no distinction between trashed and active items. As trash is only for admins and not for the end-user, the API should not return trashed items and treat them as deleted.
1.0
Don't show trashed items in API - Parent issue : #1220 Trash has been implemented using a bool column in each table (in_trash) which is True if the item has been deleted. Currently API makes no distinction between trashed and active items. As trash is only for admins and not for the end-user, the API should not return trashed items and treat them as deleted.
priority
don t show trashed items in api parent issue trash has been implemented using a bool column in each table in trash which is true if the item has been deleted currently api makes no distinction between trashed and active items as trash is only for admins and not for the end user the api should not return trashed items and treat them as deleted
1
264,521
8,316,105,312
IssuesEvent
2018-09-25 08:03:41
jsheroes/jsheroes.io
https://api.github.com/repos/jsheroes/jsheroes.io
closed
Cookie popup not closing on iPhone 8
bug high-priority
This bug was reported recently but I can't seem to reproduce it on iPhone SE (was thinking it's a problem with Safari) We should check this out asap.
1.0
Cookie popup not closing on iPhone 8 - This bug was reported recently but I can't seem to reproduce it on iPhone SE (was thinking it's a problem with Safari) We should check this out asap.
priority
cookie popup not closing on iphone this bug was reported recently but i can t seem to reproduce it on iphone se was thinking it s a problem with safari we should check this out asap
1
473,207
13,638,656,604
IssuesEvent
2020-09-25 09:42:33
status-im/status-react
https://api.github.com/repos/status-im/status-react
closed
Add aUSDC token and optimize search for Aave (LEND) (aka Ethlend)
high-priority
# Problem See [comment](https://github.com/status-im/status-react/issues/11185#issuecomment-697948911) by @j-zerah. We have an event coming up with AAVE. Given the size of their community being introduced to Status we want to make sure they can find their respective tokens. - [aUSDC ](https://etherscan.io/token/0x9bA00D6856a4eDF4665BcA2C2309936572473B7E)(they'll issue this) https://www.coingecko.com/en/coins/aave-usdc - [AAVE LEND](https://etherscan.io/token/0x80fB784B7eD66730e8b1DBd9820aFD29931aab03). EthLend is currently available, but can not be found when searching for AAVE (nice to have if this is a quick change) ## Implementation ## Acceptance Criteria - Can select aUSDC in manage assets - Can find LEND when searching for AAVE (nice to have) ## Notes ## Future Steps
1.0
Add aUSDC token and optimize search for Aave (LEND) (aka Ethlend) - # Problem See [comment](https://github.com/status-im/status-react/issues/11185#issuecomment-697948911) by @j-zerah. We have an event coming up with AAVE. Given the size of their community being introduced to Status we want to make sure they can find their respective tokens. - [aUSDC ](https://etherscan.io/token/0x9bA00D6856a4eDF4665BcA2C2309936572473B7E)(they'll issue this) https://www.coingecko.com/en/coins/aave-usdc - [AAVE LEND](https://etherscan.io/token/0x80fB784B7eD66730e8b1DBd9820aFD29931aab03). EthLend is currently available, but can not be found when searching for AAVE (nice to have if this is a quick change) ## Implementation ## Acceptance Criteria - Can select aUSDC in manage assets - Can find LEND when searching for AAVE (nice to have) ## Notes ## Future Steps
priority
add ausdc token and optimize search for aave lend aka ethlend problem see by j zerah we have an event coming up with aave given the size of their community being introduced to status we want to make sure they can find their respective tokens issue this ethlend is currently available but can not be found when searching for aave nice to have if this is a quick change implementation acceptance criteria can select ausdc in manage assets can find lend when searching for aave nice to have notes future steps
1
769,924
27,022,568,353
IssuesEvent
2023-02-11 06:53:01
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
MacOS M1 CPU/MPS tests are backlogged due to HW availability
high priority triaged ci: sev
## Current Status *ongoing* ## Error looks like Long queue times for M1 runners, which results in longer than usual merges/timeouts ## Incident timeline (all times pacific) *Include when the incident began, when it was detected, mitigated, root caused, and finally closed.* <img width="652" alt="image" src="https://user-images.githubusercontent.com/2453524/218155131-48628732-6574-4081-b56b-a3a5a2e40979.png"> ## User impact *How does this affect users of PyTorch CI?* ## Root cause Looks like we only have 16 runners for the entire org. ## Mitigation *How did we mitigate the issue?* ## Prevention/followups *How do we prevent issues like this in the future?* cc @ezyang @gchanan @zou3519
1.0
MacOS M1 CPU/MPS tests are backlogged due to HW availability - ## Current Status *ongoing* ## Error looks like Long queue times for M1 runners, which results in longer than usual merges/timeouts ## Incident timeline (all times pacific) *Include when the incident began, when it was detected, mitigated, root caused, and finally closed.* <img width="652" alt="image" src="https://user-images.githubusercontent.com/2453524/218155131-48628732-6574-4081-b56b-a3a5a2e40979.png"> ## User impact *How does this affect users of PyTorch CI?* ## Root cause Looks like we only have 16 runners for the entire org. ## Mitigation *How did we mitigate the issue?* ## Prevention/followups *How do we prevent issues like this in the future?* cc @ezyang @gchanan @zou3519
priority
macos cpu mps tests are backlogged due to hw availability current status ongoing error looks like long queue times for runners which results in longer than usual merges timeouts incident timeline all times pacific include when the incident began when it was detected mitigated root caused and finally closed img width alt image src user impact how does this affect users of pytorch ci root cause looks like we only have runners for the entire org mitigation how did we mitigate the issue prevention followups how do we prevent issues like this in the future cc ezyang gchanan
1
283,574
8,719,960,669
IssuesEvent
2018-12-08 06:43:47
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
make install / make package links for osmesa to libGL
bug likelihood medium priority reviewed severity high
To make the dynamic lib switch work. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1549 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: make install / make package links for osmesa to libGL Assigned to: Eric Brugger Category: Target version: 2.7 Author: Cyrus Harrison Start: 07/30/2013 Due date: % Done: 0 Estimated time: Created: 07/30/2013 11:51 am Updated: 08/30/2013 02:25 pm Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: trunk Impact: Expected Use: OS: All Support Group: Any Description: To make the dynamic lib switch work. Comments: Hi Everyone,For Linux make install / make package with osmesa, I added the proper symlinks that allow us to use OSMesa as libGL at runtime.Trunk commit:edge84@cyrush:svn commit -m "linux / osmesa: install symlinks so osmesa can be used as libGL dynamically" FindVisItMesa.cmakeSending FindVisItMesa.cmakeTransmitting file data .Committed revision r21734.It took quite a while b/c :INSTALL(CODE....using: \$ENV{DESTDIR}\${CMAKE_INSTALL_PREFIX}, not just ${CMAKE_INSTALL_PREFIX})-Cyrus
1.0
make install / make package links for osmesa to libGL - To make the dynamic lib switch work. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 1549 Status: Resolved Project: VisIt Tracker: Bug Priority: High Subject: make install / make package links for osmesa to libGL Assigned to: Eric Brugger Category: Target version: 2.7 Author: Cyrus Harrison Start: 07/30/2013 Due date: % Done: 0 Estimated time: Created: 07/30/2013 11:51 am Updated: 08/30/2013 02:25 pm Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: trunk Impact: Expected Use: OS: All Support Group: Any Description: To make the dynamic lib switch work. Comments: Hi Everyone,For Linux make install / make package with osmesa, I added the proper symlinks that allow us to use OSMesa as libGL at runtime.Trunk commit:edge84@cyrush:svn commit -m "linux / osmesa: install symlinks so osmesa can be used as libGL dynamically" FindVisItMesa.cmakeSending FindVisItMesa.cmakeTransmitting file data .Committed revision r21734.It took quite a while b/c :INSTALL(CODE....using: \$ENV{DESTDIR}\${CMAKE_INSTALL_PREFIX}, not just ${CMAKE_INSTALL_PREFIX})-Cyrus
priority
make install make package links for osmesa to libgl to make the dynamic lib switch work redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject make install make package links for osmesa to libgl assigned to eric brugger category target version author cyrus harrison start due date done estimated time created am updated pm likelihood occasional severity very serious found in version trunk impact expected use os all support group any description to make the dynamic lib switch work comments hi everyone for linux make install make package with osmesa i added the proper symlinks that allow us to use osmesa as libgl at runtime trunk commit cyrush svn commit m linux osmesa install symlinks so osmesa can be used as libgl dynamically findvisitmesa cmakesending findvisitmesa cmaketransmitting file data committed revision it took quite a while b c install code using env destdir cmake install prefix not just cmake install prefix cyrus
1
197,160
6,952,990,273
IssuesEvent
2017-12-06 19:23:59
Kademi/kademi-dev
https://api.github.com/repos/Kademi/kademi-dev
closed
Email sending failed
bug High priority
http://vlad32-test.admin.kademi-ci.co/emails/?q=&status=&job= With error : No MimeMessage found I've post to real and mailinator addresses - sit are the same.
1.0
Email sending failed - http://vlad32-test.admin.kademi-ci.co/emails/?q=&status=&job= With error : No MimeMessage found I've post to real and mailinator addresses - sit are the same.
priority
email sending failed with error no mimemessage found i ve post to real and mailinator addresses sit are the same
1
220,878
7,371,800,126
IssuesEvent
2018-03-13 13:01:19
OpenSRP/opensrp-server
https://api.github.com/repos/OpenSRP/opensrp-server
closed
Create /add Shipment end-point in OpenSRP server
High Priority
1. NiFi sends a POST request to the /add shipment end-point on OpenSRP server with a single Shipment object as defined in the docs linked below 2. The server attempts to create a shipment object and stores it in couchDB 3. The server returns a 201 status if successful and 500 status if it fails More at: https://openlmis.atlassian.net/wiki/spaces/OP/pages/237666309/Proof+of+Concept+Create+Nifi+API+Interactions
1.0
Create /add Shipment end-point in OpenSRP server - 1. NiFi sends a POST request to the /add shipment end-point on OpenSRP server with a single Shipment object as defined in the docs linked below 2. The server attempts to create a shipment object and stores it in couchDB 3. The server returns a 201 status if successful and 500 status if it fails More at: https://openlmis.atlassian.net/wiki/spaces/OP/pages/237666309/Proof+of+Concept+Create+Nifi+API+Interactions
priority
create add shipment end point in opensrp server nifi sends a post request to the add shipment end point on opensrp server with a single shipment object as defined in the docs linked below the server attempts to create a shipment object and stores it in couchdb the server returns a status if successful and status if it fails more at
1
52,761
3,028,942,695
IssuesEvent
2015-08-04 09:26:06
pombase/pombase-chado
https://api.github.com/repos/pombase/pombase-chado
closed
Use FYPO from Git/GitHub in Chado load
high priority next-load
The Chado load scripts are still pointing at the subversion repo on SourceForge
1.0
Use FYPO from Git/GitHub in Chado load - The Chado load scripts are still pointing at the subversion repo on SourceForge
priority
use fypo from git github in chado load the chado load scripts are still pointing at the subversion repo on sourceforge
1
357,698
10,616,948,544
IssuesEvent
2019-10-12 15:33:08
bounswe/bounswe2019group8
https://api.github.com/repos/bounswe/bounswe2019group8
opened
Auto-deploy backend app using Jenkins in a Docker container
Backend Effort: High Priority: Low Type: Enhancement
**Actions:** 1. When a push is made to a specific branch, GitHub webhook will trigger the Jenkins pipeline. 1. Jenkins will clone the repo, build image from Dockerfile, push the image to the Docker Hub and run the the app in a container from the pushed image. **Deadline:** 19.10.2019 - 23.59
1.0
Auto-deploy backend app using Jenkins in a Docker container - **Actions:** 1. When a push is made to a specific branch, GitHub webhook will trigger the Jenkins pipeline. 1. Jenkins will clone the repo, build image from Dockerfile, push the image to the Docker Hub and run the the app in a container from the pushed image. **Deadline:** 19.10.2019 - 23.59
priority
auto deploy backend app using jenkins in a docker container actions when a push is made to a specific branch github webhook will trigger the jenkins pipeline jenkins will clone the repo build image from dockerfile push the image to the docker hub and run the the app in a container from the pushed image deadline
1
131,857
5,166,425,158
IssuesEvent
2017-01-17 16:12:11
snaiperskaya96/test-import-repo
https://api.github.com/repos/snaiperskaya96/test-import-repo
opened
Add 'primary colour' to brand settings page
Accepted Enhancement High Priority
https://trello.com/c/gK8xRfFF/279-add-primary-colour-to-brand-settings-page Move brand specific CSS into the HTML skeleton view.
1.0
Add 'primary colour' to brand settings page - https://trello.com/c/gK8xRfFF/279-add-primary-colour-to-brand-settings-page Move brand specific CSS into the HTML skeleton view.
priority
add primary colour to brand settings page move brand specific css into the html skeleton view
1
496,498
14,348,697,796
IssuesEvent
2020-11-29 13:36:18
xournalpp/xournalpp
https://api.github.com/repos/xournalpp/xournalpp
closed
Zooming looses the spot in the document
bug confirmed priority::high
**Affects versions :** - Xubuntu 18.04 - Desktop: Xfce - version of libgtk 3.22.30 - Version of Xournal++: 1.0.19 - Installation method: PPA **Describe the bug** Changing from 100% zoom to fit page zoom, jumps the view around in the document. **To Reproduce** Steps to reproduce the behaviour: Open a multi page document. Do a zoom to 100%, scroll a few pages. Do a zoom to fit. The view of the document jumps back to where you started scrolling. **Expected behavior** doing the above should show the same bit of document just in different zoom depths.
1.0
Zooming looses the spot in the document - **Affects versions :** - Xubuntu 18.04 - Desktop: Xfce - version of libgtk 3.22.30 - Version of Xournal++: 1.0.19 - Installation method: PPA **Describe the bug** Changing from 100% zoom to fit page zoom, jumps the view around in the document. **To Reproduce** Steps to reproduce the behaviour: Open a multi page document. Do a zoom to 100%, scroll a few pages. Do a zoom to fit. The view of the document jumps back to where you started scrolling. **Expected behavior** doing the above should show the same bit of document just in different zoom depths.
priority
zooming looses the spot in the document affects versions xubuntu desktop xfce version of libgtk version of xournal installation method ppa describe the bug changing from zoom to fit page zoom jumps the view around in the document to reproduce steps to reproduce the behaviour open a multi page document do a zoom to scroll a few pages do a zoom to fit the view of the document jumps back to where you started scrolling expected behavior doing the above should show the same bit of document just in different zoom depths
1
748,414
26,122,093,773
IssuesEvent
2022-12-28 13:51:22
abpframework/abp
https://api.github.com/repos/abpframework/abp
closed
Angular - Settings are not visible in the menu.
bug priority:high ui-angular
**App:** Angular `7.0.0-rc.5` Settings are not visible in the menu, although the settings authority is available. ![image](https://user-images.githubusercontent.com/99716457/208830774-21be6854-6cef-4556-8c47-e491a9b00393.png)
1.0
Angular - Settings are not visible in the menu. - **App:** Angular `7.0.0-rc.5` Settings are not visible in the menu, although the settings authority is available. ![image](https://user-images.githubusercontent.com/99716457/208830774-21be6854-6cef-4556-8c47-e491a9b00393.png)
priority
angular settings are not visible in the menu app angular rc settings are not visible in the menu although the settings authority is available
1
807,258
29,985,661,485
IssuesEvent
2023-06-25 07:39:17
Unity-Technologies/com.unity.netcode.gameobjects
https://api.github.com/repos/Unity-Technologies/com.unity.netcode.gameobjects
closed
Shutdown on client side cause an error
type:bug stat:awaiting triage priority:high stat:imported
### Description A clear and concise description of what the bug is. ### Reproduce Steps 1. When a client has been connected 2. Call NetworkManager.Shutdown on Client side. 3. Will cause an error blew: 4. NullReferenceException: Object reference not set to an instance of an object Unity.Netcode.NetworkManager.Shutdown (System.Boolean discardMessageQueue) (at Library/PackageCache/com.unity.netcode.gameobjects@1.5.1/Runtime/Core/NetworkManager.cs:945) ### Environment - OS: Win11 - Unity Version: 2022.3.19f1c1 - Netcode Version: 1.51 - Netcode Commit: [e.g. https://github.com/Unity-Technologies/com.unity.netcode.gameobjects/commit/ba418fa5b600ad9eb61fab0575f12fbecc2c6520] ### Additional Context Add any other context about the problem here. Logs, code snippets would be useful here but please also consider attaching a minimal Unity project that reproduces the issue.
1.0
Shutdown on client side cause an error - ### Description A clear and concise description of what the bug is. ### Reproduce Steps 1. When a client has been connected 2. Call NetworkManager.Shutdown on Client side. 3. Will cause an error blew: 4. NullReferenceException: Object reference not set to an instance of an object Unity.Netcode.NetworkManager.Shutdown (System.Boolean discardMessageQueue) (at Library/PackageCache/com.unity.netcode.gameobjects@1.5.1/Runtime/Core/NetworkManager.cs:945) ### Environment - OS: Win11 - Unity Version: 2022.3.19f1c1 - Netcode Version: 1.51 - Netcode Commit: [e.g. https://github.com/Unity-Technologies/com.unity.netcode.gameobjects/commit/ba418fa5b600ad9eb61fab0575f12fbecc2c6520] ### Additional Context Add any other context about the problem here. Logs, code snippets would be useful here but please also consider attaching a minimal Unity project that reproduces the issue.
priority
shutdown on client side cause an error description a clear and concise description of what the bug is reproduce steps when a client has been connected call networkmanager shutdown on client side will cause an error blew nullreferenceexception object reference not set to an instance of an object unity netcode networkmanager shutdown system boolean discardmessagequeue at library packagecache com unity netcode gameobjects runtime core networkmanager cs environment os unity version netcode version netcode commit additional context add any other context about the problem here logs code snippets would be useful here but please also consider attaching a minimal unity project that reproduces the issue
1
292,552
8,959,510,827
IssuesEvent
2019-01-27 23:10:34
aa-software2112/SOEN390_SimpleCamera
https://api.github.com/repos/aa-software2112/SOEN390_SimpleCamera
closed
Setting up Travis CI
Setup CI priority: high risk: high story
As a developer, I would like to set up Travis CI for continuous integration. - [x] Integrating Travis CI to the repo #14 [1 hour] @samgo1 - [x] Write an initial passing script #15 [1 hour] @samgo1 - [x] Choose and integrate a testing framework #11 [2 hours] @KrishnaPatel1 @Lussier115 - [x] Write 3 unit tests to ensure it is functioning in the Travis build #12 [2 hours] @aa-software2112 @lkadian @anthonyle17 - [x] Choose and integrate an emulator for integration and acceptance testing in Travis #13 [2 hours] @aa-software2112 @Lussier115 - [x] Remove "clean" and "release" build within Travis @KrishnaPatel1 Please refer to the Wiki for more information on Testing. [Story points: 3] [Ideal hours: 8]
1.0
Setting up Travis CI - As a developer, I would like to set up Travis CI for continuous integration. - [x] Integrating Travis CI to the repo #14 [1 hour] @samgo1 - [x] Write an initial passing script #15 [1 hour] @samgo1 - [x] Choose and integrate a testing framework #11 [2 hours] @KrishnaPatel1 @Lussier115 - [x] Write 3 unit tests to ensure it is functioning in the Travis build #12 [2 hours] @aa-software2112 @lkadian @anthonyle17 - [x] Choose and integrate an emulator for integration and acceptance testing in Travis #13 [2 hours] @aa-software2112 @Lussier115 - [x] Remove "clean" and "release" build within Travis @KrishnaPatel1 Please refer to the Wiki for more information on Testing. [Story points: 3] [Ideal hours: 8]
priority
setting up travis ci as a developer i would like to set up travis ci for continuous integration integrating travis ci to the repo write an initial passing script choose and integrate a testing framework write unit tests to ensure it is functioning in the travis build aa lkadian choose and integrate an emulator for integration and acceptance testing in travis aa remove clean and release build within travis please refer to the wiki for more information on testing
1
320,037
9,763,956,167
IssuesEvent
2019-06-05 14:49:57
wherebyus/general-tasks
https://api.github.com/repos/wherebyus/general-tasks
closed
My Account: "Subscriptions" should be called "My Membership"
Feasibility: High Priority: Low Product: Membership Product: Site checked
![image.png](https://images.zenhubusercontent.com/5b3b8639fc589b1d091c443c/0cf44a48-4d43-4ff3-a45f-0e48971aa785) ## DOD - [x] "Subscriptions" in the My Account left navigation is now called "My Membership" - [ ] The URL endpoint is now called `my-membership` (Note this may be dependent on #486) - [ ] The "Subscription" table column header is now called "ID" - [ ] Any "Subscription" references in the order receipts are references to Memberships instead
1.0
My Account: "Subscriptions" should be called "My Membership" - ![image.png](https://images.zenhubusercontent.com/5b3b8639fc589b1d091c443c/0cf44a48-4d43-4ff3-a45f-0e48971aa785) ## DOD - [x] "Subscriptions" in the My Account left navigation is now called "My Membership" - [ ] The URL endpoint is now called `my-membership` (Note this may be dependent on #486) - [ ] The "Subscription" table column header is now called "ID" - [ ] Any "Subscription" references in the order receipts are references to Memberships instead
priority
my account subscriptions should be called my membership dod subscriptions in the my account left navigation is now called my membership the url endpoint is now called my membership note this may be dependent on the subscription table column header is now called id any subscription references in the order receipts are references to memberships instead
1
789,576
27,796,541,309
IssuesEvent
2023-03-17 12:59:04
bryntum/support
https://api.github.com/repos/bryntum/support
closed
Uncaught error in populateEventMenu when dropping an event on the first resource
bug example resolved high-priority premium forum large-account OEM
[Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=23631&p=117002#p117002) "Hello, When I drop an event on the first resource of the scheduler, the event plays a "bouncing" animation, and the event context menu fails to open on right click. The issue cannot be reproduced when dropping an event on other resources than the first one. ![populateEventMenu.png](https://forum.bryntum.com/download/file.php?id=20423&sid=ff9dd647374beefb6cd174b8fbaa7100) The attached video demonstrates that: Drag&drop of event 1 onto resource B -> working context menu on event 1 Drag&drop of event 2 onto resource A -> populateEventMenu fails on event 2 I reproduced this issue on the inline data demo, using the following code and data: ``` import { SchedulerPro } from '../../build/schedulerpro.module.js?465031'; import shared from '../_shared/shared.module.js?465031'; export const data = { "eventsData": [ { "id": 1, "name": "1", "startDate": "2022-10-02T09:12:00", "endDate": "2022-10-02T12:48:00", }, { "id": 2, "name": "2", "startDate": "2022-10-02T09:48:00", "endDate": "2022-10-02T16:24:00", } ], "resourcesData": [ { "id": 0, "name": "A", "image": false }, { "id": 1, "name": "B", "image": false } ], "assignmentsData": [ { "id": 0, "resourceId": 0, "eventId": 1 }, { "id": 1, "resourceId": 1, "eventId": 2 } ], }; const schedulerPro = new SchedulerPro({ appendTo : 'container', startDate : '2022-10-02', endDate : '2022-10-02', viewPreset : 'hourAndDay', forceFit : true, columns : [ { type : 'resourceInfo', text : 'Name', field : 'name', showEventCount : true, width : 150 } ], resources : data.resourcesData, events : data.eventsData, assignments : data.assignmentsData, dependencies : data.dependenciesData }); ``` I tried disabling the event menu to workaround this error, without success. Is there an error in my data or scheduler configuration?" https://user-images.githubusercontent.com/16693227/214033933-d4dd0bd1-33e2-4d9f-893e-6d88e38647ce.mp4
1.0
Uncaught error in populateEventMenu when dropping an event on the first resource - [Forum post](https://forum.bryntum.com/viewtopic.php?f=51&t=23631&p=117002#p117002) "Hello, When I drop an event on the first resource of the scheduler, the event plays a "bouncing" animation, and the event context menu fails to open on right click. The issue cannot be reproduced when dropping an event on other resources than the first one. ![populateEventMenu.png](https://forum.bryntum.com/download/file.php?id=20423&sid=ff9dd647374beefb6cd174b8fbaa7100) The attached video demonstrates that: Drag&drop of event 1 onto resource B -> working context menu on event 1 Drag&drop of event 2 onto resource A -> populateEventMenu fails on event 2 I reproduced this issue on the inline data demo, using the following code and data: ``` import { SchedulerPro } from '../../build/schedulerpro.module.js?465031'; import shared from '../_shared/shared.module.js?465031'; export const data = { "eventsData": [ { "id": 1, "name": "1", "startDate": "2022-10-02T09:12:00", "endDate": "2022-10-02T12:48:00", }, { "id": 2, "name": "2", "startDate": "2022-10-02T09:48:00", "endDate": "2022-10-02T16:24:00", } ], "resourcesData": [ { "id": 0, "name": "A", "image": false }, { "id": 1, "name": "B", "image": false } ], "assignmentsData": [ { "id": 0, "resourceId": 0, "eventId": 1 }, { "id": 1, "resourceId": 1, "eventId": 2 } ], }; const schedulerPro = new SchedulerPro({ appendTo : 'container', startDate : '2022-10-02', endDate : '2022-10-02', viewPreset : 'hourAndDay', forceFit : true, columns : [ { type : 'resourceInfo', text : 'Name', field : 'name', showEventCount : true, width : 150 } ], resources : data.resourcesData, events : data.eventsData, assignments : data.assignmentsData, dependencies : data.dependenciesData }); ``` I tried disabling the event menu to workaround this error, without success. Is there an error in my data or scheduler configuration?" https://user-images.githubusercontent.com/16693227/214033933-d4dd0bd1-33e2-4d9f-893e-6d88e38647ce.mp4
priority
uncaught error in populateeventmenu when dropping an event on the first resource hello when i drop an event on the first resource of the scheduler the event plays a bouncing animation and the event context menu fails to open on right click the issue cannot be reproduced when dropping an event on other resources than the first one the attached video demonstrates that drag drop of event onto resource b working context menu on event drag drop of event onto resource a populateeventmenu fails on event i reproduced this issue on the inline data demo using the following code and data import schedulerpro from build schedulerpro module js import shared from shared shared module js export const data eventsdata id name startdate enddate id name startdate enddate resourcesdata id name a image false id name b image false assignmentsdata id resourceid eventid id resourceid eventid const schedulerpro new schedulerpro appendto container startdate enddate viewpreset hourandday forcefit true columns type resourceinfo text name field name showeventcount true width resources data resourcesdata events data eventsdata assignments data assignmentsdata dependencies data dependenciesdata i tried disabling the event menu to workaround this error without success is there an error in my data or scheduler configuration
1
623,791
19,679,288,427
IssuesEvent
2022-01-11 15:20:10
openghg/openghg
https://api.github.com/repos/openghg/openghg
closed
`search` doesn't return data if ranking data is set for dates that don't overlap data
bug high-priority
I've just encountered this issue ``` ranking datrange = '2015-01-01-00:00:00+00:00_2019-01-01-00:00:00+00:00' ``` And the datasource data keys ``` data_keys = {'2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00' '2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00' '2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00'} ``` As it looks for data keys that overlap the ranking data, if the data doesn't overlap then it breaks. Really it should just return the data it finds instead.
1.0
`search` doesn't return data if ranking data is set for dates that don't overlap data - I've just encountered this issue ``` ranking datrange = '2015-01-01-00:00:00+00:00_2019-01-01-00:00:00+00:00' ``` And the datasource data keys ``` data_keys = {'2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2013-11-23-12:28:30+00:00_2013-12-30-04:51:30+00:00' '2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2014-01-01-22:36:30+00:00_2014-01-07-09:17:30+00:00' '2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00':'data/uuid/49967a55-aa96-4997-be42-642ceb1f0862/v1/2020-03-27-19:03:30+00:00_2020-06-24-09:41:30+00:00'} ``` As it looks for data keys that overlap the ranking data, if the data doesn't overlap then it breaks. Really it should just return the data it finds instead.
priority
search doesn t return data if ranking data is set for dates that don t overlap data i ve just encountered this issue ranking datrange and the datasource data keys data keys data uuid data uuid data uuid as it looks for data keys that overlap the ranking data if the data doesn t overlap then it breaks really it should just return the data it finds instead
1
762,171
26,710,658,901
IssuesEvent
2023-01-27 23:19:16
huridocs/uwazi
https://api.github.com/repos/huridocs/uwazi
closed
Public forms attachments not working on firefox.
Bug :lady_beetle: Priority: High
**Describe the bug** On Firefox, attachments or documents dropzone doesn’t open the upload file dialog on click. The dropzone itself works and files can be attached via drag and drop, but when clicking on the zone to open the upload file dialog, it doesn’t. In Chrome it works fine. **To Reproduce** Steps to reproduce the behavior: 1. Open Firefox 2. Go to https://demo.uwazi.io/en/page/hevpal3191/submit-a-document or any public form 3. Try to upload a document by clicking on 'Select files on your device'. 4. Nothing happens. **Expected behavior** Same as in Chrome, a file dialog should open. **Device (please select all that apply)** - Desktop **Browser** - Firefox **Additional context** This is on a Linux OS. Maybe for Windows or Mac it's different. On mobile Android it works fine in Firefox.
1.0
Public forms attachments not working on firefox. - **Describe the bug** On Firefox, attachments or documents dropzone doesn’t open the upload file dialog on click. The dropzone itself works and files can be attached via drag and drop, but when clicking on the zone to open the upload file dialog, it doesn’t. In Chrome it works fine. **To Reproduce** Steps to reproduce the behavior: 1. Open Firefox 2. Go to https://demo.uwazi.io/en/page/hevpal3191/submit-a-document or any public form 3. Try to upload a document by clicking on 'Select files on your device'. 4. Nothing happens. **Expected behavior** Same as in Chrome, a file dialog should open. **Device (please select all that apply)** - Desktop **Browser** - Firefox **Additional context** This is on a Linux OS. Maybe for Windows or Mac it's different. On mobile Android it works fine in Firefox.
priority
public forms attachments not working on firefox describe the bug on firefox attachments or documents dropzone doesn’t open the upload file dialog on click the dropzone itself works and files can be attached via drag and drop but when clicking on the zone to open the upload file dialog it doesn’t in chrome it works fine to reproduce steps to reproduce the behavior open firefox go to or any public form try to upload a document by clicking on select files on your device nothing happens expected behavior same as in chrome a file dialog should open device please select all that apply desktop browser firefox additional context this is on a linux os maybe for windows or mac it s different on mobile android it works fine in firefox
1
421,160
12,254,605,569
IssuesEvent
2020-05-06 08:46:54
GiftForGood/website
https://api.github.com/repos/GiftForGood/website
opened
Reset Password for NPO
c.UserStory m.MVP priority.High
# User Story <!-- https://github.com/GiftForGood/website/issues?q=is%3Aissue+label%3Ac.UserStory --> ## Describe the user story in detail. As a NPO, I want to reset my password so that I can change my password if I misplaced my previous password.
1.0
Reset Password for NPO - # User Story <!-- https://github.com/GiftForGood/website/issues?q=is%3Aissue+label%3Ac.UserStory --> ## Describe the user story in detail. As a NPO, I want to reset my password so that I can change my password if I misplaced my previous password.
priority
reset password for npo user story describe the user story in detail as a npo i want to reset my password so that i can change my password if i misplaced my previous password
1
386,415
11,438,600,697
IssuesEvent
2020-02-05 04:12:04
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Error while parsing OpenAPIDefinition when updating API definition with the same swagger
3.1.0 Priority/Highest Severity/Major Type/Bug
1. Create an API using the below customer-info-api.yaml (using the same swagger in https://github.com/wso2/product-apim/issues/7068). 2. Update the swagger, go to API Definition and click on IMPORT DEFINITION. 3. Upload the same swagger. [customer-info-api.zip](https://github.com/wso2/product-apim/files/3925652/customer-info-api.zip) `[2019-12-05 13:25:46,362] ERROR - ApisApiServiceImpl Error while parsing OpenAPI definition:attribute components.securitySchemes.default.scopes is missing ` ![Screenshot from 2019-12-05 13-24-20](https://user-images.githubusercontent.com/12208083/70215246-1c2af580-1763-11ea-94ce-48a78f7e583f.png)
1.0
Error while parsing OpenAPIDefinition when updating API definition with the same swagger - 1. Create an API using the below customer-info-api.yaml (using the same swagger in https://github.com/wso2/product-apim/issues/7068). 2. Update the swagger, go to API Definition and click on IMPORT DEFINITION. 3. Upload the same swagger. [customer-info-api.zip](https://github.com/wso2/product-apim/files/3925652/customer-info-api.zip) `[2019-12-05 13:25:46,362] ERROR - ApisApiServiceImpl Error while parsing OpenAPI definition:attribute components.securitySchemes.default.scopes is missing ` ![Screenshot from 2019-12-05 13-24-20](https://user-images.githubusercontent.com/12208083/70215246-1c2af580-1763-11ea-94ce-48a78f7e583f.png)
priority
error while parsing openapidefinition when updating api definition with the same swagger create an api using the below customer info api yaml using the same swagger in update the swagger go to api definition and click on import definition upload the same swagger error apisapiserviceimpl error while parsing openapi definition attribute components securityschemes default scopes is missing
1
235,777
7,742,974,690
IssuesEvent
2018-05-29 11:18:35
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
opened
Add swagger validation before saving the API
2.5.0 Priority/Highest Type/Improvement
**Description:** API Manager allows to save swagger file with syntax errors. For example, if we remove the HTTP method in the swagger file using the in-built swagger editor, we could save the API without resources. Refer image <img width="1323" alt="screen shot 2018-05-29 at 4 46 26 pm" src="https://user-images.githubusercontent.com/4861150/40655731-e90a0d8e-635f-11e8-89eb-97260e92508a.png"> Publishing the API (with errors) is possible after that. Add a validation to the swagger definition before continuing the flow **Affected Product Version:** APIM 2.1.0
1.0
Add swagger validation before saving the API - **Description:** API Manager allows to save swagger file with syntax errors. For example, if we remove the HTTP method in the swagger file using the in-built swagger editor, we could save the API without resources. Refer image <img width="1323" alt="screen shot 2018-05-29 at 4 46 26 pm" src="https://user-images.githubusercontent.com/4861150/40655731-e90a0d8e-635f-11e8-89eb-97260e92508a.png"> Publishing the API (with errors) is possible after that. Add a validation to the swagger definition before continuing the flow **Affected Product Version:** APIM 2.1.0
priority
add swagger validation before saving the api description api manager allows to save swagger file with syntax errors for example if we remove the http method in the swagger file using the in built swagger editor we could save the api without resources refer image img width alt screen shot at pm src publishing the api with errors is possible after that add a validation to the swagger definition before continuing the flow affected product version apim
1