Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
855
labels
stringlengths
4
721
body
stringlengths
1
261k
index
stringclasses
13 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
240k
binary_label
int64
0
1
202,292
7,046,311,220
IssuesEvent
2018-01-02 06:51:04
OperationCode/operationcode_frontend
https://api.github.com/repos/OperationCode/operationcode_frontend
closed
Stop directing traffic to Discourse
beginner friendly Priority: High Status: Available Type: Feature
<!-- Please fill out one of the sections below based on the type of issue you're creating --> # Feature ## Why is this feature being added? Discourse is not gaining traction. Let's double down on Slack and email marketing. ## What should your feature do? Don't shut down Discourse yet - instead, find all areas where it's referenced in the frontend and backend (either Discourse, or community.operationcode.org) and remove them.
1.0
Stop directing traffic to Discourse - <!-- Please fill out one of the sections below based on the type of issue you're creating --> # Feature ## Why is this feature being added? Discourse is not gaining traction. Let's double down on Slack and email marketing. ## What should your feature do? Don't shut down Discourse yet - instead, find all areas where it's referenced in the frontend and backend (either Discourse, or community.operationcode.org) and remove them.
priority
stop directing traffic to discourse feature why is this feature being added discourse is not gaining traction let s double down on slack and email marketing what should your feature do don t shut down discourse yet instead find all areas where it s referenced in the frontend and backend either discourse or community operationcode org and remove them
1
228,429
7,550,851,003
IssuesEvent
2018-04-18 18:13:06
umple/umple
https://api.github.com/repos/umple/umple
closed
variables of associations are not accessible in subclasses
Diffic-Med Priority-VHigh associations bug
Please consider the following example, ``` class A{ 0..1 sth-- * B; } class B{ } class D{ isA A; 0..1 sth -- * B; } ``` It generates the following code ``` public B getB_B(int index) { B aB = (B)bs.get(index); return aB; } ``` The variable `bs` is not accessible in the class D, so there is an error. It's a private varibale in the superclass `A`. There several other errors in other APIs in association with the variable `bs`
1.0
variables of associations are not accessible in subclasses - Please consider the following example, ``` class A{ 0..1 sth-- * B; } class B{ } class D{ isA A; 0..1 sth -- * B; } ``` It generates the following code ``` public B getB_B(int index) { B aB = (B)bs.get(index); return aB; } ``` The variable `bs` is not accessible in the class D, so there is an error. It's a private varibale in the superclass `A`. There several other errors in other APIs in association with the variable `bs`
priority
variables of associations are not accessible in subclasses please consider the following example class a sth b class b class d isa a sth b it generates the following code public b getb b int index b ab b bs get index return ab the variable bs is not accessible in the class d so there is an error it s a private varibale in the superclass a there several other errors in other apis in association with the variable bs
1
374,954
11,097,729,015
IssuesEvent
2019-12-16 13:56:26
bounswe/bounswe2019group10
https://api.github.com/repos/bounswe/bounswe2019group10
opened
Need endpoint for notification
Priority: High Relation: Backend
Need a new notification endpoint that sets all notifications read status to true. Currently checking the number of notifications on the header causes to become red preventing user from seeing new notifications.
1.0
Need endpoint for notification - Need a new notification endpoint that sets all notifications read status to true. Currently checking the number of notifications on the header causes to become red preventing user from seeing new notifications.
priority
need endpoint for notification need a new notification endpoint that sets all notifications read status to true currently checking the number of notifications on the header causes to become red preventing user from seeing new notifications
1
517,247
14,997,928,152
IssuesEvent
2021-01-29 17:37:20
visit-dav/visit
https://api.github.com/repos/visit-dav/visit
closed
We need to switch the development branch to use python 3.
enhancement impact high likelihood high priority reviewed
We are currently still building against python 2 on quartz and running the test suite using python 2. The third party libraries on quartz should be rebuilt with python 3 and the config site file updated. The CI should also be built with python 3. We should remove the python 2 stuff at some later date once we have confidence that the python 3 stuff is solid. This may be in a few months.
1.0
We need to switch the development branch to use python 3. - We are currently still building against python 2 on quartz and running the test suite using python 2. The third party libraries on quartz should be rebuilt with python 3 and the config site file updated. The CI should also be built with python 3. We should remove the python 2 stuff at some later date once we have confidence that the python 3 stuff is solid. This may be in a few months.
priority
we need to switch the development branch to use python we are currently still building against python on quartz and running the test suite using python the third party libraries on quartz should be rebuilt with python and the config site file updated the ci should also be built with python we should remove the python stuff at some later date once we have confidence that the python stuff is solid this may be in a few months
1
605,549
18,736,770,411
IssuesEvent
2021-11-04 08:44:07
AY2122S1-CS2103T-F13-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-F13-2/tp
closed
[PE-D] Deleting elderly does not remove all traces of his name
priority.High
![image.png](https://raw.githubusercontent.com/amzhy/ped/main/files/abf2d3a7-c71a-405c-947c-765e56052920.png) steps to reproduce: 1. addElderly en/John Doe a/40 g/M r/53 nn/John Beckham rs/Father p/98765432 e/mary@example.com t/friends t/owesMoney 2. addTask en/John Doe desc/wash date/2021-10-29 time/19:20 3. deleteElderly 1 The elderly is related to the task, assuming the task is meant to be completed for that elderly, but when the elderly is deleted, the tag with his name remains in the task as shown. <!--session: 1635516837992-8cbce362-68fb-4d9f-97d5-b66f233b2e1c--><!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.FunctionalityBug` original: amzhy/ped#19
1.0
[PE-D] Deleting elderly does not remove all traces of his name - ![image.png](https://raw.githubusercontent.com/amzhy/ped/main/files/abf2d3a7-c71a-405c-947c-765e56052920.png) steps to reproduce: 1. addElderly en/John Doe a/40 g/M r/53 nn/John Beckham rs/Father p/98765432 e/mary@example.com t/friends t/owesMoney 2. addTask en/John Doe desc/wash date/2021-10-29 time/19:20 3. deleteElderly 1 The elderly is related to the task, assuming the task is meant to be completed for that elderly, but when the elderly is deleted, the tag with his name remains in the task as shown. <!--session: 1635516837992-8cbce362-68fb-4d9f-97d5-b66f233b2e1c--><!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.FunctionalityBug` original: amzhy/ped#19
priority
deleting elderly does not remove all traces of his name steps to reproduce addelderly en john doe a g m r nn john beckham rs father p e mary example com t friends t owesmoney addtask en john doe desc wash date time deleteelderly the elderly is related to the task assuming the task is meant to be completed for that elderly but when the elderly is deleted the tag with his name remains in the task as shown labels severity low type functionalitybug original amzhy ped
1
223,927
7,463,655,459
IssuesEvent
2018-04-01 08:39:10
CS2103JAN2018-W14-B1/main
https://api.github.com/repos/CS2103JAN2018-W14-B1/main
closed
List command to accept arguments to request different types of contacts to be listed
Priority.high component: logic
List command to accept a "TYPE" argument for a "contacts" list or a "student" list
1.0
List command to accept arguments to request different types of contacts to be listed - List command to accept a "TYPE" argument for a "contacts" list or a "student" list
priority
list command to accept arguments to request different types of contacts to be listed list command to accept a type argument for a contacts list or a student list
1
650,925
21,443,500,969
IssuesEvent
2022-04-25 01:59:44
Bishop-Laboratory/RLoop-QC-Meta-Analysis-Miller-2022
https://api.github.com/repos/Bishop-Laboratory/RLoop-QC-Meta-Analysis-Miller-2022
closed
Analysis benefit of QC on downstream analysis
analysis needed high-priority
One of the primary distinctions between our approach and that of [R-loopBase](https://academic.oup.com/nar/advance-article/doi/10.1093/nar/gkab1103/6430826) is that they did not consider data quality. In fact, roughly 31% of the data in their analysis was marked as NEG by our quality model. To better emphasize this distinction, we will perform an analysis which demonstrates the importance of our QC approach on the downstream results in our study. We will calculate the consensus R-loop peaks +/- our high-confidence QC filter and then regenerate the following figures: - [ ] 4F - [ ] 4G - [ ] 4I - [ ] 5A - [ ] 6A
1.0
Analysis benefit of QC on downstream analysis - One of the primary distinctions between our approach and that of [R-loopBase](https://academic.oup.com/nar/advance-article/doi/10.1093/nar/gkab1103/6430826) is that they did not consider data quality. In fact, roughly 31% of the data in their analysis was marked as NEG by our quality model. To better emphasize this distinction, we will perform an analysis which demonstrates the importance of our QC approach on the downstream results in our study. We will calculate the consensus R-loop peaks +/- our high-confidence QC filter and then regenerate the following figures: - [ ] 4F - [ ] 4G - [ ] 4I - [ ] 5A - [ ] 6A
priority
analysis benefit of qc on downstream analysis one of the primary distinctions between our approach and that of is that they did not consider data quality in fact roughly of the data in their analysis was marked as neg by our quality model to better emphasize this distinction we will perform an analysis which demonstrates the importance of our qc approach on the downstream results in our study we will calculate the consensus r loop peaks our high confidence qc filter and then regenerate the following figures
1
702,582
24,126,897,502
IssuesEvent
2022-09-21 01:55:40
4paradigm/OpenMLDB
https://api.github.com/repos/4paradigm/OpenMLDB
closed
Delete GLOG message in sql_cluster_router
bug enhancement high-priority storage-engine
Log message display on CLI directly is not well. ![2e2b4fe7d6efcab8a768ba3244fee369](https://user-images.githubusercontent.com/6183996/140457660-28ef8354-6e23-4a10-a5a3-810ffce30c78.png)
1.0
Delete GLOG message in sql_cluster_router - Log message display on CLI directly is not well. ![2e2b4fe7d6efcab8a768ba3244fee369](https://user-images.githubusercontent.com/6183996/140457660-28ef8354-6e23-4a10-a5a3-810ffce30c78.png)
priority
delete glog message in sql cluster router log message display on cli directly is not well
1
300,503
9,211,292,185
IssuesEvent
2019-03-09 14:08:59
qgisissuebot/QGIS
https://api.github.com/repos/qgisissuebot/QGIS
closed
Crach on closing
Bug Priority: high Regression
--- Author Name: **Matjaž Mori** (Matjaž Mori) Original Redmine Issue: 21416, https://issues.qgis.org/issues/21416 --- ## User Feedback This crash happens everytime i close the program. ## Report Details *Crash ID*: a037e3dd18301dbfb420623bbdf24000bc039583 *Stack Trace* ``` QgsMapToolExtent::~QgsMapToolExtent : PyInit__gui : QObjectPrivate::deleteChildren : QWidget::~QWidget : QgsVectorLayerProperties::`default constructor closure' : QgisApp::~QgisApp : CPLStringList::operator char const * __ptr64 const * __ptr64 : main : BaseThreadInitThunk : RtlUserThreadStart : ``` *QGIS Info* QGIS Version: 3.4.5-Madeira QGIS code revision: commit:89ee6f6e23 Compiled against Qt: 5.11.2 Running against Qt: 5.11.2 Compiled against GDAL: 2.4.0 Running against GDAL: 2.4.0 *System Info* CPU Type: x86_64 Kernel Type: winnt Kernel Version: 6.1.7601
1.0
Crach on closing - --- Author Name: **Matjaž Mori** (Matjaž Mori) Original Redmine Issue: 21416, https://issues.qgis.org/issues/21416 --- ## User Feedback This crash happens everytime i close the program. ## Report Details *Crash ID*: a037e3dd18301dbfb420623bbdf24000bc039583 *Stack Trace* ``` QgsMapToolExtent::~QgsMapToolExtent : PyInit__gui : QObjectPrivate::deleteChildren : QWidget::~QWidget : QgsVectorLayerProperties::`default constructor closure' : QgisApp::~QgisApp : CPLStringList::operator char const * __ptr64 const * __ptr64 : main : BaseThreadInitThunk : RtlUserThreadStart : ``` *QGIS Info* QGIS Version: 3.4.5-Madeira QGIS code revision: commit:89ee6f6e23 Compiled against Qt: 5.11.2 Running against Qt: 5.11.2 Compiled against GDAL: 2.4.0 Running against GDAL: 2.4.0 *System Info* CPU Type: x86_64 Kernel Type: winnt Kernel Version: 6.1.7601
priority
crach on closing author name matjaž mori matjaž mori original redmine issue user feedback this crash happens everytime i close the program report details crash id stack trace qgsmaptoolextent qgsmaptoolextent pyinit gui qobjectprivate deletechildren qwidget qwidget qgsvectorlayerproperties default constructor closure qgisapp qgisapp cplstringlist operator char const const main basethreadinitthunk rtluserthreadstart qgis info qgis version madeira qgis code revision commit compiled against qt running against qt compiled against gdal running against gdal system info cpu type kernel type winnt kernel version
1
666,670
22,363,117,771
IssuesEvent
2022-06-15 23:08:06
microsoft/AdaptiveCards
https://api.github.com/repos/microsoft/AdaptiveCards
closed
[Accessibility] CalendarReminder: Focus order is not logical as focus is going on 'Snooze' button instead of 'Dropdown' menu items by using swipe.
Bug Area-Renderers Platform-iOS High Priority Area-Accessibility A11ySev2 HCL-E+D Product-AC HCL-AdaptiveCards-iOS
### Target Platforms iOS ### SDK Version App Version: Version 1.0 (2.3.1-beta.20210603.1) ### Application Name Visualizer ### Problem Description [33752716](https://microsoft.visualstudio.com/OS/_workitems/edit/33752716) Focus order is not logical in swipe navigation on "CalendarReminder.JSON” page. As after performing right swipe from the 'Snooze for' label it is going on the 'Snooze' button. ### Screenshots _No response_ ### Card JSON [CalendarReminder.json](https://github.com/microsoft/AdaptiveCards/blob/main/samples/v1.0/Scenarios/CalendarReminder.json) ### Sample Code Language _No response_ ### Sample Code _No response_
1.0
[Accessibility] CalendarReminder: Focus order is not logical as focus is going on 'Snooze' button instead of 'Dropdown' menu items by using swipe. - ### Target Platforms iOS ### SDK Version App Version: Version 1.0 (2.3.1-beta.20210603.1) ### Application Name Visualizer ### Problem Description [33752716](https://microsoft.visualstudio.com/OS/_workitems/edit/33752716) Focus order is not logical in swipe navigation on "CalendarReminder.JSON” page. As after performing right swipe from the 'Snooze for' label it is going on the 'Snooze' button. ### Screenshots _No response_ ### Card JSON [CalendarReminder.json](https://github.com/microsoft/AdaptiveCards/blob/main/samples/v1.0/Scenarios/CalendarReminder.json) ### Sample Code Language _No response_ ### Sample Code _No response_
priority
calendarreminder focus order is not logical as focus is going on snooze button instead of dropdown menu items by using swipe target platforms ios sdk version app version version beta application name visualizer problem description focus order is not logical in swipe navigation on calendarreminder json” page as after performing right swipe from the snooze for label it is going on the snooze button screenshots no response card json sample code language no response sample code no response
1
189,140
6,794,594,034
IssuesEvent
2017-11-01 12:50:53
metasfresh/metasfresh
https://api.github.com/repos/metasfresh/metasfresh
closed
Add greeting to partner quick creation from order
branch:master priority:high type:enhancement
### Is this a bug or feature request? FR ### What is the current behavior? field is not there #### Which are the steps to reproduce? ### What is the expected or desired behavior? ![image](https://user-images.githubusercontent.com/16079655/32235084-cc3c5eec-be5e-11e7-8385-0ef3bb95c44b.png) ![image](https://user-images.githubusercontent.com/16079655/32235113-e086da44-be5e-11e7-9aa5-0ad9332b18c6.png)
1.0
Add greeting to partner quick creation from order - ### Is this a bug or feature request? FR ### What is the current behavior? field is not there #### Which are the steps to reproduce? ### What is the expected or desired behavior? ![image](https://user-images.githubusercontent.com/16079655/32235084-cc3c5eec-be5e-11e7-8385-0ef3bb95c44b.png) ![image](https://user-images.githubusercontent.com/16079655/32235113-e086da44-be5e-11e7-9aa5-0ad9332b18c6.png)
priority
add greeting to partner quick creation from order is this a bug or feature request fr what is the current behavior field is not there which are the steps to reproduce what is the expected or desired behavior
1
458,310
13,173,068,905
IssuesEvent
2020-08-11 19:36:11
metrumresearchgroup/rbabylon
https://api.github.com/repos/metrumresearchgroup/rbabylon
closed
Add model_summary() info to run_log()
enhancement priority: high risk: medium
# Summary As a user, I would like to be able to summarize multiple models in batch and have some subset of the information in those summaries extracted into a tibble similar to `bbi_run_log_df` (the tibble output from `run_log()`). I would also like to be able to easily append that table onto a `bbi_run_log_df`. ## Technical specification This will be accomplished with the following functions: * **summary_log()** -- Return a new tibble with the "absolute_model_path" column as the primary key, plus the columns extracted from `model_summaries()`. * **add_summary()** -- Return the input tibble, with the columns extracted from `model_summaries()` joined onto it. ## Fields extracted to tibble * objective function value * model estimation method * number of (non-fixed) parameters * number of patients/observations * the boolean heuristics that are currently in `bbi_nonmem_summary$run_heuristics` ## Related Issues Note that this functionality is dependent on functionality described in https://github.com/metrumresearchgroup/rbabylon/issues/53 and will be implemented with that in mind. # Tests - tests/testthat/test-summary-log.R - summary_log() errors with malformed YAML - summary_log() returns NULL and warns when no YAML found - summary_log() works correctly with nested dirs - summary_log(.recurse = FALSE) works - add_summary() works correctly - summary_log works some failed summaries - summary_log works all failed summaries
1.0
Add model_summary() info to run_log() - # Summary As a user, I would like to be able to summarize multiple models in batch and have some subset of the information in those summaries extracted into a tibble similar to `bbi_run_log_df` (the tibble output from `run_log()`). I would also like to be able to easily append that table onto a `bbi_run_log_df`. ## Technical specification This will be accomplished with the following functions: * **summary_log()** -- Return a new tibble with the "absolute_model_path" column as the primary key, plus the columns extracted from `model_summaries()`. * **add_summary()** -- Return the input tibble, with the columns extracted from `model_summaries()` joined onto it. ## Fields extracted to tibble * objective function value * model estimation method * number of (non-fixed) parameters * number of patients/observations * the boolean heuristics that are currently in `bbi_nonmem_summary$run_heuristics` ## Related Issues Note that this functionality is dependent on functionality described in https://github.com/metrumresearchgroup/rbabylon/issues/53 and will be implemented with that in mind. # Tests - tests/testthat/test-summary-log.R - summary_log() errors with malformed YAML - summary_log() returns NULL and warns when no YAML found - summary_log() works correctly with nested dirs - summary_log(.recurse = FALSE) works - add_summary() works correctly - summary_log works some failed summaries - summary_log works all failed summaries
priority
add model summary info to run log summary as a user i would like to be able to summarize multiple models in batch and have some subset of the information in those summaries extracted into a tibble similar to bbi run log df the tibble output from run log i would also like to be able to easily append that table onto a bbi run log df technical specification this will be accomplished with the following functions summary log return a new tibble with the absolute model path column as the primary key plus the columns extracted from model summaries add summary return the input tibble with the columns extracted from model summaries joined onto it fields extracted to tibble objective function value model estimation method number of non fixed parameters number of patients observations the boolean heuristics that are currently in bbi nonmem summary run heuristics related issues note that this functionality is dependent on functionality described in and will be implemented with that in mind tests tests testthat test summary log r summary log errors with malformed yaml summary log returns null and warns when no yaml found summary log works correctly with nested dirs summary log recurse false works add summary works correctly summary log works some failed summaries summary log works all failed summaries
1
706,625
24,279,756,761
IssuesEvent
2022-09-28 16:21:22
7thbeatgames/rd
https://api.github.com/repos/7thbeatgames/rd
closed
Change the window movement to "One Screen" the first time OBS is detected to be running at the same time as RD.
Comp: In-game Priority: High Suggestion
### Please Check - [X] I searched for the issues, and made sure there were no duplicates. - [X] I agree the terms, and understand that my suggestion is not guaranteed to be added or addressed. ### What problem motivated you to submit the suggestion? Pretty much every youtuber or live streamer who plays RD will have to fumble around with the window movement settings the first time they reach 2-X. This interrupts the flow of the game and can be annoying to find. ### Suggestion / Solution The first time RD detects OBS being open at the same time as RD itself, it will automatically change the window dance settings to One Screen (only if window dance has not been activated before). The player can always change it back later and it won't be annoying for recording if 2-X has already been played. ### Alternatives & Workarounds _No response_ ### Demo & Mockup _No response_ ### Note _No response_
1.0
Change the window movement to "One Screen" the first time OBS is detected to be running at the same time as RD. - ### Please Check - [X] I searched for the issues, and made sure there were no duplicates. - [X] I agree the terms, and understand that my suggestion is not guaranteed to be added or addressed. ### What problem motivated you to submit the suggestion? Pretty much every youtuber or live streamer who plays RD will have to fumble around with the window movement settings the first time they reach 2-X. This interrupts the flow of the game and can be annoying to find. ### Suggestion / Solution The first time RD detects OBS being open at the same time as RD itself, it will automatically change the window dance settings to One Screen (only if window dance has not been activated before). The player can always change it back later and it won't be annoying for recording if 2-X has already been played. ### Alternatives & Workarounds _No response_ ### Demo & Mockup _No response_ ### Note _No response_
priority
change the window movement to one screen the first time obs is detected to be running at the same time as rd please check i searched for the issues and made sure there were no duplicates i agree the terms and understand that my suggestion is not guaranteed to be added or addressed what problem motivated you to submit the suggestion pretty much every youtuber or live streamer who plays rd will have to fumble around with the window movement settings the first time they reach x this interrupts the flow of the game and can be annoying to find suggestion solution the first time rd detects obs being open at the same time as rd itself it will automatically change the window dance settings to one screen only if window dance has not been activated before the player can always change it back later and it won t be annoying for recording if x has already been played alternatives workarounds no response demo mockup no response note no response
1
701,422
24,097,673,896
IssuesEvent
2022-09-19 20:22:04
Azordev/did-admin-panel
https://api.github.com/repos/Azordev/did-admin-panel
closed
crear funcionalidad del perfil de provedor
EE-2 priority high QA check
# EXPLANATION Al ingresar con credenciales de proveedor se abre una vista d pantalla donde esta un formulario ,este formulario se debe llenar con los datos para crear un perfil como proveedor pero al ingresar los dato no guarda la información no guarda el imagen no tiene funcionalidad # SCREESHOTS 1 Diseño asignado esta es la vista de pantalla de perfil ![image](https://user-images.githubusercontent.com/111365000/189752944-8e61e8d1-5e39-4be9-97c7-fd10df746fb2.png) 2 Crear la funcionalidad de logotipo al ingresar un imagen debe guardarse y mostrarse cuando se guarde la infomacion como un logotipo de la empresa en su perfil ![image](https://user-images.githubusercontent.com/111365000/189754249-eab3ec73-fbc3-4977-bbeb-64643cb83878.png) 3 las validaciones "Cancelar" y "Guardar" deben funcionar # USER STORY Como proveedor quiero ingresar una imagen como logo de la empresa Como proveedor quiero crea mi perfil al seleccionar la opción "Guardar"
1.0
crear funcionalidad del perfil de provedor - # EXPLANATION Al ingresar con credenciales de proveedor se abre una vista d pantalla donde esta un formulario ,este formulario se debe llenar con los datos para crear un perfil como proveedor pero al ingresar los dato no guarda la información no guarda el imagen no tiene funcionalidad # SCREESHOTS 1 Diseño asignado esta es la vista de pantalla de perfil ![image](https://user-images.githubusercontent.com/111365000/189752944-8e61e8d1-5e39-4be9-97c7-fd10df746fb2.png) 2 Crear la funcionalidad de logotipo al ingresar un imagen debe guardarse y mostrarse cuando se guarde la infomacion como un logotipo de la empresa en su perfil ![image](https://user-images.githubusercontent.com/111365000/189754249-eab3ec73-fbc3-4977-bbeb-64643cb83878.png) 3 las validaciones "Cancelar" y "Guardar" deben funcionar # USER STORY Como proveedor quiero ingresar una imagen como logo de la empresa Como proveedor quiero crea mi perfil al seleccionar la opción "Guardar"
priority
crear funcionalidad del perfil de provedor explanation al ingresar con credenciales de proveedor se abre una vista d pantalla donde esta un formulario este formulario se debe llenar con los datos para crear un perfil como proveedor pero al ingresar los dato no guarda la información no guarda el imagen no tiene funcionalidad screeshots diseño asignado esta es la vista de pantalla de perfil crear la funcionalidad de logotipo al ingresar un imagen debe guardarse y mostrarse cuando se guarde la infomacion como un logotipo de la empresa en su perfil las validaciones cancelar y guardar deben funcionar user story como proveedor quiero ingresar una imagen como logo de la empresa como proveedor quiero crea mi perfil al seleccionar la opción guardar
1
504,324
14,616,757,892
IssuesEvent
2020-12-22 13:46:03
SAP/ownid-webapp
https://api.github.com/repos/SAP/ownid-webapp
closed
Passcode login/register
Priority: High
- Once the Passcode is collected, it is being used as the encryption key to the private key (so no need to store it anywhere) - Can have unlimited attempts for entering the right Passcode in login. The design has a link to reset Passcode (we have another task for that) - Passcode set to one website should be used also for other websites. Dor suggested that can keep a string maybe in the IndexedDB that can be decrypted with the Passcode to validate the Passcode. Using user email as the string force using the same email to all websites. Other option can be to store in user profile as a hash string. Tasks: - Collect, store the PIN - Use the PIN to replace the encryption cookies SoW: 1. Implement Passcode for Login/Register
1.0
Passcode login/register - - Once the Passcode is collected, it is being used as the encryption key to the private key (so no need to store it anywhere) - Can have unlimited attempts for entering the right Passcode in login. The design has a link to reset Passcode (we have another task for that) - Passcode set to one website should be used also for other websites. Dor suggested that can keep a string maybe in the IndexedDB that can be decrypted with the Passcode to validate the Passcode. Using user email as the string force using the same email to all websites. Other option can be to store in user profile as a hash string. Tasks: - Collect, store the PIN - Use the PIN to replace the encryption cookies SoW: 1. Implement Passcode for Login/Register
priority
passcode login register once the passcode is collected it is being used as the encryption key to the private key so no need to store it anywhere can have unlimited attempts for entering the right passcode in login the design has a link to reset passcode we have another task for that passcode set to one website should be used also for other websites dor suggested that can keep a string maybe in the indexeddb that can be decrypted with the passcode to validate the passcode using user email as the string force using the same email to all websites other option can be to store in user profile as a hash string tasks collect store the pin use the pin to replace the encryption cookies sow implement passcode for login register
1
404,771
11,862,993,151
IssuesEvent
2020-03-25 18:53:59
domialex/Sidekick
https://api.github.com/repos/domialex/Sidekick
closed
Gem level filter is set to a value lower than the gem level when price checking a gem
Priority: High Status: Available Type: Bug
If I price check a gem level 21, it sets 16 in the Minimum field (0.7.0-beta). I'm pretty sure the previous version price checked using the same level.
1.0
Gem level filter is set to a value lower than the gem level when price checking a gem - If I price check a gem level 21, it sets 16 in the Minimum field (0.7.0-beta). I'm pretty sure the previous version price checked using the same level.
priority
gem level filter is set to a value lower than the gem level when price checking a gem if i price check a gem level it sets in the minimum field beta i m pretty sure the previous version price checked using the same level
1
185,956
6,732,308,681
IssuesEvent
2017-10-18 10:57:54
gear54rus/RESTED-APS
https://api.github.com/repos/gear54rus/RESTED-APS
opened
Check integration with history, collections and exports/imports
priority: high type: enhancement
The fork added a lot of APS fields, need to check that they are saved, restored and imported/exported properly. Sending both APS and main requests should save history,
1.0
Check integration with history, collections and exports/imports - The fork added a lot of APS fields, need to check that they are saved, restored and imported/exported properly. Sending both APS and main requests should save history,
priority
check integration with history collections and exports imports the fork added a lot of aps fields need to check that they are saved restored and imported exported properly sending both aps and main requests should save history
1
200,524
7,008,760,165
IssuesEvent
2017-12-19 16:41:25
kleros/kleros-interaction
https://api.github.com/repos/kleros/kleros-interaction
opened
Submit a ERC for the evidence standard
high priority
Following ERC792 https://github.com/ethereum/EIPs/issues/792 for the arbitration standard, we now need to create a standard for the way to handle evidence.
1.0
Submit a ERC for the evidence standard - Following ERC792 https://github.com/ethereum/EIPs/issues/792 for the arbitration standard, we now need to create a standard for the way to handle evidence.
priority
submit a erc for the evidence standard following for the arbitration standard we now need to create a standard for the way to handle evidence
1
68,607
3,291,434,735
IssuesEvent
2015-10-30 09:05:03
radike/issue-tracker
https://api.github.com/repos/radike/issue-tracker
closed
Split the main project into more assemblies
enhancement priority HIGH
e.g. create Entity project, use ViewModels, and use auto-mapper to map Entities and ViewModels
1.0
Split the main project into more assemblies - e.g. create Entity project, use ViewModels, and use auto-mapper to map Entities and ViewModels
priority
split the main project into more assemblies e g create entity project use viewmodels and use auto mapper to map entities and viewmodels
1
125,774
4,964,665,886
IssuesEvent
2016-12-03 22:03:58
vnaskos/lajarus
https://api.github.com/repos/vnaskos/lajarus
closed
Refactor every create method on server
point: 2 priority: highest type: refactor
Every "create" method should accept RequestBody parameters, which will contain all the necessary info. They also has to be linked with validation checks.
1.0
Refactor every create method on server - Every "create" method should accept RequestBody parameters, which will contain all the necessary info. They also has to be linked with validation checks.
priority
refactor every create method on server every create method should accept requestbody parameters which will contain all the necessary info they also has to be linked with validation checks
1
237,269
7,757,974,617
IssuesEvent
2018-05-31 18:03:26
martchellop/Entretenibit
https://api.github.com/repos/martchellop/Entretenibit
opened
Get robot working on everyones PC: specify requirements.txt
priority: high
The basic robot mentioned in #66 has already been done and now #67 needs to be done. I am trying to work o n #67 but the robot isn't working out of the box. As such, I was thinking of defining the necessary libraries and putting them in a requirements.txt witch can them in the second sprint be added to a more permanent docker solution.
1.0
Get robot working on everyones PC: specify requirements.txt - The basic robot mentioned in #66 has already been done and now #67 needs to be done. I am trying to work o n #67 but the robot isn't working out of the box. As such, I was thinking of defining the necessary libraries and putting them in a requirements.txt witch can them in the second sprint be added to a more permanent docker solution.
priority
get robot working on everyones pc specify requirements txt the basic robot mentioned in has already been done and now needs to be done i am trying to work o n but the robot isn t working out of the box as such i was thinking of defining the necessary libraries and putting them in a requirements txt witch can them in the second sprint be added to a more permanent docker solution
1
687,016
23,511,340,929
IssuesEvent
2022-08-18 16:50:37
responsible-ai-collaborative/aiid
https://api.github.com/repos/responsible-ai-collaborative/aiid
closed
Images 404ing
Type:Bug Effort: Low Priority:High
Two images on incident 66 are 404ing. <img width="480" alt="Screen Shot 2022-06-21 at 10 13 45 PM" src="https://user-images.githubusercontent.com/64780/174948812-8650ebfa-a85a-4533-a3bd-cae887d8b177.png"> <img width="862" alt="Screen Shot 2022-06-21 at 10 12 48 PM" src="https://user-images.githubusercontent.com/64780/174948725-b7315b23-7917-4f07-a152-6493c96c840d.png">
1.0
Images 404ing - Two images on incident 66 are 404ing. <img width="480" alt="Screen Shot 2022-06-21 at 10 13 45 PM" src="https://user-images.githubusercontent.com/64780/174948812-8650ebfa-a85a-4533-a3bd-cae887d8b177.png"> <img width="862" alt="Screen Shot 2022-06-21 at 10 12 48 PM" src="https://user-images.githubusercontent.com/64780/174948725-b7315b23-7917-4f07-a152-6493c96c840d.png">
priority
images two images on incident are img width alt screen shot at pm src img width alt screen shot at pm src
1
711,751
24,473,808,493
IssuesEvent
2022-10-08 00:20:47
eugenemel/maven
https://api.github.com/repos/eugenemel/maven
closed
Updates to support additional information in SRMTransition (GUI, peakdetector)
high_priority QQQ peakdetector
Sometimes, in addition to a `precursorMz` and ` productMz`, `SRMTransition`s may also have a `transitionId` number. In that case, these transitions should be split out into different groups. This has largely bee updated for peakdetector, but has not been corrected in the MAVEN gui.
1.0
Updates to support additional information in SRMTransition (GUI, peakdetector) - Sometimes, in addition to a `precursorMz` and ` productMz`, `SRMTransition`s may also have a `transitionId` number. In that case, these transitions should be split out into different groups. This has largely bee updated for peakdetector, but has not been corrected in the MAVEN gui.
priority
updates to support additional information in srmtransition gui peakdetector sometimes in addition to a precursormz and productmz srmtransition s may also have a transitionid number in that case these transitions should be split out into different groups this has largely bee updated for peakdetector but has not been corrected in the maven gui
1
686,572
23,496,118,695
IssuesEvent
2022-08-18 01:44:50
ploomber/ploomber
https://api.github.com/repos/ploomber/ploomber
closed
compatibility with IPython 8
bug good first issue high priority med effort
IPython just had a major release and one of the tests is breaking so I pinned the version in `setup.py`. Ploomber actually works fine, it seems like some of the testing config is incompatible with the new IPython internals. Even a diagnosis of next steps will be useful here!
1.0
compatibility with IPython 8 - IPython just had a major release and one of the tests is breaking so I pinned the version in `setup.py`. Ploomber actually works fine, it seems like some of the testing config is incompatible with the new IPython internals. Even a diagnosis of next steps will be useful here!
priority
compatibility with ipython ipython just had a major release and one of the tests is breaking so i pinned the version in setup py ploomber actually works fine it seems like some of the testing config is incompatible with the new ipython internals even a diagnosis of next steps will be useful here
1
43,118
2,882,788,440
IssuesEvent
2015-06-11 08:06:40
HubTurbo/HubTurbo
https://api.github.com/repos/HubTurbo/HubTurbo
closed
Selection should not change when switching back from bView to pView
priority.high type.enhancement
Currently, the selection seems to be based on index rather than a unique identifier. For example, if the 3rd issue was selected originally, the 3rd issue will be selected after switching to pView even if the originally selected issue is no longer the 3rd issue.
1.0
Selection should not change when switching back from bView to pView - Currently, the selection seems to be based on index rather than a unique identifier. For example, if the 3rd issue was selected originally, the 3rd issue will be selected after switching to pView even if the originally selected issue is no longer the 3rd issue.
priority
selection should not change when switching back from bview to pview currently the selection seems to be based on index rather than a unique identifier for example if the issue was selected originally the issue will be selected after switching to pview even if the originally selected issue is no longer the issue
1
106,384
4,271,382,688
IssuesEvent
2016-07-13 10:53:16
GluuFederation/oxAuth
https://api.github.com/repos/GluuFederation/oxAuth
closed
SuperGluu Interception Script Incorrectly Processess Registration
bug High priority
There is a problem with the service that processes the QR code request from the phone. Here is the scenario we tested in a two step authentication 1) User mike: enroll phone 1.... logout 2) Login user mike 2nd time 3) Push notification goes to mike, but Will uses his phone to scan QR code. 4) Will's phone gets "Authentication Success" and the authentication proceeds. The problem is this: post enrollment of a device, only the registered phone should be accepted. The phone should have been informed "Key does not match--authentication failed" And the oxAuth session should be terminated. I think this should be a quick fix.
1.0
SuperGluu Interception Script Incorrectly Processess Registration - There is a problem with the service that processes the QR code request from the phone. Here is the scenario we tested in a two step authentication 1) User mike: enroll phone 1.... logout 2) Login user mike 2nd time 3) Push notification goes to mike, but Will uses his phone to scan QR code. 4) Will's phone gets "Authentication Success" and the authentication proceeds. The problem is this: post enrollment of a device, only the registered phone should be accepted. The phone should have been informed "Key does not match--authentication failed" And the oxAuth session should be terminated. I think this should be a quick fix.
priority
supergluu interception script incorrectly processess registration there is a problem with the service that processes the qr code request from the phone here is the scenario we tested in a two step authentication user mike enroll phone logout login user mike time push notification goes to mike but will uses his phone to scan qr code will s phone gets authentication success and the authentication proceeds the problem is this post enrollment of a device only the registered phone should be accepted the phone should have been informed key does not match authentication failed and the oxauth session should be terminated i think this should be a quick fix
1
489,404
14,105,947,813
IssuesEvent
2020-11-06 14:14:58
onaio/reveal-frontend
https://api.github.com/repos/onaio/reveal-frontend
opened
Missing Plan for User on Thailand Local Production
Priority: High
The user **vbdu_12.1.4-1** on Thailand Local Production cannot find [this](https://mhealth.ddc.moph.go.th/plans/update/3dd52eae-2d16-5d92-ac16-b35b5d90b503) plan in monitor tab.
1.0
Missing Plan for User on Thailand Local Production - The user **vbdu_12.1.4-1** on Thailand Local Production cannot find [this](https://mhealth.ddc.moph.go.th/plans/update/3dd52eae-2d16-5d92-ac16-b35b5d90b503) plan in monitor tab.
priority
missing plan for user on thailand local production the user vbdu on thailand local production cannot find plan in monitor tab
1
146,488
5,622,897,007
IssuesEvent
2017-04-04 13:52:48
CS2103JAN2017-W14-B1/main
https://api.github.com/repos/CS2103JAN2017-W14-B1/main
closed
[UNDO] If redo is input after a command that's not undo, the state shows incorrect data
priority.high type.bug
Clear redo stack if the command entered is not "undo"
1.0
[UNDO] If redo is input after a command that's not undo, the state shows incorrect data - Clear redo stack if the command entered is not "undo"
priority
if redo is input after a command that s not undo the state shows incorrect data clear redo stack if the command entered is not undo
1
448,428
12,950,691,967
IssuesEvent
2020-07-19 14:09:50
crytic/slither
https://api.github.com/repos/crytic/slither
closed
Crash on nested try/catch if/then/else
High Priority bug
```solidity interface I{ function f() external returns(bool); } contract C{ function f(I i) public returns(bool){ bool fail; try i.f() returns (bool success){ if(success){ fail = false; } else{ fail = true; } } catch Error(string memory reason){ fail = true; } if(fail){ return fail; } } } ``` Related: https://github.com/crytic/slither/issues/511
1.0
Crash on nested try/catch if/then/else - ```solidity interface I{ function f() external returns(bool); } contract C{ function f(I i) public returns(bool){ bool fail; try i.f() returns (bool success){ if(success){ fail = false; } else{ fail = true; } } catch Error(string memory reason){ fail = true; } if(fail){ return fail; } } } ``` Related: https://github.com/crytic/slither/issues/511
priority
crash on nested try catch if then else solidity interface i function f external returns bool contract c function f i i public returns bool bool fail try i f returns bool success if success fail false else fail true catch error string memory reason fail true if fail return fail related
1
313,103
9,557,060,348
IssuesEvent
2019-05-03 10:16:59
geosolutions-it/smb-portal
https://api.github.com/repos/geosolutions-it/smb-portal
closed
GetHistogram should calculate histogram on the required interval
Priority: High backlog geoserver
At the moment the histogram start point is calculated as the first value available on the server, instead of initial value indicated in the request. For instance this request: ``` http://cloudsdi.geo-solutions.it:80/geoserver/gwc/service/wmts?service=WMTS&REQUEST=GetHistogram&resolution=P1W&histogram=time&version=1.1.0&layer=landsat8:B3&tileMatrixSet=EPSG:4326&time=2017-01-31T00:48:07.404Z/2017-07-04T21:00:22.692Z ``` Returns this response: ``` <?xml version="1.0" encoding="UTF-8"?><Histogram xmlns="http://demo.geo-solutions.it/share/wmts-multidim/wmts_multi_dimensional.xsd" xmlns:ows="http://www.opengis.net/ows/1.1"> <ows:Identifier>time</ows:Identifier> <Domain>2017-02-06T00:00:00.000Z/2017-04-11T00:00:00.000Z/P1W</Domain> <Values>4,4,3,2,0,3,2,1,1,2</Values> </Histogram> ``` So the client requires `2017-01-31T00:48:07.404Z/2017-07-04T21:00:22.692Z` and the server reply with `2017-02-06T00:00:00.000Z/2017-04-11T00:00:00.000Z/P1W`. So if you try to draw an histogram you can only draw this ![image](https://user-images.githubusercontent.com/1279510/43772976-6368bbbe-9a44-11e8-9096-49efa9a387bf.png) We should allow to start the histogram from the provided interval or add a parameter to define effective interval where to calculate histogram .
1.0
GetHistogram should calculate histogram on the required interval - At the moment the histogram start point is calculated as the first value available on the server, instead of initial value indicated in the request. For instance this request: ``` http://cloudsdi.geo-solutions.it:80/geoserver/gwc/service/wmts?service=WMTS&REQUEST=GetHistogram&resolution=P1W&histogram=time&version=1.1.0&layer=landsat8:B3&tileMatrixSet=EPSG:4326&time=2017-01-31T00:48:07.404Z/2017-07-04T21:00:22.692Z ``` Returns this response: ``` <?xml version="1.0" encoding="UTF-8"?><Histogram xmlns="http://demo.geo-solutions.it/share/wmts-multidim/wmts_multi_dimensional.xsd" xmlns:ows="http://www.opengis.net/ows/1.1"> <ows:Identifier>time</ows:Identifier> <Domain>2017-02-06T00:00:00.000Z/2017-04-11T00:00:00.000Z/P1W</Domain> <Values>4,4,3,2,0,3,2,1,1,2</Values> </Histogram> ``` So the client requires `2017-01-31T00:48:07.404Z/2017-07-04T21:00:22.692Z` and the server reply with `2017-02-06T00:00:00.000Z/2017-04-11T00:00:00.000Z/P1W`. So if you try to draw an histogram you can only draw this ![image](https://user-images.githubusercontent.com/1279510/43772976-6368bbbe-9a44-11e8-9096-49efa9a387bf.png) We should allow to start the histogram from the provided interval or add a parameter to define effective interval where to calculate histogram .
priority
gethistogram should calculate histogram on the required interval at the moment the histogram start point is calculated as the first value available on the server instead of initial value indicated in the request for instance this request returns this response histogram xmlns xmlns ows time so the client requires and the server reply with so if you try to draw an histogram you can only draw this we should allow to start the histogram from the provided interval or add a parameter to define effective interval where to calculate histogram
1
779,313
27,348,722,145
IssuesEvent
2023-02-27 07:55:29
xKDR/Survey.jl
https://api.github.com/repos/xKDR/Survey.jl
opened
Urgent main test failing
high priority
```julia Test Summary: | Pass Total Time makeshort | 6 6 0.1s gdf1 = 3×2 DataFrame Row │ stype nrow │ String1 Int64 ─────┼──────────────── 1 │ H 14 2 │ E 144 3 │ M 25 gdf2 = 2×2 DataFrame Row │ schwide nrow │ String3 Int64 ─────┼──────────────── 1 │ Yes 160 2 │ No 23 ratio.jl: Error During Test at /home/runner/work/Survey.jl/Survey.jl/test/raking.jl:1 Got exception outside of a @test ArgumentError: invalid index: "maxit" of type String ``` I dont understand why, but some tests are failing.
1.0
Urgent main test failing - ```julia Test Summary: | Pass Total Time makeshort | 6 6 0.1s gdf1 = 3×2 DataFrame Row │ stype nrow │ String1 Int64 ─────┼──────────────── 1 │ H 14 2 │ E 144 3 │ M 25 gdf2 = 2×2 DataFrame Row │ schwide nrow │ String3 Int64 ─────┼──────────────── 1 │ Yes 160 2 │ No 23 ratio.jl: Error During Test at /home/runner/work/Survey.jl/Survey.jl/test/raking.jl:1 Got exception outside of a @test ArgumentError: invalid index: "maxit" of type String ``` I dont understand why, but some tests are failing.
priority
urgent main test failing julia test summary pass total time makeshort × dataframe row │ stype nrow │ ─────┼──────────────── │ h │ e │ m × dataframe row │ schwide nrow │ ─────┼──────────────── │ yes │ no ratio jl error during test at home runner work survey jl survey jl test raking jl got exception outside of a test argumenterror invalid index maxit of type string i dont understand why but some tests are failing
1
503,061
14,578,604,572
IssuesEvent
2020-12-18 05:19:22
dogpineapple/tackyboard
https://api.github.com/repos/dogpineapple/tackyboard
closed
Navigation Bar
frontend high priority
Navigation Bar needs: - [ ] logout || sign-in, login - [ ] user settings (to allow user account deletion) - [ ] button to add new job post - [ ] the tackyboard logo that does absolutely nothing OR it should direct the user back to the dashboard if logged in. landing page if not logged in.
1.0
Navigation Bar - Navigation Bar needs: - [ ] logout || sign-in, login - [ ] user settings (to allow user account deletion) - [ ] button to add new job post - [ ] the tackyboard logo that does absolutely nothing OR it should direct the user back to the dashboard if logged in. landing page if not logged in.
priority
navigation bar navigation bar needs logout sign in login user settings to allow user account deletion button to add new job post the tackyboard logo that does absolutely nothing or it should direct the user back to the dashboard if logged in landing page if not logged in
1
469,733
13,524,598,915
IssuesEvent
2020-09-15 11:48:32
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
Using the quick search of the switch groups searches for invalid ranges
Priority: High Type: Bug
**Describe the bug** If you use the quick search of the switches in the left menu and search for a switch range (ex: 10.0.0.0/8), it searches for '10.0.0.0' and omits the mask which yields no results at all **To Reproduce** 1. Create a switch with identifier '10.0.0.0/8' 1. Try to use the quick search in the left menu of the load page to search for it **Expected behavior** Should search for '10.0.0.0/8'
1.0
Using the quick search of the switch groups searches for invalid ranges - **Describe the bug** If you use the quick search of the switches in the left menu and search for a switch range (ex: 10.0.0.0/8), it searches for '10.0.0.0' and omits the mask which yields no results at all **To Reproduce** 1. Create a switch with identifier '10.0.0.0/8' 1. Try to use the quick search in the left menu of the load page to search for it **Expected behavior** Should search for '10.0.0.0/8'
priority
using the quick search of the switch groups searches for invalid ranges describe the bug if you use the quick search of the switches in the left menu and search for a switch range ex it searches for and omits the mask which yields no results at all to reproduce create a switch with identifier try to use the quick search in the left menu of the load page to search for it expected behavior should search for
1
526,006
15,278,009,983
IssuesEvent
2021-02-23 00:30:26
jcsnorlax97/rentr
https://api.github.com/repos/jcsnorlax97/rentr
opened
[TASK] Cleaning up naming conventions & Update Listing entity table
High Priority backend database dev-task
- [ ] Rename `getUser()` to be `getUserViaId()` in UserController & UserDao - [ ] Rename `unAuthenticated()` to be `unauthenticated()` - [ ] Replace `res.status(401).json(...)` in `authenticateUser()` in `UserController` with `ApiError.unauthenticated(...)` - [ ] Ensure the following attributes are in the "Listing" entity: - [ ] Number of bedrooms (String) - [ ] Number of washrooms (String) - [ ] Price (String) - [ ] Laundry Room? (Boolean) - [ ] Pet Allowed? (Boolean) - [ ] Description (String) - [ ] Title (String) - [ ] Image (Url/Base64) (String) - [ ] Parking Available? (Boolean)
1.0
[TASK] Cleaning up naming conventions & Update Listing entity table - - [ ] Rename `getUser()` to be `getUserViaId()` in UserController & UserDao - [ ] Rename `unAuthenticated()` to be `unauthenticated()` - [ ] Replace `res.status(401).json(...)` in `authenticateUser()` in `UserController` with `ApiError.unauthenticated(...)` - [ ] Ensure the following attributes are in the "Listing" entity: - [ ] Number of bedrooms (String) - [ ] Number of washrooms (String) - [ ] Price (String) - [ ] Laundry Room? (Boolean) - [ ] Pet Allowed? (Boolean) - [ ] Description (String) - [ ] Title (String) - [ ] Image (Url/Base64) (String) - [ ] Parking Available? (Boolean)
priority
cleaning up naming conventions update listing entity table rename getuser to be getuserviaid in usercontroller userdao rename unauthenticated to be unauthenticated replace res status json in authenticateuser in usercontroller with apierror unauthenticated ensure the following attributes are in the listing entity number of bedrooms string number of washrooms string price string laundry room boolean pet allowed boolean description string title string image url string parking available boolean
1
537,941
15,757,773,562
IssuesEvent
2021-03-31 05:49:57
garden-io/garden
https://api.github.com/repos/garden-io/garden
closed
Get rid of NFS dependency for in-cluster building
enhancement priority:high stale
## Background One of the most frustrating issues with our remote building feature is the reliance on shareable (ReadWriteMany) volumes, which currently works through NFS at the moment, and optionally other RWX-capable volume provisioners such as EFS. This has turned out to cost a lot of maintenance burden and operational issues for our users. The reason we currently have this requirement is that in order to sync code from the user's local build staging directory, we rsync to an in-cluster volume, that needs to be mountable by the in-cluster builder, whether that is an in-cluster docker daemon, or kaniko. At the time we didn't see an alternative, that would be reasonably performant. We also had a performance goal in mind that may simply not be as important anymore. The raw efficiency of rsync is appealing, but in many (even most) cases the sync of code is only a small part of the overall build time. ## Proposed solution We can avoid this requirement altogether, through some added code complexity in the actual build flows, but in turn avoiding the complexity relating to managing finicky storage providers. We still rsync over to the cluster, but instead of directly mounting the sync volume, we modify the flow to allow using a simpler RWO volume. ### Build flow Depending on the build mode, we do the following: - cluster-docker - We add an rsync container to the cluster-docker Pod spec, and instead of referencing a shared PVC for the build sync volume, we have an RWO PVC that the main docker container and the rsync container share. - When building, we do the exact same thing as previously, but instead of syncing to a separate build sync Pod, we sync to the rsync container in the cluster-docker Pod, before executing the build. - Kaniko - We still create a build sync Deployment and a PVC, but the PVC can now be RWO, and we only have one sync Pod, much like the in-cluster registry. - We still sync to the build-sync Pod ahead of the build. - In the Kaniko build Pod, we create an init container, that rsyncs *from* the build-sync Pod to the container filesystem, replacing the shared build-sync PVC reference with a shared emptyDir volume within the Pod. - The Kaniko build Pod does what it did previously. ### Migration We change the name of the `build-sync` service to `build-sync-v2`, and `docker-daemon` to `docker-daemon-v2`(and the helm release names accordingly). This is to avoid conflicts during rollout since they both function differently from the prior versions, and cannot cover both client versions simultaneously. Users need to be instructed to remove the old `build-sync` and `docker-daemon` deployments and volumes manually, as well as the NFS provisioner, when their team has updated to the new version. Or uninstall and re-init completely, of course. We can print out a message to this effect in the cluster-init command. The current `storage.sync` parameters still apply to the new `build-sync-v2` volume without any changes. We still always install `build-sync-v2`, even though it isn't necessary when using the `cluster-docker` build mode, in order to avoid headaches around using cluster-docker and kaniko in different scenarios on the same cluster. If it is installed, the `cleanup-cluster-registry` script now also executes through the rsync container in the docker daemon Pod, in addition to the `build-sync-v2` Pod. ## Benefits 1. Far simpler requirements for in-cluster building. 2. Fewer bugs and support headaches. 3. Less operation overhead for customers. 4. New Kaniko flow can be made to work without even re-running cluster-init (falling back to older build-sync volume if it's available). ## Drawbacks 1. The two build modes change a bit and have somewhat different flows after the transition. 2. There is a slightly cumbersome transition, having to instruct users to clean up older system services manually, and potentially in parallel for a bit. ## Prioritization I'd say sooner is better for this, since this causes annoying operational issues for users, and support issues for the Garden team by extension.
1.0
Get rid of NFS dependency for in-cluster building - ## Background One of the most frustrating issues with our remote building feature is the reliance on shareable (ReadWriteMany) volumes, which currently works through NFS at the moment, and optionally other RWX-capable volume provisioners such as EFS. This has turned out to cost a lot of maintenance burden and operational issues for our users. The reason we currently have this requirement is that in order to sync code from the user's local build staging directory, we rsync to an in-cluster volume, that needs to be mountable by the in-cluster builder, whether that is an in-cluster docker daemon, or kaniko. At the time we didn't see an alternative, that would be reasonably performant. We also had a performance goal in mind that may simply not be as important anymore. The raw efficiency of rsync is appealing, but in many (even most) cases the sync of code is only a small part of the overall build time. ## Proposed solution We can avoid this requirement altogether, through some added code complexity in the actual build flows, but in turn avoiding the complexity relating to managing finicky storage providers. We still rsync over to the cluster, but instead of directly mounting the sync volume, we modify the flow to allow using a simpler RWO volume. ### Build flow Depending on the build mode, we do the following: - cluster-docker - We add an rsync container to the cluster-docker Pod spec, and instead of referencing a shared PVC for the build sync volume, we have an RWO PVC that the main docker container and the rsync container share. - When building, we do the exact same thing as previously, but instead of syncing to a separate build sync Pod, we sync to the rsync container in the cluster-docker Pod, before executing the build. - Kaniko - We still create a build sync Deployment and a PVC, but the PVC can now be RWO, and we only have one sync Pod, much like the in-cluster registry. - We still sync to the build-sync Pod ahead of the build. - In the Kaniko build Pod, we create an init container, that rsyncs *from* the build-sync Pod to the container filesystem, replacing the shared build-sync PVC reference with a shared emptyDir volume within the Pod. - The Kaniko build Pod does what it did previously. ### Migration We change the name of the `build-sync` service to `build-sync-v2`, and `docker-daemon` to `docker-daemon-v2`(and the helm release names accordingly). This is to avoid conflicts during rollout since they both function differently from the prior versions, and cannot cover both client versions simultaneously. Users need to be instructed to remove the old `build-sync` and `docker-daemon` deployments and volumes manually, as well as the NFS provisioner, when their team has updated to the new version. Or uninstall and re-init completely, of course. We can print out a message to this effect in the cluster-init command. The current `storage.sync` parameters still apply to the new `build-sync-v2` volume without any changes. We still always install `build-sync-v2`, even though it isn't necessary when using the `cluster-docker` build mode, in order to avoid headaches around using cluster-docker and kaniko in different scenarios on the same cluster. If it is installed, the `cleanup-cluster-registry` script now also executes through the rsync container in the docker daemon Pod, in addition to the `build-sync-v2` Pod. ## Benefits 1. Far simpler requirements for in-cluster building. 2. Fewer bugs and support headaches. 3. Less operation overhead for customers. 4. New Kaniko flow can be made to work without even re-running cluster-init (falling back to older build-sync volume if it's available). ## Drawbacks 1. The two build modes change a bit and have somewhat different flows after the transition. 2. There is a slightly cumbersome transition, having to instruct users to clean up older system services manually, and potentially in parallel for a bit. ## Prioritization I'd say sooner is better for this, since this causes annoying operational issues for users, and support issues for the Garden team by extension.
priority
get rid of nfs dependency for in cluster building background one of the most frustrating issues with our remote building feature is the reliance on shareable readwritemany volumes which currently works through nfs at the moment and optionally other rwx capable volume provisioners such as efs this has turned out to cost a lot of maintenance burden and operational issues for our users the reason we currently have this requirement is that in order to sync code from the user s local build staging directory we rsync to an in cluster volume that needs to be mountable by the in cluster builder whether that is an in cluster docker daemon or kaniko at the time we didn t see an alternative that would be reasonably performant we also had a performance goal in mind that may simply not be as important anymore the raw efficiency of rsync is appealing but in many even most cases the sync of code is only a small part of the overall build time proposed solution we can avoid this requirement altogether through some added code complexity in the actual build flows but in turn avoiding the complexity relating to managing finicky storage providers we still rsync over to the cluster but instead of directly mounting the sync volume we modify the flow to allow using a simpler rwo volume build flow depending on the build mode we do the following cluster docker we add an rsync container to the cluster docker pod spec and instead of referencing a shared pvc for the build sync volume we have an rwo pvc that the main docker container and the rsync container share when building we do the exact same thing as previously but instead of syncing to a separate build sync pod we sync to the rsync container in the cluster docker pod before executing the build kaniko we still create a build sync deployment and a pvc but the pvc can now be rwo and we only have one sync pod much like the in cluster registry we still sync to the build sync pod ahead of the build in the kaniko build pod we create an init container that rsyncs from the build sync pod to the container filesystem replacing the shared build sync pvc reference with a shared emptydir volume within the pod the kaniko build pod does what it did previously migration we change the name of the build sync service to build sync and docker daemon to docker daemon and the helm release names accordingly this is to avoid conflicts during rollout since they both function differently from the prior versions and cannot cover both client versions simultaneously users need to be instructed to remove the old build sync and docker daemon deployments and volumes manually as well as the nfs provisioner when their team has updated to the new version or uninstall and re init completely of course we can print out a message to this effect in the cluster init command the current storage sync parameters still apply to the new build sync volume without any changes we still always install build sync even though it isn t necessary when using the cluster docker build mode in order to avoid headaches around using cluster docker and kaniko in different scenarios on the same cluster if it is installed the cleanup cluster registry script now also executes through the rsync container in the docker daemon pod in addition to the build sync pod benefits far simpler requirements for in cluster building fewer bugs and support headaches less operation overhead for customers new kaniko flow can be made to work without even re running cluster init falling back to older build sync volume if it s available drawbacks the two build modes change a bit and have somewhat different flows after the transition there is a slightly cumbersome transition having to instruct users to clean up older system services manually and potentially in parallel for a bit prioritization i d say sooner is better for this since this causes annoying operational issues for users and support issues for the garden team by extension
1
98,053
4,016,260,425
IssuesEvent
2016-05-15 13:48:32
loklak/loklak_webclient
https://api.github.com/repos/loklak/loklak_webclient
closed
Show all tweets created by loklak with rich content attachments within the search results.
Attachments Feature ongoing Priority 1 - High Search
While the search results should be identical to twitter search result views, we want to make use of our own abilities and show the rich content of special tweets that are developed in https://github.com/loklak/loklak_webclient/issues/63 This means, you should share your work to visualize the rich-content tweets. This affects the following formats: - identify and attach exact geographic coordinates and maps - attach video and/or audio content - attach larger texts with simple markdown (i.e. https://guides.github.com/features/mastering-markdown/) as used in github or maybe wikitext, code citations (with code pretty-print) and other typical text rendering - attach better images (i.e. animated gif or images with EXIF data attached)
1.0
Show all tweets created by loklak with rich content attachments within the search results. - While the search results should be identical to twitter search result views, we want to make use of our own abilities and show the rich content of special tweets that are developed in https://github.com/loklak/loklak_webclient/issues/63 This means, you should share your work to visualize the rich-content tweets. This affects the following formats: - identify and attach exact geographic coordinates and maps - attach video and/or audio content - attach larger texts with simple markdown (i.e. https://guides.github.com/features/mastering-markdown/) as used in github or maybe wikitext, code citations (with code pretty-print) and other typical text rendering - attach better images (i.e. animated gif or images with EXIF data attached)
priority
show all tweets created by loklak with rich content attachments within the search results while the search results should be identical to twitter search result views we want to make use of our own abilities and show the rich content of special tweets that are developed in this means you should share your work to visualize the rich content tweets this affects the following formats identify and attach exact geographic coordinates and maps attach video and or audio content attach larger texts with simple markdown i e as used in github or maybe wikitext code citations with code pretty print and other typical text rendering attach better images i e animated gif or images with exif data attached
1
539,840
15,795,727,254
IssuesEvent
2021-04-02 13:40:51
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
Deploy Sample API option available after deploying the sample API
API-M 4.0.0 Priority/High React-UI Type/Bug
### Description: Deploy Sample API option is available even after deploying the PIzzaShackAPI. Clicking on this would cause the following error in the console and the Publisher will be loading indefinitely. ``` ERROR - ApisApiServiceImpl Error while adding new API : null-PizzaShackAPI-1.0.0 - Error occurred while adding the API. A duplicate API already exists for /pizzashack ``` ![Screenshot from 2021-03-18 08-52-28](https://user-images.githubusercontent.com/25485273/111568981-ce0afb80-87c7-11eb-87c4-dcdb41b3a184.png) ### Affected Product Version: 4.0.0-alpha
1.0
Deploy Sample API option available after deploying the sample API - ### Description: Deploy Sample API option is available even after deploying the PIzzaShackAPI. Clicking on this would cause the following error in the console and the Publisher will be loading indefinitely. ``` ERROR - ApisApiServiceImpl Error while adding new API : null-PizzaShackAPI-1.0.0 - Error occurred while adding the API. A duplicate API already exists for /pizzashack ``` ![Screenshot from 2021-03-18 08-52-28](https://user-images.githubusercontent.com/25485273/111568981-ce0afb80-87c7-11eb-87c4-dcdb41b3a184.png) ### Affected Product Version: 4.0.0-alpha
priority
deploy sample api option available after deploying the sample api description deploy sample api option is available even after deploying the pizzashackapi clicking on this would cause the following error in the console and the publisher will be loading indefinitely error apisapiserviceimpl error while adding new api null pizzashackapi error occurred while adding the api a duplicate api already exists for pizzashack affected product version alpha
1
690,614
23,665,849,180
IssuesEvent
2022-08-26 20:49:40
opendatahub-io/odh-manifests
https://api.github.com/repos/opendatahub-io/odh-manifests
closed
[BUG] Enable CI testsuite for ML Pipelines to verify functionality
ci/cd priority/high
**Describe the bug** #637 was merged with CI tests disabled due to incompatibility with the out of date OCP version used during PR tests. Once https://github.com/openshift/release/pull/31350 is merged then these tests will need to be re-enabled so we can verify ml-pipelines works with each ODH release
1.0
[BUG] Enable CI testsuite for ML Pipelines to verify functionality - **Describe the bug** #637 was merged with CI tests disabled due to incompatibility with the out of date OCP version used during PR tests. Once https://github.com/openshift/release/pull/31350 is merged then these tests will need to be re-enabled so we can verify ml-pipelines works with each ODH release
priority
enable ci testsuite for ml pipelines to verify functionality describe the bug was merged with ci tests disabled due to incompatibility with the out of date ocp version used during pr tests once is merged then these tests will need to be re enabled so we can verify ml pipelines works with each odh release
1
656,504
21,766,093,300
IssuesEvent
2022-05-13 02:10:44
kubesphere/kubesphere
https://api.github.com/repos/kubesphere/kubesphere
closed
No traffic information in service details of Service Topology
priority/high
**Describe the Bug** As shown, no traffic information in service details of Service Topology. but in fact, this service have traffic information. ![image](https://user-images.githubusercontent.com/68640256/167106394-5fa70cde-2de5-4a52-8051-b845d34641d7.png) ![image](https://user-images.githubusercontent.com/68640256/167106534-e7e4b709-fd3e-4b52-92b4-966a7e4fd702.png) **Versions Used** KubeSphere: `v3.3.0-alpha.2` /priority high /assign @weihongzhoulord
1.0
No traffic information in service details of Service Topology - **Describe the Bug** As shown, no traffic information in service details of Service Topology. but in fact, this service have traffic information. ![image](https://user-images.githubusercontent.com/68640256/167106394-5fa70cde-2de5-4a52-8051-b845d34641d7.png) ![image](https://user-images.githubusercontent.com/68640256/167106534-e7e4b709-fd3e-4b52-92b4-966a7e4fd702.png) **Versions Used** KubeSphere: `v3.3.0-alpha.2` /priority high /assign @weihongzhoulord
priority
no traffic information in service details of service topology describe the bug as shown no traffic information in service details of service topology but in fact this service have traffic information versions used kubesphere alpha priority high assign weihongzhoulord
1
386,907
11,452,586,652
IssuesEvent
2020-02-06 13:57:43
Materials-Consortia/optimade-python-tools
https://api.github.com/repos/Materials-Consortia/optimade-python-tools
closed
`response_fields` not working
bug priority/high
`response_fields` is not working currently, since some `structures` validators try to check against some values that may not be requested, i.e., they may not exist, so a `KeyError` is thrown.
1.0
`response_fields` not working - `response_fields` is not working currently, since some `structures` validators try to check against some values that may not be requested, i.e., they may not exist, so a `KeyError` is thrown.
priority
response fields not working response fields is not working currently since some structures validators try to check against some values that may not be requested i e they may not exist so a keyerror is thrown
1
642,549
20,906,849,888
IssuesEvent
2022-03-24 03:49:06
chao1224/GraphMVP
https://api.github.com/repos/chao1224/GraphMVP
closed
Request to GraphMVP pre-trained weights
High Priority
Hi Shengchao, Thanks for your great work and clear documentation! I try to download your pre-trained model with the provided link. However, I can't find it. Could you share the pre-trained models with us? Regards, Jun.
1.0
Request to GraphMVP pre-trained weights - Hi Shengchao, Thanks for your great work and clear documentation! I try to download your pre-trained model with the provided link. However, I can't find it. Could you share the pre-trained models with us? Regards, Jun.
priority
request to graphmvp pre trained weights hi shengchao thanks for your great work and clear documentation i try to download your pre trained model with the provided link however i can t find it could you share the pre trained models with us regards jun
1
646,431
21,047,425,929
IssuesEvent
2022-03-31 17:21:06
bitfoundation/bitframework
https://api.github.com/repos/bitfoundation/bitframework
closed
DatePicker unexpected experience on the year selection
area / components high priority enhancement
When you choose a year from the list of years, user expects to see months after clicking a year, but the DatePicker shows other years. This is a different experience from Fluent. ![Screenshot 2022-03-29 193315](https://user-images.githubusercontent.com/13447649/160661335-2629084e-e1e8-4d66-a2c7-a92e69f16393.png)
1.0
DatePicker unexpected experience on the year selection - When you choose a year from the list of years, user expects to see months after clicking a year, but the DatePicker shows other years. This is a different experience from Fluent. ![Screenshot 2022-03-29 193315](https://user-images.githubusercontent.com/13447649/160661335-2629084e-e1e8-4d66-a2c7-a92e69f16393.png)
priority
datepicker unexpected experience on the year selection when you choose a year from the list of years user expects to see months after clicking a year but the datepicker shows other years this is a different experience from fluent
1
123,514
4,864,185,824
IssuesEvent
2016-11-14 17:18:04
samitsolutions/atgmanager
https://api.github.com/repos/samitsolutions/atgmanager
opened
Historical Data: Too Many Entries for Single Point in Time
0 - Backlog High Priority Task
Only one entry should exist for a single point in time, not one per tank. Perhaps if you click on the single time, it will show all tanks for that time in an accordion style drop down. This is confusing: ![uploads/237490e3-ef02-4fd1-81d5-e5aa62bb8609/blob](https://s3-us-west-2.amazonaws.com/prod.huboard.com/uploads%2F237490e3-ef02-4fd1-81d5-e5aa62bb8609%2Fblob) <!--- @huboard:{"order":0.9991004498350495,"milestone_order":0.9974035067263741} -->
1.0
Historical Data: Too Many Entries for Single Point in Time - Only one entry should exist for a single point in time, not one per tank. Perhaps if you click on the single time, it will show all tanks for that time in an accordion style drop down. This is confusing: ![uploads/237490e3-ef02-4fd1-81d5-e5aa62bb8609/blob](https://s3-us-west-2.amazonaws.com/prod.huboard.com/uploads%2F237490e3-ef02-4fd1-81d5-e5aa62bb8609%2Fblob) <!--- @huboard:{"order":0.9991004498350495,"milestone_order":0.9974035067263741} -->
priority
historical data too many entries for single point in time only one entry should exist for a single point in time not one per tank perhaps if you click on the single time it will show all tanks for that time in an accordion style drop down this is confusing huboard order milestone order
1
71,667
3,367,301,650
IssuesEvent
2015-11-22 02:27:22
ParadiseSS13/Paradise
https://api.github.com/repos/ParadiseSS13/Paradise
closed
Random seeds have the potential to have the mutationtoxin and amutationtoxin reagent.
Bug Easily Fixed High Priority
https://github.com/ParadiseSS13/paradise/blob/master/code/modules/hydroponics/seed.dm#L432-L433 Also known as, oh god, why is the botanist's food making everyone into slimes.
1.0
Random seeds have the potential to have the mutationtoxin and amutationtoxin reagent. - https://github.com/ParadiseSS13/paradise/blob/master/code/modules/hydroponics/seed.dm#L432-L433 Also known as, oh god, why is the botanist's food making everyone into slimes.
priority
random seeds have the potential to have the mutationtoxin and amutationtoxin reagent also known as oh god why is the botanist s food making everyone into slimes
1
542,571
15,862,657,041
IssuesEvent
2021-04-08 11:53:29
Couchers-org/couchers
https://api.github.com/repos/Couchers-org/couchers
closed
Errors not visible in jail page when adding location
bug frontend good first issue priority: high
If you don't have a location (due to not having put one in in alpha, and you try to log in, you're taken to the jail page to add a location. (This bug might also occur on other pages with a map component) If you just click through "Save" (or whatever the button says), the coordinate will be (0,0) or something, and the backend will return an error, however, the frontend doesn't show anything, so there's no feedback on why nothing is happening. This is pretty bad as it's the first screen the user is taken to after the alpha upgrade and can cause frustration!
1.0
Errors not visible in jail page when adding location - If you don't have a location (due to not having put one in in alpha, and you try to log in, you're taken to the jail page to add a location. (This bug might also occur on other pages with a map component) If you just click through "Save" (or whatever the button says), the coordinate will be (0,0) or something, and the backend will return an error, however, the frontend doesn't show anything, so there's no feedback on why nothing is happening. This is pretty bad as it's the first screen the user is taken to after the alpha upgrade and can cause frustration!
priority
errors not visible in jail page when adding location if you don t have a location due to not having put one in in alpha and you try to log in you re taken to the jail page to add a location this bug might also occur on other pages with a map component if you just click through save or whatever the button says the coordinate will be or something and the backend will return an error however the frontend doesn t show anything so there s no feedback on why nothing is happening this is pretty bad as it s the first screen the user is taken to after the alpha upgrade and can cause frustration
1
550,914
16,134,491,984
IssuesEvent
2021-04-29 09:58:26
IgniteUI/igniteui-angular
https://api.github.com/repos/IgniteUI/igniteui-angular
closed
Time column sorting - exception when null values are present
bug grid: column-types grid: sorting priority: high status: resolved
## Description When sorting a time column, if there are null values in the current data view an exception is thrown when applying the sorting expression. * igniteui-angular version: 12.0.x * browser: all ## Steps to reproduce 1. Open [this](https://staging.infragistics.com/products/ignite-ui-angular/angular/components/grid/column-types) sample 2. Try to sort the Order Time column ## Result Sorting is not applied as an error is thrown in the browser console. ## Expected result Sorting is applied and no error is thrown. ## Attachments Attach a sample if available, and screenshots, if applicable.
1.0
Time column sorting - exception when null values are present - ## Description When sorting a time column, if there are null values in the current data view an exception is thrown when applying the sorting expression. * igniteui-angular version: 12.0.x * browser: all ## Steps to reproduce 1. Open [this](https://staging.infragistics.com/products/ignite-ui-angular/angular/components/grid/column-types) sample 2. Try to sort the Order Time column ## Result Sorting is not applied as an error is thrown in the browser console. ## Expected result Sorting is applied and no error is thrown. ## Attachments Attach a sample if available, and screenshots, if applicable.
priority
time column sorting exception when null values are present description when sorting a time column if there are null values in the current data view an exception is thrown when applying the sorting expression igniteui angular version x browser all steps to reproduce open sample try to sort the order time column result sorting is not applied as an error is thrown in the browser console expected result sorting is applied and no error is thrown attachments attach a sample if available and screenshots if applicable
1
635,478
20,403,526,447
IssuesEvent
2022-02-23 00:43:29
louismacvux/bookwyrm
https://api.github.com/repos/louismacvux/bookwyrm
opened
Create a log in page
High Priority dev task
Users need to be able to log in to the website. So we need to create a log in page.
1.0
Create a log in page - Users need to be able to log in to the website. So we need to create a log in page.
priority
create a log in page users need to be able to log in to the website so we need to create a log in page
1
229,669
7,582,576,001
IssuesEvent
2018-04-25 05:14:03
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Errors are collected twice
L1/L2 for May Priority/Highest Severity/Critical component/Compiler
**Description:** <!-- Give a brief description of the issue --> <img width="1123" alt="screen shot 2018-04-21 at 9 52 24 am" src="https://user-images.githubusercontent.com/413016/39080603-6981d1ce-454f-11e8-8c88-d4431793a483.png"> **Steps to reproduce:** Have an object and function interfaces without function implementations. **Affected Versions:** Ballerina 0.970.0-beta6-SNAPSHOT **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
Errors are collected twice - **Description:** <!-- Give a brief description of the issue --> <img width="1123" alt="screen shot 2018-04-21 at 9 52 24 am" src="https://user-images.githubusercontent.com/413016/39080603-6981d1ce-454f-11e8-8c88-d4431793a483.png"> **Steps to reproduce:** Have an object and function interfaces without function implementations. **Affected Versions:** Ballerina 0.970.0-beta6-SNAPSHOT **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
priority
errors are collected twice description img width alt screen shot at am src steps to reproduce have an object and function interfaces without function implementations affected versions ballerina snapshot os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
1
585,932
17,538,518,868
IssuesEvent
2021-08-12 09:15:19
shivam5992/textstat
https://api.github.com/repos/shivam5992/textstat
closed
Move to GitHub actions
chore priority: high
Travis is rejecting builds for some reason. I suspect it's something to with the changes at Travis around free plans. Will probably just be easier to switch to GitHub actions.
1.0
Move to GitHub actions - Travis is rejecting builds for some reason. I suspect it's something to with the changes at Travis around free plans. Will probably just be easier to switch to GitHub actions.
priority
move to github actions travis is rejecting builds for some reason i suspect it s something to with the changes at travis around free plans will probably just be easier to switch to github actions
1
761,334
26,676,681,392
IssuesEvent
2023-01-26 14:45:59
crytic/crytic-compile
https://api.github.com/repos/crytic/crytic-compile
closed
Not obvious compilation error in Slither
bug high-priority
Running `slither <contract id>` when we don't have proper `solc` version installed via `solc-select` may result in an unambiguous error of `Invalid solc compilation` which does not describe what shall we do. Example: ``` $ slither 0xB8c77482e45F1F44dE1745F52C74426C631bDD52 Expecting value: line 1 column 1 (char 0) Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 412, in _run_solc ret = json.loads(stdout) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 721, in main_impl ) = process_all(filename, args, detector_classes, printer_classes) File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 71, in process_all compilations = compile_all(target, **vars(args)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 1076, in compile_all compilations.append(CryticCompile(target, **kwargs)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 137, in __init__ self._compile(**kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 987, in _compile self._platform.compile(self, **kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/etherscan.py", line 273, in compile working_dir=working_dir, File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 417, in _run_solc raise InvalidCompilation(f"Invalid solc compilation {stderr}") crytic_compile.platform.exceptions.InvalidCompilation: Invalid solc compilation ERROR:root:None ERROR:root:Error in 0xB8c77482e45F1F44dE1745F52C74426C631bDD52 ERROR:root:Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 412, in _run_solc ret = json.loads(stdout) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 721, in main_impl ) = process_all(filename, args, detector_classes, printer_classes) File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 71, in process_all compilations = compile_all(target, **vars(args)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 1076, in compile_all compilations.append(CryticCompile(target, **kwargs)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 137, in __init__ self._compile(**kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 987, in _compile self._platform.compile(self, **kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/etherscan.py", line 273, in compile working_dir=working_dir, File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 417, in _run_solc raise InvalidCompilation(f"Invalid solc compilation {stderr}") crytic_compile.platform.exceptions.InvalidCompilation: Invalid solc compilation ``` I investigated that by adding additional prints of stdout/stderr and process.returncode in https://github.com/crytic/crytic-compile/blob/d99e5a7dc8b80b812ef953a766457dd9eb4cbde1/crytic_compile/platform/solc.py#L402-L409 It turns out that the information that solc is not installed is printed out on stdout and so is not passed further to the user: ``` STDOUT: Version '0.4.12' not installed (set by SOLC_VERSION). Run `solc-select install 0.4.12`. STDERR: RETCODE 1 ``` I think we should either make `solc-select` to print info on stdout, or, make crytic-compile to take such stdout into account, or, make so it respects return code? I didn't check that but shouldn't return code be a good indicator whether the command succeeded?
1.0
Not obvious compilation error in Slither - Running `slither <contract id>` when we don't have proper `solc` version installed via `solc-select` may result in an unambiguous error of `Invalid solc compilation` which does not describe what shall we do. Example: ``` $ slither 0xB8c77482e45F1F44dE1745F52C74426C631bDD52 Expecting value: line 1 column 1 (char 0) Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 412, in _run_solc ret = json.loads(stdout) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 721, in main_impl ) = process_all(filename, args, detector_classes, printer_classes) File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 71, in process_all compilations = compile_all(target, **vars(args)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 1076, in compile_all compilations.append(CryticCompile(target, **kwargs)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 137, in __init__ self._compile(**kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 987, in _compile self._platform.compile(self, **kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/etherscan.py", line 273, in compile working_dir=working_dir, File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 417, in _run_solc raise InvalidCompilation(f"Invalid solc compilation {stderr}") crytic_compile.platform.exceptions.InvalidCompilation: Invalid solc compilation ERROR:root:None ERROR:root:Error in 0xB8c77482e45F1F44dE1745F52C74426C631bDD52 ERROR:root:Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 412, in _run_solc ret = json.loads(stdout) File "/usr/lib/python3.6/json/__init__.py", line 354, in loads return _default_decoder.decode(s) File "/usr/lib/python3.6/json/decoder.py", line 339, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/usr/lib/python3.6/json/decoder.py", line 357, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 721, in main_impl ) = process_all(filename, args, detector_classes, printer_classes) File "/usr/local/lib/python3.6/dist-packages/slither/__main__.py", line 71, in process_all compilations = compile_all(target, **vars(args)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 1076, in compile_all compilations.append(CryticCompile(target, **kwargs)) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 137, in __init__ self._compile(**kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/crytic_compile.py", line 987, in _compile self._platform.compile(self, **kwargs) File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/etherscan.py", line 273, in compile working_dir=working_dir, File "/usr/local/lib/python3.6/dist-packages/crytic_compile/platform/solc.py", line 417, in _run_solc raise InvalidCompilation(f"Invalid solc compilation {stderr}") crytic_compile.platform.exceptions.InvalidCompilation: Invalid solc compilation ``` I investigated that by adding additional prints of stdout/stderr and process.returncode in https://github.com/crytic/crytic-compile/blob/d99e5a7dc8b80b812ef953a766457dd9eb4cbde1/crytic_compile/platform/solc.py#L402-L409 It turns out that the information that solc is not installed is printed out on stdout and so is not passed further to the user: ``` STDOUT: Version '0.4.12' not installed (set by SOLC_VERSION). Run `solc-select install 0.4.12`. STDERR: RETCODE 1 ``` I think we should either make `solc-select` to print info on stdout, or, make crytic-compile to take such stdout into account, or, make so it respects return code? I didn't check that but shouldn't return code be a good indicator whether the command succeeded?
priority
not obvious compilation error in slither running slither when we don t have proper solc version installed via solc select may result in an unambiguous error of invalid solc compilation which does not describe what shall we do example slither expecting value line column char traceback most recent call last file usr local lib dist packages crytic compile platform solc py line in run solc ret json loads stdout file usr lib json init py line in loads return default decoder decode s file usr lib json decoder py line in decode obj end self raw decode s idx w s end file usr lib json decoder py line in raw decode raise jsondecodeerror expecting value s err value from none json decoder jsondecodeerror expecting value line column char during handling of the above exception another exception occurred traceback most recent call last file usr local lib dist packages slither main py line in main impl process all filename args detector classes printer classes file usr local lib dist packages slither main py line in process all compilations compile all target vars args file usr local lib dist packages crytic compile crytic compile py line in compile all compilations append cryticcompile target kwargs file usr local lib dist packages crytic compile crytic compile py line in init self compile kwargs file usr local lib dist packages crytic compile crytic compile py line in compile self platform compile self kwargs file usr local lib dist packages crytic compile platform etherscan py line in compile working dir working dir file usr local lib dist packages crytic compile platform solc py line in run solc raise invalidcompilation f invalid solc compilation stderr crytic compile platform exceptions invalidcompilation invalid solc compilation error root none error root error in error root traceback most recent call last file usr local lib dist packages crytic compile platform solc py line in run solc ret json loads stdout file usr lib json init py line in loads return default decoder decode s file usr lib json decoder py line in decode obj end self raw decode s idx w s end file usr lib json decoder py line in raw decode raise jsondecodeerror expecting value s err value from none json decoder jsondecodeerror expecting value line column char during handling of the above exception another exception occurred traceback most recent call last file usr local lib dist packages slither main py line in main impl process all filename args detector classes printer classes file usr local lib dist packages slither main py line in process all compilations compile all target vars args file usr local lib dist packages crytic compile crytic compile py line in compile all compilations append cryticcompile target kwargs file usr local lib dist packages crytic compile crytic compile py line in init self compile kwargs file usr local lib dist packages crytic compile crytic compile py line in compile self platform compile self kwargs file usr local lib dist packages crytic compile platform etherscan py line in compile working dir working dir file usr local lib dist packages crytic compile platform solc py line in run solc raise invalidcompilation f invalid solc compilation stderr crytic compile platform exceptions invalidcompilation invalid solc compilation i investigated that by adding additional prints of stdout stderr and process returncode in it turns out that the information that solc is not installed is printed out on stdout and so is not passed further to the user stdout version not installed set by solc version run solc select install stderr retcode i think we should either make solc select to print info on stdout or make crytic compile to take such stdout into account or make so it respects return code i didn t check that but shouldn t return code be a good indicator whether the command succeeded
1
449,611
12,971,991,253
IssuesEvent
2020-07-21 11:52:33
Tancho-Welcome-Week/tPlace
https://api.github.com/repos/Tancho-Welcome-Week/tPlace
closed
Redis code must be rewritten in JavaScript.
backend highPriority
Todo: - Find JS module to use Redis - Implement in the index.js file several APIs: - Initialize canvas - Set pixel - Get canvas - Set canvas
1.0
Redis code must be rewritten in JavaScript. - Todo: - Find JS module to use Redis - Implement in the index.js file several APIs: - Initialize canvas - Set pixel - Get canvas - Set canvas
priority
redis code must be rewritten in javascript todo find js module to use redis implement in the index js file several apis initialize canvas set pixel get canvas set canvas
1
153,197
5,886,993,890
IssuesEvent
2017-05-17 05:41:42
ThoughtWorksInc/treadmill
https://api.github.com/repos/ThoughtWorksInc/treadmill
closed
Join master and nodes to FreeIPA realm and configure treadmill LDAP env vars
Feature-Security inReview Priority-High Role-Administrator Size-Small (S)
So that : Treadmill master & nodes can be configured and are in sync with central FreeIPA server. "Assumption: Issue #24 has been completed. Tasks: Update ansible master and node roles for completing FreeIPA LDAP configuration.
1.0
Join master and nodes to FreeIPA realm and configure treadmill LDAP env vars - So that : Treadmill master & nodes can be configured and are in sync with central FreeIPA server. "Assumption: Issue #24 has been completed. Tasks: Update ansible master and node roles for completing FreeIPA LDAP configuration.
priority
join master and nodes to freeipa realm and configure treadmill ldap env vars so that treadmill master nodes can be configured and are in sync with central freeipa server assumption issue has been completed tasks update ansible master and node roles for completing freeipa ldap configuration
1
506,414
14,664,494,896
IssuesEvent
2020-12-29 12:08:31
fritzing/fritzing-app
https://api.github.com/repos/fritzing/fritzing-app
closed
Autoroute crashes
Priority-High bug imported
_From [georgeko...@gmail.com](https://code.google.com/u/107656141387250685553/) on May 19, 2013 19:08:10_ What steps will reproduce the problem? 1. Deleted all traces that were autorouted before 2. Run autorouter for 100 rounds 3. Autorouter crashes after the 100th round is run and next phase of autorouting starts (not sure what's its name) I believe the problem started once I deleted all the pre-autorouted traces and vias and tried to re-run autoroute. I could be wrong. I have been working on this file for some days before and autorouting was not crashing fritzing Attach your sketch file and/or custom part files to the bug report. What is the expected output? What do you see instead? ...(loads of similar debugging msgs before) "routing net 16, subnets 2, traces 1, routed 1" "routing net 17, subnets 2, traces 1, routed 1" "routing net 18, subnets 3, traces 1, routed 1" "routing net 20, subnets 5, traces 0, routed 0" "routing net 21, subnets 2, traces 0, routed 0" "routing net 22, subnets 5, traces 0, routed 0" "traceback zero points" "traceback zero points" "overlap not found" "overlap not found" "missing source connector" Segmentation fault What version of Fritzing are you using? On what operating system? 0.7.12 (b6882 2013-02-13) [Qt 4.8.3] on Linux Mint 13 also tried it fritzing.2013.02.25.pc.zip on WinXP 32bit and confirmed same issue is present Please provide any additional information below. **Attachment:** [v1.fzz](http://code.google.com/p/fritzing/issues/detail?id=2581) _Original issue: http://code.google.com/p/fritzing/issues/detail?id=2581_
1.0
Autoroute crashes - _From [georgeko...@gmail.com](https://code.google.com/u/107656141387250685553/) on May 19, 2013 19:08:10_ What steps will reproduce the problem? 1. Deleted all traces that were autorouted before 2. Run autorouter for 100 rounds 3. Autorouter crashes after the 100th round is run and next phase of autorouting starts (not sure what's its name) I believe the problem started once I deleted all the pre-autorouted traces and vias and tried to re-run autoroute. I could be wrong. I have been working on this file for some days before and autorouting was not crashing fritzing Attach your sketch file and/or custom part files to the bug report. What is the expected output? What do you see instead? ...(loads of similar debugging msgs before) "routing net 16, subnets 2, traces 1, routed 1" "routing net 17, subnets 2, traces 1, routed 1" "routing net 18, subnets 3, traces 1, routed 1" "routing net 20, subnets 5, traces 0, routed 0" "routing net 21, subnets 2, traces 0, routed 0" "routing net 22, subnets 5, traces 0, routed 0" "traceback zero points" "traceback zero points" "overlap not found" "overlap not found" "missing source connector" Segmentation fault What version of Fritzing are you using? On what operating system? 0.7.12 (b6882 2013-02-13) [Qt 4.8.3] on Linux Mint 13 also tried it fritzing.2013.02.25.pc.zip on WinXP 32bit and confirmed same issue is present Please provide any additional information below. **Attachment:** [v1.fzz](http://code.google.com/p/fritzing/issues/detail?id=2581) _Original issue: http://code.google.com/p/fritzing/issues/detail?id=2581_
priority
autoroute crashes from on may what steps will reproduce the problem deleted all traces that were autorouted before run autorouter for rounds autorouter crashes after the round is run and next phase of autorouting starts not sure what s its name i believe the problem started once i deleted all the pre autorouted traces and vias and tried to re run autoroute i could be wrong i have been working on this file for some days before and autorouting was not crashing fritzing attach your sketch file and or custom part files to the bug report what is the expected output what do you see instead loads of similar debugging msgs before routing net subnets traces routed routing net subnets traces routed routing net subnets traces routed routing net subnets traces routed routing net subnets traces routed routing net subnets traces routed traceback zero points traceback zero points overlap not found overlap not found missing source connector segmentation fault what version of fritzing are you using on what operating system on linux mint also tried it fritzing pc zip on winxp and confirmed same issue is present please provide any additional information below attachment original issue
1
209,028
7,164,397,538
IssuesEvent
2018-01-29 11:04:24
kinvolk/habitat-operator
https://api.github.com/repos/kinvolk/habitat-operator
closed
Run e2e tests against RBAC enabled Kubernetes cluster
enhancement priority: high
Currently we run our end-to-end tests against a non RBAC enabled cluster.
1.0
Run e2e tests against RBAC enabled Kubernetes cluster - Currently we run our end-to-end tests against a non RBAC enabled cluster.
priority
run tests against rbac enabled kubernetes cluster currently we run our end to end tests against a non rbac enabled cluster
1
454,549
13,103,434,358
IssuesEvent
2020-08-04 08:33:43
gambitph/Stackable
https://api.github.com/repos/gambitph/Stackable
opened
Number of Count Up block got bigger on Mobile Preview Mode
[block] count up bug high priority
Number of Count Up block got bigger on Mobile Preview Mode ![ezgif-6-1b4caf08f4b8](https://user-images.githubusercontent.com/51441886/89272100-3a3d5280-d670-11ea-8c69-55cc4f913680.gif)
1.0
Number of Count Up block got bigger on Mobile Preview Mode - Number of Count Up block got bigger on Mobile Preview Mode ![ezgif-6-1b4caf08f4b8](https://user-images.githubusercontent.com/51441886/89272100-3a3d5280-d670-11ea-8c69-55cc4f913680.gif)
priority
number of count up block got bigger on mobile preview mode number of count up block got bigger on mobile preview mode
1
580,570
17,261,347,910
IssuesEvent
2021-07-22 08:06:27
RetroMusicPlayer/RetroMusicPlayer
https://api.github.com/repos/RetroMusicPlayer/RetroMusicPlayer
closed
Opening now playing screen from notification makes the bottom buttons unresponsive
Priority: High bug v4
Video: https://t.me/retromusicapp/174775 When you open now playing screen from notification the favorites, queue, minimize (?) and menu doesnt work in Blur, Color, Fit, Material, Normal, Plain, Simple and Tiny
1.0
Opening now playing screen from notification makes the bottom buttons unresponsive - Video: https://t.me/retromusicapp/174775 When you open now playing screen from notification the favorites, queue, minimize (?) and menu doesnt work in Blur, Color, Fit, Material, Normal, Plain, Simple and Tiny
priority
opening now playing screen from notification makes the bottom buttons unresponsive video when you open now playing screen from notification the favorites queue minimize and menu doesnt work in blur color fit material normal plain simple and tiny
1
497,000
14,360,213,575
IssuesEvent
2020-11-30 16:34:08
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Godot hangs when opening project with main scene using Label and Custom Font
bug confirmed high priority regression topic:editor
**Godot version:** 3.2.3 **OS/device including version:** Windows 10 64 **Issue description:** Project can be run doing godot.exe inside the project folder but it hangs the editor when trying to open it. **Steps to reproduce:** 1 - Add the provided project to project manager 2 - Click on run. The game runs 3 - Click on edit The editor hangs. Using --verbose seems to indicate it hangs while opening some .wav files. Remove the .import folder doesn't seem to fix the issue. **Minimal reproduction project:** <!-- A small Godot project which reproduces the issue. Drag and drop a zip archive to upload it. --> https://drive.google.com/file/d/1e9o5gMpof8JHSvb0EwcD9dvfWtkZlbHK/view?usp=sharing
1.0
Godot hangs when opening project with main scene using Label and Custom Font - **Godot version:** 3.2.3 **OS/device including version:** Windows 10 64 **Issue description:** Project can be run doing godot.exe inside the project folder but it hangs the editor when trying to open it. **Steps to reproduce:** 1 - Add the provided project to project manager 2 - Click on run. The game runs 3 - Click on edit The editor hangs. Using --verbose seems to indicate it hangs while opening some .wav files. Remove the .import folder doesn't seem to fix the issue. **Minimal reproduction project:** <!-- A small Godot project which reproduces the issue. Drag and drop a zip archive to upload it. --> https://drive.google.com/file/d/1e9o5gMpof8JHSvb0EwcD9dvfWtkZlbHK/view?usp=sharing
priority
godot hangs when opening project with main scene using label and custom font godot version os device including version windows issue description project can be run doing godot exe inside the project folder but it hangs the editor when trying to open it steps to reproduce add the provided project to project manager click on run the game runs click on edit the editor hangs using verbose seems to indicate it hangs while opening some wav files remove the import folder doesn t seem to fix the issue minimal reproduction project
1
329,848
10,025,488,607
IssuesEvent
2019-07-17 02:29:10
bbc/simorgh
https://api.github.com/repos/bbc/simorgh
closed
Snyk high vulnerability in webpack@4.34.0 & webpack-cli@4.34.0
bug high priority simorgh-core-stream
**Describe the bug** Snyk has identified a new high vulnerabilities in webpack@4.34.0 and webpack-cli@4.34.0 https://app.snyk.io/org/news-articles/project/35916eba-8cf2-4204-89f8-45455a5b9d95/ Note this vuln is not fixed in the latest version of webpack@4.35.0 or webpack-cli @4.35.0
1.0
Snyk high vulnerability in webpack@4.34.0 & webpack-cli@4.34.0 - **Describe the bug** Snyk has identified a new high vulnerabilities in webpack@4.34.0 and webpack-cli@4.34.0 https://app.snyk.io/org/news-articles/project/35916eba-8cf2-4204-89f8-45455a5b9d95/ Note this vuln is not fixed in the latest version of webpack@4.35.0 or webpack-cli @4.35.0
priority
snyk high vulnerability in webpack webpack cli describe the bug snyk has identified a new high vulnerabilities in webpack and webpack cli note this vuln is not fixed in the latest version of webpack or webpack cli
1
107,686
4,313,615,048
IssuesEvent
2016-07-22 11:14:50
osoc16/mijn-viaa
https://api.github.com/repos/osoc16/mijn-viaa
closed
Standardize API json format for error checking
High Priority
[JSend format](https://labs.omniti.com/labs/jsend) for JSON responses ## Success **Required keys:** - status: Should always be set to "success". - data: Acts as the wrapper for any data returned by the API call. If the call returns no data (as in the last example), data should be set to null. ``` { status : "success", data : { ... } | null ``` ## Fail When an API call is rejected due to invalid data or call conditions, the JSend object's data key contains an object explaining what went wrong, typically a hash of validation errors. **Required keys:** - status: Should always be set to "fail". - data: Again, provides the wrapper for the details of why the request failed. If the reasons for failure correspond to POST values, the response object's keys SHOULD correspond to those POST values. ``` { "status" : "fail", "data" : { ... } // details why failed } ``` ## Error **Required keys:** - status: Should always be set to "error". - message: A meaningful, end-user-readable (or at the least log-worthy) message, explaining what went wrong. **Optional keys:** - code: A numeric code corresponding to the error, if applicable - data: A generic container for any other information about the error, i.e. the conditions that caused the error, stack traces, etc. - ``` { "status" : "error", "message" : "..." } ```
1.0
Standardize API json format for error checking - [JSend format](https://labs.omniti.com/labs/jsend) for JSON responses ## Success **Required keys:** - status: Should always be set to "success". - data: Acts as the wrapper for any data returned by the API call. If the call returns no data (as in the last example), data should be set to null. ``` { status : "success", data : { ... } | null ``` ## Fail When an API call is rejected due to invalid data or call conditions, the JSend object's data key contains an object explaining what went wrong, typically a hash of validation errors. **Required keys:** - status: Should always be set to "fail". - data: Again, provides the wrapper for the details of why the request failed. If the reasons for failure correspond to POST values, the response object's keys SHOULD correspond to those POST values. ``` { "status" : "fail", "data" : { ... } // details why failed } ``` ## Error **Required keys:** - status: Should always be set to "error". - message: A meaningful, end-user-readable (or at the least log-worthy) message, explaining what went wrong. **Optional keys:** - code: A numeric code corresponding to the error, if applicable - data: A generic container for any other information about the error, i.e. the conditions that caused the error, stack traces, etc. - ``` { "status" : "error", "message" : "..." } ```
priority
standardize api json format for error checking for json responses success required keys status should always be set to success data acts as the wrapper for any data returned by the api call if the call returns no data as in the last example data should be set to null status success data null fail when an api call is rejected due to invalid data or call conditions the jsend object s data key contains an object explaining what went wrong typically a hash of validation errors required keys status should always be set to fail data again provides the wrapper for the details of why the request failed if the reasons for failure correspond to post values the response object s keys should correspond to those post values status fail data details why failed error required keys status should always be set to error message a meaningful end user readable or at the least log worthy message explaining what went wrong optional keys code a numeric code corresponding to the error if applicable data a generic container for any other information about the error i e the conditions that caused the error stack traces etc status error message
1
603,335
18,541,402,181
IssuesEvent
2021-10-21 16:35:12
roq-trading/roq-issues
https://api.github.com/repos/roq-trading/roq-issues
opened
[roq-deribit] Round quantity and price to nearest tick-size / step-size
enhancement high priority
This change is prompted by Deribit's release note for the 22nd Oct. Order actions will be rejected if quantity or price isn't accurate. It appears the release note only relates to the REST/WS interfaces. As a precautionary measure, we implement this change for FIX. The following examples are **after** the change has been implemented (field 38 is quantity and 44 is price): CreateOrder ```fix 8=FIX.4.4|9=0000157|35=D|49=ROQ_TRADING|56=DERIBITSERVER|34=5|52=20211021-16:28:27.447|11=kAAC6QMAAQAA-bWA9hMX|54=1|38=1|44=63493.0|55=BTC-PERPETUAL|40=2|59=1|100010=roq-2-1001|10=077| ``` ModifyOrder ```fix 8=FIX.4.4|9=0000173|35=G|49=ROQ_TRADING|56=DERIBITSERVER|34=7|52=20211021-16:28:30.465|41=6759839183|11=BAAC6QMAAgAANsOu9hMX|60=20211021-16:28:27.449|54=1|38=1|40=2|44=63427.5|55=BTC-PERPETUAL|10=079| ```
1.0
[roq-deribit] Round quantity and price to nearest tick-size / step-size - This change is prompted by Deribit's release note for the 22nd Oct. Order actions will be rejected if quantity or price isn't accurate. It appears the release note only relates to the REST/WS interfaces. As a precautionary measure, we implement this change for FIX. The following examples are **after** the change has been implemented (field 38 is quantity and 44 is price): CreateOrder ```fix 8=FIX.4.4|9=0000157|35=D|49=ROQ_TRADING|56=DERIBITSERVER|34=5|52=20211021-16:28:27.447|11=kAAC6QMAAQAA-bWA9hMX|54=1|38=1|44=63493.0|55=BTC-PERPETUAL|40=2|59=1|100010=roq-2-1001|10=077| ``` ModifyOrder ```fix 8=FIX.4.4|9=0000173|35=G|49=ROQ_TRADING|56=DERIBITSERVER|34=7|52=20211021-16:28:30.465|41=6759839183|11=BAAC6QMAAgAANsOu9hMX|60=20211021-16:28:27.449|54=1|38=1|40=2|44=63427.5|55=BTC-PERPETUAL|10=079| ```
priority
round quantity and price to nearest tick size step size this change is prompted by deribit s release note for the oct order actions will be rejected if quantity or price isn t accurate it appears the release note only relates to the rest ws interfaces as a precautionary measure we implement this change for fix the following examples are after the change has been implemented field is quantity and is price createorder fix fix d roq trading deribitserver btc perpetual roq modifyorder fix fix g roq trading deribitserver btc perpetual
1
825,236
31,280,512,197
IssuesEvent
2023-08-22 09:17:22
kubermatic/kubermatic
https://api.github.com/repos/kubermatic/kubermatic
closed
Configure tags to be set on vSphere folder created by KKP
priority/high customer-request kind/feature sig/cluster-management
**Important**: Before closing this issue, consider this also needs implementation on the kubermatic/dashboard side to be usable by end users. Please provide details in https://github.com/kubermatic/dashboard/issues/6169 so it can be implemented. ### Description of the feature you would like to add / User story <!-- We've provided an example format of the user story below. You're free to use any other format as well. --> As a user creating a vSphere cluster I would like to configure a set of tags that get applied to the folder auto-created by KKP in order to have proper tagging on the top-level resource for my user cluster. ### Solution details <!-- Please, provide a bullet-pointed list or a few sentences of requirements that have to be met to mark the requested feature (user story) complete. --> - A field for vSphere folder tags to `ClusterSpec`. this could live in `spec.cloud.vsphere.tags` (so extending the `VsphereTags` struct) or in a separate `spec.cloud.vsphere.folderTags`. - Tags set this way are applied to folder objects created by KKP. - Only folders created by KKP are tagged, pre-created folders are not. - Optionally: An option to inherit those tags into MachineDeployments. If that is explored, the labelling should probably change from "folder tags" to "cluster tags" or something. ### Alternative approaches <!-- Optional: A clear and concise description of any alternative solutions or features you've considered. --> n/a ### Use cases <!-- Optional: Please try to explain some use cases why this feature would be valuable for you. If you can, please also explain what metrics would be improved for you with having this feature. --> n/a ### Additional information <!-- Optional: Additional information about the feature you're reporting. -->
1.0
Configure tags to be set on vSphere folder created by KKP - **Important**: Before closing this issue, consider this also needs implementation on the kubermatic/dashboard side to be usable by end users. Please provide details in https://github.com/kubermatic/dashboard/issues/6169 so it can be implemented. ### Description of the feature you would like to add / User story <!-- We've provided an example format of the user story below. You're free to use any other format as well. --> As a user creating a vSphere cluster I would like to configure a set of tags that get applied to the folder auto-created by KKP in order to have proper tagging on the top-level resource for my user cluster. ### Solution details <!-- Please, provide a bullet-pointed list or a few sentences of requirements that have to be met to mark the requested feature (user story) complete. --> - A field for vSphere folder tags to `ClusterSpec`. this could live in `spec.cloud.vsphere.tags` (so extending the `VsphereTags` struct) or in a separate `spec.cloud.vsphere.folderTags`. - Tags set this way are applied to folder objects created by KKP. - Only folders created by KKP are tagged, pre-created folders are not. - Optionally: An option to inherit those tags into MachineDeployments. If that is explored, the labelling should probably change from "folder tags" to "cluster tags" or something. ### Alternative approaches <!-- Optional: A clear and concise description of any alternative solutions or features you've considered. --> n/a ### Use cases <!-- Optional: Please try to explain some use cases why this feature would be valuable for you. If you can, please also explain what metrics would be improved for you with having this feature. --> n/a ### Additional information <!-- Optional: Additional information about the feature you're reporting. -->
priority
configure tags to be set on vsphere folder created by kkp important before closing this issue consider this also needs implementation on the kubermatic dashboard side to be usable by end users please provide details in so it can be implemented description of the feature you would like to add user story as a user creating a vsphere cluster i would like to configure a set of tags that get applied to the folder auto created by kkp in order to have proper tagging on the top level resource for my user cluster solution details a field for vsphere folder tags to clusterspec this could live in spec cloud vsphere tags so extending the vspheretags struct or in a separate spec cloud vsphere foldertags tags set this way are applied to folder objects created by kkp only folders created by kkp are tagged pre created folders are not optionally an option to inherit those tags into machinedeployments if that is explored the labelling should probably change from folder tags to cluster tags or something alternative approaches n a use cases n a additional information
1
415,406
12,129,016,211
IssuesEvent
2020-04-22 21:37:24
LBNL-ETA/BEDES-Manager
https://api.github.com/repos/LBNL-ETA/BEDES-Manager
closed
Application Term DataType
bug high priority
Instead of the textbox labeled "Type of Term" on the Application Mapping page (see attached image), the textbox should be labeled "Data Type" and the contents should be the DataType of the Application/BEDES term (e.g., Decimal, Integer) [Application Term DataType Issue.docx](https://github.com/LBNL-ETA/BEDES-Manager/files/3933824/Application.Term.DataType.Issue.docx)
1.0
Application Term DataType - Instead of the textbox labeled "Type of Term" on the Application Mapping page (see attached image), the textbox should be labeled "Data Type" and the contents should be the DataType of the Application/BEDES term (e.g., Decimal, Integer) [Application Term DataType Issue.docx](https://github.com/LBNL-ETA/BEDES-Manager/files/3933824/Application.Term.DataType.Issue.docx)
priority
application term datatype instead of the textbox labeled type of term on the application mapping page see attached image the textbox should be labeled data type and the contents should be the datatype of the application bedes term e g decimal integer
1
598,623
18,249,151,666
IssuesEvent
2021-10-02 00:06:22
datalab-dev/covid_worksite_exposure
https://api.github.com/repos/datalab-dev/covid_worksite_exposure
opened
Map not playing
High Priority
Map appears but no data (current or historical) displays. First noticed on 10/1/2021. Issue appears on both the blog post embedding and at https://datalab-dev.github.io/covid_worksite_exposure/.
1.0
Map not playing - Map appears but no data (current or historical) displays. First noticed on 10/1/2021. Issue appears on both the blog post embedding and at https://datalab-dev.github.io/covid_worksite_exposure/.
priority
map not playing map appears but no data current or historical displays first noticed on issue appears on both the blog post embedding and at
1
812,022
30,312,308,123
IssuesEvent
2023-07-10 13:33:49
scaleway/terraform-provider-scaleway
https://api.github.com/repos/scaleway/terraform-provider-scaleway
closed
Updating a private network on a redis cluster forces its recreation
bug redis priority:highest
<!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * scaleway_redis_cluster ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "scaleway_vpc_private_network" "pn" {} resource "scaleway_redis_cluster" "main" { name = "test_redis" version = "7.0.5" node_type = "RED1-XS" user_name = "my_initial_user" password = "thiZ_is_v&ry_s3cret" cluster_size = 1 private_network { id = scaleway_vpc_private_network.pn.id service_ips = [ "10.12.2.0/20", ] } } ``` ### Debug Output After adding a subnet "10.12.3.0/20": ``` # scaleway_redis_cluster.main must be replaced -/+ resource "scaleway_redis_cluster" "main" { + certificate = (known after apply) ~ created_at = "2023-06-23T14:59:07Z" -> (known after apply) ~ id = "fr-par-1/cfdff96e-c164-4313-9d8e-2ebaedb08c13" -> (known after apply) name = "test_redis" - settings = {} -> null ~ updated_at = "2023-06-23T15:00:52Z" -> (known after apply) ~ zone = "fr-par-1" -> (known after apply) # (5 unchanged attributes hidden) - private_network { # forces replacement - endpoint_id = "82589763-e116-400b-9a16-2f3f70e4648e" -> null - id = "fr-par-1/5ee648a6-5307-452c-a98d-ebc6b5a070ed" -> null - service_ips = [ - "10.12.2.0/20", ] -> null - zone = "fr-par-1" -> null } + private_network { # forces replacement + endpoint_id = (known after apply) + id = "fr-par-1/5ee648a6-5307-452c-a98d-ebc6b5a070ed" + service_ips = [ + "10.12.2.0/20", + "10.12.3.0/20", ] + zone = (known after apply) } + public_network { + id = (known after apply) + ips = (known after apply) + port = (known after apply) } } Plan: 1 to add, 0 to change, 1 to destroy. ``` ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior <!--- What should have happened? ---> ### Actual Behavior <!--- What actually happened? ---> ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: ---> * #0000
1.0
Updating a private network on a redis cluster forces its recreation - <!--- Please note the following potential times when an issue might be in Terraform core: * [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues * [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues * [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues * [Registry](https://registry.terraform.io/) issues * Spans resources across multiple providers If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead. ---> <!--- Please keep this note for the community ---> ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * scaleway_redis_cluster ### Terraform Configuration Files <!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code ---> ```hcl resource "scaleway_vpc_private_network" "pn" {} resource "scaleway_redis_cluster" "main" { name = "test_redis" version = "7.0.5" node_type = "RED1-XS" user_name = "my_initial_user" password = "thiZ_is_v&ry_s3cret" cluster_size = 1 private_network { id = scaleway_vpc_private_network.pn.id service_ips = [ "10.12.2.0/20", ] } } ``` ### Debug Output After adding a subnet "10.12.3.0/20": ``` # scaleway_redis_cluster.main must be replaced -/+ resource "scaleway_redis_cluster" "main" { + certificate = (known after apply) ~ created_at = "2023-06-23T14:59:07Z" -> (known after apply) ~ id = "fr-par-1/cfdff96e-c164-4313-9d8e-2ebaedb08c13" -> (known after apply) name = "test_redis" - settings = {} -> null ~ updated_at = "2023-06-23T15:00:52Z" -> (known after apply) ~ zone = "fr-par-1" -> (known after apply) # (5 unchanged attributes hidden) - private_network { # forces replacement - endpoint_id = "82589763-e116-400b-9a16-2f3f70e4648e" -> null - id = "fr-par-1/5ee648a6-5307-452c-a98d-ebc6b5a070ed" -> null - service_ips = [ - "10.12.2.0/20", ] -> null - zone = "fr-par-1" -> null } + private_network { # forces replacement + endpoint_id = (known after apply) + id = "fr-par-1/5ee648a6-5307-452c-a98d-ebc6b5a070ed" + service_ips = [ + "10.12.2.0/20", + "10.12.3.0/20", ] + zone = (known after apply) } + public_network { + id = (known after apply) + ips = (known after apply) + port = (known after apply) } } Plan: 1 to add, 0 to change, 1 to destroy. ``` ### Panic Output <!--- If Terraform produced a panic, please provide a link to a GitHub Gist containing the output of the `crash.log`. ---> ### Expected Behavior <!--- What should have happened? ---> ### Actual Behavior <!--- What actually happened? ---> ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. `terraform apply` ### Important Factoids <!--- Are there anything atypical about your accounts that we should know? ---> ### References <!--- Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor documentation? For example: ---> * #0000
priority
updating a private network on a redis cluster forces its recreation please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version affected resource s scaleway redis cluster terraform configuration files hcl resource scaleway vpc private network pn resource scaleway redis cluster main name test redis version node type xs user name my initial user password thiz is v ry cluster size private network id scaleway vpc private network pn id service ips debug output after adding a subnet scaleway redis cluster main must be replaced resource scaleway redis cluster main certificate known after apply created at known after apply id fr par known after apply name test redis settings null updated at known after apply zone fr par known after apply unchanged attributes hidden private network forces replacement endpoint id null id fr par null service ips null zone fr par null private network forces replacement endpoint id known after apply id fr par service ips zone known after apply public network id known after apply ips known after apply port known after apply plan to add to change to destroy panic output expected behavior actual behavior steps to reproduce terraform apply important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor documentation for example
1
791,836
27,879,349,027
IssuesEvent
2023-03-21 18:09:32
rstudio/gt
https://api.github.com/repos/rstudio/gt
closed
gtsave to accept lists for pagination V1
Difficulty: [3] Advanced Effort: [3] High Priority: [3] High Type: ★ Enhancement Focus: Word Output Focus: Pharma
## Prework - [x] Read and abide by **gt**'s [code of conduct](https://www.contributor-covenant.org/version/2/0/code_of_conduct/) and [contributing guidelines](https://github.com/rstudio/gt/blob/master/.github/CONTRIBUTING.md). - [x] Search for duplicates among the [existing issues](https://github.com/rstudio/gt/issues) (both open and closed). ## Proposal Describe the new feature clearly and concisely. If applicable, write a minimal example in R code or pseudo-code to show input, usage, and desired output. This is a suggestion to allow gtsave to accept a list of gts (and perhaps some meta information such as entries into tab_options or special titles, etc too?) to created paginated reports of gts. This is mainly useful for the rtf, word, and pdf outputs, but I can see it being useful for all the outputs to be able to generate multiple tables in one go to a single document. This is to support work going towards pagination & was previously discussed with @rich-iannone An example of how this might look code wise is: ```{r} ## tables by car make library(tidyverse) library(gt) gtcars |> group_split(mfr) |> map(function(x){ x |> select(-mfr) |> gt() |> tab_caption(paste("Cars by",unique(x$mfr)) }) |> gtsave("tables_by_mfr.pdf") ```
1.0
gtsave to accept lists for pagination V1 - ## Prework - [x] Read and abide by **gt**'s [code of conduct](https://www.contributor-covenant.org/version/2/0/code_of_conduct/) and [contributing guidelines](https://github.com/rstudio/gt/blob/master/.github/CONTRIBUTING.md). - [x] Search for duplicates among the [existing issues](https://github.com/rstudio/gt/issues) (both open and closed). ## Proposal Describe the new feature clearly and concisely. If applicable, write a minimal example in R code or pseudo-code to show input, usage, and desired output. This is a suggestion to allow gtsave to accept a list of gts (and perhaps some meta information such as entries into tab_options or special titles, etc too?) to created paginated reports of gts. This is mainly useful for the rtf, word, and pdf outputs, but I can see it being useful for all the outputs to be able to generate multiple tables in one go to a single document. This is to support work going towards pagination & was previously discussed with @rich-iannone An example of how this might look code wise is: ```{r} ## tables by car make library(tidyverse) library(gt) gtcars |> group_split(mfr) |> map(function(x){ x |> select(-mfr) |> gt() |> tab_caption(paste("Cars by",unique(x$mfr)) }) |> gtsave("tables_by_mfr.pdf") ```
priority
gtsave to accept lists for pagination prework read and abide by gt s and search for duplicates among the both open and closed proposal describe the new feature clearly and concisely if applicable write a minimal example in r code or pseudo code to show input usage and desired output this is a suggestion to allow gtsave to accept a list of gts and perhaps some meta information such as entries into tab options or special titles etc too to created paginated reports of gts this is mainly useful for the rtf word and pdf outputs but i can see it being useful for all the outputs to be able to generate multiple tables in one go to a single document this is to support work going towards pagination was previously discussed with rich iannone an example of how this might look code wise is r tables by car make library tidyverse library gt gtcars group split mfr map function x x select mfr gt tab caption paste cars by unique x mfr gtsave tables by mfr pdf
1
592,625
17,912,596,927
IssuesEvent
2021-09-09 07:42:36
grafana/k6-docs
https://api.github.com/repos/grafana/k6-docs
opened
Rewrite the execution context variables page
Area: Community Content Type: Enhancement Priority: High Status: Available Area: Cloud Content
As I mentioned in https://github.com/grafana/k6-docs/pull/396, now that we have the `k6/execution` API, we should rewrite the [execution context variables](https://k6.io/docs/using-k6/execution-context-variables/) docs page. In essence, we should highly discourage the usage of `__VU` and `__ITER` and instead encourage their respective `k6/execution` alternatives, or better yet, the test-wide unique identifiers that will work equally well locally and in the cloud
1.0
Rewrite the execution context variables page - As I mentioned in https://github.com/grafana/k6-docs/pull/396, now that we have the `k6/execution` API, we should rewrite the [execution context variables](https://k6.io/docs/using-k6/execution-context-variables/) docs page. In essence, we should highly discourage the usage of `__VU` and `__ITER` and instead encourage their respective `k6/execution` alternatives, or better yet, the test-wide unique identifiers that will work equally well locally and in the cloud
priority
rewrite the execution context variables page as i mentioned in now that we have the execution api we should rewrite the docs page in essence we should highly discourage the usage of vu and iter and instead encourage their respective execution alternatives or better yet the test wide unique identifiers that will work equally well locally and in the cloud
1
692,548
23,739,677,959
IssuesEvent
2022-08-31 11:17:40
nestauk/ai_genomics
https://api.github.com/repos/nestauk/ai_genomics
closed
Identify AI Genomics people/organisations
Epic 2: epic lite 1 HIGH PRIORITY
Produce a list of AI Genomics people/organisations. From Harry's email: We are looking for a mixture of people, but especially those who: * Understand and work on genomic science, and particularly the application of AI to genomics; * Understand and have insight into the larger trends in private and public funding for genomics research and R&D; or * Understand and have insight into the larger trends in terms of how government, industry and other actors are thinking about researching and deploying genomic science. I was wondering if it would be possible to use the data so far generated by the scientometric analysis to help identify relevant figures. Specifically, is it possible to look through the data * To identify people who match or look likely to match the above criteria; and/or * to identify companies, research centres and other entities that may be home of people worth including on the longlist?
1.0
Identify AI Genomics people/organisations - Produce a list of AI Genomics people/organisations. From Harry's email: We are looking for a mixture of people, but especially those who: * Understand and work on genomic science, and particularly the application of AI to genomics; * Understand and have insight into the larger trends in private and public funding for genomics research and R&D; or * Understand and have insight into the larger trends in terms of how government, industry and other actors are thinking about researching and deploying genomic science. I was wondering if it would be possible to use the data so far generated by the scientometric analysis to help identify relevant figures. Specifically, is it possible to look through the data * To identify people who match or look likely to match the above criteria; and/or * to identify companies, research centres and other entities that may be home of people worth including on the longlist?
priority
identify ai genomics people organisations produce a list of ai genomics people organisations from harry s email we are looking for a mixture of people but especially those who understand and work on genomic science and particularly the application of ai to genomics understand and have insight into the larger trends in private and public funding for genomics research and r d or understand and have insight into the larger trends in terms of how government industry and other actors are thinking about researching and deploying genomic science i was wondering if it would be possible to use the data so far generated by the scientometric analysis to help identify relevant figures specifically is it possible to look through the data to identify people who match or look likely to match the above criteria and or to identify companies research centres and other entities that may be home of people worth including on the longlist
1
60,875
3,135,066,644
IssuesEvent
2015-09-10 13:43:29
syl20bnr/spacemacs
https://api.github.com/repos/syl20bnr/spacemacs
closed
Error starting spacemacs when non-existent layer is used
Core Fixed in `develop` High Priority regression :-(
Having a non-existent layer in `dotspacemacs-configuration-layers` causes an error during emacs startup: "Wrong type argument: stringp, nil", as reported in #2951. Spacemacs home buffer is shown and a warning buffer, but otherwise it seems like spacemacs didn't load - no key-bindings and no elpa packages have been loaded. To reproduce, add `foo` to `dotspacemacs-configuratoin-layers` and start emacs. This situation can happen if a user misspells a layer name, or if a layer is removed or renamed. In my case, I had the old `slime` layer enabled. I had the `slime` layer enabled for quite a time now, using develop, so probably a recent change is the source this issue. Debugging showed the error rises from `configuration-layer//declare-layers`. I may have time later to look into this further and submit a patch, but I can't guarantee it.
1.0
Error starting spacemacs when non-existent layer is used - Having a non-existent layer in `dotspacemacs-configuration-layers` causes an error during emacs startup: "Wrong type argument: stringp, nil", as reported in #2951. Spacemacs home buffer is shown and a warning buffer, but otherwise it seems like spacemacs didn't load - no key-bindings and no elpa packages have been loaded. To reproduce, add `foo` to `dotspacemacs-configuratoin-layers` and start emacs. This situation can happen if a user misspells a layer name, or if a layer is removed or renamed. In my case, I had the old `slime` layer enabled. I had the `slime` layer enabled for quite a time now, using develop, so probably a recent change is the source this issue. Debugging showed the error rises from `configuration-layer//declare-layers`. I may have time later to look into this further and submit a patch, but I can't guarantee it.
priority
error starting spacemacs when non existent layer is used having a non existent layer in dotspacemacs configuration layers causes an error during emacs startup wrong type argument stringp nil as reported in spacemacs home buffer is shown and a warning buffer but otherwise it seems like spacemacs didn t load no key bindings and no elpa packages have been loaded to reproduce add foo to dotspacemacs configuratoin layers and start emacs this situation can happen if a user misspells a layer name or if a layer is removed or renamed in my case i had the old slime layer enabled i had the slime layer enabled for quite a time now using develop so probably a recent change is the source this issue debugging showed the error rises from configuration layer declare layers i may have time later to look into this further and submit a patch but i can t guarantee it
1
668,716
22,594,851,310
IssuesEvent
2022-06-29 01:01:38
avuserow/endgame-singularity
https://api.github.com/repos/avuserow/endgame-singularity
closed
Power States
auto-migrated Priority-High Component-Logic Type-Enhancement Milestone-0.31
``` Currently there are two hidden power states, Sleep and Active. Sleep is what makes the Sleep task work, and all other tasks are Active. This system should be expanded and given its own GUI element. Desired power states: 1. Stasis. The base is completely shut down and put into stasis, dropping its maintenance cost by 50% (rounded down) and its detection chance by 90%. It takes 10 minutes per size to put a base into or take a base out of stasis, during which time no bonus is given. Bases entering, exiting, or in Stasis do not support the AI. 2. Sleep. The base is running in standby, using less power than normal. 50% reduction in detection chance. 3. Active. The base functions normally. 4. Overclocked. The base gets a 50% bonus to CPU time, but has a 50% higher detection chance as well. 5. Suicidal. The base gets a 200% bonus to CPU time, but the detection chance is set to 50%/day. All modes except Active should have associated technologies. ``` Original issue reported on code.google.com by `funnyman3595` on 5 May 2008 at 7:13
1.0
Power States - ``` Currently there are two hidden power states, Sleep and Active. Sleep is what makes the Sleep task work, and all other tasks are Active. This system should be expanded and given its own GUI element. Desired power states: 1. Stasis. The base is completely shut down and put into stasis, dropping its maintenance cost by 50% (rounded down) and its detection chance by 90%. It takes 10 minutes per size to put a base into or take a base out of stasis, during which time no bonus is given. Bases entering, exiting, or in Stasis do not support the AI. 2. Sleep. The base is running in standby, using less power than normal. 50% reduction in detection chance. 3. Active. The base functions normally. 4. Overclocked. The base gets a 50% bonus to CPU time, but has a 50% higher detection chance as well. 5. Suicidal. The base gets a 200% bonus to CPU time, but the detection chance is set to 50%/day. All modes except Active should have associated technologies. ``` Original issue reported on code.google.com by `funnyman3595` on 5 May 2008 at 7:13
priority
power states currently there are two hidden power states sleep and active sleep is what makes the sleep task work and all other tasks are active this system should be expanded and given its own gui element desired power states stasis the base is completely shut down and put into stasis dropping its maintenance cost by rounded down and its detection chance by it takes minutes per size to put a base into or take a base out of stasis during which time no bonus is given bases entering exiting or in stasis do not support the ai sleep the base is running in standby using less power than normal reduction in detection chance active the base functions normally overclocked the base gets a bonus to cpu time but has a higher detection chance as well suicidal the base gets a bonus to cpu time but the detection chance is set to day all modes except active should have associated technologies original issue reported on code google com by on may at
1
475,175
13,688,258,873
IssuesEvent
2020-09-30 11:25:42
AY2021S1-CS2103T-W13-1/tp
https://api.github.com/repos/AY2021S1-CS2103T-W13-1/tp
closed
Draft an About Us page
priority.High type.Task
This is to inform people who land on this codebase about the team behind Eva.
1.0
Draft an About Us page - This is to inform people who land on this codebase about the team behind Eva.
priority
draft an about us page this is to inform people who land on this codebase about the team behind eva
1
43,157
2,883,627,486
IssuesEvent
2015-06-11 13:19:35
curationexperts/mira
https://api.github.com/repos/curationexperts/mira
closed
Mint release candidate
high priority
Since the batch changes are breaking, let's call this MIRA 4.0 **ACCEPTANCE** - [x] Initial release candidate of tufts_models 4.0.0.rc1 - [x] Initial release candidate of MIRA 4.0.0.rc1- pin to tufts_models 4.0.0.rc1 - [x] Initial release candidate of TROVE 1.1.0.rc1 - pin to tufts_models 4.0.0.rc1
1.0
Mint release candidate - Since the batch changes are breaking, let's call this MIRA 4.0 **ACCEPTANCE** - [x] Initial release candidate of tufts_models 4.0.0.rc1 - [x] Initial release candidate of MIRA 4.0.0.rc1- pin to tufts_models 4.0.0.rc1 - [x] Initial release candidate of TROVE 1.1.0.rc1 - pin to tufts_models 4.0.0.rc1
priority
mint release candidate since the batch changes are breaking let s call this mira acceptance initial release candidate of tufts models initial release candidate of mira pin to tufts models initial release candidate of trove pin to tufts models
1
128,797
5,076,058,840
IssuesEvent
2016-12-27 23:25:52
nilsschmidt1337/ldparteditor
https://api.github.com/repos/nilsschmidt1337/ldparteditor
closed
When you insert a protractor, the order of the vertex triplet is not deterministic.
bug high-priority
_Philo wrote:_ > Looks like Protractor has his own will: sometimes angle appears on first selected vertex, sometimes second, sometimes third... Maybe the best way would be to display all 3 angles?
1.0
When you insert a protractor, the order of the vertex triplet is not deterministic. - _Philo wrote:_ > Looks like Protractor has his own will: sometimes angle appears on first selected vertex, sometimes second, sometimes third... Maybe the best way would be to display all 3 angles?
priority
when you insert a protractor the order of the vertex triplet is not deterministic philo wrote looks like protractor has his own will sometimes angle appears on first selected vertex sometimes second sometimes third maybe the best way would be to display all angles
1
544,236
15,891,391,405
IssuesEvent
2021-04-10 19:03:29
nilshah98/Knowledge-Lake
https://api.github.com/repos/nilshah98/Knowledge-Lake
closed
Decouple Google Sheets and Data
🟥 priority: high
**Context:** Currently the flow is as such -> 1. Extract `JSON` data from Pocket API, Twitter, Hypothesis 2. Clean this `JSON` data 3. Convert this cleaned `JSON` data to form suitable for Google Sheets, ie. `[ [ A1,B1 ], [ A2,B2 ] ]` form, which is array of arrays 4. Update the data on Google Sheets **Problem:** - Currently the cleaned `JSON` data is not stored anywhere, and so is not persistent - We are directly converting to array of arrays in-memory and updating on Google Sheets - Hence, if any other service wants to consume the same, they cannot **Requirement:** - [ ] Store the cleaned JSON locally - [ ] Update data on Google Sheets by consuming this stored data **Future:** - Create an API out of the locally stored JSON - Consume data directly on Google Sheets using Google Sheet Scripts or finding some JSON importer on Google Sheets Marketplace
1.0
Decouple Google Sheets and Data - **Context:** Currently the flow is as such -> 1. Extract `JSON` data from Pocket API, Twitter, Hypothesis 2. Clean this `JSON` data 3. Convert this cleaned `JSON` data to form suitable for Google Sheets, ie. `[ [ A1,B1 ], [ A2,B2 ] ]` form, which is array of arrays 4. Update the data on Google Sheets **Problem:** - Currently the cleaned `JSON` data is not stored anywhere, and so is not persistent - We are directly converting to array of arrays in-memory and updating on Google Sheets - Hence, if any other service wants to consume the same, they cannot **Requirement:** - [ ] Store the cleaned JSON locally - [ ] Update data on Google Sheets by consuming this stored data **Future:** - Create an API out of the locally stored JSON - Consume data directly on Google Sheets using Google Sheet Scripts or finding some JSON importer on Google Sheets Marketplace
priority
decouple google sheets and data context currently the flow is as such extract json data from pocket api twitter hypothesis clean this json data convert this cleaned json data to form suitable for google sheets ie form which is array of arrays update the data on google sheets problem currently the cleaned json data is not stored anywhere and so is not persistent we are directly converting to array of arrays in memory and updating on google sheets hence if any other service wants to consume the same they cannot requirement store the cleaned json locally update data on google sheets by consuming this stored data future create an api out of the locally stored json consume data directly on google sheets using google sheet scripts or finding some json importer on google sheets marketplace
1
511,045
14,852,395,893
IssuesEvent
2021-01-18 08:32:01
Joeb3219/recitations
https://api.github.com/repos/Joeb3219/recitations
closed
Add Weeks Settings Component
Feature High Priority
Weeks settings component allows an administrator to define the weeks that the course will be occurring for, as well as specify default lesson plans for each week, or if the week is not occurring (Spring break, fall break, etc) Key features: - [ ] Define semester weeks - [ ] Specify lesson plans for each week
1.0
Add Weeks Settings Component - Weeks settings component allows an administrator to define the weeks that the course will be occurring for, as well as specify default lesson plans for each week, or if the week is not occurring (Spring break, fall break, etc) Key features: - [ ] Define semester weeks - [ ] Specify lesson plans for each week
priority
add weeks settings component weeks settings component allows an administrator to define the weeks that the course will be occurring for as well as specify default lesson plans for each week or if the week is not occurring spring break fall break etc key features define semester weeks specify lesson plans for each week
1
354,349
10,565,908,032
IssuesEvent
2019-10-05 15:02:52
geneontology/go-annotation
https://api.github.com/repos/geneontology/go-annotation
reopened
Annotation review: GO:0004012 phospholipid-translocating ATPase activity
annotation review high priority
~@marcfeuermann and I still need to add suggestions~ Suggestions now added - see spreadsheet https://docs.google.com/spreadsheets/d/1Pzr5O9-9ZTsbG75IfDH1pho1GqQ708hN_UNrAAofTx8/edit#gid=0
1.0
Annotation review: GO:0004012 phospholipid-translocating ATPase activity - ~@marcfeuermann and I still need to add suggestions~ Suggestions now added - see spreadsheet https://docs.google.com/spreadsheets/d/1Pzr5O9-9ZTsbG75IfDH1pho1GqQ708hN_UNrAAofTx8/edit#gid=0
priority
annotation review go phospholipid translocating atpase activity marcfeuermann and i still need to add suggestions suggestions now added see spreadsheet
1
814,741
30,519,933,826
IssuesEvent
2023-07-19 07:21:12
Ecological-Complexity-Lab/emln_package
https://api.github.com/repos/Ecological-Complexity-Lab/emln_package
closed
Multiples
High priority
Consult with @Noa164 What we decided to do with the strange network definition "multiples". Then search ALL the files for this and fix.
1.0
Multiples - Consult with @Noa164 What we decided to do with the strange network definition "multiples". Then search ALL the files for this and fix.
priority
multiples consult with what we decided to do with the strange network definition multiples then search all the files for this and fix
1
360,249
10,685,899,682
IssuesEvent
2019-10-22 13:32:09
onaio/reveal-frontend
https://api.github.com/repos/onaio/reveal-frontend
closed
Assign Teams to FI Plans
Priority: High enhancement
Focus investigations should include a multi-select dropdown to select organizations to be assigned to the single plan-jurisdiction. * ~~Render (Async?)Select component in New (FI) Plan Form~~ * ~~Render (Async?)Select component in Update (FI) Plan Form~~ * ~~[POST deprecated Assignments to OpenSRP to retire them](https://github.com/onaio/reveal-frontend/issues/411#issuecomment-538620498)~~ * ~~[POST new Assignments to OpenSRP to create new ones](https://github.com/onaio/reveal-frontend/issues/411#issuecomment-538620498)~~
1.0
Assign Teams to FI Plans - Focus investigations should include a multi-select dropdown to select organizations to be assigned to the single plan-jurisdiction. * ~~Render (Async?)Select component in New (FI) Plan Form~~ * ~~Render (Async?)Select component in Update (FI) Plan Form~~ * ~~[POST deprecated Assignments to OpenSRP to retire them](https://github.com/onaio/reveal-frontend/issues/411#issuecomment-538620498)~~ * ~~[POST new Assignments to OpenSRP to create new ones](https://github.com/onaio/reveal-frontend/issues/411#issuecomment-538620498)~~
priority
assign teams to fi plans focus investigations should include a multi select dropdown to select organizations to be assigned to the single plan jurisdiction render async select component in new fi plan form render async select component in update fi plan form
1
712,255
24,488,301,032
IssuesEvent
2022-10-09 18:36:30
onulstore/Onulstore-BE
https://api.github.com/repos/onulstore/Onulstore-BE
closed
Feat Order
Priority : High Type: Feature Status : In Progress
## 설명 Order관련 기능 추가 ## 할일 - [x] 주문 상태 추가 - [x] 주문 상태 변경 - [x] 배송 방법 추가 - [x] 환불 - [x] 하나의 주문 내역 조회
1.0
Feat Order - ## 설명 Order관련 기능 추가 ## 할일 - [x] 주문 상태 추가 - [x] 주문 상태 변경 - [x] 배송 방법 추가 - [x] 환불 - [x] 하나의 주문 내역 조회
priority
feat order 설명 order관련 기능 추가 할일 주문 상태 추가 주문 상태 변경 배송 방법 추가 환불 하나의 주문 내역 조회
1
796,933
28,132,474,188
IssuesEvent
2023-04-01 02:16:31
AY2223S2-CS2103T-F12-3/tp
https://api.github.com/repos/AY2223S2-CS2103T-F12-3/tp
opened
Improve `edituser` command
priority.High type.Bug severity.Medium
* Documentation in help window has `INDEX` as a parameter, causing confusion * #245
1.0
Improve `edituser` command - * Documentation in help window has `INDEX` as a parameter, causing confusion * #245
priority
improve edituser command documentation in help window has index as a parameter causing confusion
1
472,067
13,615,869,217
IssuesEvent
2020-09-23 14:56:58
ansible/galaxy_ng
https://api.github.com/repos/ansible/galaxy_ng
closed
Importer: Implement runner that downloads artifacts directly
area/importer priority/high sprint/6 status/ready-for-release type/enhancement
To comply with AppSRE requirements galaxy-importer must move away from building images dynamically via BuildConfigs.
1.0
Importer: Implement runner that downloads artifacts directly - To comply with AppSRE requirements galaxy-importer must move away from building images dynamically via BuildConfigs.
priority
importer implement runner that downloads artifacts directly to comply with appsre requirements galaxy importer must move away from building images dynamically via buildconfigs
1
349,719
10,472,370,571
IssuesEvent
2019-09-23 10:01:29
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Need to add a new config for STSTimeToLive in new config model
Priority/High config
In the old config model, If you want to configure an expiration time for the security token, you need to add the following configuration in the <IS_HOME>/repository/conf/carbon.xml file, under the <Server> element: `<STSTimeToLive>1800000</STSTimeToLive>` Refer doc (https://is.docs.wso2.com/en/5.9.0/learn/configuring-ws-federation-single-sign-on/) But there is no similar config can be found for new config model (deployment.toml).
1.0
Need to add a new config for STSTimeToLive in new config model - In the old config model, If you want to configure an expiration time for the security token, you need to add the following configuration in the <IS_HOME>/repository/conf/carbon.xml file, under the <Server> element: `<STSTimeToLive>1800000</STSTimeToLive>` Refer doc (https://is.docs.wso2.com/en/5.9.0/learn/configuring-ws-federation-single-sign-on/) But there is no similar config can be found for new config model (deployment.toml).
priority
need to add a new config for ststimetolive in new config model in the old config model if you want to configure an expiration time for the security token you need to add the following configuration in the repository conf carbon xml file under the element refer doc but there is no similar config can be found for new config model deployment toml
1
766,722
26,896,143,081
IssuesEvent
2023-02-06 12:34:20
fractal-analytics-platform/fractal-server
https://api.github.com/repos/fractal-analytics-platform/fractal-server
closed
Separate server/user job-execution folders
High Priority
(based on several ongoing discussions, and edited accordingly) --- # Files Two kinds of files should be "exchanged" between server and cluster user 1. Files written by fractal (AKA the user running the server), from python: * JSON file with task input &#x1F534; * Sbatch scripts * Pickle file with cfut job input * Log files 2. Files written by test01 * Log files for the tasks (written by test01, from python) * JSON file with task output (written by test01, from python) * SLURM stdout/stderr (written by SLURM) * Pickle file with cfut job output (written by test01, from python) # sudo requirements: * fractal can run `sudo -u test01 sbatch something.sh` * fractal can run `sudo -u test01 cat something.sh` * fractal can run `sudo -u test01 ls something.sh` # Folder structure ### Current version JOBDIR and USERJOBDIR are the same folder, owned by fractal and with 777 permissions. All the rest is the same as in future version. ### Future version: Folders owned by fractal: * ARTIFACTS=/some/share/fractal/folder/artifacts 755, visible to all users * JOBDIR=/some/share/fractal/folder/artifacts/workflow_01_job_02 755, visible to all users * TASKDIR=/some/share/fractal/folder/tasks 755, visible to all users Folder owned by user (i.e. test01): * PROJECT_DIR=/some/share/test01/folder 700, but visible to fractal through sudo-cat * USERJOBDIR=/some/share/test01/folder/tmp_wf_01_job_01 700, but visible to fractal through sudo-cat Notes: * PROJECT_DIR must exist (check upon project creation) * USERJOBDIR must exist (mkdir at the beginning of submit.sh) &#x1F534; # Workflow execution ## Job preparation ### Current version * fractal writes args.json and submit.sh input.pickle to JOBDIR (and also logs, but that doesn't matter because test01 has no access to workflow.log) ### Future version * same as current version ## Job execution (by user) ### Current version Steps: * fractal runs `sudo -u test01 sbatch JOBDIR/submit.sh` (OK, since JOBDIR is 777) * test01 (via sbatch) runs `srun /some/path/python -m cfut JOBDIR/input.pickle` (OK, since JOBDIR is 777) * test01 (via srun) runs `TASKDIR/some/path/python TASKDIR/some/path/task.py --input JOBDIR/args.json` (OK, since JOBDIR is 777) * test01 (via sbatch) writes slurm err/out files to JOBDIR (OK, since JOBDIR is 777) ### Future version `JOBDIR/submit.sh` will look like ``` mkdir -p USERJOBDIR &#x1F534; srun /some/path/python -m cfut JOBDIR/input.pickle --output-dir USERJOBDIR ``` Steps: * fractal runs `sudo -u test01 sbatch JOBDIR/submit.sh` (OK, since JOBDIR is 755) * test01 (via sbatch) runs `mkdir -p USERJOBDIR` &#x1F534; (OK, since USERJOBDIR belongs to test01) * test01 (via sbatch) runs `srun /some/path/python -m cfut JOBDIR/input.pickle --output-dir USERJOBDIR` (OK, since JOBDIR is 755 and USERJOBDIR belongs to test01) * test01 (via srun) runs `TASKDIR/some/path/python TASKDIR/some/path/task.py --input JOBDIR/args.json --output USERJOBDIR/output.json` (OK, since JOBDIR is 755 and USERJOBDIR belongs to test01) * test01 (via sbatch) writes slurm err/out files to USERJOBDIR (OK, since USERJOBDIR belongs to test01) ## Post job execution (by fractal) ### Current version * fractal directly (AKA from python) polls existence of JOBDIR/out.pickle (which fractal has access to) every second, and proceeds as soon as the file is there * if JOBDIR/out.pickle is missing AND squeue polling (taking place e.g. every minute) says job is over, then proceed with error handling ### Future version * fractal indirectly (AKA subproces.run("sudo -u test01 ls ..")) polls USERJOBDIR/out.pickle every second, and proceeds as soon as the file is there * as soon as USERJOBDIR/out.pickle exists: - trigger copy of all task-related files from USERJOBDIR into JOBDIR (via sudo-cat) * if USERJOBDIR/out.pickle is missing AND squeue polling (taking place e.g. every minute) says SLURM job is over, then - trigger copy of all task-related files from USERJOBDIR into JOBDIR (via sudo-cat) - proceed with error handling ---- # Comment about abstractions We may say that the proposed update goes one step further towards abstracting out the data exchange between server and user. **Note**: this is not an abstraction in the sense that we will then be able to just replace one implementation with another one (via a standardized interface, like what we did for concurrent.futures standardized Executor), but rather an abstraction as in "we roughly know which code blocks will need to be replaced with new logic in future versions". V0 (current): * server->user is trivial, thanks to 777 folder * user->server is trivial, thanks to 777 folder V1 (proposed change) * server->user is trivial, thanks to 755 folder * user->server is non-trivial, and requires sudo privileges V2 (one day) * Both server->user and user->server are implemented via custom script send_data.py V3 (one day) * Both server->user and user->server are implemented with open sockets on both sides, pushing/receiving data back and forth --- # Open questions: * Is `local` backend meant as a test for the slurm one, or rather as an independently useful backend? We think that the local backend is important in itself, and not just for testing. * Based on the previous point, the current proposed solution is to make the backend-independent runner component flexible, and then only implement the proposed change for the slurm backend. * Number of files in job folders is currently too large, because each task will have something like 7 files (out/err/in.json/out.json/slurm.sh/slurm.out/slurm.err). If a dataset has 400 wells and a workflow has 5 tasks, this means 14k files, which is not acceptable. Ref #474
1.0
Separate server/user job-execution folders - (based on several ongoing discussions, and edited accordingly) --- # Files Two kinds of files should be "exchanged" between server and cluster user 1. Files written by fractal (AKA the user running the server), from python: * JSON file with task input &#x1F534; * Sbatch scripts * Pickle file with cfut job input * Log files 2. Files written by test01 * Log files for the tasks (written by test01, from python) * JSON file with task output (written by test01, from python) * SLURM stdout/stderr (written by SLURM) * Pickle file with cfut job output (written by test01, from python) # sudo requirements: * fractal can run `sudo -u test01 sbatch something.sh` * fractal can run `sudo -u test01 cat something.sh` * fractal can run `sudo -u test01 ls something.sh` # Folder structure ### Current version JOBDIR and USERJOBDIR are the same folder, owned by fractal and with 777 permissions. All the rest is the same as in future version. ### Future version: Folders owned by fractal: * ARTIFACTS=/some/share/fractal/folder/artifacts 755, visible to all users * JOBDIR=/some/share/fractal/folder/artifacts/workflow_01_job_02 755, visible to all users * TASKDIR=/some/share/fractal/folder/tasks 755, visible to all users Folder owned by user (i.e. test01): * PROJECT_DIR=/some/share/test01/folder 700, but visible to fractal through sudo-cat * USERJOBDIR=/some/share/test01/folder/tmp_wf_01_job_01 700, but visible to fractal through sudo-cat Notes: * PROJECT_DIR must exist (check upon project creation) * USERJOBDIR must exist (mkdir at the beginning of submit.sh) &#x1F534; # Workflow execution ## Job preparation ### Current version * fractal writes args.json and submit.sh input.pickle to JOBDIR (and also logs, but that doesn't matter because test01 has no access to workflow.log) ### Future version * same as current version ## Job execution (by user) ### Current version Steps: * fractal runs `sudo -u test01 sbatch JOBDIR/submit.sh` (OK, since JOBDIR is 777) * test01 (via sbatch) runs `srun /some/path/python -m cfut JOBDIR/input.pickle` (OK, since JOBDIR is 777) * test01 (via srun) runs `TASKDIR/some/path/python TASKDIR/some/path/task.py --input JOBDIR/args.json` (OK, since JOBDIR is 777) * test01 (via sbatch) writes slurm err/out files to JOBDIR (OK, since JOBDIR is 777) ### Future version `JOBDIR/submit.sh` will look like ``` mkdir -p USERJOBDIR &#x1F534; srun /some/path/python -m cfut JOBDIR/input.pickle --output-dir USERJOBDIR ``` Steps: * fractal runs `sudo -u test01 sbatch JOBDIR/submit.sh` (OK, since JOBDIR is 755) * test01 (via sbatch) runs `mkdir -p USERJOBDIR` &#x1F534; (OK, since USERJOBDIR belongs to test01) * test01 (via sbatch) runs `srun /some/path/python -m cfut JOBDIR/input.pickle --output-dir USERJOBDIR` (OK, since JOBDIR is 755 and USERJOBDIR belongs to test01) * test01 (via srun) runs `TASKDIR/some/path/python TASKDIR/some/path/task.py --input JOBDIR/args.json --output USERJOBDIR/output.json` (OK, since JOBDIR is 755 and USERJOBDIR belongs to test01) * test01 (via sbatch) writes slurm err/out files to USERJOBDIR (OK, since USERJOBDIR belongs to test01) ## Post job execution (by fractal) ### Current version * fractal directly (AKA from python) polls existence of JOBDIR/out.pickle (which fractal has access to) every second, and proceeds as soon as the file is there * if JOBDIR/out.pickle is missing AND squeue polling (taking place e.g. every minute) says job is over, then proceed with error handling ### Future version * fractal indirectly (AKA subproces.run("sudo -u test01 ls ..")) polls USERJOBDIR/out.pickle every second, and proceeds as soon as the file is there * as soon as USERJOBDIR/out.pickle exists: - trigger copy of all task-related files from USERJOBDIR into JOBDIR (via sudo-cat) * if USERJOBDIR/out.pickle is missing AND squeue polling (taking place e.g. every minute) says SLURM job is over, then - trigger copy of all task-related files from USERJOBDIR into JOBDIR (via sudo-cat) - proceed with error handling ---- # Comment about abstractions We may say that the proposed update goes one step further towards abstracting out the data exchange between server and user. **Note**: this is not an abstraction in the sense that we will then be able to just replace one implementation with another one (via a standardized interface, like what we did for concurrent.futures standardized Executor), but rather an abstraction as in "we roughly know which code blocks will need to be replaced with new logic in future versions". V0 (current): * server->user is trivial, thanks to 777 folder * user->server is trivial, thanks to 777 folder V1 (proposed change) * server->user is trivial, thanks to 755 folder * user->server is non-trivial, and requires sudo privileges V2 (one day) * Both server->user and user->server are implemented via custom script send_data.py V3 (one day) * Both server->user and user->server are implemented with open sockets on both sides, pushing/receiving data back and forth --- # Open questions: * Is `local` backend meant as a test for the slurm one, or rather as an independently useful backend? We think that the local backend is important in itself, and not just for testing. * Based on the previous point, the current proposed solution is to make the backend-independent runner component flexible, and then only implement the proposed change for the slurm backend. * Number of files in job folders is currently too large, because each task will have something like 7 files (out/err/in.json/out.json/slurm.sh/slurm.out/slurm.err). If a dataset has 400 wells and a workflow has 5 tasks, this means 14k files, which is not acceptable. Ref #474
priority
separate server user job execution folders based on several ongoing discussions and edited accordingly files two kinds of files should be exchanged between server and cluster user files written by fractal aka the user running the server from python json file with task input sbatch scripts pickle file with cfut job input log files files written by log files for the tasks written by from python json file with task output written by from python slurm stdout stderr written by slurm pickle file with cfut job output written by from python sudo requirements fractal can run sudo u sbatch something sh fractal can run sudo u cat something sh fractal can run sudo u ls something sh folder structure current version jobdir and userjobdir are the same folder owned by fractal and with permissions all the rest is the same as in future version future version folders owned by fractal artifacts some share fractal folder artifacts visible to all users jobdir some share fractal folder artifacts workflow job visible to all users taskdir some share fractal folder tasks visible to all users folder owned by user i e project dir some share folder but visible to fractal through sudo cat userjobdir some share folder tmp wf job but visible to fractal through sudo cat notes project dir must exist check upon project creation userjobdir must exist mkdir at the beginning of submit sh workflow execution job preparation current version fractal writes args json and submit sh input pickle to jobdir and also logs but that doesn t matter because has no access to workflow log future version same as current version job execution by user current version steps fractal runs sudo u sbatch jobdir submit sh ok since jobdir is via sbatch runs srun some path python m cfut jobdir input pickle ok since jobdir is via srun runs taskdir some path python taskdir some path task py input jobdir args json ok since jobdir is via sbatch writes slurm err out files to jobdir ok since jobdir is future version jobdir submit sh will look like mkdir p userjobdir srun some path python m cfut jobdir input pickle output dir userjobdir steps fractal runs sudo u sbatch jobdir submit sh ok since jobdir is via sbatch runs mkdir p userjobdir ok since userjobdir belongs to via sbatch runs srun some path python m cfut jobdir input pickle output dir userjobdir ok since jobdir is and userjobdir belongs to via srun runs taskdir some path python taskdir some path task py input jobdir args json output userjobdir output json ok since jobdir is and userjobdir belongs to via sbatch writes slurm err out files to userjobdir ok since userjobdir belongs to post job execution by fractal current version fractal directly aka from python polls existence of jobdir out pickle which fractal has access to every second and proceeds as soon as the file is there if jobdir out pickle is missing and squeue polling taking place e g every minute says job is over then proceed with error handling future version fractal indirectly aka subproces run sudo u ls polls userjobdir out pickle every second and proceeds as soon as the file is there as soon as userjobdir out pickle exists trigger copy of all task related files from userjobdir into jobdir via sudo cat if userjobdir out pickle is missing and squeue polling taking place e g every minute says slurm job is over then trigger copy of all task related files from userjobdir into jobdir via sudo cat proceed with error handling comment about abstractions we may say that the proposed update goes one step further towards abstracting out the data exchange between server and user note this is not an abstraction in the sense that we will then be able to just replace one implementation with another one via a standardized interface like what we did for concurrent futures standardized executor but rather an abstraction as in we roughly know which code blocks will need to be replaced with new logic in future versions current server user is trivial thanks to folder user server is trivial thanks to folder proposed change server user is trivial thanks to folder user server is non trivial and requires sudo privileges one day both server user and user server are implemented via custom script send data py one day both server user and user server are implemented with open sockets on both sides pushing receiving data back and forth open questions is local backend meant as a test for the slurm one or rather as an independently useful backend we think that the local backend is important in itself and not just for testing based on the previous point the current proposed solution is to make the backend independent runner component flexible and then only implement the proposed change for the slurm backend number of files in job folders is currently too large because each task will have something like files out err in json out json slurm sh slurm out slurm err if a dataset has wells and a workflow has tasks this means files which is not acceptable ref
1
186,012
6,732,760,320
IssuesEvent
2017-10-18 12:45:36
salesagility/SuiteCRM
https://api.github.com/repos/salesagility/SuiteCRM
closed
Cannot edit Layouts in Module Builder
bug Fix Proposed High Priority Resolved: Next Release
#### Issue I've created a new module but I cannot edit the Layouts. A PHP error is reported in the web server logs. #### Expected Behavior In the Module Builder, create a new Package, then a Module within that Package, then expand Layouts in the navigation tree and click on (eg) "EditView". I expect the view editor to appear. #### Actual Behavior Instead nothing happens and the following error is logged in the web server log file: `[error] 3560#3560: *1 FastCGI sent in stderr: "PHP message: PHP Fatal error: Declaration of UndeployedMetaDataImplementation::getFileName($view, $moduleName, $packageName, $type = MB_BASEMETADATALOCATION) must be compatible with AbstractMetaDataImplementation::getFileName($view, $moduleName, $type = MB_CUSTOMMETADATALOCATION) in /var/www/crm.eatworkart.com/modules/ModuleBuilder/parsers/views/UndeployedMetaDataImplementation.php on line 49" while reading response header from upstream, client: 2001:8b0:3b1:beef:ca60:ff:fec2:b8cf, server: crm.eatworkart.com, request: "POST /index.php?to_pdf=1&sugar_body_only=1&module=ModuleBuilder&MB=true&action=editLayout&view=editview&view_module=Colours&view_package=WRS_TestPackage HTTP/2.0", upstream: "fastcgi://unix:/var/run/php/php7.0-fpm.sock:", host: "crm.eatworkart.com", referrer: "https://crm.eatworkart.com/index.php?module=ModuleBuilder&action=index&type=mb"` #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.9.5 * Browser name and version: Chrome for Linux, 64-bit, Version 60.0.3112.113 * Environment name and version (e.g. MySQL, PHP 7): MariaDB 10.1.26, PHP 7.0.19 * Operating System and version (e.g Ubuntu 16.04): Debian 9.1 "stretch"
1.0
Cannot edit Layouts in Module Builder - #### Issue I've created a new module but I cannot edit the Layouts. A PHP error is reported in the web server logs. #### Expected Behavior In the Module Builder, create a new Package, then a Module within that Package, then expand Layouts in the navigation tree and click on (eg) "EditView". I expect the view editor to appear. #### Actual Behavior Instead nothing happens and the following error is logged in the web server log file: `[error] 3560#3560: *1 FastCGI sent in stderr: "PHP message: PHP Fatal error: Declaration of UndeployedMetaDataImplementation::getFileName($view, $moduleName, $packageName, $type = MB_BASEMETADATALOCATION) must be compatible with AbstractMetaDataImplementation::getFileName($view, $moduleName, $type = MB_CUSTOMMETADATALOCATION) in /var/www/crm.eatworkart.com/modules/ModuleBuilder/parsers/views/UndeployedMetaDataImplementation.php on line 49" while reading response header from upstream, client: 2001:8b0:3b1:beef:ca60:ff:fec2:b8cf, server: crm.eatworkart.com, request: "POST /index.php?to_pdf=1&sugar_body_only=1&module=ModuleBuilder&MB=true&action=editLayout&view=editview&view_module=Colours&view_package=WRS_TestPackage HTTP/2.0", upstream: "fastcgi://unix:/var/run/php/php7.0-fpm.sock:", host: "crm.eatworkart.com", referrer: "https://crm.eatworkart.com/index.php?module=ModuleBuilder&action=index&type=mb"` #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.9.5 * Browser name and version: Chrome for Linux, 64-bit, Version 60.0.3112.113 * Environment name and version (e.g. MySQL, PHP 7): MariaDB 10.1.26, PHP 7.0.19 * Operating System and version (e.g Ubuntu 16.04): Debian 9.1 "stretch"
priority
cannot edit layouts in module builder issue i ve created a new module but i cannot edit the layouts a php error is reported in the web server logs expected behavior in the module builder create a new package then a module within that package then expand layouts in the navigation tree and click on eg editview i expect the view editor to appear actual behavior instead nothing happens and the following error is logged in the web server log file fastcgi sent in stderr php message php fatal error declaration of undeployedmetadataimplementation getfilename view modulename packagename type mb basemetadatalocation must be compatible with abstractmetadataimplementation getfilename view modulename type mb custommetadatalocation in var www crm eatworkart com modules modulebuilder parsers views undeployedmetadataimplementation php on line while reading response header from upstream client beef ff server crm eatworkart com request post index php to pdf sugar body only module modulebuilder mb true action editlayout view editview view module colours view package wrs testpackage http upstream fastcgi unix var run php fpm sock host crm eatworkart com referrer your environment suitecrm version used browser name and version chrome for linux bit version environment name and version e g mysql php mariadb php operating system and version e g ubuntu debian stretch
1
278,010
8,635,044,597
IssuesEvent
2018-11-22 19:57:39
slarse/repomate
https://api.github.com/repos/slarse/repomate
opened
hook results even if no user-specified plugins are loaded
bug high priority
As there is only a check for whether there are any plugins loaded at all, the default plugin now causes hook results to be printed even if the user has not specified any plugins. As the default plugin does not have any reporting plugins, nothing interesting is printed. It's only bloat.
1.0
hook results even if no user-specified plugins are loaded - As there is only a check for whether there are any plugins loaded at all, the default plugin now causes hook results to be printed even if the user has not specified any plugins. As the default plugin does not have any reporting plugins, nothing interesting is printed. It's only bloat.
priority
hook results even if no user specified plugins are loaded as there is only a check for whether there are any plugins loaded at all the default plugin now causes hook results to be printed even if the user has not specified any plugins as the default plugin does not have any reporting plugins nothing interesting is printed it s only bloat
1
203,350
7,060,116,665
IssuesEvent
2018-01-05 06:53:07
wso2/message-broker
https://api.github.com/repos/wso2/message-broker
opened
Subscription related REST API
Module/broker-core Priority/High Severity/Major Type/New Feature
### Description Implement a REST service to handle subscription related admin tasks.
1.0
Subscription related REST API - ### Description Implement a REST service to handle subscription related admin tasks.
priority
subscription related rest api description implement a rest service to handle subscription related admin tasks
1
586,733
17,595,815,752
IssuesEvent
2021-08-17 04:51:38
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
closed
[Челенджі.Карусель] Challenge> Incorrect text alignment in tasks
bug UI Priority: High
Environment: Windows 10, version 92.0.4515.107, (64) Reproducible: always Build found: last commit **Steps to reproduce** Go to https://speak-ukrainian.org.ua/dev/ Login as admin@gmail.com, 'Password'="admin"; Go to 'Челенджі' > 'Мовомаратон' Click on content box 'День 1' from Завдання челенджу ![day1](https://user-images.githubusercontent.com/87965983/129439296-cd39a85f-c2ed-4585-bccf-60039fc03796.png) **Actual result** Incorrect text alignment. Text overlay. ![text alignment](https://user-images.githubusercontent.com/87965983/129439413-e4f2ab1c-bf04-4b93-ad95-f726870e7312.png) **Expected result** Text alignment according to mock up. ![mock up](https://user-images.githubusercontent.com/87965983/129439550-cb59b756-deb5-4ddd-afab-0d8e95322b73.png)
1.0
[Челенджі.Карусель] Challenge> Incorrect text alignment in tasks - Environment: Windows 10, version 92.0.4515.107, (64) Reproducible: always Build found: last commit **Steps to reproduce** Go to https://speak-ukrainian.org.ua/dev/ Login as admin@gmail.com, 'Password'="admin"; Go to 'Челенджі' > 'Мовомаратон' Click on content box 'День 1' from Завдання челенджу ![day1](https://user-images.githubusercontent.com/87965983/129439296-cd39a85f-c2ed-4585-bccf-60039fc03796.png) **Actual result** Incorrect text alignment. Text overlay. ![text alignment](https://user-images.githubusercontent.com/87965983/129439413-e4f2ab1c-bf04-4b93-ad95-f726870e7312.png) **Expected result** Text alignment according to mock up. ![mock up](https://user-images.githubusercontent.com/87965983/129439550-cb59b756-deb5-4ddd-afab-0d8e95322b73.png)
priority
challenge incorrect text alignment in tasks environment windows version reproducible always build found last commit steps to reproduce go to login as admin gmail com password admin go to челенджі мовомаратон click on content box день from завдання челенджу actual result incorrect text alignment text overlay expected result text alignment according to mock up
1
383,105
11,350,358,937
IssuesEvent
2020-01-24 08:35:04
Cxbx-Reloaded/Cxbx-Reloaded
https://api.github.com/repos/Cxbx-Reloaded/Cxbx-Reloaded
closed
DSound's Stream Packet Lockup Issue
HLE bug high-priority sound
Title says it all. The affected titles are: - Grand Theft Auto III **(confirmed)** - After garage checkpoint, audio kept looping which prevent first mission checkpoint to start. - Grand Theft Auto Vice City **(confirmed)** - During mission intro then cause a lockup. - Azurik - Rise of Perathia **(confirmed)** - At start of gameplay. - ~~Project Gotham Racing 1 & 2 **(likely)**~~ Unrelative to stream's packet issue - After intros, become stuck at loading screen. Remove silence.wma file allow title to progress further. **UPDATE:** Found out the source of cause. Stream's single packet buffer size is larger than the host's buffer size. This is the cause of lockup for same packet.
1.0
DSound's Stream Packet Lockup Issue - Title says it all. The affected titles are: - Grand Theft Auto III **(confirmed)** - After garage checkpoint, audio kept looping which prevent first mission checkpoint to start. - Grand Theft Auto Vice City **(confirmed)** - During mission intro then cause a lockup. - Azurik - Rise of Perathia **(confirmed)** - At start of gameplay. - ~~Project Gotham Racing 1 & 2 **(likely)**~~ Unrelative to stream's packet issue - After intros, become stuck at loading screen. Remove silence.wma file allow title to progress further. **UPDATE:** Found out the source of cause. Stream's single packet buffer size is larger than the host's buffer size. This is the cause of lockup for same packet.
priority
dsound s stream packet lockup issue title says it all the affected titles are grand theft auto iii confirmed after garage checkpoint audio kept looping which prevent first mission checkpoint to start grand theft auto vice city confirmed during mission intro then cause a lockup azurik rise of perathia confirmed at start of gameplay project gotham racing likely unrelative to stream s packet issue after intros become stuck at loading screen remove silence wma file allow title to progress further update found out the source of cause stream s single packet buffer size is larger than the host s buffer size this is the cause of lockup for same packet
1
564,756
16,740,459,489
IssuesEvent
2021-06-11 09:08:24
codetapacademy/codetap.academy
https://api.github.com/repos/codetapacademy/codetap.academy
closed
feat: display completed percentage in the course view page
Priority: High Status: Available Type: Enhancement
display completed percentage in the course view page
1.0
feat: display completed percentage in the course view page - display completed percentage in the course view page
priority
feat display completed percentage in the course view page display completed percentage in the course view page
1
208,909
7,162,814,227
IssuesEvent
2018-01-29 03:04:50
IfyAniefuna/experiment_metadata
https://api.github.com/repos/IfyAniefuna/experiment_metadata
opened
No volume on electron acceptor
enhancement high priority
Remove both form for specifying volume and validation of volume with spreadsheet input.
1.0
No volume on electron acceptor - Remove both form for specifying volume and validation of volume with spreadsheet input.
priority
no volume on electron acceptor remove both form for specifying volume and validation of volume with spreadsheet input
1
444,659
12,815,429,337
IssuesEvent
2020-07-05 02:31:44
getting-things-gnome/gtg
https://api.github.com/repos/getting-things-gnome/gtg
opened
Traceback when trying to send a task via email (plugin) in the flatpak version
packaging priority:high reproducible-in-git
When you try to use the "Send by email" action in the task editor's "…" menu when the plugin is activated, you get this traceback (doesn't occur when running the git version manually, IIRC): ``` File "/app/lib/python3.7/site-packages/GTG/plugins/send_email/sendEmail.py", line 69, in onTbTaskButton Gio.app_info_get_default_for_uri_scheme('mailto').launch_uris( AttributeError: 'NoneType' object has no attribute 'launch_uris' ``` Is it just an error in terms of the Flatpak package needing to have permissions for a portal for this, or this requires changes in the plugin's code?
1.0
Traceback when trying to send a task via email (plugin) in the flatpak version - When you try to use the "Send by email" action in the task editor's "…" menu when the plugin is activated, you get this traceback (doesn't occur when running the git version manually, IIRC): ``` File "/app/lib/python3.7/site-packages/GTG/plugins/send_email/sendEmail.py", line 69, in onTbTaskButton Gio.app_info_get_default_for_uri_scheme('mailto').launch_uris( AttributeError: 'NoneType' object has no attribute 'launch_uris' ``` Is it just an error in terms of the Flatpak package needing to have permissions for a portal for this, or this requires changes in the plugin's code?
priority
traceback when trying to send a task via email plugin in the flatpak version when you try to use the send by email action in the task editor s … menu when the plugin is activated you get this traceback doesn t occur when running the git version manually iirc file app lib site packages gtg plugins send email sendemail py line in ontbtaskbutton gio app info get default for uri scheme mailto launch uris attributeerror nonetype object has no attribute launch uris is it just an error in terms of the flatpak package needing to have permissions for a portal for this or this requires changes in the plugin s code
1
647,971
21,161,029,127
IssuesEvent
2022-04-07 09:22:47
Betarena/scores
https://api.github.com/repos/Betarena/scores
opened
Reducing svg file size
enhancement high priority
The following image is occupying 262Kb, its a very large size it needs to be compressed, check the alternatives attached. ![begambleawareorg_blac](https://user-images.githubusercontent.com/37311649/162166682-bd8df3d7-0970-4b90-9e29-b85164a8e554.png) ![begambleawareorg_blac](https://user-images.githubusercontent.com/37311649/162166688-9479f179-3c51-4051-b8cb-f548907c0229.svg)
1.0
Reducing svg file size - The following image is occupying 262Kb, its a very large size it needs to be compressed, check the alternatives attached. ![begambleawareorg_blac](https://user-images.githubusercontent.com/37311649/162166682-bd8df3d7-0970-4b90-9e29-b85164a8e554.png) ![begambleawareorg_blac](https://user-images.githubusercontent.com/37311649/162166688-9479f179-3c51-4051-b8cb-f548907c0229.svg)
priority
reducing svg file size the following image is occupying its a very large size it needs to be compressed check the alternatives attached
1
344,772
10,349,639,050
IssuesEvent
2019-09-04 23:17:55
oslc-op/jira-migration-landfill
https://api.github.com/repos/oslc-op/jira-migration-landfill
closed
Does a ServiceProvider have one Service per oslc:domain value, or does it reflect the server&apos;s desired container structure.
Core: Main Spec Priority: High Xtra: Jira
In [OSLCCORE-23](https://issues.oasis-open.org/browse/OSLCCORE-23 "Should a ServiceProviderCatalog or ServiceProvider resource be an LDPC?") we agreed that a Service resource should be an LDPC, but didn‘t state what its members were. My reading of that ticket‘s proposal is that the members are defined by the domain. In recent emails, there has been a discussion about what the members of a Service should be. My latest comments that triggered me to raise this ticket: According to [OSLCCORE-23](https://issues.oasis-open.org/browse/OSLCCORE-23 "Should a ServiceProviderCatalog or ServiceProvider resource be an LDPC?"), we agreed that "Service is the point at which Domain specifications specify their specific service capabilities." - which suggests the tie between domains and Service resources. Of course, it doesn‘t have to be an OSLC-defined domain, but in my opinion it must be a value of the oslc:domain property on the Service. And that it is unreasonable to expect clients to be able to work with two Service resources with the same oslc:domain value, unless explicitly permitted by that domain. So I think we have a question to answer, which probably requires its own ticket: Do we keep the one-to-one relationship between Service resources and oslc:domain values from v2 (within the context of a single SP, and if that understanding of v2 is correct), or do we redefine it and suggest/require that a Service is one-to-one with an LDPC (if not exactly the same resource) in the server‘s desired organisation of containers? The benefit of the former is that clients have fewer options to present to users, and that v3 servers are more likely to work with v2 clients (although that could do with verification). The benefit of the latter is that the server‘s organisation of containers is exposed in the OSLC data, but this comes at the cost of complexity for the clients. --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-53 (opened by @oslc-bot; previously assigned to @jamsden)_
1.0
Does a ServiceProvider have one Service per oslc:domain value, or does it reflect the server&apos;s desired container structure. - In [OSLCCORE-23](https://issues.oasis-open.org/browse/OSLCCORE-23 "Should a ServiceProviderCatalog or ServiceProvider resource be an LDPC?") we agreed that a Service resource should be an LDPC, but didn‘t state what its members were. My reading of that ticket‘s proposal is that the members are defined by the domain. In recent emails, there has been a discussion about what the members of a Service should be. My latest comments that triggered me to raise this ticket: According to [OSLCCORE-23](https://issues.oasis-open.org/browse/OSLCCORE-23 "Should a ServiceProviderCatalog or ServiceProvider resource be an LDPC?"), we agreed that "Service is the point at which Domain specifications specify their specific service capabilities." - which suggests the tie between domains and Service resources. Of course, it doesn‘t have to be an OSLC-defined domain, but in my opinion it must be a value of the oslc:domain property on the Service. And that it is unreasonable to expect clients to be able to work with two Service resources with the same oslc:domain value, unless explicitly permitted by that domain. So I think we have a question to answer, which probably requires its own ticket: Do we keep the one-to-one relationship between Service resources and oslc:domain values from v2 (within the context of a single SP, and if that understanding of v2 is correct), or do we redefine it and suggest/require that a Service is one-to-one with an LDPC (if not exactly the same resource) in the server‘s desired organisation of containers? The benefit of the former is that clients have fewer options to present to users, and that v3 servers are more likely to work with v2 clients (although that could do with verification). The benefit of the latter is that the server‘s organisation of containers is exposed in the OSLC data, but this comes at the cost of complexity for the clients. --- _Migrated from https://issues.oasis-open.org/browse/OSLCCORE-53 (opened by @oslc-bot; previously assigned to @jamsden)_
priority
does a serviceprovider have one service per oslc domain value or does it reflect the server apos s desired container structure in should a serviceprovidercatalog or serviceprovider resource be an ldpc we agreed that a service resource should be an ldpc but didn‘t state what its members were my reading of that ticket‘s proposal is that the members are defined by the domain in recent emails there has been a discussion about what the members of a service should be my latest comments that triggered me to raise this ticket according to should a serviceprovidercatalog or serviceprovider resource be an ldpc we agreed that service is the point at which domain specifications specify their specific service capabilities which suggests the tie between domains and service resources of course it doesn‘t have to be an oslc defined domain but in my opinion it must be a value of the oslc domain property on the service and that it is unreasonable to expect clients to be able to work with two service resources with the same oslc domain value unless explicitly permitted by that domain so i think we have a question to answer which probably requires its own ticket do we keep the one to one relationship between service resources and oslc domain values from within the context of a single sp and if that understanding of is correct or do we redefine it and suggest require that a service is one to one with an ldpc if not exactly the same resource in the server‘s desired organisation of containers the benefit of the former is that clients have fewer options to present to users and that servers are more likely to work with clients although that could do with verification the benefit of the latter is that the server‘s organisation of containers is exposed in the oslc data but this comes at the cost of complexity for the clients migrated from opened by oslc bot previously assigned to jamsden
1
611,380
18,953,682,498
IssuesEvent
2021-11-18 17:38:54
nash-io/openlimits
https://api.github.com/repos/nash-io/openlimits
closed
Separate supported exchanges by features
Type: Enhancement Priority: High
It would be good to separate the exchanges by features. This would help if someone using the library don't want the Binance support, for example. ```toml # Cargo.toml [features] binance = [] coinbase = [] ``` Then lib.rs would be something like ```rust // --snip-- #[cfg(feature = "binance")] pub mod binance; #[cfg(feature = "coinbase")] pub mod coinbase; ```
1.0
Separate supported exchanges by features - It would be good to separate the exchanges by features. This would help if someone using the library don't want the Binance support, for example. ```toml # Cargo.toml [features] binance = [] coinbase = [] ``` Then lib.rs would be something like ```rust // --snip-- #[cfg(feature = "binance")] pub mod binance; #[cfg(feature = "coinbase")] pub mod coinbase; ```
priority
separate supported exchanges by features it would be good to separate the exchanges by features this would help if someone using the library don t want the binance support for example toml cargo toml binance coinbase then lib rs would be something like rust snip pub mod binance pub mod coinbase
1
531,770
15,504,786,083
IssuesEvent
2021-03-11 14:40:21
ioos/catalog-ckan
https://api.github.com/repos/ioos/catalog-ckan
closed
CKAN Harvest job backlog and monitoring/alerting options
High Priority
@benjwadams I was trying to understand why Google DSS shows different dataset time coverage for the results in [this search](https://datasetsearch.research.google.com/search?query=site%3A%20data.ioos.us%20UW157-20190916T0000&docid=H8kTGBwNfihXa6vjAAAAAA%3D%3D) for a Glider DAC dataset (these all in fact point to the same source dataset, just from different origins). Both records from the IOOS Catalog (CeNCOOS and Glider DAC) are from Dec 3, whereas the CeNCOOS ERDDAP is more recent (Jan 21). A little troubleshooting led me to see that we have some hung harvest jobs again for both of these sources at least (others as well no doubt): https://data.ioos.us/harvest/admin/cencoos-waf https://data.ioos.us/harvest/admin/glider-dac-waf Can we look at options for monitoring harvest job status in CKAN, and possibly set up email alerts if a harvest job doesn't finish for more than a week? This would give us a minimal, backstop notice when things are really amiss like this, and hopefully would prevent cases like this where we just don't notice for awhile. Of course, any one harvest job should complete much sooner than a week, but a threshold like this should prevent over-alerting.
1.0
CKAN Harvest job backlog and monitoring/alerting options - @benjwadams I was trying to understand why Google DSS shows different dataset time coverage for the results in [this search](https://datasetsearch.research.google.com/search?query=site%3A%20data.ioos.us%20UW157-20190916T0000&docid=H8kTGBwNfihXa6vjAAAAAA%3D%3D) for a Glider DAC dataset (these all in fact point to the same source dataset, just from different origins). Both records from the IOOS Catalog (CeNCOOS and Glider DAC) are from Dec 3, whereas the CeNCOOS ERDDAP is more recent (Jan 21). A little troubleshooting led me to see that we have some hung harvest jobs again for both of these sources at least (others as well no doubt): https://data.ioos.us/harvest/admin/cencoos-waf https://data.ioos.us/harvest/admin/glider-dac-waf Can we look at options for monitoring harvest job status in CKAN, and possibly set up email alerts if a harvest job doesn't finish for more than a week? This would give us a minimal, backstop notice when things are really amiss like this, and hopefully would prevent cases like this where we just don't notice for awhile. Of course, any one harvest job should complete much sooner than a week, but a threshold like this should prevent over-alerting.
priority
ckan harvest job backlog and monitoring alerting options benjwadams i was trying to understand why google dss shows different dataset time coverage for the results in for a glider dac dataset these all in fact point to the same source dataset just from different origins both records from the ioos catalog cencoos and glider dac are from dec whereas the cencoos erddap is more recent jan a little troubleshooting led me to see that we have some hung harvest jobs again for both of these sources at least others as well no doubt can we look at options for monitoring harvest job status in ckan and possibly set up email alerts if a harvest job doesn t finish for more than a week this would give us a minimal backstop notice when things are really amiss like this and hopefully would prevent cases like this where we just don t notice for awhile of course any one harvest job should complete much sooner than a week but a threshold like this should prevent over alerting
1
172,515
6,506,889,914
IssuesEvent
2017-08-24 10:53:33
VirtoCommerce/vc-platform
https://api.github.com/repos/VirtoCommerce/vc-platform
closed
Fix localization on checkout page
bug frontend Priority: High
Should fix (add) localization ![default](https://user-images.githubusercontent.com/7639413/28643085-edff45ae-7254-11e7-9910-ed5077386d0d.png) ### Expected behavior See real text
1.0
Fix localization on checkout page - Should fix (add) localization ![default](https://user-images.githubusercontent.com/7639413/28643085-edff45ae-7254-11e7-9910-ed5077386d0d.png) ### Expected behavior See real text
priority
fix localization on checkout page should fix add localization expected behavior see real text
1
128,598
5,071,871,918
IssuesEvent
2016-12-26 16:59:23
AlbatrossAvionics/Alba-2017
https://api.github.com/repos/AlbatrossAvionics/Alba-2017
opened
Androidアプリの作成
high priority
機体搭載用Androidアプリの作成を行う。mbedとBlueToothの相互通信で接続する。主な機能は以下の通り。 - [ ] Android内の各種センサ値の取得、保存 - [ ] 電装のセンサで取れた値もBlueToothで送ってしまって保存して、SDロガーの代わりにしてSDロガーを廃止したい。 - [ ] パイロットにセンサの値を伝える表示画面として使用する。 - [ ] カメラ機能を使用して、フライト動画を取りたい。 - [ ] ロールアラームの機能もここに統合してしまいたい。音量的にロールアラームを廃止できなかったとしても両方鳴らす。 - [ ] mbedからはString型の文字列を送ってAndroidを操作。 データ保存にはSQLiteを使用する。
1.0
Androidアプリの作成 - 機体搭載用Androidアプリの作成を行う。mbedとBlueToothの相互通信で接続する。主な機能は以下の通り。 - [ ] Android内の各種センサ値の取得、保存 - [ ] 電装のセンサで取れた値もBlueToothで送ってしまって保存して、SDロガーの代わりにしてSDロガーを廃止したい。 - [ ] パイロットにセンサの値を伝える表示画面として使用する。 - [ ] カメラ機能を使用して、フライト動画を取りたい。 - [ ] ロールアラームの機能もここに統合してしまいたい。音量的にロールアラームを廃止できなかったとしても両方鳴らす。 - [ ] mbedからはString型の文字列を送ってAndroidを操作。 データ保存にはSQLiteを使用する。
priority
androidアプリの作成 機体搭載用androidアプリの作成を行う。mbedとbluetoothの相互通信で接続する。主な機能は以下の通り。 android内の各種センサ値の取得、保存 電装のセンサで取れた値もbluetoothで送ってしまって保存して、sdロガーの代わりにしてsdロガーを廃止したい。 パイロットにセンサの値を伝える表示画面として使用する。 カメラ機能を使用して、フライト動画を取りたい。 ロールアラームの機能もここに統合してしまいたい。音量的にロールアラームを廃止できなかったとしても両方鳴らす。 mbedからはstring型の文字列を送ってandroidを操作。 データ保存にはsqliteを使用する。
1
745,267
25,977,177,426
IssuesEvent
2022-12-19 15:44:04
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
tests: kernel: timer: starve: DTS failure stm32f3_seco_d23
bug priority: high area: Devicetree area: Timer platform: STM32 Regression
**Describe the bug** Running CI on PR for zephyr: scripts/twister -p stm32f3_seco_d23 -s zephyr/tests/kernel/timer/starve/kernel.timer.starve Observed for - stm32f3_seco_d23 **To Reproduce** Steps to reproduce the behavior: 1. scripts/twister -p stm32f3_seco_d23 -s zephyr/tests/kernel/timer/starve/kernel.timer.starve **Expected behavior** Valid console output **Impact** Not clear **Logs and console output** ``` ERROR - Cmake build failure: /__w/zephyr/zephyr/tests/kernel/timer/starve for stm32f3_seco_d23 INFO - 56/510 stm32f3_seco_d23 zephyr/tests/kernel/timer/starve/kernel.timer.starve FAILED Cmake build failure (build) INFO - /__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/build.log ERROR - Loading Zephyr default modules (Zephyr base). -- Application: /__w/zephyr/zephyr/tests/kernel/timer/starve -- Found Python3: /usr/bin/python3.8 (found suitable exact version "3.8.10") found components: Interpreter -- Cache files will be written to: /__w/zephyr/zephyr/.cache -- Zephyr version: 3.2.99 (/__w/zephyr/zephyr) -- Found west (found suitable version "0.14.0", minimum required is "0.7.1") -- Board: stm32f3_seco_d23 -- Found host-tools: zephyr 0.15.2 (/opt/toolchains/zephyr-sdk-0.15.2) -- Found toolchain: zephyr 0.15.2 (/opt/toolchains/zephyr-sdk-0.15.2) -- Found Dtc: /opt/toolchains/zephyr-sdk-0.15.2/sysroots/x86_64-pokysdk-linux/usr/bin/dtc (found suitable version "1.6.0", minimum required is "1.4.6") -- Found BOARD.dts: /__w/zephyr/zephyr/boards/arm/stm32f3_seco_d23/stm32f3_seco_d23.dts devicetree error: 'resets' is marked as required in 'properties:' in /__w/zephyr/zephyr/dts/bindings/timer/st,stm32-timers.yaml, but does not appear in <Node /soc/timers@40000800 in '/__w/zephyr/zephyr/misc/empty_file.c'> -- In: /__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr, command: /usr/bin/python3.8;/__w/zephyr/zephyr/scripts/dts/gen_defines.py;--dts;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/zephyr.dts.pre;--dtc-flags;'';--bindings-dirs;/__w/zephyr/zephyr/dts/bindings;--header-out;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/include/generated/devicetree_generated.h.new;--dts-out;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/zephyr.dts.new;--edt-pickle-out;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/edt.pickle;--edtlib-Werror;--vendor-prefixes;/__w/zephyr/zephyr/dts/bindings/vendor-prefixes.txt CMake Error at /__w/zephyr/zephyr/cmake/modules/dts.cmake:231 (message): gen_defines.py failed with return code: 1 Call Stack (most recent call first): /__w/zephyr/zephyr/cmake/modules/zephyr_default.cmake:[108](https://github.com/zephyrproject-rtos/zephyr/actions/runs/3732036217/jobs/6331025248#step:13:109) (include) /__w/zephyr/zephyr/share/zephyr-package/cmake/ZephyrConfig.cmake:66 (include) /__w/zephyr/zephyr/share/zephyr-package/cmake/ZephyrConfig.cmake:92 (include_boilerplate) CMakeLists.txt:4 (find_package) -- Configuring incomplete, errors occurred! ``` **Environment (please complete the following information):** - OS: Ubuntu 20.04.1 LTS - Toolchain Zephyr SDK 0.15.1 - Commit SHA or Version used: 3c9aa927ee55c77153cbac4ffe28cc4f36a73257
1.0
tests: kernel: timer: starve: DTS failure stm32f3_seco_d23 - **Describe the bug** Running CI on PR for zephyr: scripts/twister -p stm32f3_seco_d23 -s zephyr/tests/kernel/timer/starve/kernel.timer.starve Observed for - stm32f3_seco_d23 **To Reproduce** Steps to reproduce the behavior: 1. scripts/twister -p stm32f3_seco_d23 -s zephyr/tests/kernel/timer/starve/kernel.timer.starve **Expected behavior** Valid console output **Impact** Not clear **Logs and console output** ``` ERROR - Cmake build failure: /__w/zephyr/zephyr/tests/kernel/timer/starve for stm32f3_seco_d23 INFO - 56/510 stm32f3_seco_d23 zephyr/tests/kernel/timer/starve/kernel.timer.starve FAILED Cmake build failure (build) INFO - /__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/build.log ERROR - Loading Zephyr default modules (Zephyr base). -- Application: /__w/zephyr/zephyr/tests/kernel/timer/starve -- Found Python3: /usr/bin/python3.8 (found suitable exact version "3.8.10") found components: Interpreter -- Cache files will be written to: /__w/zephyr/zephyr/.cache -- Zephyr version: 3.2.99 (/__w/zephyr/zephyr) -- Found west (found suitable version "0.14.0", minimum required is "0.7.1") -- Board: stm32f3_seco_d23 -- Found host-tools: zephyr 0.15.2 (/opt/toolchains/zephyr-sdk-0.15.2) -- Found toolchain: zephyr 0.15.2 (/opt/toolchains/zephyr-sdk-0.15.2) -- Found Dtc: /opt/toolchains/zephyr-sdk-0.15.2/sysroots/x86_64-pokysdk-linux/usr/bin/dtc (found suitable version "1.6.0", minimum required is "1.4.6") -- Found BOARD.dts: /__w/zephyr/zephyr/boards/arm/stm32f3_seco_d23/stm32f3_seco_d23.dts devicetree error: 'resets' is marked as required in 'properties:' in /__w/zephyr/zephyr/dts/bindings/timer/st,stm32-timers.yaml, but does not appear in <Node /soc/timers@40000800 in '/__w/zephyr/zephyr/misc/empty_file.c'> -- In: /__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr, command: /usr/bin/python3.8;/__w/zephyr/zephyr/scripts/dts/gen_defines.py;--dts;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/zephyr.dts.pre;--dtc-flags;'';--bindings-dirs;/__w/zephyr/zephyr/dts/bindings;--header-out;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/include/generated/devicetree_generated.h.new;--dts-out;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/zephyr.dts.new;--edt-pickle-out;/__w/zephyr/zephyr/twister-out/stm32f3_seco_d23/zephyr/tests/kernel/timer/starve/kernel.timer.starve/zephyr/edt.pickle;--edtlib-Werror;--vendor-prefixes;/__w/zephyr/zephyr/dts/bindings/vendor-prefixes.txt CMake Error at /__w/zephyr/zephyr/cmake/modules/dts.cmake:231 (message): gen_defines.py failed with return code: 1 Call Stack (most recent call first): /__w/zephyr/zephyr/cmake/modules/zephyr_default.cmake:[108](https://github.com/zephyrproject-rtos/zephyr/actions/runs/3732036217/jobs/6331025248#step:13:109) (include) /__w/zephyr/zephyr/share/zephyr-package/cmake/ZephyrConfig.cmake:66 (include) /__w/zephyr/zephyr/share/zephyr-package/cmake/ZephyrConfig.cmake:92 (include_boilerplate) CMakeLists.txt:4 (find_package) -- Configuring incomplete, errors occurred! ``` **Environment (please complete the following information):** - OS: Ubuntu 20.04.1 LTS - Toolchain Zephyr SDK 0.15.1 - Commit SHA or Version used: 3c9aa927ee55c77153cbac4ffe28cc4f36a73257
priority
tests kernel timer starve dts failure seco describe the bug running ci on pr for zephyr scripts twister p seco s zephyr tests kernel timer starve kernel timer starve observed for seco to reproduce steps to reproduce the behavior scripts twister p seco s zephyr tests kernel timer starve kernel timer starve expected behavior valid console output impact not clear logs and console output error cmake build failure w zephyr zephyr tests kernel timer starve for seco info seco zephyr tests kernel timer starve kernel timer starve failed cmake build failure build info w zephyr zephyr twister out seco zephyr tests kernel timer starve kernel timer starve build log error loading zephyr default modules zephyr base application w zephyr zephyr tests kernel timer starve found usr bin found suitable exact version found components interpreter cache files will be written to w zephyr zephyr cache zephyr version w zephyr zephyr found west found suitable version minimum required is board seco found host tools zephyr opt toolchains zephyr sdk found toolchain zephyr opt toolchains zephyr sdk found dtc opt toolchains zephyr sdk sysroots pokysdk linux usr bin dtc found suitable version minimum required is found board dts w zephyr zephyr boards arm seco seco dts devicetree error resets is marked as required in properties in w zephyr zephyr dts bindings timer st timers yaml but does not appear in in w zephyr zephyr twister out seco zephyr tests kernel timer starve kernel timer starve zephyr command usr bin w zephyr zephyr scripts dts gen defines py dts w zephyr zephyr twister out seco zephyr tests kernel timer starve kernel timer starve zephyr zephyr dts pre dtc flags bindings dirs w zephyr zephyr dts bindings header out w zephyr zephyr twister out seco zephyr tests kernel timer starve kernel timer starve zephyr include generated devicetree generated h new dts out w zephyr zephyr twister out seco zephyr tests kernel timer starve kernel timer starve zephyr zephyr dts new edt pickle out w zephyr zephyr twister out seco zephyr tests kernel timer starve kernel timer starve zephyr edt pickle edtlib werror vendor prefixes w zephyr zephyr dts bindings vendor prefixes txt cmake error at w zephyr zephyr cmake modules dts cmake message gen defines py failed with return code call stack most recent call first w zephyr zephyr cmake modules zephyr default cmake include w zephyr zephyr share zephyr package cmake zephyrconfig cmake include w zephyr zephyr share zephyr package cmake zephyrconfig cmake include boilerplate cmakelists txt find package configuring incomplete errors occurred environment please complete the following information os ubuntu lts toolchain zephyr sdk commit sha or version used
1
404,358
11,855,985,563
IssuesEvent
2020-03-25 06:12:53
input-output-hk/cardano-db-sync
https://api.github.com/repos/input-output-hk/cardano-db-sync
closed
cardano-db-sync-extended: epochs with 0 transactions not present in table
bug priority high
See https://github.com/input-output-hk/cardano-graphql/issues/126 for a query result aimed at `testnet`
1.0
cardano-db-sync-extended: epochs with 0 transactions not present in table - See https://github.com/input-output-hk/cardano-graphql/issues/126 for a query result aimed at `testnet`
priority
cardano db sync extended epochs with transactions not present in table see for a query result aimed at testnet
1
134,785
5,234,287,307
IssuesEvent
2017-01-30 15:18:53
Angblah/The-Comparator
https://api.github.com/repos/Angblah/The-Comparator
closed
Login Functionality
Priority: High Stack: Backend Status: Review Needed Type: Feature
As a registered user, I want to login so I can access account-related storage, such as saving templates
1.0
Login Functionality - As a registered user, I want to login so I can access account-related storage, such as saving templates
priority
login functionality as a registered user i want to login so i can access account related storage such as saving templates
1
31,607
2,734,565,111
IssuesEvent
2015-04-17 20:44:58
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
opened
sct_dmri_separate_b0_and_dwi does not work with parameter -m
bug priority: high
data used : /Volumes/folder_shared/sct_issues/20150313_allan command: sct_dmri_separate_b0_and_dwi -i dmrir.nii.gz -b bvecs.txt -m bvals.txt Return: Get dimensions data... .. 128 x 128 x 12 x 30 Identify b=0 and DWI images... Traceback (most recent call last): File "/Users/tamag/spinalcordtoolbox/bin/sct_dmri_separate_b0_and_dwi", line 301, in <module> main() File "/Users/tamag/spinalcordtoolbox/bin/sct_dmri_separate_b0_and_dwi", line 131, in main index_b0, index_dwi, nb_b0, nb_dwi = identify_b0(fname_bvecs, fname_bvals, param.bval_min, verbose) File "/Users/tamag/spinalcordtoolbox/bin/sct_dmri_separate_b0_and_dwi", line 232, in identify_b0 with open(fname_bvals) as f: IOError: [Errno 2] No such file or directory: 'bvals.txt'
1.0
sct_dmri_separate_b0_and_dwi does not work with parameter -m - data used : /Volumes/folder_shared/sct_issues/20150313_allan command: sct_dmri_separate_b0_and_dwi -i dmrir.nii.gz -b bvecs.txt -m bvals.txt Return: Get dimensions data... .. 128 x 128 x 12 x 30 Identify b=0 and DWI images... Traceback (most recent call last): File "/Users/tamag/spinalcordtoolbox/bin/sct_dmri_separate_b0_and_dwi", line 301, in <module> main() File "/Users/tamag/spinalcordtoolbox/bin/sct_dmri_separate_b0_and_dwi", line 131, in main index_b0, index_dwi, nb_b0, nb_dwi = identify_b0(fname_bvecs, fname_bvals, param.bval_min, verbose) File "/Users/tamag/spinalcordtoolbox/bin/sct_dmri_separate_b0_and_dwi", line 232, in identify_b0 with open(fname_bvals) as f: IOError: [Errno 2] No such file or directory: 'bvals.txt'
priority
sct dmri separate and dwi does not work with parameter m data used volumes folder shared sct issues allan command sct dmri separate and dwi i dmrir nii gz b bvecs txt m bvals txt return get dimensions data x x x identify b and dwi images traceback most recent call last file users tamag spinalcordtoolbox bin sct dmri separate and dwi line in main file users tamag spinalcordtoolbox bin sct dmri separate and dwi line in main index index dwi nb nb dwi identify fname bvecs fname bvals param bval min verbose file users tamag spinalcordtoolbox bin sct dmri separate and dwi line in identify with open fname bvals as f ioerror no such file or directory bvals txt
1